infering by multi-model session but get wrong output
#8
by
enlei
- opened
When you use this model in LM Studio - you need to use the included ChatML preset.
Then in Settings (Right hand side chat screen) Go to -> Model Initialization -> Flash Attention -> Turn it on
jklj077
changed discussion status to
closed