All Qwen3 uploads updated with fixed chat templates!

#14
by shimmyshimmer - opened
Unsloth AI org

Thanks to a r/localllama user, they found out there was a bug in Qwen's chat template so, we've now updated all the Qwen3 GGUFs and safetensors with the fixed chat template! Should improve results slightly.

Original post for Qwen3 chat template bug findings: https://www.reddit.com/r/LocalLLaMA/comments/1klltt4/the_qwen3_chat_template_is_still_bugged/

shimmyshimmer pinned discussion

Not directly related to the issue, but just if somebody experience issues with jinja/ollama templates there is online tool to quickly render the template:
https://eugene-kamenev.github.io/ollama-template-test

Unsloth AI org

For a full list of changes:

  1. Updated quants due to chat template not working in llama.cpp / lm studio due to [::-1] and other jinja template issues - now worked for llama.cpp
  2. Updated again since lm studio didn't like llama.cpp's chat template - will work with lm studio in the future to test templates
  3. Updated with an updated dynamic 2.0 quant methodology (2.1) upgrading our dataset to over 1 million tokens with both short and long context lengths to improve accuracy. Also fixed 235B imatrix quants - in fact we're the only provider for imatrix 235B quants.
  4. Updated again due to tool calling issues as mentioned in https://www.reddit.com/r/LocalLLaMA/comments/1klltt4/the_qwen3_chat_template_is_still_bugged/ - other people's quants I think are still buggy
  5. Updated all quants due to speculative decoding not working (BOS tokens mismatched)
  6. Should now be fully stable

Sign up or log in to comment