• Extracted a 64 Rank Lora from DeepSeek-R1-Distill-Qwen-32B
  • Merged & Quantized into Q4_K_M

Note: The model seems to be somewhat working with the R1's weird template too but it repeats random Chinese characters and the quality seems to be consistently worse.

Maybe try using the R1 tokenizer.

Downloads last month
20
GGUF
Model size
32.8B params
Architecture
qwen2

4-bit

Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model's library.

Model tree for Ba2han/qwen-coder-thinker-q4_k_m

Quantized
(71)
this model