Code ability is not as good as qwq32b, otherwise it's fine
Not a bad model.
Not a bad model.
how about cogito-v1-preview-qwen-32B-GGUF? I switched to it because its thinking time is muuuch less than qwq32b. And i don't see it performing worse.
Not a bad model.
how about cogito-v1-preview-qwen-32B-GGUF? I switched to it because its thinking time is muuuch less than qwq32b. And i don't see it performing worse.
Ironically, if you have a weaker hardware and have to use lower quants (Q2_K in particular), it's better to use QwQ-32B because its longer CoT will make up for the loss of intelligence due to quantization. Same quant of Cogito will probably struggle with the same tasks and likely fail. I'm speaking from my own experience here. BUT to be completely honest and fair towards Cogito, its 14B version is a little beast and I absolutely love it. That one I can run in Q8_0, it's not the fastest inference in this high quant, but the increased quality of the output makes up for the lower inference speed.
Not a bad model.
how about cogito-v1-preview-qwen-32B-GGUF? I switched to it because its thinking time is muuuch less than qwq32b. And i don't see it performing worse.
qwq32b better