Request to Release the Base Model for Qwen3-32B
1
#25 opened about 13 hours ago
by
eramax

How to control thinking length?
#24 opened about 17 hours ago
by
lidh15
Qwen3 does not deploy on Endpoints
#23 opened 1 day ago
by
zenfiric

The model's instructions follow too poorly
1
3
#22 opened 7 days ago
by
xldistance
Update README.md
#21 opened 7 days ago
by
Logical-Transcendence84

please release AWQ version
#20 opened 8 days ago
by
classdemo
Collections of Bad Cases User Reviews and Comments of Qwen3 32B model
#19 opened 12 days ago
by
DeepNLP
Potential issue with large context sizes - can someone confirm?
13
#18 opened 14 days ago
by
Thireus
Qwen 3 presence of tools affect output length?
#17 opened 14 days ago
by
evetsagg
"/no_think" control is unstable
1
#16 opened 14 days ago
by
Smorty100
LICENSE files missing
1
#14 opened 14 days ago
by
johndoe2001
After setting /nothinking or enable_thinking=False, can the empty <thinking> tag be omitted from the response?
3
2
#13 opened 15 days ago
by
pteromyini

Feedback: It's a good model, however it hallucinates very badly at local facts (Germany)
9
2
#12 opened 15 days ago
by
Dampfinchen
The correct way of fine-tuning on multi-turn trajectories
5
1
#11 opened 15 days ago
by
hr0nix
Providing a GPTQ version
3
12
#10 opened 15 days ago
by
blueteamqq1
how to set, enable_thinking=False, on ollama
6
2
#9 opened 15 days ago
by
TatsuhiroC
🚀[Fine-tuning] Implementation and Best Practices for Qwen3 CPT/SFT/DPO/GRPO Training👋
3
#7 opened 15 days ago
by
study-hjt

Reasoning or Non-reasoning model?
4
#6 opened 15 days ago
by
dipta007

Local Installation Video and Testing - Step by Step
#5 opened 15 days ago
by
fahdmirzac

【Evaluation】Best practice for evaluating Qwen3 !!
5
#4 opened 15 days ago
by
wangxingjun778

Base Model?
4
8
#3 opened 15 days ago
by
Downtown-Case
Is this multimodal?
1
#2 opened 15 days ago
by
pbarker

Add languages tag
2
#1 opened 15 days ago
by
de-francophones
