Zheng Han
traphix
ยท
AI & ML interests
None yet
Recent Activity
new activity
2 days ago
Qwen/Qwen3-30B-A3B-FP8:Remove vLLM FP8 Limitation
new activity
5 days ago
RedHatAI/Qwen3-235B-A22B-FP8-dynamic:Error running on A100?
new activity
16 days ago
RedHatAI/Qwen3-235B-A22B-FP8-dynamic:Any plans for int8 quantized.w8a8?
Organizations
None yet
traphix's activity
Remove vLLM FP8 Limitation
9
#2 opened 26 days ago
by
simon-mo
Error running on A100?
2
#4 opened 16 days ago
by
traphix
Any plans for int8 quantized.w8a8?
#5 opened 16 days ago
by
traphix
How about int8 quantization?
#3 opened 16 days ago
by
traphix
How many RAM in GBs when quantizing Qwen3-235B-A22B?
#2 opened 19 days ago
by
traphix
Where are the safetensors?
๐
1
1
#1 opened 19 days ago
by
traphix
What is the difference between Qwen/Qwen3-32B-FP8 and this quatinized model๏ผ
4
#1 opened 22 days ago
by
traphix
Does vllm 0.8.4 support this quantized model?
1
#1 opened about 1 month ago
by
traphix
This model beats Qwen Max!
๐
1
7
#33 opened 3 months ago
by
MrDevolver

why "MLA is not supported with awq_marlin quantization. Disabling MLA." with 4090 * 32 (4 node / vllm 0.7.2)
๐
1
3
#14 opened 3 months ago
by
FightLLM
Is there any accuracy results comparing to original DeepSeek-R1๏ผ
2
#15 opened 3 months ago
by
traphix
Has anyone evaluated the performance of the AWQ version of the model on benchmarks?
4
#8 opened 3 months ago
by
liuqianchao
skips the thinking process
11
#5 opened 4 months ago
by
muzizon
Deployment framework
27
#2 opened 4 months ago
by
xro7
vllm support a100
17
#2 opened 4 months ago
by
HuggingLianWang
Can it run on A100/A800 with VLLM?
3
#1 opened 10 months ago
by
Parkerlambert123