This is an HQQ all 8-bit (group-size=64) quantized Qwen3-32B model.

Downloads last month
262
Safetensors
Model size
33.7B params
Tensor type
I64
·
FP16
·
U8
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for ArtusDev/Qwen3-32b_8bitgs64_hqq

Base model

Qwen/Qwen3-32B
Quantized
(88)
this model

Collection including ArtusDev/Qwen3-32b_8bitgs64_hqq