Model Card for Model ID

This model is a quantized version of QwQ3-32B converted to AWQ Q4 format using the mlx library for efficient inference. It retains the core capabilities of QwQ 32b while optimizing for resource constraints.

Downloads last month
81
Safetensors
Model size
5.17B params
Tensor type
BF16
·
U32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Goraint/QwQ-32b-MLX-AWQ-4bit

Base model

Qwen/Qwen2.5-32B
Finetuned
Qwen/QwQ-32B
Finetuned
(74)
this model