This model is a quantized version of QwQ3-32B converted to AWQ Q4 format using the mlx library for efficient inference. It retains the core capabilities of QwQ 32b while optimizing for resource constraints.
Chat template
Files info
Base model