Model details

This is Qwen/Qwen2.5-14B-Instruct quantized with AutoRound (symmetric quantization) and serialized with the GPTQ format in 4-bit.

Downloads last month
0
Safetensors
Model size
3.33B params
Tensor type
I32
·
FP16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including Siddharth63/Qwen2.5-14B-Instruct-AutoRound-GPTQ-4bit