EXL3 quantization of II-Search-CIR-4B, 8 bits per weight, including output layers.

HumanEval (argmax)

Model Q4 Q6 Q8 FP16
II-Search-CIR-4B-exl3-4bpw 81.7 79.3 78.7 79.9
II-Search-CIR-4B-exl3-6bpw 80.5 81.1 81.1 81.7
II-Search-CIR-4B-exl3-8bpw-h8 83.5 83.5 82.3 82.9
Qwen3-4B-exl3-4bpw 80.5 81.1 81.7 80.5
Qwen3-4B-exl3-6bpw 80.5 85.4 86.0 86.0
Qwen3-4B-exl3-8bpw-h8 82.3 84.8 83.5 82.9
Downloads last month
4
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for isogen/II-Search-CIR-4B-exl3-8bpw-h8

Quantized
(2)
this model