EXL3 quantization of Jan-v1-4B, 4 bits per weight.

HumanEval (argmax)

Model Q4 Q6 Q8 FP16
Jan-v1-4B-exl3-4bpw 82.3 79.3 78.0 78.0
Jan-v1-4B-exl3-6bpw 78.0 76.8 77.4 76.8
Jan-v1-4B-exl3-8bpw-h8 79.9 78.7 78.0 77.4
Downloads last month
4
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for isogen/Jan-v1-4B-exl3-4bpw

Finetuned
janhq/Jan-v1-4B
Quantized
(46)
this model