EXL3 quantization of Jan-v1-4B, 8 bits per weight, including output layers.

HumanEval (argmax)

Model Q4 Q6 Q8 FP16
Jan-v1-4B-exl3-4bpw 82.3 79.3 78.0 78.0
Jan-v1-4B-exl3-6bpw 78.0 76.8 77.4 76.8
Jan-v1-4B-exl3-8bpw-h8 79.9 78.7 78.0 77.4
Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for isogen/Jan-v1-4B-exl3-8bpw-h8

Finetuned
janhq/Jan-v1-4B
Quantized
(32)
this model