EXL3 quantization of gemma-3-4b-it, 8 bits per weight, including output layers.

HumanEval (argmax)

Model Q4 Q6 Q8 FP16
gemma-3-4b-it-exl3-8bpw-h8 72.0 73.2 71.3 70.1
Downloads last month
12
Safetensors
Model size
3.03B params
Tensor type
BF16
FP16
I16
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support

Model tree for isogen/gemma-3-4b-it-exl3-8bpw-h8

Quantized
(96)
this model