EXL3 quantization of gemma-3-4b-it, 8 bits per weight, including output layers.

HumanEval (argmax)

Model Q4 Q6 Q8 FP16
gemma-3-4b-it-exl3-8bpw-h8 72.0 73.2 71.3 70.1
Downloads last month
9
Safetensors
Model size
3.03B params
Tensor type
BF16
·
F16
·
I16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for isogen/gemma-3-4b-it-exl3-8bpw-h8

Quantized
(131)
this model