EXL3 quantization of gemma-3-4b-it, 8 bits per weight, including output layers.

HumanEval (argmax)

Model Q4 Q6 Q8 FP16
gemma-3-4b-it-exl3-8bpw-h8 72.0 73.2 71.3 70.1
Downloads last month
14
Safetensors
Model size
3.03B params
Tensor type
BF16
F16
I16
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support

Model tree for isogen/gemma-3-4b-it-exl3-8bpw-h8

Quantized
(97)
this model