EXL3 quantization of gemma-3-4b-it, 8 bits per weight, including output layers.

HumanEval (argmax)

Model Q4 Q6 Q8 FP16
gemma-3-4b-it-exl3-8bpw-h8 72.0 73.2 71.3 70.1
Downloads last month
4
Safetensors
Model size
3B params
Tensor type
BF16
F16
I16
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support

Model tree for isogen/gemma-3-4b-it-exl3-8bpw-h8

Quantized
(156)
this model