EXL3 quantization of Mistral-Small-Instruct-2409, 4 bits per weight.

HumanEval (argmax)

Model Q4 Q6 Q8 FP16
Mistral-Small-Instruct-2409-exl3-3bpw 76.8 74.4 76.2 75.6
Mistral-Small-Instruct-2409-exl3-3.5bpw 73.8 75.6 75.0 75.6
Mistral-Small-Instruct-2409-exl3-4bpw 78.7 78.7 79.3 79.3
Mistral-Nemo-Instruct-2407-exl3-4bpw 74.4 72.6 73.2 72.0
Mistral-Nemo-Instruct-2407-exl3-6bpw 70.7 69.5 69.5 68.9
Downloads last month
4
Safetensors
Model size
5.74B params
Tensor type
F16
·
I16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for isogen/Mistral-Small-Instruct-2409-exl3-4bpw

Quantized
(47)
this model