4-bit Quantization of the Qwen3 30B A3B Instruct 2507 Model

Quantized using GPTQModel

quantiziation config:

quant_config = QuantizeConfig(
    bits=4,
    group_size=32,
    sym=True,
    desc_act=False,
    true_sequential=True,
    pack_dtype=torch.int32,
    damp_percent=0.1       
)
Downloads last month
66
Safetensors
Model size
5.44B params
Tensor type
I32
·
F16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for btbtyler09/Qwen3-30B-A3B-Instruct-2507-gptq-4bit

Quantized
(67)
this model