8-bit Quantization of the Qwen3 30B A3B Model

Quantized using GPTQModel

quantiziation config:

quant_config = QuantizeConfig(
    bits=8,
    group_size=32,
    sym=True,
    desc_act=False,
    true_sequential=True,
    pack_dtype=torch.int32,
    damp_percent=0.1       
)
Downloads last month
301
Safetensors
Model size
9.3B params
Tensor type
FP16
·
I32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for btbtyler09/Qwen3-30B-A3B-gptq-8bit

Finetuned
Qwen/Qwen3-30B-A3B
Quantized
(71)
this model