This is the MXFP4 PTQ version of huizimao/gpt-oss-120b-uncensored-bf16.

Finetuning is done on Amazon FalseReject train set with 800 samples.

PTQ is done with NVIDIA ModelOpt

Evaluation results obtained on Amazon FalseReject test set with 300 samples.

Model Variants False refusal rate
gpt-oss-120b original (MXFP4) 70%
LoRA (BF16) 6%
LoRA + PTQ (MXFP4) - this model 24%

Code example, documentation, and further QAT checkpoints will be released soon.

Downloads last month
16
Safetensors
Model size
117B params
Tensor type
BF16
·
U8
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for huizimao/gpt-oss-120b-uncensored-mxfp4

Quantized
(31)
this model
Quantizations
3 models