Model Card for radm/Qwen2.5-32B-simpo-FP8

Model Details

Improved quality on hard tasks by 25 percent relative to the base model Qwen2.5-32B-Instruct. Improved multilingual support.

Fine-tuning on A100 in 4-bit with unsloth using SIMPO and internal dataset

Eval results

Eval results on ZebraLogic

image/png

Downloads last month
19
Safetensors
Model size
32.8B params
Tensor type
BF16
·
F8_E4M3
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for radm/Qwen2.5-32B-simpo-FP8

Base model

Qwen/Qwen2.5-32B
Finetuned
(162)
this model