Trained from anthracite-core/Mistral-Small-3.2-24B-Instruct-2506-ChatML for convenience (no Pixtral compatibility needed). No vision adapter currently.

Test model trained at 16k context on 50M tokens of long-form human writing (mostly books).

Haven't tested yet but regular Tekken v7 instruct will work and samplers are probably the same as you'd use for 3.2 Instruct.

Downloads last month
16
Safetensors
Model size
23.6B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for ToastyPigeon/ms3.2-24b-longform