Trained from anthracite-core/Mistral-Small-3.2-24B-Instruct-2506-ChatML for convenience (no Pixtral compatibility needed). No vision adapter currently.
Test model trained at 16k context on 50M tokens of long-form human writing (mostly books).
Haven't tested yet but regular Tekken v7 instruct will work and samplers are probably the same as you'd use for 3.2 Instruct.
- Downloads last month
- 16
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for ToastyPigeon/ms3.2-24b-longform
Base model
mistralai/Mistral-Small-3.1-24B-Base-2503