Text Generation
Safetensors
English
llama

OctoThinker: Mid-training Incentivizes Reinforcement Learning Scaling

OctoThinker-1B-Short-Base

The OctoThinker family is built on carefully studied mid-training insights, starting from the Llama-3 family, to create a reinforcement learning–friendly base language model.

Training Recipe

Data Pipeline

Evaluation Results

Note that we adopt the few-shot prompting evaluation for these base language models.

Data Pipeline

More about OctoThinker

Data Pipeline

Citation

Check out our paper for more details. If you use our models, datasets or find our work useful, please cite

@article{wang2025octothinker,
  title={OctoThinker: Mid-training Incentivizes Reinforcement Learning Scaling},
  author={Wang, Zengzhi and Zhou, Fan and Li, Xuefeng and Liu, Pengfei},
  year={2025},
  journal={arXiv preprint arXiv:2506.20512},
  note={Preprint}
}
Downloads last month
10
Safetensors
Model size
1.24B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for OctoThinker/OctoThinker-1B-Short-Base

Finetuned
(454)
this model

Datasets used to train OctoThinker/OctoThinker-1B-Short-Base

Collection including OctoThinker/OctoThinker-1B-Short-Base