The base Qwen2.5-Math-7B model used by HAPO. We extend the context window to 32k.

Citation

If you find our model, data, or evaluation code useful, please kindly cite our paper:

@misc{liu2025uniformheterogeneoustailoringpolicy,
      title={From Uniform to Heterogeneous: Tailoring Policy Optimization to Every Token's Nature}, 
      author={Zheng Liu and Mengjie Liu and Siwei Wen and Mengzhang Cai and Bin Cui and Conghui He and Wentao Zhang},
      year={2025},
      eprint={2509.16591},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2509.16591}, 
}
Downloads last month
8
Safetensors
Model size
7.62B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Collection including starriver030515/Qwen2.5-Math-7B-32k