Qwen2.5-Math-7B-RoPE-300k

This model is a variant of Qwen/Qwen2.5-Math-7B-Instruct, with its RoPE base frequency raised from 10k to 300k, enabling the context length to expand from 4k to 32k tokens.

Citation

If you find this model useful, please cite the original Qwen2.5-Math paper:

@article{yang2024qwen2,
  title={Qwen2. 5-math technical report: Toward mathematical expert model via self-improvement},
  author={Yang, An and Zhang, Beichen and Hui, Binyuan and Gao, Bofei and Yu, Bowen and Li, Chengpeng and Liu, Dayiheng and Tu, Jianhong and Zhou, Jingren and Lin, Junyang and others},
  journal={arXiv preprint arXiv:2409.12122},
  year={2024}
}
Downloads last month
7
Safetensors
Model size
7.62B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for QizhiPei/Qwen2.5-Math-7B-Instruct-RoPE-300k

Base model

Qwen/Qwen2.5-7B
Finetuned
(900)
this model
Finetunes
1 model

Collection including QizhiPei/Qwen2.5-Math-7B-Instruct-RoPE-300k