YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

CoRT-Prompt-Hint-1.5B-RL

Model Description

CoRT-Prompt-Hint-1.5B-RL is a 1.5B parameter model trained using the CoRT (Code-integrated Reasoning within Thinking) framework. This model specializes in mathematical reasoning by effectively integrating natural language reasoning with Python code execution.

This model uses the Prompt-Hint approach, which strategically inserts hints at the beginning of the reasoning process to encourage code usage throughout problem-solving.

Key Features

  • High Performance: Achieves 58.3% average accuracy across mathematical reasoning benchmarks
  • Code Integration: Seamlessly combines natural language reasoning with Python code execution
  • Multi-turn Tool-Integrated Reasoning: Supports interactive code execution within reasoning chains
  • Optimized for Mathematics: Specifically trained on mathematical problem-solving tasks

Performance

Benchmark Accuracy
AIME24 43.1%
AIME25 30.2%
AMC23 73.8%
MATH500 87.3%
Olympiad 57.1%
Average 58.3%

Model Architecture

  • Base Model: DeepSeek-R1-Distill-Qwen-1.5B
  • Training Method: Supervised Fine-tuning (SFT) → Reinforcement Learning (RL)
  • Framework: CoRT (Code-integrated Reasoning within Thinking)

Usage

⚠️ Important: This model requires multi-turn tool-integrated reasoning capabilities. Please use our specialized inference script from the CoRT GitHub repository for optimal performance.

Installation

First, clone and install the CoRT repository:

git clone https://github.com/ChengpengLi1003/CoRT.git
cd CoRT
# Follow installation instructions in the repository

Inference

TOKENIZERS_PARALLELISM=false VLLM_USE_V1=1 python -m infer.inference_vllm_dp_mj \
    --input_file <path_to_input_file_in_jsonl> \
    --start 0 \
    --end 0 \
    --output_dir <path_to_output_dir> \
    --model_name_or_path <path_to_this_model> \
    --engine vllm \
    --temperature 0.6 \
    --top_p 0.95 \
    --n_sampling 16 \
    --stop_tokens_mode normal_code_block_end \
    --max_tokens_per_call 32768 \
    --max_model_len 32768 \
    --max_func_call 15 \
    --func_call_mode jupyter \
    --data_parallel_size 1 \
    --tensor_parallel_size 1

Input Format

The input should be a JSONL file where each line contains a JSON object with a prompt field:

{
    "prompt": "Every morning Aya goes for a $9$-kilometer-long walk and stops at a coffee shop afterwards. When she walks at a constant speed of $s$ kilometers per hour, the walk takes her 4 hours, including $t$ minutes spent in the coffee shop. When she walks $s+2$ kilometers per hour, the walk takes her 2 hours and 24 minutes, including $t$ minutes spent in the coffee shop. Suppose Aya walks at $s+\\frac{1}{2}$ kilometers per hour. Find the number of minutes the walk takes her, including the $t$ minutes spent in the coffee shop.\nPlease integrate natural language reasoning with python programs to solve the problem above, and put your final answer within \\boxed{}."
}

Related Resources

Citation

If you find our work useful for your research, please cite our paper:

@misc{li2025cortcodeintegratedreasoningthinking,
      title={CoRT: Code-integrated Reasoning within Thinking}, 
      author={Chengpeng Li and Zhengyang Tang and Ziniu Li and Mingfeng Xue and Keqin Bao and Tian Ding and Ruoyu Sun and Benyou Wang and Xiang Wang and Junyang Lin and Dayiheng Liu},
      year={2025},
      eprint={2506.09820},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2506.09820}, 
}
Downloads last month
4
Safetensors
Model size
1.78B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for theshyustc/CoRT-Prompt-Hint-1.5B-RL

Quantizations
1 model