SFT or RL? An Early Investigation into Training R1-Like Reasoning Large Vision-Language Models
This model, VLAA-Thinker-Qwen2VL-7B, is a vision-language model fine-tuned on the VLAA-Thinking dataset. As described in , it leverages a combination of supervised fine-tuning (SFT) and reinforcement learning (RL) to improve reasoning capabilities in LLMs. The model excels in multimodal reasoning tasks, achieving state-of-the-art performance on the OpenCompass Multimodal Reasoning Leaderboard as of April 7th, 2025.
π Project Page
β’
Arxiv
β’ π» Code
Both VLAA-Thinker-Qwen2.5-3B and VLAA-Thinker-Qwen2.5-7B achieve SOTA performance on OpenCompass Multimodal Reasoning Leaderboard as of April 7th, 2025.

Quick Start π
Inference
Run python inference.py
. Note that our model is trained with a system prompt. Please ensure that it is included for inference.
Dataset Download
Run bash ./utils/download_dataset.sh
. Specify the dataset root with absolute path. The dataset should be ordered as follows:
βββ VLAA-Thinking-SFT-126K.json
βββ VLAA-Thinking-GRPO-25K.json
βββ images
βββ allava_laion
βββ arxivqa
βββ chartqa
βββ clevr_math
βββ coco
β βββ train2017
βββ docvqa
βββ geoqa170k
βββ synthesis
βββ vg
β βββ VG_100K
β βββ VG_100K_2
βββ vizwiz
Training
Code coming soon!
(Rest of the README content can be kept as is)
- Downloads last month
- 8