Model Overview
We present Visual Game Learning (ViGaL), a novel post-training paradigm where multimodal large language models (MLLMs) develop out-of-domain generalization of multimodal reasoning through playing arcade-like games.
ViGaL-7B demonstrates that training a 7B-parameter MLLM via reinforcement learning on simple arcade-like games like Snake significantly enhances its downstream performance on multimodal math benchmarks like MathVista, and on multi-discipline questions like MMMU, without seeing any worked solutions, equations, or diagrams during RL, suggesting the capture of transferable reasoning skills.
Resources
For details of our approach and performance comparison, please see our paper.
For details of training and evaluation, please see our code repo.
| π Project Page | π Paper | π GitHub | π€ Training Data (Coming Soon) |
Citation
If you feel this model useful, please give us a free cite:
@article{xie2025play,
title = {Play to Generalize: Learning to Reason Through Game Play},
author = {Xie, Yunfei and Ma, Yinsong and Lan, Shiyi and Yuille, Alan and Xiao, Junfei and Wei, Chen},
journal = {arXiv preprint arXiv:2506.08011},
year = {2025},
}
- Downloads last month
- 7