Better TAMA Models with Limited Data
In [1], we reveal that with limited instruction tuning data, we can achieve competitive performance on table tasks. This compact setup enables quick instruction tuning with advanced base models.
We present TAMA models built on Qwen 2.5 and Qwen 3. These models achieve strong results on the MMTU benchmark [2], outperforming recent table reasoning models [3] and competitive table LLMs like Table-GPT 2 [4], which is tuned on 2.36M datapoints.
Notably, TAMA-QWen3 achieves the best overall performance of 33.9, surpassing QWen-3-8B (32.9) and TableGPT-2 (30.0).
Evaluation Details
We adopt the official MMTU evaluation script to compute scores. For overall performance, we use the evaluation function described here. Category scores are the arithmetic mean across datasets in that category. For QWen 3 model and TAMA-QWen3, we turned off the thinking mode.
References
- Downloads last month
- 4