Model Card for LLaVA_MORE-phi_4-finetuning

🔥 LLaVA-MORE 🔥

A Comparative Study of LLMs and Visual Backbones
for Enhanced Visual Instruction Tuning

This model is part of the LLaVA-MORE family of Multimodal Large Language Models (MLLMs), presented in the paper LLaVA-MORE: A Comparative Study of LLMs and Visual Backbones for Enhanced Visual Instruction Tuning.

LLaVA-MORE integrates recent language models with diverse visual backbones. It employs a unified training protocol applied consistently across all architectures to ensure fair comparisons and systematically explore the trade-offs between model size, architecture, and performance. This model, LLaVA_MORE-phi_4-finetuning, uses Phi-4 Instruct as its LLM backbone and is finetuned on the LLaVA-Instruct-665K dataset.

It is designed for multimodal reasoning, generation, and instruction following, and provides insights into the design of more effective MLLMs.

Citation

If you make use of our work, please cite our paper:

@inproceedings{cocchi2025llava,
      title={{LLaVA-MORE: A Comparative Study of LLMs and Visual Backbones for Enhanced Visual Instruction Tuning}},
      author={Cocchi, Federico and Moratelli, Nicholas and Caffagni, Davide and Sarto, Sara and Baraldi, Lorenzo and Cornia, Marcella and Cucchiara, Rita},
      booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops},
      year={2025}
}

Model Details

Model Description

This is a checkpoint from the LLaVA-MORE family of MLLMs. It integrates the Phi-4 Instruct Large Language Model with a visual backbone (specifically, openai/clip-vit-large-patch14-336 as per config.json). It has been finetuned on the LLaVA-Instruct-665K dataset. The project aims to provide a reproducible evaluation framework to guide future model development by systematically studying the impact of different LLMs and visual encoders, as well as factors like image resolution and pre-training datasets.

  • Developed by: Federico Cocchi, Nicholas Moratelli, Davide Caffagni, Sara Sarto, Lorenzo Baraldi, Marcella Cornia and Rita Cucchiara (AImageLab, University of Modena and Reggio Emilia).
  • Model type: Multimodal Large Language Model (MLLM) / Vision-Language Model
  • Language(s): English
  • License: Apache-2.0
  • Finetuned from model: microsoft/Phi-4-mini-instruct

Model Sources

Uses

Direct Use

This model is intended for various multimodal reasoning, generation, and instruction-following tasks. It can be used to process visual inputs in conjunction with textual prompts to generate informative and relevant text responses. Typical applications include visual question answering, image captioning, and conversational AI involving images.

Out-of-Scope Use

This model is not intended for generating harmful content, engaging in misinformation, or being deployed in applications without proper human oversight. As an AI model, it may hallucinate or produce factually incorrect information. It should not be used in safety-critical applications without thorough domain-specific evaluation and mitigation strategies.

Bias, Risks, and Limitations

Given that the model is trained on large datasets, it may inherit biases present in the data, leading to biased outputs. Potential risks include generating offensive, inaccurate, or harmful content. Like all generative models, it may also hallucinate or provide factually incorrect information.

Recommendations

Users should be aware of the inherent biases and limitations of MLLMs. It is recommended to apply human review to outputs, especially in sensitive applications. Further research and evaluation are needed to fully understand and mitigate potential societal impacts.

How to Get Started with the Model

To get started with inference, you can use the transformers library along with the provided run_llava.py script from the project's GitHub repository or integrate it directly using Python as shown below.

First, install the necessary packages as described in the GitHub Installation section:

conda create -n more python==3.8.16
conda activate more
pip install -r requirements.txt # Refer to the GitHub repo for the exact requirements.txt

Using the run_llava.py script (recommended for full functionality):

cd ~/LLaVA-MORE # Navigate to the cloned LLaVA-MORE repository
source activate more
export PYTHONPATH=.

model_path=aimagelab/LLaVA_MORE-phi_4-finetuning # Adjust to the specific model path
model_architecture=llava_phi # Based on config.json
conversation=phi_4 # This might vary based on tokenizer config, check original LLaVA-MORE code for best match

export HF_TOKEN=hf_read_token # Replace with your Hugging Face read token if needed
export TOKENIZER_PATH=$model_path

python -u src/llava/eval/run_llava.py --model-path $model_path --model-architecture $model_architecture --conv-mode $conversation

Training Details

Training Data

The LLaVA-MORE models are typically trained in two stages:

Training Procedure

The training employs a unified protocol consistently applied across all architectures to ensure fair comparisons and enhance reproducibility. The project publicly releases the source code and bash scripts for distributed training on HPC facilities with a SLURM scheduler. More details on the training procedure and hyperparameters can be found in the Training section of the GitHub repository.

Evaluation

Benchmarks and Comparisons on Instruction Multimodal Datasets in the Literature

The table below presents the performance of LLaVA-MORE variants, including this model, compared to other LLaVA versions across various multimodal datasets. For the most up-to-date and complete evaluation results, please refer to the Performance section in the GitHub repository.

Model Name Text-VQA* Science-QA AI2D SEED-vid SEED-all SEED-img MMMU MMBench-Cn MMBench-En POPE GQA MME-P MME-C
LLaVA-v1.5-7B 58.2 69.0 56.4 42.0 61.6 66.8 34.2 56.5 65.3 85.6 62.4 1474.3 314.6
LLaVA-v1.5-LLaMA3-8B 57.6 74.2 60.7 42.0 64.3 70.1 37.3 65.4 70.3 85.4 63.5 1544.4 330.3
LLaVA-v1.5-LLaMA3_1-8B 58.4 76.3 61.8 42.4 64.1 69.8 39.4 68.2 72.4 85.1 63.6 1531.5 353.3
LLaVA-v1.5-LLaMA3_1-8B-S2 60.9 76.7 62.2 42.3 64.2 69.9 38.7 65.8 71.1 86.5 64.5 1563.8 293.2
LLaVA-v1.5-LLaMA3_1-8B-siglip 62.1 77.5 63.6 46.1 65.8 71.0 39.8 68.2 73.1 86.1 64.6 1531.0 315.4
LLaVA-v1.5-LLaMA3_1-8B-S2-siglip 63.5 77.1 62.7 44.7 65.5 71.0 40.0 68.0 71.8 86.0 64.9 1541.4 336.4
LLaVA-v1.5-Phi_4-4B 54.0 71.3 61.1 42.3 63.5 69.1 38.8 64.2 69.2 85.9 62.1 1372.2 281.1
LLaVA-v1.5-gemma_2-9B 60.7 75.4 64.8 44.1 64.5 69.9 37.9 65.9 71.9 86.8 64.2 1522.5 307.5
LLaVA-v1.5-gemma_2-9B-siglip2 66.7 76.2 65.3 46.0 67.5 73.1 38.7 68.0 72.0 86.1 65.6 1510.9 308.2
LLaVA-v1.5-Distill-LLaMA-8B 56.3 74.5 58.8 43.5 63.5 68.6 38.1 66.8 61.3 85.1 63.0 1495.1 295.0

* The results of TextVQA are computed with OCR token in the input prompt. The models in bold represent LLaVA-MORE variants.

Checkpoints

For a complete list of all LLaVA-MORE checkpoints, you can refer to the Hugging Face model collection.

Acknowledgments

We thank the LLaVA team for open-sourcing a modular codebase to extend and train different models within the LLaVA family. We are also happy users of the lmms-eval library, which has significantly reduced the evaluation time of our checkpoints across different datasets.

We also thank CINECA for the availability of high-performance computing resources used to train LLaVA-MORE. This work is supported by the PNRR-M4C2 project FAIR - Future Artificial Intelligence Research and by the PNRR project ITSERR - Italian Strengthening of Esfri RI Resilience.

Downloads last month
10
Safetensors
Model size
4.15B params
Tensor type
F16
·
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for aimagelab/LLaVA_MORE-phi_4-finetuning

Finetuned
(40)
this model

Collection including aimagelab/LLaVA_MORE-phi_4-finetuning