wjpoom commited on
Commit
0d1e383
Β·
verified Β·
1 Parent(s): 6036340

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -3
README.md CHANGED
@@ -160,11 +160,18 @@ model-index:
160
 
161
  ---
162
 
163
- # LLaVA-Next-Inst-It-Vicuna-7B: A Multimodal Model that Excels at Instance-level Understanding
164
- introduced in the paper [Inst-IT: Boosting Multimodal Instance Understanding via Explicit Visual Prompt Instruction Tuning](https://huggingface.co/papers/2412.03565)
165
-
166
  [**🌐 Homepage**](https://inst-it.github.io/) | [**Code**](https://github.com/inst-it/inst-it) | [**πŸ€— Paper**](https://huggingface.co/papers/2412.03565) | [**πŸ“– arXiv**](https://arxiv.org/abs/2412.03565)
167
 
 
 
 
 
 
 
 
 
 
168
  ## Quick Start
169
  **Install**
170
 
 
160
 
161
  ---
162
 
163
+ # LLaVA-Next-Inst-It-Vicuna-7B
 
 
164
  [**🌐 Homepage**](https://inst-it.github.io/) | [**Code**](https://github.com/inst-it/inst-it) | [**πŸ€— Paper**](https://huggingface.co/papers/2412.03565) | [**πŸ“– arXiv**](https://arxiv.org/abs/2412.03565)
165
 
166
+ LLaVA-Next-Inst-It-Vicuna-7B is a multimodal model that excels at instance-level understanding,
167
+ which is introduced in the paper [Inst-IT: Boosting Multimodal Instance Understanding via Explicit Visual Prompt Instruction Tuning](https://huggingface.co/papers/2412.03565)
168
+
169
+ * **Architecture**: clip-vit-large-patch14-336 + Vicuna-7B
170
+ * **Initialized Model**: LLaVA-NeXT
171
+ * **Data**: LLaVA-NeXT-Data / Inst-IT-Dataset
172
+ * **Precision**: bfloat16
173
+
174
+
175
  ## Quick Start
176
  **Install**
177