Update README.md
Browse files
README.md
CHANGED
@@ -1,11 +1,11 @@
|
|
1 |
---
|
2 |
base_model:
|
3 |
- meta-llama/Llama-3.1-8B-Instruct
|
4 |
-
- google/
|
5 |
tags:
|
6 |
- captioning
|
7 |
---
|
8 |
-
# Model Card for Llama JoyCaption
|
9 |
|
10 |
[Github](https://github.com/fpgaminer/joycaption)
|
11 |
|
@@ -103,7 +103,7 @@ with torch.no_grad():
|
|
103 |
vLLM provides the highest performance inference for JoyCaption, and an OpenAI compatible API so JoyCaption can be used like any other VLMs. Example usage:
|
104 |
|
105 |
```
|
106 |
-
vllm serve fancyfeast/llama-joycaption-
|
107 |
```
|
108 |
|
109 |
VLMs are a bit finicky on vLLM, and vLLM is memory hungry, so you may have to adjust settings for your particular environment, such as forcing eager mode, adjusting max-model-len, adjusting gpu_memory_utilization, etc.
|
|
|
1 |
---
|
2 |
base_model:
|
3 |
- meta-llama/Llama-3.1-8B-Instruct
|
4 |
+
- google/siglip2-so400m-patch14-384
|
5 |
tags:
|
6 |
- captioning
|
7 |
---
|
8 |
+
# Model Card for Llama JoyCaption Beta One
|
9 |
|
10 |
[Github](https://github.com/fpgaminer/joycaption)
|
11 |
|
|
|
103 |
vLLM provides the highest performance inference for JoyCaption, and an OpenAI compatible API so JoyCaption can be used like any other VLMs. Example usage:
|
104 |
|
105 |
```
|
106 |
+
vllm serve fancyfeast/llama-joycaption-beta-one-hf-llava --max-model-len 4096 --enable-prefix-caching
|
107 |
```
|
108 |
|
109 |
VLMs are a bit finicky on vLLM, and vLLM is memory hungry, so you may have to adjust settings for your particular environment, such as forcing eager mode, adjusting max-model-len, adjusting gpu_memory_utilization, etc.
|