fancyfeast commited on
Commit
537fbaf
·
verified ·
1 Parent(s): 76f8453

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -1,11 +1,11 @@
1
  ---
2
  base_model:
3
  - meta-llama/Llama-3.1-8B-Instruct
4
- - google/siglip-so400m-patch14-384
5
  tags:
6
  - captioning
7
  ---
8
- # Model Card for Llama JoyCaption Alpha Two
9
 
10
  [Github](https://github.com/fpgaminer/joycaption)
11
 
@@ -103,7 +103,7 @@ with torch.no_grad():
103
  vLLM provides the highest performance inference for JoyCaption, and an OpenAI compatible API so JoyCaption can be used like any other VLMs. Example usage:
104
 
105
  ```
106
- vllm serve fancyfeast/llama-joycaption-alpha-two-hf-llava --max-model-len 4096 --enable-prefix-caching
107
  ```
108
 
109
  VLMs are a bit finicky on vLLM, and vLLM is memory hungry, so you may have to adjust settings for your particular environment, such as forcing eager mode, adjusting max-model-len, adjusting gpu_memory_utilization, etc.
 
1
  ---
2
  base_model:
3
  - meta-llama/Llama-3.1-8B-Instruct
4
+ - google/siglip2-so400m-patch14-384
5
  tags:
6
  - captioning
7
  ---
8
+ # Model Card for Llama JoyCaption Beta One
9
 
10
  [Github](https://github.com/fpgaminer/joycaption)
11
 
 
103
  vLLM provides the highest performance inference for JoyCaption, and an OpenAI compatible API so JoyCaption can be used like any other VLMs. Example usage:
104
 
105
  ```
106
+ vllm serve fancyfeast/llama-joycaption-beta-one-hf-llava --max-model-len 4096 --enable-prefix-caching
107
  ```
108
 
109
  VLMs are a bit finicky on vLLM, and vLLM is memory hungry, so you may have to adjust settings for your particular environment, such as forcing eager mode, adjusting max-model-len, adjusting gpu_memory_utilization, etc.