Spaces:
Running
on
Zero
Running
on
Zero
fancyfeast
commited on
Commit
·
27c9477
1
Parent(s):
89e9fac
Update version and tweak the UI
Browse files- README.md +1 -1
- app.py +5 -3
- requirements.txt +1 -2
README.md
CHANGED
|
@@ -4,7 +4,7 @@ emoji: 🖼️💬
|
|
| 4 |
colorFrom: yellow
|
| 5 |
colorTo: blue
|
| 6 |
sdk: gradio
|
| 7 |
-
sdk_version: 5.
|
| 8 |
app_file: app.py
|
| 9 |
pinned: false
|
| 10 |
---
|
|
|
|
| 4 |
colorFrom: yellow
|
| 5 |
colorTo: blue
|
| 6 |
sdk: gradio
|
| 7 |
+
sdk_version: 5.29.0
|
| 8 |
app_file: app.py
|
| 9 |
pinned: false
|
| 10 |
---
|
app.py
CHANGED
|
@@ -8,11 +8,10 @@ from typing import Generator
|
|
| 8 |
|
| 9 |
|
| 10 |
MODEL_PATH = "fancyfeast/llama-joycaption-beta-one-hf-llava"
|
| 11 |
-
TITLE = "<h1><center>JoyCaption Beta One - (2025-05-10a)</center></h1>"
|
| 12 |
DESCRIPTION = """
|
| 13 |
<div>
|
| 14 |
<p></p>
|
| 15 |
-
<p>**This model cannot see any chat history.**</p>
|
| 16 |
<p>🚨🚨🚨 If the "Help improve JoyCaption" box is checked, the _text_ query you write will be logged and I _might_ use it to help improve JoyCaption.
|
| 17 |
It does not log images, user data, etc; only the text query. I cannot see what images you send, and frankly, I don't want to. But knowing what kinds of instructions
|
| 18 |
and queries users want JoyCaption to handle will help guide me in building JoyCaption's dataset. This dataset will be made public. As always, the model itself is completely
|
|
@@ -33,7 +32,7 @@ CAPTION_TYPE_MAP = {
|
|
| 33 |
"Write a descriptive caption for this image in a casual tone within {word_count} words.",
|
| 34 |
"Write a {length} descriptive caption for this image in a casual tone.",
|
| 35 |
],
|
| 36 |
-
"
|
| 37 |
"Write a stable diffusion prompt for this image.",
|
| 38 |
"Write a stable diffusion prompt for this image within {word_count} words.",
|
| 39 |
"Write a {length} stable diffusion prompt for this image.",
|
|
@@ -238,6 +237,9 @@ with gr.Blocks() as demo:
|
|
| 238 |
outputs=output_caption,
|
| 239 |
)
|
| 240 |
|
|
|
|
|
|
|
|
|
|
| 241 |
gr.Markdown(DESCRIPTION)
|
| 242 |
|
| 243 |
|
|
|
|
| 8 |
|
| 9 |
|
| 10 |
MODEL_PATH = "fancyfeast/llama-joycaption-beta-one-hf-llava"
|
| 11 |
+
TITLE = "<h1><center>JoyCaption Beta One - (2025-05-10a)</center></h1>JoyCaption is an image captioning model"
|
| 12 |
DESCRIPTION = """
|
| 13 |
<div>
|
| 14 |
<p></p>
|
|
|
|
| 15 |
<p>🚨🚨🚨 If the "Help improve JoyCaption" box is checked, the _text_ query you write will be logged and I _might_ use it to help improve JoyCaption.
|
| 16 |
It does not log images, user data, etc; only the text query. I cannot see what images you send, and frankly, I don't want to. But knowing what kinds of instructions
|
| 17 |
and queries users want JoyCaption to handle will help guide me in building JoyCaption's dataset. This dataset will be made public. As always, the model itself is completely
|
|
|
|
| 32 |
"Write a descriptive caption for this image in a casual tone within {word_count} words.",
|
| 33 |
"Write a {length} descriptive caption for this image in a casual tone.",
|
| 34 |
],
|
| 35 |
+
"Stable Diffusion Prompt": [
|
| 36 |
"Write a stable diffusion prompt for this image.",
|
| 37 |
"Write a stable diffusion prompt for this image within {word_count} words.",
|
| 38 |
"Write a {length} stable diffusion prompt for this image.",
|
|
|
|
| 237 |
outputs=output_caption,
|
| 238 |
)
|
| 239 |
|
| 240 |
+
# Initial prompt
|
| 241 |
+
prompt_box.value = build_prompt(caption_type.value, caption_length.value, extra_options.value, name_input.value)
|
| 242 |
+
|
| 243 |
gr.Markdown(DESCRIPTION)
|
| 244 |
|
| 245 |
|
requirements.txt
CHANGED
|
@@ -3,5 +3,4 @@ accelerate
|
|
| 3 |
torch
|
| 4 |
transformers==4.51.0
|
| 5 |
sentencepiece
|
| 6 |
-
torchvision
|
| 7 |
-
pydantic==2.10.6
|
|
|
|
| 3 |
torch
|
| 4 |
transformers==4.51.0
|
| 5 |
sentencepiece
|
| 6 |
+
torchvision
|
|
|