fancyfeast commited on
Commit
27c9477
·
1 Parent(s): 89e9fac

Update version and tweak the UI

Browse files
Files changed (3) hide show
  1. README.md +1 -1
  2. app.py +5 -3
  3. requirements.txt +1 -2
README.md CHANGED
@@ -4,7 +4,7 @@ emoji: 🖼️💬
4
  colorFrom: yellow
5
  colorTo: blue
6
  sdk: gradio
7
- sdk_version: 5.6.0
8
  app_file: app.py
9
  pinned: false
10
  ---
 
4
  colorFrom: yellow
5
  colorTo: blue
6
  sdk: gradio
7
+ sdk_version: 5.29.0
8
  app_file: app.py
9
  pinned: false
10
  ---
app.py CHANGED
@@ -8,11 +8,10 @@ from typing import Generator
8
 
9
 
10
  MODEL_PATH = "fancyfeast/llama-joycaption-beta-one-hf-llava"
11
- TITLE = "<h1><center>JoyCaption Beta One - (2025-05-10a)</center></h1>"
12
  DESCRIPTION = """
13
  <div>
14
  <p></p>
15
- <p>**This model cannot see any chat history.**</p>
16
  <p>🚨🚨🚨 If the "Help improve JoyCaption" box is checked, the _text_ query you write will be logged and I _might_ use it to help improve JoyCaption.
17
  It does not log images, user data, etc; only the text query. I cannot see what images you send, and frankly, I don't want to. But knowing what kinds of instructions
18
  and queries users want JoyCaption to handle will help guide me in building JoyCaption's dataset. This dataset will be made public. As always, the model itself is completely
@@ -33,7 +32,7 @@ CAPTION_TYPE_MAP = {
33
  "Write a descriptive caption for this image in a casual tone within {word_count} words.",
34
  "Write a {length} descriptive caption for this image in a casual tone.",
35
  ],
36
- "Training Prompt": [
37
  "Write a stable diffusion prompt for this image.",
38
  "Write a stable diffusion prompt for this image within {word_count} words.",
39
  "Write a {length} stable diffusion prompt for this image.",
@@ -238,6 +237,9 @@ with gr.Blocks() as demo:
238
  outputs=output_caption,
239
  )
240
 
 
 
 
241
  gr.Markdown(DESCRIPTION)
242
 
243
 
 
8
 
9
 
10
  MODEL_PATH = "fancyfeast/llama-joycaption-beta-one-hf-llava"
11
+ TITLE = "<h1><center>JoyCaption Beta One - (2025-05-10a)</center></h1>JoyCaption is an image captioning model"
12
  DESCRIPTION = """
13
  <div>
14
  <p></p>
 
15
  <p>🚨🚨🚨 If the "Help improve JoyCaption" box is checked, the _text_ query you write will be logged and I _might_ use it to help improve JoyCaption.
16
  It does not log images, user data, etc; only the text query. I cannot see what images you send, and frankly, I don't want to. But knowing what kinds of instructions
17
  and queries users want JoyCaption to handle will help guide me in building JoyCaption's dataset. This dataset will be made public. As always, the model itself is completely
 
32
  "Write a descriptive caption for this image in a casual tone within {word_count} words.",
33
  "Write a {length} descriptive caption for this image in a casual tone.",
34
  ],
35
+ "Stable Diffusion Prompt": [
36
  "Write a stable diffusion prompt for this image.",
37
  "Write a stable diffusion prompt for this image within {word_count} words.",
38
  "Write a {length} stable diffusion prompt for this image.",
 
237
  outputs=output_caption,
238
  )
239
 
240
+ # Initial prompt
241
+ prompt_box.value = build_prompt(caption_type.value, caption_length.value, extra_options.value, name_input.value)
242
+
243
  gr.Markdown(DESCRIPTION)
244
 
245
 
requirements.txt CHANGED
@@ -3,5 +3,4 @@ accelerate
3
  torch
4
  transformers==4.51.0
5
  sentencepiece
6
- torchvision
7
- pydantic==2.10.6
 
3
  torch
4
  transformers==4.51.0
5
  sentencepiece
6
+ torchvision