Fucius commited on
Commit
98eeca7
·
verified ·
1 Parent(s): 60ccacc

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -2
app.py CHANGED
@@ -63,7 +63,8 @@ title = r"""
63
  """
64
 
65
  description = r"""
66
- <b>Official 🤗 Gradio demo</b> for <a href='https://github.com/kongzhecn/OMG/' target='_blank'><b>OMG: Occlusion-friendly Personalized Multi-concept Generation In Diffusion Models</b></a>.<br>
 
67
  How to use:<br>
68
  1. Select two characters.
69
  2. Enter a text prompt as done in normal text-to-image models.
@@ -366,7 +367,7 @@ def main(device, segment_type):
366
  image = Image.fromarray((image * 255.0).clip(0, 255).astype(np.uint8))
367
  return image
368
 
369
- @spaces.GPU(duration=210)
370
  def generate_image(prompt1, negative_prompt, man, woman, resolution, local_prompt1, local_prompt2, seed, condition, condition_img1, style):
371
  # try:
372
  path1 = lorapath_man[man]
 
63
  """
64
 
65
  description = r"""
66
+ <b>Official 🤗 Gradio demo</b> for <a href='https://github.com/kongzhecn/OMG/' target='_blank'><b>OMG: Occlusion-friendly Personalized Multi-concept Generation In Diffusion Models</b></a>.<be>
67
+ <a href='https://kongzhecn.github.io/omg-project/' target='_blank'><b>Project</b></a>.<a href='https://github.com/kongzhecn/OMG/' target='_blank'><b>Code</b></a>.<a href='https://arxiv.org/abs/2403.10983/' target='_blank'><b>Arxiv</b></a>.<br>
68
  How to use:<br>
69
  1. Select two characters.
70
  2. Enter a text prompt as done in normal text-to-image models.
 
367
  image = Image.fromarray((image * 255.0).clip(0, 255).astype(np.uint8))
368
  return image
369
 
370
+ @spaces.GPU(duration=240)
371
  def generate_image(prompt1, negative_prompt, man, woman, resolution, local_prompt1, local_prompt2, seed, condition, condition_img1, style):
372
  # try:
373
  path1 = lorapath_man[man]