Profakerr commited on
Commit
ba73128
·
verified ·
1 Parent(s): 72dd1ab

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +13 -5
app.py CHANGED
@@ -62,7 +62,7 @@ def is_civitai_url(url):
62
 
63
  @spaces.GPU
64
  def generate_image(prompt, negative_prompt, lora_url, num_inference_steps=30, guidance_scale=7.0,
65
- model="Real6.0", num_images=1, width=512, height=512):
66
 
67
  if model == "Real5.0":
68
  model_id = "SG161222/Realistic_Vision_V5.0_noVAE"
@@ -147,6 +147,11 @@ def generate_image(prompt, negative_prompt, lora_url, num_inference_steps=30, gu
147
  use_karras_sigmas=True
148
  )
149
 
 
 
 
 
 
150
  text_inputs = tokenizer(
151
  prompt,
152
  padding="max_length",
@@ -175,10 +180,11 @@ def generate_image(prompt, negative_prompt, lora_url, num_inference_steps=30, gu
175
  guidance_scale=guidance_scale,
176
  width=width,
177
  height=height,
178
- num_images_per_prompt=num_images
 
179
  )
180
 
181
- return result.images
182
 
183
  def clean_lora_cache():
184
  """Clean the LoRA cache directory"""
@@ -259,6 +265,7 @@ with gr.Blocks() as demo:
259
  step=0.5,
260
  label="Guidance Scale"
261
  )
 
262
 
263
  with gr.Column():
264
  # Output component
@@ -269,13 +276,14 @@ with gr.Blocks() as demo:
269
  columns=2,
270
  rows=2
271
  )
 
272
 
273
  # Connect the interface to the generation function
274
  generate_button.click(
275
  fn=generate_image,
276
  inputs=[prompt, negative_prompt, lora_input, steps_slider, guidance_slider,
277
- model, num_images, width, height],
278
- outputs=gallery
279
  )
280
 
281
  # Connect clear cache button
 
62
 
63
  @spaces.GPU
64
  def generate_image(prompt, negative_prompt, lora_url, num_inference_steps=30, guidance_scale=7.0,
65
+ model="Real6.0", num_images=1, width=512, height=512,seed=None):
66
 
67
  if model == "Real5.0":
68
  model_id = "SG161222/Realistic_Vision_V5.0_noVAE"
 
147
  use_karras_sigmas=True
148
  )
149
 
150
+ if seed is None:
151
+ seed = random.randint(0, 2**32 - 1)
152
+
153
+ generator = torch.manual_seed(seed)
154
+
155
  text_inputs = tokenizer(
156
  prompt,
157
  padding="max_length",
 
180
  guidance_scale=guidance_scale,
181
  width=width,
182
  height=height,
183
+ num_images_per_prompt=num_images,
184
+ generator=generator
185
  )
186
 
187
+ return result.images,seed
188
 
189
  def clean_lora_cache():
190
  """Clean the LoRA cache directory"""
 
265
  step=0.5,
266
  label="Guidance Scale"
267
  )
268
+ seed_input = gr.Number(value=random.randint(0, 2**32 - 1), label="Seed (optional)")
269
 
270
  with gr.Column():
271
  # Output component
 
276
  columns=2,
277
  rows=2
278
  )
279
+ seed_display = gr.Textbox(label="Seed Used", interactive=False)
280
 
281
  # Connect the interface to the generation function
282
  generate_button.click(
283
  fn=generate_image,
284
  inputs=[prompt, negative_prompt, lora_input, steps_slider, guidance_slider,
285
+ model, num_images, width, height,seed_input],
286
+ outputs=[gallery,seed]
287
  )
288
 
289
  # Connect clear cache button