Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
@@ -87,6 +87,7 @@ def generate(
|
|
87 |
decoder_pipeline.to(device)
|
88 |
|
89 |
generator = torch.Generator().manual_seed(seed)
|
|
|
90 |
prior_output = prior_pipeline(
|
91 |
prompt=prompt,
|
92 |
height=height,
|
@@ -272,8 +273,8 @@ with gr.Blocks() as demo:
|
|
272 |
with gr.Blocks(css="style.css") as demo_with_history:
|
273 |
with gr.Tab("App"):
|
274 |
demo.render()
|
275 |
-
|
276 |
-
|
277 |
|
278 |
if __name__ == "__main__":
|
279 |
demo_with_history.queue(max_size=20).launch()
|
|
|
87 |
decoder_pipeline.to(device)
|
88 |
|
89 |
generator = torch.Generator().manual_seed(seed)
|
90 |
+
print("prior_num_inference_steps: ", prior_num_inference_steps)
|
91 |
prior_output = prior_pipeline(
|
92 |
prompt=prompt,
|
93 |
height=height,
|
|
|
273 |
with gr.Blocks(css="style.css") as demo_with_history:
|
274 |
with gr.Tab("App"):
|
275 |
demo.render()
|
276 |
+
with gr.Tab("Past generations"):
|
277 |
+
user_history.render()
|
278 |
|
279 |
if __name__ == "__main__":
|
280 |
demo_with_history.queue(max_size=20).launch()
|