Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -224,7 +224,7 @@ def llm_call(question_prompt, model_name,
|
|
224 |
top_p=1, n_samples=64, stop=None):
|
225 |
if HUGGINGFACE:
|
226 |
model_inputs = hug_tokenizer([question_prompt], return_tensors="pt").to('cuda')
|
227 |
-
generated_ids = hug_model.generate(**model_inputs, max_length=1400, temperature=1, num_return_sequences=
|
228 |
responses = hug_tokenizer.batch_decode(generated_ids, skip_special_tokens=True)
|
229 |
codes = []
|
230 |
for response in responses:
|
@@ -443,7 +443,7 @@ def main():
|
|
443 |
with gr.Column(scale=1):
|
444 |
canvas = gr.Sketchpad(canvas_size=(512,512), brush=Brush(colors=["black"], default_size=2, color_mode='fixed'))
|
445 |
submit_button = gr.Button("Submit")
|
446 |
-
with gr.Column(scale=
|
447 |
output_gallery = gr.Gallery(
|
448 |
label="Generated images", show_label=False, elem_id="gallery"
|
449 |
, columns=[5], rows=[2], object_fit="contain", height="auto")
|
|
|
224 |
top_p=1, n_samples=64, stop=None):
|
225 |
if HUGGINGFACE:
|
226 |
model_inputs = hug_tokenizer([question_prompt], return_tensors="pt").to('cuda')
|
227 |
+
generated_ids = hug_model.generate(**model_inputs, max_length=1400, temperature=1, num_return_sequences=12, do_sample=True)
|
228 |
responses = hug_tokenizer.batch_decode(generated_ids, skip_special_tokens=True)
|
229 |
codes = []
|
230 |
for response in responses:
|
|
|
443 |
with gr.Column(scale=1):
|
444 |
canvas = gr.Sketchpad(canvas_size=(512,512), brush=Brush(colors=["black"], default_size=2, color_mode='fixed'))
|
445 |
submit_button = gr.Button("Submit")
|
446 |
+
with gr.Column(scale=4):
|
447 |
output_gallery = gr.Gallery(
|
448 |
label="Generated images", show_label=False, elem_id="gallery"
|
449 |
, columns=[5], rows=[2], object_fit="contain", height="auto")
|