Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -79,18 +79,8 @@ def bot_streaming(message, history, max_new_tokens=250):
|
|
79 |
yield buffer
|
80 |
|
81 |
|
82 |
-
demo = gr.ChatInterface(fn=bot_streaming,
|
83 |
-
|
84 |
-
200],
|
85 |
-
[{"text": "Where do the droughts happen according to this diagram?", "files":["./examples/weather_events.png"]},
|
86 |
-
250],
|
87 |
-
[{"text": "What happens when you take out white cat from this chain?", "files":["./examples/ai2d_test.jpg"]},
|
88 |
-
250],
|
89 |
-
[{"text": "How long does it take from invoice date to due date? Be short and concise.", "files":["./examples/invoice.png"]},
|
90 |
-
250],
|
91 |
-
[{"text": "Where to find this monument? Can you give me other recommendations around the area?", "files":["./examples/wat_arun.jpg"]},
|
92 |
-
250],
|
93 |
-
],
|
94 |
textbox=gr.MultimodalTextbox(),
|
95 |
additional_inputs = [gr.Slider(
|
96 |
minimum=10,
|
@@ -101,7 +91,7 @@ demo = gr.ChatInterface(fn=bot_streaming, title="Multimodal Llama", examples=[
|
|
101 |
)
|
102 |
],
|
103 |
cache_examples=False,
|
104 |
-
description="
|
105 |
stop_btn="Stop Generation",
|
106 |
fill_height=True,
|
107 |
multimodal=True)
|
|
|
79 |
yield buffer
|
80 |
|
81 |
|
82 |
+
demo = gr.ChatInterface(fn=bot_streaming,
|
83 |
+
title="Multimodal Llama 3.2 Vision Instruct 11b, by Meta",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
84 |
textbox=gr.MultimodalTextbox(),
|
85 |
additional_inputs = [gr.Slider(
|
86 |
minimum=10,
|
|
|
91 |
)
|
92 |
],
|
93 |
cache_examples=False,
|
94 |
+
description="Upload an image, and start chatting about it, or simply try one of the examples below. ",
|
95 |
stop_btn="Stop Generation",
|
96 |
fill_height=True,
|
97 |
multimodal=True)
|