Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -70,18 +70,7 @@ def bot_streaming(message, history, max_new_tokens=250):
|
|
70 |
yield buffer
|
71 |
|
72 |
|
73 |
-
demo = gr.ChatInterface(fn=bot_streaming, title="Multimodal Llama",
|
74 |
-
[{"text": "Which era does this piece belong to? Give details about the era.", "files":["./examples/rococo.jpg"]},
|
75 |
-
200],
|
76 |
-
[{"text": "Where do the droughts happen according to this diagram?", "files":["./examples/weather_events.png"]},
|
77 |
-
250],
|
78 |
-
[{"text": "What happens when you take out white cat from this chain?", "files":["./examples/ai2d_test.jpg"]},
|
79 |
-
250],
|
80 |
-
[{"text": "How long does it take from invoice date to due date? Be short and concise.", "files":["./examples/invoice.png"]},
|
81 |
-
250],
|
82 |
-
[{"text": "Where to find this monument? Can you give me other recommendations around the area?", "files":["./examples/wat_arun.jpg"]},
|
83 |
-
250],
|
84 |
-
],
|
85 |
textbox=gr.MultimodalTextbox(),
|
86 |
additional_inputs = [gr.Slider(
|
87 |
minimum=10,
|
|
|
70 |
yield buffer
|
71 |
|
72 |
|
73 |
+
demo = gr.ChatInterface(fn=bot_streaming, title="Multimodal Llama",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
74 |
textbox=gr.MultimodalTextbox(),
|
75 |
additional_inputs = [gr.Slider(
|
76 |
minimum=10,
|