Create app.py
Browse files
app.py
ADDED
@@ -0,0 +1,73 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
import gradio as gr
|
2 |
+
import aiohttp
|
3 |
+
|
4 |
+
async def fetch(session, url, context, question):
|
5 |
+
payload = {
|
6 |
+
"context": context,
|
7 |
+
"question": question
|
8 |
+
}
|
9 |
+
async with session.post(url, json=payload) as response:
|
10 |
+
response.raise_for_status()
|
11 |
+
data = await response.json()
|
12 |
+
result = data.get("result", "No result")
|
13 |
+
inference_time = data.get("inference_time", 0)
|
14 |
+
return result, f"{inference_time:.2f} seconds"
|
15 |
+
|
16 |
+
async def send_message(context, question, url):
|
17 |
+
async with aiohttp.ClientSession() as session:
|
18 |
+
result, time = await fetch(session, url, context, question)
|
19 |
+
return result, time
|
20 |
+
|
21 |
+
async def send_to_qwen(context, question):
|
22 |
+
url = "https://qwen.nexaai.com/inference"
|
23 |
+
return await send_message(context, question, url)
|
24 |
+
|
25 |
+
async def send_to_dolphin(context, question):
|
26 |
+
url = "https://dolphin.nexaai.com/inference"
|
27 |
+
return await send_message(context, question, url)
|
28 |
+
|
29 |
+
predefined_questions = [
|
30 |
+
"Tell me where the Nexa AI office is?",
|
31 |
+
"Tell me who founded Nexa AI?",
|
32 |
+
"Tell me what Nexa AI is famous for?",
|
33 |
+
"Tell me what is the mission of Nexa AI?"
|
34 |
+
]
|
35 |
+
|
36 |
+
default_context = """Nexa AI is a Cupertino-based company founded in May 2023 that researches and develops models and tools for on-device AI applications. The company is founded by Alex and Zack. The company is known for its Octopus-series models, which rival large-scale language models in capabilities such as function-calling, multimodality, and action-planning, while remaining efficient and compact for edge device deployment. Nexa AI's mission is to advance on-device AI in collaboration with the global developer community. To this end, the company has created an on-device model hub for users to find, share, and collaborate on open-source AI models optimized for edge devices, as well as an SDK for developers to run and deploy AI models locally."""
|
37 |
+
|
38 |
+
with gr.Blocks() as demo:
|
39 |
+
gr.Markdown("# AI Inference Comparison")
|
40 |
+
|
41 |
+
context = gr.Textbox(value=default_context, label="Background Context", lines=5)
|
42 |
+
question_dropdown = gr.Dropdown(choices=predefined_questions, label="Select a question")
|
43 |
+
custom_question = gr.Textbox(placeholder="Or enter your custom question here...", label="Custom Question")
|
44 |
+
|
45 |
+
with gr.Row():
|
46 |
+
output1 = gr.Textbox(label="Output from Qwen Model", interactive=False)
|
47 |
+
output2 = gr.Textbox(label="Output from Dolphin Model", interactive=False)
|
48 |
+
with gr.Row():
|
49 |
+
total_time1 = gr.Textbox(label="Total Time for Qwen Model", interactive=False)
|
50 |
+
total_time2 = gr.Textbox(label="Total Time for Dolphin Model", interactive=False)
|
51 |
+
|
52 |
+
def update_custom_question(choice):
|
53 |
+
return choice
|
54 |
+
|
55 |
+
question_dropdown.change(update_custom_question, inputs=question_dropdown, outputs=custom_question)
|
56 |
+
|
57 |
+
with gr.Row():
|
58 |
+
send_qwen_button = gr.Button("Send to Qwen")
|
59 |
+
send_dolphin_button = gr.Button("Send to Dolphin")
|
60 |
+
|
61 |
+
send_qwen_button.click(
|
62 |
+
send_to_qwen,
|
63 |
+
inputs=[context, custom_question],
|
64 |
+
outputs=[output1, total_time1]
|
65 |
+
)
|
66 |
+
|
67 |
+
send_dolphin_button.click(
|
68 |
+
send_to_dolphin,
|
69 |
+
inputs=[context, custom_question],
|
70 |
+
outputs=[output2, total_time2]
|
71 |
+
)
|
72 |
+
|
73 |
+
demo.launch()
|