zackli4ai commited on
Commit
cb1707d
·
verified ·
1 Parent(s): da05a35

Create app.py

Browse files
Files changed (1) hide show
  1. app.py +73 -0
app.py ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ import aiohttp
3
+
4
+ async def fetch(session, url, context, question):
5
+ payload = {
6
+ "context": context,
7
+ "question": question
8
+ }
9
+ async with session.post(url, json=payload) as response:
10
+ response.raise_for_status()
11
+ data = await response.json()
12
+ result = data.get("result", "No result")
13
+ inference_time = data.get("inference_time", 0)
14
+ return result, f"{inference_time:.2f} seconds"
15
+
16
+ async def send_message(context, question, url):
17
+ async with aiohttp.ClientSession() as session:
18
+ result, time = await fetch(session, url, context, question)
19
+ return result, time
20
+
21
+ async def send_to_qwen(context, question):
22
+ url = "https://qwen.nexaai.com/inference"
23
+ return await send_message(context, question, url)
24
+
25
+ async def send_to_dolphin(context, question):
26
+ url = "https://dolphin.nexaai.com/inference"
27
+ return await send_message(context, question, url)
28
+
29
+ predefined_questions = [
30
+ "Tell me where the Nexa AI office is?",
31
+ "Tell me who founded Nexa AI?",
32
+ "Tell me what Nexa AI is famous for?",
33
+ "Tell me what is the mission of Nexa AI?"
34
+ ]
35
+
36
+ default_context = """Nexa AI is a Cupertino-based company founded in May 2023 that researches and develops models and tools for on-device AI applications. The company is founded by Alex and Zack. The company is known for its Octopus-series models, which rival large-scale language models in capabilities such as function-calling, multimodality, and action-planning, while remaining efficient and compact for edge device deployment. Nexa AI's mission is to advance on-device AI in collaboration with the global developer community. To this end, the company has created an on-device model hub for users to find, share, and collaborate on open-source AI models optimized for edge devices, as well as an SDK for developers to run and deploy AI models locally."""
37
+
38
+ with gr.Blocks() as demo:
39
+ gr.Markdown("# AI Inference Comparison")
40
+
41
+ context = gr.Textbox(value=default_context, label="Background Context", lines=5)
42
+ question_dropdown = gr.Dropdown(choices=predefined_questions, label="Select a question")
43
+ custom_question = gr.Textbox(placeholder="Or enter your custom question here...", label="Custom Question")
44
+
45
+ with gr.Row():
46
+ output1 = gr.Textbox(label="Output from Qwen Model", interactive=False)
47
+ output2 = gr.Textbox(label="Output from Dolphin Model", interactive=False)
48
+ with gr.Row():
49
+ total_time1 = gr.Textbox(label="Total Time for Qwen Model", interactive=False)
50
+ total_time2 = gr.Textbox(label="Total Time for Dolphin Model", interactive=False)
51
+
52
+ def update_custom_question(choice):
53
+ return choice
54
+
55
+ question_dropdown.change(update_custom_question, inputs=question_dropdown, outputs=custom_question)
56
+
57
+ with gr.Row():
58
+ send_qwen_button = gr.Button("Send to Qwen")
59
+ send_dolphin_button = gr.Button("Send to Dolphin")
60
+
61
+ send_qwen_button.click(
62
+ send_to_qwen,
63
+ inputs=[context, custom_question],
64
+ outputs=[output1, total_time1]
65
+ )
66
+
67
+ send_dolphin_button.click(
68
+ send_to_dolphin,
69
+ inputs=[context, custom_question],
70
+ outputs=[output2, total_time2]
71
+ )
72
+
73
+ demo.launch()