Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
@@ -5,7 +5,7 @@ import pytz
|
|
5 |
import yaml
|
6 |
from tools.final_answer import FinalAnswerTool
|
7 |
|
8 |
-
from Gradio_UI import GradioUI
|
9 |
|
10 |
# Below is an example of a tool that does nothing. Amaze us with your creativity !
|
11 |
@tool
|
@@ -39,12 +39,12 @@ final_answer = FinalAnswerTool()
|
|
39 |
# If the agent does not answer, the model is overloaded, please use another model or the following Hugging Face Endpoint that also contains qwen2.5 coder:
|
40 |
# model_id='https://pflgm2locj2t89co.us-east-1.aws.endpoints.huggingface.cloud'
|
41 |
|
42 |
-
model = HfApiModel(
|
43 |
-
max_tokens=2096,
|
44 |
-
temperature=0.5,
|
45 |
-
model_id='Qwen/Qwen2.5-Coder-32B-Instruct',# it is possible that this model may be overloaded
|
46 |
-
custom_role_conversions=None,
|
47 |
-
)
|
48 |
|
49 |
|
50 |
# Import tool from Hub
|
@@ -56,18 +56,20 @@ web_browser = VisitWebpageTool()
|
|
56 |
|
57 |
with open("prompts.yaml", 'r') as stream:
|
58 |
prompt_templates = yaml.safe_load(stream)
|
59 |
-
|
60 |
-
agent = CodeAgent(
|
61 |
-
model=model,
|
62 |
-
tools=[final_answer, get_current_time_in_timezone, web_search_engine, web_browser], ## add your tools here (don't remove final answer)
|
63 |
-
max_steps=6,
|
64 |
-
verbosity_level=1,
|
65 |
-
grammar=None,
|
66 |
-
planning_interval=None,
|
67 |
-
name=None,
|
68 |
-
description=None,
|
69 |
-
prompt_templates=prompt_templates
|
70 |
-
)
|
71 |
|
|
|
72 |
|
73 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5 |
import yaml
|
6 |
from tools.final_answer import FinalAnswerTool
|
7 |
|
8 |
+
from Gradio_UI import GradioUI, AgentCloner
|
9 |
|
10 |
# Below is an example of a tool that does nothing. Amaze us with your creativity !
|
11 |
@tool
|
|
|
39 |
# If the agent does not answer, the model is overloaded, please use another model or the following Hugging Face Endpoint that also contains qwen2.5 coder:
|
40 |
# model_id='https://pflgm2locj2t89co.us-east-1.aws.endpoints.huggingface.cloud'
|
41 |
|
42 |
+
# model = HfApiModel(
|
43 |
+
# max_tokens=2096,
|
44 |
+
# temperature=0.5,
|
45 |
+
# model_id='Qwen/Qwen2.5-Coder-32B-Instruct',# it is possible that this model may be overloaded
|
46 |
+
# custom_role_conversions=None,
|
47 |
+
# )
|
48 |
|
49 |
|
50 |
# Import tool from Hub
|
|
|
56 |
|
57 |
with open("prompts.yaml", 'r') as stream:
|
58 |
prompt_templates = yaml.safe_load(stream)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
59 |
|
60 |
+
tools = [final_answer, get_current_time_in_timezone, web_search_engine, web_browser], ## add your tools here (don't remove final answer)
|
61 |
|
62 |
+
# agent = CodeAgent(
|
63 |
+
# model=model,
|
64 |
+
# tools=[final_answer, get_current_time_in_timezone, web_search_engine, web_browser], ## add your tools here (don't remove final answer)
|
65 |
+
# max_steps=6,
|
66 |
+
# verbosity_level=1,
|
67 |
+
# grammar=None,
|
68 |
+
# planning_interval=None,
|
69 |
+
# name=None,
|
70 |
+
# description=None,
|
71 |
+
# prompt_templates=prompt_templates
|
72 |
+
# )
|
73 |
+
|
74 |
+
|
75 |
+
GradioUI(AgentCloner(tools, prompt_templates)).launch()
|