Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
from PIL import Image, ImageDraw, ImageFont
|
| 2 |
import tempfile
|
| 3 |
import gradio as gr
|
| 4 |
-
from smolagents import CodeAgent, InferenceClientModel, TransformersModel
|
| 5 |
from smolagents import DuckDuckGoSearchTool, Tool
|
| 6 |
from huggingface_hub import InferenceClient
|
| 7 |
from diffusers import DiffusionPipeline
|
|
@@ -131,11 +131,15 @@ image_generation_tool = Tool.from_space(
|
|
| 131 |
search_tool = DuckDuckGoSearchTool()
|
| 132 |
print('iiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiii')
|
| 133 |
#llm_engine = InferenceClientModel("Qwen/Qwen2.5-72B-Instruct")
|
| 134 |
-
llm_engine = TransformersModel(
|
| 135 |
-
model_id="Qwen/Qwen2.5-72B-Instruct",
|
| 136 |
-
device="cuda",
|
| 137 |
-
max_new_tokens=5000,
|
| 138 |
-
)
|
|
|
|
|
|
|
|
|
|
|
|
|
| 139 |
print('aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa')
|
| 140 |
|
| 141 |
agent = CodeAgent(tools=[image_generation_tool, search_tool], model=llm_engine)
|
|
|
|
| 1 |
from PIL import Image, ImageDraw, ImageFont
|
| 2 |
import tempfile
|
| 3 |
import gradio as gr
|
| 4 |
+
from smolagents import CodeAgent, InferenceClientModel, TransformersModel
|
| 5 |
from smolagents import DuckDuckGoSearchTool, Tool
|
| 6 |
from huggingface_hub import InferenceClient
|
| 7 |
from diffusers import DiffusionPipeline
|
|
|
|
| 131 |
search_tool = DuckDuckGoSearchTool()
|
| 132 |
print('iiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiii')
|
| 133 |
#llm_engine = InferenceClientModel("Qwen/Qwen2.5-72B-Instruct")
|
| 134 |
+
#llm_engine = TransformersModel(
|
| 135 |
+
# model_id="Qwen/Qwen2.5-72B-Instruct",
|
| 136 |
+
# device="cuda",
|
| 137 |
+
# max_new_tokens=5000,
|
| 138 |
+
#)
|
| 139 |
+
|
| 140 |
+
from smolagents import LiteLLMModel
|
| 141 |
+
|
| 142 |
+
llm_engine = LiteLLMModel(model_id="Qwen/Qwen2.5-72B-Instruct", temperature=0.2, max_tokens=5000)
|
| 143 |
print('aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa')
|
| 144 |
|
| 145 |
agent = CodeAgent(tools=[image_generation_tool, search_tool], model=llm_engine)
|