Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
add new gemini model
Browse files- app.py +21 -22
- app_gemini_coder.py +1 -1
- pyproject.toml +1 -1
- requirements.txt +76 -5
app.py
CHANGED
@@ -1,10 +1,11 @@
|
|
1 |
from app_huggingface import demo as demo_huggingface
|
|
|
2 |
from utils import get_app
|
3 |
import gradio as gr
|
4 |
|
5 |
# Create mapping of providers to their code snippets
|
6 |
PROVIDER_SNIPPETS = {
|
7 |
-
"Hugging Face": """
|
8 |
import gradio as gr
|
9 |
import ai_gradio
|
10 |
gr.load(
|
@@ -13,38 +14,36 @@ gr.load(
|
|
13 |
coder=True,
|
14 |
provider="together"
|
15 |
).launch()""",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
}
|
17 |
# Create mapping of providers to their demos
|
18 |
PROVIDERS = {
|
19 |
"Hugging Face": demo_huggingface,
|
|
|
20 |
}
|
21 |
|
22 |
# Modified get_app implementation
|
23 |
demo = gr.Blocks()
|
24 |
with demo:
|
25 |
-
gr.Markdown("#
|
|
|
|
|
|
|
26 |
|
27 |
-
provider_dropdown = gr.Dropdown(
|
28 |
-
choices=list(PROVIDERS.keys()),
|
29 |
-
value="Hugging Face",
|
30 |
-
label="Select code snippet"
|
31 |
-
)
|
32 |
-
code_display = gr.Code(
|
33 |
-
label="Provider Code Snippet",
|
34 |
-
language="python",
|
35 |
-
value=PROVIDER_SNIPPETS["Hugging Face"]
|
36 |
-
)
|
37 |
-
|
38 |
def update_code(provider):
|
39 |
return PROVIDER_SNIPPETS.get(provider, "Code snippet not available")
|
40 |
-
|
41 |
-
|
42 |
-
|
43 |
-
fn=update_code,
|
44 |
-
inputs=[provider_dropdown],
|
45 |
-
outputs=[code_display]
|
46 |
-
)
|
47 |
-
|
48 |
selected_demo = get_app(
|
49 |
models=list(PROVIDERS.keys()),
|
50 |
default_model="Hugging Face",
|
@@ -53,4 +52,4 @@ with demo:
|
|
53 |
)
|
54 |
|
55 |
if __name__ == "__main__":
|
56 |
-
demo.queue(api_open=False).launch(show_api=False)
|
|
|
1 |
from app_huggingface import demo as demo_huggingface
|
2 |
+
from app_gemini_coder import demo as demo_gemini
|
3 |
from utils import get_app
|
4 |
import gradio as gr
|
5 |
|
6 |
# Create mapping of providers to their code snippets
|
7 |
PROVIDER_SNIPPETS = {
|
8 |
+
"Hugging Face": """
|
9 |
import gradio as gr
|
10 |
import ai_gradio
|
11 |
gr.load(
|
|
|
14 |
coder=True,
|
15 |
provider="together"
|
16 |
).launch()""",
|
17 |
+
"Gemini Coder": """
|
18 |
+
import gradio as gr
|
19 |
+
import ai_gradio
|
20 |
+
gr.load(
|
21 |
+
name='gemini:gemini-2.5-pro-exp-03-25',
|
22 |
+
src=ai_gradio.registry,
|
23 |
+
coder=True,
|
24 |
+
provider="together"
|
25 |
+
).launch()
|
26 |
+
""",
|
27 |
}
|
28 |
# Create mapping of providers to their demos
|
29 |
PROVIDERS = {
|
30 |
"Hugging Face": demo_huggingface,
|
31 |
+
"Gemini Coder": demo_gemini,
|
32 |
}
|
33 |
|
34 |
# Modified get_app implementation
|
35 |
demo = gr.Blocks()
|
36 |
with demo:
|
37 |
+
gr.Markdown("#Anycoder")
|
38 |
+
|
39 |
+
provider_dropdown = gr.Dropdown(choices=list(PROVIDERS.keys()), value="Hugging Face", label="Select code snippet")
|
40 |
+
code_display = gr.Code(label="Provider Code Snippet", language="python", value=PROVIDER_SNIPPETS["Hugging Face"])
|
41 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
42 |
def update_code(provider):
|
43 |
return PROVIDER_SNIPPETS.get(provider, "Code snippet not available")
|
44 |
+
|
45 |
+
provider_dropdown.change(fn=update_code, inputs=[provider_dropdown], outputs=[code_display])
|
46 |
+
|
|
|
|
|
|
|
|
|
|
|
47 |
selected_demo = get_app(
|
48 |
models=list(PROVIDERS.keys()),
|
49 |
default_model="Hugging Face",
|
|
|
52 |
)
|
53 |
|
54 |
if __name__ == "__main__":
|
55 |
+
demo.queue(api_open=False).launch(show_api=False)
|
app_gemini_coder.py
CHANGED
@@ -11,7 +11,7 @@ GEMINI_MODELS_DISPLAY = [k.replace("gemini:", "") for k in GEMINI_MODELS_FULL]
|
|
11 |
# Create and launch the interface using get_app utility
|
12 |
demo = get_app(
|
13 |
models=GEMINI_MODELS_FULL, # Use the full names with prefix
|
14 |
-
default_model=GEMINI_MODELS_FULL[
|
15 |
dropdown_label="Select Gemini Model",
|
16 |
choices=GEMINI_MODELS_DISPLAY, # Display names without prefix
|
17 |
src=ai_gradio.registry,
|
|
|
11 |
# Create and launch the interface using get_app utility
|
12 |
demo = get_app(
|
13 |
models=GEMINI_MODELS_FULL, # Use the full names with prefix
|
14 |
+
default_model=GEMINI_MODELS_FULL[0],
|
15 |
dropdown_label="Select Gemini Model",
|
16 |
choices=GEMINI_MODELS_DISPLAY, # Display names without prefix
|
17 |
src=ai_gradio.registry,
|
pyproject.toml
CHANGED
@@ -5,7 +5,7 @@ description = ""
|
|
5 |
readme = "README.md"
|
6 |
requires-python = ">=3.10"
|
7 |
dependencies = [
|
8 |
-
"ai-gradio[huggingface]==0.2.
|
9 |
]
|
10 |
|
11 |
[tool.uv.sources]
|
|
|
5 |
readme = "README.md"
|
6 |
requires-python = ">=3.10"
|
7 |
dependencies = [
|
8 |
+
"ai-gradio[huggingface, gemini]==0.2.55",
|
9 |
]
|
10 |
|
11 |
[tool.uv.sources]
|
requirements.txt
CHANGED
@@ -2,7 +2,7 @@
|
|
2 |
# uv pip compile pyproject.toml -o requirements.txt
|
3 |
accelerate==1.5.2
|
4 |
# via ai-gradio
|
5 |
-
ai-gradio==0.2.
|
6 |
# via anychat (pyproject.toml)
|
7 |
aiofiles==23.2.1
|
8 |
# via gradio
|
@@ -37,6 +37,8 @@ av==13.1.0
|
|
37 |
# via aiortc
|
38 |
bitsandbytes==0.42.0
|
39 |
# via ai-gradio
|
|
|
|
|
40 |
certifi==2025.1.31
|
41 |
# via
|
42 |
# httpcore
|
@@ -85,8 +87,35 @@ fsspec==2025.3.0
|
|
85 |
# gradio-client
|
86 |
# huggingface-hub
|
87 |
# torch
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
88 |
google-crc32c==1.7.0
|
89 |
# via aiortc
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
90 |
gradio==5.23.0
|
91 |
# via
|
92 |
# ai-gradio
|
@@ -98,12 +127,22 @@ gradio-webrtc==0.0.31
|
|
98 |
# via ai-gradio
|
99 |
groovy==0.1.2
|
100 |
# via gradio
|
|
|
|
|
|
|
|
|
|
|
|
|
101 |
h11==0.14.0
|
102 |
# via
|
103 |
# httpcore
|
104 |
# uvicorn
|
105 |
httpcore==1.0.7
|
106 |
# via httpx
|
|
|
|
|
|
|
|
|
107 |
httpx==0.28.1
|
108 |
# via
|
109 |
# gradio
|
@@ -199,6 +238,7 @@ pandas==2.2.3
|
|
199 |
pillow==11.1.0
|
200 |
# via
|
201 |
# ai-gradio
|
|
|
202 |
# gradio
|
203 |
platformdirs==4.3.7
|
204 |
# via pooch
|
@@ -208,15 +248,34 @@ propcache==0.3.0
|
|
208 |
# via
|
209 |
# aiohttp
|
210 |
# yarl
|
211 |
-
|
212 |
-
# via
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
213 |
psutil==7.0.0
|
214 |
# via accelerate
|
|
|
|
|
|
|
|
|
|
|
|
|
215 |
pycparser==2.22
|
216 |
# via cffi
|
217 |
pydantic==2.10.6
|
218 |
# via
|
219 |
# fastapi
|
|
|
|
|
220 |
# gradio
|
221 |
pydantic-core==2.27.2
|
222 |
# via pydantic
|
@@ -234,6 +293,8 @@ pylibsrtp==0.11.0
|
|
234 |
# via aiortc
|
235 |
pyopenssl==25.0.0
|
236 |
# via aiortc
|
|
|
|
|
237 |
python-dateutil==2.9.0.post0
|
238 |
# via pandas
|
239 |
python-dotenv==1.0.1
|
@@ -249,11 +310,15 @@ pyyaml==6.0.2
|
|
249 |
# huggingface-hub
|
250 |
requests==2.32.3
|
251 |
# via
|
|
|
|
|
252 |
# huggingface-hub
|
253 |
# pooch
|
254 |
# twilio
|
255 |
rich==13.9.4
|
256 |
# via typer
|
|
|
|
|
257 |
ruff==0.11.2
|
258 |
# via gradio
|
259 |
safehttpx==0.1.6
|
@@ -296,7 +361,9 @@ torch==2.6.0
|
|
296 |
# accelerate
|
297 |
# ai-gradio
|
298 |
tqdm==4.67.1
|
299 |
-
# via
|
|
|
|
|
300 |
twilio==9.5.1
|
301 |
# via ai-gradio
|
302 |
typer==0.15.2
|
@@ -305,6 +372,7 @@ typing-extensions==4.12.2
|
|
305 |
# via
|
306 |
# anyio
|
307 |
# fastapi
|
|
|
308 |
# gradio
|
309 |
# gradio-client
|
310 |
# huggingface-hub
|
@@ -320,13 +388,16 @@ typing-extensions==4.12.2
|
|
320 |
# uvicorn
|
321 |
tzdata==2025.2
|
322 |
# via pandas
|
|
|
|
|
323 |
urllib3==2.3.0
|
324 |
# via requests
|
325 |
uvicorn==0.34.0
|
326 |
# via gradio
|
327 |
-
websockets==
|
328 |
# via
|
329 |
# ai-gradio
|
|
|
330 |
# gradio-client
|
331 |
yarl==1.18.3
|
332 |
# via aiohttp
|
|
|
2 |
# uv pip compile pyproject.toml -o requirements.txt
|
3 |
accelerate==1.5.2
|
4 |
# via ai-gradio
|
5 |
+
ai-gradio==0.2.55
|
6 |
# via anychat (pyproject.toml)
|
7 |
aiofiles==23.2.1
|
8 |
# via gradio
|
|
|
37 |
# via aiortc
|
38 |
bitsandbytes==0.42.0
|
39 |
# via ai-gradio
|
40 |
+
cachetools==5.5.2
|
41 |
+
# via google-auth
|
42 |
certifi==2025.1.31
|
43 |
# via
|
44 |
# httpcore
|
|
|
87 |
# gradio-client
|
88 |
# huggingface-hub
|
89 |
# torch
|
90 |
+
google-ai-generativelanguage==0.6.15
|
91 |
+
# via google-generativeai
|
92 |
+
google-api-core==2.24.2
|
93 |
+
# via
|
94 |
+
# google-ai-generativelanguage
|
95 |
+
# google-api-python-client
|
96 |
+
# google-generativeai
|
97 |
+
google-api-python-client==2.165.0
|
98 |
+
# via google-generativeai
|
99 |
+
google-auth==2.38.0
|
100 |
+
# via
|
101 |
+
# google-ai-generativelanguage
|
102 |
+
# google-api-core
|
103 |
+
# google-api-python-client
|
104 |
+
# google-auth-httplib2
|
105 |
+
# google-genai
|
106 |
+
# google-generativeai
|
107 |
+
google-auth-httplib2==0.2.0
|
108 |
+
# via google-api-python-client
|
109 |
google-crc32c==1.7.0
|
110 |
# via aiortc
|
111 |
+
google-genai==0.3.0
|
112 |
+
# via ai-gradio
|
113 |
+
google-generativeai==0.8.4
|
114 |
+
# via ai-gradio
|
115 |
+
googleapis-common-protos==1.69.2
|
116 |
+
# via
|
117 |
+
# google-api-core
|
118 |
+
# grpcio-status
|
119 |
gradio==5.23.0
|
120 |
# via
|
121 |
# ai-gradio
|
|
|
127 |
# via ai-gradio
|
128 |
groovy==0.1.2
|
129 |
# via gradio
|
130 |
+
grpcio==1.71.0
|
131 |
+
# via
|
132 |
+
# google-api-core
|
133 |
+
# grpcio-status
|
134 |
+
grpcio-status==1.71.0
|
135 |
+
# via google-api-core
|
136 |
h11==0.14.0
|
137 |
# via
|
138 |
# httpcore
|
139 |
# uvicorn
|
140 |
httpcore==1.0.7
|
141 |
# via httpx
|
142 |
+
httplib2==0.22.0
|
143 |
+
# via
|
144 |
+
# google-api-python-client
|
145 |
+
# google-auth-httplib2
|
146 |
httpx==0.28.1
|
147 |
# via
|
148 |
# gradio
|
|
|
238 |
pillow==11.1.0
|
239 |
# via
|
240 |
# ai-gradio
|
241 |
+
# google-genai
|
242 |
# gradio
|
243 |
platformdirs==4.3.7
|
244 |
# via pooch
|
|
|
248 |
# via
|
249 |
# aiohttp
|
250 |
# yarl
|
251 |
+
proto-plus==1.26.1
|
252 |
+
# via
|
253 |
+
# google-ai-generativelanguage
|
254 |
+
# google-api-core
|
255 |
+
protobuf==5.29.4
|
256 |
+
# via
|
257 |
+
# google-ai-generativelanguage
|
258 |
+
# google-api-core
|
259 |
+
# google-generativeai
|
260 |
+
# googleapis-common-protos
|
261 |
+
# grpcio-status
|
262 |
+
# onnxruntime
|
263 |
+
# proto-plus
|
264 |
psutil==7.0.0
|
265 |
# via accelerate
|
266 |
+
pyasn1==0.6.1
|
267 |
+
# via
|
268 |
+
# pyasn1-modules
|
269 |
+
# rsa
|
270 |
+
pyasn1-modules==0.4.1
|
271 |
+
# via google-auth
|
272 |
pycparser==2.22
|
273 |
# via cffi
|
274 |
pydantic==2.10.6
|
275 |
# via
|
276 |
# fastapi
|
277 |
+
# google-genai
|
278 |
+
# google-generativeai
|
279 |
# gradio
|
280 |
pydantic-core==2.27.2
|
281 |
# via pydantic
|
|
|
293 |
# via aiortc
|
294 |
pyopenssl==25.0.0
|
295 |
# via aiortc
|
296 |
+
pyparsing==3.2.3
|
297 |
+
# via httplib2
|
298 |
python-dateutil==2.9.0.post0
|
299 |
# via pandas
|
300 |
python-dotenv==1.0.1
|
|
|
310 |
# huggingface-hub
|
311 |
requests==2.32.3
|
312 |
# via
|
313 |
+
# google-api-core
|
314 |
+
# google-genai
|
315 |
# huggingface-hub
|
316 |
# pooch
|
317 |
# twilio
|
318 |
rich==13.9.4
|
319 |
# via typer
|
320 |
+
rsa==4.9
|
321 |
+
# via google-auth
|
322 |
ruff==0.11.2
|
323 |
# via gradio
|
324 |
safehttpx==0.1.6
|
|
|
361 |
# accelerate
|
362 |
# ai-gradio
|
363 |
tqdm==4.67.1
|
364 |
+
# via
|
365 |
+
# google-generativeai
|
366 |
+
# huggingface-hub
|
367 |
twilio==9.5.1
|
368 |
# via ai-gradio
|
369 |
typer==0.15.2
|
|
|
372 |
# via
|
373 |
# anyio
|
374 |
# fastapi
|
375 |
+
# google-generativeai
|
376 |
# gradio
|
377 |
# gradio-client
|
378 |
# huggingface-hub
|
|
|
388 |
# uvicorn
|
389 |
tzdata==2025.2
|
390 |
# via pandas
|
391 |
+
uritemplate==4.1.1
|
392 |
+
# via google-api-python-client
|
393 |
urllib3==2.3.0
|
394 |
# via requests
|
395 |
uvicorn==0.34.0
|
396 |
# via gradio
|
397 |
+
websockets==14.2
|
398 |
# via
|
399 |
# ai-gradio
|
400 |
+
# google-genai
|
401 |
# gradio-client
|
402 |
yarl==1.18.3
|
403 |
# via aiohttp
|