Spaces:
Running on CPU Upgrade

app.py CHANGED
@@ -1,56 +1,71 @@
1
- from app_huggingface import demo as demo_huggingface
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  from utils import get_app
3
- import gradio as gr
4
 
5
- # Create mapping of providers to their code snippets
6
- PROVIDER_SNIPPETS = {
7
- "Hugging Face": """
8
- import gradio as gr
9
- import ai_gradio
10
- gr.load(
11
- name='huggingface:deepseek-ai/DeepSeek-R1',
12
- src=ai_gradio.registry,
13
- coder=True,
14
- provider="together"
15
- ).launch()""",
16
- }
17
  # Create mapping of providers to their demos
18
  PROVIDERS = {
19
- "Hugging Face": demo_huggingface,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
  }
21
 
22
- # Modified get_app implementation
23
- demo = gr.Blocks()
24
- with demo:
25
- gr.Markdown("# Anychat")
26
-
27
- provider_dropdown = gr.Dropdown(
28
- choices=list(PROVIDERS.keys()),
29
- value="Hugging Face",
30
- label="Select code snippet"
31
- )
32
- code_display = gr.Code(
33
- label="Provider Code Snippet",
34
- language="python",
35
- value=PROVIDER_SNIPPETS["Hugging Face"]
36
- )
37
-
38
- def update_code(provider):
39
- return PROVIDER_SNIPPETS.get(provider, "Code snippet not available")
40
-
41
-
42
- provider_dropdown.change(
43
- fn=update_code,
44
- inputs=[provider_dropdown],
45
- outputs=[code_display]
46
- )
47
-
48
- selected_demo = get_app(
49
- models=list(PROVIDERS.keys()),
50
- default_model="Hugging Face",
51
- src=PROVIDERS,
52
- dropdown_label="Select Provider",
53
- )
54
 
55
  if __name__ == "__main__":
56
- demo.queue(api_open=False).launch(show_api=False)
 
1
+ from app_lumaai import demo as demo_lumaai
2
+ from app_claude import demo as demo_claude
3
+ from app_cohere import demo as demo_cohere
4
+ from app_fal import demo as demo_fal
5
+ from app_fireworks import demo as demo_fireworks
6
+ from app_meta import demo as demo_meta
7
+ from app_perplexity import demo as demo_perplexity
8
+ from app_playai import demo as demo_playai
9
+ from app_replicate import demo as demo_replicate
10
+ from app_sambanova import demo as demo_sambanova
11
+ from app_together import demo as demo_together
12
+ from app_xai import demo as demo_grok
13
+ from app_qwen import demo as demo_qwen
14
+ from app_crew import demo as demo_crew
15
+ from app_hyperbolic import demo as demo_hyperbolic
16
+ from app_gemini_coder import demo as demo_gemini_coder
17
+ from app_gemini import demo as demo_gemini
18
+ from app_hyperbolic_coder import demo as demo_hyperbolic_coder
19
+ from app_smolagents import demo as demo_smolagents
20
+ from app_groq import demo as demo_groq
21
+ from app_groq_coder import demo as demo_groq_coder
22
+ from app_openai_coder import demo as demo_openai_coder
23
+ from app_langchain import demo as demo_langchain
24
+ from app_mistral import demo as demo_mistral
25
+ from app_minimax import demo as demo_minimax
26
+ from app_minimax_coder import demo as demo_minimax_coder
27
+ from app_nvidia import demo as demo_nvidia
28
+ from app_deepseek import demo as demo_deepseek
29
  from utils import get_app
 
30
 
 
 
 
 
 
 
 
 
 
 
 
 
31
  # Create mapping of providers to their demos
32
  PROVIDERS = {
33
+ "Gemini Coder": demo_gemini_coder,
34
+ "Groq Coder": demo_groq_coder,
35
+ "DeepSeek Coder": demo_deepseek,
36
+ "Minimax Coder": demo_minimax_coder,
37
+ "NVIDIA": demo_nvidia,
38
+ "Minimax": demo_minimax,
39
+ "Mistral": demo_mistral,
40
+ "Langchain Agent": demo_langchain,
41
+ "OpenAI Coder": demo_openai_coder,
42
+ "Gemini": demo_gemini,
43
+ "Hyperbolic Coder": demo_hyperbolic_coder,
44
+ "SmolAgents": demo_smolagents,
45
+ "Qwen": demo_qwen,
46
+ "Hyperbolic": demo_hyperbolic,
47
+ "CrewAI": demo_crew,
48
+ "LumaAI": demo_lumaai,
49
+ "Grok": demo_grok,
50
+ "Cohere": demo_cohere,
51
+ "SambaNova": demo_sambanova,
52
+ "Fireworks": demo_fireworks,
53
+ "Together": demo_together,
54
+ "Groq": demo_groq,
55
+ "Meta Llama": demo_meta,
56
+ "Replicate": demo_replicate,
57
+ "Fal": demo_fal,
58
+ "PlayAI": demo_playai,
59
+ "Claude": demo_claude,
60
+ "Perplexity": demo_perplexity,
61
  }
62
 
63
+ demo = get_app(
64
+ models=list(PROVIDERS.keys()),
65
+ default_model="Gemini Coder",
66
+ src=PROVIDERS,
67
+ dropdown_label="Select Provider",
68
+ )
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
69
 
70
  if __name__ == "__main__":
71
+ demo.queue(api_open=False).launch(show_api=False)
app_gemini_coder.py CHANGED
@@ -11,7 +11,7 @@ GEMINI_MODELS_DISPLAY = [k.replace("gemini:", "") for k in GEMINI_MODELS_FULL]
11
  # Create and launch the interface using get_app utility
12
  demo = get_app(
13
  models=GEMINI_MODELS_FULL, # Use the full names with prefix
14
- default_model=GEMINI_MODELS_FULL[-2],
15
  dropdown_label="Select Gemini Model",
16
  choices=GEMINI_MODELS_DISPLAY, # Display names without prefix
17
  src=ai_gradio.registry,
 
11
  # Create and launch the interface using get_app utility
12
  demo = get_app(
13
  models=GEMINI_MODELS_FULL, # Use the full names with prefix
14
+ default_model=GEMINI_MODELS_FULL[-1],
15
  dropdown_label="Select Gemini Model",
16
  choices=GEMINI_MODELS_DISPLAY, # Display names without prefix
17
  src=ai_gradio.registry,
app_huggingface.py CHANGED
@@ -1,21 +1,67 @@
1
- import ai_gradio
 
2
 
3
- from utils_ai_gradio import get_app
4
 
5
- # Get the hyperbolic models but keep their full names for loading
6
- HUGGINGFACE_MODELS_FULL = [k for k in ai_gradio.registry.keys() if k.startswith("huggingface:")]
7
 
8
- # Create display names without the prefix
9
- HUGGINGFACE_MODELS_DISPLAY = [k.replace("huggingface:", "") for k in HUGGINGFACE_MODELS_FULL]
 
 
 
10
 
 
 
11
 
12
- # Create and launch the interface using get_app utility
13
- demo = get_app(
14
- models=HUGGINGFACE_MODELS_FULL, # Use the full names with prefix
15
- default_model=HUGGINGFACE_MODELS_FULL[0],
16
- dropdown_label="Select Huggingface Model",
17
- choices=HUGGINGFACE_MODELS_DISPLAY, # Display names without prefix
18
- fill_height=True,
19
- coder=True,
20
- provider="fireworks-ai"
21
- )
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ from gradio_client import Client, handle_file
3
 
4
+ MODELS = {"SmolVLM-Instruct": "akhaliq/SmolVLM-Instruct"}
5
 
 
 
6
 
7
+ def create_chat_fn(client):
8
+ def chat(message, history):
9
+ # Extract text and files from the message
10
+ text = message.get("text", "")
11
+ files = message.get("files", [])
12
 
13
+ # Handle file uploads if present
14
+ processed_files = [handle_file(f) for f in files]
15
 
16
+ response = client.predict(
17
+ message={"text": text, "files": processed_files},
18
+ system_prompt="You are a helpful AI assistant.",
19
+ temperature=0.7,
20
+ max_new_tokens=1024,
21
+ top_k=40,
22
+ repetition_penalty=1.1,
23
+ top_p=0.95,
24
+ api_name="/chat",
25
+ )
26
+ return response
27
+
28
+ return chat
29
+
30
+
31
+ def set_client_for_session(model_name, request: gr.Request):
32
+ headers = {}
33
+ if request and hasattr(request, "headers"):
34
+ x_ip_token = request.headers.get("x-ip-token")
35
+ if x_ip_token:
36
+ headers["X-IP-Token"] = x_ip_token
37
+
38
+ return Client(MODELS[model_name], headers=headers)
39
+
40
+
41
+ def safe_chat_fn(message, history, client):
42
+ if client is None:
43
+ return "Error: Client not initialized. Please refresh the page."
44
+ try:
45
+ return create_chat_fn(client)(message, history)
46
+ except Exception as e:
47
+ print(f"Error during chat: {e!s}")
48
+ return f"Error during chat: {e!s}"
49
+
50
+
51
+ with gr.Blocks() as demo:
52
+ client = gr.State()
53
+
54
+ model_dropdown = gr.Dropdown(
55
+ choices=list(MODELS.keys()), value="SmolVLM-Instruct", label="Select Model", interactive=True
56
+ )
57
+
58
+ chat_interface = gr.ChatInterface(fn=safe_chat_fn, additional_inputs=[client], multimodal=True)
59
+
60
+ # Update client when model changes
61
+ model_dropdown.change(fn=set_client_for_session, inputs=[model_dropdown], outputs=[client])
62
+
63
+ # Initialize client on page load
64
+ demo.load(fn=set_client_for_session, inputs=[gr.State("SmolVLM-Instruct")], outputs=[client])
65
+
66
+ if __name__ == "__main__":
67
+ demo.launch()
app_hyperbolic.py CHANGED
@@ -12,7 +12,7 @@ HYPERBOLIC_MODELS_DISPLAY = [k.replace("hyperbolic:", "") for k in HYPERBOLIC_MO
12
  # Create and launch the interface using get_app utility
13
  demo = get_app(
14
  models=HYPERBOLIC_MODELS_FULL, # Use the full names with prefix
15
- default_model=HYPERBOLIC_MODELS_FULL[-2],
16
  dropdown_label="Select Hyperbolic Model",
17
  choices=HYPERBOLIC_MODELS_DISPLAY, # Display names without prefix
18
  fill_height=True,
 
12
  # Create and launch the interface using get_app utility
13
  demo = get_app(
14
  models=HYPERBOLIC_MODELS_FULL, # Use the full names with prefix
15
+ default_model=HYPERBOLIC_MODELS_FULL[-1],
16
  dropdown_label="Select Hyperbolic Model",
17
  choices=HYPERBOLIC_MODELS_DISPLAY, # Display names without prefix
18
  fill_height=True,
app_hyperbolic_coder.py CHANGED
@@ -12,7 +12,7 @@ HYPERBOLIC_MODELS_DISPLAY = [k.replace("hyperbolic:", "") for k in HYPERBOLIC_MO
12
  # Create and launch the interface using get_app utility
13
  demo = get_app(
14
  models=HYPERBOLIC_MODELS_FULL, # Use the full names with prefix
15
- default_model=HYPERBOLIC_MODELS_FULL[-2],
16
  dropdown_label="Select Hyperbolic Model",
17
  choices=HYPERBOLIC_MODELS_DISPLAY, # Display names without prefix
18
  fill_height=True,
 
12
  # Create and launch the interface using get_app utility
13
  demo = get_app(
14
  models=HYPERBOLIC_MODELS_FULL, # Use the full names with prefix
15
+ default_model=HYPERBOLIC_MODELS_FULL[-1],
16
  dropdown_label="Select Hyperbolic Model",
17
  choices=HYPERBOLIC_MODELS_DISPLAY, # Display names without prefix
18
  fill_height=True,
app_nvidia_coder.py DELETED
@@ -1,23 +0,0 @@
1
- import ai_gradio
2
-
3
- from utils_ai_gradio import get_app
4
-
5
- # Get the nvidia models but keep their full names for loading
6
- NVIDIA_MODELS_FULL = [k for k in ai_gradio.registry.keys() if k.startswith("nvidia:")]
7
-
8
- # Create display names without the prefix
9
- NVIDIA_MODELS_DISPLAY = [k.replace("nvidia:", "") for k in NVIDIA_MODELS_FULL]
10
-
11
-
12
- # Create and launch the interface using get_app utility
13
- demo = get_app(
14
- models=NVIDIA_MODELS_FULL, # Use the full names with prefix
15
- default_model=NVIDIA_MODELS_FULL[-1],
16
- dropdown_label="Select Nvidia Model",
17
- choices=NVIDIA_MODELS_DISPLAY, # Display names without prefix
18
- fill_height=True,
19
- coder=True
20
- )
21
-
22
- if __name__ == "__main__":
23
- demo.launch()
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
app_openai_coder.py CHANGED
@@ -11,7 +11,7 @@ OPENAI_MODELS_DISPLAY = [k.replace("openai:", "") for k in OPENAI_MODELS_FULL]
11
  # Create and launch the interface using get_app utility
12
  demo = get_app(
13
  models=OPENAI_MODELS_FULL, # Use the full names with prefix
14
- default_model=OPENAI_MODELS_FULL[-1],
15
  dropdown_label="Select OpenAI Model",
16
  choices=OPENAI_MODELS_DISPLAY, # Display names without prefix
17
  fill_height=True,
 
11
  # Create and launch the interface using get_app utility
12
  demo = get_app(
13
  models=OPENAI_MODELS_FULL, # Use the full names with prefix
14
+ default_model=OPENAI_MODELS_FULL[5],
15
  dropdown_label="Select OpenAI Model",
16
  choices=OPENAI_MODELS_DISPLAY, # Display names without prefix
17
  fill_height=True,
app_openrouter.py DELETED
@@ -1,22 +0,0 @@
1
- import ai_gradio
2
-
3
- from utils_ai_gradio import get_app
4
-
5
- # Get the OpenAI models but keep their full names for loading
6
- OPENROUTER_MODELS_FULL = [k for k in ai_gradio.registry.keys() if k.startswith("openrouter:")]
7
-
8
- # Create display names without the prefix
9
- OPENROUTER_MODELS_DISPLAY = [k.replace("openrouter:", "") for k in OPENROUTER_MODELS_FULL]
10
-
11
- # Create and launch the interface using get_app utility
12
- demo = get_app(
13
- models=OPENROUTER_MODELS_FULL, # Use the full names with prefix
14
- default_model=OPENROUTER_MODELS_FULL[-1],
15
- dropdown_label="Select OpenRouter Model",
16
- choices=OPENROUTER_MODELS_DISPLAY, # Display names without prefix
17
- fill_height=True,
18
- coder=True,
19
- )
20
-
21
- if __name__ == "__main__":
22
- demo.launch()
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
app_qwen.py CHANGED
@@ -1,19 +1,27 @@
1
- import ai_gradio
2
 
3
- from utils_ai_gradio import get_app
4
 
5
- # Get the qwen models but keep their full names for loading
6
- QWEN_MODELS_FULL = [k for k in ai_gradio.registry.keys() if k.startswith("qwen:")]
7
 
8
- # Create display names without the prefix
9
- QWEN_MODELS_DISPLAY = [k.replace("qwen:", "") for k in QWEN_MODELS_FULL]
10
-
11
-
12
- # Create and launch the interface using get_app utility
13
  demo = get_app(
14
- models=QWEN_MODELS_FULL, # Use the full names with prefix
15
- default_model=QWEN_MODELS_FULL[-1],
16
- dropdown_label="Select Qwen Model",
17
- choices=QWEN_MODELS_DISPLAY, # Display names without prefix
18
- fill_height=True,
 
 
 
 
 
 
 
 
 
 
 
19
  )
 
 
 
 
1
+ import os
2
 
3
+ import dashscope_gradio
4
 
5
+ from utils import get_app
 
6
 
 
 
 
 
 
7
  demo = get_app(
8
+ models=[
9
+ "qwen-turbo-latest",
10
+ "qwen-turbo",
11
+ "qwen-plus",
12
+ "qwen-max",
13
+ "qwen1.5-110b-chat",
14
+ "qwen1.5-72b-chat",
15
+ "qwen1.5-32b-chat",
16
+ "qwen1.5-14b-chat",
17
+ "qwen1.5-7b-chat",
18
+ "qwq-32b-preview",
19
+ "qvq-72b-preview",
20
+ ],
21
+ default_model="qvq-72b-preview",
22
+ src=dashscope_gradio.registry,
23
+ accept_token=not os.getenv("DASHSCOPE_API_KEY"),
24
  )
25
+
26
+ if __name__ == "__main__":
27
+ demo.launch()
app_qwen_coder.py DELETED
@@ -1,20 +0,0 @@
1
- import ai_gradio
2
-
3
- from utils_ai_gradio import get_app
4
-
5
- # Get the qwen models but keep their full names for loading
6
- QWEN_MODELS_FULL = [k for k in ai_gradio.registry.keys() if k.startswith("qwen:")]
7
-
8
- # Create display names without the prefix
9
- QWEN_MODELS_DISPLAY = [k.replace("qwen:", "") for k in QWEN_MODELS_FULL]
10
-
11
-
12
- # Create and launch the interface using get_app utility
13
- demo = get_app(
14
- models=QWEN_MODELS_FULL, # Use the full names with prefix
15
- default_model=QWEN_MODELS_FULL[-1],
16
- dropdown_label="Select Qwen Model",
17
- choices=QWEN_MODELS_DISPLAY, # Display names without prefix
18
- fill_height=True,
19
- coder=True,
20
- )
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
app_sambanova.py CHANGED
@@ -1,19 +1,28 @@
1
- import ai_gradio
2
 
3
- from utils_ai_gradio import get_app
4
 
5
- # Get the hyperbolic models but keep their full names for loading
6
- SAMBANOVA_MODELS_FULL = [k for k in ai_gradio.registry.keys() if k.startswith("sambanova:")]
7
 
8
- # Create display names without the prefix
9
- SAMBANOVA_MODELS_DISPLAY = [k.replace("sambanova:", "") for k in SAMBANOVA_MODELS_FULL]
10
-
11
-
12
- # Create and launch the interface using get_app utility
13
  demo = get_app(
14
- models=SAMBANOVA_MODELS_FULL, # Use the full names with prefix
15
- default_model=SAMBANOVA_MODELS_FULL[-1],
16
- dropdown_label="Select Sambanova Model",
17
- choices=SAMBANOVA_MODELS_DISPLAY, # Display names without prefix
18
- fill_height=True,
19
- )
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os
2
 
3
+ import sambanova_gradio
4
 
5
+ from utils import get_app
 
6
 
 
 
 
 
 
7
  demo = get_app(
8
+ models=[
9
+ "Meta-Llama-3.2-1B-Instruct",
10
+ "Meta-Llama-3.2-3B-Instruct",
11
+ "Llama-3.2-11B-Vision-Instruct",
12
+ "Llama-3.2-90B-Vision-Instruct",
13
+ "Meta-Llama-3.1-8B-Instruct",
14
+ "Meta-Llama-3.1-70B-Instruct",
15
+ "Meta-Llama-3.1-405B-Instruct",
16
+ "Qwen2.5-72B-Instruct",
17
+ "Qwen2.5-Coder-32B-Instruct",
18
+ "Meta-Llama-3.3-70B-Instruct",
19
+ "QwQ-32B-Preview",
20
+ ],
21
+ default_model="QwQ-32B-Preview",
22
+ src=sambanova_gradio.registry,
23
+ accept_token=not os.getenv("SAMBANOVA_API_KEY"),
24
+ multimodal=True,
25
+ )
26
+
27
+ if __name__ == "__main__":
28
+ demo.launch()
app_sambanova_coder.py DELETED
@@ -1,20 +0,0 @@
1
- import ai_gradio
2
-
3
- from utils_ai_gradio import get_app
4
-
5
- # Get the hyperbolic models but keep their full names for loading
6
- SAMBANOVA_MODELS_FULL = [k for k in ai_gradio.registry.keys() if k.startswith("sambanova:")]
7
-
8
- # Create display names without the prefix
9
- SAMBANOVA_MODELS_DISPLAY = [k.replace("sambanova:", "") for k in SAMBANOVA_MODELS_FULL]
10
-
11
-
12
- # Create and launch the interface using get_app utility
13
- demo = get_app(
14
- models=SAMBANOVA_MODELS_FULL, # Use the full names with prefix
15
- default_model=SAMBANOVA_MODELS_FULL[-1],
16
- dropdown_label="Select Sambanova Model",
17
- choices=SAMBANOVA_MODELS_DISPLAY, # Display names without prefix
18
- fill_height=True,
19
- coder=True,
20
- )
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
pyproject.toml CHANGED
@@ -5,11 +5,54 @@ description = ""
5
  readme = "README.md"
6
  requires-python = ">=3.10"
7
  dependencies = [
8
- "ai-gradio[huggingface]==0.2.54",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  ]
10
 
11
  [tool.uv.sources]
12
-
 
 
 
 
 
 
 
 
 
 
13
 
14
  [tool.ruff]
15
  line-length = 119
 
5
  readme = "README.md"
6
  requires-python = ">=3.10"
7
  dependencies = [
8
+ "anthropic-gradio",
9
+ "cerebras-gradio",
10
+ "dashscope-gradio",
11
+ "fal-gradio",
12
+ "fireworks-gradio",
13
+ "gemini-gradio>=0.0.3",
14
+ "gradio-webrtc>=0.0.28",
15
+ "gradio>=5.12.0",
16
+ "groq-gradio>=0.0.2",
17
+ "hyperbolic-gradio>=0.0.5",
18
+ "mistral-gradio",
19
+ "nvidia-gradio",
20
+ "openai-gradio>=0.0.6",
21
+ "perplexity-gradio>=0.0.1",
22
+ "python-dotenv>=1.0.1",
23
+ "replicate-gradio",
24
+ "sambanova-gradio>=0.1.9",
25
+ "together-gradio>=0.0.1",
26
+ "websockets>=14.1",
27
+ "xai-gradio>=0.0.2",
28
+ "numba>=0.60.0",
29
+ "llvmlite>=0.43.0",
30
+ "librosa>=0.10.2.post1",
31
+ "twilio>=9.4.2",
32
+ "playai-gradio",
33
+ "lumaai-gradio",
34
+ "cohere-gradio",
35
+ "crewai>=0.95.0",
36
+ "instructor>=1.7.2",
37
+ "crewai-tools>=0.25.8",
38
+ "langchain>=0.3.14",
39
+ "chromadb>=0.5.23",
40
+ "openai>=1.55.0",
41
+ "ai-gradio[crewai,deepseek,gemini,groq,hyperbolic,openai,smolagents,transformers, langchain, mistral,minimax,nvidia]>=0.2.28",
42
  ]
43
 
44
  [tool.uv.sources]
45
+ anthropic-gradio = { git = "https://github.com/AK391/anthropic-gradio.git" }
46
+ fireworks-gradio = { git = "https://github.com/AK391/fireworks-ai-gradio.git" }
47
+ cerebras-gradio = { git = "https://github.com/gradio-app/cerebras_gradio.git" }
48
+ nvidia-gradio = { git = "https://github.com/AK391/nvidia-gradio.git" }
49
+ dashscope-gradio = { git = "https://github.com/AK391/dashscope-gradio.git" }
50
+ fal-gradio = { git = "https://github.com/AK391/fal-gradio.git" }
51
+ replicate-gradio = { git = "https://github.com/AK391/replicate-gradio.git" }
52
+ mistral-gradio = { git = "https://github.com/AK391/mistral-gradio.git" }
53
+ playai-gradio = { git = "https://github.com/AK391/playai-gradio.git" }
54
+ lumaai-gradio = { git = "https://github.com/AK391/lumaai-gradio.git" }
55
+ cohere-gradio = { git = "https://github.com/AK391/cohere-gradio.git" }
56
 
57
  [tool.ruff]
58
  line-length = 119
requirements.txt CHANGED
@@ -1,332 +1,1289 @@
1
  # This file was autogenerated by uv via the following command:
2
  # uv pip compile pyproject.toml -o requirements.txt
3
- accelerate==1.5.2
4
  # via ai-gradio
5
- ai-gradio==0.2.54
6
  # via anychat (pyproject.toml)
7
  aiofiles==23.2.1
8
  # via gradio
9
- aiohappyeyeballs==2.6.1
10
  # via aiohttp
11
- aiohttp==3.11.14
12
  # via
13
  # aiohttp-retry
 
 
 
 
 
 
 
14
  # twilio
15
- aiohttp-retry==2.9.1
16
  # via twilio
17
  aioice==0.9.0
18
  # via aiortc
19
- aiortc==1.10.1
20
  # via gradio-webrtc
21
  aiosignal==1.3.2
22
  # via aiohttp
 
 
23
  annotated-types==0.7.0
24
  # via pydantic
25
- anyio==4.9.0
 
 
 
 
26
  # via
 
 
27
  # gradio
 
28
  # httpx
 
 
 
29
  # starlette
30
- async-timeout==5.0.1
31
- # via aiohttp
32
- attrs==25.3.0
33
- # via aiohttp
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
34
  audioread==3.0.1
35
  # via librosa
36
- av==13.1.0
 
 
 
 
37
  # via aiortc
 
 
 
 
 
 
 
 
 
 
38
  bitsandbytes==0.42.0
39
  # via ai-gradio
40
- certifi==2025.1.31
 
 
 
 
 
 
 
 
 
 
 
 
41
  # via
42
  # httpcore
43
  # httpx
 
44
  # requests
 
45
  cffi==1.17.1
46
  # via
47
  # aiortc
48
  # cryptography
49
  # pylibsrtp
 
 
50
  # soundfile
51
  charset-normalizer==3.4.1
52
- # via requests
 
 
 
 
 
 
 
 
 
 
53
  click==8.1.8
54
  # via
 
 
 
55
  # typer
56
  # uvicorn
 
 
 
 
 
 
 
57
  coloredlogs==15.0.1
58
  # via onnxruntime
59
- cryptography==44.0.2
 
 
 
 
 
 
 
 
 
60
  # via
61
  # aiortc
 
 
 
62
  # pyopenssl
63
- decorator==5.2.1
64
- # via librosa
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
65
  dnspython==2.7.0
66
  # via aioice
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
67
  exceptiongroup==1.2.2
68
- # via anyio
69
- fastapi==0.115.12
70
- # via gradio
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
71
  ffmpy==0.5.0
72
  # via gradio
73
- filelock==3.18.0
74
  # via
75
  # huggingface-hub
76
  # torch
77
- flatbuffers==25.2.10
 
 
 
 
 
78
  # via onnxruntime
79
  frozenlist==1.5.0
80
  # via
81
  # aiohttp
82
  # aiosignal
83
- fsspec==2025.3.0
84
  # via
85
  # gradio-client
86
  # huggingface-hub
87
  # torch
88
- google-crc32c==1.7.0
89
- # via aiortc
90
- gradio==5.23.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
91
  # via
 
92
  # ai-gradio
 
 
 
 
 
 
 
93
  # gradio-webrtc
 
 
 
 
94
  # modelscope-studio
95
- gradio-client==1.8.0
96
- # via gradio
97
- gradio-webrtc==0.0.31
98
- # via ai-gradio
99
- groovy==0.1.2
 
 
 
 
 
100
  # via gradio
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
101
  h11==0.14.0
102
  # via
103
  # httpcore
104
  # uvicorn
105
- httpcore==1.0.7
 
106
  # via httpx
107
- httpx==0.28.1
 
 
 
 
 
 
 
 
 
 
 
 
 
108
  # via
 
 
 
 
 
 
 
 
109
  # gradio
110
  # gradio-client
 
 
 
 
 
 
 
 
 
 
111
  # safehttpx
112
- huggingface-hub==0.29.3
 
 
 
 
 
 
 
 
 
 
113
  # via
114
  # accelerate
115
- # ai-gradio
116
  # gradio
117
  # gradio-client
 
 
118
  humanfriendly==10.0
119
  # via coloredlogs
 
 
 
 
120
  idna==3.10
121
  # via
122
  # anyio
123
  # httpx
124
  # requests
 
125
  # yarl
126
  ifaddr==0.2.0
127
  # via aioice
128
- jinja2==3.1.6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
129
  # via
130
  # gradio
 
 
 
 
131
  # torch
 
 
 
 
 
132
  joblib==1.4.2
133
  # via
134
  # librosa
135
  # scikit-learn
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
136
  lazy-loader==0.4
137
  # via librosa
138
- librosa==0.11.0
139
  # via
 
140
  # ai-gradio
141
  # gradio-webrtc
 
 
 
 
142
  llvmlite==0.43.0
143
- # via numba
 
 
 
 
 
 
 
 
 
 
144
  markdown-it-py==3.0.0
145
  # via rich
146
- markupsafe==3.0.2
 
 
147
  # via
148
  # gradio
149
  # jinja2
 
 
 
 
 
150
  mdurl==0.1.2
151
  # via markdown-it-py
152
- modelscope-studio==1.2.1
 
 
 
 
 
 
 
 
 
 
153
  # via ai-gradio
 
 
154
  mpmath==1.3.0
155
  # via sympy
156
  msgpack==1.1.0
157
  # via librosa
158
- multidict==6.2.0
159
  # via
160
  # aiohttp
161
  # yarl
 
 
162
  networkx==3.4.2
163
- # via torch
 
 
 
 
164
  numba==0.60.0
165
  # via
 
166
  # ai-gradio
167
  # librosa
168
- numpy==2.0.2
 
169
  # via
170
  # accelerate
171
  # ai-gradio
 
 
 
 
172
  # gradio
 
 
173
  # librosa
174
  # numba
175
  # onnxruntime
 
176
  # opencv-python
177
  # pandas
 
 
178
  # scikit-learn
179
  # scipy
 
180
  # soundfile
181
  # soxr
182
- onnxruntime==1.21.0
183
- # via gradio-webrtc
184
- opencv-python==4.11.0.86
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
185
  # via ai-gradio
186
- orjson==3.10.16
187
- # via gradio
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
188
  packaging==24.2
189
  # via
190
  # accelerate
 
 
 
 
 
191
  # gradio
192
  # gradio-client
193
  # huggingface-hub
 
 
194
  # lazy-loader
 
195
  # onnxruntime
 
196
  # pooch
 
 
 
197
  pandas==2.2.3
198
- # via gradio
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
199
  pillow==11.1.0
200
  # via
201
  # ai-gradio
 
 
202
  # gradio
203
- platformdirs==4.3.7
 
 
 
 
 
204
  # via pooch
 
 
 
 
205
  pooch==1.8.2
206
  # via librosa
207
- propcache==0.3.0
 
 
 
 
 
 
 
 
 
 
 
208
  # via
209
  # aiohttp
210
  # yarl
211
- protobuf==6.30.1
212
- # via onnxruntime
213
- psutil==7.0.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
214
  # via accelerate
 
 
 
 
 
 
 
 
 
 
 
 
215
  pycparser==2.22
216
  # via cffi
217
- pydantic==2.10.6
218
  # via
 
 
 
 
 
 
219
  # fastapi
 
 
 
 
220
  # gradio
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
221
  pydantic-core==2.27.2
222
- # via pydantic
 
 
 
 
 
223
  pydub==0.25.1
224
  # via
225
  # ai-gradio
226
  # gradio
227
- pyee==13.0.0
 
228
  # via aiortc
229
  pygments==2.19.1
230
- # via rich
 
 
231
  pyjwt==2.10.1
232
- # via twilio
233
- pylibsrtp==0.11.0
 
 
 
 
234
  # via aiortc
235
- pyopenssl==25.0.0
236
  # via aiortc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
237
  python-dateutil==2.9.0.post0
238
- # via pandas
 
 
 
 
 
 
239
  python-dotenv==1.0.1
240
- # via ai-gradio
 
 
 
 
 
 
 
 
 
 
241
  python-multipart==0.0.20
242
  # via gradio
243
- pytz==2025.1
244
- # via pandas
 
 
 
 
 
 
 
 
 
 
245
  pyyaml==6.0.2
246
  # via
247
  # accelerate
 
248
  # gradio
249
  # huggingface-hub
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
250
  requests==2.32.3
251
  # via
 
 
 
 
 
 
 
 
 
 
252
  # huggingface-hub
 
 
 
 
 
 
253
  # pooch
 
 
 
 
 
 
 
 
 
 
254
  # twilio
 
 
 
 
255
  rich==13.9.4
256
- # via typer
257
- ruff==0.11.2
 
 
 
 
 
 
 
 
 
 
 
258
  # via gradio
259
  safehttpx==0.1.6
260
  # via gradio
261
- safetensors==0.5.3
262
- # via accelerate
 
 
 
 
 
 
263
  scikit-learn==1.6.1
264
  # via librosa
265
- scipy==1.15.2
266
  # via
267
  # bitsandbytes
268
  # librosa
269
  # scikit-learn
 
 
 
 
270
  semantic-version==2.10.0
271
  # via gradio
 
 
 
 
 
 
272
  shellingham==1.5.4
273
  # via typer
274
  six==1.17.0
275
- # via python-dateutil
 
 
 
 
 
 
276
  sniffio==1.3.1
277
- # via anyio
278
- soundfile==0.13.1
 
 
 
 
 
 
 
 
 
 
279
  # via librosa
 
 
280
  soxr==0.5.0.post1
281
  # via librosa
282
- starlette==0.46.1
 
 
 
 
 
 
 
 
 
 
 
283
  # via
284
  # fastapi
285
  # gradio
286
- sympy==1.13.1
287
  # via
288
  # onnxruntime
289
  # torch
290
- threadpoolctl==3.6.0
 
 
 
 
 
 
 
 
 
 
 
 
291
  # via scikit-learn
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
292
  tomlkit==0.13.2
293
  # via gradio
294
- torch==2.6.0
295
  # via
296
  # accelerate
297
  # ai-gradio
298
- tqdm==4.67.1
299
- # via huggingface-hub
300
- twilio==9.5.1
301
  # via ai-gradio
302
- typer==0.15.2
303
- # via gradio
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
304
  typing-extensions==4.12.2
305
  # via
 
 
306
  # anyio
 
 
 
 
 
307
  # fastapi
 
 
308
  # gradio
309
  # gradio-client
 
310
  # huggingface-hub
 
 
311
  # librosa
 
312
  # multidict
 
 
313
  # pydantic
314
  # pydantic-core
315
  # pyee
316
- # pyopenssl
 
 
317
  # rich
 
 
318
  # torch
319
  # typer
 
320
  # uvicorn
321
- tzdata==2025.2
 
 
 
 
322
  # via pandas
 
 
323
  urllib3==2.3.0
324
- # via requests
 
 
 
 
 
 
 
 
 
325
  uvicorn==0.34.0
326
- # via gradio
327
- websockets==15.0.1
328
  # via
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
329
  # ai-gradio
 
 
330
  # gradio-client
 
 
 
 
 
 
 
 
 
 
 
331
  yarl==1.18.3
332
  # via aiohttp
 
 
 
1
  # This file was autogenerated by uv via the following command:
2
  # uv pip compile pyproject.toml -o requirements.txt
3
+ accelerate==1.2.1
4
  # via ai-gradio
5
+ ai-gradio==0.2.28
6
  # via anychat (pyproject.toml)
7
  aiofiles==23.2.1
8
  # via gradio
9
+ aiohappyeyeballs==2.4.4
10
  # via aiohttp
11
+ aiohttp==3.11.11
12
  # via
13
  # aiohttp-retry
14
+ # auth0-python
15
+ # instructor
16
+ # langchain
17
+ # langchain-community
18
+ # litellm
19
+ # scrapegraph-py
20
+ # spider-client
21
  # twilio
22
+ aiohttp-retry==2.8.3
23
  # via twilio
24
  aioice==0.9.0
25
  # via aiortc
26
+ aiortc==1.9.0
27
  # via gradio-webrtc
28
  aiosignal==1.3.2
29
  # via aiohttp
30
+ alembic==1.14.0
31
+ # via embedchain
32
  annotated-types==0.7.0
33
  # via pydantic
34
+ anthropic==0.42.0
35
+ # via anthropic-gradio
36
+ anthropic-gradio @ git+https://github.com/AK391/anthropic-gradio.git@34e5622031487ea207073177b4ec1f3067936295
37
+ # via anychat (pyproject.toml)
38
+ anyio==4.8.0
39
  # via
40
+ # anthropic
41
+ # cerebras-cloud-sdk
42
  # gradio
43
+ # groq
44
  # httpx
45
+ # httpx-ws
46
+ # lumaai
47
+ # openai
48
  # starlette
49
+ # watchfiles
50
+ appdirs==1.4.4
51
+ # via crewai
52
+ asgiref==3.8.1
53
+ # via opentelemetry-instrumentation-asgi
54
+ asttokens==3.0.0
55
+ # via stack-data
56
+ async-timeout==4.0.3
57
+ # via
58
+ # aiohttp
59
+ # langchain
60
+ attrs==24.3.0
61
+ # via
62
+ # aiohttp
63
+ # e2b
64
+ # e2b-code-interpreter
65
+ # jsonschema
66
+ # outcome
67
+ # referencing
68
+ # trio
69
  audioread==3.0.1
70
  # via librosa
71
+ auth0-python==4.7.2
72
+ # via crewai
73
+ authlib==1.3.1
74
+ # via weaviate-client
75
+ av==12.3.0
76
  # via aiortc
77
+ backoff==2.2.1
78
+ # via posthog
79
+ bcrypt==4.2.1
80
+ # via chromadb
81
+ beautifulsoup4==4.12.3
82
+ # via
83
+ # crewai-tools
84
+ # embedchain
85
+ # markdownify
86
+ # scrapegraph-py
87
  bitsandbytes==0.42.0
88
  # via ai-gradio
89
+ blinker==1.9.0
90
+ # via crewai
91
+ build==1.2.2.post1
92
+ # via chromadb
93
+ cachetools==5.5.0
94
+ # via
95
+ # google-auth
96
+ # gptcache
97
+ cerebras-cloud-sdk==1.16.0
98
+ # via cerebras-gradio
99
+ cerebras-gradio @ git+https://github.com/gradio-app/cerebras_gradio.git@0135ea37048281ca4fae2d6a90311be3b38954ae
100
+ # via anychat (pyproject.toml)
101
+ certifi==2024.12.14
102
  # via
103
  # httpcore
104
  # httpx
105
+ # kubernetes
106
  # requests
107
+ # selenium
108
  cffi==1.17.1
109
  # via
110
  # aiortc
111
  # cryptography
112
  # pylibsrtp
113
+ # pyvips
114
+ # pyvips-binary
115
  # soundfile
116
  charset-normalizer==3.4.1
117
+ # via
118
+ # pdfminer-six
119
+ # requests
120
+ chroma-hnswlib==0.7.6
121
+ # via chromadb
122
+ chromadb==0.5.23
123
+ # via
124
+ # anychat (pyproject.toml)
125
+ # crewai
126
+ # crewai-tools
127
+ # embedchain
128
  click==8.1.8
129
  # via
130
+ # crewai
131
+ # duckduckgo-search
132
+ # litellm
133
  # typer
134
  # uvicorn
135
+ cohere==5.13.6
136
+ # via
137
+ # cohere-gradio
138
+ # embedchain
139
+ # langchain-cohere
140
+ cohere-gradio @ git+https://github.com/AK391/cohere-gradio.git@f514496532579574b9bdd55c07e2eb14f2ddf1b0
141
+ # via anychat (pyproject.toml)
142
  coloredlogs==15.0.1
143
  # via onnxruntime
144
+ crewai==0.95.0
145
+ # via
146
+ # anychat (pyproject.toml)
147
+ # ai-gradio
148
+ # crewai-tools
149
+ crewai-tools==0.25.8
150
+ # via
151
+ # anychat (pyproject.toml)
152
+ # ai-gradio
153
+ cryptography==43.0.3
154
  # via
155
  # aiortc
156
+ # auth0-python
157
+ # authlib
158
+ # pdfminer-six
159
  # pyopenssl
160
+ dashscope-gradio @ git+https://github.com/AK391/dashscope-gradio.git@afe2a389fd194e71ca4c12e6162f3f0fd3cf534a
161
+ # via anychat (pyproject.toml)
162
+ dataclasses-json==0.6.7
163
+ # via langchain-community
164
+ decorator==5.1.1
165
+ # via
166
+ # ipython
167
+ # librosa
168
+ deprecated==1.2.15
169
+ # via
170
+ # opentelemetry-api
171
+ # opentelemetry-exporter-otlp-proto-grpc
172
+ # opentelemetry-exporter-otlp-proto-http
173
+ # opentelemetry-semantic-conventions
174
+ deprecation==2.1.0
175
+ # via lancedb
176
+ distro==1.9.0
177
+ # via
178
+ # anthropic
179
+ # cerebras-cloud-sdk
180
+ # groq
181
+ # lumaai
182
+ # openai
183
  dnspython==2.7.0
184
  # via aioice
185
+ docker==7.1.0
186
+ # via crewai-tools
187
+ docstring-parser==0.16
188
+ # via
189
+ # google-cloud-aiplatform
190
+ # instructor
191
+ docx2txt==0.8
192
+ # via crewai-tools
193
+ duckduckgo-search==7.2.1
194
+ # via smolagents
195
+ durationpy==0.9
196
+ # via kubernetes
197
+ e2b==1.0.5
198
+ # via e2b-code-interpreter
199
+ e2b-code-interpreter==1.0.3
200
+ # via smolagents
201
+ einops==0.8.0
202
+ # via ai-gradio
203
+ embedchain==0.1.126
204
+ # via crewai-tools
205
+ et-xmlfile==2.0.0
206
+ # via openpyxl
207
+ eval-type-backport==0.2.2
208
+ # via mistralai
209
  exceptiongroup==1.2.2
210
+ # via
211
+ # anyio
212
+ # ipython
213
+ # pytest
214
+ # trio
215
+ # trio-websocket
216
+ executing==2.1.0
217
+ # via stack-data
218
+ fal-client==0.5.6
219
+ # via fal-gradio
220
+ fal-gradio @ git+https://github.com/AK391/fal-gradio.git@f4d47158564df06ebb4dc8782dad676c4c539e47
221
+ # via anychat (pyproject.toml)
222
+ fastapi==0.115.6
223
+ # via
224
+ # chromadb
225
+ # gradio
226
+ fastavro==1.10.0
227
+ # via cohere
228
  ffmpy==0.5.0
229
  # via gradio
230
+ filelock==3.16.1
231
  # via
232
  # huggingface-hub
233
  # torch
234
+ # transformers
235
+ fireworks-ai==0.15.11
236
+ # via fireworks-gradio
237
+ fireworks-gradio @ git+https://github.com/AK391/fireworks-ai-gradio.git@5805959ebe511052c417569f056139b4de350a57
238
+ # via anychat (pyproject.toml)
239
+ flatbuffers==24.12.23
240
  # via onnxruntime
241
  frozenlist==1.5.0
242
  # via
243
  # aiohttp
244
  # aiosignal
245
+ fsspec==2024.12.0
246
  # via
247
  # gradio-client
248
  # huggingface-hub
249
  # torch
250
+ gemini-gradio==0.0.3
251
+ # via anychat (pyproject.toml)
252
+ google-ai-generativelanguage==0.6.10
253
+ # via google-generativeai
254
+ google-api-core==2.24.0
255
+ # via
256
+ # google-ai-generativelanguage
257
+ # google-api-python-client
258
+ # google-cloud-aiplatform
259
+ # google-cloud-bigquery
260
+ # google-cloud-core
261
+ # google-cloud-resource-manager
262
+ # google-cloud-storage
263
+ # google-generativeai
264
+ google-api-python-client==2.158.0
265
+ # via google-generativeai
266
+ google-auth==2.37.0
267
+ # via
268
+ # google-ai-generativelanguage
269
+ # google-api-core
270
+ # google-api-python-client
271
+ # google-auth-httplib2
272
+ # google-cloud-aiplatform
273
+ # google-cloud-bigquery
274
+ # google-cloud-core
275
+ # google-cloud-resource-manager
276
+ # google-cloud-storage
277
+ # google-genai
278
+ # google-generativeai
279
+ # kubernetes
280
+ google-auth-httplib2==0.2.0
281
+ # via google-api-python-client
282
+ google-cloud-aiplatform==1.76.0
283
+ # via embedchain
284
+ google-cloud-bigquery==3.27.0
285
+ # via google-cloud-aiplatform
286
+ google-cloud-core==2.4.1
287
+ # via
288
+ # google-cloud-bigquery
289
+ # google-cloud-storage
290
+ google-cloud-resource-manager==1.14.0
291
+ # via google-cloud-aiplatform
292
+ google-cloud-storage==2.19.0
293
+ # via google-cloud-aiplatform
294
+ google-crc32c==1.6.0
295
+ # via
296
+ # aiortc
297
+ # google-cloud-storage
298
+ # google-resumable-media
299
+ google-genai==0.3.0
300
+ # via ai-gradio
301
+ google-generativeai==0.8.3
302
+ # via
303
+ # ai-gradio
304
+ # gemini-gradio
305
+ google-resumable-media==2.7.2
306
+ # via
307
+ # google-cloud-bigquery
308
+ # google-cloud-storage
309
+ googleapis-common-protos==1.66.0
310
+ # via
311
+ # google-api-core
312
+ # grpc-google-iam-v1
313
+ # grpcio-status
314
+ # opentelemetry-exporter-otlp-proto-grpc
315
+ # opentelemetry-exporter-otlp-proto-http
316
+ gptcache==0.1.44
317
+ # via embedchain
318
+ gradio==5.12.0
319
  # via
320
+ # anychat (pyproject.toml)
321
  # ai-gradio
322
+ # anthropic-gradio
323
+ # cerebras-gradio
324
+ # cohere-gradio
325
+ # dashscope-gradio
326
+ # fal-gradio
327
+ # fireworks-gradio
328
+ # gemini-gradio
329
  # gradio-webrtc
330
+ # groq-gradio
331
+ # hyperbolic-gradio
332
+ # lumaai-gradio
333
+ # mistral-gradio
334
  # modelscope-studio
335
+ # nvidia-gradio
336
+ # openai-gradio
337
+ # perplexity-gradio
338
+ # playai-gradio
339
+ # replicate-gradio
340
+ # sambanova-gradio
341
+ # smolagents
342
+ # together-gradio
343
+ # xai-gradio
344
+ gradio-client==1.5.4
345
  # via gradio
346
+ gradio-webrtc==0.0.28
347
+ # via
348
+ # anychat (pyproject.toml)
349
+ # ai-gradio
350
+ # gemini-gradio
351
+ # openai-gradio
352
+ groq==0.15.0
353
+ # via groq-gradio
354
+ groq-gradio==0.0.2
355
+ # via anychat (pyproject.toml)
356
+ grpc-google-iam-v1==0.14.0
357
+ # via google-cloud-resource-manager
358
+ grpcio==1.69.0
359
+ # via
360
+ # chromadb
361
+ # google-api-core
362
+ # googleapis-common-protos
363
+ # grpc-google-iam-v1
364
+ # grpcio-health-checking
365
+ # grpcio-status
366
+ # grpcio-tools
367
+ # opentelemetry-exporter-otlp-proto-grpc
368
+ # qdrant-client
369
+ # weaviate-client
370
+ grpcio-health-checking==1.69.0
371
+ # via weaviate-client
372
+ grpcio-status==1.69.0
373
+ # via google-api-core
374
+ grpcio-tools==1.69.0
375
+ # via
376
+ # qdrant-client
377
+ # weaviate-client
378
  h11==0.14.0
379
  # via
380
  # httpcore
381
  # uvicorn
382
+ # wsproto
383
+ h2==4.1.0
384
  # via httpx
385
+ hpack==4.0.0
386
+ # via h2
387
+ httpcore==1.0.7
388
+ # via
389
+ # e2b
390
+ # httpx
391
+ # httpx-ws
392
+ httplib2==0.22.0
393
+ # via
394
+ # google-api-python-client
395
+ # google-auth-httplib2
396
+ httptools==0.6.4
397
+ # via uvicorn
398
+ httpx==0.27.2
399
  # via
400
+ # anthropic
401
+ # cerebras-cloud-sdk
402
+ # chromadb
403
+ # cohere
404
+ # e2b
405
+ # e2b-code-interpreter
406
+ # fal-client
407
+ # fireworks-ai
408
  # gradio
409
  # gradio-client
410
+ # groq
411
+ # httpx-ws
412
+ # langsmith
413
+ # linkup-sdk
414
+ # litellm
415
+ # lumaai
416
+ # mistralai
417
+ # openai
418
+ # qdrant-client
419
+ # replicate
420
  # safehttpx
421
+ # tavily-python
422
+ # weaviate-client
423
+ httpx-sse==0.4.0
424
+ # via
425
+ # cohere
426
+ # fal-client
427
+ # fireworks-ai
428
+ # langchain-community
429
+ httpx-ws==0.7.1
430
+ # via fireworks-ai
431
+ huggingface-hub==0.27.1
432
  # via
433
  # accelerate
 
434
  # gradio
435
  # gradio-client
436
+ # tokenizers
437
+ # transformers
438
  humanfriendly==10.0
439
  # via coloredlogs
440
+ hyperbolic-gradio==0.0.5
441
+ # via anychat (pyproject.toml)
442
+ hyperframe==6.0.1
443
+ # via h2
444
  idna==3.10
445
  # via
446
  # anyio
447
  # httpx
448
  # requests
449
+ # trio
450
  # yarl
451
  ifaddr==0.2.0
452
  # via aioice
453
+ ijson==3.3.0
454
+ # via spider-client
455
+ importlib-metadata==8.5.0
456
+ # via
457
+ # litellm
458
+ # opentelemetry-api
459
+ importlib-resources==6.5.2
460
+ # via chromadb
461
+ iniconfig==2.0.0
462
+ # via pytest
463
+ instructor==1.7.2
464
+ # via
465
+ # anychat (pyproject.toml)
466
+ # crewai
467
+ ipython==8.31.0
468
+ # via pyvis
469
+ jedi==0.19.2
470
+ # via ipython
471
+ jinja2==3.1.5
472
  # via
473
  # gradio
474
+ # instructor
475
+ # litellm
476
+ # pyvis
477
+ # smolagents
478
  # torch
479
+ jiter==0.8.2
480
+ # via
481
+ # anthropic
482
+ # instructor
483
+ # openai
484
  joblib==1.4.2
485
  # via
486
  # librosa
487
  # scikit-learn
488
+ json-repair==0.35.0
489
+ # via crewai
490
+ jsonpatch==1.33
491
+ # via langchain-core
492
+ jsonpath-python==1.0.6
493
+ # via mistralai
494
+ jsonpickle==4.0.1
495
+ # via pyvis
496
+ jsonpointer==3.0.0
497
+ # via jsonpatch
498
+ jsonref==1.1.0
499
+ # via crewai
500
+ jsonschema==4.23.0
501
+ # via litellm
502
+ jsonschema-specifications==2024.10.1
503
+ # via jsonschema
504
+ kubernetes==31.0.0
505
+ # via chromadb
506
+ lancedb==0.17.0
507
+ # via crewai-tools
508
+ langchain==0.3.14
509
+ # via
510
+ # anychat (pyproject.toml)
511
+ # ai-gradio
512
+ # embedchain
513
+ # langchain-community
514
+ langchain-cohere==0.3.4
515
+ # via embedchain
516
+ langchain-community==0.3.14
517
+ # via
518
+ # ai-gradio
519
+ # embedchain
520
+ # langchain-experimental
521
+ langchain-core==0.3.29
522
+ # via
523
+ # ai-gradio
524
+ # langchain
525
+ # langchain-cohere
526
+ # langchain-community
527
+ # langchain-experimental
528
+ # langchain-openai
529
+ # langchain-text-splitters
530
+ langchain-experimental==0.3.4
531
+ # via langchain-cohere
532
+ langchain-openai==0.2.14
533
+ # via
534
+ # ai-gradio
535
+ # embedchain
536
+ langchain-text-splitters==0.3.5
537
+ # via langchain
538
+ langsmith==0.1.147
539
+ # via
540
+ # embedchain
541
+ # langchain
542
+ # langchain-community
543
+ # langchain-core
544
  lazy-loader==0.4
545
  # via librosa
546
+ librosa==0.10.2.post1
547
  # via
548
+ # anychat (pyproject.toml)
549
  # ai-gradio
550
  # gradio-webrtc
551
+ linkup-sdk==0.2.2
552
+ # via crewai-tools
553
+ litellm==1.57.7
554
+ # via crewai
555
  llvmlite==0.43.0
556
+ # via
557
+ # anychat (pyproject.toml)
558
+ # numba
559
+ lumaai==1.2.2
560
+ # via lumaai-gradio
561
+ lumaai-gradio @ git+https://github.com/AK391/lumaai-gradio.git@b7856b8e7f1977e3f003a4e101b12eae2c232227
562
+ # via anychat (pyproject.toml)
563
+ lxml==5.3.0
564
+ # via duckduckgo-search
565
+ mako==1.3.8
566
+ # via alembic
567
  markdown-it-py==3.0.0
568
  # via rich
569
+ markdownify==0.14.1
570
+ # via smolagents
571
+ markupsafe==2.1.5
572
  # via
573
  # gradio
574
  # jinja2
575
+ # mako
576
+ marshmallow==3.25.0
577
+ # via dataclasses-json
578
+ matplotlib-inline==0.1.7
579
+ # via ipython
580
  mdurl==0.1.2
581
  # via markdown-it-py
582
+ mem0ai==0.1.41
583
+ # via embedchain
584
+ mistral-gradio @ git+https://github.com/AK391/mistral-gradio.git@dfef7dc871ea35100743a415fde8a57a30c49fcb
585
+ # via anychat (pyproject.toml)
586
+ mistralai==1.2.6
587
+ # via
588
+ # ai-gradio
589
+ # mistral-gradio
590
+ mmh3==5.0.1
591
+ # via chromadb
592
+ modelscope-studio==1.0.2
593
  # via ai-gradio
594
+ monotonic==1.6
595
+ # via posthog
596
  mpmath==1.3.0
597
  # via sympy
598
  msgpack==1.1.0
599
  # via librosa
600
+ multidict==6.1.0
601
  # via
602
  # aiohttp
603
  # yarl
604
+ mypy-extensions==1.0.0
605
+ # via typing-inspect
606
  networkx==3.4.2
607
+ # via
608
+ # pyvis
609
+ # torch
610
+ nodeenv==1.9.1
611
+ # via pyright
612
  numba==0.60.0
613
  # via
614
+ # anychat (pyproject.toml)
615
  # ai-gradio
616
  # librosa
617
+ # openai-gradio
618
+ numpy==1.26.4
619
  # via
620
  # accelerate
621
  # ai-gradio
622
+ # chroma-hnswlib
623
+ # chromadb
624
+ # gemini-gradio
625
+ # gptcache
626
  # gradio
627
+ # langchain
628
+ # langchain-community
629
  # librosa
630
  # numba
631
  # onnxruntime
632
+ # openai-gradio
633
  # opencv-python
634
  # pandas
635
+ # pylance
636
+ # qdrant-client
637
  # scikit-learn
638
  # scipy
639
+ # shapely
640
  # soundfile
641
  # soxr
642
+ # torchvision
643
+ # transformers
644
+ nvidia-gradio @ git+https://github.com/AK391/nvidia-gradio.git@735cc0ba06afb44eeef789d8c0c35b1fc61fee16
645
+ # via anychat (pyproject.toml)
646
+ oauthlib==3.2.2
647
+ # via
648
+ # kubernetes
649
+ # requests-oauthlib
650
+ onnxruntime==1.20.1
651
+ # via
652
+ # chromadb
653
+ # gradio-webrtc
654
+ openai==1.59.6
655
+ # via
656
+ # anychat (pyproject.toml)
657
+ # ai-gradio
658
+ # crewai
659
+ # crewai-tools
660
+ # dashscope-gradio
661
+ # embedchain
662
+ # fireworks-gradio
663
+ # hyperbolic-gradio
664
+ # instructor
665
+ # langchain-openai
666
+ # litellm
667
+ # mem0ai
668
+ # nvidia-gradio
669
+ # openai-gradio
670
+ # perplexity-gradio
671
+ # sambanova-gradio
672
+ # smolagents
673
+ # together-gradio
674
+ # xai-gradio
675
+ openai-gradio==0.0.6
676
+ # via anychat (pyproject.toml)
677
+ opencv-python==4.10.0.84
678
  # via ai-gradio
679
+ openpyxl==3.1.5
680
+ # via crewai
681
+ opentelemetry-api==1.29.0
682
+ # via
683
+ # chromadb
684
+ # crewai
685
+ # opentelemetry-exporter-otlp-proto-grpc
686
+ # opentelemetry-exporter-otlp-proto-http
687
+ # opentelemetry-instrumentation
688
+ # opentelemetry-instrumentation-asgi
689
+ # opentelemetry-instrumentation-fastapi
690
+ # opentelemetry-sdk
691
+ # opentelemetry-semantic-conventions
692
+ opentelemetry-exporter-otlp-proto-common==1.29.0
693
+ # via
694
+ # opentelemetry-exporter-otlp-proto-grpc
695
+ # opentelemetry-exporter-otlp-proto-http
696
+ opentelemetry-exporter-otlp-proto-grpc==1.29.0
697
+ # via chromadb
698
+ opentelemetry-exporter-otlp-proto-http==1.29.0
699
+ # via crewai
700
+ opentelemetry-instrumentation==0.50b0
701
+ # via
702
+ # opentelemetry-instrumentation-asgi
703
+ # opentelemetry-instrumentation-fastapi
704
+ opentelemetry-instrumentation-asgi==0.50b0
705
+ # via opentelemetry-instrumentation-fastapi
706
+ opentelemetry-instrumentation-fastapi==0.50b0
707
+ # via chromadb
708
+ opentelemetry-proto==1.29.0
709
+ # via
710
+ # opentelemetry-exporter-otlp-proto-common
711
+ # opentelemetry-exporter-otlp-proto-grpc
712
+ # opentelemetry-exporter-otlp-proto-http
713
+ opentelemetry-sdk==1.29.0
714
+ # via
715
+ # chromadb
716
+ # crewai
717
+ # opentelemetry-exporter-otlp-proto-grpc
718
+ # opentelemetry-exporter-otlp-proto-http
719
+ opentelemetry-semantic-conventions==0.50b0
720
+ # via
721
+ # opentelemetry-instrumentation
722
+ # opentelemetry-instrumentation-asgi
723
+ # opentelemetry-instrumentation-fastapi
724
+ # opentelemetry-sdk
725
+ opentelemetry-util-http==0.50b0
726
+ # via
727
+ # opentelemetry-instrumentation-asgi
728
+ # opentelemetry-instrumentation-fastapi
729
+ orjson==3.10.14
730
+ # via
731
+ # chromadb
732
+ # gradio
733
+ # langsmith
734
+ outcome==1.3.0.post0
735
+ # via trio
736
+ overrides==7.7.0
737
+ # via
738
+ # chromadb
739
+ # lancedb
740
  packaging==24.2
741
  # via
742
  # accelerate
743
+ # build
744
+ # deprecation
745
+ # e2b
746
+ # google-cloud-aiplatform
747
+ # google-cloud-bigquery
748
  # gradio
749
  # gradio-client
750
  # huggingface-hub
751
+ # lancedb
752
+ # langchain-core
753
  # lazy-loader
754
+ # marshmallow
755
  # onnxruntime
756
+ # opentelemetry-instrumentation
757
  # pooch
758
+ # pytest
759
+ # replicate
760
+ # transformers
761
  pandas==2.2.3
762
+ # via
763
+ # gradio
764
+ # langchain-cohere
765
+ # smolagents
766
+ parameterized==0.9.0
767
+ # via cohere
768
+ parso==0.8.4
769
+ # via jedi
770
+ pdfminer-six==20231228
771
+ # via pdfplumber
772
+ pdfplumber==0.11.5
773
+ # via crewai
774
+ perplexity-gradio==0.0.1
775
+ # via anychat (pyproject.toml)
776
+ pexpect==4.9.0
777
+ # via ipython
778
  pillow==11.1.0
779
  # via
780
  # ai-gradio
781
+ # fireworks-ai
782
+ # google-genai
783
  # gradio
784
+ # pdfplumber
785
+ # smolagents
786
+ # torchvision
787
+ pkgconfig==1.5.5
788
+ # via pyvips
789
+ platformdirs==4.3.6
790
  # via pooch
791
+ playai-gradio @ git+https://github.com/AK391/playai-gradio.git@0958ca289652e231fd12ebbf0f40895a049b0ad1
792
+ # via anychat (pyproject.toml)
793
+ pluggy==1.5.0
794
+ # via pytest
795
  pooch==1.8.2
796
  # via librosa
797
+ portalocker==2.10.1
798
+ # via qdrant-client
799
+ posthog==3.7.5
800
+ # via
801
+ # chromadb
802
+ # embedchain
803
+ # mem0ai
804
+ primp==0.10.0
805
+ # via duckduckgo-search
806
+ prompt-toolkit==3.0.48
807
+ # via ipython
808
+ propcache==0.2.1
809
  # via
810
  # aiohttp
811
  # yarl
812
+ proto-plus==1.25.0
813
+ # via
814
+ # google-ai-generativelanguage
815
+ # google-api-core
816
+ # google-cloud-aiplatform
817
+ # google-cloud-resource-manager
818
+ protobuf==5.29.3
819
+ # via
820
+ # e2b
821
+ # google-ai-generativelanguage
822
+ # google-api-core
823
+ # google-cloud-aiplatform
824
+ # google-cloud-resource-manager
825
+ # google-generativeai
826
+ # googleapis-common-protos
827
+ # grpc-google-iam-v1
828
+ # grpcio-health-checking
829
+ # grpcio-status
830
+ # grpcio-tools
831
+ # onnxruntime
832
+ # opentelemetry-proto
833
+ # proto-plus
834
+ psutil==6.1.1
835
  # via accelerate
836
+ ptyprocess==0.7.0
837
+ # via pexpect
838
+ pure-eval==0.2.3
839
+ # via stack-data
840
+ pyarrow==18.1.0
841
+ # via pylance
842
+ pyasn1==0.6.1
843
+ # via
844
+ # pyasn1-modules
845
+ # rsa
846
+ pyasn1-modules==0.4.1
847
+ # via google-auth
848
  pycparser==2.22
849
  # via cffi
850
+ pydantic==2.10.5
851
  # via
852
+ # anthropic
853
+ # cerebras-cloud-sdk
854
+ # chromadb
855
+ # cohere
856
+ # crewai
857
+ # crewai-tools
858
  # fastapi
859
+ # fireworks-ai
860
+ # google-cloud-aiplatform
861
+ # google-genai
862
+ # google-generativeai
863
  # gradio
864
+ # groq
865
+ # instructor
866
+ # lancedb
867
+ # langchain
868
+ # langchain-cohere
869
+ # langchain-core
870
+ # langsmith
871
+ # linkup-sdk
872
+ # litellm
873
+ # lumaai
874
+ # mem0ai
875
+ # mistralai
876
+ # openai
877
+ # pydantic-settings
878
+ # qdrant-client
879
+ # replicate
880
+ # scrapegraph-py
881
+ # weaviate-client
882
  pydantic-core==2.27.2
883
+ # via
884
+ # cohere
885
+ # instructor
886
+ # pydantic
887
+ pydantic-settings==2.7.1
888
+ # via langchain-community
889
  pydub==0.25.1
890
  # via
891
  # ai-gradio
892
  # gradio
893
+ # openai-gradio
894
+ pyee==12.1.1
895
  # via aiortc
896
  pygments==2.19.1
897
+ # via
898
+ # ipython
899
+ # rich
900
  pyjwt==2.10.1
901
+ # via
902
+ # auth0-python
903
+ # twilio
904
+ pylance==0.20.0
905
+ # via lancedb
906
+ pylibsrtp==0.10.0
907
  # via aiortc
908
+ pyopenssl==24.3.0
909
  # via aiortc
910
+ pyparsing==3.2.1
911
+ # via httplib2
912
+ pypdf==5.1.0
913
+ # via embedchain
914
+ pypdfium2==4.30.1
915
+ # via pdfplumber
916
+ pypika==0.48.9
917
+ # via chromadb
918
+ pyproject-hooks==1.2.0
919
+ # via build
920
+ pyright==1.1.391
921
+ # via crewai-tools
922
+ pysbd==0.3.4
923
+ # via embedchain
924
+ pysocks==1.7.1
925
+ # via urllib3
926
+ pytest==8.3.4
927
+ # via crewai-tools
928
  python-dateutil==2.9.0.post0
929
+ # via
930
+ # e2b
931
+ # google-cloud-bigquery
932
+ # kubernetes
933
+ # mistralai
934
+ # pandas
935
+ # posthog
936
  python-dotenv==1.0.1
937
+ # via
938
+ # anychat (pyproject.toml)
939
+ # ai-gradio
940
+ # crewai
941
+ # embedchain
942
+ # litellm
943
+ # openai-gradio
944
+ # pydantic-settings
945
+ # scrapegraph-py
946
+ # smolagents
947
+ # uvicorn
948
  python-multipart==0.0.20
949
  # via gradio
950
+ pytube==15.0.0
951
+ # via crewai-tools
952
+ pytz==2024.2
953
+ # via
954
+ # mem0ai
955
+ # pandas
956
+ pyvips==2.2.3
957
+ # via ai-gradio
958
+ pyvips-binary==8.16.0
959
+ # via ai-gradio
960
+ pyvis==0.3.2
961
+ # via crewai
962
  pyyaml==6.0.2
963
  # via
964
  # accelerate
965
+ # chromadb
966
  # gradio
967
  # huggingface-hub
968
+ # kubernetes
969
+ # langchain
970
+ # langchain-community
971
+ # langchain-core
972
+ # transformers
973
+ # uvicorn
974
+ qdrant-client==1.12.2
975
+ # via mem0ai
976
+ referencing==0.35.1
977
+ # via
978
+ # jsonschema
979
+ # jsonschema-specifications
980
+ regex==2024.11.6
981
+ # via
982
+ # crewai
983
+ # tiktoken
984
+ # transformers
985
+ replicate==1.0.4
986
+ # via replicate-gradio
987
+ replicate-gradio @ git+https://github.com/AK391/replicate-gradio.git@691c397515f5b58cec99b38d28830d31879b8d11
988
+ # via anychat (pyproject.toml)
989
  requests==2.32.3
990
  # via
991
+ # ai-gradio
992
+ # auth0-python
993
+ # cohere
994
+ # crewai-tools
995
+ # docker
996
+ # google-api-core
997
+ # google-cloud-bigquery
998
+ # google-cloud-storage
999
+ # google-genai
1000
+ # gptcache
1001
  # huggingface-hub
1002
+ # instructor
1003
+ # kubernetes
1004
+ # langchain
1005
+ # langchain-community
1006
+ # langsmith
1007
+ # opentelemetry-exporter-otlp-proto-http
1008
  # pooch
1009
+ # posthog
1010
+ # requests-oauthlib
1011
+ # requests-toolbelt
1012
+ # scrapegraph-py
1013
+ # serpapi
1014
+ # smolagents
1015
+ # spider-client
1016
+ # tavily-python
1017
+ # tiktoken
1018
+ # transformers
1019
  # twilio
1020
+ requests-oauthlib==2.0.0
1021
+ # via kubernetes
1022
+ requests-toolbelt==1.0.0
1023
+ # via langsmith
1024
  rich==13.9.4
1025
+ # via
1026
+ # chromadb
1027
+ # embedchain
1028
+ # instructor
1029
+ # smolagents
1030
+ # typer
1031
+ rpds-py==0.22.3
1032
+ # via
1033
+ # jsonschema
1034
+ # referencing
1035
+ rsa==4.9
1036
+ # via google-auth
1037
+ ruff==0.9.1
1038
  # via gradio
1039
  safehttpx==0.1.6
1040
  # via gradio
1041
+ safetensors==0.5.2
1042
+ # via
1043
+ # accelerate
1044
+ # transformers
1045
+ sambanova-gradio==0.1.9
1046
+ # via anychat (pyproject.toml)
1047
+ schema==0.7.7
1048
+ # via embedchain
1049
  scikit-learn==1.6.1
1050
  # via librosa
1051
+ scipy==1.15.1
1052
  # via
1053
  # bitsandbytes
1054
  # librosa
1055
  # scikit-learn
1056
+ scrapegraph-py==1.9.0
1057
+ # via crewai-tools
1058
+ selenium==4.27.1
1059
+ # via crewai-tools
1060
  semantic-version==2.10.0
1061
  # via gradio
1062
+ serpapi==0.1.5
1063
+ # via crewai-tools
1064
+ setuptools==75.8.0
1065
+ # via grpcio-tools
1066
+ shapely==2.0.6
1067
+ # via google-cloud-aiplatform
1068
  shellingham==1.5.4
1069
  # via typer
1070
  six==1.17.0
1071
+ # via
1072
+ # kubernetes
1073
+ # markdownify
1074
+ # posthog
1075
+ # python-dateutil
1076
+ smolagents==1.2.2
1077
+ # via ai-gradio
1078
  sniffio==1.3.1
1079
+ # via
1080
+ # anthropic
1081
+ # anyio
1082
+ # cerebras-cloud-sdk
1083
+ # groq
1084
+ # httpx
1085
+ # lumaai
1086
+ # openai
1087
+ # trio
1088
+ sortedcontainers==2.4.0
1089
+ # via trio
1090
+ soundfile==0.13.0
1091
  # via librosa
1092
+ soupsieve==2.6
1093
+ # via beautifulsoup4
1094
  soxr==0.5.0.post1
1095
  # via librosa
1096
+ spider-client==0.1.25
1097
+ # via crewai-tools
1098
+ sqlalchemy==2.0.37
1099
+ # via
1100
+ # alembic
1101
+ # embedchain
1102
+ # langchain
1103
+ # langchain-community
1104
+ # mem0ai
1105
+ stack-data==0.6.3
1106
+ # via ipython
1107
+ starlette==0.41.3
1108
  # via
1109
  # fastapi
1110
  # gradio
1111
+ sympy==1.13.3
1112
  # via
1113
  # onnxruntime
1114
  # torch
1115
+ tabulate==0.9.0
1116
+ # via langchain-cohere
1117
+ tavily-python==0.5.0
1118
+ # via ai-gradio
1119
+ tenacity==9.0.0
1120
+ # via
1121
+ # chromadb
1122
+ # instructor
1123
+ # langchain
1124
+ # langchain-community
1125
+ # langchain-core
1126
+ # spider-client
1127
+ threadpoolctl==3.5.0
1128
  # via scikit-learn
1129
+ tiktoken==0.7.0
1130
+ # via
1131
+ # embedchain
1132
+ # langchain-openai
1133
+ # litellm
1134
+ # tavily-python
1135
+ together-gradio==0.0.1
1136
+ # via anychat (pyproject.toml)
1137
+ tokenizers==0.20.3
1138
+ # via
1139
+ # chromadb
1140
+ # cohere
1141
+ # litellm
1142
+ # transformers
1143
+ tomli==2.2.1
1144
+ # via
1145
+ # build
1146
+ # crewai
1147
+ # pytest
1148
+ tomli-w==1.1.0
1149
+ # via crewai
1150
  tomlkit==0.13.2
1151
  # via gradio
1152
+ torch==2.4.1
1153
  # via
1154
  # accelerate
1155
  # ai-gradio
1156
+ # torchvision
1157
+ torchvision==0.19.1
 
1158
  # via ai-gradio
1159
+ tqdm==4.67.1
1160
+ # via
1161
+ # chromadb
1162
+ # google-generativeai
1163
+ # huggingface-hub
1164
+ # lancedb
1165
+ # openai
1166
+ # transformers
1167
+ traitlets==5.14.3
1168
+ # via
1169
+ # ipython
1170
+ # matplotlib-inline
1171
+ transformers==4.46.3
1172
+ # via
1173
+ # ai-gradio
1174
+ # smolagents
1175
+ trio==0.28.0
1176
+ # via
1177
+ # selenium
1178
+ # trio-websocket
1179
+ trio-websocket==0.11.1
1180
+ # via selenium
1181
+ twilio==9.4.2
1182
+ # via
1183
+ # anychat (pyproject.toml)
1184
+ # ai-gradio
1185
+ # gemini-gradio
1186
+ # openai-gradio
1187
+ typer==0.15.1
1188
+ # via
1189
+ # chromadb
1190
+ # gradio
1191
+ # instructor
1192
+ types-requests==2.32.0.20241016
1193
+ # via cohere
1194
  typing-extensions==4.12.2
1195
  # via
1196
+ # alembic
1197
+ # anthropic
1198
  # anyio
1199
+ # asgiref
1200
+ # cerebras-cloud-sdk
1201
+ # chromadb
1202
+ # cohere
1203
+ # e2b
1204
  # fastapi
1205
+ # google-cloud-aiplatform
1206
+ # google-generativeai
1207
  # gradio
1208
  # gradio-client
1209
+ # groq
1210
  # huggingface-hub
1211
+ # ipython
1212
+ # langchain-core
1213
  # librosa
1214
+ # lumaai
1215
  # multidict
1216
+ # openai
1217
+ # opentelemetry-sdk
1218
  # pydantic
1219
  # pydantic-core
1220
  # pyee
1221
+ # pypdf
1222
+ # pyright
1223
+ # replicate
1224
  # rich
1225
+ # selenium
1226
+ # sqlalchemy
1227
  # torch
1228
  # typer
1229
+ # typing-inspect
1230
  # uvicorn
1231
+ typing-inspect==0.9.0
1232
+ # via
1233
+ # dataclasses-json
1234
+ # mistralai
1235
+ tzdata==2024.2
1236
  # via pandas
1237
+ uritemplate==4.1.1
1238
+ # via google-api-python-client
1239
  urllib3==2.3.0
1240
+ # via
1241
+ # auth0-python
1242
+ # docker
1243
+ # kubernetes
1244
+ # qdrant-client
1245
+ # requests
1246
+ # selenium
1247
+ # types-requests
1248
+ uv==0.5.17
1249
+ # via crewai
1250
  uvicorn==0.34.0
 
 
1251
  # via
1252
+ # chromadb
1253
+ # gradio
1254
+ uvloop==0.21.0
1255
+ # via uvicorn
1256
+ validators==0.34.0
1257
+ # via weaviate-client
1258
+ watchfiles==1.0.4
1259
+ # via uvicorn
1260
+ wcwidth==0.2.13
1261
+ # via prompt-toolkit
1262
+ weaviate-client==4.10.2
1263
+ # via crewai-tools
1264
+ websocket-client==1.8.0
1265
+ # via
1266
+ # kubernetes
1267
+ # selenium
1268
+ websockets==14.1
1269
+ # via
1270
+ # anychat (pyproject.toml)
1271
  # ai-gradio
1272
+ # gemini-gradio
1273
+ # google-genai
1274
  # gradio-client
1275
+ # uvicorn
1276
+ wrapt==1.17.1
1277
+ # via
1278
+ # deprecated
1279
+ # opentelemetry-instrumentation
1280
+ wsproto==1.2.0
1281
+ # via
1282
+ # httpx-ws
1283
+ # trio-websocket
1284
+ xai-gradio==0.0.2
1285
+ # via anychat (pyproject.toml)
1286
  yarl==1.18.3
1287
  # via aiohttp
1288
+ zipp==3.21.0
1289
+ # via importlib-metadata