Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -26,7 +26,7 @@ def test_api_connection(api_key):
|
|
26 |
client = Groq(api_key=api_key)
|
27 |
# Making a minimal API call to test the connection
|
28 |
client.chat.completions.create(
|
29 |
-
model="
|
30 |
messages=[{"role": "user", "content": "test"}],
|
31 |
max_tokens=5
|
32 |
)
|
@@ -246,8 +246,8 @@ def generate_analytics():
|
|
246 |
|
247 |
# Define available models
|
248 |
models = [
|
249 |
-
"
|
250 |
-
"
|
251 |
"mistral-saba-24b",
|
252 |
"gemma2-9b-it",
|
253 |
"allam-2-7b"
|
@@ -268,13 +268,13 @@ with gr.Blocks(title="Groq AI Chat Playground") as app:
|
|
268 |
gr.Markdown("""
|
269 |
### Available Models and Use Cases
|
270 |
|
271 |
-
**
|
272 |
- Meta's most powerful language model
|
273 |
- 70 billion parameters with 8192 token context window
|
274 |
- Best for: Complex reasoning, sophisticated content generation, creative writing, and detailed analysis
|
275 |
- Optimal for users needing the highest quality AI responses
|
276 |
|
277 |
-
**
|
278 |
- Lighter version of Llama 3
|
279 |
- 8 billion parameters with 8192 token context window
|
280 |
- Best for: Faster responses, everyday tasks, simpler queries
|
@@ -320,7 +320,7 @@ with gr.Blocks(title="Groq AI Chat Playground") as app:
|
|
320 |
model_dropdown = gr.Dropdown(
|
321 |
choices=models,
|
322 |
label="Select Model",
|
323 |
-
value="
|
324 |
)
|
325 |
with gr.Column(scale=1):
|
326 |
template_dropdown = gr.Dropdown(
|
|
|
26 |
client = Groq(api_key=api_key)
|
27 |
# Making a minimal API call to test the connection
|
28 |
client.chat.completions.create(
|
29 |
+
model="llama-3.3-70b-versatile",
|
30 |
messages=[{"role": "user", "content": "test"}],
|
31 |
max_tokens=5
|
32 |
)
|
|
|
246 |
|
247 |
# Define available models
|
248 |
models = [
|
249 |
+
"llama-3.3-70b-versatile",
|
250 |
+
"llama-3.1-8b-instant",
|
251 |
"mistral-saba-24b",
|
252 |
"gemma2-9b-it",
|
253 |
"allam-2-7b"
|
|
|
268 |
gr.Markdown("""
|
269 |
### Available Models and Use Cases
|
270 |
|
271 |
+
**llama-3.3-70b-versatile**
|
272 |
- Meta's most powerful language model
|
273 |
- 70 billion parameters with 8192 token context window
|
274 |
- Best for: Complex reasoning, sophisticated content generation, creative writing, and detailed analysis
|
275 |
- Optimal for users needing the highest quality AI responses
|
276 |
|
277 |
+
**llama-3.1-8b-instant**
|
278 |
- Lighter version of Llama 3
|
279 |
- 8 billion parameters with 8192 token context window
|
280 |
- Best for: Faster responses, everyday tasks, simpler queries
|
|
|
320 |
model_dropdown = gr.Dropdown(
|
321 |
choices=models,
|
322 |
label="Select Model",
|
323 |
+
value="llama-3.3-70b-versatile"
|
324 |
)
|
325 |
with gr.Column(scale=1):
|
326 |
template_dropdown = gr.Dropdown(
|