Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
|
@@ -26,7 +26,7 @@ def test_api_connection(api_key):
|
|
| 26 |
client = Groq(api_key=api_key)
|
| 27 |
# Making a minimal API call to test the connection
|
| 28 |
client.chat.completions.create(
|
| 29 |
-
model="
|
| 30 |
messages=[{"role": "user", "content": "test"}],
|
| 31 |
max_tokens=5
|
| 32 |
)
|
|
@@ -246,8 +246,8 @@ def generate_analytics():
|
|
| 246 |
|
| 247 |
# Define available models
|
| 248 |
models = [
|
| 249 |
-
"
|
| 250 |
-
"
|
| 251 |
"mistral-saba-24b",
|
| 252 |
"gemma2-9b-it",
|
| 253 |
"allam-2-7b"
|
|
@@ -268,13 +268,13 @@ with gr.Blocks(title="Groq AI Chat Playground") as app:
|
|
| 268 |
gr.Markdown("""
|
| 269 |
### Available Models and Use Cases
|
| 270 |
|
| 271 |
-
**
|
| 272 |
- Meta's most powerful language model
|
| 273 |
- 70 billion parameters with 8192 token context window
|
| 274 |
- Best for: Complex reasoning, sophisticated content generation, creative writing, and detailed analysis
|
| 275 |
- Optimal for users needing the highest quality AI responses
|
| 276 |
|
| 277 |
-
**
|
| 278 |
- Lighter version of Llama 3
|
| 279 |
- 8 billion parameters with 8192 token context window
|
| 280 |
- Best for: Faster responses, everyday tasks, simpler queries
|
|
@@ -320,7 +320,7 @@ with gr.Blocks(title="Groq AI Chat Playground") as app:
|
|
| 320 |
model_dropdown = gr.Dropdown(
|
| 321 |
choices=models,
|
| 322 |
label="Select Model",
|
| 323 |
-
value="
|
| 324 |
)
|
| 325 |
with gr.Column(scale=1):
|
| 326 |
template_dropdown = gr.Dropdown(
|
|
|
|
| 26 |
client = Groq(api_key=api_key)
|
| 27 |
# Making a minimal API call to test the connection
|
| 28 |
client.chat.completions.create(
|
| 29 |
+
model="llama-3.3-70b-versatile",
|
| 30 |
messages=[{"role": "user", "content": "test"}],
|
| 31 |
max_tokens=5
|
| 32 |
)
|
|
|
|
| 246 |
|
| 247 |
# Define available models
|
| 248 |
models = [
|
| 249 |
+
"llama-3.3-70b-versatile",
|
| 250 |
+
"llama-3.1-8b-instant",
|
| 251 |
"mistral-saba-24b",
|
| 252 |
"gemma2-9b-it",
|
| 253 |
"allam-2-7b"
|
|
|
|
| 268 |
gr.Markdown("""
|
| 269 |
### Available Models and Use Cases
|
| 270 |
|
| 271 |
+
**llama-3.3-70b-versatile**
|
| 272 |
- Meta's most powerful language model
|
| 273 |
- 70 billion parameters with 8192 token context window
|
| 274 |
- Best for: Complex reasoning, sophisticated content generation, creative writing, and detailed analysis
|
| 275 |
- Optimal for users needing the highest quality AI responses
|
| 276 |
|
| 277 |
+
**llama-3.1-8b-instant**
|
| 278 |
- Lighter version of Llama 3
|
| 279 |
- 8 billion parameters with 8192 token context window
|
| 280 |
- Best for: Faster responses, everyday tasks, simpler queries
|
|
|
|
| 320 |
model_dropdown = gr.Dropdown(
|
| 321 |
choices=models,
|
| 322 |
label="Select Model",
|
| 323 |
+
value="llama-3.3-70b-versatile"
|
| 324 |
)
|
| 325 |
with gr.Column(scale=1):
|
| 326 |
template_dropdown = gr.Dropdown(
|