Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
|
@@ -38,6 +38,7 @@ MODELS = [
|
|
| 38 |
"@cf/meta/llama-3.1-8b-instruct",
|
| 39 |
"mistralai/Mistral-Nemo-Instruct-2407",
|
| 40 |
"meta-llama/Meta-Llama-3.1-8B-Instruct",
|
|
|
|
| 41 |
"duckduckgo/gpt-4o-mini",
|
| 42 |
"duckduckgo/claude-3-haiku",
|
| 43 |
"duckduckgo/llama-3.1-70b",
|
|
@@ -595,7 +596,7 @@ Write a detailed and complete response that answers the following user question:
|
|
| 595 |
logging.info(f"API call {i+1}/{num_calls}")
|
| 596 |
for message in client.chat_completion(
|
| 597 |
messages=[{"role": "user", "content": prompt}],
|
| 598 |
-
max_tokens=
|
| 599 |
temperature=temperature,
|
| 600 |
stream=True,
|
| 601 |
):
|
|
|
|
| 38 |
"@cf/meta/llama-3.1-8b-instruct",
|
| 39 |
"mistralai/Mistral-Nemo-Instruct-2407",
|
| 40 |
"meta-llama/Meta-Llama-3.1-8B-Instruct",
|
| 41 |
+
"mattshumer/Reflection-Llama-3.1-70B",
|
| 42 |
"duckduckgo/gpt-4o-mini",
|
| 43 |
"duckduckgo/claude-3-haiku",
|
| 44 |
"duckduckgo/llama-3.1-70b",
|
|
|
|
| 596 |
logging.info(f"API call {i+1}/{num_calls}")
|
| 597 |
for message in client.chat_completion(
|
| 598 |
messages=[{"role": "user", "content": prompt}],
|
| 599 |
+
max_tokens=10000,
|
| 600 |
temperature=temperature,
|
| 601 |
stream=True,
|
| 602 |
):
|