Spaces:
Sleeping
Sleeping
Add gemini-2.0-flash-exp LLM
Browse files- app.py +1 -1
- global_config.py +7 -2
app.py
CHANGED
|
@@ -159,7 +159,7 @@ with st.sidebar:
|
|
| 159 |
else:
|
| 160 |
# The LLMs
|
| 161 |
llm_provider_to_use = st.sidebar.selectbox(
|
| 162 |
-
label='2: Select
|
| 163 |
options=[f'{k} ({v["description"]})' for k, v in GlobalConfig.VALID_MODELS.items()],
|
| 164 |
index=GlobalConfig.DEFAULT_MODEL_INDEX,
|
| 165 |
help=GlobalConfig.LLM_PROVIDER_HELP,
|
|
|
|
| 159 |
else:
|
| 160 |
# The LLMs
|
| 161 |
llm_provider_to_use = st.sidebar.selectbox(
|
| 162 |
+
label='2: Select a suitable LLM to use:\n\n(Gemini and Mistral-Nemo are recommended)',
|
| 163 |
options=[f'{k} ({v["description"]})' for k, v in GlobalConfig.VALID_MODELS.items()],
|
| 164 |
index=GlobalConfig.DEFAULT_MODEL_INDEX,
|
| 165 |
help=GlobalConfig.LLM_PROVIDER_HELP,
|
global_config.py
CHANGED
|
@@ -34,7 +34,12 @@ class GlobalConfig:
|
|
| 34 |
'paid': True,
|
| 35 |
},
|
| 36 |
'[gg]gemini-1.5-flash-002': {
|
| 37 |
-
'description': 'faster
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 38 |
'max_new_tokens': 8192,
|
| 39 |
'paid': True,
|
| 40 |
},
|
|
@@ -55,7 +60,7 @@ class GlobalConfig:
|
|
| 55 |
'- **[gg]**: Google Gemini API\n'
|
| 56 |
'- **[hf]**: Hugging Face Inference API\n'
|
| 57 |
)
|
| 58 |
-
DEFAULT_MODEL_INDEX =
|
| 59 |
LLM_MODEL_TEMPERATURE = 0.2
|
| 60 |
LLM_MODEL_MIN_OUTPUT_LENGTH = 100
|
| 61 |
LLM_MODEL_MAX_INPUT_LENGTH = 400 # characters
|
|
|
|
| 34 |
'paid': True,
|
| 35 |
},
|
| 36 |
'[gg]gemini-1.5-flash-002': {
|
| 37 |
+
'description': 'faster, detailed',
|
| 38 |
+
'max_new_tokens': 8192,
|
| 39 |
+
'paid': True,
|
| 40 |
+
},
|
| 41 |
+
'[gg]gemini-2.0-flash-exp': {
|
| 42 |
+
'description': 'fast, detailed',
|
| 43 |
'max_new_tokens': 8192,
|
| 44 |
'paid': True,
|
| 45 |
},
|
|
|
|
| 60 |
'- **[gg]**: Google Gemini API\n'
|
| 61 |
'- **[hf]**: Hugging Face Inference API\n'
|
| 62 |
)
|
| 63 |
+
DEFAULT_MODEL_INDEX = 3
|
| 64 |
LLM_MODEL_TEMPERATURE = 0.2
|
| 65 |
LLM_MODEL_MIN_OUTPUT_LENGTH = 100
|
| 66 |
LLM_MODEL_MAX_INPUT_LENGTH = 400 # characters
|