Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -160,15 +160,31 @@ def search_glossary(query):
|
|
| 160 |
# 🕵️♂️ Searching the glossary for: query
|
| 161 |
all_results = ""
|
| 162 |
st.markdown(f"- {query}")
|
| 163 |
-
|
|
|
|
|
|
|
|
|
|
| 164 |
# 🔍 Run 1 - ArXiv RAG researcher expert ~-<>-~ Paper Summary & Ask LLM
|
| 165 |
client = Client("awacke1/Arxiv-Paper-Search-And-QA-RAG-Pattern")
|
| 166 |
response2 = client.predict(
|
| 167 |
-
query, # str in 'parameter_13' Textbox component
|
| 168 |
-
|
| 169 |
-
|
| 170 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 171 |
)
|
|
|
|
| 172 |
st.write('🔍 Run of Multi-Agent System Paper Summary Spec is Complete')
|
| 173 |
st.markdown(response2)
|
| 174 |
|
|
|
|
| 160 |
# 🕵️♂️ Searching the glossary for: query
|
| 161 |
all_results = ""
|
| 162 |
st.markdown(f"- {query}")
|
| 163 |
+
|
| 164 |
+
#database_choice Literal['Semantic Search', 'Arxiv Search - Latest - (EXPERIMENTAL)'] Default: "Semantic Search"
|
| 165 |
+
#llm_model_picked Literal['mistralai/Mixtral-8x7B-Instruct-v0.1', 'mistralai/Mistral-7B-Instruct-v0.2', 'google/gemma-7b-it', 'None'] Default: "mistralai/Mistral-7B-Instruct-v0.2"
|
| 166 |
+
|
| 167 |
# 🔍 Run 1 - ArXiv RAG researcher expert ~-<>-~ Paper Summary & Ask LLM
|
| 168 |
client = Client("awacke1/Arxiv-Paper-Search-And-QA-RAG-Pattern")
|
| 169 |
response2 = client.predict(
|
| 170 |
+
message=query, # str in 'parameter_13' Textbox component
|
| 171 |
+
llm_results_use=5,
|
| 172 |
+
database_choice="Semantic Search",
|
| 173 |
+
llm_model_picked="mistralai/Mistral-7B-Instruct-v0.2",
|
| 174 |
+
api_name="/update_with_rag_md"
|
| 175 |
+
)
|
| 176 |
+
|
| 177 |
+
#llm_model_picked Literal['mistralai/Mixtral-8x7B-Instruct-v0.1', 'mistralai/Mistral-7B-Instruct-v0.2', 'google/gemma-7b-it', 'None'] Default: "mistralai/Mistral-7B-Instruct-v0.2"
|
| 178 |
+
|
| 179 |
+
|
| 180 |
+
client = Client("awacke1/Arxiv-Paper-Search-And-QA-RAG-Pattern")
|
| 181 |
+
result = client.predict(
|
| 182 |
+
prompt=query,
|
| 183 |
+
llm_model_picked="mistralai/Mistral-7B-Instruct-v0.2",
|
| 184 |
+
stream_outputs=True,
|
| 185 |
+
api_name="/ask_llm"
|
| 186 |
)
|
| 187 |
+
|
| 188 |
st.write('🔍 Run of Multi-Agent System Paper Summary Spec is Complete')
|
| 189 |
st.markdown(response2)
|
| 190 |
|