Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -335,6 +335,7 @@ def search_glossary(query):
|
|
| 335 |
api_name="/ask_llm"
|
| 336 |
)
|
| 337 |
st.markdown(result)
|
|
|
|
| 338 |
st.code(result, language="python", line_numbers=True)
|
| 339 |
# ๐ ArXiv RAG researcher expert ~-<>-~ Paper Summary & Ask LLM - api_name: /ask_llm
|
| 340 |
result2 = client.predict(
|
|
@@ -344,6 +345,7 @@ def search_glossary(query):
|
|
| 344 |
api_name="/ask_llm"
|
| 345 |
)
|
| 346 |
st.markdown(result2)
|
|
|
|
| 347 |
st.code(result2, language="python", line_numbers=True)
|
| 348 |
# ๐ ArXiv RAG researcher expert ~-<>-~ Paper Summary & Ask LLM - api_name: /ask_llm
|
| 349 |
result3 = client.predict(
|
|
@@ -353,6 +355,7 @@ def search_glossary(query):
|
|
| 353 |
api_name="/ask_llm"
|
| 354 |
)
|
| 355 |
st.markdown(result3)
|
|
|
|
| 356 |
st.code(result3, language="python", line_numbers=True)
|
| 357 |
# ๐ ArXiv RAG researcher expert ~-<>-~ Paper Summary & Ask LLM - api_name: /update_with_rag_md
|
| 358 |
response2 = client.predict(
|
|
@@ -363,9 +366,31 @@ def search_glossary(query):
|
|
| 363 |
api_name="/update_with_rag_md"
|
| 364 |
)
|
| 365 |
st.markdown(response2[0])
|
| 366 |
-
st.code(response2[0], language="python", line_numbers=True, wrap_lines=True)
|
|
|
|
| 367 |
st.markdown(response2[1])
|
| 368 |
st.code(response2[1], language="python", line_numbers=True, wrap_lines=True)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 369 |
return result, result2, result3, response2
|
| 370 |
|
| 371 |
|
|
|
|
| 335 |
api_name="/ask_llm"
|
| 336 |
)
|
| 337 |
st.markdown(result)
|
| 338 |
+
|
| 339 |
st.code(result, language="python", line_numbers=True)
|
| 340 |
# ๐ ArXiv RAG researcher expert ~-<>-~ Paper Summary & Ask LLM - api_name: /ask_llm
|
| 341 |
result2 = client.predict(
|
|
|
|
| 345 |
api_name="/ask_llm"
|
| 346 |
)
|
| 347 |
st.markdown(result2)
|
| 348 |
+
|
| 349 |
st.code(result2, language="python", line_numbers=True)
|
| 350 |
# ๐ ArXiv RAG researcher expert ~-<>-~ Paper Summary & Ask LLM - api_name: /ask_llm
|
| 351 |
result3 = client.predict(
|
|
|
|
| 355 |
api_name="/ask_llm"
|
| 356 |
)
|
| 357 |
st.markdown(result3)
|
| 358 |
+
|
| 359 |
st.code(result3, language="python", line_numbers=True)
|
| 360 |
# ๐ ArXiv RAG researcher expert ~-<>-~ Paper Summary & Ask LLM - api_name: /update_with_rag_md
|
| 361 |
response2 = client.predict(
|
|
|
|
| 366 |
api_name="/update_with_rag_md"
|
| 367 |
)
|
| 368 |
st.markdown(response2[0])
|
| 369 |
+
st.code(response2[0], language="python", line_numbers=True, wrap_lines=True)
|
| 370 |
+
|
| 371 |
st.markdown(response2[1])
|
| 372 |
st.code(response2[1], language="python", line_numbers=True, wrap_lines=True)
|
| 373 |
+
|
| 374 |
+
|
| 375 |
+
# Persist AI Results to Markdown Files
|
| 376 |
+
filename = generate_filename(result, "md")
|
| 377 |
+
create_file(filename, query, result)
|
| 378 |
+
st.session_state.chat_history.append({"assistant": query, "ArXiV": result})
|
| 379 |
+
|
| 380 |
+
filename = generate_filename(result2, "md")
|
| 381 |
+
create_file(filename, query, result2)
|
| 382 |
+
st.session_state.chat_history.append({"assistant": query, "ArXiV": result2})
|
| 383 |
+
|
| 384 |
+
filename = generate_filename(result3, "md")
|
| 385 |
+
create_file(filename, query, result3)
|
| 386 |
+
st.session_state.chat_history.append({"assistant": query, "ArXiV": result3})
|
| 387 |
+
|
| 388 |
+
filename = generate_filename(response2, "md")
|
| 389 |
+
create_file(filename, query, response2)
|
| 390 |
+
st.session_state.chat_history.append({"assistant": query, "ArXiV": response2})
|
| 391 |
+
|
| 392 |
+
|
| 393 |
+
|
| 394 |
return result, result2, result3, response2
|
| 395 |
|
| 396 |
|