Spaces:
				
			
			
	
			
			
		Runtime error
		
	
	
	
			
			
	
	
	
	
		
		
		Runtime error
		
	Commit 
							
							Β·
						
						4a448eb
	
1
								Parent(s):
							
							65935d6
								
Working all
Browse files- .streamlit/config.toml +3 -0
 - app.py +52 -109
 - backend_utils.py +120 -0
 - my_faiss_index.faiss β data/my_faiss_index.faiss +0 -0
 - my_faiss_index.json β data/my_faiss_index.json +0 -0
 - data/sample_1.txt +0 -1
 - data/sample_2.txt +0 -1
 - my_faiss_config.json +0 -1
 - requirements.txt +2 -1
 
    	
        .streamlit/config.toml
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            [theme]
         
     | 
| 2 | 
         
            +
            base = "light"
         
     | 
| 3 | 
         
            +
            font="monospace"
         
     | 
    	
        app.py
    CHANGED
    
    | 
         @@ -1,116 +1,59 @@ 
     | 
|
| 1 | 
         
            -
            import glob
         
     | 
| 2 | 
         
            -
            import os
         
     | 
| 3 | 
         
            -
            import logging
         
     | 
| 4 | 
         
            -
            import sys
         
     | 
| 5 | 
         
            -
             
     | 
| 6 | 
         
             
            import streamlit as st
         
     | 
| 7 | 
         
            -
            from  
     | 
| 8 | 
         
            -
             
     | 
| 9 | 
         
            -
             
     | 
| 10 | 
         
            -
             
     | 
| 11 | 
         
            -
             
     | 
| 12 | 
         
            -
             
     | 
| 13 | 
         
            -
             
     | 
| 14 | 
         
            -
             
     | 
| 15 | 
         
            -
             
     | 
| 16 | 
         
            -
             
     | 
| 17 | 
         
            -
             
     | 
| 18 | 
         
            -
             
     | 
| 19 | 
         
            -
             
     | 
| 20 | 
         
            -
             
     | 
| 21 | 
         
            -
             
     | 
| 22 | 
         
            -
             
     | 
| 23 | 
         
            -
                 
     | 
| 24 | 
         
            -
             
     | 
| 25 | 
         
            -
             
     | 
| 26 | 
         
            -
             
     | 
| 27 | 
         
            -
             
     | 
| 28 | 
         
            -
             
     | 
| 29 | 
         
            -
             
     | 
| 30 | 
         
            -
             
     | 
| 31 | 
         
            -
             
     | 
| 32 | 
         
            -
             
     | 
| 33 | 
         
            -
             
     | 
| 34 | 
         
            -
             
     | 
| 35 | 
         
            -
             
     | 
| 36 | 
         
            -
             
     | 
| 37 | 
         
            -
             
     | 
| 38 | 
         
            -
             
     | 
| 39 | 
         
            -
             
     | 
| 40 | 
         
            -
             
     | 
| 41 | 
         
            -
                )
         
     | 
| 42 | 
         
            -
             
     | 
| 43 | 
         
            -
             
     | 
| 44 | 
         
            -
             
     | 
| 45 | 
         
            -
             
     | 
| 46 | 
         
            -
             
     | 
| 47 | 
         
            -
             
     | 
| 48 | 
         
            -
             
     | 
| 49 | 
         
            -
             
     | 
| 50 | 
         
            -
             
     | 
| 51 | 
         
            -
             
     | 
| 52 | 
         
            -
             
     | 
| 53 | 
         
            -
             
     | 
| 54 | 
         
            -
             
     | 
| 55 | 
         
            -
                 
     | 
| 56 | 
         
            -
                pipe.add_node(component=shaper, name="shaper", inputs=["retriever"])
         
     | 
| 57 | 
         
            -
                pipe.add_node(component=node, name="prompt_node", inputs=["shaper"])
         
     | 
| 58 | 
         
            -
                return pipe
         
     | 
| 59 | 
         
            -
             
     | 
| 60 | 
         
            -
             
     | 
| 61 | 
         
            -
            def get_web_ret_pipeline():
         
     | 
| 62 | 
         
            -
                search_key = st.secrets["WEBRET_API_KEY"]
         
     | 
| 63 | 
         
            -
                web_retriever = WebRetriever(api_key=search_key, search_engine_provider="SerperDev")
         
     | 
| 64 | 
         
            -
                shaper = Shaper(func="join_documents", inputs={"documents": "documents"}, outputs=["documents"])
         
     | 
| 65 | 
         
            -
                default_template = PromptTemplate(
         
     | 
| 66 | 
         
            -
                    name="question-answering",
         
     | 
| 67 | 
         
            -
                    prompt_text="Given the context please answer the question. Context: $documents; Question: "
         
     | 
| 68 | 
         
            -
                                "$query; Answer:",
         
     | 
| 69 | 
         
            -
                )
         
     | 
| 70 | 
         
            -
                # Let's initiate the PromptNode
         
     | 
| 71 | 
         
            -
                node = PromptNode("text-davinci-003", default_prompt_template=default_template,
         
     | 
| 72 | 
         
            -
                                  api_key=st.secrets["OPENAI_API_KEY"], max_length=500)
         
     | 
| 73 | 
         
            -
                # Let's create a pipeline with Shaper and PromptNode
         
     | 
| 74 | 
         
            -
                pipe = Pipeline()
         
     | 
| 75 | 
         
            -
                pipe.add_node(component=web_retriever, name='retriever', inputs=['Query'])
         
     | 
| 76 | 
         
            -
                pipe.add_node(component=shaper, name="shaper", inputs=["retriever"])
         
     | 
| 77 | 
         
            -
                pipe.add_node(component=node, name="prompt_node", inputs=["shaper"])
         
     | 
| 78 | 
         
            -
                return pipe
         
     | 
| 79 | 
         
            -
             
     | 
| 80 | 
         
            -
            def app_init():
         
     | 
| 81 | 
         
            -
                os.environ["OPENAI_API_KEY"] = st.secrets["OPENAI_API_KEY"]
         
     | 
| 82 | 
         
            -
                p1 = get_plain_pipeline()
         
     | 
| 83 | 
         
            -
                p2 = get_ret_aug_pipeline()
         
     | 
| 84 | 
         
            -
                p3 = get_web_ret_pipeline()
         
     | 
| 85 | 
         
            -
                return p1, p2, p3
         
     | 
| 86 | 
         
            -
             
     | 
| 87 | 
         
            -
             
     | 
| 88 | 
         
            -
            def main():
         
     | 
| 89 | 
         
             
                p1, p2, p3 = app_init()
         
     | 
| 90 | 
         
            -
                st.title("Haystack Demo")
         
     | 
| 91 | 
         
            -
                input = st.text_input("Query ...", "Did SVB collapse?")
         
     | 
| 92 | 
         
            -
             
     | 
| 93 | 
         
            -
                query_type = st.radio("Type",
         
     | 
| 94 | 
         
            -
                                      ("Retrieval Augmented", "Retrieval Augmented with Web Search"))
         
     | 
| 95 | 
         
            -
                # col_1, col_2 = st.columns(2)
         
     | 
| 96 | 
         
            -
             
     | 
| 97 | 
         
            -
                if st.button("Random Question"):
         
     | 
| 98 | 
         
            -
                    new_text = "Streamlit is great!"
         
     | 
| 99 | 
         
            -
                    input.value = new_text
         
     | 
| 100 | 
         
            -
             
     | 
| 101 | 
         
            -
                # with col_1:
         
     | 
| 102 | 
         
            -
                # st.text("PLAIN")
         
     | 
| 103 | 
         
             
                answers = p1.run(input)
         
     | 
| 104 | 
         
            -
                 
     | 
| 105 | 
         | 
| 106 | 
         
            -
                 
     | 
| 107 | 
         
            -
                # st.write(query_type.upper())
         
     | 
| 108 | 
         
            -
                if query_type == "Retrieval Augmented":
         
     | 
| 109 | 
         
             
                    answers_2 = p2.run(input)
         
     | 
| 110 | 
         
             
                else:
         
     | 
| 111 | 
         
             
                    answers_2 = p3.run(input)
         
     | 
| 112 | 
         
            -
                 
     | 
| 113 | 
         
            -
             
     | 
| 114 | 
         
            -
             
     | 
| 115 | 
         
            -
            if __name__ == "__main__":
         
     | 
| 116 | 
         
            -
                main()
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 1 | 
         
             
            import streamlit as st
         
     | 
| 2 | 
         
            +
            from backend_utils import app_init, set_q1, set_q2, set_q3, set_q4, set_q5
         
     | 
| 3 | 
         
            +
             
     | 
| 4 | 
         
            +
            st.markdown("<center> <h1> Haystack Demo </h1> </center>", unsafe_allow_html=True)
         
     | 
| 5 | 
         
            +
             
     | 
| 6 | 
         
            +
            if st.session_state.get('pipelines_loaded', False):
         
     | 
| 7 | 
         
            +
                with st.spinner('Loading pipelines...'):
         
     | 
| 8 | 
         
            +
                    p1, p2, p3 = app_init()
         
     | 
| 9 | 
         
            +
                    st.success('Pipelines are loaded', icon="β
")
         
     | 
| 10 | 
         
            +
                    st.session_state['pipelines_loaded'] = True
         
     | 
| 11 | 
         
            +
             
     | 
| 12 | 
         
            +
            placeholder = st.empty()
         
     | 
| 13 | 
         
            +
            with placeholder:
         
     | 
| 14 | 
         
            +
                search_bar, button = st.columns([3, 1])
         
     | 
| 15 | 
         
            +
                with search_bar:
         
     | 
| 16 | 
         
            +
                    username = st.text_area(f"", max_chars=200, key='query')
         
     | 
| 17 | 
         
            +
             
     | 
| 18 | 
         
            +
                with button:
         
     | 
| 19 | 
         
            +
                    st.write("")
         
     | 
| 20 | 
         
            +
                    st.write("")
         
     | 
| 21 | 
         
            +
                    run_pressed = st.button("Run")
         
     | 
| 22 | 
         
            +
             
     | 
| 23 | 
         
            +
            st.radio("Type", ("Retrieval Augmented", "Retrieval Augmented with Web Search"), key="query_type")
         
     | 
| 24 | 
         
            +
             
     | 
| 25 | 
         
            +
            # st.sidebar.selectbox(
         
     | 
| 26 | 
         
            +
            #      "Example Questions:",
         
     | 
| 27 | 
         
            +
            #      QUERIES,
         
     | 
| 28 | 
         
            +
            #      key='q_drop_down', on_change=set_question)
         
     | 
| 29 | 
         
            +
             
     | 
| 30 | 
         
            +
            c1, c2, c3, c4, c5 = st.columns(5)
         
     | 
| 31 | 
         
            +
            with c1:
         
     | 
| 32 | 
         
            +
                st.button('Example Q1', on_click=set_q1)
         
     | 
| 33 | 
         
            +
            with c2:
         
     | 
| 34 | 
         
            +
                st.button('Example Q2', on_click=set_q2)
         
     | 
| 35 | 
         
            +
            with c3:
         
     | 
| 36 | 
         
            +
                st.button('Example Q3', on_click=set_q3)
         
     | 
| 37 | 
         
            +
            with c4:
         
     | 
| 38 | 
         
            +
                st.button('Example Q4', on_click=set_q4)
         
     | 
| 39 | 
         
            +
            with c5:
         
     | 
| 40 | 
         
            +
                st.button('Example Q5', on_click=set_q5)
         
     | 
| 41 | 
         
            +
             
     | 
| 42 | 
         
            +
            st.markdown("<h4> Answer with PLAIN GPT </h4>", unsafe_allow_html=True)
         
     | 
| 43 | 
         
            +
            placeholder_plain_gpt = st.empty()
         
     | 
| 44 | 
         
            +
            st.text("")
         
     | 
| 45 | 
         
            +
            st.text("")
         
     | 
| 46 | 
         
            +
            st.markdown(f"<h4> Answer with {st.session_state['query_type'].upper()} </h4>", unsafe_allow_html=True)
         
     | 
| 47 | 
         
            +
            placeholder_retrieval_augmented = st.empty()
         
     | 
| 48 | 
         
            +
             
     | 
| 49 | 
         
            +
            if st.session_state.get('query') and run_pressed:
         
     | 
| 50 | 
         
            +
                input = st.session_state['query']
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 51 | 
         
             
                p1, p2, p3 = app_init()
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 52 | 
         
             
                answers = p1.run(input)
         
     | 
| 53 | 
         
            +
                placeholder_plain_gpt.markdown(answers['results'][0])
         
     | 
| 54 | 
         | 
| 55 | 
         
            +
                if st.session_state.get("query_type", "Retrieval Augmented") == "Retrieval Augmented":
         
     | 
| 
         | 
|
| 
         | 
|
| 56 | 
         
             
                    answers_2 = p2.run(input)
         
     | 
| 57 | 
         
             
                else:
         
     | 
| 58 | 
         
             
                    answers_2 = p3.run(input)
         
     | 
| 59 | 
         
            +
                placeholder_retrieval_augmented.markdown(answers_2['results'][0])
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
    	
        backend_utils.py
    ADDED
    
    | 
         @@ -0,0 +1,120 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            import os
         
     | 
| 2 | 
         
            +
             
     | 
| 3 | 
         
            +
            import streamlit as st
         
     | 
| 4 | 
         
            +
            from haystack import Pipeline
         
     | 
| 5 | 
         
            +
            from haystack.document_stores import FAISSDocumentStore
         
     | 
| 6 | 
         
            +
            from haystack.nodes import Shaper, PromptNode, PromptTemplate, PromptModel, EmbeddingRetriever
         
     | 
| 7 | 
         
            +
            from haystack.nodes.retriever.web import WebRetriever
         
     | 
| 8 | 
         
            +
             
     | 
| 9 | 
         
            +
             
     | 
| 10 | 
         
            +
            QUERIES = [
         
     | 
| 11 | 
         
            +
                "Did SVB collapse?",
         
     | 
| 12 | 
         
            +
                "Why did SVB collapse?",
         
     | 
| 13 | 
         
            +
                "What does SVB failure mean for our economy?",
         
     | 
| 14 | 
         
            +
                "Who is responsible for SVC collapse?",
         
     | 
| 15 | 
         
            +
                "When did SVB collapse?"
         
     | 
| 16 | 
         
            +
            ]
         
     | 
| 17 | 
         
            +
             
     | 
| 18 | 
         
            +
            def ChangeWidgetFontSize(wgt_txt, wch_font_size = '12px'):
         
     | 
| 19 | 
         
            +
                htmlstr = """<script>var elements = window.parent.document.querySelectorAll('*'), i;
         
     | 
| 20 | 
         
            +
                                for (i = 0; i < elements.length; ++i) { if (elements[i].innerText == |wgt_txt|) 
         
     | 
| 21 | 
         
            +
                                    { elements[i].style.fontSize='""" + wch_font_size + """';} } </script>  """
         
     | 
| 22 | 
         
            +
             
     | 
| 23 | 
         
            +
                htmlstr = htmlstr.replace('|wgt_txt|', "'" + wgt_txt + "'")
         
     | 
| 24 | 
         
            +
             
     | 
| 25 | 
         
            +
             
     | 
| 26 | 
         
            +
            def get_plain_pipeline():
         
     | 
| 27 | 
         
            +
                prompt_open_ai = PromptModel(model_name_or_path="text-davinci-003", api_key=st.secrets["OPENAI_API_KEY"])
         
     | 
| 28 | 
         
            +
                # Now let make one PromptNode use the default model and the other one the OpenAI model:
         
     | 
| 29 | 
         
            +
                plain_llm_template = PromptTemplate(name="plain_llm", prompt_text="Answer the following question: $query")
         
     | 
| 30 | 
         
            +
                node_openai = PromptNode(prompt_open_ai, default_prompt_template=plain_llm_template, max_length=300)
         
     | 
| 31 | 
         
            +
                pipeline = Pipeline()
         
     | 
| 32 | 
         
            +
                pipeline.add_node(component=node_openai, name="prompt_node", inputs=["Query"])
         
     | 
| 33 | 
         
            +
                return pipeline
         
     | 
| 34 | 
         
            +
             
     | 
| 35 | 
         
            +
             
     | 
| 36 | 
         
            +
            def get_retrieval_augmented_pipeline():
         
     | 
| 37 | 
         
            +
                ds = FAISSDocumentStore(faiss_index_path="data/my_faiss_index.faiss",
         
     | 
| 38 | 
         
            +
                                        faiss_config_path="data/my_faiss_index.json")
         
     | 
| 39 | 
         
            +
             
     | 
| 40 | 
         
            +
                retriever = EmbeddingRetriever(
         
     | 
| 41 | 
         
            +
                    document_store=ds,
         
     | 
| 42 | 
         
            +
                    embedding_model="sentence-transformers/multi-qa-mpnet-base-dot-v1",
         
     | 
| 43 | 
         
            +
                    model_format="sentence_transformers",
         
     | 
| 44 | 
         
            +
                    top_k=2
         
     | 
| 45 | 
         
            +
                )
         
     | 
| 46 | 
         
            +
                shaper = Shaper(func="join_documents", inputs={"documents": "documents"}, outputs=["documents"])
         
     | 
| 47 | 
         
            +
             
     | 
| 48 | 
         
            +
                default_template = PromptTemplate(
         
     | 
| 49 | 
         
            +
                    name="question-answering",
         
     | 
| 50 | 
         
            +
                    prompt_text="Given the context please answer the question. Context: $documents; Question: "
         
     | 
| 51 | 
         
            +
                                "$query; Answer:",
         
     | 
| 52 | 
         
            +
                )
         
     | 
| 53 | 
         
            +
                # Let's initiate the PromptNode
         
     | 
| 54 | 
         
            +
                node = PromptNode("text-davinci-003", default_prompt_template=default_template,
         
     | 
| 55 | 
         
            +
                                  api_key=st.secrets["OPENAI_API_KEY"], max_length=500)
         
     | 
| 56 | 
         
            +
             
     | 
| 57 | 
         
            +
                # Let's create a pipeline with Shaper and PromptNode
         
     | 
| 58 | 
         
            +
                pipeline = Pipeline()
         
     | 
| 59 | 
         
            +
                pipeline.add_node(component=retriever, name='retriever', inputs=['Query'])
         
     | 
| 60 | 
         
            +
                pipeline.add_node(component=shaper, name="shaper", inputs=["retriever"])
         
     | 
| 61 | 
         
            +
                pipeline.add_node(component=node, name="prompt_node", inputs=["shaper"])
         
     | 
| 62 | 
         
            +
                return pipeline
         
     | 
| 63 | 
         
            +
             
     | 
| 64 | 
         
            +
             
     | 
| 65 | 
         
            +
            def get_web_retrieval_augmented_pipeline():
         
     | 
| 66 | 
         
            +
                search_key = st.secrets["WEBRET_API_KEY"]
         
     | 
| 67 | 
         
            +
                web_retriever = WebRetriever(api_key=search_key, search_engine_provider="SerperDev")
         
     | 
| 68 | 
         
            +
                shaper = Shaper(func="join_documents", inputs={"documents": "documents"}, outputs=["documents"])
         
     | 
| 69 | 
         
            +
                default_template = PromptTemplate(
         
     | 
| 70 | 
         
            +
                    name="question-answering",
         
     | 
| 71 | 
         
            +
                    prompt_text="Given the context please answer the question. Context: $documents; Question: "
         
     | 
| 72 | 
         
            +
                                "$query; Answer:",
         
     | 
| 73 | 
         
            +
                )
         
     | 
| 74 | 
         
            +
                # Let's initiate the PromptNode
         
     | 
| 75 | 
         
            +
                node = PromptNode("text-davinci-003", default_prompt_template=default_template,
         
     | 
| 76 | 
         
            +
                                  api_key=st.secrets["OPENAI_API_KEY"], max_length=500)
         
     | 
| 77 | 
         
            +
                # Let's create a pipeline with Shaper and PromptNode
         
     | 
| 78 | 
         
            +
                pipeline = Pipeline()
         
     | 
| 79 | 
         
            +
                pipeline.add_node(component=web_retriever, name='retriever', inputs=['Query'])
         
     | 
| 80 | 
         
            +
                pipeline.add_node(component=shaper, name="shaper", inputs=["retriever"])
         
     | 
| 81 | 
         
            +
                pipeline.add_node(component=node, name="prompt_node", inputs=["shaper"])
         
     | 
| 82 | 
         
            +
                return pipeline
         
     | 
| 83 | 
         
            +
             
     | 
| 84 | 
         
            +
             
     | 
| 85 | 
         
            +
            @st.cache_resource(show_spinner=False)
         
     | 
| 86 | 
         
            +
            def app_init():
         
     | 
| 87 | 
         
            +
                os.environ["OPENAI_API_KEY"] = st.secrets["OPENAI_API_KEY"]
         
     | 
| 88 | 
         
            +
                p1 = get_plain_pipeline()
         
     | 
| 89 | 
         
            +
                p2 = get_retrieval_augmented_pipeline()
         
     | 
| 90 | 
         
            +
                p3 = get_web_retrieval_augmented_pipeline()
         
     | 
| 91 | 
         
            +
                return p1, p2, p3
         
     | 
| 92 | 
         
            +
             
     | 
| 93 | 
         
            +
             
     | 
| 94 | 
         
            +
            if 'query' not in st.session_state:
         
     | 
| 95 | 
         
            +
                st.session_state['query'] = ""
         
     | 
| 96 | 
         
            +
             
     | 
| 97 | 
         
            +
             
     | 
| 98 | 
         
            +
            def set_question():
         
     | 
| 99 | 
         
            +
                st.session_state['query'] = st.session_state['q_drop_down']
         
     | 
| 100 | 
         
            +
             
     | 
| 101 | 
         
            +
             
     | 
| 102 | 
         
            +
            def set_q1():
         
     | 
| 103 | 
         
            +
                st.session_state['query'] = QUERIES[0]
         
     | 
| 104 | 
         
            +
             
     | 
| 105 | 
         
            +
             
     | 
| 106 | 
         
            +
            def set_q2():
         
     | 
| 107 | 
         
            +
                st.session_state['query'] = QUERIES[1]
         
     | 
| 108 | 
         
            +
             
     | 
| 109 | 
         
            +
             
     | 
| 110 | 
         
            +
            def set_q3():
         
     | 
| 111 | 
         
            +
                st.session_state['query'] = QUERIES[2]
         
     | 
| 112 | 
         
            +
             
     | 
| 113 | 
         
            +
             
     | 
| 114 | 
         
            +
            def set_q4():
         
     | 
| 115 | 
         
            +
                st.session_state['query'] = QUERIES[3]
         
     | 
| 116 | 
         
            +
             
     | 
| 117 | 
         
            +
             
     | 
| 118 | 
         
            +
            def set_q5():
         
     | 
| 119 | 
         
            +
                st.session_state['query'] = QUERIES[4]
         
     | 
| 120 | 
         
            +
             
     | 
    	
        my_faiss_index.faiss β data/my_faiss_index.faiss
    RENAMED
    
    | 
         
            File without changes
         
     | 
    	
        my_faiss_index.json β data/my_faiss_index.json
    RENAMED
    
    | 
         
            File without changes
         
     | 
    	
        data/sample_1.txt
    DELETED
    
    | 
         @@ -1 +0,0 @@ 
     | 
|
| 1 | 
         
            -
            Hello World 1!
         
     | 
| 
         | 
|
| 
         | 
    	
        data/sample_2.txt
    DELETED
    
    | 
         @@ -1 +0,0 @@ 
     | 
|
| 1 | 
         
            -
            Hello World 2!
         
     | 
| 
         | 
|
| 
         | 
    	
        my_faiss_config.json
    DELETED
    
    | 
         @@ -1 +0,0 @@ 
     | 
|
| 1 | 
         
            -
            {"faiss_config_path": "my_faiss_config.json", "embedding_dim": 768}
         
     | 
| 
         | 
|
| 
         | 
    	
        requirements.txt
    CHANGED
    
    | 
         @@ -2,4 +2,5 @@ git+https://github.com/deepset-ai/haystack.git@ffd02c29f7cc83a119b6440bfbabaacda 
     | 
|
| 2 | 
         
             
            faiss-cpu==1.7.2
         
     | 
| 3 | 
         
             
            sqlalchemy>=1.4.2,<2
         
     | 
| 4 | 
         
             
            sqlalchemy_utils
         
     | 
| 5 | 
         
            -
            psycopg2-binary
         
     | 
| 
         | 
| 
         | 
|
| 2 | 
         
             
            faiss-cpu==1.7.2
         
     | 
| 3 | 
         
             
            sqlalchemy>=1.4.2,<2
         
     | 
| 4 | 
         
             
            sqlalchemy_utils
         
     | 
| 5 | 
         
            +
            psycopg2-binary
         
     | 
| 6 | 
         
            +
            streamlit==1.19.0
         
     |