Update app.py
Browse files
app.py
CHANGED
@@ -56,14 +56,23 @@ retriever = vectordb.as_retriever(
|
|
56 |
search_type="similarity", search_kwargs={"k": 2}
|
57 |
)
|
58 |
|
59 |
-
from langchain.chains import RetrievalQA
|
60 |
-
|
61 |
|
|
|
|
|
|
|
62 |
|
63 |
READER_MODEL="HuggingFaceH4/zephyr-7b-beta"
|
64 |
#HuggingFaceH4/zephyr-7b-beta
|
65 |
#qa = ConversationalRetrievalChain.from_llm(llm=READER_MODEL,retriever=retriever,memory=memory)
|
66 |
-
qa = RetrievalQA.from_chain_type(llm=READER_MODEL,chain_type="map_reduce",retriever=retriever,verbose=True)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
67 |
result = qa(question)
|
68 |
import gradio as gr
|
69 |
gr.load("READER_MODEL").launch()
|
|
|
56 |
search_type="similarity", search_kwargs={"k": 2}
|
57 |
)
|
58 |
|
59 |
+
#from langchain.chains import RetrievalQA
|
|
|
60 |
|
61 |
+
from langchain.chains import create_retrieval_chain
|
62 |
+
from langchain.chains.combine_documents import create_stuff_documents_chain
|
63 |
+
from langchain import hub
|
64 |
|
65 |
READER_MODEL="HuggingFaceH4/zephyr-7b-beta"
|
66 |
#HuggingFaceH4/zephyr-7b-beta
|
67 |
#qa = ConversationalRetrievalChain.from_llm(llm=READER_MODEL,retriever=retriever,memory=memory)
|
68 |
+
#qa = RetrievalQA.from_chain_type(llm=READER_MODEL,chain_type="map_reduce",retriever=retriever,verbose=True)
|
69 |
+
|
70 |
+
retrieval_qa_chat_prompt = hub.pull("langchain-ai/retrieval-qa-chat")
|
71 |
+
combine_docs_chain = create_stuff_documents_chain(
|
72 |
+
READER_MODEL, retrieval_qa_chat_prompt
|
73 |
+
)
|
74 |
+
qa = create_retrieval_chain(retriever, combine_docs_chain)
|
75 |
+
|
76 |
result = qa(question)
|
77 |
import gradio as gr
|
78 |
gr.load("READER_MODEL").launch()
|