Update app.py
Browse files
app.py
CHANGED
@@ -59,14 +59,14 @@ retriever = vectordb.as_retriever(
|
|
59 |
#from langchain.chains import RetrievalQA
|
60 |
from langchain_core.prompts import ChatPromptTemplate
|
61 |
|
62 |
-
from langchain import create_retrieval_chain
|
63 |
from langchain.chains.combine_documents import create_stuff_documents_chain
|
64 |
#from langchain import hub
|
65 |
|
66 |
|
67 |
-
|
68 |
#HuggingFaceH4/zephyr-7b-beta
|
69 |
-
READER_MODEL=Ollama(model="meta-llama/Meta-Llama-Guard-2-8B")
|
70 |
#qa = ConversationalRetrievalChain.from_llm(llm=READER_MODEL,retriever=retriever,memory=memory)
|
71 |
#qa = RetrievalQA.from_chain_type(llm=READER_MODEL,chain_type="map_reduce",retriever=retriever,verbose=True)
|
72 |
|
|
|
59 |
#from langchain.chains import RetrievalQA
|
60 |
from langchain_core.prompts import ChatPromptTemplate
|
61 |
|
62 |
+
from langchain.chains import create_retrieval_chain
|
63 |
from langchain.chains.combine_documents import create_stuff_documents_chain
|
64 |
#from langchain import hub
|
65 |
|
66 |
|
67 |
+
READER_MODEL="HuggingFaceH4/zephyr-7b-beta"
|
68 |
#HuggingFaceH4/zephyr-7b-beta
|
69 |
+
#READER_MODEL=Ollama(model="meta-llama/Meta-Llama-Guard-2-8B")
|
70 |
#qa = ConversationalRetrievalChain.from_llm(llm=READER_MODEL,retriever=retriever,memory=memory)
|
71 |
#qa = RetrievalQA.from_chain_type(llm=READER_MODEL,chain_type="map_reduce",retriever=retriever,verbose=True)
|
72 |
|