Update app.py
Browse files
app.py
CHANGED
@@ -68,17 +68,15 @@ def load_db():
|
|
68 |
return vectordb
|
69 |
|
70 |
# Initialize langchain LLM chain
|
71 |
-
|
72 |
-
|
73 |
-
|
74 |
-
|
75 |
-
|
76 |
-
|
77 |
-
|
78 |
-
|
79 |
-
|
80 |
-
|
81 |
-
#return qa_chain
|
82 |
|
83 |
|
84 |
|
@@ -155,9 +153,7 @@ def conversation(qa_chain, message, history):
|
|
155 |
#document = os.listdir(list_file_obj)
|
156 |
vector_db, collection_name = initialize_database(list_file_obj)
|
157 |
#qa_chain =
|
158 |
-
|
159 |
-
qa_chain = ConversationalRetrievalChain.from_llm(llm,retriever=retriever,chain_type="stuff",
|
160 |
-
memory=memory,return_source_documents=True,verbose=False,)
|
161 |
print('qa chain and vector_db done')
|
162 |
|
163 |
def demo():
|
|
|
68 |
return vectordb
|
69 |
|
70 |
# Initialize langchain LLM chain
|
71 |
+
llm = HuggingFaceHub(repo_id = llm_model,
|
72 |
+
model_kwargs={"temperature": temperature,
|
73 |
+
"max_new_tokens": max_tokens,
|
74 |
+
"top_k": top_k,
|
75 |
+
"load_in_8bit": True})
|
76 |
+
retriever=vector_db.as_retriever()
|
77 |
+
memory = ConversationBufferMemory(memory_key="chat_history", output_key='answer', return_messages=True)
|
78 |
+
qa_chain = ConversationalRetrievalChain.from_llm(llm_model,retriever=retriever,chain_type="stuff",
|
79 |
+
memory=memory,return_source_documents=True,verbose=False,)
|
|
|
|
|
80 |
|
81 |
|
82 |
|
|
|
153 |
#document = os.listdir(list_file_obj)
|
154 |
vector_db, collection_name = initialize_database(list_file_obj)
|
155 |
#qa_chain =
|
156 |
+
|
|
|
|
|
157 |
print('qa chain and vector_db done')
|
158 |
|
159 |
def demo():
|