Update app.py
Browse files
app.py
CHANGED
@@ -128,14 +128,13 @@ def initialize_llmchain(temperature, max_tokens, top_k, vector_db, progress=gr.P
|
|
128 |
"load_in_8bit": True})
|
129 |
|
130 |
progress(0.75, desc="Defining buffer memory...")
|
131 |
-
|
132 |
#memory = ConversationTokenBufferMemory(llm = llm, max_token_limit=100)
|
133 |
# retriever=vector_db.as_retriever(search_type="similarity", search_kwargs={'k': 3})
|
134 |
retriever=vector_db.as_retriever()
|
135 |
progress(0.8, desc="Defining retrieval chain...")
|
136 |
qa_chain = ConversationalRetrievalChain.from_llm(llm,retriever=retriever,chain_type="stuff",
|
137 |
-
|
138 |
-
return_source_documents=True,verbose=False)
|
139 |
progress(0.9, desc="Done!")
|
140 |
return qa_chain
|
141 |
|
@@ -221,9 +220,6 @@ def upload_file(file_obj):
|
|
221 |
# initialize_database(file_path, progress)
|
222 |
return list_file_path
|
223 |
|
224 |
-
|
225 |
-
|
226 |
-
|
227 |
def demo():
|
228 |
with gr.Blocks(theme=gr.themes.Soft()) as demo:
|
229 |
vector_db = gr.State()
|
|
|
128 |
"load_in_8bit": True})
|
129 |
|
130 |
progress(0.75, desc="Defining buffer memory...")
|
131 |
+
memory = ConversationBufferMemory(memory_key="chat_history",output_key='answer',return_messages=True)
|
132 |
#memory = ConversationTokenBufferMemory(llm = llm, max_token_limit=100)
|
133 |
# retriever=vector_db.as_retriever(search_type="similarity", search_kwargs={'k': 3})
|
134 |
retriever=vector_db.as_retriever()
|
135 |
progress(0.8, desc="Defining retrieval chain...")
|
136 |
qa_chain = ConversationalRetrievalChain.from_llm(llm,retriever=retriever,chain_type="stuff",
|
137 |
+
memory=memory,return_source_documents=True,verbose=False)
|
|
|
138 |
progress(0.9, desc="Done!")
|
139 |
return qa_chain
|
140 |
|
|
|
220 |
# initialize_database(file_path, progress)
|
221 |
return list_file_path
|
222 |
|
|
|
|
|
|
|
223 |
def demo():
|
224 |
with gr.Blocks(theme=gr.themes.Soft()) as demo:
|
225 |
vector_db = gr.State()
|