Update app.py
Browse files
app.py
CHANGED
@@ -208,16 +208,16 @@ def conversation(qa_chain, message, history):
|
|
208 |
response_source2 = response_sources[1].page_content.strip()
|
209 |
response_source3 = response_sources[2].page_content.strip()
|
210 |
# Langchain sources are zero-based
|
211 |
-
response_source1_page = response_sources[0].metadata["page"] + 1
|
212 |
-
response_source2_page = response_sources[1].metadata["page"] + 1
|
213 |
-
response_source3_page = response_sources[2].metadata["page"] + 1
|
214 |
# print ('chat response: ', response_answer)
|
215 |
# print('DB source', response_sources)
|
216 |
|
217 |
# Append user message and response to chat history
|
218 |
new_history = history + [(message, response_answer)]
|
219 |
# return gr.update(value=""), new_history, response_sources[0], response_sources[1]
|
220 |
-
return qa_chain, gr.update(value=""), new_history
|
221 |
|
222 |
|
223 |
def upload_file(file_obj):
|
@@ -294,7 +294,9 @@ def demo():
|
|
294 |
chatbot.change(preprocess = english_to_indian, postprocess = indian_to_english)
|
295 |
|
296 |
with gr.Row():
|
297 |
-
msg = gr.Textbox(placeholder="Type message", container=True)
|
|
|
|
|
298 |
with gr.Accordion("References", open=False):
|
299 |
with gr.Row():
|
300 |
doc_source1 = gr.Textbox(label="Reference 1", lines=2, container=True, scale=20)
|
@@ -305,7 +307,7 @@ def demo():
|
|
305 |
with gr.Row():
|
306 |
doc_source3 = gr.Textbox(label="Reference 3", lines=2, container=True, scale=20)
|
307 |
source3_page = gr.Number(label="Page", scale=1)
|
308 |
-
|
309 |
with gr.Row():
|
310 |
submit_btn = gr.Button("Submit")
|
311 |
clear_btn = gr.ClearButton([msg, chatbot])
|
@@ -319,21 +321,29 @@ def demo():
|
|
319 |
inputs=[slider_temperature, slider_maxtokens, slider_topk, vector_db], \
|
320 |
outputs=[qa_chain, llm_progress]).then(lambda:[None,"",0,"",0,"",0], \
|
321 |
inputs=None, \
|
322 |
-
outputs=[chatbot
|
|
|
|
|
323 |
queue=False)
|
324 |
|
325 |
# Chatbot events
|
326 |
msg.submit(conversation, \
|
327 |
inputs=[qa_chain, msg, chatbot], \
|
328 |
-
outputs=[qa_chain, msg, chatbot
|
|
|
|
|
329 |
queue=False)
|
330 |
submit_btn.click(conversation, \
|
331 |
inputs=[qa_chain, msg, chatbot], \
|
332 |
-
outputs=[qa_chain, msg, chatbot
|
|
|
|
|
333 |
queue=False)
|
334 |
clear_btn.click(lambda:[None,"",0,"",0,"",0], \
|
335 |
inputs=None, \
|
336 |
-
outputs=[chatbot
|
|
|
|
|
337 |
queue=False)
|
338 |
demo.queue().launch(debug=True)
|
339 |
|
|
|
208 |
response_source2 = response_sources[1].page_content.strip()
|
209 |
response_source3 = response_sources[2].page_content.strip()
|
210 |
# Langchain sources are zero-based
|
211 |
+
#response_source1_page = response_sources[0].metadata["page"] + 1
|
212 |
+
#response_source2_page = response_sources[1].metadata["page"] + 1
|
213 |
+
#response_source3_page = response_sources[2].metadata["page"] + 1
|
214 |
# print ('chat response: ', response_answer)
|
215 |
# print('DB source', response_sources)
|
216 |
|
217 |
# Append user message and response to chat history
|
218 |
new_history = history + [(message, response_answer)]
|
219 |
# return gr.update(value=""), new_history, response_sources[0], response_sources[1]
|
220 |
+
return qa_chain, gr.update(value=""), new_history#, response_source1, response_source1_page, response_source2, response_source2_page, response_source3, response_source3_page
|
221 |
|
222 |
|
223 |
def upload_file(file_obj):
|
|
|
294 |
chatbot.change(preprocess = english_to_indian, postprocess = indian_to_english)
|
295 |
|
296 |
with gr.Row():
|
297 |
+
msg = gr.Textbox(placeholder="Type message", container=True)
|
298 |
+
|
299 |
+
'''
|
300 |
with gr.Accordion("References", open=False):
|
301 |
with gr.Row():
|
302 |
doc_source1 = gr.Textbox(label="Reference 1", lines=2, container=True, scale=20)
|
|
|
307 |
with gr.Row():
|
308 |
doc_source3 = gr.Textbox(label="Reference 3", lines=2, container=True, scale=20)
|
309 |
source3_page = gr.Number(label="Page", scale=1)
|
310 |
+
'''
|
311 |
with gr.Row():
|
312 |
submit_btn = gr.Button("Submit")
|
313 |
clear_btn = gr.ClearButton([msg, chatbot])
|
|
|
321 |
inputs=[slider_temperature, slider_maxtokens, slider_topk, vector_db], \
|
322 |
outputs=[qa_chain, llm_progress]).then(lambda:[None,"",0,"",0,"",0], \
|
323 |
inputs=None, \
|
324 |
+
outputs=[chatbot
|
325 |
+
#, doc_source1, source1_page, doc_source2, source2_page, doc_source3, source3_page
|
326 |
+
], \
|
327 |
queue=False)
|
328 |
|
329 |
# Chatbot events
|
330 |
msg.submit(conversation, \
|
331 |
inputs=[qa_chain, msg, chatbot], \
|
332 |
+
outputs=[qa_chain, msg, chatbot
|
333 |
+
#, doc_source1, source1_page, doc_source2, source2_page, doc_source3, source3_page
|
334 |
+
], \
|
335 |
queue=False)
|
336 |
submit_btn.click(conversation, \
|
337 |
inputs=[qa_chain, msg, chatbot], \
|
338 |
+
outputs=[qa_chain, msg, chatbot
|
339 |
+
#, doc_source1, source1_page, doc_source2, source2_page, doc_source3, source3_page
|
340 |
+
], \
|
341 |
queue=False)
|
342 |
clear_btn.click(lambda:[None,"",0,"",0,"",0], \
|
343 |
inputs=None, \
|
344 |
+
outputs=[chatbot
|
345 |
+
#, doc_source1, source1_page, doc_source2, source2_page, doc_source3, source3_page
|
346 |
+
], \
|
347 |
queue=False)
|
348 |
demo.queue().launch(debug=True)
|
349 |
|