vishwask commited on
Commit
2e01859
·
verified ·
1 Parent(s): c407349

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -4
app.py CHANGED
@@ -73,7 +73,7 @@ def load_db():
73
 
74
 
75
  # Initialize langchain LLM chain
76
- def initialize_llmchain(llm_model, temperature, max_tokens, top_k, vector_db, progress=gr.Progress()):
77
  progress(0.1, desc="Initializing HF tokenizer...")
78
 
79
  # HuggingFaceHub uses HF inference endpoints
@@ -140,11 +140,11 @@ def initialize_database(list_file_obj, chunk_size, chunk_overlap, progress=gr.Pr
140
  return vector_db, collection_name, "Complete!"
141
 
142
 
143
- def initialize_LLM(llm_option, llm_temperature, max_tokens, top_k, vector_db, progress=gr.Progress()):
144
  # print("llm_option",llm_option)
145
  llm_name = llm_model
146
  print("llm_name: ",llm_name)
147
- qa_chain = initialize_llmchain(llm_name, llm_temperature, max_tokens, top_k, vector_db, progress)
148
  return qa_chain, "Complete!"
149
 
150
 
@@ -266,7 +266,7 @@ def demo():
266
  inputs=[document, slider_chunk_size, slider_chunk_overlap], \
267
  outputs=[vector_db, collection_name, db_progress])
268
  qachain_btn.click(initialize_LLM, \
269
- inputs=[llm_model, slider_temperature, slider_maxtokens, slider_topk, vector_db], \
270
  outputs=[qa_chain, llm_progress]).then(lambda:[None,"",0,"",0,"",0], \
271
  inputs=None, \
272
  outputs=[chatbot, doc_source1, source1_page, doc_source2, source2_page, doc_source3, source3_page], \
 
73
 
74
 
75
  # Initialize langchain LLM chain
76
+ def initialize_llmchain(temperature, max_tokens, top_k, vector_db, progress=gr.Progress()):
77
  progress(0.1, desc="Initializing HF tokenizer...")
78
 
79
  # HuggingFaceHub uses HF inference endpoints
 
140
  return vector_db, collection_name, "Complete!"
141
 
142
 
143
+ def initialize_LLM(llm_temperature, max_tokens, top_k, vector_db, progress=gr.Progress()):
144
  # print("llm_option",llm_option)
145
  llm_name = llm_model
146
  print("llm_name: ",llm_name)
147
+ qa_chain = initialize_llmchain(llm_temperature, max_tokens, top_k, vector_db, progress)
148
  return qa_chain, "Complete!"
149
 
150
 
 
266
  inputs=[document, slider_chunk_size, slider_chunk_overlap], \
267
  outputs=[vector_db, collection_name, db_progress])
268
  qachain_btn.click(initialize_LLM, \
269
+ inputs=[slider_temperature, slider_maxtokens, slider_topk, vector_db], \
270
  outputs=[qa_chain, llm_progress]).then(lambda:[None,"",0,"",0,"",0], \
271
  inputs=None, \
272
  outputs=[chatbot, doc_source1, source1_page, doc_source2, source2_page, doc_source3, source3_page], \