wholewhale commited on
Commit
355932a
·
1 Parent(s): b1c579e

increase chunk size

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -16,14 +16,14 @@ def loading_pdf():
16
  def pdf_changes(pdf_doc):
17
  loader = OnlinePDFLoader(pdf_doc.name)
18
  documents = loader.load()
19
- text_splitter = CharacterTextSplitter(chunk_size=500, chunk_overlap=50)
20
  texts = text_splitter.split_documents(documents)
21
  embeddings = OpenAIEmbeddings()
22
  db = Chroma.from_documents(texts, embeddings)
23
  retriever = db.as_retriever()
24
  global qa
25
  qa = ConversationalRetrievalChain.from_llm(
26
- llm=OpenAI(temperature=0.5),
27
  retriever=retriever,
28
  return_source_documents=False)
29
  return "Ready"
@@ -39,7 +39,7 @@ def add_text(history, text):
39
 
40
  def bot(history):
41
  response = infer(history[-1][0], history)
42
- formatted_response = "**Bot:** \n" + ' \n'.join(response.split('. '))
43
  history[-1][1] = ""
44
 
45
  for character in formatted_response:
 
16
  def pdf_changes(pdf_doc):
17
  loader = OnlinePDFLoader(pdf_doc.name)
18
  documents = loader.load()
19
+ text_splitter = CharacterTextSplitter(chunk_size=1000, chunk_overlap=100)
20
  texts = text_splitter.split_documents(documents)
21
  embeddings = OpenAIEmbeddings()
22
  db = Chroma.from_documents(texts, embeddings)
23
  retriever = db.as_retriever()
24
  global qa
25
  qa = ConversationalRetrievalChain.from_llm(
26
+ llm=OpenAI(temperature=0.2, max_tokens=8000),
27
  retriever=retriever,
28
  return_source_documents=False)
29
  return "Ready"
 
39
 
40
  def bot(history):
41
  response = infer(history[-1][0], history)
42
+ formatted_response = "**AI:** \n" + ' \n'.join(response.split('. '))
43
  history[-1][1] = ""
44
 
45
  for character in formatted_response: