Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
Commit
·
355932a
1
Parent(s):
b1c579e
increase chunk size
Browse files
app.py
CHANGED
@@ -16,14 +16,14 @@ def loading_pdf():
|
|
16 |
def pdf_changes(pdf_doc):
|
17 |
loader = OnlinePDFLoader(pdf_doc.name)
|
18 |
documents = loader.load()
|
19 |
-
text_splitter = CharacterTextSplitter(chunk_size=
|
20 |
texts = text_splitter.split_documents(documents)
|
21 |
embeddings = OpenAIEmbeddings()
|
22 |
db = Chroma.from_documents(texts, embeddings)
|
23 |
retriever = db.as_retriever()
|
24 |
global qa
|
25 |
qa = ConversationalRetrievalChain.from_llm(
|
26 |
-
llm=OpenAI(temperature=0.
|
27 |
retriever=retriever,
|
28 |
return_source_documents=False)
|
29 |
return "Ready"
|
@@ -39,7 +39,7 @@ def add_text(history, text):
|
|
39 |
|
40 |
def bot(history):
|
41 |
response = infer(history[-1][0], history)
|
42 |
-
formatted_response = "**
|
43 |
history[-1][1] = ""
|
44 |
|
45 |
for character in formatted_response:
|
|
|
16 |
def pdf_changes(pdf_doc):
|
17 |
loader = OnlinePDFLoader(pdf_doc.name)
|
18 |
documents = loader.load()
|
19 |
+
text_splitter = CharacterTextSplitter(chunk_size=1000, chunk_overlap=100)
|
20 |
texts = text_splitter.split_documents(documents)
|
21 |
embeddings = OpenAIEmbeddings()
|
22 |
db = Chroma.from_documents(texts, embeddings)
|
23 |
retriever = db.as_retriever()
|
24 |
global qa
|
25 |
qa = ConversationalRetrievalChain.from_llm(
|
26 |
+
llm=OpenAI(temperature=0.2, max_tokens=8000),
|
27 |
retriever=retriever,
|
28 |
return_source_documents=False)
|
29 |
return "Ready"
|
|
|
39 |
|
40 |
def bot(history):
|
41 |
response = infer(history[-1][0], history)
|
42 |
+
formatted_response = "**AI:** \n" + ' \n'.join(response.split('. '))
|
43 |
history[-1][1] = ""
|
44 |
|
45 |
for character in formatted_response:
|