Update app.py
Browse files
app.py
CHANGED
@@ -1,5 +1,6 @@
|
|
1 |
from langchain.prompts.prompt import PromptTemplate
|
2 |
-
from langchain.llms import OpenAI
|
|
|
3 |
from langchain.chains import ChatVectorDBChain
|
4 |
from langchain.embeddings import HuggingFaceEmbeddings, HuggingFaceInstructEmbeddings
|
5 |
from langchain.callbacks.base import CallbackManager
|
@@ -11,6 +12,8 @@ import gradio as gr
|
|
11 |
import pickle
|
12 |
from threading import Lock
|
13 |
|
|
|
|
|
14 |
model_options = {'all-mpnet-base-v2': "sentence-transformers/all-mpnet-base-v2",
|
15 |
'instructor-base': "hkunlp/instructor-base"}
|
16 |
|
@@ -64,7 +67,8 @@ QA_PROMPT = PromptTemplate(template=template, input_variables=["question", "cont
|
|
64 |
|
65 |
|
66 |
def get_chain(vectorstore):
|
67 |
-
llm =
|
|
|
68 |
qa_chain = ChatVectorDBChain.from_llm(
|
69 |
llm,
|
70 |
vectorstore,
|
|
|
1 |
from langchain.prompts.prompt import PromptTemplate
|
2 |
+
# from langchain.llms import OpenAI
|
3 |
+
from langchain.llms import OpenAIChat
|
4 |
from langchain.chains import ChatVectorDBChain
|
5 |
from langchain.embeddings import HuggingFaceEmbeddings, HuggingFaceInstructEmbeddings
|
6 |
from langchain.callbacks.base import CallbackManager
|
|
|
12 |
import pickle
|
13 |
from threading import Lock
|
14 |
|
15 |
+
prefix_messages = [{"role": "system", "content": "You are a helpful assistant that is very good at answering questions about investments using the information given."}]
|
16 |
+
|
17 |
model_options = {'all-mpnet-base-v2': "sentence-transformers/all-mpnet-base-v2",
|
18 |
'instructor-base': "hkunlp/instructor-base"}
|
19 |
|
|
|
67 |
|
68 |
|
69 |
def get_chain(vectorstore):
|
70 |
+
llm = OpenAIChat(streaming=True, callback_manager=CallbackManager([StreamingStdOutCallbackHandler()]), verbose=True, temperature=0,\
|
71 |
+
prefix_messages=prefix_messages)
|
72 |
qa_chain = ChatVectorDBChain.from_llm(
|
73 |
llm,
|
74 |
vectorstore,
|