Commit
·
2ac93d3
1
Parent(s):
9c8b73f
Add demo
Browse files- .gitattributes +2 -0
- README.md +1 -1
- __pycache__/gradio_demo.cpython-311.pyc +0 -0
- __pycache__/prompt.cpython-311.pyc +0 -0
- __pycache__/utils.cpython-311.pyc +0 -0
- gradio_demo.py +127 -0
- icon_1-modified.png +3 -0
- icon_1.jpeg +3 -0
- icon_2-modified.png +3 -0
- icon_2.jpeg +3 -0
- prompt.py +202 -0
- requirements.txt +6 -0
- utils.py +217 -0
.gitattributes
CHANGED
@@ -7,6 +7,7 @@
|
|
7 |
*.gz filter=lfs diff=lfs merge=lfs -text
|
8 |
*.h5 filter=lfs diff=lfs merge=lfs -text
|
9 |
*.joblib filter=lfs diff=lfs merge=lfs -text
|
|
|
10 |
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
11 |
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
12 |
*.model filter=lfs diff=lfs merge=lfs -text
|
@@ -19,6 +20,7 @@
|
|
19 |
*.pb filter=lfs diff=lfs merge=lfs -text
|
20 |
*.pickle filter=lfs diff=lfs merge=lfs -text
|
21 |
*.pkl filter=lfs diff=lfs merge=lfs -text
|
|
|
22 |
*.pt filter=lfs diff=lfs merge=lfs -text
|
23 |
*.pth filter=lfs diff=lfs merge=lfs -text
|
24 |
*.rar filter=lfs diff=lfs merge=lfs -text
|
|
|
7 |
*.gz filter=lfs diff=lfs merge=lfs -text
|
8 |
*.h5 filter=lfs diff=lfs merge=lfs -text
|
9 |
*.joblib filter=lfs diff=lfs merge=lfs -text
|
10 |
+
*.jpeg filter=lfs diff=lfs merge=lfs -text
|
11 |
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
12 |
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
13 |
*.model filter=lfs diff=lfs merge=lfs -text
|
|
|
20 |
*.pb filter=lfs diff=lfs merge=lfs -text
|
21 |
*.pickle filter=lfs diff=lfs merge=lfs -text
|
22 |
*.pkl filter=lfs diff=lfs merge=lfs -text
|
23 |
+
*.png filter=lfs diff=lfs merge=lfs -text
|
24 |
*.pt filter=lfs diff=lfs merge=lfs -text
|
25 |
*.pth filter=lfs diff=lfs merge=lfs -text
|
26 |
*.rar filter=lfs diff=lfs merge=lfs -text
|
README.md
CHANGED
@@ -5,7 +5,7 @@ colorFrom: green
|
|
5 |
colorTo: pink
|
6 |
sdk: gradio
|
7 |
sdk_version: 3.32.0
|
8 |
-
app_file:
|
9 |
pinned: false
|
10 |
---
|
11 |
|
|
|
5 |
colorTo: pink
|
6 |
sdk: gradio
|
7 |
sdk_version: 3.32.0
|
8 |
+
app_file: gradio_demo.py
|
9 |
pinned: false
|
10 |
---
|
11 |
|
__pycache__/gradio_demo.cpython-311.pyc
ADDED
Binary file (8.57 kB). View file
|
|
__pycache__/prompt.cpython-311.pyc
ADDED
Binary file (10.1 kB). View file
|
|
__pycache__/utils.cpython-311.pyc
ADDED
Binary file (10.8 kB). View file
|
|
gradio_demo.py
ADDED
@@ -0,0 +1,127 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
import gradio as gr
|
2 |
+
from langchain.chat_models import ChatOpenAI
|
3 |
+
from langchain.schema import HumanMessage
|
4 |
+
from utils import SyncStreamingLLMCallbackHandler
|
5 |
+
from anyio.from_thread import start_blocking_portal
|
6 |
+
from queue import Queue
|
7 |
+
import prompt, tiktoken
|
8 |
+
|
9 |
+
def num_token(string: str) -> int:
|
10 |
+
"""Returns the number of tokens in a text string."""
|
11 |
+
encoding = tiktoken.get_encoding('cl100k_base')
|
12 |
+
num_tokens = len(encoding.encode(string))
|
13 |
+
return num_tokens
|
14 |
+
|
15 |
+
def send_message(history, temp, top_p, fp, pp):
|
16 |
+
q = Queue()
|
17 |
+
job_done = object()
|
18 |
+
def task():
|
19 |
+
llm = ChatOpenAI(
|
20 |
+
model_name="gpt-3.5-turbo",
|
21 |
+
streaming=True,
|
22 |
+
temperature=temp,
|
23 |
+
model_kwargs = {
|
24 |
+
'top_p':top_p,
|
25 |
+
'frequency_penalty':fp,
|
26 |
+
'presence_penalty':pp
|
27 |
+
}
|
28 |
+
)
|
29 |
+
query_llm = ChatOpenAI(
|
30 |
+
model_name="gpt-3.5-turbo",
|
31 |
+
temperature = 0
|
32 |
+
)
|
33 |
+
query_template = prompt.search_query_gradio(history)
|
34 |
+
search_query = query_llm([HumanMessage(content=query_template)]).content
|
35 |
+
print('query: ', search_query)
|
36 |
+
answer_template = prompt.question_answer_gradio(history, search_query)
|
37 |
+
print('final temp: \n\n', answer_template, f'\nprompt total: {num_token(answer_template)}')
|
38 |
+
llm([HumanMessage(content=answer_template)], callbacks=[SyncStreamingLLMCallbackHandler(q)])
|
39 |
+
q.put(job_done)
|
40 |
+
return
|
41 |
+
|
42 |
+
with start_blocking_portal() as portal:
|
43 |
+
portal.start_task_soon(task)
|
44 |
+
|
45 |
+
content = ""
|
46 |
+
while True:
|
47 |
+
next_token = q.get(True, timeout=10)
|
48 |
+
if next_token is job_done:
|
49 |
+
return history
|
50 |
+
|
51 |
+
content += next_token
|
52 |
+
latest = history[-1]
|
53 |
+
latest[1] = content
|
54 |
+
yield history
|
55 |
+
|
56 |
+
|
57 |
+
def append_user_message(message, history):
|
58 |
+
"Add user message to chain/history and clear input field"
|
59 |
+
print(history)
|
60 |
+
history.append([message, None])
|
61 |
+
return "", history
|
62 |
+
|
63 |
+
css = """
|
64 |
+
.contain { display: flex; flex-direction: column; height: 100%;}
|
65 |
+
#chatbot { flex-grow: 1;}
|
66 |
+
#chatbot .user {text-align: right;}
|
67 |
+
|
68 |
+
.contain { display: flex; flex-direction: column; height: 100%;}
|
69 |
+
#query_ref { flex-grow: 1; }
|
70 |
+
"""
|
71 |
+
|
72 |
+
|
73 |
+
with gr.Blocks(css=css, theme=gr.themes.Default(primary_hue="orange")) as demo:
|
74 |
+
|
75 |
+
with gr.Row(elem_id="whole_page"):
|
76 |
+
with gr.Column(elem_id="sidebar", scale=0.2, min_width=0):
|
77 |
+
# TODO temperature?
|
78 |
+
gr.Markdown('## City of Lake Elsinore Proposal Demo', show_label=False)
|
79 |
+
with gr.Row(elem_id="logo_section1"):
|
80 |
+
with gr.Column(elem_id="logo_col1", scale=0.5, min_width=0):
|
81 |
+
gr.HTML("""<div style="display: flex; justify-content: center; align-items: center"><img src="./file=icon_1-modified.png" alt="ThinkCol" width="75" height="87" /></a></div>""")
|
82 |
+
with gr.Column(elem_id="logo_col2", scale=0.5, min_width=0):
|
83 |
+
gr.HTML("""<div style="display: flex; justify-content: center; align-items: center"><img src="./file=icon_2-modified.png" alt="ThinkCol" width="75" height="87" /></a></div>""")
|
84 |
+
gr.Markdown('### Chatbot Settings', show_label=False)
|
85 |
+
temp = gr.Slider(maximum=1, label='Temperature')
|
86 |
+
top_p = gr.Slider(maximum=1, label='Top_p', value=1, interactive=True, visible='primary')
|
87 |
+
fp = gr.Slider(maximum=2, label='Frequency Penalty')
|
88 |
+
pp = gr.Slider(maximum=2, label='Presence Penalty')
|
89 |
+
|
90 |
+
with gr.Column(elem_id="main_chat_interface", scale=0.5):
|
91 |
+
chatbot = gr.Chatbot([], elem_id="chatbot")
|
92 |
+
with gr.Row(elem_id="box_split"):
|
93 |
+
with gr.Column(elem_id="enter_box",scale=0.85):
|
94 |
+
txt = gr.Textbox(
|
95 |
+
elem_id='input',
|
96 |
+
show_label=False,
|
97 |
+
placeholder="Enter text and press enter, or upload an image"
|
98 |
+
).style(container=False)
|
99 |
+
txt.submit(append_user_message, [txt, chatbot], [txt, chatbot]) \
|
100 |
+
.then(send_message, [chatbot, temp, top_p, fp, pp], [chatbot])
|
101 |
+
|
102 |
+
with gr.Column(elem_id="send_box",scale=0.15, min_width=0):
|
103 |
+
btn = gr.Button('Send', elem_id='send', variant='primary')
|
104 |
+
btn.click(append_user_message, [txt, chatbot], [txt, chatbot]) \
|
105 |
+
.then(send_message, [chatbot, temp, top_p, fp, pp], [chatbot])
|
106 |
+
|
107 |
+
|
108 |
+
# with gr.Column(elem_id="main_chat_interface", scale=0.3):
|
109 |
+
# with gr.Tab("Search Query"):
|
110 |
+
# query_ref = gr.TextArea(
|
111 |
+
# value='',
|
112 |
+
# interactive=False,
|
113 |
+
# elem_id='query_ref',
|
114 |
+
# show_label=False,
|
115 |
+
# ).style(container=False)
|
116 |
+
# with gr.Tab("Prompt"):
|
117 |
+
# prompt_ref = gr.TextArea(
|
118 |
+
# value='',
|
119 |
+
# interactive=False,
|
120 |
+
# elem_id='prompt_ref',
|
121 |
+
# show_label=False,
|
122 |
+
# ).style(container=False)
|
123 |
+
# print(query_ref, prompt_ref)
|
124 |
+
|
125 |
+
demo.queue()
|
126 |
+
if __name__ == "__main__":
|
127 |
+
demo.launch()
|
icon_1-modified.png
ADDED
![]() |
Git LFS Details
|
icon_1.jpeg
ADDED
![]() |
Git LFS Details
|
icon_2-modified.png
ADDED
![]() |
Git LFS Details
|
icon_2.jpeg
ADDED
![]() |
Git LFS Details
|
prompt.py
ADDED
@@ -0,0 +1,202 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
import os
|
2 |
+
import tiktoken
|
3 |
+
import requests, json
|
4 |
+
|
5 |
+
|
6 |
+
def num_token(string: str) -> int:
|
7 |
+
"""Returns the number of tokens in a text string."""
|
8 |
+
encoding = tiktoken.get_encoding('cl100k_base')
|
9 |
+
num_tokens = len(encoding.encode(string))
|
10 |
+
return num_tokens
|
11 |
+
|
12 |
+
def qa_prompt(context_str, query_str):
|
13 |
+
temp = (
|
14 |
+
"Context information is below. \n"
|
15 |
+
"---------------------\n"
|
16 |
+
"{context_str}"
|
17 |
+
"\n---------------------\n"
|
18 |
+
"Given the context information and not prior knowledge, "
|
19 |
+
"answer the question: {query_str}\n"
|
20 |
+
)
|
21 |
+
temp = temp.format(context_str=context_str, query_str=query_str)
|
22 |
+
print('qa_prompt token: ', num_token(temp))
|
23 |
+
return temp
|
24 |
+
|
25 |
+
|
26 |
+
|
27 |
+
def refine_prompt(context_str, query_str, existing_answer):
|
28 |
+
temp = (
|
29 |
+
"The original question is as follows: {query_str}\n"
|
30 |
+
"We have provided an existing answer: {existing_answer}\n"
|
31 |
+
"Given the new context, refine the original answer to better "
|
32 |
+
"answer the question. However, "
|
33 |
+
"------------\n"
|
34 |
+
"{context_str}\n"
|
35 |
+
"------------\n"
|
36 |
+
"If the context isn't useful, return the original answer as new answer."
|
37 |
+
"New answer:"
|
38 |
+
)
|
39 |
+
temp = temp.format(
|
40 |
+
context_str=context_str,
|
41 |
+
query_str=query_str,
|
42 |
+
existing_answer=existing_answer
|
43 |
+
)
|
44 |
+
print('qa_prompt token: ', num_token(temp))
|
45 |
+
return temp
|
46 |
+
|
47 |
+
|
48 |
+
def qa_prompt_t(query_str):
|
49 |
+
temp = (
|
50 |
+
"Context information is below. \n"
|
51 |
+
"---------------------\n"
|
52 |
+
"{text}"
|
53 |
+
"\n---------------------\n"
|
54 |
+
"Given the context information and not prior knowledge, "
|
55 |
+
"answer the question: {query_str}\n"
|
56 |
+
)
|
57 |
+
return temp.replace('{query_str}', query_str)
|
58 |
+
|
59 |
+
|
60 |
+
def refine_prompt_t(query_str):
|
61 |
+
temp = (
|
62 |
+
"The original question is as follows: {query_str}\n"
|
63 |
+
"We have provided an existing answer: {existing_answer}\n"
|
64 |
+
"Given the new context, refine the original answer to better "
|
65 |
+
"answer the question. However, "
|
66 |
+
"------------\n"
|
67 |
+
"{text}\n"
|
68 |
+
"------------\n"
|
69 |
+
"If the context isn't useful, return the original answer as new answer."
|
70 |
+
"New answer:"
|
71 |
+
)
|
72 |
+
return temp.replace('{query_str}', query_str)
|
73 |
+
|
74 |
+
def language_detection(query_str):
|
75 |
+
detect_template = (
|
76 |
+
"Determine if the QUERY want to be answered in English/繁體中文/简体中文 (ONLY choose one).\n"
|
77 |
+
f"QUERY: {query_str}"
|
78 |
+
"LANGUAGE of the answer should be (default English if you don't know):"
|
79 |
+
)
|
80 |
+
return detect_template
|
81 |
+
|
82 |
+
# rephrased text is the current text.
|
83 |
+
def rephrase_query(query_str, history, company, language):
|
84 |
+
rephrase_template = (
|
85 |
+
f"Given the following conversation and a new input from user, rephrase it (in {language}) to a standalone input.\n\n"
|
86 |
+
|
87 |
+
"Chat History:\n"
|
88 |
+
f"{history}\n"
|
89 |
+
|
90 |
+
f"User (New Input): {query_str}\n"
|
91 |
+
"Standalone Input (rephrase according to User's perspective):"
|
92 |
+
)
|
93 |
+
return rephrase_template
|
94 |
+
|
95 |
+
|
96 |
+
def rephrase_query_hkas(query_str, history, language):
|
97 |
+
rephrase_template = (
|
98 |
+
f"Rewrite the new QUERY in {language} given the context of the previous QA history, "
|
99 |
+
"such that the objective and subject should be clear in the REWRITED QUERY.\n\n "
|
100 |
+
|
101 |
+
"QA history:\n"
|
102 |
+
"===\n"
|
103 |
+
f"{history}\n"
|
104 |
+
"===\n\n"
|
105 |
+
|
106 |
+
f"QUERY: {query_str} \n"
|
107 |
+
f"REWRITED QUERY ONLY (no description):"
|
108 |
+
)
|
109 |
+
return rephrase_template
|
110 |
+
|
111 |
+
def rephrase_query_chat(query, history, language):
|
112 |
+
rephrase_template = (
|
113 |
+
f"Given the following conversation and a new input from user, rephrase it (in {language}) to a new standalone input.\n\n"
|
114 |
+
|
115 |
+
"Chat History:\n"
|
116 |
+
f"{history}\n"
|
117 |
+
|
118 |
+
f"New Input: {query}\n"
|
119 |
+
"Standalone Input (New Input perspective):"
|
120 |
+
)
|
121 |
+
return rephrase_template
|
122 |
+
|
123 |
+
|
124 |
+
def search_query_gradio(chat_history):
|
125 |
+
history = []
|
126 |
+
for i in chat_history:
|
127 |
+
if i[1]:
|
128 |
+
history.append(f'User: {i[0]}')
|
129 |
+
history.append(f'Assistant: {i[1]}')
|
130 |
+
else:
|
131 |
+
question = i[0]
|
132 |
+
num = [num_token(record) for record in history]
|
133 |
+
cum_sum = [abs(sum(num[-i-1::]) - 700) for i in range(len(num))]
|
134 |
+
cum_sum = cum_sum[::-1]
|
135 |
+
if cum_sum:
|
136 |
+
idx = cum_sum.index(min(cum_sum))
|
137 |
+
format_history = '\n'.join(i for i in history[idx:])
|
138 |
+
else:
|
139 |
+
format_history = ''
|
140 |
+
print(format_history)
|
141 |
+
template = """
|
142 |
+
Below is a history of the conversation so far, and a new question asked by the user that needs to be answered by searching in a knowledge base. Generate a search query based on the conversation and the new question. Do not include cited source filenames and document names e.g info.txt or doc.pdf or https://web.com in the search query terms. Do not include any text inside [] or <<>> in the search query terms.
|
143 |
+
|
144 |
+
Chat History:
|
145 |
+
{format_history}
|
146 |
+
|
147 |
+
Question:
|
148 |
+
{question}
|
149 |
+
|
150 |
+
|
151 |
+
Search query in the language used in Question:
|
152 |
+
"""
|
153 |
+
|
154 |
+
return template.format(format_history=format_history, question=question)
|
155 |
+
|
156 |
+
|
157 |
+
def question_answer_gradio(chat_history, search_query):
|
158 |
+
|
159 |
+
docs_string = requests.get(f'http://{os.environ["QUESTION_ANSWER"]}/retrieve_docs/comsec/{search_query}')
|
160 |
+
docs_dict = json.loads(docs_string.json())
|
161 |
+
source_lst = []
|
162 |
+
for i, k in docs_dict.items():
|
163 |
+
source_lst.append(f"{i}: {k}")
|
164 |
+
print('can do source lst: ', source_lst)
|
165 |
+
format_source = '\n\n'.join(source_lst)
|
166 |
+
# TODO number of docs
|
167 |
+
|
168 |
+
|
169 |
+
|
170 |
+
|
171 |
+
history = []
|
172 |
+
for i in chat_history:
|
173 |
+
if i[1]:
|
174 |
+
history.append(f'User: {i[0]}')
|
175 |
+
history.append(f'Assistant: {i[1]}')
|
176 |
+
else:
|
177 |
+
history.append(f'User: {i[0]}')
|
178 |
+
num = [num_token(record) for record in history]
|
179 |
+
cum_sum = [abs(sum(num[-i-1::]) - 700) for i in range(len(num))]
|
180 |
+
cum_sum = cum_sum[::-1]
|
181 |
+
if cum_sum:
|
182 |
+
idx = cum_sum.index(min(cum_sum))
|
183 |
+
format_history = '\n'.join(i for i in history[idx:])
|
184 |
+
else:
|
185 |
+
format_history = ''
|
186 |
+
print('history ok: ', format_history)
|
187 |
+
# Each source has a name followed by colon and the actual information, ALWAYS include the source name for each fact you use in the response. Use square brakets to reference the source, e.g. [info1.txt]. Don't combine sources, list each source separately, e.g. [info1.txt][info2.pdf].
|
188 |
+
template = """<|im_start|>
|
189 |
+
Assistant helps the user with their questions about City of Lake Elsinore Proposal. Be brief in your answers with appropriate tone and emotion. Answer ONLY with the facts listed in the list of sources below, where each source has a name followed by colon and the actual information. If there isn't enough information below, say you don't know. Do not generate answers that don't use the sources below. If asking a clarifying question to the user would help, ask the question. For tabular information return it as an html table. Do not return markdown format.
|
190 |
+
|
191 |
+
Sources:
|
192 |
+
{sources}
|
193 |
+
<|im_end|>
|
194 |
+
|
195 |
+
{chat_history}
|
196 |
+
Assistant:
|
197 |
+
"""
|
198 |
+
|
199 |
+
token_for_doc = 3700 - num_token(format_history) - num_token(template)
|
200 |
+
print('token for doc: ', token_for_doc)
|
201 |
+
|
202 |
+
return template.format(sources=format_source, chat_history=format_history)
|
requirements.txt
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
gradio==3.32.0
|
2 |
+
langchain==0.0.183
|
3 |
+
tiktoken==0.3.1
|
4 |
+
requests==2.28.2
|
5 |
+
anyio==3.6.2
|
6 |
+
openai
|
utils.py
ADDED
@@ -0,0 +1,217 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
from langchain.llms import OpenAI
|
2 |
+
from langchain.prompts import PromptTemplate
|
3 |
+
from langchain.chains.qa_with_sources import load_qa_with_sources_chain
|
4 |
+
from langchain.chat_models import ChatOpenAI
|
5 |
+
from langchain.schema import HumanMessage
|
6 |
+
import tiktoken
|
7 |
+
|
8 |
+
|
9 |
+
def num_token(string: str) -> int:
|
10 |
+
"""Returns the number of tokens in a text string."""
|
11 |
+
encoding = tiktoken.get_encoding('cl100k_base')
|
12 |
+
num_tokens = len(encoding.encode(string))
|
13 |
+
return num_tokens
|
14 |
+
|
15 |
+
|
16 |
+
def retrive_doc_on_token(rephrased_query, db):
|
17 |
+
top_docs = db.similarity_search(rephrased_query, k=10)
|
18 |
+
num = [num_token(doc.page_content) for doc in top_docs]
|
19 |
+
cum_sum = [abs(sum(num[:i+1]) - 2700) for i in range(len(num))]
|
20 |
+
idx = cum_sum.index(min(cum_sum))
|
21 |
+
|
22 |
+
return top_docs[:idx+1]
|
23 |
+
|
24 |
+
|
25 |
+
|
26 |
+
|
27 |
+
def multi_docs_qa(query, rephrased_query, db, company, language, temperature):
|
28 |
+
"""
|
29 |
+
Return an answer to the query based on multiple documents limited
|
30 |
+
by total token of 3000.
|
31 |
+
"""
|
32 |
+
print('temperature: ', temperature)
|
33 |
+
template = """<|im_start|>
|
34 |
+
Manulife's assistant helps the user with their questions about products and services. Be brief in your answers with appropriate tone and emotion. Answer ONLY with the facts listed in the list of sources below. If there isn't enough information below, say you don't know. Do not generate answers that don't use the sources below. If asking a clarifying question to the user would help, ask the question. For tabular information return it as an html table. Do not return markdown format. Each source has a name followed by colon and the actual information, ALWAYS include the source name for each fact you use in the response. Use square brakets to reference the source, e.g. [info1.txt]. Don't combine sources, list each source separately, e.g. [info1.txt][info2.pdf].
|
35 |
+
|
36 |
+
Sources:
|
37 |
+
{sources}
|
38 |
+
<|im_end|>
|
39 |
+
|
40 |
+
{chat_history}
|
41 |
+
"""
|
42 |
+
|
43 |
+
docs = retrive_doc_on_token(query, db)
|
44 |
+
sources = []
|
45 |
+
for i in docs:
|
46 |
+
source_txt = i.metadata['source']
|
47 |
+
source_content = i.page_content
|
48 |
+
add = f"{source_txt}: {source_content}"
|
49 |
+
sources.append(add)
|
50 |
+
|
51 |
+
s_txt = '\n\n'.join(sources)
|
52 |
+
# print('this is docs: ', docs)
|
53 |
+
ch = rephrased_query + f'\nUser: {query}' + '\nAssistant: '
|
54 |
+
final_template = template.format(sources=s_txt, chat_history=ch)
|
55 |
+
print('\n\n', final_template, '\n\n')
|
56 |
+
# PROMPT = PromptTemplate(template=template, input_variables=["summaries", "question"])
|
57 |
+
# llm = OpenAI(temperature=temperature, model_name='gpt-3.5-turbo')
|
58 |
+
# chain = load_qa_with_sources_chain(llm, chain_type="stuff", prompt=PROMPT)
|
59 |
+
# response = chain({"input_documents": docs, "question": query})
|
60 |
+
|
61 |
+
llm = ChatOpenAI(model_name="gpt-3.5-turbo", temperature=temperature)
|
62 |
+
response = llm([HumanMessage(content=final_template)]).content
|
63 |
+
|
64 |
+
|
65 |
+
# return response['output_text'], docs
|
66 |
+
return response, ''
|
67 |
+
|
68 |
+
|
69 |
+
|
70 |
+
def multi_docs_qa_hkas(query, rephrased_query, db, language, temperature):
|
71 |
+
"""
|
72 |
+
Return an answer to the query based on multiple documents limited
|
73 |
+
by total token of 3000.
|
74 |
+
"""
|
75 |
+
|
76 |
+
template = (
|
77 |
+
f"Create a comprehensive and truthful final response in {language}. "
|
78 |
+
"Ask for clearification before answering if the QUESTION is not clear.\n\n"
|
79 |
+
|
80 |
+
"Context (may or may not be useful)"
|
81 |
+
"===\n"
|
82 |
+
"{summaries}\n"
|
83 |
+
"===\n\n"
|
84 |
+
|
85 |
+
"Query: "
|
86 |
+
"===\n"
|
87 |
+
"{question} "
|
88 |
+
f"({rephrased_query})\n"
|
89 |
+
"===\n\n"
|
90 |
+
|
91 |
+
f"FINAL RESPONSE (in complete sentence):"
|
92 |
+
)
|
93 |
+
docs = retrive_doc_on_token(query+ f" ({rephrased_query})", db)
|
94 |
+
|
95 |
+
|
96 |
+
PROMPT = PromptTemplate(template=template, input_variables=["summaries", "question"])
|
97 |
+
llm = OpenAI(temperature=temperature, model_name='gpt-3.5-turbo')
|
98 |
+
chain = load_qa_with_sources_chain(llm, chain_type="stuff", prompt=PROMPT)
|
99 |
+
response = chain({"input_documents": docs, "question": query})
|
100 |
+
|
101 |
+
return response['output_text'], docs
|
102 |
+
# return response['output_text'].replace('Manulife', 'Company A').replace('manulife', 'Company A'), docs
|
103 |
+
|
104 |
+
|
105 |
+
|
106 |
+
def summary(query, context):
|
107 |
+
template = (
|
108 |
+
"Use the following portion of a long document to see if any of the text is relevant to answer the question. "
|
109 |
+
"Return any relevant text verbatim and 'SOURCE'.\n\n"
|
110 |
+
|
111 |
+
"===\n"
|
112 |
+
"{context}"
|
113 |
+
"===\n\n"
|
114 |
+
|
115 |
+
"Question: {question}\n"
|
116 |
+
"Relevant text, if any:"
|
117 |
+
)
|
118 |
+
return template.format(context=context,question=query)
|
119 |
+
|
120 |
+
import asyncio
|
121 |
+
|
122 |
+
async def multi_reponse(temperature, messages, docs):
|
123 |
+
chat = ChatOpenAI(temperature=temperature, model_name="gpt-3.5-turbo")
|
124 |
+
responses = await chat.agenerate(messages=messages)
|
125 |
+
|
126 |
+
text = ''
|
127 |
+
|
128 |
+
for i, r in enumerate(responses.generations):
|
129 |
+
if 'N/A' not in r[0].text:
|
130 |
+
text += f"{r[0].text}\nSOURCE: {docs[i].metadata['source']}\n\n"
|
131 |
+
|
132 |
+
print(text)
|
133 |
+
|
134 |
+
return text
|
135 |
+
|
136 |
+
|
137 |
+
from typing import Any, Dict, List, Optional, Union
|
138 |
+
from types import GeneratorType
|
139 |
+
from langchain.callbacks.base import AsyncCallbackHandler, BaseCallbackHandler
|
140 |
+
from langchain.schema import AgentAction, AgentFinish, LLMResult
|
141 |
+
|
142 |
+
class SyncStreamingLLMCallbackHandler(BaseCallbackHandler):
|
143 |
+
"""Callback handler for streaming LLM responses to a queue."""
|
144 |
+
|
145 |
+
def __init__(self, q):
|
146 |
+
self.q = q
|
147 |
+
|
148 |
+
def on_llm_start(
|
149 |
+
self, serialized: Dict[str, Any], prompts: List[str], **kwargs: Any
|
150 |
+
) -> None:
|
151 |
+
"""Do nothing."""
|
152 |
+
pass
|
153 |
+
|
154 |
+
def on_llm_new_token(self, token: str, **kwargs: Any) -> None:
|
155 |
+
self.q.put(token)
|
156 |
+
|
157 |
+
def on_llm_end(self, response: LLMResult, **kwargs: Any) -> None:
|
158 |
+
"""Do nothing."""
|
159 |
+
pass
|
160 |
+
|
161 |
+
def on_llm_error(
|
162 |
+
self, error: Union[Exception, KeyboardInterrupt], **kwargs: Any
|
163 |
+
) -> None:
|
164 |
+
"""Do nothing."""
|
165 |
+
pass
|
166 |
+
|
167 |
+
def on_chain_start(
|
168 |
+
self, serialized: Dict[str, Any], inputs: Dict[str, Any], **kwargs: Any
|
169 |
+
) -> None:
|
170 |
+
"""Do nothing."""
|
171 |
+
pass
|
172 |
+
|
173 |
+
def on_chain_end(self, outputs: Dict[str, Any], **kwargs: Any) -> None:
|
174 |
+
"""Do nothing."""
|
175 |
+
pass
|
176 |
+
|
177 |
+
def on_chain_error(
|
178 |
+
self, error: Union[Exception, KeyboardInterrupt], **kwargs: Any
|
179 |
+
) -> None:
|
180 |
+
"""Do nothing."""
|
181 |
+
pass
|
182 |
+
|
183 |
+
def on_tool_start(
|
184 |
+
self,
|
185 |
+
serialized: Dict[str, Any],
|
186 |
+
input_str: str,
|
187 |
+
**kwargs: Any,
|
188 |
+
) -> None:
|
189 |
+
"""Do nothing."""
|
190 |
+
pass
|
191 |
+
|
192 |
+
def on_tool_end(
|
193 |
+
self,
|
194 |
+
output: str,
|
195 |
+
color: Optional[str] = None,
|
196 |
+
observation_prefix: Optional[str] = None,
|
197 |
+
llm_prefix: Optional[str] = None,
|
198 |
+
**kwargs: Any,
|
199 |
+
) -> None:
|
200 |
+
"""Do nothing."""
|
201 |
+
pass
|
202 |
+
|
203 |
+
def on_tool_error(
|
204 |
+
self, error: Union[Exception, KeyboardInterrupt], **kwargs: Any
|
205 |
+
) -> None:
|
206 |
+
"""Do nothing."""
|
207 |
+
pass
|
208 |
+
|
209 |
+
def on_agent_action(self, action: AgentAction, **kwargs: Any) -> Any:
|
210 |
+
"""Run on agent action."""
|
211 |
+
pass
|
212 |
+
|
213 |
+
def on_agent_finish(
|
214 |
+
self, finish: AgentFinish, color: Optional[str] = None, **kwargs: Any
|
215 |
+
) -> None:
|
216 |
+
"""Run on agent end."""
|
217 |
+
pass
|