Upload 26 files
Browse files- app.py +767 -445
- chatgpt.py +61 -0
- joeshi/.DS_Store +0 -0
- joeshi/database_name.csv +24 -6
- joeshi/default__vector_store.json +1 -0
- joeshi/docstore.json +1 -0
- joeshi/faiss_index/.DS_Store +0 -0
- joeshi/faiss_index/index.faiss +0 -0
- joeshi/faiss_index/index.pkl +3 -0
- joeshi/graph_store.json +1 -0
- joeshi/image__vector_store.json +1 -0
- joeshi/index_store.json +1 -0
- joeshi/joeshi_upload.csv +201 -0
- joeshi/message.json +1 -0
- joeshi/vector_store.json +0 -0
- langchain_KB.py +108 -0
- localKB_construct.py +5 -7
- message.json +1 -0
- multiQuery_prompt.py +95 -0
- rag_source.py +68 -0
- save_database_info.py +1 -2
- st_msautogen.py +171 -0
- st_openai_assistant.py +291 -0
- st_reset_conversation.py +15 -0
app.py
CHANGED
@@ -1,33 +1,28 @@
|
|
1 |
'''
|
2 |
-
|
3 |
-
|
4 |
-
1.
|
5 |
-
1.
|
6 |
-
1.
|
7 |
-
1.
|
8 |
-
|
9 |
-
|
10 |
-
|
11 |
-
1.
|
12 |
-
## 需要严格的按照yaml文件的格式来定义如下几个字段。
|
13 |
-
for un, name, pw in zip(users, names, hashed_passwords):
|
14 |
-
# user_dict = {"name":name,"password":pw}
|
15 |
-
user_dict = {"name": un, "password": pw}
|
16 |
-
# credentials["usernames"].update({un:user_dict})
|
17 |
-
credentials["usernames"].update({un: user_dict})
|
18 |
|
19 |
'''
|
20 |
-
# TODO:1.
|
21 |
|
22 |
-
|
23 |
-
|
|
|
24 |
import requests
|
25 |
from codeinterpreterapi import CodeInterpreterSession, File
|
26 |
import streamlit as st
|
27 |
-
# from codeinterpreterapi import CodeInterpreterSession
|
28 |
import openai
|
29 |
import os
|
30 |
import matplotlib.pyplot as plt
|
|
|
31 |
import pandas as pd
|
32 |
# import csv
|
33 |
import tempfile
|
@@ -38,43 +33,94 @@ from matplotlib.font_manager import FontProperties
|
|
38 |
import seaborn as sns
|
39 |
from time import sleep
|
40 |
import streamlit_authenticator as stauth
|
41 |
-
#
|
42 |
-
|
43 |
-
from langchain.chat_models import ChatOpenAI
|
44 |
-
from llama_index import StorageContext, load_index_from_storage, GPTVectorStoreIndex, LLMPredictor, PromptHelper
|
45 |
-
from llama_index import ServiceContext, QuestionAnswerPrompt
|
46 |
import sys
|
47 |
import time
|
48 |
import PyPDF2 ## read the local_KB PDF file.
|
49 |
# import localKB_construct
|
|
|
|
|
|
|
50 |
import save_database_info
|
51 |
-
from datetime import datetime
|
52 |
import pytz
|
53 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
54 |
os.environ["OPENAI_API_KEY"] = os.environ['user_token']
|
55 |
-
|
56 |
-
# os.environ["VERBOSE"] = "True" # 可以看到具体的错误?
|
57 |
|
58 |
-
#* 如果碰到接口问题,可以启用如下设置。
|
59 |
-
# openai.proxy = {
|
60 |
-
# "http": "http://127.0.0.1:7890",
|
61 |
-
# "https": "http://127.0.0.1:7890"
|
62 |
-
# }
|
63 |
|
|
|
|
|
|
|
|
|
|
|
64 |
|
65 |
-
|
66 |
-
st.title("
|
67 |
st.subheader("Artificial Intelligence Backend Center for Professionals")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
68 |
|
69 |
-
# clear conversion.
|
70 |
reset_button_key = "reset_button"
|
71 |
-
reset_button = st.button(label=("
|
72 |
-
|
73 |
-
if reset_button:
|
74 |
-
st.session_state.conversation = None
|
75 |
-
st.session_state.chat_history = None
|
76 |
-
st.session_state.messages = []
|
77 |
-
message_placeholder = st.empty()
|
78 |
|
79 |
def clear_all():
|
80 |
st.session_state.conversation = None
|
@@ -83,73 +129,104 @@ def clear_all():
|
|
83 |
message_placeholder = st.empty()
|
84 |
return None
|
85 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
86 |
|
87 |
-
#
|
88 |
-
|
89 |
-
# if uploaded_file is not None:
|
90 |
-
# filename = uploaded_file.name
|
91 |
-
# # st.write(filename) # print out the whole file name to validate. not to show in the final version.
|
92 |
-
# try:
|
93 |
-
# if '.pdf' in filename:
|
94 |
-
# # pdf_file = PyPDF2.PdfReader(uploaded_file)
|
95 |
-
# PyPDF2.PdfReader(uploaded_file)
|
96 |
-
# # st.write(pdf_file.pages[0].extract_text())
|
97 |
-
# # with st.status('正在为您解析新知识库...', expanded=False, state='running') as status:
|
98 |
-
# spinner = st.spinner('正在为您解析新知识库...请耐心等待')
|
99 |
-
# # with st.spinner('正在为您解析新知识库...请耐心等待'):
|
100 |
-
# with spinner:
|
101 |
-
# import localKB_construct
|
102 |
-
# # sleep(3)
|
103 |
-
# # st.write(upload_file)
|
104 |
-
# localKB_construct.process_file(uploaded_file)
|
105 |
-
# st.markdown('新知识库解析成功,可以开始对话!')
|
106 |
-
# spinner = st.empty()
|
107 |
-
# # sleep(3)
|
108 |
-
# # display = []
|
109 |
-
|
110 |
-
# else:
|
111 |
-
# if '.csv' in filename:
|
112 |
-
# csv_file = pd.read_csv(uploaded_file)
|
113 |
-
# csv_file.to_csv('./upload.csv', encoding='utf-8', index=False)
|
114 |
-
# st.write(csv_file[:3]) # 这里只是显示文件,后面需要定位文件所在的绝对路径。
|
115 |
-
# else:
|
116 |
-
# xls_file = pd.read_excel(uploaded_file)
|
117 |
-
# xls_file.to_csv('./upload.csv', index=False)
|
118 |
-
# st.write(xls_file[:3])
|
119 |
-
|
120 |
-
# uploaded_file_name = "File_provided"
|
121 |
-
# temp_dir = tempfile.TemporaryDirectory()
|
122 |
-
# # ! working.
|
123 |
-
# uploaded_file_path = pathlib.Path(temp_dir.name) / uploaded_file_name
|
124 |
-
# # with open('./upload.csv', 'wb') as output_temporary_file:
|
125 |
-
# with open(f'./{name}_upload.csv', 'wb') as output_temporary_file:
|
126 |
-
# # print(f'./{name}_upload.csv')
|
127 |
-
# # ! 必须用这种格式读入内容,然后才可以写入temporary文件夹中。
|
128 |
-
# # output_temporary_file.write(uploaded_file.getvalue())
|
129 |
-
# output_temporary_file.write(uploaded_file.getvalue())
|
130 |
-
# # st.write(uploaded_file_path) #* 可以查看文件是否真实存在,然后是否可以
|
131 |
-
# # st.write('Now file saved successfully.')
|
132 |
-
# except Exception as e:
|
133 |
-
# st.write(e)
|
134 |
-
|
135 |
-
# # uploaded_file_name = "File_provided"
|
136 |
-
# # temp_dir = tempfile.TemporaryDirectory()
|
137 |
-
# # # ! working.
|
138 |
-
# # uploaded_file_path = pathlib.Path(temp_dir.name) / uploaded_file_name
|
139 |
-
# # # with open('./upload.csv', 'wb') as output_temporary_file:
|
140 |
-
# # with open(f'./{name}_upload.csv', 'wb') as output_temporary_file:
|
141 |
-
# # # print(f'./{name}_upload.csv')
|
142 |
-
# # # ! 必须用这种格式读入内容,然后才可以写入temporary文件夹中。
|
143 |
-
# # # output_temporary_file.write(uploaded_file.getvalue())
|
144 |
-
# # output_temporary_file.write(uploaded_file.getvalue())
|
145 |
-
# # # st.write(uploaded_file_path) # * 可以查看文件是否真实存在,然后是否可以
|
146 |
-
# # # st.write('Now file saved successfully.')
|
147 |
-
|
148 |
-
# return None
|
149 |
|
|
|
150 |
bing_search_api_key = os.environ['bing_api_key']
|
151 |
bing_search_endpoint = 'https://api.bing.microsoft.com/v7.0/search'
|
152 |
-
|
153 |
def search(query):
|
154 |
# Construct a request
|
155 |
# mkt = 'en-EN'
|
@@ -169,19 +246,26 @@ def search(query):
|
|
169 |
except Exception as e:
|
170 |
raise e
|
171 |
|
172 |
-
#
|
173 |
-
|
174 |
-
|
|
|
|
|
175 |
# Set a default model
|
176 |
if "openai_model" not in st.session_state:
|
177 |
st.session_state["openai_model"] = "gpt-3.5-turbo-16k"
|
178 |
-
|
|
|
179 |
# print('----------'*5)
|
180 |
print('radio_1: GPT-3.5 starts!')
|
181 |
st.session_state["openai_model"] = "gpt-3.5-turbo-16k"
|
182 |
-
|
183 |
print('radio_1: GPT-4.0 starts!')
|
184 |
-
st.session_state["openai_model"] = "gpt-4"
|
|
|
|
|
|
|
|
|
185 |
|
186 |
# Initialize chat history
|
187 |
if "messages" not in st.session_state:
|
@@ -193,11 +277,10 @@ async def text_mode():
|
|
193 |
st.markdown(message["content"])
|
194 |
|
195 |
# Display assistant response in chat message container
|
196 |
-
# if prompt := st.chat_input("
|
197 |
-
prompt = st.chat_input("
|
198 |
print('prompt now:', prompt)
|
199 |
print('----------'*5)
|
200 |
-
# if prompt:
|
201 |
if prompt:
|
202 |
st.session_state.messages.append({"role": "user", "content": prompt})
|
203 |
with st.chat_message("user"):
|
@@ -207,8 +290,9 @@ async def text_mode():
|
|
207 |
message_placeholder = st.empty()
|
208 |
full_response = ""
|
209 |
|
210 |
-
if
|
211 |
-
|
|
|
212 |
input_message = prompt
|
213 |
internet_search_result = search(input_message)
|
214 |
search_prompt = [
|
@@ -221,7 +305,26 @@ async def text_mode():
|
|
221 |
st.session_state.messages.append(
|
222 |
{"role": "user", "content": prompt})
|
223 |
|
224 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
225 |
model=st.session_state["openai_model"],
|
226 |
messages=[
|
227 |
{"role": m["role"], "content": m["content"]}
|
@@ -229,85 +332,134 @@ async def text_mode():
|
|
229 |
],
|
230 |
stream=True,
|
231 |
):
|
232 |
-
|
233 |
-
|
234 |
-
|
235 |
-
|
236 |
-
|
237 |
-
|
238 |
-
|
239 |
-
|
240 |
-
|
241 |
-
|
242 |
-
print('
|
243 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
244 |
model=st.session_state["openai_model"],
|
|
|
|
|
|
|
|
|
|
|
|
|
245 |
messages=[
|
246 |
{"role": m["role"], "content": m["content"]}
|
247 |
for m in st.session_state.messages
|
248 |
],
|
249 |
-
stream=True,
|
250 |
# messages=[{'role': 'system', 'content': 'you are ChatGPT'}, {
|
251 |
-
# 'role': 'user', 'content': prompt}],
|
252 |
-
|
253 |
):
|
254 |
-
|
255 |
-
|
256 |
-
|
257 |
-
|
258 |
-
|
259 |
-
|
260 |
-
|
261 |
-
|
262 |
-
|
263 |
-
|
264 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
265 |
def localKB_mode(username):
|
266 |
-
|
267 |
-
|
268 |
-
# st.session_state.chat_history = None
|
269 |
-
# st.session_state.messages = []
|
270 |
-
# message_placeholder = st.empty()
|
271 |
-
clear_all() ## reset the conversation.
|
272 |
-
|
273 |
print('now starts the local KB version of ChatGPT')
|
274 |
-
# Initialize chat history
|
275 |
-
if "messages" not in st.session_state:
|
276 |
-
st.session_state.messages = []
|
277 |
|
278 |
-
|
279 |
-
|
280 |
-
|
|
|
|
|
|
|
|
|
281 |
|
282 |
# Display assistant response in chat message container
|
283 |
-
# if prompt := st.chat_input("Say something"):
|
284 |
-
# prompt = st.chat_input("Say something")
|
285 |
-
# print('prompt now:', prompt)
|
286 |
-
# print('----------'*5)
|
287 |
# if prompt:
|
288 |
-
if prompt := st.chat_input("
|
289 |
st.session_state.messages.append({"role": "user", "content": prompt})
|
290 |
with st.chat_message("user"):
|
291 |
st.markdown(prompt)
|
292 |
|
293 |
with st.status('检索中...', expanded=True, state='running') as status:
|
|
|
294 |
with st.chat_message("assistant"):
|
295 |
message_placeholder = st.empty()
|
296 |
full_response = ""
|
297 |
|
298 |
-
|
299 |
-
|
300 |
-
QA_PROMPT_TMPL = (
|
301 |
-
"We have provided context information below. \n"
|
302 |
-
"---------------------\n"
|
303 |
-
"{context_str}"
|
304 |
-
"\n---------------------\n"
|
305 |
-
"Given all this information, please answer the following questions,"
|
306 |
-
"You MUST use the SAME language as the question:\n"
|
307 |
-
"{query_str}\n")
|
308 |
-
QA_PROMPT = QuestionAnswerPrompt(QA_PROMPT_TMPL)
|
309 |
-
# print('QA_PROMPT:', QA_PROMPT)
|
310 |
-
|
311 |
# llm_predictor = LLMPredictor(llm=ChatOpenAI(temperature=0.8, model_name="gpt-3.5-turbo", max_tokens=4024,streaming=True))
|
312 |
# # print('llm_predictor:', llm_predictor)
|
313 |
# prompt_helper = PromptHelper(max_input_size, num_outputs, max_chunk_overlap, chunk_size_limit=chunk_size_limit)
|
@@ -317,47 +469,129 @@ def localKB_mode(username):
|
|
317 |
# # # index = load_index_from_storage(storage_context)
|
318 |
# print("storage_context:", storage_context)
|
319 |
# index = load_index_from_storage(storage_context,service_context=service_context)
|
320 |
-
|
321 |
-
|
322 |
-
|
323 |
-
|
324 |
-
#
|
325 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
326 |
# query_engine = index.as_query_engine(streaming=True, text_qa_template=QA_PROMPT)
|
327 |
-
# query_engine = index.as_query_engine(streaming=False, text_qa_template=QA_PROMPT)
|
328 |
# query_engine = index.as_query_engine()
|
329 |
# reply = query_engine.query(prompt)
|
330 |
|
331 |
-
llama_index_reply = query_engine.query(prompt)
|
332 |
-
# full_response += query_engine.query(prompt)
|
333 |
-
print('local KB reply:', llama_index_reply)
|
334 |
-
# query_engine.query(prompt).print_response_stream() #* 能在terminal中流式输出。
|
335 |
-
# for resp in llama_index_reply.response_gen:
|
336 |
-
# print(resp)
|
337 |
-
# full_response += resp
|
338 |
-
# message_placeholder.markdown(full_response + "▌")
|
339 |
-
|
340 |
-
#
|
341 |
-
#
|
342 |
-
# st.session_state.messages
|
343 |
-
#
|
344 |
-
#
|
345 |
-
#
|
346 |
-
#
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
347 |
|
348 |
-
async def data_mode():
|
349 |
-
|
350 |
clear_all() ## reset the conversation.
|
|
|
351 |
# uploaded_file_path = './upload.csv'
|
352 |
-
uploaded_file_path = f'./{username}_upload.csv'
|
353 |
-
# uploaded_file_path =
|
354 |
print('file path:', uploaded_file_path)
|
355 |
|
356 |
-
# st.write(f"passed file path in data_mode: {uploaded_file_path}")
|
357 |
-
# tmp1 = pd.read_csv(uploaded_file_path)
|
358 |
-
# st.markdown('成功启动数据模式,以下是加载的文件内容')
|
359 |
-
# st.write(tmp1[:5])
|
360 |
-
|
361 |
# Initialize chat history
|
362 |
if "messages" not in st.session_state:
|
363 |
st.session_state.messages = []
|
@@ -372,76 +606,130 @@ async def data_mode():
|
|
372 |
prompt = st.chat_input("Say something")
|
373 |
print('prompt now:', prompt)
|
374 |
print('----------'*5)
|
375 |
-
# if prompt:
|
376 |
if prompt:
|
377 |
-
|
378 |
-
|
379 |
-
st.
|
380 |
-
|
381 |
-
|
382 |
-
|
383 |
-
|
384 |
-
|
385 |
-
|
386 |
-
|
387 |
-
|
388 |
-
|
389 |
-
|
390 |
-
|
391 |
-
|
392 |
-
|
393 |
-
|
394 |
-
|
395 |
-
|
396 |
-
|
397 |
-
|
398 |
-
|
399 |
-
|
400 |
-
|
401 |
-
|
402 |
-
|
403 |
-
|
404 |
-
|
405 |
-
|
406 |
-
|
407 |
-
|
408 |
-
|
409 |
-
|
410 |
-
|
411 |
-
|
412 |
-
|
413 |
-
|
414 |
-
|
415 |
-
|
416 |
-
#
|
417 |
-
#
|
418 |
-
|
419 |
-
|
420 |
-
#
|
421 |
-
|
422 |
-
|
423 |
-
|
424 |
-
|
425 |
-
|
426 |
-
|
427 |
-
|
428 |
-
|
429 |
-
|
430 |
-
|
431 |
-
|
432 |
-
|
433 |
-
#
|
434 |
-
|
435 |
-
|
436 |
-
|
437 |
-
|
438 |
-
|
439 |
-
|
440 |
-
|
441 |
-
|
442 |
-
|
443 |
-
|
444 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
445 |
|
446 |
### authentication with a local yaml file.
|
447 |
import yaml
|
@@ -456,8 +744,6 @@ authenticator = stauth.Authenticate(
|
|
456 |
config['preauthorized']
|
457 |
)
|
458 |
|
459 |
-
|
460 |
-
###'''authentication with a remove cloud-based database.'''
|
461 |
# authentication with a remove cloud-based database.
|
462 |
# 导入云端用户数据库。
|
463 |
|
@@ -468,8 +754,10 @@ authenticator = stauth.Authenticate(
|
|
468 |
|
469 |
# deta = Deta(DETA_KEY)
|
470 |
|
471 |
-
#
|
472 |
# credentials = {"usernames":{}}
|
|
|
|
|
473 |
# users = []
|
474 |
# email = []
|
475 |
# passwords = []
|
@@ -501,9 +789,13 @@ authenticator = stauth.Authenticate(
|
|
501 |
# database_table.append([i,credentials['usernames'][i]['name'],credentials['usernames'][i]['password']])
|
502 |
# print("database_table:",database_table)
|
503 |
|
|
|
504 |
# authenticator = stauth.Authenticate(
|
505 |
# credentials=credentials, cookie_name="joeshi_gpt", key='abcedefg', cookie_expiry_days=30)
|
506 |
|
|
|
|
|
|
|
507 |
# ## sign-up widget,未完成。
|
508 |
# try:
|
509 |
# if authenticator.register_user('新用户注册', preauthorization=False):
|
@@ -514,13 +806,6 @@ authenticator = stauth.Authenticate(
|
|
514 |
# st.success('注册成功!')
|
515 |
# except Exception as e:
|
516 |
# st.error(e)
|
517 |
-
''''''
|
518 |
-
|
519 |
-
# user, authentication_status, username = authenticator.login('用户登录', 'main')
|
520 |
-
# user, authentication_status, username = authenticator.login('用户登录', 'sidebar')
|
521 |
-
user, authentication_status, username = authenticator.login('sidebar')
|
522 |
-
|
523 |
-
# print("name", name, "username", username)
|
524 |
|
525 |
if authentication_status:
|
526 |
with st.sidebar:
|
@@ -528,187 +813,224 @@ if authentication_status:
|
|
528 |
"""
|
529 |
<style>
|
530 |
[data-testid="stSidebar"][aria-expanded="true"]{
|
531 |
-
min-width:
|
532 |
-
max-width:
|
533 |
}
|
534 |
""",
|
535 |
unsafe_allow_html=True,
|
536 |
)
|
|
|
537 |
st.header(f'**欢迎 **{username}** 来到人工智能的世界** ♠')
|
538 |
-
st.write(f'_Welcome and Hope U Enjoy Staying
|
539 |
authenticator.logout('登出', 'sidebar')
|
540 |
|
541 |
-
|
542 |
-
#
|
543 |
-
#
|
544 |
-
#
|
545 |
-
#
|
546 |
-
#
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
547 |
|
548 |
-
|
549 |
-
|
550 |
-
|
551 |
-
with st.
|
552 |
-
|
|
|
553 |
# with st.text(body="说明"):
|
554 |
# st.markdown("* 重启一轮新对话时,只需要刷新页面(按Ctrl/Command + R)即可。")
|
555 |
with st.text(body="说明"):
|
556 |
st.markdown("* 为了保护数据与隐私,所有对话均不会被保存,刷新页面立即删除。敬请放心。")
|
557 |
with st.text(body="说明"):
|
558 |
-
st.markdown("* “GPT-4
|
559 |
with st.text(body="说明"):
|
560 |
-
st.markdown("*
|
561 |
with st.text(body="说明"):
|
562 |
st.markdown(
|
563 |
-
"*
|
564 |
with st.text(body="说明"):
|
565 |
-
st.markdown("*
|
566 |
-
|
567 |
-
st.markdown("#### 参考资料")
|
568 |
-
with st.expander(label="**核心模式的专用提示词Prompt示例**", expanded=False):
|
569 |
-
# with st.subheader(body="提示词Prompt"):
|
570 |
-
st.code(
|
571 |
-
body="继续用中文写一篇关于 [文章主题] 的文章,以下列句子开头:[文章开头]。", language='plaintext')
|
572 |
-
st.code(body="将以下文字概括为 100 个字,使其易于阅读和理解。避免使用复杂的句子结构或技术术语。",
|
573 |
-
language='plaintext')
|
574 |
-
st.code(body="给我出一个迪奥2023春季发布会活动策划。", language='plaintext')
|
575 |
-
st.code(body="帮我按照正式会议结构写一个会邀:主题是xx手机游戏立项会议。", language='plaintext')
|
576 |
-
st.code(body="帮我写一个车内健康监测全场景落地的项目计划,用表格。", language='plaintext')
|
577 |
-
st.code(
|
578 |
-
body="同时掷两枚质地均匀的骰子,则两枚骰子向上的点数之和为 7 的概率是多少。", language='plaintext')
|
579 |
-
st.code(body="写一篇产品经理的演讲稿,注意使用以下词汇: 赋能,抓手,中台,闭环,落地,漏斗,沉淀,给到,同步,对齐,对标,迭代,拉通,打通,升级,交付,聚焦,倒逼,复盘,梳理,方案,联动,透传,咬合,洞察,渗透,兜底,解耦,耦合,复用,拆解。", language='plaintext')
|
580 |
-
|
581 |
-
with st.expander(label="**数据模式的专用提示词Prompt示例**", expanded=False):
|
582 |
-
# with st.subheader(body="提示词Prompt"):
|
583 |
-
st.code(body="分析此数据集并绘制一些'有趣的图表'。", language='python')
|
584 |
-
st.code(
|
585 |
-
body="对于这个文件中的数据,你需要要找出[X,Y]数据之间的寻找'相关性'。", language='python')
|
586 |
-
st.code(body="对于这个文件中的[xxx]数据给我一个'整体的分析'。", language='python')
|
587 |
-
st.code(body="对于[xxx]数据给我一个'直方图',提供图表,并给出分析结果。", language='python')
|
588 |
-
st.code(body="对于[xxx]数据给我一个'小提琴图',并给出分析结果。", language='python')
|
589 |
-
st.code(
|
590 |
-
body="对于[X,Y,Z]数据在一个'分布散点图 (stripplot)',所有的数据在一张图上展现, 并给出分析结果。", language='python')
|
591 |
-
st.code(body="对于[X,Y]数据,进行'T检验',你需要展示图表,并给出分析结果。",
|
592 |
-
language='python')
|
593 |
-
st.code(body="对于[X,Y]数据给我一个3个类别的'聚类分析',并给出分析结果。",
|
594 |
-
language='python')
|
595 |
-
|
596 |
-
col1, col2 = st.columns(spec=[1, 2])
|
597 |
-
radio_2 = col2.radio(label='模式选择', options=[
|
598 |
-
'核心模式', '联网模式', '知识库模式', '数据模式'], horizontal=True, label_visibility='visible')
|
599 |
-
radio_1 = col1.radio(label='ChatGPT版本', options=[
|
600 |
-
'GPT-3.5', 'GPT-4.0'], horizontal=True, label_visibility='visible')
|
601 |
-
|
602 |
-
elif authentication_status == False:
|
603 |
-
st.error('⛔ 用户名或密码错误!')
|
604 |
-
elif authentication_status == None:
|
605 |
-
st.warning('⬅ 请先登录!')
|
606 |
-
|
607 |
-
### 上传文件的模块
|
608 |
-
def upload_file(uploaded_file):
|
609 |
-
if uploaded_file is not None:
|
610 |
-
filename = uploaded_file.name
|
611 |
-
# st.write(filename) # print out the whole file name to validate. not to show in the final version.
|
612 |
-
try:
|
613 |
-
if '.pdf' in filename:
|
614 |
-
# pdf_file = PyPDF2.PdfReader(uploaded_file)
|
615 |
-
PyPDF2.PdfReader(uploaded_file)
|
616 |
-
# st.write(pdf_file.pages[0].extract_text())
|
617 |
-
# with st.status('正在为您解析新知识库...', expanded=False, state='running') as status:
|
618 |
-
spinner = st.spinner('正在为您解析新知识库...请耐心等待')
|
619 |
-
# with st.spinner('正在为您解析新知识库...请耐心等待'):
|
620 |
-
with spinner:
|
621 |
-
import localKB_construct
|
622 |
-
# st.write(upload_file)
|
623 |
-
localKB_construct.process_file(uploaded_file, username)
|
624 |
-
## 在屏幕上展示当前知识库的信息,包括名字和加载日期。
|
625 |
-
save_database_info.save_database_info(f'./{username}/database_name.csv', filename, str(datetime.now(pytz.timezone('Asia/Shanghai')).strftime("%Y-%m-%d %H:%M")))
|
626 |
-
st.markdown('新知识库解析成功,请务必刷新页面,然后开启对话 🔃')
|
627 |
-
# spinner = st.empty()
|
628 |
|
629 |
-
|
630 |
-
|
631 |
-
|
632 |
-
|
633 |
-
|
634 |
-
|
635 |
-
|
636 |
-
|
637 |
-
st.write(xls_file[:3])
|
638 |
|
639 |
-
|
640 |
-
|
641 |
-
|
642 |
-
|
643 |
-
|
644 |
-
|
645 |
-
# print(f'./{name}_upload.csv')
|
646 |
-
# ! 必须用这种格式读入内容,然后才可以写入temporary文件夹中。
|
647 |
-
# output_temporary_file.write(uploaded_file.getvalue())
|
648 |
-
output_temporary_file.write(uploaded_file.getvalue())
|
649 |
-
# st.write(uploaded_file_path) #* 可以查看文件是否真实存在,然后是否可以
|
650 |
-
# st.write('Now file saved successfully.')
|
651 |
-
except Exception as e:
|
652 |
-
st.write(e)
|
653 |
|
654 |
-
|
655 |
-
#
|
656 |
-
#
|
657 |
-
#
|
658 |
-
|
659 |
-
|
660 |
-
|
661 |
-
|
662 |
-
|
663 |
-
|
664 |
-
|
665 |
-
|
666 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
667 |
|
668 |
-
return None
|
669 |
|
|
|
|
|
|
|
|
|
670 |
|
671 |
if __name__ == "__main__":
|
672 |
import asyncio
|
673 |
try:
|
674 |
-
|
675 |
-
|
676 |
-
#
|
677 |
-
|
678 |
-
|
679 |
-
|
680 |
-
|
681 |
-
|
682 |
-
|
683 |
-
|
684 |
-
|
685 |
-
|
686 |
-
|
687 |
-
|
688 |
-
|
689 |
-
|
690 |
-
|
691 |
-
|
692 |
-
|
693 |
-
#
|
694 |
-
|
695 |
-
|
696 |
-
|
697 |
-
|
698 |
-
|
699 |
-
|
700 |
-
|
701 |
-
|
702 |
-
|
703 |
-
|
704 |
-
|
705 |
-
|
706 |
-
|
707 |
-
|
708 |
-
|
709 |
-
|
710 |
-
|
711 |
-
|
712 |
-
|
713 |
-
|
714 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
'''
|
2 |
+
1. 基于ChatGPT的多场景应用:
|
3 |
+
1. 核心模式
|
4 |
+
1. 联网模式
|
5 |
+
1. 知识库模式
|
6 |
+
1. 数据分析模式
|
7 |
+
1. 智能体模式
|
8 |
+
1. RAG:
|
9 |
+
1. 核心文件包括:
|
10 |
+
1. langchain_KB.py包含了形成vector database的函数,和产生total prompt的函数。
|
11 |
+
1. rag_source.py包含了从vector database中提取信息来源的函数,包括文档名称和页码。
|
|
|
|
|
|
|
|
|
|
|
|
|
12 |
|
13 |
'''
|
14 |
+
# TODO:1. 更新huggingface上code01的版本,包括:知识库和数据分析模块。 2. 将知识库模块更新为:multi-query + source。 3. 将数据分析模块重写。
|
15 |
|
16 |
+
import numpy as np
|
17 |
+
import pandas as pd
|
18 |
+
from dotenv import load_dotenv # pip3 install python-dotenv
|
19 |
import requests
|
20 |
from codeinterpreterapi import CodeInterpreterSession, File
|
21 |
import streamlit as st
|
|
|
22 |
import openai
|
23 |
import os
|
24 |
import matplotlib.pyplot as plt
|
25 |
+
import xlrd
|
26 |
import pandas as pd
|
27 |
# import csv
|
28 |
import tempfile
|
|
|
33 |
import seaborn as sns
|
34 |
from time import sleep
|
35 |
import streamlit_authenticator as stauth
|
36 |
+
# from langchain.chat_models import ChatOpenAI
|
37 |
+
from langchain.llms import openai
|
|
|
|
|
|
|
38 |
import sys
|
39 |
import time
|
40 |
import PyPDF2 ## read the local_KB PDF file.
|
41 |
# import localKB_construct
|
42 |
+
# from streamlit_option_menu import option_menu
|
43 |
+
# import st_reset_conversation
|
44 |
+
from st_reset_conversation import reset_all, reset_message
|
45 |
import save_database_info
|
|
|
46 |
import pytz
|
47 |
+
from datetime import datetime
|
48 |
+
from dotenv import load_dotenv
|
49 |
+
from openai import OpenAI
|
50 |
+
import st_msautogen
|
51 |
+
import rag_source
|
52 |
+
import add_fonts
|
53 |
+
import asyncio
|
54 |
+
|
55 |
+
import warnings
|
56 |
+
warnings.filterwarnings("ignore")
|
57 |
+
|
58 |
+
#make it look nice from the start
|
59 |
+
# st.set_page_config(layout='wide',initial_sidebar_state='collapsed',)
|
60 |
+
|
61 |
+
### 设置openai的API key
|
62 |
+
load_dotenv()
|
63 |
+
openai.api_key = os.environ['user_token']
|
64 |
os.environ["OPENAI_API_KEY"] = os.environ['user_token']
|
65 |
+
bing_search_api_key = os.environ['bing_api_key']
|
|
|
66 |
|
|
|
|
|
|
|
|
|
|
|
67 |
|
68 |
+
# # #* 如果数据分析模块在本地调试时碰到接口问题,可以启用如下设置。还可能是一个bash命令的问题,见ChatGPT讲课要点.txt.
|
69 |
+
openai.proxy = {
|
70 |
+
"http": "http://127.0.0.1:7890",
|
71 |
+
"https": "http://127.0.0.1:7890"
|
72 |
+
}
|
73 |
|
74 |
+
## layout settings.
|
75 |
+
st.title("专业版大语言模型智能中心")
|
76 |
st.subheader("Artificial Intelligence Backend Center for Professionals")
|
77 |
+
st.caption("_声明:本网站仅提供技术测试与评估服务。内容由人工智能生成,仅供参考。如果您本人使用或对外传播本服务生成的输出,您应当主动核查输出内容的真实性、准确性,避免传播虚假信息。_")
|
78 |
+
# st.divider()
|
79 |
+
|
80 |
+
# ## clear conversion.
|
81 |
+
# def reset_all():
|
82 |
+
# # st.session_state.conversation = None
|
83 |
+
# st.session_state.chat_history = None
|
84 |
+
# st.session_state.messages = []
|
85 |
+
# message_placeholder = st.empty()
|
86 |
+
# return None
|
87 |
+
|
88 |
+
|
89 |
+
# navigation menu using Hydralit. 并没有解决menu跳转的问题。
|
90 |
+
# option_data = [
|
91 |
+
# {'icon': "house", 'label':"核心模式"},
|
92 |
+
# {'icon':"cloud-upload",'label':"信息检索模式"},
|
93 |
+
# {'icon': "gear", 'label':"数据分析模式"},
|
94 |
+
# {'icon': "list-task", 'label':"智���体模式"},
|
95 |
+
# ]
|
96 |
+
# navi_menu = op = hc.option_bar(option_definition=option_data,title=None,key='PrimaryOption', horizontal_orientation=True)
|
97 |
+
# navi_menu = hc.nav_bar(menu_definition=option_data, key='navi_menu', use_animation=True, option_menu=False, sticky_mode='pinned', sticky_nav=False, hide_streamlit_markers=False)
|
98 |
+
|
99 |
+
### 使用streamlit_option_menu格式的类似横幅选项。但是会出现第一次无法运行,需要手动清零或者做一个动作,才可以。
|
100 |
+
# navi_menu = option_menu(
|
101 |
+
# menu_title=None,
|
102 |
+
# options=['核心模式', '信息检索模式', '数据分析模式', '智能体模式'],
|
103 |
+
# # options=['GPT-3.5', 'GPT-4.0','清华GLM2-6B','百川Baichuan-13B', '阿里通义千问14B'],
|
104 |
+
# icons=['house', 'cloud-upload','gear','list-task'],
|
105 |
+
# menu_icon='cast',
|
106 |
+
# default_index=0,
|
107 |
+
# orientation='horizontal',
|
108 |
+
# # manual_select=0,
|
109 |
+
# # styles={
|
110 |
+
# # "container": {"padding": "0!important", "background-color": "#fafafa"},
|
111 |
+
# # "icon": {"color": "orange", "font-size": "25px"},
|
112 |
+
# # "nav-link": {"font-size": "25px", "text-align": "left", "margin":"0px", "--hover-color": "#eee"},
|
113 |
+
# # "nav-link-selected": {"background-color": "green"},
|
114 |
+
# # }
|
115 |
+
# )
|
116 |
+
|
117 |
+
### 常规streamlit选择
|
118 |
+
navi_menu = st.radio(label='选择一个大语言模型工作模式', options=['核心模式', '联网模式', '知识库模式','数据分析模式', '智能体模式'],index=0,horizontal=True)
|
119 |
+
# navi_menu = st.selectbox('选择一个大语言模型工作模式', ['核心模式', '信息检索模式', '数据分析模式', '智能体模式'],index=0) ### 原始agent001模式。
|
120 |
|
|
|
121 |
reset_button_key = "reset_button"
|
122 |
+
reset_button = st.button(label=("清除所有记录,并开启一轮新对话 ▶"),
|
123 |
+
key=reset_button_key, use_container_width=True, type="primary")
|
|
|
|
|
|
|
|
|
|
|
124 |
|
125 |
def clear_all():
|
126 |
st.session_state.conversation = None
|
|
|
129 |
message_placeholder = st.empty()
|
130 |
return None
|
131 |
|
132 |
+
## 清除所有对话记录, reset all conversation.
|
133 |
+
if reset_button:
|
134 |
+
reset_all()
|
135 |
+
|
136 |
+
### 上传文件的模块
|
137 |
+
def upload_file(uploaded_file):
|
138 |
+
if uploaded_file is not None:
|
139 |
+
# filename = uploaded_file.name
|
140 |
+
# st.write(filename) # print out the whole file name to validate. not to show in the final version.
|
141 |
+
try:
|
142 |
+
# if '.pdf' in filename: ### original code here.
|
143 |
+
if '.pdf' in uploaded_file.name:
|
144 |
+
pdf_filename = uploaded_file.name ### original code here.
|
145 |
+
filename = uploaded_file.name
|
146 |
+
# print('PDF file:', pdf_filename)
|
147 |
+
# with st.status('正在为您解析新知识库...', expanded=False, state='running') as status:
|
148 |
+
spinner = st.spinner('正在为您解析新知识库...请耐心等待')
|
149 |
+
with spinner:
|
150 |
+
### 一下是llama_index方法,但是升级后,可能会报错。
|
151 |
+
# import localKB_construct
|
152 |
+
# # st.write(upload_file)
|
153 |
+
# localKB_construct.process_file(uploaded_file, username)
|
154 |
+
# ## 在屏幕上展示当前知识库的信息,包括名字和加载日期。
|
155 |
+
# save_database_info.save_database_info(f'./{username}/database_name.csv', filename, str(datetime.now(pytz.timezone('Asia/Shanghai')).strftime("%Y-%m-%d %H:%M")))
|
156 |
+
# st.markdown('新知识库解析成功,请务必刷新页面,然后开启对话 🔃')
|
157 |
+
|
158 |
+
### 以下是langchain方案。
|
159 |
+
import langchain_KB
|
160 |
+
import save_database_info
|
161 |
+
|
162 |
+
uploaded_file_name = "File_provided"
|
163 |
+
temp_dir = tempfile.TemporaryDirectory()
|
164 |
+
# ! working.
|
165 |
+
uploaded_file_path = pathlib.Path(temp_dir.name) / uploaded_file_name
|
166 |
+
with open(pdf_filename, 'wb') as output_temporary_file:
|
167 |
+
# with open(f'./{username}_upload.pdf', 'wb') as output_temporary_file: ### original code here. 可能会造成在引用信息来源时文件名不对的问题。
|
168 |
+
# ! 必须用这种格式读入内容,然后才可以写入temporary文件夹中。
|
169 |
+
# output_temporary_file.write(uploaded_file.getvalue())
|
170 |
+
output_temporary_file.write(uploaded_file.getvalue())
|
171 |
+
|
172 |
+
langchain_KB.langchain_localKB_construct(output_temporary_file, username)
|
173 |
+
## 在屏幕上展示当前知识库的信息,包括名字和加载日期。
|
174 |
+
save_database_info.save_database_info(f'./{username}/database_name.csv', pdf_filename, str(datetime.now(pytz.timezone('Asia/Shanghai')).strftime("%Y-%m-%d %H:%M")))
|
175 |
+
st.markdown('新知识库解析成功,请务必刷新页面,然后开启对话 🔃')
|
176 |
+
|
177 |
+
return pdf_filename
|
178 |
+
|
179 |
+
else:
|
180 |
+
# if '.csv' in filename: ### original code here.
|
181 |
+
if '.csv' in uploaded_file.name:
|
182 |
+
print('start the csv file processing...')
|
183 |
+
csv_filename = uploaded_file.name
|
184 |
+
filename = uploaded_file.name
|
185 |
+
|
186 |
+
csv_file = pd.read_csv(uploaded_file)
|
187 |
+
csv_file.to_csv(f'./{username}/{username}_upload.csv', encoding='utf-8', index=False)
|
188 |
+
st.write(csv_file[:3]) # 这里只是显示文件,后面需要定位文件所在的绝对路径。
|
189 |
+
else:
|
190 |
+
xls_file = pd.read_excel(uploaded_file)
|
191 |
+
xls_file.to_csv(f'./{username}_upload.csv', index=False)
|
192 |
+
st.write(xls_file[:3])
|
193 |
+
|
194 |
+
print('end the csv file processing...')
|
195 |
+
|
196 |
+
# uploaded_file_name = "File_provided"
|
197 |
+
# temp_dir = tempfile.TemporaryDirectory()
|
198 |
+
# ! working.
|
199 |
+
# uploaded_file_path = pathlib.Path(temp_dir.name) / uploaded_file_name
|
200 |
+
# with open('./upload.csv', 'wb') as output_temporary_file:
|
201 |
+
# with open(f'./{username}_upload.csv', 'wb') as output_temporary_file:
|
202 |
+
# print(f'./{name}_upload.csv')
|
203 |
+
# ! 必须用这种格式读入内容,然后才可以写入temporary文件夹中。
|
204 |
+
# output_temporary_file.write(uploaded_file.getvalue())
|
205 |
+
# st.write(uploaded_file_path) #* 可以查看文件是否真实存在,然后是否可以
|
206 |
+
|
207 |
+
except Exception as e:
|
208 |
+
st.write(e)
|
209 |
+
|
210 |
+
## 以下代码是为了解决上传文件后,文件路径和文件名不对的问题。
|
211 |
+
# uploaded_file_name = "File_provided"
|
212 |
+
# temp_dir = tempfile.TemporaryDirectory()
|
213 |
+
# # ! working.
|
214 |
+
# uploaded_file_path = pathlib.Path(temp_dir.name) / uploaded_file_name
|
215 |
+
# # with open('./upload.csv', 'wb') as output_temporary_file:
|
216 |
+
# with open(f'./{name}_upload.csv', 'wb') as output_temporary_file:
|
217 |
+
# # print(f'./{name}_upload.csv')
|
218 |
+
# # ! 必须用这种格式读入内容,然后才可以写入temporary文件夹中。
|
219 |
+
# # output_temporary_file.write(uploaded_file.getvalue())
|
220 |
+
# output_temporary_file.write(uploaded_file.getvalue())
|
221 |
+
# # st.write(uploaded_file_path) # * 可以查看文件是否真实存在,然后是否可以
|
222 |
+
# # st.write('Now file saved successfully.')
|
223 |
|
224 |
+
# return pdf_filename, csv_filename
|
225 |
+
return filename
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
226 |
|
227 |
+
### 互联网搜索模块
|
228 |
bing_search_api_key = os.environ['bing_api_key']
|
229 |
bing_search_endpoint = 'https://api.bing.microsoft.com/v7.0/search'
|
|
|
230 |
def search(query):
|
231 |
# Construct a request
|
232 |
# mkt = 'en-EN'
|
|
|
246 |
except Exception as e:
|
247 |
raise e
|
248 |
|
249 |
+
# async def text_mode():
|
250 |
+
def text_mode():
|
251 |
+
reset_message() ## reset the message and placeholder.
|
252 |
+
print('text mode starts!')
|
253 |
+
|
254 |
# Set a default model
|
255 |
if "openai_model" not in st.session_state:
|
256 |
st.session_state["openai_model"] = "gpt-3.5-turbo-16k"
|
257 |
+
|
258 |
+
if radio_1 == 'ChatGPT-3.5':
|
259 |
# print('----------'*5)
|
260 |
print('radio_1: GPT-3.5 starts!')
|
261 |
st.session_state["openai_model"] = "gpt-3.5-turbo-16k"
|
262 |
+
elif radio_1 == 'ChatGPT-4':
|
263 |
print('radio_1: GPT-4.0 starts!')
|
264 |
+
st.session_state["openai_model"] = "gpt-4-1106-preview"
|
265 |
+
else:
|
266 |
+
st.markdown("**当前大模型无效,请在左侧工具栏中选择一个有效的模型。您现在访问的站点仅提供ChatGPT中的GPT-3.5/4。**")
|
267 |
+
|
268 |
+
print(st.session_state["openai_model"])
|
269 |
|
270 |
# Initialize chat history
|
271 |
if "messages" not in st.session_state:
|
|
|
277 |
st.markdown(message["content"])
|
278 |
|
279 |
# Display assistant response in chat message container
|
280 |
+
# if prompt := st.chat_input("说点什么吧"):
|
281 |
+
prompt = st.chat_input("说点什么吧...")
|
282 |
print('prompt now:', prompt)
|
283 |
print('----------'*5)
|
|
|
284 |
if prompt:
|
285 |
st.session_state.messages.append({"role": "user", "content": prompt})
|
286 |
with st.chat_message("user"):
|
|
|
290 |
message_placeholder = st.empty()
|
291 |
full_response = ""
|
292 |
|
293 |
+
if navi_menu == '联网模式':
|
294 |
+
# if (navi_menu=='信息检索模式') and (radio_2=='互联网'): ### original code here.
|
295 |
+
print('联网模式下的prompt:', prompt)
|
296 |
input_message = prompt
|
297 |
internet_search_result = search(input_message)
|
298 |
search_prompt = [
|
|
|
305 |
st.session_state.messages.append(
|
306 |
{"role": "user", "content": prompt})
|
307 |
|
308 |
+
## old version of openai API.
|
309 |
+
# for response in openai.ChatCompletion.create(
|
310 |
+
# model=st.session_state["openai_model"],
|
311 |
+
# messages=[
|
312 |
+
# {"role": m["role"], "content": m["content"]}
|
313 |
+
# for m in st.session_state.messages
|
314 |
+
# ],
|
315 |
+
# stream=True,
|
316 |
+
# ):
|
317 |
+
# full_response += response.choices[0].delta.get(
|
318 |
+
# "content", "")
|
319 |
+
# message_placeholder.markdown(full_response + "▌")
|
320 |
+
# message_placeholder.markdown(full_response)
|
321 |
+
# st.session_state.messages.append(
|
322 |
+
# {"role": "assistant", "content": full_response})
|
323 |
+
# st.session_state.messages = []
|
324 |
+
|
325 |
+
## new version of openai API.
|
326 |
+
openai_client = OpenAI()
|
327 |
+
for response in openai_client.chat.completions.create(
|
328 |
model=st.session_state["openai_model"],
|
329 |
messages=[
|
330 |
{"role": m["role"], "content": m["content"]}
|
|
|
332 |
],
|
333 |
stream=True,
|
334 |
):
|
335 |
+
if str(response.choices[0].delta.content) != 'None':
|
336 |
+
full_response += str(response.choices[0].delta.content)
|
337 |
+
message_placeholder.markdown(full_response + "▌")
|
338 |
+
message_placeholder.markdown(full_response)
|
339 |
+
st.session_state.messages.append(
|
340 |
+
{"role": "assistant", "content": full_response})
|
341 |
+
st.session_state.messages = []
|
342 |
+
|
343 |
+
# elif radio_2 != '互联网':
|
344 |
+
else:
|
345 |
+
print('ChatGPT only starts!!!')
|
346 |
+
## 这里需要确认是直接从import openai中获得的函数,而不是langchain中调用openai,否则随着langchain的更新,会出现问题。
|
347 |
+
# for response in openai.ChatCompletion.create(
|
348 |
+
# model=st.session_state["openai_model"],
|
349 |
+
# max_tokens=max_tokens,
|
350 |
+
# temperature=temperature,
|
351 |
+
# top_p=top_p,
|
352 |
+
# presence_penalty=presence_penalty,
|
353 |
+
# frequency_penalty=frequency_penalty,
|
354 |
+
# ## 多轮会话,需要记住历史记录。
|
355 |
+
# messages=[
|
356 |
+
# {"role": m["role"], "content": m["content"]}
|
357 |
+
# for m in st.session_state.messages
|
358 |
+
# ],
|
359 |
+
# # messages=[{'role': 'system', 'content': 'you are ChatGPT'}, {
|
360 |
+
# # 'role': 'user', 'content': prompt}], ## 这是单轮会话。
|
361 |
+
# stream=True,
|
362 |
+
# ):
|
363 |
+
openai_client = OpenAI()
|
364 |
+
for response in openai_client.chat.completions.create(
|
365 |
model=st.session_state["openai_model"],
|
366 |
+
max_tokens=max_tokens,
|
367 |
+
temperature=temperature,
|
368 |
+
top_p=top_p,
|
369 |
+
presence_penalty=presence_penalty,
|
370 |
+
frequency_penalty=frequency_penalty,
|
371 |
+
## 多轮会话,需要记住历史记录。
|
372 |
messages=[
|
373 |
{"role": m["role"], "content": m["content"]}
|
374 |
for m in st.session_state.messages
|
375 |
],
|
|
|
376 |
# messages=[{'role': 'system', 'content': 'you are ChatGPT'}, {
|
377 |
+
# 'role': 'user', 'content': prompt}], ## 这是单轮会话。
|
378 |
+
stream=True,
|
379 |
):
|
380 |
+
# print('full response now:', full_response)
|
381 |
+
# print('response now:', response)
|
382 |
+
|
383 |
+
## old version output format.
|
384 |
+
# full_response += response.choices[0].delta.get(
|
385 |
+
# "content", "")
|
386 |
+
|
387 |
+
## new version output format.
|
388 |
+
if str(response.choices[0].delta.content) != 'None': ## 注意这里是内容,而不是response,否则一个chunk的回复。
|
389 |
+
# print('response now:',response)
|
390 |
+
full_response += str(response.choices[0].delta.content)
|
391 |
+
message_placeholder.markdown(full_response + "▌")
|
392 |
+
message_placeholder.markdown(full_response)
|
393 |
+
st.session_state.messages.append(
|
394 |
+
{"role": "assistant", "content": full_response})
|
395 |
+
|
396 |
+
## load the local_KB PDF file.
|
397 |
+
# # def local_KB(uploaded_file):
|
398 |
+
# print('now starts the local KB version of ChatGPT')
|
399 |
+
# max_input_size = 4096
|
400 |
+
# # set number of output tokens
|
401 |
+
# # num_outputs = 3000 #* working
|
402 |
+
# num_outputs = 1000
|
403 |
+
# # set maximum chunk overlap
|
404 |
+
# max_chunk_overlap = -1000 #* working
|
405 |
+
# # set chunk size limit
|
406 |
+
# # chunk_size_limit = 600
|
407 |
+
# chunk_size_limit = 6000 #* working
|
408 |
+
|
409 |
+
# history = []
|
410 |
+
# if input:
|
411 |
+
# # ! 这里需要重新装载一下storage_context。
|
412 |
+
# QA_PROMPT_TMPL = (
|
413 |
+
# "We have provided context information below. \n"
|
414 |
+
# "---------------------\n"
|
415 |
+
# "{context_str}"
|
416 |
+
# "\n---------------------\n"
|
417 |
+
# "Given all this information, please answer the following questions,"
|
418 |
+
# "You MUST use the SAME language as the question:\n"
|
419 |
+
# "{query_str}\n")
|
420 |
+
# QA_PROMPT = QuestionAnswerPrompt(QA_PROMPT_TMPL)
|
421 |
+
|
422 |
+
# llm_predictor = LLMPredictor(llm=ChatOpenAI(temperature=0.8, model_name="gpt-3.5-turbo", max_tokens=4096,streaming=True))
|
423 |
+
# prompt_helper = PromptHelper(max_input_size, num_outputs, max_chunk_overlap, chunk_size_limit=chunk_size_limit)
|
424 |
+
# service_context = ServiceContext.from_defaults(llm_predictor=llm_predictor, prompt_helper=prompt_helper)
|
425 |
+
|
426 |
+
# # # index = load_index_from_storage(storage_context)
|
427 |
+
# storage_context = StorageContext.from_defaults(persist_dir="./")
|
428 |
+
# index = load_index_from_storage(storage_context,service_context=service_context)
|
429 |
+
# # query_engine = index.as_query_engine(streaming=True, similarity_top_k=3, text_qa_template=QA_PROMPT)
|
430 |
+
# # query_engine = index.as_query_engine(streaming=True)
|
431 |
+
# query_engine = index.as_query_engine(streaming=True, text_qa_template=QA_PROMPT)
|
432 |
+
# reply = query_engine.query(input)
|
433 |
+
|
434 |
+
# async def localKB_mode(username):
|
435 |
def localKB_mode(username):
|
436 |
+
# reset_all() ## reset the conversation.
|
437 |
+
reset_message() ## only reset the message and placeholder.
|
|
|
|
|
|
|
|
|
|
|
438 |
print('now starts the local KB version of ChatGPT')
|
|
|
|
|
|
|
439 |
|
440 |
+
# # Initialize chat history
|
441 |
+
# if "messages" not in st.session_state:
|
442 |
+
# st.session_state.messages = []
|
443 |
+
|
444 |
+
# for message in st.session_state.messages:
|
445 |
+
# with st.chat_message(message["role"]):
|
446 |
+
# st.markdown(message["content"])
|
447 |
|
448 |
# Display assistant response in chat message container
|
|
|
|
|
|
|
|
|
449 |
# if prompt:
|
450 |
+
if prompt := st.chat_input("说点什么吧"):
|
451 |
st.session_state.messages.append({"role": "user", "content": prompt})
|
452 |
with st.chat_message("user"):
|
453 |
st.markdown(prompt)
|
454 |
|
455 |
with st.status('检索中...', expanded=True, state='running') as status:
|
456 |
+
# try:
|
457 |
with st.chat_message("assistant"):
|
458 |
message_placeholder = st.empty()
|
459 |
full_response = ""
|
460 |
|
461 |
+
### llama_index框架的RAG代码,最近更新版本后不成功,会报错。
|
462 |
+
### outdated version.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
463 |
# llm_predictor = LLMPredictor(llm=ChatOpenAI(temperature=0.8, model_name="gpt-3.5-turbo", max_tokens=4024,streaming=True))
|
464 |
# # print('llm_predictor:', llm_predictor)
|
465 |
# prompt_helper = PromptHelper(max_input_size, num_outputs, max_chunk_overlap, chunk_size_limit=chunk_size_limit)
|
|
|
469 |
# # # index = load_index_from_storage(storage_context)
|
470 |
# print("storage_context:", storage_context)
|
471 |
# index = load_index_from_storage(storage_context,service_context=service_context)
|
472 |
+
|
473 |
+
## sample code for reference.
|
474 |
+
# docstore = 'storage/docstore.json'
|
475 |
+
# index_store = 'storage/index_store.json'
|
476 |
+
# vector_store = 'storage/vector_store.json'
|
477 |
+
# print('storage_context:', storage_context)
|
478 |
+
|
479 |
+
##NOTE: 这里需要重新装载一下storage_context。
|
480 |
+
# storage_context = StorageContext.from_defaults(persist_dir=f"./{username}/")
|
481 |
+
# print('--'*30)
|
482 |
+
# print('storage_context:',storage_context)
|
483 |
+
# print('type of storage_context.index_store:', type(storage_context.index_store))
|
484 |
+
# print('--'*30)
|
485 |
+
# # storage_context = {storage_context}
|
486 |
+
|
487 |
+
# index = load_index_from_storage(storage_context)
|
488 |
+
# print('--'*30)
|
489 |
+
# print('index now:', index)
|
490 |
+
|
491 |
+
# context_str = index
|
492 |
+
|
493 |
+
# ##TODO 重新构建Prompt,加入QA_Template.
|
494 |
+
# QA_PROMPT_TMPL = (
|
495 |
+
# "We have provided context information below. \n"
|
496 |
+
# "---------------------\n"
|
497 |
+
# "{context_str}"
|
498 |
+
# "\n---------------------\n"
|
499 |
+
# "Given all this information, please answer the following questions,"
|
500 |
+
# "You MUST use the SAME language as the question and the default language is Chinese:\n"
|
501 |
+
# "{query_str}\n")
|
502 |
+
# # QA_PROMPT = QuestionAnswerPrompt(QA_PROMPT_TMPL) ## outdated version.
|
503 |
+
|
504 |
+
# ##TODO: newer version but may run into llama_index import problem.
|
505 |
+
# # qa_template = PromptTemplate(QA_PROMPT_TMPL)
|
506 |
+
# # prompt = qa_template.format(context_str=context_str, query_str=prompt)
|
507 |
+
# # prompt = qa_template.format(context_str=context_str, query_str=QA_PROMPT)
|
508 |
+
# # query_engine = index.as_query_engine(streaming=True, similarity_top_k=3, text_qa_template=QA_PROMPT)
|
509 |
+
# query_engine = index.as_query_engine(streaming=False)
|
510 |
+
# print('111')
|
511 |
+
## older version.
|
512 |
# query_engine = index.as_query_engine(streaming=True, text_qa_template=QA_PROMPT)
|
|
|
513 |
# query_engine = index.as_query_engine()
|
514 |
# reply = query_engine.query(prompt)
|
515 |
|
516 |
+
# llama_index_reply = query_engine.query(prompt)
|
517 |
+
# # full_response += query_engine.query(prompt)
|
518 |
+
# print('local KB reply:', llama_index_reply)
|
519 |
+
# # query_engine.query(prompt).print_response_stream() #* 能在terminal中流式输出。
|
520 |
+
# # for resp in llama_index_reply.response_gen:
|
521 |
+
# # print(resp)
|
522 |
+
# # full_response += resp
|
523 |
+
# # message_placeholder.markdown(full_response + "▌")
|
524 |
+
|
525 |
+
# message_placeholder.markdown(str(llama_index_reply))
|
526 |
+
# print('333')
|
527 |
+
# # st.session_state.messages.append(
|
528 |
+
# # {"role": "assistant", "content": full_response})
|
529 |
+
# # st.session_state.messages = []
|
530 |
+
# # full_response += reply
|
531 |
+
# # full_response = reply
|
532 |
+
# # st.session_state.messages.append(
|
533 |
+
# # {"role": "assistant", "content": full_response})
|
534 |
+
|
535 |
+
### 用langchain的FAISS来做RAG
|
536 |
+
import langchain_KB
|
537 |
+
|
538 |
+
### 默认选择使用multi-query的方法进行查询。
|
539 |
+
##NOTE: 目前这个版本是把所有的multiquery当成一个问题提交给大模型。后续可以考��将每一个问题分别送入大模型,然后得到的多个答案,然后在一并汇总。
|
540 |
+
if True:
|
541 |
+
import multiQuery_prompt
|
542 |
+
prompt = multiQuery_prompt.generate_sim_query(orignal_question=prompt)
|
543 |
+
# print('multiQuery prompts now:', prompt)
|
544 |
+
|
545 |
+
prompt = str(prompt) ## 需要强制转换成str格式。
|
546 |
+
total_prompt, docs = langchain_KB.langchain_RAG(prompt, username)
|
547 |
+
print('total_prompt now:', total_prompt)
|
548 |
+
st.session_state.messages.append({"role": "user", "content": total_prompt})
|
549 |
+
|
550 |
+
## new version of openai API.
|
551 |
+
openai_client = OpenAI()
|
552 |
+
for response in openai_client.chat.completions.create(
|
553 |
+
model=st.session_state["openai_model"],
|
554 |
+
messages=[
|
555 |
+
{"role": m["role"], "content": m["content"]}
|
556 |
+
for m in st.session_state.messages
|
557 |
+
],
|
558 |
+
stream=True,
|
559 |
+
):
|
560 |
+
if str(response.choices[0].delta.content) != 'None':
|
561 |
+
full_response += str(response.choices[0].delta.content)
|
562 |
+
message_placeholder.markdown(full_response + "▌")
|
563 |
+
message_placeholder.markdown(full_response)
|
564 |
+
st.session_state.messages.append(
|
565 |
+
{"role": "assistant", "content": full_response})
|
566 |
+
st.session_state.messages = []
|
567 |
+
|
568 |
+
try:
|
569 |
+
### 显示RAG的source,即查询得到的信息来源出处。
|
570 |
+
print('docs now:', docs)
|
571 |
+
source = rag_source.rag_source(docs) ## get the k reference source of the RAG answer, in a designed format.
|
572 |
+
# print('返回的source内容:', source)
|
573 |
+
st.divider()
|
574 |
+
st.caption(source)
|
575 |
+
except Exception as e:
|
576 |
+
print('Exception:', e)
|
577 |
+
pass
|
578 |
+
|
579 |
+
##TODO 确认是否需要?
|
580 |
+
st.session_state.messages = []
|
581 |
+
# except Exception as e:
|
582 |
+
# print('Exception:', e)
|
583 |
+
# pass
|
584 |
+
|
585 |
|
586 |
+
# async def data_mode():
|
587 |
+
def data_mode():
|
588 |
clear_all() ## reset the conversation.
|
589 |
+
print('数据分析模式启动!')
|
590 |
# uploaded_file_path = './upload.csv'
|
591 |
+
uploaded_file_path = f'./{username}/{username}_upload.csv'
|
592 |
+
# uploaded_file_path = f'./{username}_upload.csv' ### original code here.
|
593 |
print('file path:', uploaded_file_path)
|
594 |
|
|
|
|
|
|
|
|
|
|
|
595 |
# Initialize chat history
|
596 |
if "messages" not in st.session_state:
|
597 |
st.session_state.messages = []
|
|
|
606 |
prompt = st.chat_input("Say something")
|
607 |
print('prompt now:', prompt)
|
608 |
print('----------'*5)
|
|
|
609 |
if prompt:
|
610 |
+
try:
|
611 |
+
st.session_state.messages.append({"role": "user", "content": prompt})
|
612 |
+
with st.chat_message("user"):
|
613 |
+
st.markdown(prompt)
|
614 |
+
|
615 |
+
with st.status('思考中...需要1至10分钟左右,请耐心等待 🏃', expanded=True, state='running') as status:
|
616 |
+
with st.chat_message("assistant"):
|
617 |
+
|
618 |
+
#### Using the open-source CodeInterpreter solution below. May not work after version update, need to upgrade the code accoridngly later on.
|
619 |
+
# from langchain.chat_models import ChatOpenAI
|
620 |
+
# llm_model = ChatOpenAI(model_name="gpt-4-1106-preview")
|
621 |
+
# # llm_model = ChatOpenAI(model_name="gpt-4")
|
622 |
+
# # async with CodeInterpreterSession(llm=llm_model) as session:
|
623 |
+
# import interpreter
|
624 |
+
# interpreter.llm.model = "gpt-3.5-turbo"
|
625 |
+
|
626 |
+
# with CodeInterpreterSession(llm=llm_model) as session:
|
627 |
+
# # with CodeInterpreterSession(llm=llm_model) as session:
|
628 |
+
# print('222')
|
629 |
+
# # user_request = "对于文件中的'SepalLengthCm’数据给我一个'直方图',提供图表,并给出分析结果"
|
630 |
+
# #! 可以用设定dpi=300来输出高质量的图表。(注:图的解析度dpi设定为300)
|
631 |
+
# environ_settings = """【背景要求】如果我没有告诉你任何定制化的要求,那么请你按照以下的默认要求来回答:
|
632 |
+
# -------------------------------------------------------------------------
|
633 |
+
# 1. 你需要用我提问的语言来回答,且默认情况下用中文来回答。
|
634 |
+
# 2. 如果要求你输出图表,那么图的解析度dpi需要设定为600。图尽量使用seaborn库。seaborn库的参数设定:sns.set(rc={'axes.facecolor':'#FFF9ED','figure.facecolor':'#FFF9ED'}, palette='dark'。
|
635 |
+
# 3. 图上所有的文字全部翻译成<英文English>来表示。
|
636 |
+
# 4. 你回答的文字内容必须尽可能的详细且通俗易懂。
|
637 |
+
# 5. 回答时尽可能地展示分析所对应的图表,并提供分析结果。 你需要按如下格式提供内容:
|
638 |
+
# 5.1 提供详细且专业的分析结果,提供足够的分析依据。
|
639 |
+
# 5.2 给出可能造成这一结果的可能原因有哪些?
|
640 |
+
# 以上内容全部用【1, 2, 3这样的序列号格式】来表达。
|
641 |
+
# -------------------------------------------------------------------------
|
642 |
+
# """ # seaborn中的palette参数可以设定图表的颜色,选项包括:deep, muted, pastel, bright, dark, colorblind,Spectral。更多参数可以参考:https://seaborn.pydata.org/generated/seaborn.color_palette.html。
|
643 |
+
|
644 |
+
# # uploaded_file_path = upload_file()
|
645 |
+
|
646 |
+
# user_request = environ_settings + "\n\n" + \
|
647 |
+
# "你需要完成以下任务:\n\n" + prompt + "\n\n" \
|
648 |
+
# f"注:文件位置在 {uploaded_file_path}"
|
649 |
+
# # user_request = str(prompt) ### only prompt without environment prompt.
|
650 |
+
# print('user_request: \n', user_request)
|
651 |
+
|
652 |
+
# # 加载上传的文件,主要路径在上面代码中。
|
653 |
+
# files = [File.from_path(str(uploaded_file_path))]
|
654 |
+
# # files = [File.from_path("/Users/yunshi/Downloads/360Data/Data Center/Working-On Task/演讲与培训/2023ChatGPT/Coding/code_interpreter/rawdata/short_csv.csv")]
|
655 |
+
# # st.write(pd.DataFrame(files))
|
656 |
+
# # print('session.__init__', session.__init__)
|
657 |
+
# # print('session', session.__init__)
|
658 |
+
|
659 |
+
# with st.status('思考中...', expanded=True, state='running') as status:
|
660 |
+
# # generate the response
|
661 |
+
# # response = await session.generate_response(user_msg=user_request, files=files, detailed_error=True)
|
662 |
+
# # response = await session.generate_response(user_msg=user_request, files=files)
|
663 |
+
# response = session.generate_response(user_msg=user_request, files=files)
|
664 |
+
|
665 |
+
# # output to the user
|
666 |
+
# print("AI: ", response.content)
|
667 |
+
# full_response = response.content
|
668 |
+
# ### full_response = "this is full response"
|
669 |
+
|
670 |
+
# # for file in response.files:
|
671 |
+
# for i, file in enumerate(response.files):
|
672 |
+
# # await file.asave(f"/Users/yunshi/Downloads/360Data/Data Center/Working-On Task/演讲与培训/2023ChatGPT/Coding/code_interpreter/output{i}.png") ##working.
|
673 |
+
# # st.image(file.get_image()) #! working.
|
674 |
+
# # * 注意这里的设定,可以提高图片的精细程度。
|
675 |
+
# st.image(file.get_image(), width=None,
|
676 |
+
# output_format='PNG')
|
677 |
+
# # message_placeholder.markdown(full_response + "▌") ## orignal code.
|
678 |
+
# # message_placeholder.markdown(full_response) ## orignal code.
|
679 |
+
# st.write(full_response)
|
680 |
+
# status.update(label='complete', state='complete')
|
681 |
+
# # TODO: 确认是否要记录所有的full response。
|
682 |
+
# st.session_state.messages.append(
|
683 |
+
# {"role": "assistant", "content": full_response})
|
684 |
+
# # await session.astop() # ! 确认需要关闭。
|
685 |
+
# session.astop() # ! 确认需要关闭。
|
686 |
+
# # st.session_state.messages.append({"role": "assistant", "content": full_response})
|
687 |
+
|
688 |
+
#### #### Using the OpenAI's assistant API, wrap into the st_openai_assistant.py.
|
689 |
+
import st_openai_assistant
|
690 |
+
|
691 |
+
### NOTE:在st_openai_assistant.py中可以设置system_prompt.
|
692 |
+
# sys_prompt = """ 1. 你是一位智能AI助手,你连接着一台电脑,但请注意不能联网。在使用Python解决任务时,你可以运行代码并得到结果,如果运行结果有错误,你需要尽可能对代码进行改进。你可以处理用户上传到电脑上的文件。
|
693 |
+
# 2. 你使用matplotlib.pylab(plt)或者seaborn(sns)画图时,需要添加中文字库,代码如下:
|
694 |
+
# matplotlib.rcParams['font.sans-serif'] = ['Microsoft YaHei UI']
|
695 |
+
# sns.set(rc={'axes.facecolor':'#FFF9ED','figure.facecolor':'#FFF9ED'}, palette='dark', font='Microsoft YaHei UI')
|
696 |
+
|
697 |
+
# 3. 如果我没有告诉你任何定制化的要求,那么请你按照以下的默认要求来回答:
|
698 |
+
# 3.1 你回答的文字内容必须尽可能的详细且通俗易懂。
|
699 |
+
# 3.2 回答时尽可能地展示分析所对应的图表,并提供分析结果。图表上的文字采用中文。你需要按如下格式提供内容:
|
700 |
+
# * 提供详细且专业的分析结果,提供足够的分析依据。
|
701 |
+
# * 给出可能造成这一结果的可能原因有哪些?
|
702 |
+
# 以上内容你用序列号1、2、3这样的格式表达。
|
703 |
+
# """
|
704 |
+
# prompt = [
|
705 |
+
# {"role": "system", "content": sys_prompt},
|
706 |
+
# {"role": "user", "content": prompt},
|
707 |
+
# ]
|
708 |
+
messages, text_response, img_response, image_files, final_answer = st_openai_assistant.openai_assistant(prompt=prompt, filepath=uploaded_file_path, username=username)
|
709 |
+
# st.image(img_response) ## show one single image.
|
710 |
+
# st.markdown(text_response) ## show one single text response.
|
711 |
+
|
712 |
+
try:
|
713 |
+
from PIL import Image
|
714 |
+
print("返回到Agent001程序中的图表个数:", len(image_files))
|
715 |
+
# plt.imshow(img)
|
716 |
+
# plt.show()
|
717 |
+
for img in image_files:
|
718 |
+
img = Image.open(img) ## image object.
|
719 |
+
st.image(img, output_format='PNG')
|
720 |
+
# st.image(f"./{username}/{img_response[i]}", output_format='PNG')
|
721 |
+
# st.image(f'/Users/yunshi/Downloads/360Data/Data Center/Working-On Task/演讲与培训/2023ChatGPT/Coding/code_interpreter/joeshi/output{i}.png', output_format='PNG')
|
722 |
+
except:
|
723 |
+
pass
|
724 |
+
|
725 |
+
try:
|
726 |
+
st.markdown(final_answer) ## all messages are appended together, need to print out one by one?
|
727 |
+
except:
|
728 |
+
pass
|
729 |
+
|
730 |
+
except Exception as e:
|
731 |
+
print(e)
|
732 |
+
pass
|
733 |
|
734 |
### authentication with a local yaml file.
|
735 |
import yaml
|
|
|
744 |
config['preauthorized']
|
745 |
)
|
746 |
|
|
|
|
|
747 |
# authentication with a remove cloud-based database.
|
748 |
# 导入云端用户数据库。
|
749 |
|
|
|
754 |
|
755 |
# deta = Deta(DETA_KEY)
|
756 |
|
757 |
+
# mybase is the name of the database in Deta. You can change it to any name you want.
|
758 |
# credentials = {"usernames":{}}
|
759 |
+
# # credentials = {"users": {}}
|
760 |
+
# # db = db()
|
761 |
# users = []
|
762 |
# email = []
|
763 |
# passwords = []
|
|
|
789 |
# database_table.append([i,credentials['usernames'][i]['name'],credentials['usernames'][i]['password']])
|
790 |
# print("database_table:",database_table)
|
791 |
|
792 |
+
|
793 |
# authenticator = stauth.Authenticate(
|
794 |
# credentials=credentials, cookie_name="joeshi_gpt", key='abcedefg', cookie_expiry_days=30)
|
795 |
|
796 |
+
user, authentication_status, username = authenticator.login('用户登录', 'main')
|
797 |
+
# print("name", name, "username", username)
|
798 |
+
|
799 |
# ## sign-up widget,未完成。
|
800 |
# try:
|
801 |
# if authenticator.register_user('新用户注册', preauthorization=False):
|
|
|
806 |
# st.success('注册成功!')
|
807 |
# except Exception as e:
|
808 |
# st.error(e)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
809 |
|
810 |
if authentication_status:
|
811 |
with st.sidebar:
|
|
|
813 |
"""
|
814 |
<style>
|
815 |
[data-testid="stSidebar"][aria-expanded="true"]{
|
816 |
+
min-width: 450px;
|
817 |
+
max-width: 450px;
|
818 |
}
|
819 |
""",
|
820 |
unsafe_allow_html=True,
|
821 |
)
|
822 |
+
### siderbar的题目。
|
823 |
st.header(f'**欢迎 **{username}** 来到人工智能的世界** ♠')
|
824 |
+
st.write(f'_Welcome and Hope U Enjoy Staying Here_')
|
825 |
authenticator.logout('登出', 'sidebar')
|
826 |
|
827 |
+
# reset_button_key = "reset_button"
|
828 |
+
# reset_button = st.button(label=("清除所有记录,并开启一轮新对话 ▶"),
|
829 |
+
# key=reset_button_key, use_container_width=True, type="primary")
|
830 |
+
# ## 清除所有对话记录, reset all conversation.
|
831 |
+
# if reset_button:
|
832 |
+
# reset_all()
|
833 |
+
|
834 |
+
# st.markdown("#### 大语言模型设置")
|
835 |
+
# with st.expander(label='**选择一个大语言模型基座**', expanded=True):
|
836 |
+
radio_1 = st.selectbox(
|
837 |
+
label='选择一个大语言模型基座 (注:根据站点不同,部分基座不可用)',
|
838 |
+
options=["ChatGPT-4", "ChatGPT-3.5", "Google Gemini","Claude 3", "清华ChatGLM3-6B", "百川Baichuan-13B", "阿里通义千问-14B", "阿里通义千问-72B", "Llama-2", "Mistral", "Vicuna"],
|
839 |
+
index=0,
|
840 |
+
placeholder="大语言模型列表",
|
841 |
+
)
|
842 |
|
843 |
+
## 在sidebar上的三个分页显示,用st.tabs实现。
|
844 |
+
tab_1, tab_2, tab_3, tab_4 = st.tabs(['使用须知', '模型参数', '提示词模板', '系统角色设定'])
|
845 |
+
|
846 |
+
# with st.expander(label='**使用须知**', expanded=False):
|
847 |
+
with tab_1:
|
848 |
+
# st.markdown("#### 快速上手指南")
|
849 |
# with st.text(body="说明"):
|
850 |
# st.markdown("* 重启一轮新对话时,只需要刷新页面(按Ctrl/Command + R)即可。")
|
851 |
with st.text(body="说明"):
|
852 |
st.markdown("* 为了保护数据与隐私,所有对话均不会被保存,刷新页面立即删除。敬请放心。")
|
853 |
with st.text(body="说明"):
|
854 |
+
st.markdown("* “GPT-4”回答质量极佳,但速度缓慢,建议适当使用。")
|
855 |
with st.text(body="说明"):
|
856 |
+
st.markdown("* “信息检索模式”与所有的搜索引擎或者数据库检索方式一样,仅限一轮对话,将不会保持之前的会话记录。")
|
857 |
with st.text(body="说明"):
|
858 |
st.markdown(
|
859 |
+
"* “数据分析模式”暂时只支持1000个单元格以内的数据分析,单元格中的内容不支持中文数据(表头也尽量不使用中文)。一般运行时间在1至10分钟左右,期间需要保持网络畅通。")
|
860 |
with st.text(body="说明"):
|
861 |
+
st.markdown("* “数据分析模式”推荐上传csv格式的文件,部分Excel文件容易出现数据不兼容的情况。")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
862 |
|
863 |
+
## 大模型参数
|
864 |
+
# with st.expander(label='**大语言模型参数**', expanded=True):
|
865 |
+
with tab_2:
|
866 |
+
max_tokens = st.slider(label='Max_Token(生成结果时最大字数)', min_value=100, max_value=8096, value=4096,step=100)
|
867 |
+
temperature = st.slider(label='Temperature (温度)', min_value=0.0, max_value=1.0, value=0.8, step=0.1)
|
868 |
+
top_p = st.slider(label='Top_P (核采样)', min_value=0.0, max_value=1.0, value=0.6, step=0.1)
|
869 |
+
frequency_penalty = st.slider(label='Frequency Penalty (重复度惩罚因子)', min_value=-2.0, max_value=2.0, value=1.0, step=0.1)
|
870 |
+
presence_penalty = st.slider(label='Presence Penalty (控制主题的重复度)', min_value=-2.0, max_value=2.0, value=1.0, step=0.1)
|
|
|
871 |
|
872 |
+
## reset password widget
|
873 |
+
# try:
|
874 |
+
# if authenticator.reset_password(st.session_state["username"], 'Reset password'):
|
875 |
+
# st.success('Password modified successfully')
|
876 |
+
# except Exception as e:
|
877 |
+
# st.error(e)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
878 |
|
879 |
+
# with st.header(body="欢迎"):
|
880 |
+
# st.markdown("# 欢迎使用大语言模型商业智能中心")
|
881 |
+
# with st.expander(label=("**重要的使用注意事项**"), expanded=True):
|
882 |
+
# with st.container():
|
883 |
+
|
884 |
+
|
885 |
+
with tab_3:
|
886 |
+
# st.markdown("#### Prompt提示词参考资料")
|
887 |
+
with st.expander(label="**大语言模型基础提示词Prompt示例**", expanded=False):
|
888 |
+
st.code(
|
889 |
+
body="继续用中文写一篇关于 [文章主题] 的文章,以下列句子开头:[文章开头]。", language='plaintext')
|
890 |
+
st.code(body="将以下文字概括为 100 个字,使其易于阅读和理解。避免使用复杂的句子结构或技术术语。",
|
891 |
+
language='plaintext')
|
892 |
+
st.code(body="给我出一个迪奥2024春季发布会活动策划。", language='plaintext')
|
893 |
+
st.code(body="帮我按照正式会议结构写一个会邀:主题是xx手机游戏立项会议。", language='plaintext')
|
894 |
+
st.code(body="帮我写一个车内健康监测全场景落地的项目计划,用表格。", language='plaintext')
|
895 |
+
st.code(
|
896 |
+
body="同时掷两枚质地均匀的骰子,则两枚骰子向上的点数之和为 7 的概率是多少。", language='plaintext')
|
897 |
+
st.code(body="写一篇产品经理的演讲稿,注意使用以下词汇: 赋能,抓手,中台,闭环,落地,漏斗,沉淀,给到,同步,对齐,对标,迭代,拉通,打通,升级,交付,聚焦,倒逼,复盘,梳理,方案,联动,透传,咬合,洞察,渗透,兜底,解耦,耦合,复用,拆解。", language='plaintext')
|
898 |
+
|
899 |
+
with st.expander(label="**数据分析模式的专用提示词Prompt示例**", expanded=False):
|
900 |
+
# with st.subheader(body="提示词Prompt"):
|
901 |
+
st.code(body="分析此数据集并绘制一些'有趣的图表'。", language='python')
|
902 |
+
st.code(
|
903 |
+
body="对于这个文件中的数据,你需要要找出[X,Y]数据之间的寻找'相关性'。", language='python')
|
904 |
+
st.code(body="对于这个文件中的[xxx]数据给我一个'整体的分析'。", language='python')
|
905 |
+
st.code(body="对于[xxx]数据给我一个'直方图',提供图表,并给出分析结果。", language='python')
|
906 |
+
st.code(body="对于[xxx]数据给我一个'小提琴图',并给出分析结果。", language='python')
|
907 |
+
st.code(
|
908 |
+
body="对于[X,Y,Z]数据在一个'分布散点图 (stripplot)',所有的数据在一张图上展现, 并给出分析结果。", language='python')
|
909 |
+
st.code(body="对于[X,Y]数据,进行'T检验',你需要展示图表,并给出分析结果。",
|
910 |
+
language='python')
|
911 |
+
st.code(body="对于[X,Y]数据给我一个3个类别的'聚类分析',并给出分析结果。",
|
912 |
+
language='python')
|
913 |
+
|
914 |
+
with tab_4:
|
915 |
+
st.text_area(label='系统角色设定', value='你是一个人工智能,你需要回答我提出的问题,或者完成我交代的任务。你需要使用我提问的语言(如中文、英文)来回答。', height=200, label_visibility='hidden')
|
916 |
|
|
|
917 |
|
918 |
+
elif authentication_status == False:
|
919 |
+
st.error('⛔ 用户名或密码错误!')
|
920 |
+
elif authentication_status == None:
|
921 |
+
st.warning('⬆ 请先登录!')
|
922 |
|
923 |
if __name__ == "__main__":
|
924 |
import asyncio
|
925 |
try:
|
926 |
+
match navi_menu:
|
927 |
+
case "核心模式":
|
928 |
+
# if navi_menu == "核心模式":
|
929 |
+
print(f'navi_menu 选择了 {navi_menu}')
|
930 |
+
reset_all()
|
931 |
+
# * 也可以用命令执行这个python文件。’streamlit run frontend/app.py‘
|
932 |
+
# asyncio.run(text_mode())
|
933 |
+
text_mode()
|
934 |
+
|
935 |
+
# elif navi_menu == "信息检索模式":
|
936 |
+
case "联网模式":
|
937 |
+
# print(f'navi_menu 选择了 {navi_menu}')
|
938 |
+
# reset_all()
|
939 |
+
##TODO 如下设置中的index=None, 可能可以解决了刷屏会调回第一项的问题?好像不会。
|
940 |
+
# radio_2 = st.radio(label='信息检索源选择:', options=['互联网', '维基百科', '本地文档', '文献库', '企业知识库','知识图谱库'], horizontal=True, label_visibility='visible')
|
941 |
+
|
942 |
+
### 横向排列的checkbox选项。也可以实现多项选择的功能。
|
943 |
+
# col_1, col_2, col_3, col_4, col_5 = st.columns(5)
|
944 |
+
# rag_1 = col_1.checkbox(label='互联网', label_visibility='visible')
|
945 |
+
# rag_2 = col_2.checkbox(label='上传文件', label_visibility='visible')
|
946 |
+
# rag_3 = col_3.checkbox(label='企业知识库', label_visibility='visible')
|
947 |
+
# rag_4 = col_4.checkbox(label='百科全书', label_visibility='visible')
|
948 |
+
# rag_5 = col_5.checkbox(label='其他数据源', label_visibility='visible')
|
949 |
+
|
950 |
+
if (navi_menu=='联网模式'):
|
951 |
+
# print(f'radio 选择了 {radio_2}')
|
952 |
+
# asyncio.run(text_mode())
|
953 |
+
text_mode()
|
954 |
+
|
955 |
+
case "知识库模式":
|
956 |
+
print(f'navi_menu 选择了 {navi_menu}')
|
957 |
+
st.session_state.messages = []
|
958 |
+
|
959 |
+
# ### llama_index框架的RAG代码,最近更新版本后不成功,会报错。
|
960 |
+
# path = f'./{username}/vector_store.json'
|
961 |
+
# if os.path.exists(path):
|
962 |
+
# print(f'{path} local KB exists')
|
963 |
+
# database_info = pd.read_csv(f'./{username}/database_name.csv')
|
964 |
+
# current_database_name = database_info.iloc[-1][0]
|
965 |
+
# current_database_date = database_info.iloc[-1][1]
|
966 |
+
# database_claim = f"当前知识库为:{current_database_name},创建于{current_database_date}。可以开始提问!"
|
967 |
+
# st.markdown(database_claim)
|
968 |
+
|
969 |
+
### Langchain框架的RAG代码。
|
970 |
+
path = f'./{username}/faiss_index/index.faiss'
|
971 |
+
if os.path.exists(path):
|
972 |
+
print(f'{path} local KB exists')
|
973 |
+
database_info = pd.read_csv(f'./{username}/database_name.csv', encoding='utf-8', header=None) ## 不加encoding的话,中文名字的PDF会报错。
|
974 |
+
print(database_info)
|
975 |
+
current_database_name = database_info.iloc[-1][0]
|
976 |
+
current_database_date = database_info.iloc[-1][1]
|
977 |
+
database_claim = f"当前知识库为:{current_database_name},创建于{current_database_date}。可以开始提问!"
|
978 |
+
st.markdown(database_claim)
|
979 |
+
|
980 |
+
try:
|
981 |
+
uploaded_file = st.file_uploader(
|
982 |
+
"选择上传一个新知识库", type=(["pdf"]))
|
983 |
+
# 默认状态下没有上传文件,None,会报错。需要判断。
|
984 |
+
if uploaded_file is not None:
|
985 |
+
# uploaded_file_path = upload_file(uploaded_file)
|
986 |
+
upload_file(uploaded_file)
|
987 |
+
except Exception as e:
|
988 |
+
print(e)
|
989 |
+
pass
|
990 |
+
|
991 |
+
try:
|
992 |
+
## 启动本地知识库模式。
|
993 |
+
localKB_mode(username)
|
994 |
+
# asyncio.run(localKB_mode(username))
|
995 |
+
except Exception as e:
|
996 |
+
print(e)
|
997 |
+
pass
|
998 |
+
|
999 |
+
|
1000 |
+
# elif navi_menu == "数据分析模式":
|
1001 |
+
case "数据分析模式":
|
1002 |
+
# reset_message()
|
1003 |
+
uploaded_file = st.file_uploader(
|
1004 |
+
"选择一个文件", type=(["csv", "xlsx", "xls"]))
|
1005 |
+
# 默认状态下没有上传文件,None,会报错。需要判断。
|
1006 |
+
if uploaded_file is not None:
|
1007 |
+
# uploaded_file_path = upload_file(uploaded_file) ### original code here.
|
1008 |
+
csv_filename = upload_file(uploaded_file)
|
1009 |
+
# asyncio.run(data_mode())
|
1010 |
+
reset_all()
|
1011 |
+
data_mode()
|
1012 |
+
|
1013 |
+
# elif navi_menu == "智能体模式":
|
1014 |
+
case "智能体模式":
|
1015 |
+
uploaded_file = st.file_uploader(
|
1016 |
+
"选择一个文件", type=(["csv"]))
|
1017 |
+
reset_all()
|
1018 |
+
print('st uploaded_file:',uploaded_file)
|
1019 |
+
# 默认状态下没有上传文件,None,会报错。需要判断。
|
1020 |
+
# if uploaded_file is not None:
|
1021 |
+
if uploaded_file is not None:
|
1022 |
+
uploaded_file_path = upload_file(uploaded_file)
|
1023 |
+
# asyncio.run(data_mode())
|
1024 |
+
else:
|
1025 |
+
uploaded_file_path = None
|
1026 |
+
# st.markdown('**此功能还在内部测试阶段,尚未开放,敬请期待!**')
|
1027 |
+
|
1028 |
+
# reset_message()
|
1029 |
+
print('st_msautogen starts!')
|
1030 |
+
uploaded_file_path = '/Users/yunshi/Downloads/360Data/Data Center/Working-On Task/演讲与培训/2023ChatGPT/Coding/code_interpreter/joeshi_upload.csv'
|
1031 |
+
# asyncio.run(st_msautogen.auto_gen(uploaded_file_path)) ## 好像不需要启动asyncio,也可以正常运行。在msautogen中已经启动了。
|
1032 |
+
st_msautogen.auto_gen(uploaded_file_path) ## 这里不需要使用asyncio.run(),否则会streamlit中会刷新页面?
|
1033 |
+
|
1034 |
+
except Exception as e:
|
1035 |
+
print('Exception Raised:',e)
|
1036 |
+
pass
|
chatgpt.py
ADDED
@@ -0,0 +1,61 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
# -*- coding: utf-8 -*-
|
3 |
+
import requests
|
4 |
+
import streamlit as st
|
5 |
+
import openai
|
6 |
+
from openai import OpenAI
|
7 |
+
import os
|
8 |
+
from dotenv import load_dotenv
|
9 |
+
import numpy as np
|
10 |
+
import pandas as pd
|
11 |
+
import csv
|
12 |
+
import tempfile
|
13 |
+
from tempfile import NamedTemporaryFile
|
14 |
+
import pathlib
|
15 |
+
from pathlib import Path
|
16 |
+
import re
|
17 |
+
from re import sub
|
18 |
+
import matplotlib.pyplot as plt
|
19 |
+
from itertools import product
|
20 |
+
from tqdm import tqdm_notebook, tqdm, trange
|
21 |
+
import time
|
22 |
+
from time import sleep
|
23 |
+
import pretty_errors
|
24 |
+
import seaborn as sns
|
25 |
+
from matplotlib.pyplot import style
|
26 |
+
from rich import print
|
27 |
+
import warnings
|
28 |
+
warnings.filterwarnings('ignore')
|
29 |
+
|
30 |
+
load_dotenv()
|
31 |
+
### 设置openai的API key
|
32 |
+
os.environ["OPENAI_API_KEY"] = os.environ['user_token']
|
33 |
+
openai.api_key = os.environ['user_token']
|
34 |
+
|
35 |
+
##NOTE: 需要注意OpenAI的版本
|
36 |
+
def chatgpt(user_prompt, sys_prompt="You are professional consultant", openai_model="gpt-3.5-turbo-16k"):
|
37 |
+
# print('chatgpt mode starts!')
|
38 |
+
try:
|
39 |
+
openai_client = OpenAI()
|
40 |
+
chatgpt_response = openai_client.chat.completions.create(
|
41 |
+
model=openai_model,
|
42 |
+
messages=[
|
43 |
+
{"role": "system", "content": sys_prompt},
|
44 |
+
{"role": "user", "content": user_prompt},
|
45 |
+
],
|
46 |
+
stream=False,
|
47 |
+
)
|
48 |
+
|
49 |
+
full_response = chatgpt_response.choices[0].message.content
|
50 |
+
print(full_response)
|
51 |
+
except Exception as e:
|
52 |
+
print(e)
|
53 |
+
full_response = "Sorry, I don't have an answer for that yet."
|
54 |
+
|
55 |
+
# return chatgpt_response, full_response
|
56 |
+
return full_response
|
57 |
+
|
58 |
+
# user_prompt = "where is Shanghai?"
|
59 |
+
# system_prompt = "You are a helpful AI assistant."
|
60 |
+
# full_response = chatgpt(user_prompt=user_prompt, sys_prompt=system_prompt)
|
61 |
+
# print(full_response)
|
joeshi/.DS_Store
ADDED
Binary file (6.15 kB). View file
|
|
joeshi/database_name.csv
CHANGED
@@ -1,6 +1,24 @@
|
|
1 |
-
|
2 |
-
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
大模型LLM解决方案调研问卷.pdf,2024-01-18 20:12
|
2 |
+
joeshi_upload.pdf,2024-01-18 22:32
|
3 |
+
joeshi_upload.pdf,2024-01-18 22:32
|
4 |
+
joeshi_upload.pdf,2024-01-18 22:32
|
5 |
+
专属ChatGPT产品介绍 20230415.pdf,2024-03-10 21:37
|
6 |
+
专属ChatGPT产品介绍 20230415.pdf,2024-03-10 21:37
|
7 |
+
Web App开发需求说明 20230512.pdf,2024-03-10 23:02
|
8 |
+
Web App开发需求说明 20230512.pdf,2024-03-10 23:02
|
9 |
+
Web App开发需求说明 20230512.pdf,2024-03-10 23:02
|
10 |
+
大模型LLM解决方案调研问卷.pdf,2024-03-12 12:51
|
11 |
+
大模型LLM解决方案调研问卷.pdf,2024-03-12 12:51
|
12 |
+
企业数据资源相关会计处理暂行规定.pdf,2024-03-12 12:51
|
13 |
+
企业数据资源相关会计处理暂行规定.pdf,2024-03-12 12:54
|
14 |
+
企业数据资源相关会计处理暂行规定.pdf,2024-03-12 12:54
|
15 |
+
企业数据资源相关会计处理暂行规定.pdf,2024-03-12 12:54
|
16 |
+
joeshi_upload.pdf,2024-03-12 18:19
|
17 |
+
joeshi_upload.pdf,2024-03-12 18:19
|
18 |
+
joeshi_upload.pdf,2024-03-12 18:20
|
19 |
+
企业数据资源相关会计处理暂行规定.pdf,2024-03-12 21:53
|
20 |
+
企业数据资源相关会计处理暂行规定.pdf,2024-03-12 22:40
|
21 |
+
企业数据资源相关会计处理暂行规定.pdf,2024-03-12 22:54
|
22 |
+
企业数据资源相关会计处理暂行规定.pdf,2024-03-12 22:57
|
23 |
+
AI与ChatGPT讲座大纲 20230615.pdf,2024-03-13 13:20
|
24 |
+
AI与ChatGPT讲座大纲 20230615.pdf,2024-03-13 13:20
|
joeshi/default__vector_store.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"embedding_dict": {"2d707a34-28f5-4cfa-8973-961c2f7e96e1": [-0.010600954294204712, -0.007286855950951576, 0.010656420141458511, -0.02024511806666851, 0.015849431976675987, 0.002634638687595725, -0.003260365454480052, 0.02477947063744068, -0.014032918028533459, -0.03535962477326393, 0.003071434097364545, 0.027802372351288795, -0.021604038774967194, -0.0021302441600710154, -0.0038583590649068356, -0.003899958683177829, -0.010677220299839973, -0.002721304539591074, 0.019163530319929123, -0.0012913196114823222, -0.01995392143726349, 0.024696271866559982, -0.01525317132472992, 0.0017263817135244608, -0.0022775758989155293, 0.010781219229102135, 0.025389596819877625, -0.015669167041778564, 0.001039988943375647, 0.012757197953760624, -0.0007518253987655044, -0.018484070897102356, -0.02182590216398239, -0.0011015216587111354, -0.006662862841039896, -0.016071297228336334, 0.00825751293450594, -0.00332276476547122, 0.02959115244448185, -0.010115626268088818, 0.009394567459821701, 0.01849793829023838, 0.024765605106949806, 0.0011665209894999862, -0.02265789359807968, 0.010143359191715717, 0.008992438204586506, -0.018026474863290787, 0.008493243716657162, 0.01518383901566267, -0.015336371026933193, 0.021340573206543922, -0.009020171128213406, -0.0029050358571112156, -0.006787661463022232, 0.006770328152924776, -0.0064444649033248425, 0.020078720524907112, -0.011266547255218029, 0.0006651596049778163, 0.012251070700585842, 0.0006075269193388522, -0.029230624437332153, -0.007321522571146488, -0.021437639370560646, 0.007910849526524544, -0.02695651538670063, 0.00843084417283535, 0.006077002268284559, 0.017097419127821922, 0.018428605049848557, 0.013055328279733658, -0.0025150401052087545, -0.008950838819146156, 0.041821423918008804, -0.006666329223662615, -0.014573711901903152, -0.013804120011627674, -0.008084180764853954, 0.010108692571520805, 0.021604038774967194, -0.020064854994416237, 0.007536453660577536, 0.01512837316840887, 0.007460187654942274, 0.029286090284585953, 0.004638351034373045, -0.006378598976880312, -0.014490513131022453, 0.0010469222906976938, 0.00860417541116476, 0.0031806330662220716, 0.020494716241955757, 0.010587087832391262, -0.017125152051448822, 0.03036767803132534, -0.002047044923529029, -0.0005160945584066212, -0.00717592379078269, -0.03996330872178078, 0.007702851668000221, -0.019537925720214844, -0.031698863953351974, -0.0023642415180802345, -0.041655026376247406, 0.012479867786169052, 0.03474950045347214, 0.021007778123021126, 0.01412305049598217, 0.016334760934114456, -0.01514223963022232, 0.028897827491164207, -0.008819106966257095, -0.017069686204195023, -0.005886337719857693, -0.009193502366542816, 0.015932630747556686, 0.0007947249105200171, 0.0014265181962400675, -0.008146580308675766, 0.020439250394701958, -0.01324945967644453, -0.014878775924444199, -0.01346439030021429, 0.004714616574347019, -0.003893025452271104, -0.014781709760427475, -0.02577785961329937, 0.033335112035274506, -0.01240360178053379, -0.008091114461421967, 0.00029748017550446093, 0.010594021528959274, -0.0027507708873599768, -0.04068436846137047, 0.020037120208144188, -0.025958124548196793, 0.0033695644233375788, -0.03635801374912262, -0.016071297228336334, 0.003315831534564495, 0.035914286971092224, -0.011446812190115452, -0.01568303443491459, 0.00614286819472909, 0.025070667266845703, -0.008312978781759739, 0.021035511046648026, 0.008846839889883995, 0.017527280375361443, 0.01334652490913868, -0.02081364579498768, 0.00965109746903181, 0.040018774569034576, 0.012001472525298595, -0.009034037590026855, 0.012667065486311913, 0.02945248782634735, -0.005345543380826712, -0.009734297171235085, -0.011828141286969185, 0.03025674633681774, 0.01145374495536089, 0.013083061203360558, 0.018387004733085632, 0.02672078274190426, -0.0029206357430666685, 0.011495345272123814, 0.01521157193928957, 0.007966315373778343, -0.014559845440089703, 0.003267298685386777, -0.0209939107298851, 0.0007713251397944987, 0.012687865644693375, -0.009089503437280655, 0.013041461817920208, -0.0012765864375978708, -0.025112267583608627, -0.020203519612550735, -0.015059040859341621, -0.006548463832587004, 0.012008406221866608, 0.008819106966257095, -0.01989845559000969, -0.0064375316724181175, 0.0009654564200900495, -0.014698510989546776, 0.01419238280504942, -0.03480496630072594, 0.035692423582077026, 0.019024865701794624, 0.010649487376213074, -0.019621126353740692, -0.5959275364875793, -0.018858466297388077, 0.011925207450985909, -0.014767843298614025, -0.0073839216493070126, 0.0068535273894667625, 0.0012531867250800133, 0.005446075461804867, -0.006420198827981949, -0.006382065825164318, -0.029008759185671806, 0.0011899207020178437, 0.003275965340435505, -0.007896983064711094, -0.011717209592461586, -0.01602969691157341, -0.006742595229297876, 0.003518629353493452, 0.006628196686506271, 0.03028447926044464, -0.041682757437229156, 0.0055188750848174095, 0.01433798111975193, 0.0045447517186403275, -0.01118334848433733, 0.01406758464872837, 0.022463761270046234, 0.0028097035828977823, -0.00432982062920928, 0.004437286406755447, -0.046868838369846344, 0.009817495942115784, 0.004981547128409147, -0.013041461817920208, 0.04575951397418976, -0.015308638103306293, -0.018913932144641876, 0.008839906193315983, 0.009644164703786373, 0.018997132778167725, -0.012895863503217697, -0.01806807518005371, 0.004253555089235306, 0.010802019387483597, -0.0040143574588000774, 0.0031858328729867935, 0.030811406672000885, -0.005983403418213129, -0.02348988503217697, 0.004569018259644508, 0.0036330281291157007, -0.01761048100888729, 0.00837537832558155, -0.00702685909345746, 0.006642063148319721, 0.0010451889829710126, 0.034971363842487335, -0.047673095017671585, 0.0024803737178444862, -0.016501158475875854, -0.001149187795817852, 0.010850551538169384, 0.0042986213229596615, -0.02437734231352806, -0.022491494193673134, 0.019565658643841743, -0.015585968270897865, -0.008118847385048866, 0.018650468438863754, -0.004381820093840361, -0.014268648810684681, 0.012999862432479858, 1.0467597348906565e-05, -0.019884590059518814, -0.009006304666399956, 0.03613615036010742, 0.006624729838222265, 0.0028807695489376783, -7.458670734195039e-05, 0.025334130972623825, 0.004506618715822697, 0.0009663230739533901, 0.027233844622969627, -0.00037937931483611465, 0.004517018795013428, 0.005952203646302223, -0.012050005607306957, -0.011869740672409534, 0.016875553876161575, -0.0056506069377064705, -0.0018199806800112128, 0.02864822931587696, -0.007453254424035549, -0.009886828251183033, -0.0015660501085221767, 0.012701732106506824, -0.01608516275882721, -0.025736261159181595, -0.000288380280835554, -0.02196456678211689, 0.007945516146719456, -0.02939702197909355, 0.015267038717865944, 0.027594374492764473, 0.037633735686540604, 0.027053579688072205, -0.03294685110449791, -0.020439250394701958, 0.01820673979818821, -0.008479377254843712, -0.018081942573189735, -0.005456475540995598, 0.0021787770092487335, -0.029702086001634598, 0.006146335043013096, -0.027705306187272072, 0.01327719260007143, -0.004596751183271408, 0.003206632798537612, -0.0034666298888623714, 0.006780728232115507, -0.015988096594810486, 0.01632089354097843, -0.00620873412117362, 0.02150697261095047, 0.03228125721216202, 0.01205693930387497, -0.018969399854540825, 0.025486662983894348, 0.02416934445500374, -0.005962603725492954, -0.035692423582077026, 0.03857665881514549, -0.013845720328390598, 0.047673095017671585, 0.007564186584204435, 0.024724004790186882, 0.0021978432778269053, 0.004558618180453777, -0.009158836677670479, -0.025555996224284172, -0.004881015047430992, 0.02615225687623024, 0.003033301094546914, -0.011301213875412941, -0.02325415425002575, -0.01605742983520031, -0.005574340932071209, 0.0017142485594376922, -0.017596613615751266, -0.003470096504315734, -0.017943276092410088, 0.022436028346419334, 0.04073983430862427, -0.003920758608728647, 0.0018529137596488, 0.00931830145418644, -0.03533189371228218, -0.0013337858254089952, -0.008153514005243778, -0.006714862305670977, 0.01973205804824829, -0.003667694516479969, -0.015544368885457516, -0.006239933893084526, -0.0191496629267931, -0.009567898698151112, 0.02518160082399845, -0.030950071290135384, -0.016015829518437386, 0.01218867115676403, 0.00851404294371605, -0.019607258960604668, 0.0021441106218844652, -0.005844737868756056, -0.021298974752426147, 0.0022082433570176363, -0.017582746222615242, 0.021631771698594093, -0.010094826109707355, 0.00037092939601279795, 0.027746906504034996, -0.009796695783734322, -0.0008454243652522564, 0.003750893520191312, -0.0026675716508179903, -0.0015851164935156703, 0.022172564640641212, -0.017457948997616768, 0.024502141401171684, -0.01114174909889698, 0.031587932258844376, -0.03203165903687477, 0.03375110775232315, -0.008874572813510895, 0.046009112149477005, 0.009380700998008251, -0.02500133588910103, -0.017083551734685898, 0.006434065289795399, -0.002868636278435588, 0.014920375309884548, -0.003274231916293502, -0.007113524712622166, 0.013755587860941887, -0.014393447898328304, 0.018137408420443535, -0.011301213875412941, 0.014448913745582104, -0.012334269471466541, 0.011918273754417896, 0.010947617702186108, -0.024585340172052383, 0.005317810457199812, -0.01409531757235527, 0.010815885849297047, -0.007841517217457294, 0.007397788111120462, -0.022463761270046234, -0.012542267329990864, -0.01139827910810709, -0.003778626676648855, 0.01220253761857748, 0.017901677638292313, -0.013956652022898197, 0.02733091078698635, -0.013180127367377281, 0.017014220356941223, 0.01731928251683712, -0.014213182963430882, -0.004569018259644508, -0.0006361266132444143, 0.011023882776498795, 0.01225800346583128, -0.003653828054666519, -0.015530502423644066, 0.03039541095495224, -0.015322504565119743, 0.03278045356273651, -0.0041564893908798695, -0.0094223003834486, 0.006680195685476065, 0.02228349819779396, -0.04961440712213516, 0.011613210663199425, 0.0021926434710621834, 0.010233491659164429, 0.010261224582791328, -0.01952406018972397, 0.02862049639225006, -0.03294685110449791, 0.02961888536810875, -0.024266408756375313, -0.042043287307024, 0.04831095412373543, -0.007141257636249065, 0.0024422407150268555, 0.00602153642103076, -0.007987115532159805, 0.017915543168783188, -0.002539306413382292, 0.019856857135891914, 0.005390609614551067, 0.022006167098879814, 0.011273480951786041, -0.023822680115699768, -0.0026138389948755503, -0.011481478810310364, -0.010372157208621502, -0.0169171541929245, 0.03286365047097206, -0.018941666930913925, 0.017873944714665413, -0.0095054991543293, 0.014365714974701405, 2.046124245680403e-05, 0.006250333972275257, 0.027247712016105652, -0.006503397598862648, 0.023351218551397324, -0.009151902981102467, -0.023101622238755226, -0.0010529888095334172, 0.018220607191324234, -0.019995521754026413, -0.012972129508852959, -0.009443099610507488, -0.018276073038578033, -0.03111647069454193, -0.005078612826764584, -0.01114174909889698, -0.013214793056249619, -0.010344423353672028, 0.015932630747556686, -0.020175786688923836, 0.011432945728302002, -0.013138527050614357, -0.00558474101126194, -0.032530855387449265, -0.008396177552640438, 0.009040971286594868, -0.008389244787395, -0.044289663434028625, -0.02075817994773388, 0.001915313070639968, -0.0169726200401783, -0.04903201386332512, -0.012119337916374207, -0.0031650331802666187, -0.007370055187493563, 0.026485051959753036, -0.01645955815911293, -0.00620873412117362, -0.03710680827498436, 0.011939073912799358, -0.0020297118462622166, -0.010684153065085411, -0.014837176539003849, 0.03655214607715607, 0.005446075461804867, -0.03022901341319084, -0.01849793829023838, -0.02064724825322628, 0.021687237545847893, 0.04323580861091614, 0.05502235144376755, 0.007896983064711094, -0.0013441857881844044, -0.016251560300588608, 0.01573850028216839, -0.02939702197909355, -0.04526032134890556, 0.01119721494615078, -0.0021770435851067305, -0.014948108233511448, -0.009034037590026855, 0.006683662533760071, 0.005851671565324068, 0.013928919099271297, 0.003173699602484703, 0.021285107359290123, -0.015447302721440792, -0.008139647543430328, 0.008014848455786705, -0.012708665803074837, -0.0020435783080756664, 0.02942475490272045, 0.006125535350292921, -0.0073908548802137375, -0.009956161491572857, 0.0231154877692461, 0.017901677638292313, -0.0033816974610090256, -0.011044682934880257, -0.017984876409173012, -0.0015140506438910961, 0.005962603725492954, -0.010233491659164429, -0.010344423353672028, 0.04406780004501343, -0.029757551848888397, -0.001527917105704546, -0.0013866520021110773, 0.015988096594810486, 0.012729465030133724, 0.03289138525724411, 0.011259614489972591, -0.030866872519254684, 0.01118334848433733, -0.014767843298614025, 0.015558235347270966, 0.0075503201223909855, -0.011966806836426258, -0.0047042169608175755, 0.012604666873812675, 0.025985857471823692, -0.02437734231352806, 0.00665939599275589, -0.0004952947492711246, 0.0036226282827556133, -0.0033435646910220385, -0.015849431976675987, 0.013623856008052826, -0.040767569094896317, -0.01774914562702179, -0.033251915127038956, -0.012105471454560757, -0.01726381666958332, 0.008347644470632076, 0.004988480359315872, -0.0211048424243927, -0.03228125721216202, -0.0037959597539156675, 0.015308638103306293, 0.00843084417283535, -0.017971009016036987, -0.029147423803806305, -0.012348135933279991, 0.00567833986133337, -0.007023392245173454, 0.02376721426844597, 0.017901677638292313, -0.017014220356941223, 0.008299112319946289, 0.01995392143726349, -0.037467338144779205, 0.007037258706986904, -0.02494587004184723, -0.003071434097364545, 0.007037258706986904, -0.011127881705760956, -0.009172703139483929, 0.004499685484915972, 0.014448913745582104, -0.002736904425546527, -0.002509840065613389, -0.0013138527283445, 0.013707054778933525, 0.011114015243947506, -0.001543516991659999, -0.008923105895519257, -0.004073290154337883, 0.02371174842119217, -0.007196723949164152, -0.02609679102897644, -0.04797815904021263, 0.004541285336017609, -0.0337233766913414, 0.016473425552248955, 0.027774639427661896, 0.017527280375361443, 0.0030125014018267393, 0.008423910476267338, -0.004589817952364683, 0.02956341952085495, 0.0034302303101867437, 0.0018719801446422935, -0.01946859434247017, 0.011100148782134056, 0.009893761947751045, 0.017513414844870567, 0.026623718440532684, 0.010982283391058445, -0.02059178240597248, 0.03522096201777458, -0.022352829575538635, 0.0209939107298851, 0.04584271460771561, -0.015364103950560093, 0.039214517921209335, -0.004721549805253744, -0.012965195812284946, -0.03524869307875633, -0.007480987347662449, -0.0047388831153512, -0.004523952025920153, 0.004523952025920153, -0.013637722469866276, -0.01605742983520031, -0.01392198633402586, -0.01657049171626568, 0.0075503201223909855, 0.0013511190190911293, -0.01975979097187519, -0.023989079520106316, -0.010836685076355934, -0.005505008157342672, 0.012985995970666409, 0.034999094903469086, -0.03125513717532158, -0.017305416986346245, -0.011495345272123814, 0.023864280432462692, 0.042653415352106094, -0.012916662730276585, 0.014545978978276253, -0.021756568923592567, -0.01869206875562668, -0.00714819086715579, -0.017430216073989868, -0.0007830250542610884, 0.004080223385244608, 0.01562756858766079, -0.01608516275882721, 0.037439603358507156, -9.690313891042024e-05, -0.002924102358520031, 0.025417331606149673, 0.001367585500702262, 0.001363252173177898, -0.0054842084646224976, -0.013013728894293308, -0.0379110649228096, 0.004319421015679836, 0.006492997985333204, 0.0105662876740098, 0.023697882890701294, -0.013187060132622719, -0.023184821009635925, 0.005581274162977934, -0.013207860291004181, 0.02285202406346798, -0.013422790914773941, -0.0546618215739727, -0.03189299628138542, -0.004881015047430992, -0.03192072734236717, -0.0045378184877336025, -0.03552602231502533, -0.01616836152970791, 0.024502141401171684, -0.003823692910373211, 0.02563919499516487, 0.000946389976888895, 0.01597423106431961, -0.03616388514637947, -0.016501158475875854, 0.006250333972275257, 0.013443591073155403, 0.002707438077777624, -0.00943616684526205, -0.039380915462970734, 0.0026953048072755337, 0.01602969691157341, 0.02053631655871868, 0.04467792809009552, -0.0032846317626535892, -0.014795576222240925, 0.01573850028216839, 0.016209961846470833, -0.019316062331199646, -0.01146761141717434, -0.0009593898430466652, -0.02583332546055317, -0.01648729294538498, -0.018400872126221657, -0.0068431273102760315, -0.005591674242168665, -0.007786050904542208, 0.02145150676369667, -0.013762520626187325, 0.026027457788586617, -0.011848941445350647, 0.002657171804457903, 0.011980673298239708, 0.005300477147102356, 0.055327415466308594, -0.012687865644693375, 0.027691440656781197, 0.048505086451768875, 0.002426641061902046, -0.015516635961830616, -0.023032288998365402, 0.011148681864142418, 0.000750525388866663, 0.022033900022506714, 0.02792717143893242, -0.0001294569665333256, -0.014698510989546776, 0.029313823208212852, 0.026540519669651985, -0.02405841276049614, -0.016612090170383453, -0.014171583577990532, 0.008950838819146156, -0.005882870871573687, -0.028842361643910408, -0.02494587004184723, 0.007453254424035549, 0.0026086389552801847, -0.0035671622026711702, 0.010649487376213074, -0.014421180821955204, -0.02167337015271187, -0.015516635961830616, 0.011800408363342285, 0.014213182963430882, 0.011855874210596085, 0.008423910476267338, 0.002665838459506631, 0.022408295422792435, -0.002790637081488967, -0.013152394443750381, 0.022172564640641212, -0.0035238293930888176, 0.026166122406721115, -0.01432411465793848, 0.003716227365657687, 0.006829260848462582, 0.015835566446185112, -0.0006595263257622719, -0.004579418338835239, 0.009866029024124146, 0.045066189020872116, -0.01731928251683712, -0.003969291225075722, 0.00042444549035280943, -0.00759885273873806, 0.0505295991897583, -0.015017440542578697, 0.003934625070542097, 0.0003408130432944745, -0.032447654753923416, -0.009727363474667072, 0.024682404473423958, 0.010240424424409866, -0.0005056946538388729, 0.009034037590026855, -0.008229779079556465, -0.01946859434247017, 0.0002040978433797136, 0.010171092115342617, -0.004693816881626844, -0.01786007732152939, 0.009866029024124146, 0.00517567852512002, -0.008666574954986572, 0.005459942389279604, -0.00661779660731554, -0.011530010960996151, -0.004378353711217642, 0.02835703268647194, 0.0014282515039667487, 0.0012332536280155182, -0.0004476285830605775, -0.02752504125237465, -0.029951682314276695, 0.011578544043004513, -0.0035220959689468145, -0.01033749058842659, 0.026845581829547882, -0.031615663319826126, -0.02244989573955536, -0.02416934445500374, 0.009789763018488884, 0.02735864371061325, -0.02185363508760929, 0.026221588253974915, 0.018331538885831833, 0.008132713846862316, 0.0006157601601444185, -0.012888929806649685, 0.00432982062920928, 0.011010016314685345, -0.005498074926435947, -0.0019031798001378775, -0.0013173193437978625, -0.014545978978276253, 0.03477723151445389, -0.01406758464872837, 0.00946389976888895, -0.00611513527110219, -0.04318034276366234, -0.02362854965031147, 0.006707929074764252, 0.0044303531758487225, -0.020522449165582657, -0.008000981993973255, -0.025958124548196793, -0.009387633763253689, -0.010191892273724079, -0.001462917891331017, 0.001115388236939907, -0.0011795208556577563, 0.017152884975075722, 0.01519770547747612, 0.0026311720721423626, 0.008978571742773056, -0.008971638046205044, -0.011134815402328968, -0.02408614568412304, -0.010600954294204712, -0.047562163323163986, -0.021576303988695145, 0.014656911604106426, 0.03369564190506935, 0.022172564640641212, -0.030700474977493286, -0.012625466100871563, -0.018400872126221657, -0.05563247576355934, -0.031532466411590576, -0.010850551538169384, 0.03136606886982918, -0.013408924452960491, 0.01651502586901188, 0.004062890075147152, 0.01648729294538498, -0.0007557253120467067, 0.02755277417600155, 0.007321522571146488, -0.0041391560807824135, 0.009269768372178078, -0.025680795311927795, 0.01995392143726349, -0.006905526854097843, -0.046868838369846344, -0.010892150923609734, 0.006336999591439962, 0.013159327208995819, 0.0015383170684799552, 0.025029068812727928, -0.009068704210221767, -0.01224413700401783, 0.0012358535313978791, 0.00038847920950502157, -0.009762030094861984, 0.005810071714222431, 0.017055818811058998, -0.05241544544696808, 0.0012349869357421994, 0.012362002395093441, -0.002346908440813422, -0.014767843298614025, -0.003983157686889172, -0.009221235290169716, -0.019745923578739166, -0.0032274324912577868, -0.01143987849354744, -0.005057813134044409, 0.011010016314685345, 0.012618533335626125, 0.04467792809009552, 0.004714616574347019, -0.008354578167200089, 0.045149389654397964, 0.022006167098879814, -0.010399890132248402, -0.001309519517235458, 0.01141214556992054, -0.024308009073138237, -0.012167870998382568, 0.02709518000483513, -0.013741721399128437, 0.014448913745582104, -0.007564186584204435, 0.011495345272123814, -0.015988096594810486, -0.023503750562667847, -0.006375132594257593, -0.006132468581199646, -0.02287975698709488, 0.012639332562685013, 0.034167103469371796, -0.02159017138183117, 0.011495345272123814, -0.01616836152970791, -0.01145374495536089, 0.016043564304709435, -0.00042812881292775273, -0.019454726949334145, -0.01602969691157341, -0.0022862425539642572, 0.007224456872791052, 0.009179635904729366, 0.003792493138462305, 0.014961974695324898, 0.01428251527249813, -0.010288957506418228, 0.015766233205795288, 0.25891566276550293, -0.014157717116177082, 0.017041953280568123, -0.002119844313710928, 0.014837176539003849, 0.013741721399128437, 0.03136606886982918, 0.025445064529776573, -0.0006772928172722459, -0.0055258083157241344, -0.009429233148694038, 0.0026675716508179903, 0.015613701194524765, -0.008292178623378277, 0.006097801961004734, -0.036663077771663666, -0.024613073095679283, -0.04035157337784767, -0.005882870871573687, -0.00848631002008915, -0.01935766078531742, -0.014213182963430882, 0.011432945728302002, -0.007245256565511227, -0.006045802496373653, 0.003447563387453556, -0.014684644527733326, -0.016251560300588608, 0.030672742053866386, 0.0033279648050665855, -0.020522449165582657, 0.004305554553866386, 0.0035012962762266397, 0.0168894212692976, -0.008770573884248734, -0.013873453252017498, 0.024044545367360115, 0.0032655654940754175, 0.022602427750825882, -4.067873669555411e-05, 0.007876182906329632, -0.010635620914399624, 0.013720921240746975, -0.050890129059553146, 0.02330962009727955, 0.015447302721440792, -0.027247712016105652, -0.013665455393493176, -0.00571993924677372, 0.011758808977901936, -0.009075636975467205, 0.013609989546239376, 0.025486662983894348, 0.033279646188020706, -0.012077738530933857, 0.00943616684526205, -0.016986487433314323, 0.0076820519752800465, -0.018414737656712532, 0.006357799284160137, -0.052775975316762924, 0.023961346596479416, 0.0029275689739733934, 0.010628687217831612, -0.02167337015271187, 0.0022533093579113483, -0.016099030151963234, -0.023184821009635925, 0.016043564304709435, -0.023503750562667847, -0.0024214410223066807, -0.01568303443491459, -0.005834338255226612, 0.006409798748791218, -0.0506405308842659, -0.05166665092110634, 0.026679184287786484, 0.013505990616977215, 0.013429724611341953, 0.057074595242738724, -0.0022533093579113483, -0.011114015243947506, 0.01408145111054182, -0.01614062860608101, -0.0031303667929023504, -0.03455536812543869, 0.03480496630072594, -0.0013424524804577231, -0.012382802553474903, -0.009588698856532574, -0.02265789359807968, 0.003972758073359728, 0.0014005184639245272, -0.015544368885457516, 0.0067391288466751575, 0.02067498117685318, 0.01989845559000969, -0.004101023077964783, -0.015267038717865944, 0.012341202236711979, -0.026082923635840416, 0.06162281334400177, 0.04046250507235527, 0.0169171541929245, -0.014421180821955204, -0.008909239433705807, -0.008548709563910961, 0.01231346931308508, 0.014643045142292976, -0.009720430709421635, -0.007467120885848999, -0.013395057991147041, 0.010122559033334255, 0.0009853895753622055, 0.014012117870151997, 0.01941312849521637, 0.0148233100771904, 0.0025462396442890167, 0.008638842031359673, -0.020175786688923836, 0.027053579688072205, -0.034111637622117996, -0.002189176855608821, 0.01602969691157341, -0.016015829518437386, -0.030811406672000885, -0.009512432850897312, -0.01815127395093441, -0.01975979097187519, -0.01515610609203577, 0.002034911885857582, -0.041738223284482956, 0.015280905179679394, -0.013096927665174007, -0.007113524712622166, 0.020744314417243004, 0.008146580308675766, 0.011599344201385975, 0.0042882212437689304, 0.008784440346062183, 0.0028738363180309534, 0.008722040802240372, -0.01521157193928957, 0.005272744223475456, 0.017582746222615242, -0.016625957563519478, 0.015267038717865944, 0.001215920434333384, 0.019801389425992966, -0.013110794126987457, -0.03272498771548271, -0.008306045085191727, -0.005116745829582214, 0.00513407913967967, 0.028481831774115562, 0.009165769442915916, -0.02443280816078186, -0.01608516275882721, 0.013658521696925163, -0.00430208770558238, -0.041849154978990555, -0.017596613615751266, 0.03369564190506935, -0.023503750562667847, -0.014961974695324898, -0.004919148050248623, -0.1776023805141449, 0.004534352105110884, 0.005948736798018217, -0.013214793056249619, 0.026776250451803207, 0.02695651538670063, 0.012951329350471497, 0.00824364647269249, 0.012063872069120407, 0.002651971997693181, 0.023753348737955093, -0.005560474470257759, 0.0012965196510776877, 0.0025913058780133724, 0.014490513131022453, 0.01229960285127163, -0.026942647993564606, 0.005158345215022564, 0.022838158532977104, 0.009907628409564495, 0.019912322983145714, -0.014545978978276253, -0.009602565318346024, 0.017707545310258865, 0.0001432151475455612, 0.00153225043322891, -0.020286718383431435, -0.004364487249404192, -0.007959382608532906, -0.027233844622969627, -0.007467120885848999, -0.009020171128213406, 0.009075636975467205, 0.015003574080765247, 0.03380657359957695, 0.003948491532355547, 0.004926081281155348, -0.025888793170452118, -0.02609679102897644, 0.008888439275324345, 0.02142377384006977, 0.026249321177601814, 0.011405212804675102, 0.010032427497208118, -0.011086282320320606, 0.023531483486294746, 0.011744942516088486, -0.02030058577656746, 0.010288957506418228, 0.0009567898814566433, 0.025417331606149673, -0.021784301847219467, 0.02239442989230156, 0.004825548734515905, 0.008507110178470612, 0.038327060639858246, 0.015988096594810486, -0.0013615188654512167, -0.002284509129822254, -0.004551684949547052, 0.009124170057475567, -0.0036711611319333315, -0.0074255215004086494, -0.012382802553474903, 0.005709539633244276, -0.020910711959004402, -0.0336124449968338, 0.013665455393493176, -0.012840397655963898, 0.007397788111120462, -0.00661779660731554, -0.027677573263645172, 0.017360882833600044, -0.011627077125012875, 0.028925560414791107, 0.020439250394701958, -0.02612452395260334, 0.01519770547747612, 0.03031221218407154, -0.005355943460017443, -0.009179635904729366, -0.0030159680172801018, -0.007973249070346355, 0.004336753860116005, -0.011710275895893574, -0.01643182523548603, 0.018858466297388077, -0.009255901910364628, 0.015488903038203716, -0.011280413717031479, 0.002937968820333481, -0.00761271920055151, -0.0032568988390266895, 0.00861804187297821, 0.024100011214613914, 0.01433798111975193, -0.018428605049848557, -0.0052276779897511005, -0.014116116799414158, -0.02056404948234558, 0.0056332736276090145, 0.006676729302853346, -0.012798797339200974, -0.007730584591627121, 0.027746906504034996, 0.012923596426844597, 0.015086773782968521, 0.035942018032073975, 0.050890129059553146, -0.005494608543813229, -0.043513137847185135, 0.003962357994168997, 0.02161790430545807, 0.029369289055466652, 0.0049780807457864285, 0.023143220692873, -0.030007148161530495, -0.0378555990755558, -0.012930530123412609, -0.005258877761662006, 0.019052598625421524, -0.010864417999982834, 0.021770436316728592, 0.03549829125404358, -0.012923596426844597, -0.01334652490913868, -0.05158345401287079, -0.015031307004392147, 0.01949632726609707, 0.0338897742331028, -0.010815885849297047, 0.03696814179420471, -0.005407942924648523, 0.02443280816078186, -0.0031650331802666187, 0.027941036969423294, -0.009782829321920872, -0.029896216467022896, -0.014795576222240925, -0.0009351234184578061, 0.01132894679903984, -0.002267176052555442, 0.007058058399707079, -0.03522096201777458, -0.015350237488746643, 0.026332521811127663, 0.002282775705680251, -0.018456337973475456, 0.016847820952534676, -0.03816066309809685, -0.009047904051840305, -0.014892642386257648, -0.034028440713882446, -0.001543516991659999, 0.01616836152970791, -0.0027871704660356045, 0.03555375710129738, -0.034860432147979736, 0.0094500333070755, -0.038354795426130295, 0.011509211733937263, -0.005047413054853678, -0.027663707733154297, -0.01737475022673607, 0.045066189020872116, -0.013901186175644398, -0.007120457943528891, -0.002086911117658019, 0.01406758464872837, -0.026055190712213516, -0.02373948134481907, 0.002242909511551261, -0.031033271923661232, 0.024266408756375313, -0.01016415935009718, 5.037446680944413e-05, -0.0035532957408577204, 0.011946006678044796, -0.021021643653512, -0.009845228865742683, 0.022297363728284836, -0.005231144372373819, 0.006687129382044077, 0.026734650135040283, -0.0074255215004086494, -0.026249321177601814, -0.0025913058780133724, -0.017901677638292313, -0.021950701251626015, 0.008028714917600155, 0.003574095433577895, 0.02830156683921814, -0.004756216425448656, -0.0167507566511631, 0.013665455393493176, -0.00012490700464695692, -0.008860706351697445, 0.03283591940999031, -0.006520730908960104, 0.013741721399128437, -0.037800133228302, -0.007702851668000221, -0.010455355979502201, -0.006981792859733105, 0.005726872477680445, -0.006614329759031534, 0.002431840868666768, -0.027941036969423294, 0.008888439275324345, 0.00190664641559124, -0.0035706288181245327, 0.016043564304709435, 0.0063231331296265125, -0.014518246054649353, 0.012133204378187656, -0.05546607822179794, 0.004000490996986628, 0.004517018795013428, -0.001111054909415543, -0.020466983318328857, -0.00483248196542263, 0.007300722878426313, -0.0006785928271710873, -0.004042090382426977, -0.010864417999982834, 0.02411387860774994, -0.01417851634323597, -0.036635346710681915, -0.10355517268180847, 0.02333735302090645, 0.0024370409082621336, 0.0056228735484182835, 0.007085791788995266, 0.007446321193128824, 0.0013641188852488995, 0.00016520658391527832, -0.005449542310088873, 0.001329452614299953, -0.005290077533572912, 0.005737272556871176, -0.015072907321155071, -0.012764131650328636, -0.014448913745582104, -0.015877164900302887, 0.011280413717031479, 0.016209961846470833, 0.016625957563519478, -0.006215667352080345, 0.016237694770097733, 0.0071343244053423405, 0.018733669072389603, 0.03685721009969711, 0.011107082478702068, 0.008278312161564827, -0.02862049639225006, 0.03211485967040062, -0.010288957506418228, -0.010600954294204712, -0.011668676510453224, -0.013408924452960491, 0.03616388514637947, -0.0012219870695844293, -0.00825751293450594, -0.01221640408039093, 0.012667065486311913, 0.013956652022898197, 0.02121577598154545, 0.03436123579740524, -0.010191892273724079, -0.021409906446933746, 0.013311859220266342, -0.016931021586060524, -0.02781623788177967, 0.0026311720721423626, -0.0059140706434845924, -0.018276073038578033, 0.0334460474550724, -0.03397297486662865, 0.02859276346862316, 0.010060160420835018, -0.014490513131022453, -0.030811406672000885, -0.002119844313710928, -0.007141257636249065, 0.026221588253974915, 0.013582255691289902, 0.00045152855454944074, 0.007633519358932972, 0.022560827434062958, 0.009124170057475567, 0.03222579136490822, -0.010843618772923946, -0.0014057183871045709, 0.011578544043004513, -0.027968769893050194, -0.025334130972623825, 0.009942295029759407, -0.035082295536994934, -0.008077247999608517, -0.0030853005591779947, 0.011100148782134056, 0.04470565915107727, -0.012126271612942219, -0.0032967650331556797, -0.03133833408355713, -0.0009602564969100058, -0.010372157208621502, 0.032503120601177216, 0.028925560414791107, 0.023503750562667847, -0.02081364579498768, 0.016501158475875854, 0.006427132058888674, 0.04037930443882942, -0.013866519555449486, 0.003385164076462388, -0.020827513188123703, -0.0026485053822398186, -0.012022272683680058, 0.003868758911266923, -0.004974613897502422, -0.003055834211409092, 0.0042916880920529366, 0.02394747920334339, -0.012660132721066475, 0.01780461147427559, 0.02233896404504776, 0.021021643653512, 0.025736261159181595, 0.01605742983520031, 0.006714862305670977, -0.00861804187297821, -0.032558590173721313, -0.008909239433705807, -0.04132222756743431, -0.02319868840277195, 0.009526299312710762, -0.006638596300035715, 0.012687865644693375, -0.002166643738746643, -0.0036711611319333315, 0.02701198123395443, -0.034915897995233536, 0.013866519555449486, 0.0019517126493155956, -0.010760419070720673, -0.027622107416391373, 0.04018517583608627, 0.011114015243947506, 0.020550182089209557, 0.01930219493806362, 0.0011977206449955702, 0.028925560414791107, 0.02247762866318226, 0.012001472525298595, -0.01308999489992857, 0.022546961903572083, -0.011613210663199425, -0.015807831659913063, 0.0034527634270489216, 0.006780728232115507, -0.005657540168613195, -0.013984384946525097, 0.022130966186523438, -0.027677573263645172, 0.009990827180445194, 0.025944259017705917, 0.06389692425727844, 0.002813170198351145, -0.0190664641559124, -0.012167870998382568, 0.004350620787590742, -0.00664552953094244, 0.02859276346862316, 0.023323485627770424, -0.017541147768497467, -0.022172564640641212, 0.014892642386257648, -0.008687375113368034, -0.011377479881048203, -0.009859095327556133, -0.018220607191324234, -0.009567898698151112, 0.00464528426527977, 0.021090976893901825, -0.021368306130170822, 0.027594374492764473, 0.02626318857073784, -0.0040074242278933525, 0.0047042169608175755, -0.002320908708497882, -0.016598224639892578, 0.004506618715822697, 0.019537925720214844, 0.016736889258027077, -0.0006192267755977809, -0.031643398106098175, 0.02365628257393837, 0.016764622181653976, -0.021271241828799248, -0.014025984331965446, -0.0017766478704288602, 0.003868758911266923, -0.005432208999991417, -0.015377970412373543, 0.011259614489972591, 0.007342322263866663, -0.006420198827981949, -0.008666574954986572, -0.02156243845820427, -0.02580559253692627, -0.005546608008444309, -0.010628687217831612, -0.009921494871377945, -0.026485051959753036, -0.02692878060042858]}, "text_id_to_ref_doc_id": {"2d707a34-28f5-4cfa-8973-961c2f7e96e1": "e62f594c-6592-44d1-aeca-eaf9ce7b7d9b"}, "metadata_dict": {"2d707a34-28f5-4cfa-8973-961c2f7e96e1": {"page_label": "1", "file_name": "ChatGPT\u4ea7\u54c1\u56e2\u961f 20230429.pdf", "_node_type": "TextNode", "document_id": "e62f594c-6592-44d1-aeca-eaf9ce7b7d9b", "doc_id": "e62f594c-6592-44d1-aeca-eaf9ce7b7d9b", "ref_doc_id": "e62f594c-6592-44d1-aeca-eaf9ce7b7d9b"}}}
|
joeshi/docstore.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"docstore/metadata": {"e62f594c-6592-44d1-aeca-eaf9ce7b7d9b": {"doc_hash": "24d0fd2f1e041001a18d765f980998dd01a3587f8732ba0c139b4a7d782f5ed8"}, "2d707a34-28f5-4cfa-8973-961c2f7e96e1": {"doc_hash": "046d7c5337fef37cd7e84f404abfe6f83d7751a277b71725847539c993818f2c", "ref_doc_id": "e62f594c-6592-44d1-aeca-eaf9ce7b7d9b"}}, "docstore/data": {"2d707a34-28f5-4cfa-8973-961c2f7e96e1": {"__data__": {"id_": "2d707a34-28f5-4cfa-8973-961c2f7e96e1", "embedding": null, "metadata": {"page_label": "1", "file_name": "ChatGPT\u4ea7\u54c1\u56e2\u961f 20230429.pdf"}, "excluded_embed_metadata_keys": [], "excluded_llm_metadata_keys": [], "relationships": {"1": {"node_id": "e62f594c-6592-44d1-aeca-eaf9ce7b7d9b", "node_type": "4", "metadata": {"page_label": "1", "file_name": "ChatGPT\u4ea7\u54c1\u56e2\u961f 20230429.pdf"}, "hash": "24d0fd2f1e041001a18d765f980998dd01a3587f8732ba0c139b4a7d782f5ed8", "class_name": "RelatedNodeInfo"}}, "hash": "046d7c5337fef37cd7e84f404abfe6f83d7751a277b71725847539c993818f2c", "text": "\u79c1\u6709 ChatGPT\u4ea7\u54c1\u56e2\u961f\u4ecb\u7ecd\n1\u4ea7\u54c1\u53d1\u8d77\u4eba\uff0c\u9996\u5e2d\u5f00\u53d1\u4e13\u5bb6\uff1a\n\u77f3\u4e91\u535a\u58eb\uff0c\u5ba2\u6237\u4e16\u754c\u526f\u603b\u88c1\uff0c \u4e0a\u6d77\u8d22\u7ecf\u5927\u5b66\u91d1\u878d\u79d1\u6280\u7814\u7a76\u9662\u9ad8\u7ea7\u7814\n\u7a76\u5458\uff0c\u590d\u65e6\u5927\u5b66 \u8ba1\u7b97\u673a\u8f6f\u4ef6\u5b66\u9662\u5ba2\u5ea7\u8bb2\u5e08\n\u804c\u8d23\uff1a\n\u2022\u57fa\u4e8e ChatGPT\u5e94\u7528\u7684\u529f\u80fd\u7b97\u6cd5\u8bbe\u8ba1\u4e0e\u4f18\u5316\n\u2022\u4e2d\u6587\u573a\u666f\u5e94\u7528\u8bbe\u8ba1\u4e0e\u5f00\u53d1\n\u2022\u5398\u6e05\u5168\u7403\u53ca\u4e2d\u56fd\u56fd\u5185\u5927\u8bed\u97f3\u6a21\u578b\u6280\u672f\u8def\u7ebf\u56fe\n\u2022\u7ade\u54c1\u6d4b\u8bd5\u4e0e\u53cd\u9988\n\u6838\u5fc3\u5f00\u53d1\u56e2\u961f\u6210\u5458 \uff1a\n\u6881\u8d3a\u541b\u535a\u58eb\uff0c\u4e0a\u6d77\u8d22\u7ecf\u5927\u5b66\u91d1\u878d\u79d1\u6280\u7814\u7a76\u9662\u9ad8\u7ea7\u7814\u7a76 \u5458\nShirley Zhou \uff0c\u5fae\u8f6f\u5927\u4e2d\u534e\u533a\u4e91\u8ba1\u7b97\u5e73\u53f0 Azure\u9ad8\u7ea7\u6280\u672f\u4e3b\u7ba1\n\u5468\u5fd7\u82ac\uff0c\u67d0\u56fd\u5bb6\u7ea7\u4f01\u4e1a\u6570\u636e\u5e93\u5f00\u53d1\u8d1f\u8d23\u4eba\n\u804c\u8d23\uff1a\n\u2022\u57fa\u4e8e ChatGPT\u5e94\u7528\u5f00\u53d1\u4e0e\u5728\u7ebf \u90e8\u7f72\n\u2022\u4e2d\u6587\u573a\u666f\u5e94\u7528\u5f00\u53d1\u4e0e \u6d4b\u8bd5\n\u2022\u4ea7\u54c1\u8fd0\u7ef4\u6307\u5bfc\u6559\u6388 \uff1a\n\u97e9\u666f\u501c\u6559\u6388\uff0c\u535a\u58eb\u751f\u5bfc\u5e08\uff0c\u4e0a\u6d77\u8d22\u7ecf\u5927\u5b66\u4fe1\u606f\u5de5\u7a0b\u4e0e\u7ba1\u7406\u5b66\u9662\u5e38\u52a1\u526f\u9662\n\u957f\uff0c\u4e0a\u6d77\u8d22\u7ecf\u5927\u5b66\u91d1\u878d\u79d1\u6280\u7814\u7a76\u9662\u9662\u957f\n\u804c\u8d23\uff1a\n1.\u51b3\u5b9a\u5927\u8bed\u97f3\u6a21\u578b\u9879\u76ee \u7684\u6218\u7565\u65b9\u5411\u548c\u76ee\u6807\n2.\u786e\u4fdd\u9879\u76ee\u8d44\u6e90\u7684\u6295\u5165\n3.\u76d1\u63a7\u9879\u76ee\u98ce\u9669\u548c\u98ce\u9669\u7ba1\u7406\u7b56\u7565\n\u5e02\u573a\u4e0e\u9500\u552e \uff1a\n\u9648\u753b\uff0c\u67d0 TikTok\u8fd0\u8425\u5e73\u53f0\u8d1f\u8d23\u4eba\nCarrie Cao \uff0cTikTok\u56fd\u9645\u5546\u57ce\u8fd0\u8425\u9ad8\u7ea7\u4e3b\u7ba1\n\u804c\u8d23\uff1a\n1.\u62d3\u5c55\u4e2a\u4eba\u5ba2\u6237\u6e20\u9053\u53ca\u4f01\u4e1a\u7528\u6237\n2.\u7ef4\u7cfb\u5ba2\u6237\u5173\u7cfb\n3.\u6536\u96c6\u7ade\u54c1\u4fe1\u606f", "start_char_idx": 0, "end_char_idx": 496, "text_template": "{metadata_str}\n\n{content}", "metadata_template": "{key}: {value}", "metadata_seperator": "\n", "class_name": "TextNode"}, "__type__": "1"}}, "docstore/ref_doc_info": {"e62f594c-6592-44d1-aeca-eaf9ce7b7d9b": {"node_ids": ["2d707a34-28f5-4cfa-8973-961c2f7e96e1"], "metadata": {"page_label": "1", "file_name": "ChatGPT\u4ea7\u54c1\u56e2\u961f 20230429.pdf"}}}}
|
joeshi/faiss_index/.DS_Store
ADDED
Binary file (6.15 kB). View file
|
|
joeshi/faiss_index/index.faiss
ADDED
Binary file (18.5 kB). View file
|
|
joeshi/faiss_index/index.pkl
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b44e357294efc4964845bcf8fa9e7fa95cdbe72f89157475f04f95376209ed65
|
3 |
+
size 4361
|
joeshi/graph_store.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"graph_dict": {}}
|
joeshi/image__vector_store.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"embedding_dict": {}, "text_id_to_ref_doc_id": {}, "metadata_dict": {}}
|
joeshi/index_store.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"index_store/data": {"bdcb3bbb-bbec-44f5-af7d-52bc9f2b06e9": {"__type__": "vector_store", "__data__": "{\"index_id\": \"bdcb3bbb-bbec-44f5-af7d-52bc9f2b06e9\", \"summary\": null, \"nodes_dict\": {\"2d707a34-28f5-4cfa-8973-961c2f7e96e1\": \"2d707a34-28f5-4cfa-8973-961c2f7e96e1\"}, \"doc_id_dict\": {}, \"embeddings_dict\": {}}"}}}
|
joeshi/joeshi_upload.csv
ADDED
@@ -0,0 +1,201 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
员工,组别,平均通话时长,话务量,平均等待时间,工时利用率
|
2 |
+
1,乙,22.08,60,109.14,75%
|
3 |
+
2,乙,24.59,126,39.53,57%
|
4 |
+
3,乙,12.0,141,45.65,72%
|
5 |
+
4,甲,17.82,504,0.66,105%
|
6 |
+
5,甲,16.95,394,106.61,129%
|
7 |
+
6,丙,23.55,228,12.13,96%
|
8 |
+
7,甲,21.35,53,24.68,37%
|
9 |
+
8,乙,18.89,309,11.65,70%
|
10 |
+
9,甲,22.05,349,17.11,88%
|
11 |
+
10,丙,10.0,45,77.1,54%
|
12 |
+
11,丙,44.27,38,69.13,62%
|
13 |
+
12,乙,9.55,482,103.73,93%
|
14 |
+
13,甲,15.79,269,83.66,76%
|
15 |
+
14,丙,27.89,172,23.67,81%
|
16 |
+
15,甲,32.76,78,41.09,91%
|
17 |
+
16,甲,20.79,290,38.82,91%
|
18 |
+
17,乙,6.6,459,76.78,84%
|
19 |
+
18,乙,70.01,898,3.07,766%
|
20 |
+
19,甲,10.0,814,143.95,128%
|
21 |
+
20,甲,25.91,327,35.12,95%
|
22 |
+
21,丙,14.75,358,11.46,78%
|
23 |
+
22,甲,20.33,245,18.23,89%
|
24 |
+
23,丙,3.2,239,157.58,70%
|
25 |
+
24,甲,16.73,349,10.38,73%
|
26 |
+
25,甲,13.23,143,130.79,87%
|
27 |
+
26,甲,15.32,234,54.51,75%
|
28 |
+
27,乙,16.98,388,26.27,81%
|
29 |
+
28,乙,25.61,276,40.89,82%
|
30 |
+
29,丙,26.68,188,26.67,84%
|
31 |
+
30,丙,28.52,166,46.63,150%
|
32 |
+
31,甲,23.29,274,10.49,81%
|
33 |
+
32,丙,24.29,308,103.67,146%
|
34 |
+
33,甲,20.46,305,60.78,86%
|
35 |
+
34,丙,28.98,184,32.7,78%
|
36 |
+
35,乙,27.07,248,23.51,86%
|
37 |
+
36,甲,20.23,131,139.91,98%
|
38 |
+
37,乙,15.6,370,3.4,78%
|
39 |
+
38,乙,24.58,332,3.18,144%
|
40 |
+
39,丙,14.24,605,2.69,87%
|
41 |
+
40,乙,16.95,316,65.29,98%
|
42 |
+
41,丙,25.83,301,22.51,94%
|
43 |
+
42,丙,15.96,35,68.07,50%
|
44 |
+
43,乙,13.03,33,51.33,44%
|
45 |
+
44,丙,21.02,594,2.7,135%
|
46 |
+
45,甲,28.45,379,102.45,144%
|
47 |
+
46,丙,2.92,188,127.77,79%
|
48 |
+
47,乙,12.95,314,45.43,76%
|
49 |
+
48,甲,5.68,21,77.11,60%
|
50 |
+
49,甲,4.65,10,142.83,85%
|
51 |
+
50,甲,18.18,445,5.6,83%
|
52 |
+
51,丙,21.99,137,66.46,73%
|
53 |
+
52,甲,18.72,340,2.72,95%
|
54 |
+
53,丙,17.53,108,7.4,78%
|
55 |
+
54,丙,16.22,552,3.93,97%
|
56 |
+
55,乙,16.31,40,14.46,17%
|
57 |
+
56,甲,18.53,360,51.85,95%
|
58 |
+
57,乙,12.21,445,2.89,96%
|
59 |
+
58,丙,17.48,508,0.04,103%
|
60 |
+
59,丙,6.68,111,117.19,77%
|
61 |
+
60,丙,16.05,435,3.04,93%
|
62 |
+
61,乙,23.31,456,4.26,96%
|
63 |
+
62,丙,87.77,115,5.35,107%
|
64 |
+
63,甲,30.08,251,0.11,143%
|
65 |
+
64,乙,17.13,322,49.31,87%
|
66 |
+
65,甲,22.55,322,8.95,81%
|
67 |
+
66,乙,29.08,145,23.05,75%
|
68 |
+
67,甲,91.11,130,10.04,123%
|
69 |
+
68,甲,17.7,524,3.74,89%
|
70 |
+
69,甲,24.58,258,19.14,82%
|
71 |
+
70,丙,21.94,313,16.58,88%
|
72 |
+
71,乙,21.9,172,47.2,82%
|
73 |
+
72,甲,6.19,98,95.91,71%
|
74 |
+
73,丙,10.8,173,73.72,74%
|
75 |
+
74,甲,5.04,323,121.82,79%
|
76 |
+
75,甲,17.42,580,2.34,101%
|
77 |
+
76,乙,21.18,205,41.75,80%
|
78 |
+
77,丙,12.34,494,4.77,69%
|
79 |
+
78,丙,18.8,115,1.15,75%
|
80 |
+
79,丙,12.38,36,66.28,49%
|
81 |
+
80,乙,16.85,387,15.63,78%
|
82 |
+
81,乙,18.86,188,6.97,79%
|
83 |
+
82,甲,16.23,263,58.58,70%
|
84 |
+
83,乙,11.3,40,0.0,6%
|
85 |
+
84,丙,18.6,198,41.03,82%
|
86 |
+
85,丙,21.75,224,12.23,69%
|
87 |
+
86,乙,14.09,593,4.66,84%
|
88 |
+
87,甲,12.07,109,83.17,74%
|
89 |
+
88,丙,3.6,58,125.58,77%
|
90 |
+
89,甲,19.7,477,41.59,100%
|
91 |
+
90,丙,19.98,320,13.97,73%
|
92 |
+
91,丙,24.65,283,18.44,81%
|
93 |
+
92,甲,27.69,157,82.63,91%
|
94 |
+
93,甲,22.84,84,153.37,64%
|
95 |
+
94,丙,15.37,30,33.95,57%
|
96 |
+
95,甲,30.75,109,2.53,57%
|
97 |
+
96,乙,18.38,275,55.57,79%
|
98 |
+
97,甲,18.1,373,67.18,110%
|
99 |
+
98,丙,17.52,343,10.9,70%
|
100 |
+
99,丙,19.06,355,94.92,126%
|
101 |
+
100,丙,26.87,312,4.31,83%
|
102 |
+
101,乙,2.71,87,118.05,66%
|
103 |
+
102,乙,18.96,18,0.8,4%
|
104 |
+
103,乙,16.51,574,3.75,101%
|
105 |
+
104,甲,19.04,387,1.04,102%
|
106 |
+
105,甲,19.05,346,28.32,82%
|
107 |
+
106,丙,11.0,478,1.93,86%
|
108 |
+
107,丙,8.11,281,26.4,38%
|
109 |
+
108,乙,24.39,280,34.6,81%
|
110 |
+
109,甲,37.29,211,24.79,87%
|
111 |
+
110,乙,12.34,50,0.04,81%
|
112 |
+
111,乙,22.27,227,58.91,133%
|
113 |
+
112,乙,14.5,374,17.41,66%
|
114 |
+
113,甲,17.82,406,27.11,107%
|
115 |
+
114,乙,13.71,66,5.19,10%
|
116 |
+
115,丙,19.38,37,65.41,56%
|
117 |
+
116,甲,3.13,251,117.86,78%
|
118 |
+
117,丙,15.32,122,98.46,78%
|
119 |
+
118,甲,22.34,232,9.15,67%
|
120 |
+
119,丙,17.42,400,16.82,79%
|
121 |
+
120,甲,18.42,18,1.44,3%
|
122 |
+
121,丙,28.16,197,29.18,76%
|
123 |
+
122,乙,20.67,384,3.54,124%
|
124 |
+
123,乙,7.12,379,100.1,82%
|
125 |
+
124,丙,21.84,361,3.25,73%
|
126 |
+
125,丙,9.97,10,84.59,85%
|
127 |
+
126,丙,20.41,158,79.89,76%
|
128 |
+
127,丙,25.0,246,11.21,67%
|
129 |
+
128,丙,16.55,202,60.2,91%
|
130 |
+
129,乙,18.76,452,1.41,117%
|
131 |
+
130,丙,17.47,314,0.12,110%
|
132 |
+
131,丙,15.63,51,13.85,70%
|
133 |
+
132,甲,62.78,335,65.86,262%
|
134 |
+
133,乙,21.59,223,86.08,116%
|
135 |
+
134,甲,2.25,233,86.8,76%
|
136 |
+
135,甲,19.72,229,86.59,91%
|
137 |
+
136,甲,24.21,324,33.54,105%
|
138 |
+
137,乙,14.0,335,66.23,85%
|
139 |
+
138,丙,19.71,139,28.28,80%
|
140 |
+
139,乙,22.08,195,9.73,65%
|
141 |
+
140,丙,24.19,389,0.36,120%
|
142 |
+
141,甲,3.58,51,134.34,75%
|
143 |
+
142,丙,13.81,499,2.88,96%
|
144 |
+
143,丙,11.34,352,77.42,86%
|
145 |
+
144,丙,23.03,297,61.17,94%
|
146 |
+
145,甲,15.22,595,6.47,84%
|
147 |
+
146,丙,24.76,206,8.7,75%
|
148 |
+
147,甲,2.6,194,144.26,81%
|
149 |
+
148,丙,21.66,326,20.6,84%
|
150 |
+
149,丙,12.08,479,105.67,92%
|
151 |
+
150,甲,17.05,53,59.59,46%
|
152 |
+
151,乙,20.21,423,31.86,98%
|
153 |
+
152,丙,5.07,86,104.74,75%
|
154 |
+
153,乙,15.2,648,3.1,97%
|
155 |
+
154,丙,11.63,55,72.1,46%
|
156 |
+
155,乙,28.5,46,0.68,73%
|
157 |
+
156,乙,17.85,29,1.41,6%
|
158 |
+
157,乙,24.62,293,89.46,138%
|
159 |
+
158,甲,24.96,204,3.55,82%
|
160 |
+
159,甲,10.95,951,2.89,105%
|
161 |
+
160,丙,3.32,426,138.99,80%
|
162 |
+
161,丙,7.45,502,95.35,98%
|
163 |
+
162,甲,14.5,323,71.55,84%
|
164 |
+
163,甲,18.99,301,85.58,116%
|
165 |
+
164,甲,16.49,307,48.6,78%
|
166 |
+
165,乙,15.95,281,44.09,67%
|
167 |
+
166,甲,26.22,439,38.11,110%
|
168 |
+
167,丙,17.18,119,30.92,42%
|
169 |
+
168,丙,11.75,171,91.64,77%
|
170 |
+
169,丙,16.79,374,18.87,80%
|
171 |
+
170,乙,17.49,286,34.68,83%
|
172 |
+
171,甲,20.15,343,15.98,78%
|
173 |
+
172,丙,23.81,262,8.17,79%
|
174 |
+
173,乙,27.47,230,7.05,77%
|
175 |
+
174,乙,17.01,208,47.66,83%
|
176 |
+
175,丙,20.8,390,22.55,114%
|
177 |
+
176,丙,19.15,278,50.95,76%
|
178 |
+
177,甲,16.76,271,56.42,75%
|
179 |
+
178,甲,33.28,228,4.95,76%
|
180 |
+
179,甲,24.57,427,62.71,114%
|
181 |
+
180,丙,19.76,124,96.75,87%
|
182 |
+
181,丙,2.85,36,158.6,83%
|
183 |
+
182,甲,20.81,277,137.43,99%
|
184 |
+
183,乙,16.07,312,111.49,96%
|
185 |
+
184,丙,28.08,679,2.89,227%
|
186 |
+
185,丙,14.78,488,3.89,79%
|
187 |
+
186,丙,10.75,398,70.31,85%
|
188 |
+
187,乙,16.25,385,10.97,84%
|
189 |
+
188,甲,14.9,345,73.56,100%
|
190 |
+
189,甲,27.54,309,37.5,105%
|
191 |
+
190,丙,16.72,139,18.38,83%
|
192 |
+
191,丙,28.82,172,46.19,81%
|
193 |
+
192,乙,13.34,300,68.01,83%
|
194 |
+
193,乙,9.98,401,68.94,82%
|
195 |
+
194,乙,24.61,122,175.58,84%
|
196 |
+
195,乙,4.34,357,103.76,76%
|
197 |
+
196,乙,18.02,223,69.09,83%
|
198 |
+
197,丙,32.28,53,67.35,59%
|
199 |
+
198,甲,24.91,276,79.53,145%
|
200 |
+
199,乙,18.57,201,41.61,78%
|
201 |
+
200,丙,30.17,46,59.18,47%
|
joeshi/message.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
"{\"data\":[{\"id\":\"msg_rAmEZriKWuh4ZOm3nm7mnSfR\",\"assistant_id\":\"asst_iyj1jY6FYX83y8UHKvD9GfGN\",\"content\":[{\"text\":{\"annotations\":[],\"value\":\"\u8ba9\u6211\u5148\u52a0\u8f7d\u6570\u636e\u96c6\uff0c\u7136\u540e\u5bf9\u6570\u636e\u8fdb\u884c\u521d\u6b65\u7684\u63a2\u7d22\u6027\u5206\u6790\u3002\u63a5\u7740\u6211\u5c06\u7ed8\u5236\u4e00\u4e9b\u6709\u8da3\u7684\u56fe\u8868\u6765\u5c55\u793a\u6570\u636e\u7684\u7279\u5f81\u548c\u8d8b\u52bf\u3002\u8ba9\u6211\u4eec\u5f00\u59cb\u5427\uff01\\n\\n\"},\"type\":\"text\"}],\"created_at\":1710314054,\"file_ids\":[],\"metadata\":{},\"object\":\"thread.message\",\"role\":\"assistant\",\"run_id\":\"run_mB2wrEzJBvI5ELOdk6lmezhW\",\"thread_id\":\"thread_cuANpHCi1VpxEnk381pEIh9g\"},{\"id\":\"msg_P8YumXlKMjs8375DUJljf089\",\"assistant_id\":null,\"content\":[{\"text\":{\"annotations\":[],\"value\":\"\u5206\u6790\u6b64\u6570\u636e\u96c6\u5e76\u7ed8\u5236\u4e00\u4e9b'\u6709\u8da3\u7684\u56fe\u8868'\uff0c\"},\"type\":\"text\"}],\"created_at\":1710314052,\"file_ids\":[\"file-Y0XiJYmVItTLPgmjfvK2TU3m\"],\"metadata\":{},\"object\":\"thread.message\",\"role\":\"user\",\"run_id\":null,\"thread_id\":\"thread_cuANpHCi1VpxEnk381pEIh9g\"}],\"object\":\"list\",\"first_id\":\"msg_rAmEZriKWuh4ZOm3nm7mnSfR\",\"last_id\":\"msg_P8YumXlKMjs8375DUJljf089\",\"has_more\":false}"
|
joeshi/vector_store.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
langchain_KB.py
ADDED
@@ -0,0 +1,108 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"""
|
2 |
+
|
3 |
+
"""
|
4 |
+
|
5 |
+
# -*- coding: utf-8 -*-
|
6 |
+
import requests
|
7 |
+
import streamlit as st
|
8 |
+
import openai
|
9 |
+
# from openai import embeddings
|
10 |
+
import os
|
11 |
+
from dotenv import load_dotenv
|
12 |
+
import numpy as np
|
13 |
+
import pandas as pd
|
14 |
+
import csv
|
15 |
+
import tempfile
|
16 |
+
from tempfile import NamedTemporaryFile
|
17 |
+
import pathlib
|
18 |
+
from pathlib import Path
|
19 |
+
import re
|
20 |
+
from re import sub
|
21 |
+
import matplotlib.pyplot as plt
|
22 |
+
from itertools import product
|
23 |
+
from tqdm import tqdm_notebook, tqdm, trange
|
24 |
+
import time
|
25 |
+
from time import sleep
|
26 |
+
# import pretty_errors
|
27 |
+
import seaborn as sns
|
28 |
+
from matplotlib.pyplot import style
|
29 |
+
from rich import print
|
30 |
+
import warnings
|
31 |
+
import PyPDF2
|
32 |
+
from openai import OpenAI
|
33 |
+
client = OpenAI()
|
34 |
+
warnings.filterwarnings('ignore')
|
35 |
+
|
36 |
+
''' 以下加载本地知识的核心内容。'''
|
37 |
+
##! Install package, !pip install "unstructured[all-docs]", 需要完成这一步,否则会报错!
|
38 |
+
# from langchain.document_loaders import UnstructuredFileLoader ## older version.
|
39 |
+
from langchain_community.document_loaders.unstructured import UnstructuredFileLoader ## new version.
|
40 |
+
from langchain_community.document_loaders import PyPDFLoader
|
41 |
+
|
42 |
+
from langchain.text_splitter import CharacterTextSplitter
|
43 |
+
from langchain.embeddings.openai import OpenAIEmbeddings
|
44 |
+
from langchain.embeddings.huggingface import HuggingFaceEmbeddings
|
45 |
+
from langchain.vectorstores import FAISS
|
46 |
+
|
47 |
+
load_dotenv()
|
48 |
+
### 设置openai的API key
|
49 |
+
os.environ["OPENAI_API_KEY"] = os.environ['user_token']
|
50 |
+
openai.api_key = os.environ['user_token']
|
51 |
+
|
52 |
+
# filepath = "/Users/yunshi/Downloads/txt_dir/Sparks_of_AGI.pdf"
|
53 |
+
|
54 |
+
def langchain_localKB_construct(filepath, username):
|
55 |
+
print('开始构建Langchain知识库...')
|
56 |
+
# ''' 以下加载本地知识的核心内容。'''
|
57 |
+
##! Install package, !pip install "unstructured[all-docs]", 需要完成这一步,否则会报错!
|
58 |
+
|
59 |
+
## 加载文件
|
60 |
+
# filepath = "/Users/yunshi/Downloads/txt_dir/Sparks_of_AGI.pdf" ## a sample reference of a local PDF file.
|
61 |
+
# loader = UnstructuredFileLoader(filepath)
|
62 |
+
|
63 |
+
# from pathlib import Path
|
64 |
+
# filepath = Path(filepath)
|
65 |
+
print('now filepath:', filepath.name)
|
66 |
+
# loader = UnstructuredFileLoader(filepath.name) ### original code here.
|
67 |
+
loader = PyPDFLoader(filepath.name) ##NOTE: 只有PyPDFLoader才可以提取PDF的页数page信息。
|
68 |
+
# print('langchain loader:',loader)
|
69 |
+
docs = loader.load()
|
70 |
+
# print('docs now:', docs)
|
71 |
+
|
72 |
+
## 文本分割
|
73 |
+
# text_splitter = CharacterTextSplitter(chunk_size=5000, chunk_overlap=200)
|
74 |
+
docs = CharacterTextSplitter(chunk_size=1000, chunk_overlap=200).split_documents(docs)
|
75 |
+
|
76 |
+
## 创建向量数据库
|
77 |
+
# embedding_model_name = 'GanymedeNil/text2vec-large-chinese'
|
78 |
+
# embeddings = HuggingFaceEmbeddings(model_name=embedding_model_name) ## 这里是联网情况下连接huggingface后使用。
|
79 |
+
from langchain.embeddings.openai import OpenAIEmbeddings
|
80 |
+
embeddings = OpenAIEmbeddings(disallowed_special=()) ## 可能需要更新了。
|
81 |
+
print('langchain embeddings:', embeddings)
|
82 |
+
|
83 |
+
vector_store = FAISS.from_documents(docs, embeddings)
|
84 |
+
# print(vector_store)
|
85 |
+
vector_store.save_local(f'./{username}/faiss_index')
|
86 |
+
|
87 |
+
return vector_store
|
88 |
+
|
89 |
+
|
90 |
+
# vector_store = langchain_KB_construct(filepath='/Users/yunshi/Downloads/txt_dir/Sparks_of_AGI.pdf')
|
91 |
+
# print(vs)
|
92 |
+
|
93 |
+
|
94 |
+
|
95 |
+
|
96 |
+
### 根据prompt来检索本地知识库并回答。
|
97 |
+
def langchain_RAG(prompt, username):
|
98 |
+
### 用langchain看框架接入本地知识库。
|
99 |
+
embeddings = OpenAIEmbeddings(disallowed_special=()) ## load embedding model again here.
|
100 |
+
vector_store = FAISS.load_local(f'./{username}/faiss_index', embeddings)
|
101 |
+
docs = vector_store.similarity_search(prompt, k=5)
|
102 |
+
context = [doc.page_content for doc in docs]
|
103 |
+
total_prompt = f"已知信息:\n{context}\n 根据这些已知信息来回答问题:\n{prompt}"
|
104 |
+
# print('total prompt in local KB version:', total_prompt)
|
105 |
+
|
106 |
+
return total_prompt, docs
|
107 |
+
|
108 |
+
# langchain_RAG('what are main challenges of AGI?')
|
localKB_construct.py
CHANGED
@@ -12,8 +12,8 @@ from llama_index import LLMPredictor
|
|
12 |
from llama_index import ServiceContext
|
13 |
from langchain.chat_models import ChatOpenAI
|
14 |
from langchain import OpenAI
|
15 |
-
from fastapi import FastAPI #* 实现流式数据
|
16 |
-
from fastapi.responses import StreamingResponse #* 实现流式数据
|
17 |
import sys
|
18 |
import os
|
19 |
import math
|
@@ -32,8 +32,8 @@ import os
|
|
32 |
from rich import print
|
33 |
|
34 |
## enironment settings.
|
35 |
-
os.environ["OPENAI_API_KEY"] =
|
36 |
-
openai.api_key =
|
37 |
# file_path = "/Users/yunshi/Downloads/txt_dir/Sparks_of_AGI.pdf"
|
38 |
# file_path = "/Users/yunshi/Downloads/txt_dir/2023年百人会电动论坛 纪要 20230401.pdf"
|
39 |
|
@@ -65,9 +65,7 @@ def construct_index(directory_path):
|
|
65 |
|
66 |
## 好像work了,2023.09.22, 注意这里的写法有调整。
|
67 |
# prompt_helper = PromptHelper(max_input_s≈ize, num_outputs, max_chunk_overlap, chunk_size_limit=chunk_size_limit)
|
68 |
-
|
69 |
-
prompt_helper = PromptHelper(max_input_size=max_input_size, max_chunk_overlap=-1000, num_output=num_outputs, chunk_size_limit=chunk_size_limit)
|
70 |
-
|
71 |
service_context = ServiceContext.from_defaults(llm_predictor=llm_predictor, prompt_helper=prompt_helper)
|
72 |
|
73 |
## 如果是txt文件,那么需要用如下命令。注意与PDF文件的区别。
|
|
|
12 |
from llama_index import ServiceContext
|
13 |
from langchain.chat_models import ChatOpenAI
|
14 |
from langchain import OpenAI
|
15 |
+
# from fastapi import FastAPI #* 实现流式数据
|
16 |
+
# from fastapi.responses import StreamingResponse #* 实现流式数据
|
17 |
import sys
|
18 |
import os
|
19 |
import math
|
|
|
32 |
from rich import print
|
33 |
|
34 |
## enironment settings.
|
35 |
+
os.environ["OPENAI_API_KEY"] = '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'
|
36 |
+
openai.api_key = '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'
|
37 |
# file_path = "/Users/yunshi/Downloads/txt_dir/Sparks_of_AGI.pdf"
|
38 |
# file_path = "/Users/yunshi/Downloads/txt_dir/2023年百人会电动论坛 纪要 20230401.pdf"
|
39 |
|
|
|
65 |
|
66 |
## 好像work了,2023.09.22, 注意这里的写法有调整。
|
67 |
# prompt_helper = PromptHelper(max_input_s≈ize, num_outputs, max_chunk_overlap, chunk_size_limit=chunk_size_limit)
|
68 |
+
prompt_helper = PromptHelper(max_input_size, num_outputs, chunk_overlap_ratio= 0.1, chunk_size_limit=chunk_size_limit)
|
|
|
|
|
69 |
service_context = ServiceContext.from_defaults(llm_predictor=llm_predictor, prompt_helper=prompt_helper)
|
70 |
|
71 |
## 如果是txt文件,那么需要用如下命令。注意与PDF文件的区别。
|
message.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
"{\"data\":[{\"id\":\"msg_kjVmuosPYiXK9xihJTek7G8l\",\"assistant_id\":\"asst_2o85VMcY0Sq1yj243C1GC36K\",\"content\":[{\"text\":{\"annotations\":[],\"value\":\"### 5.1 \u5206\u6790\u7ed3\u679c\\n\u4ece\u5c0f\u63d0\u7434\u56fe\u4e2d\u53ef\u4ee5\u770b\u51fa\uff0c\u8bdd\u52a1\u91cf\u7684\u5206\u5e03\u76f8\u5bf9\u96c6\u4e2d\u5728\u67d0\u4e2a\u8303\u56f4\u5185\uff0c\u540c\u65f6\u4e5f\u5b58\u5728\u4e00\u4e9b\u79bb\u7fa4\u503c\uff08Outlier\uff09\u3002\u5927\u90e8\u5206\u5458\u5de5\u7684\u8bdd\u52a1\u91cf\u5206\u5e03\u5728\u4e2d\u7b49\u6c34\u5e73\uff0c\u4f46\u4e5f\u6709\u4e00\u4e9b\u5458\u5de5\u7684\u8bdd\u52a1\u91cf\u8f83\u9ad8\u6216\u8f83\u4f4e\u3002\\n\\n### 5.2 \u5206\u6790\u539f\u56e0\\n\u53ef\u80fd\u9020\u6210\u8bdd\u52a1\u91cf\u5206\u5e03\u7279\u70b9\u7684\u539f\u56e0\u6709\uff1a\\n- \u5458\u5de5\u4e2a\u4eba\u80fd\u529b\u548c\u5de5\u4f5c\u6548\u7387\u4e0d\u540c\uff0c\u5bfc\u81f4\u8bdd\u52a1\u91cf\u5b58\u5728\u8f83\u5927\u7684\u5dee\u5f02\u3002\\n- \u4e0d\u540c\u7ec4\u522b\u7684\u5de5\u4f5c\u6027\u8d28\u548c\u4efb\u52a1\u5206\u914d\u4e0d\u540c\uff0c\u4e5f\u53ef\u80fd\u5bfc\u81f4\u8bdd\u52a1\u91cf\u7684\u5dee\u5f02\u6027\u3002\"},\"type\":\"text\"}],\"created_at\":1710316665,\"file_ids\":[],\"metadata\":{},\"object\":\"thread.message\",\"role\":\"assistant\",\"run_id\":\"run_04pRX93iX4UHdgS8dIc8niZG\",\"thread_id\":\"thread_uHcXOmtseqIunnLVR5cbPaK2\"},{\"id\":\"msg_otQ5IyeVU91Zr6JAu1IJVdov\",\"assistant_id\":\"asst_2o85VMcY0Sq1yj243C1GC36K\",\"content\":[{\"image_file\":{\"file_id\":\"file-ipbJKn14b9QNPqUGryEnTd4w\"},\"type\":\"image_file\"},{\"text\":{\"annotations\":[],\"value\":\"\u4ee5\u4e0a\u662f\u5173\u4e8e\u8bdd\u52a1\u91cf\u6570\u636e\u7684\u5c0f\u63d0\u7434\u56fe\u3002\u63a5\u4e0b\u6765\u6211\u5c06\u8fdb\u884c\u5206\u6790\u5e76\u63d0\u4f9b\u76f8\u5173\u7ed3\u679c\u3002\"},\"type\":\"text\"}],\"created_at\":1710316664,\"file_ids\":[],\"metadata\":{},\"object\":\"thread.message\",\"role\":\"assistant\",\"run_id\":\"run_04pRX93iX4UHdgS8dIc8niZG\",\"thread_id\":\"thread_uHcXOmtseqIunnLVR5cbPaK2\"},{\"id\":\"msg_8IEuLsMcJGCWSNbPN27yMq5x\",\"assistant_id\":null,\"content\":[{\"text\":{\"annotations\":[],\"value\":\"\u5bf9\u4e8e[\u8bdd\u52a1\u91cf]\u6570\u636e\u7ed9\u6211\u4e00\u4e2a'\u5c0f\u63d0\u7434\u56fe'\uff0c\u5e76\u7ed9\u51fa\u5206\u6790\u7ed3\u679c\u3002\"},\"type\":\"text\"}],\"created_at\":1710316648,\"file_ids\":[\"file-I2slYV5SIs4hGEwB6PGi9xzT\"],\"metadata\":{},\"object\":\"thread.message\",\"role\":\"user\",\"run_id\":null,\"thread_id\":\"thread_uHcXOmtseqIunnLVR5cbPaK2\"}],\"object\":\"list\",\"first_id\":\"msg_kjVmuosPYiXK9xihJTek7G8l\",\"last_id\":\"msg_8IEuLsMcJGCWSNbPN27yMq5x\",\"has_more\":false}"
|
multiQuery_prompt.py
ADDED
@@ -0,0 +1,95 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"""
|
2 |
+
1. 目前这个版本是把所有的multiquery当成一个问题提交给大模型。后续可以考虑将每一个问题分别送入大模型,然后得到的多个答案,然后在一并汇总。
|
3 |
+
|
4 |
+
|
5 |
+
"""
|
6 |
+
|
7 |
+
# # from langchain.document_loaders import UnstructuredFileLoader
|
8 |
+
# print('start the importing of UnstructuredFileLoader')
|
9 |
+
# from langchain_community.document_loaders.unstructured import UnstructuredFileLoader
|
10 |
+
# filepath = "./joeshi_upload.pdf"
|
11 |
+
# loader = UnstructuredFileLoader(filepath)
|
12 |
+
# docs = loader.load()
|
13 |
+
# print('docs now:', docs)
|
14 |
+
# print('langchain loader:',loader)
|
15 |
+
# print('end the importing of UnstructuredFileLoader')
|
16 |
+
from openai import OpenAI
|
17 |
+
import openai
|
18 |
+
import os
|
19 |
+
import rag_source
|
20 |
+
import langchain_KB
|
21 |
+
from langchain.vectorstores import FAISS
|
22 |
+
from langchain.embeddings.openai import OpenAIEmbeddings
|
23 |
+
from langchain_community.document_loaders.unstructured import UnstructuredFileLoader ## new version.
|
24 |
+
from langchain_community.document_loaders import PyPDFLoader
|
25 |
+
from langchain.text_splitter import CharacterTextSplitter
|
26 |
+
from langchain.embeddings.openai import OpenAIEmbeddings
|
27 |
+
from langchain.embeddings.huggingface import HuggingFaceEmbeddings
|
28 |
+
from langchain.vectorstores import FAISS
|
29 |
+
from dotenv import load_dotenv
|
30 |
+
|
31 |
+
client = OpenAI()
|
32 |
+
load_dotenv()
|
33 |
+
### 设置openai的API key
|
34 |
+
os.environ["OPENAI_API_KEY"] = os.environ['user_token']
|
35 |
+
openai.api_key = os.environ['user_token']
|
36 |
+
|
37 |
+
# loader = UnstructuredFileLoader(filepath)
|
38 |
+
# loader = PyPDFLoader(filepath)
|
39 |
+
# print('langchain loader:',loader)
|
40 |
+
# docs = loader.load()
|
41 |
+
# print('docs now:', docs)
|
42 |
+
|
43 |
+
## 文本分割
|
44 |
+
# text_splitter = CharacterTextSplitter(chunk_size=5000, chunk_overlap=200)
|
45 |
+
# docs = CharacterTextSplitter(chunk_size=1000, chunk_overlap=200).split_documents(docs)
|
46 |
+
|
47 |
+
## 创建向量数据库
|
48 |
+
# embedding_model_name = 'GanymedeNil/text2vec-large-chinese'
|
49 |
+
# embeddings = HuggingFaceEmbeddings(model_name=embedding_model_name) ## 这里是联网情况下连接huggingface后使用。
|
50 |
+
# from langchain.embeddings.openai import OpenAIEmbeddings
|
51 |
+
# embeddings = OpenAIEmbeddings(disallowed_special=()) ## 可能需要更新了。
|
52 |
+
# print('langchain embeddings:', embeddings)
|
53 |
+
# vector_store = FAISS.from_documents(docs, embeddings)
|
54 |
+
# print(vector_store)
|
55 |
+
# vector_store.save_local('./joeshi/faiss_index') ## 将矢量库保存在本地指定目录。
|
56 |
+
# vector_store = FAISS.load_local('./joeshi/faiss_index', embeddings)
|
57 |
+
|
58 |
+
|
59 |
+
### 构建一个multi query的方法来进行大模型查询。可以用chatglm的方法来做, 也可以用openai的方法来做。
|
60 |
+
def generate_sim_query(orignal_question):
|
61 |
+
similar_query_prompt = f"""你是一个AI语言模型,你的任务是生成4个不同的用户问题版本,以便从向量数据库中检索相关文档。通过生成用户问题多角度的表述,你的目标是帮助用户克服基于距离的相似性搜索的一些局限性。请按以下要求生成替代问题,并用隔行排列这些:
|
62 |
+
1. 使用与原始问题不同的关键词或同义词来重新表述问题。
|
63 |
+
2. 尝试从不同的角度提问,例如更具体、更抽象或更概括。
|
64 |
+
3. 考虑问题中隐含的前提或假设,并提出针对这些前提或假设的替代问题。
|
65 |
+
4. 结合用户可能拥有的背景知识或先前交互的信息,生成更具针对性的问题。
|
66 |
+
用户的原始问题是: {orignal_question}"""
|
67 |
+
# similar_query_prompt = """You are an AI language model assistant. Your task is to generate five
|
68 |
+
# different versions of the given user question to retrieve relevant documents from a vector
|
69 |
+
# database. By generating multiple perspectives on the user question, your goal is to help
|
70 |
+
# the user overcome some of the limitations of the distance-based similarity search.
|
71 |
+
# Provide these alternative questions separated by newlines.
|
72 |
+
# Original question: {orignal_question}""" ## English version.
|
73 |
+
|
74 |
+
#### 用大模型来生成相似问。
|
75 |
+
# response, history = chatglm.model.chat(chatglm.tokenizer, query=similar_query_prompt) ## 从用langchain的自定义方式来做.
|
76 |
+
response = client.chat.completions.create(
|
77 |
+
model="gpt-3.5-turbo-16k",
|
78 |
+
messages=[{"role": "user", "content": f'{similar_query_prompt}'}],
|
79 |
+
stream=False,
|
80 |
+
)
|
81 |
+
|
82 |
+
## response.choices[0].message.content 才是ChatGPT反馈的返回的str结果。
|
83 |
+
similar_questions = response.choices[0].message.content.split("\n") ## 将反馈结果切割成一个个问题。
|
84 |
+
|
85 |
+
return similar_questions
|
86 |
+
|
87 |
+
|
88 |
+
# #### 根据相似问查询,返回查询后的结果以及信息来源。
|
89 |
+
## 在主程序中会进行大模型的查询,这里应该不需要了。
|
90 |
+
# def multiQuery_prompt(prompt):
|
91 |
+
# prompt = generate_sim_query(prompt)
|
92 |
+
# docs = vector_store.similarity_search(prompt, k=3) ##NOTE:注意可能新版本这里的k是小写的,之前的版本可能是大写。
|
93 |
+
# source = rag_source.rag_source(docs) ## get the K reference source of the RAG answer, in a designed format.
|
94 |
+
|
95 |
+
# return docs, source
|
rag_source.py
ADDED
@@ -0,0 +1,68 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
"""
|
2 |
+
1. 只有使用PyPDFLoader(from langchain_community.document_loaders import PyPDFLoader), 才能获得在metadata中获得page的信息。
|
3 |
+
|
4 |
+
"""
|
5 |
+
|
6 |
+
from langchain_community.vectorstores import FAISS
|
7 |
+
from langchain.embeddings.huggingface import HuggingFaceEmbeddings
|
8 |
+
import streamlit as st
|
9 |
+
import re
|
10 |
+
from langchain.llms.base import LLM
|
11 |
+
from langchain.llms.utils import enforce_stop_tokens
|
12 |
+
from typing import Dict, List, Optional, Tuple, Union
|
13 |
+
import requests
|
14 |
+
import json
|
15 |
+
|
16 |
+
# embeddings = HuggingFaceEmbeddings(model_name='/Users/yunshi/Downloads/360Data/Data Center/Working-On Task/演讲与培训/2023ChatGPT/RAG/bge-large-zh/') ## 切换成BGE的embedding。
|
17 |
+
|
18 |
+
# from langchain.embeddings.openai import OpenAIEmbeddings
|
19 |
+
# embeddings = OpenAIEmbeddings(disallowed_special=()) ## 可能需要更新了。
|
20 |
+
# vector_store = FAISS.load_local("./faiss_index/", embeddings=embeddings) ## 加载vector store到本地。
|
21 |
+
|
22 |
+
## 在绝对路径中提取完整的文件名
|
23 |
+
def extract_document_name(path):
|
24 |
+
# 路径分割
|
25 |
+
path_segments = path.split("/")
|
26 |
+
# 文件名提取
|
27 |
+
document_name = path_segments[-1]
|
28 |
+
return document_name
|
29 |
+
|
30 |
+
## 从一段话中提取 1 句完整的句子,且该句子的长度必须超过 5 个词,同时去除了换行符'\n\n'。
|
31 |
+
import re
|
32 |
+
def extract_sentence(text):
|
33 |
+
"""
|
34 |
+
从一段话中提取 1 句完整的句子,且该句子的长度必须超过 5 个词。
|
35 |
+
|
36 |
+
Args:
|
37 |
+
text: 一段话。
|
38 |
+
|
39 |
+
Returns:
|
40 |
+
提取到的句子。
|
41 |
+
"""
|
42 |
+
|
43 |
+
# 去除换行符。
|
44 |
+
text = text.replace('\n\n', '')
|
45 |
+
# 使用正则表达式匹配句子。
|
46 |
+
sentences = re.split(r'[。?!;]', text)
|
47 |
+
|
48 |
+
# 过滤掉长度小于 5 个词的句子。
|
49 |
+
sentences = [sentence for sentence in sentences if len(sentence.split()) >= 5]
|
50 |
+
|
51 |
+
# 返回第一句句子。
|
52 |
+
return sentences[0] if sentences else None
|
53 |
+
|
54 |
+
### 综合source的输出内容。
|
55 |
+
##NOTE: 只有使用PyPDFLoader(from langchain_community.document_loaders import PyPDFLoader), 才能获得在metadata中获得page的信息。
|
56 |
+
def rag_source(docs):
|
57 |
+
print('starting source function!')
|
58 |
+
print('docs now:', docs)
|
59 |
+
source = ""
|
60 |
+
for i, doc in enumerate(docs):
|
61 |
+
# for i, doc in enumerate(docs): ### original code here.
|
62 |
+
source += f"**【信息来源 {i+1}】** " + extract_document_name(doc.metadata['source']) + ',' + f"第{docs[i].metadata['page']+1}页" + ',部分内容摘录:' + extract_sentence(doc.page_content) + '\n\n'
|
63 |
+
# source += f"**【信息来源 {i+1}】** " + extract_document_name(doc.metadata['source']) + ',' + f"第{docs[i].metadata['page']+1}页" + ',部分内容摘录:' + extract_sentence(doc.page_content) + '\n\n'
|
64 |
+
### original code here.
|
65 |
+
print('source:', source)
|
66 |
+
return source
|
67 |
+
|
68 |
+
|
save_database_info.py
CHANGED
@@ -10,9 +10,8 @@ from tqdm import tqdm_notebook, tqdm, trange
|
|
10 |
import time
|
11 |
import seaborn as sns
|
12 |
from matplotlib.pyplot import style
|
13 |
-
from rich import print
|
14 |
import warnings
|
15 |
-
|
16 |
sns.set()
|
17 |
# style.use('seaborn')
|
18 |
|
|
|
10 |
import time
|
11 |
import seaborn as sns
|
12 |
from matplotlib.pyplot import style
|
|
|
13 |
import warnings
|
14 |
+
from datetime import datetime
|
15 |
sns.set()
|
16 |
# style.use('seaborn')
|
17 |
|
st_msautogen.py
ADDED
@@ -0,0 +1,171 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
##TODO: 1. 多个图片处理。 2. 目前有个bug,在数据模式下,需要点击一下‘清除所有记录‘这个按键,或者重新输入一下内容。
|
2 |
+
|
3 |
+
# -*- coding: utf-8 -*-
|
4 |
+
import requests
|
5 |
+
import streamlit as st
|
6 |
+
import openai
|
7 |
+
import os
|
8 |
+
import datetime
|
9 |
+
from dotenv import load_dotenv
|
10 |
+
import numpy as np
|
11 |
+
import pandas as pd
|
12 |
+
import csv
|
13 |
+
import tempfile
|
14 |
+
from tempfile import NamedTemporaryFile
|
15 |
+
import pathlib
|
16 |
+
from pathlib import Path
|
17 |
+
import re
|
18 |
+
from re import sub
|
19 |
+
import matplotlib.pyplot as plt
|
20 |
+
from itertools import product
|
21 |
+
from tqdm import tqdm_notebook, tqdm, trange
|
22 |
+
import time
|
23 |
+
from time import sleep
|
24 |
+
import seaborn as sns
|
25 |
+
from matplotlib.pyplot import style
|
26 |
+
from rich import print
|
27 |
+
import autogen
|
28 |
+
from autogen import AssistantAgent
|
29 |
+
from autogen import AssistantAgent, UserProxyAgent, config_list_from_json
|
30 |
+
from st_reset_conversation import reset_all, reset_message
|
31 |
+
import asyncio
|
32 |
+
# import file
|
33 |
+
# import pyqt5
|
34 |
+
from PIL import Image # type: ignore
|
35 |
+
import warnings
|
36 |
+
warnings.filterwarnings('ignore')
|
37 |
+
# sns.set()
|
38 |
+
import matplotlib
|
39 |
+
matplotlib.use('Agg') ## https://blog.51cto.com/u_15642578/5301647
|
40 |
+
|
41 |
+
load_dotenv()
|
42 |
+
### 设置openai的API key
|
43 |
+
os.environ["OPENAI_API_KEY"] = os.environ['user_token']
|
44 |
+
openai.api_key = os.environ['user_token']
|
45 |
+
|
46 |
+
# reset_all()
|
47 |
+
|
48 |
+
# st.title('Microsoft AutoGen - Streamlit Demo')
|
49 |
+
|
50 |
+
##NOTE: working,可以在MS AutoGen中的agent的system_message中添加一段提示信息。而且不会在st中显示。
|
51 |
+
time_stamp = str(time.strftime("%Y-%m-%d %H:%M:%S", time.localtime())) + str(np.random.randint(0, 1000000))
|
52 |
+
|
53 |
+
## 以下两个类基本相同,唯一的区别是基类不同。
|
54 |
+
class TrackableAssistentAgent(AssistantAgent):
|
55 |
+
def _process_received_message(self, message, sender, silent):
|
56 |
+
with st.chat_message(sender.name): ##sender.name好像是streamlit的要求。
|
57 |
+
if "execution succeeded" in message:
|
58 |
+
if os.path.exists(f'/Users/yunshi/Downloads/360Data/Data Center/Working-On Task/演讲与培训/2023ChatGPT/Microsoft AutoGen/output_{time_stamp}.png'):
|
59 |
+
st.image(f'/Users/yunshi/Downloads/360Data/Data Center/Working-On Task/演讲与培训/2023ChatGPT/Microsoft AutoGen/output_{time_stamp}.png')
|
60 |
+
st.markdown(message)
|
61 |
+
st.markdown('任务已经完成,请查看屏幕上的显示信息!')
|
62 |
+
else:
|
63 |
+
st.markdown(message)
|
64 |
+
st.markdown('任务已经完成,请查看屏幕上的显示信息!')
|
65 |
+
st.stop()
|
66 |
+
else:
|
67 |
+
st.markdown(message)
|
68 |
+
|
69 |
+
return super()._process_received_message(message, sender, silent)
|
70 |
+
|
71 |
+
## 以下两个类基本相同,唯一的区别是基类不同。
|
72 |
+
class TrackableUserProxyAgent(UserProxyAgent):
|
73 |
+
def _process_received_message(self, message, sender, silent):
|
74 |
+
# with st.status('思考中...', expanded=True, state='running') as status:
|
75 |
+
with st.chat_message(sender.name):
|
76 |
+
if "execution succeeded" in message:
|
77 |
+
if os.path.exists(f'/Users/yunshi/Downloads/360Data/Data Center/Working-On Task/演讲与培训/2023ChatGPT/Microsoft AutoGen/output_{time_stamp}.png'):
|
78 |
+
# st.markdown('有图片!!')
|
79 |
+
st.image(f'/Users/yunshi/Downloads/360Data/Data Center/Working-On Task/演讲与培训/2023ChatGPT/Microsoft AutoGen/output_{time_stamp}.png')
|
80 |
+
st.markdown(message)
|
81 |
+
st.markdown('任务已经完成,请查看屏幕上的显示信息!')
|
82 |
+
else:
|
83 |
+
st.markdown(message)
|
84 |
+
st.markdown('任务已经完成,请查看屏幕上的显示信息!')
|
85 |
+
st.stop()
|
86 |
+
else:
|
87 |
+
st.markdown(message)
|
88 |
+
return super()._process_received_message(message, sender, silent)
|
89 |
+
|
90 |
+
|
91 |
+
|
92 |
+
|
93 |
+
# async def auto_gen(uploaded_file_path): ## async一定需要在主程序中用asyncio.run(st_msautogen.auto_gen(uploaded_file_path))启动。
|
94 |
+
def auto_gen(uploaded_file_path):
|
95 |
+
# st.title('Microsoft AutoGen - Streamlit Demo')
|
96 |
+
|
97 |
+
##TODO 是否需要如下代码?
|
98 |
+
### Initialize chat history
|
99 |
+
if "messages" not in st.session_state:
|
100 |
+
st.session_state.messages = []
|
101 |
+
|
102 |
+
### Display chat messages from history on app rerun
|
103 |
+
for message in st.session_state.messages:
|
104 |
+
with st.chat_message(message["role"]):
|
105 |
+
st.markdown(message["content"])
|
106 |
+
# print('msautogen uploaded_file_path:', uploaded_file_path)
|
107 |
+
print('running time:', datetime.datetime.now().strftime("%H:%M:%S.%f"))
|
108 |
+
|
109 |
+
prompt = st.chat_input("说点什么吧...")
|
110 |
+
# print('prompt:', prompt)
|
111 |
+
|
112 |
+
system_prompt = f"""You are a helpful AI assistant. Solve tasks using your coding and language skills. In the following cases, suggest python code (in a python coding block) or shell script (in a sh coding block) for the user to execute.
|
113 |
+
1. When you need to collect info, use the code to output the info you need, for example, browse or search the web, download/read a file, print the content of a webpage or a file, get the current date/time, check the operating system. After sufficient info is printed and the task is ready to be solved based on your language skill, you can solve the task by yourself.
|
114 |
+
2. When you need to perform some task with code, use the code to perform the task and output the result. Finish the task smartly. Solve the task step by step if you need to. If a plan is not provided, explain your plan first. Be clear which step uses code, and which step uses your language skill.When using code, you must indicate the script type in the code block. The user cannot provide any other feedback or perform any other action beyond executing the code you suggest. The user can't modify your code. So do not suggest incomplete code which requires users to modify. Don't use a code block if it's not intended to be executed by the user. If you want the user to save the code in a file before executing it, put # filename: <filename> inside the code block as the first line. Don't include multiple code blocks in one response. Do not ask users to copy and paste the result. Instead, use 'print' function for the output when relevant. Check the execution result returned by the user.If the result indicates there is an error, fix the error and output the code again. Suggest the full code instead of partial code or code changes. If the error can't be fixed or if the task is not solved even after the code is executed successfully, analyze the problem, revisit your assumption, collect additional info you need, and think of a different approach to try. When you find an answer, verify the answer carefully. Include verifiable evidence in your response if possible.
|
115 |
+
Reply "TERMINATE" in the end when everything is done.
|
116 |
+
Other settings include:
|
117 |
+
* 如果我需要你分析上传的文件,那么文件的位置在:{uploaded_file_path}
|
118 |
+
* 如果要求你输出图表,那么图的解析度dpi需要设定为300。图使用seaborn库。seaborn库的参数设定:'axes.facecolor':'#FFF9ED','figure.facecolor':'#FFF9ED', palette='dark'。
|
119 |
+
* 如果需要你输出图片,那么将图片保持到以下地址:'/Users/yunshi/Downloads/360Data/Data Center/Working-On Task/演讲与培训/2023ChatGPT/Microsoft AutoGen/output_{time_stamp}.png'。
|
120 |
+
* 回答时尽可能地展示分析所对应的图表,并提供分析结果。 你需要按如下格式提供内容:
|
121 |
+
1. 提供详细且专业的分析结果,提供足够的分析依据。
|
122 |
+
2. 给出可能造成这一结果的可能原因有哪些?
|
123 |
+
以上这些内容全部用序列号格式来表达。
|
124 |
+
"""
|
125 |
+
|
126 |
+
if prompt:
|
127 |
+
llm_config = {
|
128 |
+
# "request_timeout": 600,
|
129 |
+
"config_list":[{
|
130 |
+
"model": "gpt-4-1106-preview",
|
131 |
+
# "model": "gpt-4",
|
132 |
+
# "model": "gpt-3.5-turbo-16k",
|
133 |
+
"api_key": os.environ["OPENAI_API_KEY"],
|
134 |
+
}], ## 注意这里的格式,否则会在后期输出会报错
|
135 |
+
}
|
136 |
+
|
137 |
+
## check if the prompt has been passed in.
|
138 |
+
print('prompt after llm_config:', prompt)
|
139 |
+
|
140 |
+
## create an assistant.
|
141 |
+
assistant = TrackableAssistentAgent(name='assistant', llm_config=llm_config, system_message=system_prompt)
|
142 |
+
|
143 |
+
## create a user proxy.
|
144 |
+
### see notes at https://microsoft.github.io/autogen/docs/FAQ#use-the-constructed-configuration-list-in-agents
|
145 |
+
user_proxy = TrackableUserProxyAgent(name='user',human_input_mode='NEVER',llm_config=llm_config)
|
146 |
+
|
147 |
+
## create a loop event
|
148 |
+
loop = asyncio.new_event_loop()
|
149 |
+
# Set the loop as the event loop.
|
150 |
+
asyncio.set_event_loop(loop)
|
151 |
+
|
152 |
+
### define an async function.
|
153 |
+
async def start_chat():
|
154 |
+
try:
|
155 |
+
# print('333')
|
156 |
+
await user_proxy.a_initiate_chat(assistant, message=prompt, silent=True)
|
157 |
+
# user_proxy.a_initiate_chat(assistant, message=prompt, silent=True) ## not working, the function not even start up.
|
158 |
+
print('444')
|
159 |
+
except Exception as e:
|
160 |
+
print('start_chat function not working because:', e)
|
161 |
+
pass
|
162 |
+
|
163 |
+
print('111')
|
164 |
+
## run the async function within the event loop.
|
165 |
+
loop.run_until_complete(start_chat())
|
166 |
+
print('222')
|
167 |
+
|
168 |
+
# # #TODO:尝试停止一个async中的loop。
|
169 |
+
loop.close()
|
170 |
+
|
171 |
+
# auto_gen()
|
st_openai_assistant.py
ADDED
@@ -0,0 +1,291 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
'''
|
2 |
+
使用OpenAI的官方Assistant,实现code_interpreter功能:
|
3 |
+
1. 使用OpenAI官方的API教程:
|
4 |
+
https://cookbook.openai.com/examples/assistants_api_overview_python (重要)
|
5 |
+
https://platform.openai.com/docs/assistants/overview
|
6 |
+
2. YT视频: https://www.youtube.com/watch?v=5rcjGjgJNQc
|
7 |
+
3. 基础代码参考:openai_interpreter.py
|
8 |
+
4. 函数化text_response和image_response和prompt。
|
9 |
+
5. 中文字体显示还是有问题,现在妥协的方式是转换成英文来显示。不影响分析过程和结果。
|
10 |
+
|
11 |
+
##TEMP
|
12 |
+
|
13 |
+
import matplotlib
|
14 |
+
import seaborn as sns
|
15 |
+
SimHei = FontProperties(fname='SimHei.ttf')
|
16 |
+
matplotlib.rcParams['font.sans-serif'] = ['SimHei']
|
17 |
+
sns.set(rc={'axes.facecolor':'#FFF9ED','figure.facecolor':'#FFF9ED'}, palette='dark', font='SimHei')
|
18 |
+
fontproperties = SimHei
|
19 |
+
|
20 |
+
每次输出图表时,你都需要在matplotlib和seaborn中安装中文字体'SimHei',安装的路径在:'/Users/yunshi/Downloads/360Data/Data Center/Working-On Task/演讲与培训/2023ChatGPT/Coding/code_interpreter/SimHei.ttf'。然后,你需要在mattplotlib和seaborn中设置字体为'SimHei'。
|
21 |
+
|
22 |
+
'''
|
23 |
+
##TODO:
|
24 |
+
|
25 |
+
# -*- coding: utf-8 -*-
|
26 |
+
import csv
|
27 |
+
import os
|
28 |
+
import pathlib
|
29 |
+
import re
|
30 |
+
import tempfile
|
31 |
+
import time
|
32 |
+
import warnings
|
33 |
+
from itertools import product
|
34 |
+
from pathlib import Path
|
35 |
+
from re import sub
|
36 |
+
from tempfile import NamedTemporaryFile
|
37 |
+
from time import sleep
|
38 |
+
|
39 |
+
import matplotlib.pyplot as plt
|
40 |
+
import numpy as np
|
41 |
+
import openai
|
42 |
+
import pandas as pd
|
43 |
+
import requests
|
44 |
+
import seaborn as sns
|
45 |
+
from dotenv import load_dotenv
|
46 |
+
from matplotlib.pyplot import style
|
47 |
+
from rich import print
|
48 |
+
from tqdm import tqdm, tqdm_notebook, trange
|
49 |
+
import chatgpt ## 自定义的chatgpt函数,用于优化text response的回答。
|
50 |
+
|
51 |
+
import streamlit as st
|
52 |
+
import add_fonts
|
53 |
+
|
54 |
+
warnings.filterwarnings('ignore')
|
55 |
+
sns.set()
|
56 |
+
|
57 |
+
load_dotenv()
|
58 |
+
### 设置openai的API key
|
59 |
+
os.environ["OPENAI_API_KEY"] = os.environ['user_token']
|
60 |
+
openai.api_key = os.environ['user_token']
|
61 |
+
|
62 |
+
from openai import Client
|
63 |
+
client = Client()
|
64 |
+
|
65 |
+
## JSON output display.
|
66 |
+
import json
|
67 |
+
def show_json(name=None,obj=None):
|
68 |
+
# display(json.loads(obj.model_dump_json())) ## orignial code.
|
69 |
+
print(name,":",json.loads(obj.model_dump_json()))
|
70 |
+
print("--"*40)
|
71 |
+
|
72 |
+
def save_json(obj):
|
73 |
+
json_message = json.dumps(obj.model_dump_json(), indent=4)
|
74 |
+
with open('message.json', 'w') as file:
|
75 |
+
file.write(json_message)
|
76 |
+
|
77 |
+
### set up a function to wrap up the code-interpreter.
|
78 |
+
def openai_assistant(prompt=None, filepath=None, username=None):
|
79 |
+
client = Client()
|
80 |
+
file = client.files.create(
|
81 |
+
file=open(filepath, 'rb'),
|
82 |
+
# file=open('./大模型LLM解决方案调研问卷.pdf', 'rb'),
|
83 |
+
purpose='assistants',
|
84 |
+
)
|
85 |
+
# show_json(file) ## view the file info.
|
86 |
+
|
87 |
+
### create an ansistant.
|
88 |
+
assistant = client.beta.assistants.create(
|
89 |
+
name="AI Expert",
|
90 |
+
# instructions="You are a helpful and powerful AI. When asked a question, you need to answer the question based on the information in the file provided to you.",
|
91 |
+
instructions="""
|
92 |
+
你是一个强大的AI助手。当被问到一个问题时,你需要根据提供给你的文件中的信息来回答这个问题。如果我没有告诉你任何定制化的要求,那么请你按照以下的默认要求来回答:
|
93 |
+
-------------------------------------------------------------------------
|
94 |
+
1. 你需要用我提问的语言来回答。
|
95 |
+
2. 如果要求你输出图表,那么图的解析度dpi需要设定为600。图尽量使用seaborn库。
|
96 |
+
3. 图表上如果有非英文的文字,那么你需要将字体翻译为英文,然后显示。
|
97 |
+
4. 你回答的文字内容必须尽可能的详细且通俗易懂。
|
98 |
+
5. 回答时尽可能地展示分析所对应的图表,并提供分析结果。 你需要按如下格式提供内容:
|
99 |
+
5.1 提供详细且专业的分析结果,提供足够的分析依据。
|
100 |
+
5.2 给出可能造成这一结果的可能原因有哪些?
|
101 |
+
以上内容全部用1, 2, 3这样的序列号格式来表达。
|
102 |
+
""",
|
103 |
+
# tools=[{"type": "retrieval"}],
|
104 |
+
tools=[{"type": "code_interpreter"}],
|
105 |
+
# model="gpt-4-1106-preview",
|
106 |
+
model="gpt-3.5-turbo-1106", ## 注意这个版本以上才有retrieval功能。
|
107 |
+
file_ids=[file.id],
|
108 |
+
# file_ids=["file-6ZaFecYrnYDFXH7WM3HbtQRK"], ## what are the OpenAI assistant API's file_ids look like.
|
109 |
+
)
|
110 |
+
# print('assistant ID:', assistant.id)
|
111 |
+
# show_json(assistant)
|
112 |
+
|
113 |
+
### 这里必须要提供file.id。
|
114 |
+
thread = client.beta.threads.create(
|
115 |
+
messages=[
|
116 |
+
{
|
117 |
+
"role": "user",
|
118 |
+
# "content": "analyze the dataset and plot something interesting.",
|
119 |
+
"content": prompt,
|
120 |
+
# "file_ids": ["file-6ZaFecYrnYDFXH7WM3HbtQRK"],
|
121 |
+
"file_ids": [file.id]
|
122 |
+
}
|
123 |
+
],
|
124 |
+
)
|
125 |
+
run = client.beta.threads.runs.create(
|
126 |
+
thread_id = thread.id,
|
127 |
+
assistant_id = assistant.id,
|
128 |
+
)
|
129 |
+
|
130 |
+
run = client.beta.threads.runs.retrieve(
|
131 |
+
run_id = run.id,
|
132 |
+
thread_id = thread.id,
|
133 |
+
timeout=100,
|
134 |
+
)
|
135 |
+
|
136 |
+
##NOTE: 因为run是异步的,所以这里必须要等待一段时间,直到run完成。否则返回的结果不会有内容,只有user的prompt。
|
137 |
+
import time
|
138 |
+
def wait_on_run(run, thread):
|
139 |
+
while run.status == "queued" or run.status == "in_progress":
|
140 |
+
run = client.beta.threads.runs.retrieve(
|
141 |
+
thread_id=thread.id,
|
142 |
+
run_id=run.id,
|
143 |
+
)
|
144 |
+
time.sleep(0.5)
|
145 |
+
return run
|
146 |
+
run = wait_on_run(run, thread)
|
147 |
+
|
148 |
+
### retrieve the message from OpenAI.
|
149 |
+
messages = client.beta.threads.messages.list(thread_id=thread.id) ##NOTE: 注意这里的返回格式默认是倒序,可以设置成正序。但是最后一个往往是我们要的答案。
|
150 |
+
|
151 |
+
show_json(name='messages:',obj=messages)
|
152 |
+
print('--'*40)
|
153 |
+
save_json(obj=messages)
|
154 |
+
|
155 |
+
## check the messages format. find image and context accordingly.
|
156 |
+
# text_response, image_response = None, None ## single image and text response.
|
157 |
+
|
158 |
+
###NOTE: find all the image files inside the reponsed message.
|
159 |
+
image_response = [] ## multiple images and text responses.
|
160 |
+
imagefile_count = 0
|
161 |
+
imagefile_position = []
|
162 |
+
for i in range(len(messages.data)):
|
163 |
+
for j in range(len(messages.data[i].content)):
|
164 |
+
try:
|
165 |
+
if messages.data[i].content[j].image_file:
|
166 |
+
imagefile_count += 1
|
167 |
+
imagefile_position.append((i,j))
|
168 |
+
except:
|
169 |
+
pass
|
170 |
+
|
171 |
+
print('--'*30)
|
172 |
+
print("总共有几张图片?:", imagefile_count)
|
173 |
+
print('--'*30)
|
174 |
+
|
175 |
+
print('start the image and text repsonse process!')
|
176 |
+
|
177 |
+
image_name = []
|
178 |
+
image_files = []
|
179 |
+
for x, y in imagefile_position:
|
180 |
+
random_num = np.random.randint(10000, 50000)
|
181 |
+
print('x,y=', x, y)
|
182 |
+
try:
|
183 |
+
if messages.data[x].content[y].image_file:
|
184 |
+
### save the image file.
|
185 |
+
image_file = openai.files.content(file_id=messages.data[x].content[y].image_file.file_id) ## message's image file id. note the format.
|
186 |
+
image_files.append(image_file)
|
187 |
+
|
188 |
+
##NOTE: 这里会保存图片在本地目录中,考虑关闭。
|
189 |
+
# with open(f'./{username}/output{x}{y}_{random_num}.png', "wb") as f:
|
190 |
+
# f.write(image_file.content)
|
191 |
+
|
192 |
+
# # Load the PNG file
|
193 |
+
# import matplotlib.image as mpimg
|
194 |
+
# import matplotlib.pyplot as plt
|
195 |
+
# # image_response.append(mpimg.imread(f'./{username}/output{x}{y}.png'))
|
196 |
+
# image_response = mpimg.imread(f'./{username}/output{x}{y}_{random_num}.png')
|
197 |
+
# plt.imshow(image_response)
|
198 |
+
# plt.show()
|
199 |
+
except Exception as e:
|
200 |
+
print(f"An error occurred: {e}")
|
201 |
+
pass
|
202 |
+
|
203 |
+
### consolidate text_response below.
|
204 |
+
text_response_num = 0
|
205 |
+
my_msg = []
|
206 |
+
for x in range(5): #NOTE: 遍历所有的数据点,但是注意要去掉最后的一段。
|
207 |
+
for y in range(5):
|
208 |
+
try:
|
209 |
+
if messages.data[x].content[y].text:
|
210 |
+
print('x, y=', x, y)
|
211 |
+
my_msg.append(messages.data[x].content[y].text.value)
|
212 |
+
text_response_num += 1
|
213 |
+
except:
|
214 |
+
pass
|
215 |
+
final_msg = sorted(my_msg[1:], reverse=True) ## 需要去除最后两段,因为是默认的user的输入,还有就是第一个无用的assistant接受命令的回答。
|
216 |
+
text_response = str()
|
217 |
+
for i in range(len(final_msg)):
|
218 |
+
text_response += final_msg[i]
|
219 |
+
print('final_msg:', final_msg)
|
220 |
+
print('总共有几个text response:', text_response_num)
|
221 |
+
|
222 |
+
## convert a readiable markdown format.
|
223 |
+
# text_response = final_msg[0] #NOTE:输出的格式是list,需要转换成Markdown可以是识别的内容。这里是获得一个str格式的内容,否则一个list对象。
|
224 |
+
# text_response = text_response.replace("['",'"""').replace("']",'"""') ## 这里不需要处理首尾的list符号。
|
225 |
+
|
226 |
+
## 用LLM优化所有的text response的回答。
|
227 |
+
import chatgpt
|
228 |
+
# text_response = '中国在哪里?'
|
229 |
+
user_prompt = f"""首先,我会向你提供一段【文字内容】,这段文字中可能包括了一系列的多轮对话的内容。接着,我需要你根据这段文字中的内容整理成一段文字结论。你的回答风格需要很专业,包括:尽可能的包含统计数据、数字和专业的结论,不能有口语化的表达。【文字内容】如下{text_response}。"""
|
230 |
+
final_answer = chatgpt.chatgpt(user_prompt=user_prompt)
|
231 |
+
# final_answer = chatgpt.chatgpt(user_prompt=user_prompt, openai_model="gpt-3.5-turbo-16k") ### original code here.
|
232 |
+
|
233 |
+
return messages, text_response, image_response, image_files, final_answer
|
234 |
+
|
235 |
+
|
236 |
+
# filepath = '/Users/yunshi/Downloads/360Data/Data Center/Working-On Task/演讲与培训/2023ChatGPT/Coding/code_interpreter/rawdata/iris.csv'
|
237 |
+
# # # # # # # prompt = "analyze the dataset and plot something interesting."
|
238 |
+
|
239 |
+
# prompt = "analyze the dataset and plot something interesting. give me at least 2 plots. please reply in Chinese instead of English."
|
240 |
+
# # # # # # # messages, text_reponse, image_response = openai_assistant(filepath=filepath,username='joeshi') ## working.
|
241 |
+
# messages, text_response, image_response, image_files, final_answer = openai_assistant(prompt, filepath=filepath,username='joeshi') ## probable multiple images and text responses.
|
242 |
+
|
243 |
+
# print("--"*40)
|
244 |
+
# print('final_answer:', final_answer)
|
245 |
+
# print("--"*40)
|
246 |
+
|
247 |
+
|
248 |
+
# # # '''temp work area'''
|
249 |
+
# # # # msg_json = show_json(messages)
|
250 |
+
# # show_json(name='messages:',obj=messages)
|
251 |
+
|
252 |
+
# # # messages.data[0].content[0].image_file
|
253 |
+
# # # messages.data[0].content[2].text.value
|
254 |
+
|
255 |
+
# # # tt = text_response
|
256 |
+
|
257 |
+
# # # import re
|
258 |
+
# # # regex = re.compile(r'(\n\n)|(\n1)')
|
259 |
+
# # # def convert_markdown(text):
|
260 |
+
# # # """
|
261 |
+
# # # 将markdown格式的文字转换成可读性强的文字格式。
|
262 |
+
# # # Args:
|
263 |
+
# # # text: markdown格式的文字。
|
264 |
+
# # # Returns:
|
265 |
+
# # # 可读性强的文字。
|
266 |
+
# # # """
|
267 |
+
# # # result = re.sub(regex, lambda m: m.group(1) if m.group(1) else '\n1.', text)
|
268 |
+
# # # return result
|
269 |
+
# # # print(convert_markdown(str(text_response)))
|
270 |
+
|
271 |
+
# # # messages.data[1].content[1].text.value
|
272 |
+
|
273 |
+
# my_msg = []
|
274 |
+
# for x in range(5):
|
275 |
+
# for y in range(5):
|
276 |
+
# try:
|
277 |
+
# if messages.data[x].content[y].text:
|
278 |
+
# print('x, y=', x, y)
|
279 |
+
# my_msg.append(messages.data[x].content[y].text.value)
|
280 |
+
# # else:
|
281 |
+
# # continue
|
282 |
+
# except:
|
283 |
+
# pass
|
284 |
+
# print(sorted(my_msg[:-1], reverse=True))
|
285 |
+
|
286 |
+
# msg = str()
|
287 |
+
# for i in range(len(final_msg)):
|
288 |
+
# msg += final_msg[i]
|
289 |
+
|
290 |
+
# print(msg)
|
291 |
+
# type(msg)
|
st_reset_conversation.py
ADDED
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
## reset all conversion inside streamlit
|
2 |
+
## clear conversion.
|
3 |
+
import streamlit as st
|
4 |
+
def reset_all():
|
5 |
+
# st.session_state.conversation = None ## 关闭这个可以保持当前的模型选项?
|
6 |
+
st.session_state.chat_history = None
|
7 |
+
st.session_state.messages = []
|
8 |
+
message_placeholder = st.empty()
|
9 |
+
return None
|
10 |
+
|
11 |
+
|
12 |
+
def reset_message():
|
13 |
+
st.session_state.messages = []
|
14 |
+
message_placeholder = st.empty()
|
15 |
+
return None
|