File size: 13,301 Bytes
21f9899
 
 
 
7fe89de
21f9899
7fe89de
 
 
 
b02fbd4
 
 
 
9c1093f
21f9899
b02fbd4
21f9899
9c1093f
21f9899
 
 
 
 
 
7fe89de
21f9899
7fe89de
21f9899
 
 
 
9a54855
 
a0054c1
21f9899
645d499
 
b6c88e7
7fe89de
e5f946e
 
 
 
21f9899
7fe89de
 
 
 
9c1093f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a0054c1
9c1093f
 
 
a0054c1
9c1093f
 
 
 
 
 
 
 
a0054c1
9c1093f
a0054c1
9c1093f
a0054c1
b02fbd4
9c1093f
 
645d499
9c1093f
 
 
7fe89de
 
 
 
 
 
 
 
 
9c1093f
 
7fe89de
 
 
 
 
 
 
21f9899
 
 
7fe89de
9c1093f
 
21f9899
7fe89de
 
9c1093f
7fe89de
 
 
 
2933858
21f9899
7fe89de
21f9899
 
 
7fe89de
21f9899
 
 
 
7fe89de
9c1093f
 
 
 
 
 
21f9899
 
 
 
 
 
 
 
9c1093f
7fe89de
 
 
9c1093f
 
 
 
 
 
7fe89de
9c1093f
 
7fe89de
 
 
 
 
 
 
 
 
9c1093f
 
7fe89de
 
 
 
 
9c1093f
7fe89de
9c1093f
7fe89de
 
9c1093f
 
 
 
 
7fe89de
 
9c1093f
 
7fe89de
9c1093f
7fe89de
 
 
9c1093f
 
b02fbd4
 
a0054c1
 
b02fbd4
 
9c1093f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
b82bf8a
6db74e2
 
 
 
 
 
9c1093f
 
 
 
7fe89de
9c1093f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
21f9899
 
 
9c1093f
 
 
 
 
 
 
 
 
 
b02fbd4
9c1093f
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
'''
参考: https://github.com/shroominic/codeinterpreter-api

1. 可以存在本地,然后再调出来。 working. 
1. 可以在临时文件夹中读取文件。
1. 可以直接在内存中读出图片。
1. 中文字体成功。
    from matplotlib.font_manager import FontProperties
    myfont=FontProperties(fname='/Users/yunshi/Downloads/360Data/Data Center/Working-On Task/演讲与培训/2023ChatGPT/Coding/code_interpreter/rawdata/SimHei.ttf')
    sns.set_style('whitegrid',{'font.sans-serif':['simhei','Arial']})      
1. app.py:
    1. openai key, bing key.
    1. proxy
    1. 

'''
# TODO:1. remove the table display in the second round. 2. 

import requests
from codeinterpreterapi import CodeInterpreterSession, File
import streamlit as st
from codeinterpreterapi import CodeInterpreterSession
import openai
import os
import matplotlib.pyplot as plt
import xlrd
import pandas as pd
# import csv
import tempfile
from tempfile import NamedTemporaryFile
import pathlib
from pathlib import Path
from matplotlib.font_manager import FontProperties
import seaborn as sns
from time import sleep

os.environ["OPENAI_API_KEY"] = os.environ['user_token']
openai.api_key = os.environ['user_token']

# #* 如果碰到接口问题,可以启用如下设置。
# openai.proxy = {
#     "http": "http://127.0.0.1:7890",
#             "https": "http://127.0.0.1:7890"
# }

# layout settings.
st.title("个人大语言模型商业智能中心")
st.subheader("Artificial Intelligence Backend Center for Individuals")

# tab1, tab2 = st.tabs(["传统模式", "数据模型"])

# with tab1:
col1, col2 = st.columns(spec=[1, 2])
radio_1 = col1.radio(label='ChatGPT版本', options=[
    'GPT-3.5', 'GPT-4.0'], horizontal=True, label_visibility='visible')
radio_2 = col2.radio(label='模式选择', options=[
    '核心模式', '联网模式', '数据模式'], horizontal=True, label_visibility='visible')

# with tab2:
def upload_file(uploaded_file):
    if uploaded_file is not None:
        filename = uploaded_file.name
        st.write(filename) ## print out the whole file name to validate.
        try:
            if '.csv' in filename:
                csv_file = pd.read_csv(uploaded_file)
                csv_file.to_csv('./upload.csv', encoding='utf-8', index=False)
                st.write(csv_file[:3])  # 这里只是显示文件,后面需要定位文件所在的绝对路径。
            else:
                xls_file = pd.read_excel(uploaded_file)
                xls_file.to_csv('./upload.csv', index=False)
                st.write(xls_file[:3])
        except Exception as e:
            st.write(e)

        uploaded_file_name = "File_provided"
        temp_dir = tempfile.TemporaryDirectory()
        # ! working.
        uploaded_file_path = pathlib.Path(temp_dir.name) / uploaded_file_name
        with open('./upload.csv', 'wb') as output_temporary_file:
            # ! 必须用这种格式读入内容,然后才可以写入temporary文件夹中。
            # output_temporary_file.write(uploaded_file.getvalue())
            output_temporary_file.write(uploaded_file.getvalue())

        return None


bing_search_api_key = os.environ['bing_api_key']
bing_search_endpoint = 'https://api.bing.microsoft.com/v7.0/search'


def search(query):
    # Construct a request
    # mkt = 'en-EN'
    mkt = 'zh-CN'
    params = {'q': query, 'mkt': mkt}
    headers = {'Ocp-Apim-Subscription-Key': bing_search_api_key}

    # Call the API
    try:
        response = requests.get(bing_search_endpoint,
                                headers=headers, params=params)
        response.raise_for_status()
        json = response.json()
        return json["webPages"]["value"]
        # print("\nJSON Response:\n")
        # pprint(response.json())
    except Exception as e:
        raise e

# openai.api_key = st.secrets["OPENAI_API_KEY"]

# async def main():
async def text_mode():
    # Set a default model
    if "openai_model" not in st.session_state:
        st.session_state["openai_model"] = "gpt-3.5-turbo-16k"
    if radio_1 == 'GPT-3.5':
        # print('----------'*5)
        print('radio_1: GPT-3.5 starts!')
        st.session_state["openai_model"] = "gpt-3.5-turbo-16k"
    else:
        print('radio_1: GPT-4.0 starts!')
        st.session_state["openai_model"] = "gpt-4"

    # Initialize chat history
    if "messages" not in st.session_state:
        st.session_state.messages = []

    # Display chat messages from history on app rerun
    for message in st.session_state.messages:
        with st.chat_message(message["role"]):
            st.markdown(message["content"])

    # Display assistant response in chat message container
    # if prompt := st.chat_input("Say something"):
    prompt = st.chat_input("Say something")
    print('prompt now:', prompt)
    print('----------'*5)
    # if prompt:
    if prompt:
        st.session_state.messages.append({"role": "user", "content": prompt})
        with st.chat_message("user"):
            st.markdown(prompt)

        with st.chat_message("assistant"):
            message_placeholder = st.empty()
            full_response = ""

            if radio_2 == '联网模式':
                # print('联网模式入口,prompt:', prompt)
                input_message = prompt
                internet_search_result = search(input_message)
                search_prompt = [
                    f"Source:\nTitle: {result['name']}\nURL: {result['url']}\nContent: {result['snippet']}" for result in internet_search_result]
                prompt = "基于如下的互联网公开信息, 回答问题:\n\n" + \
                    "\n\n".join(search_prompt[:3]) + "\n\n问题: " + input_message + \
                    "你需要注意的是回答问题时必须用提问的语言(如英文或者中文)来提示:'答案基于互联网公开信息。'" + "\n\n答案: "  # 限制了只有3个搜索结果。
                # prompt = "Use these sources to answer the question:\n\n" + "\n\n".join(search_prompt[0:3]) + "\n\nQuestion: " + input_message + "(注意:回答问题时请提示'以下答案基于互联网公开信息。')\n\n" + "\n\nAnswer: "

                st.session_state.messages.append(
                    {"role": "user", "content": prompt})

                for response in openai.ChatCompletion.create(
                    model=st.session_state["openai_model"],
                    messages=[
                        {"role": m["role"], "content": m["content"]}
                        for m in st.session_state.messages
                    ],
                    stream=True,
                ):
                    full_response += response.choices[0].delta.get(
                        "content", "")
                    message_placeholder.markdown(full_response + "▌")
                message_placeholder.markdown(full_response)
                st.session_state.messages.append(
                    {"role": "assistant", "content": full_response})

            if radio_2 == '核心模式':
                print('GPT only starts!!!')
                print('messages:', st.session_state['messages'])
                for response in openai.ChatCompletion.create(
                    model=st.session_state["openai_model"],
                    # messages=[
                    #     {"role": m["role"], "content": m["content"]}
                    #     for m in st.session_state.messages
                    # ],
                    messages=[{'role': 'system', 'content': 'you are ChatGPT'}, {'role': 'user', 'content': prompt}],
                    stream=True,
                ):
                    full_response += response.choices[0].delta.get(
                        "content", "")
                    message_placeholder.markdown(full_response + "▌")
                    print('session completed!')
                message_placeholder.markdown(full_response)
            st.session_state.messages.append(
                {"role": "assistant", "content": full_response})


async def data_mode():
    # print('数据分析模式启动!')
    uploaded_file_path = './upload.csv'
    # st.write(f"passed file path in data_mode: {uploaded_file_path}")
    # tmp1 = pd.read_csv('./upload.csv')
    # st.write(tmp1[:5])

    # Initialize chat history
    if "messages" not in st.session_state:
        st.session_state.messages = []

    # Display chat messages from history on app rerun
    for message in st.session_state.messages:
        with st.chat_message(message["role"]):
            st.markdown(message["content"])

    # Display assistant response in chat message container
    # if prompt := st.chat_input("Say something"):
    prompt = st.chat_input("Say something")
    print('prompt now:', prompt)
    print('----------'*5)
    # if prompt:
    if prompt:
        st.session_state.messages.append({"role": "user", "content": prompt})
        with st.chat_message("user"):
            st.markdown(prompt)

        with st.chat_message("assistant"):
            async with CodeInterpreterSession() as session:
                # user_request = "对于文件中的'SepalLengthCm’数据给我一个'直方图',提供图表,并给出分析结果"
                #! 可以用设定dpi=300来输出高质量的图表。(注:图的解析度dpi设定为300)
                environ_settings = """【背景要求】如果我没有告诉你任何定制化的要求,那么请你按照以下的默认要求来回答:
                -------------------------------------------------------------------------
                1. 你需要用提问的语言来回答(如:中文提问你就用中文来回答,英文提问你就用英文来回答)。
                2. 如果要求你输出图表,那么图的解析度dpi需要设定为600。图尽量使用seaborn库。seaborn库的参数设定:sns.set(rc={'axes.facecolor':'#FFF9ED','figure.facecolor':'#FFF9ED'}, palette='dark'。
                3. 在图上显示中文时,需要按如下要求操作:
                    myfont=FontProperties(fname='./SimHei.ttf')
                    plt.rcParams['font.sans-serif'] = ['myfont']
                    plt.rcParams['font.family']='sans-serif'
                    sns.set_style({'font.sans-serif':['myfont']})

                -------------------------------------------------------------------------
                """  # seaborn中的palette参数可以设定图表的颜色,选项包括:deep, muted, pastel, bright, dark, colorblind,Spectral。更多参数可以参考:https://seaborn.pydata.org/generated/seaborn.color_palette.html。

                # uploaded_file_path = upload_file()
                
                user_request = environ_settings + "\n\n" + \
                    "你需要完成以下任务:\n\n" + prompt + "\n\n" \
                    f"注:文件位置在{uploaded_file_path}"
                print('user_request: \n', user_request)

                # 加载上传的文件,主要路径在上面代码中。
                files = [File.from_path(str(uploaded_file_path))]

                with st.status('thinking...', expanded=True, state='running') as status:
                    # generate the response
                    response = await session.generate_response(
                        user_request, files=files
                    )

                    # output to the user
                    print("AI: ", response.content)
                    full_response = response.content
                    ### full_response = "this is full response"

                    # for file in response.files:
                    for i, file in enumerate(response.files):
                        # await file.asave(f"/Users/yunshi/Downloads/360Data/Data Center/Working-On Task/演讲与培训/2023ChatGPT/Coding/code_interpreter/output{i}.png") ##working.
                        # st.image(file.get_image()) #! working.
                        # * 注意这里的设定,可以提高图片的精细程度。
                        st.image(file.get_image(), width=None,
                                    output_format='PNG')

                    # message_placeholder.markdown(full_response + "▌") ## orignal code.
                # message_placeholder.markdown(full_response) ## orignal code.
                st.write(full_response)
                status.update(label='complete', state='complete')
                # st.session_state.messages.append(
                #     {"role": "assistant", "content": full_response})

                await session.astop()  # ! 确认需要关闭。
        st.session_state.messages.append({"role": "assistant", "content": full_response})

if __name__ == "__main__":
    import asyncio
    if radio_2 != "数据模式":
        # * 也可以用命令执行这个python文件。’streamlit run frontend/app.py‘
        asyncio.run(text_mode())
        # upload_file()
    elif radio_2 == "数据模式":
        uploaded_file = st.file_uploader(
        "选择一个文件", type=(["csv", "xlsx", "xls"]))
        ## 默认状态下没有上传文件,None,会报错。需要判断。
        if uploaded_file is not None:
            uploaded_file_path = upload_file(uploaded_file)
            asyncio.run(data_mode())