Spaces:
Runtime error
Runtime error
Fix formatting and response template issues
Browse files
app.py
CHANGED
@@ -3,31 +3,30 @@ import replicate
|
|
3 |
import os
|
4 |
|
5 |
# App title
|
6 |
-
st.set_page_config(page_title="
|
7 |
|
8 |
# Replicate Credentials
|
9 |
with st.sidebar:
|
10 |
-
st.title('
|
11 |
if 'REPLICATE_API_TOKEN' in st.secrets:
|
12 |
-
st.success('API token loaded!', icon='✅')
|
13 |
replicate_api = st.secrets['REPLICATE_API_TOKEN']
|
14 |
else:
|
15 |
replicate_api = st.text_input('Enter Replicate API token:', type='password')
|
16 |
if not (replicate_api.startswith('r8_') and len(replicate_api)==40):
|
17 |
st.warning('Please enter your Replicate API token.', icon='⚠️')
|
18 |
st.markdown("**Don't have an API token?** Head over to [Replicate](https://replicate.com) to sign up for one.")
|
19 |
-
else:
|
20 |
-
|
21 |
|
22 |
os.environ['REPLICATE_API_TOKEN'] = replicate_api
|
23 |
st.subheader("Adjust model parameters")
|
24 |
temperature = st.sidebar.slider('temperature', min_value=0.01, max_value=5.0, value=0.6, step=0.01)
|
25 |
top_p = st.sidebar.slider('top_p', min_value=0.01, max_value=1.0, value=0.9, step=0.01)
|
26 |
-
max_length = st.sidebar.slider('max_length', min_value=32, max_value=128, value=120, step=8)
|
27 |
|
28 |
# Store LLM-generated responses
|
29 |
if "messages" not in st.session_state.keys():
|
30 |
-
st.session_state.messages = [{"role": "assistant", "content": "Ask me anything"}]
|
31 |
|
32 |
# Display or clear chat messages
|
33 |
for message in st.session_state.messages:
|
@@ -35,32 +34,29 @@ for message in st.session_state.messages:
|
|
35 |
st.write(message["content"])
|
36 |
|
37 |
def clear_chat_history():
|
38 |
-
st.session_state.messages = [{"role": "assistant", "content": "Ask me anything"}]
|
39 |
st.sidebar.button('Clear chat history', on_click=clear_chat_history)
|
40 |
|
|
|
|
|
41 |
# Function for generating Snowflake Arctic response
|
42 |
-
def generate_arctic_response(
|
43 |
-
|
44 |
for dict_message in st.session_state.messages:
|
45 |
if dict_message["role"] == "user":
|
46 |
-
|
47 |
else:
|
48 |
-
|
|
|
|
|
|
|
49 |
|
50 |
for event in replicate.stream("snowflake/snowflake-arctic-instruct",
|
51 |
-
input={"prompt":
|
52 |
-
<|im_start|>system
|
53 |
-
You're a helpful assistant<|im_end|>
|
54 |
-
<|im_start|>user
|
55 |
-
{string_dialogue}
|
56 |
-
<|im_end|>
|
57 |
-
<|im_start|>assistant
|
58 |
-
""",
|
59 |
"prompt_template": r"{prompt}",
|
60 |
"temperature": temperature,
|
61 |
"top_p": top_p,
|
62 |
-
|
63 |
-
"repetition_penalty":1}):
|
64 |
yield str(event)
|
65 |
|
66 |
# User-provided prompt
|
@@ -72,8 +68,7 @@ if prompt := st.chat_input(disabled=not replicate_api):
|
|
72 |
# Generate a new response if last message is not from assistant
|
73 |
if st.session_state.messages[-1]["role"] != "assistant":
|
74 |
with st.chat_message("assistant"):
|
75 |
-
|
76 |
-
|
77 |
-
full_response = st.write_stream(response)
|
78 |
message = {"role": "assistant", "content": full_response}
|
79 |
st.session_state.messages.append(message)
|
|
|
3 |
import os
|
4 |
|
5 |
# App title
|
6 |
+
st.set_page_config(page_title="Snowflake Arctic")
|
7 |
|
8 |
# Replicate Credentials
|
9 |
with st.sidebar:
|
10 |
+
st.title('Snowflake Arctic')
|
11 |
if 'REPLICATE_API_TOKEN' in st.secrets:
|
12 |
+
#st.success('API token loaded!', icon='✅')
|
13 |
replicate_api = st.secrets['REPLICATE_API_TOKEN']
|
14 |
else:
|
15 |
replicate_api = st.text_input('Enter Replicate API token:', type='password')
|
16 |
if not (replicate_api.startswith('r8_') and len(replicate_api)==40):
|
17 |
st.warning('Please enter your Replicate API token.', icon='⚠️')
|
18 |
st.markdown("**Don't have an API token?** Head over to [Replicate](https://replicate.com) to sign up for one.")
|
19 |
+
#else:
|
20 |
+
# st.success('API token loaded!', icon='✅')
|
21 |
|
22 |
os.environ['REPLICATE_API_TOKEN'] = replicate_api
|
23 |
st.subheader("Adjust model parameters")
|
24 |
temperature = st.sidebar.slider('temperature', min_value=0.01, max_value=5.0, value=0.6, step=0.01)
|
25 |
top_p = st.sidebar.slider('top_p', min_value=0.01, max_value=1.0, value=0.9, step=0.01)
|
|
|
26 |
|
27 |
# Store LLM-generated responses
|
28 |
if "messages" not in st.session_state.keys():
|
29 |
+
st.session_state.messages = [{"role": "assistant", "content": "Hi. I'm Arctic, a new, efficient, intelligent, and truly open language model created by Snowflake AI Research. Ask me anything."}]
|
30 |
|
31 |
# Display or clear chat messages
|
32 |
for message in st.session_state.messages:
|
|
|
34 |
st.write(message["content"])
|
35 |
|
36 |
def clear_chat_history():
|
37 |
+
st.session_state.messages = [{"role": "assistant", "content": "Hi. I'm Arctic, a new, efficient, intelligent, and truly open language model created by Snowflake AI Research. Ask me anything."}]
|
38 |
st.sidebar.button('Clear chat history', on_click=clear_chat_history)
|
39 |
|
40 |
+
st.sidebar.caption('Built by [Snowflake](https://snowflake.com/) to demonstrate [Snowflake Arctic](https://www.snowflake.com/blog/arctic-open-and-efficient-foundation-language-models-snowflake). App hosted on [Streamlit Community Cloud](https://streamlit.io/cloud). Model hosted by [Replicate](https://replicate.com/snowflake/snowflake-arctic-instruct).')
|
41 |
+
|
42 |
# Function for generating Snowflake Arctic response
|
43 |
+
def generate_arctic_response():
|
44 |
+
prompt = []
|
45 |
for dict_message in st.session_state.messages:
|
46 |
if dict_message["role"] == "user":
|
47 |
+
prompt.append("<|im_start|>user\n" + dict_message["content"] + "<|im_end|>")
|
48 |
else:
|
49 |
+
prompt.append("<|im_start|>assistant\n" + dict_message["content"] + "<|im_end|>")
|
50 |
+
|
51 |
+
prompt.append("<|im_start|>assistant")
|
52 |
+
prompt.append("")
|
53 |
|
54 |
for event in replicate.stream("snowflake/snowflake-arctic-instruct",
|
55 |
+
input={"prompt": "\n".join(prompt),
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
56 |
"prompt_template": r"{prompt}",
|
57 |
"temperature": temperature,
|
58 |
"top_p": top_p,
|
59 |
+
}):
|
|
|
60 |
yield str(event)
|
61 |
|
62 |
# User-provided prompt
|
|
|
68 |
# Generate a new response if last message is not from assistant
|
69 |
if st.session_state.messages[-1]["role"] != "assistant":
|
70 |
with st.chat_message("assistant"):
|
71 |
+
response = generate_arctic_response()
|
72 |
+
full_response = st.write_stream(response)
|
|
|
73 |
message = {"role": "assistant", "content": full_response}
|
74 |
st.session_state.messages.append(message)
|