Spaces:
Runtime error
Runtime error
import streamlit as st | |
from transformers import pipeline | |
from transformers import AutoConfig, pipeline | |
from langchain_community.llms import Ollama | |
import time | |
st.title('*ChatBot clone*') | |
llm = Ollama(model='llama3:latest') | |
def response_generator(prompt): | |
response = llm.invoke(prompt, stop=['<|eot_id|>']) | |
for word in response.split(): | |
yield word + " " | |
time.sleep(0.05) | |
# init chat history | |
if "messages" not in st.session_state: | |
st.session_state.messages = [] | |
# display chat history | |
for message in st.session_state.messages: | |
with st.chat_message(message['role']): | |
st.markdown(message['content']) | |
# accept user input | |
if prompt := st.chat_input("What is up?"): | |
# add user message to user history | |
st.session_state.messages.append({'role':'user','content':prompt}) | |
# display user message | |
with st.chat_message('user'): | |
st.markdown(prompt) | |
# display assistant response | |
with st.chat_message('assistant'): | |
ans = llm.invoke(prompt, stop=['<|eot_id|>']) | |
respose = st.write_stream(response_generator(prompt)) | |
st.session_state.messages.append({'role':'assistant', 'content':respose}) | |