import streamlit as st import torch from transformers import pipeline def main(): # Set up the page st.set_page_config(page_title="Nudge Generator Demo - tinyllama 1b", page_icon="orYx logo.png") # Title and logo col1, col2 = st.columns([3, 1]) with col1: st.title("Pirate Chatbot") with col2: st.image("orYx logo.png", use_column_width=True) # Chat interface st.markdown("---") st.header("Chat Interface") # Input for user-provided message user_message = st.text_area("Enter your message:") if st.button("Generate Response"): if user_message.strip(): with st.spinner("Generating response..."): # Load the pipeline pipe = pipeline("text-generation", model="TinyLlama/TinyLlama-1.1B-Chat-v1.0", torch_dtype=torch.bfloat16, device_map="auto") # Define the message structure messages = [ { "role": "system", "content": "You are a personal corporate trainer that teaches me how to perform well in my organization.", }, {"role": "user", "content": user_message}, ] # Generate the prompt prompt = pipe.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) # Generate the response outputs = pipe(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) # Display the response st.text_area("Chatbot Response:", outputs[0]["generated_text"], height=200) else: st.warning("Please enter a message to get a response.") if __name__ == "__main__": main()