File size: 1,095 Bytes
47d4754
 
 
 
 
8f8b696
47d4754
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
import streamlit as st
from ctransformers import AutoModelForCausalLM

# Load the model
llm = AutoModelForCausalLM.from_pretrained(
    model_path_or_repo_id="my-model/mistral-7b-instruct-v0.2.Q2_K.gguf",
    model_type="mistral",
)

st.title("Conversational Chat with Mistral 🦙🗨️")


# Function to generate response
def generate_response(user_query):
    prompt = f"""The user query is '{user_query}'"""
    args = {
        "prompt": prompt,
        "stream": True,
        "max_new_tokens": 2048,
        "temperature": 0,
    }

    response_placeholder = st.empty()  # Placeholder for displaying response chunks

    response_so_far = ""  # Initialize empty string to store cumulative response

    for chunk in llm(**args):
        response_so_far += chunk  # Append current chunk to cumulative response
        response_placeholder.write(response_so_far)  # Display cumulative response

    return  # No need to return anything


# User input
user_query = st.text_input("Enter your query:", "")

if user_query:
    # Generate and display response
    generate_response(user_query)