Spaces:
Runtime error
Runtime error
File size: 4,031 Bytes
07ff0bc 793769c e4a5ec9 dc79961 1ad8ba9 0766a90 dc79961 f7670c1 fe58fe8 dc79961 f7670c1 dc79961 f7670c1 fe58fe8 f7670c1 ddda3d2 e377924 f7670c1 fe58fe8 e4a5ec9 0766a90 fe58fe8 42039d6 1ad8ba9 42039d6 1ad8ba9 42039d6 1ad8ba9 42039d6 1ad8ba9 42039d6 1ad8ba9 42039d6 1ad8ba9 42039d6 1ad8ba9 42039d6 1ad8ba9 42039d6 1ad8ba9 42039d6 1ad8ba9 42039d6 1ad8ba9 42039d6 1ad8ba9 42039d6 1ad8ba9 42039d6 1ad8ba9 42039d6 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 |
import streamlit as st
from streamlit_chat import message
from langchain.text_splitter import RecursiveCharacterTextSplitter
from langchain.embeddings import HuggingFaceEmbeddings
from langchain.vectorstores import FAISS
from langchain.llms import CTransformers
from langchain.memory import ConversationBufferMemory
from langchain.chains import ConversationalRetrievalChain
import sys
import tempfile
# Initialize the CSVLoader to load the uploaded CSV file
from langchain.document_loaders.csv_loader import CSVLoader
DB_FAISS_PATH = 'vectorstore/db_faiss'
from transformers import pipeline
pipe = pipeline("text-generation",model="mistralai/Mistral-7B-v0.1",model_type="llama",max_new_tokens=512,temperature=0.1 )
# Display the title of the web page
st.title("Chat with CSV using open source LLM Inference Point π¦π¦")
# Display a markdown message with additional information
st.markdown("<h3 style='text-align: center; color: white;'>Built by <a href='https://github.com/AIAnytime'>AI Anytime with β€οΈ </a></h3>", unsafe_allow_html=True)
# Allow users to upload a CSV file
uploaded_file = st.sidebar.file_uploader("Upload your Data", type="csv")
if uploaded_file:
# Initialize the CSVLoader to load the uploaded CSV file
with tempfile.NamedTemporaryFile(delete=False) as tmp_file:
tmp_file.write(uploaded_file.getvalue())
tmp_file_path = tmp_file.name
# Initialize the CSVLoader to load the uploaded CSV file
loader = CSVLoader(file_path=tmp_file_path, encoding="utf-8", csv_args={'delimiter': ','})
data = loader.load()
embeddings = HuggingFaceEmbeddings(model_name='sentence-transformers/all-MiniLM-L6-v2',model_kwargs={'device': 'cpu'})
db = FAISS.from_documents(data, embeddings)
db.save_local(DB_FAISS_PATH)
llm = load_llm()
chain = ConversationalRetrievalChain.from_llm(llm=llm, retriever=db.as_retriever())
def conversational_chat(query):
# Maintain and display the chat history
result = chain({"question": query, "chat_history": st.session_state['history']})
# Maintain and display the chat history
st.session_state['history'].append((query, result["answer"]))
return result["answer"]
# Maintain and display the chat history
if 'history' not in st.session_state:
# Maintain and display the chat history
st.session_state['history'] = []
# Maintain and display the chat history
if 'generated' not in st.session_state:
# Maintain and display the chat history
st.session_state['generated'] = ["Hello ! Ask me anything about " + uploaded_file.name + " π€"]
# Maintain and display the chat history
if 'past' not in st.session_state:
# Maintain and display the chat history
st.session_state['past'] = ["Hey ! π"]
#container for the chat history
response_container = st.container()
#container for the user's text input
container = st.container()
with container:
with st.form(key='my_form', clear_on_submit=True):
user_input = st.text_input("Query:", placeholder="Talk to your csv data here (:", key='input')
submit_button = st.form_submit_button(label='Send')
if submit_button and user_input:
output = conversational_chat(user_input)
# Maintain and display the chat history
st.session_state['past'].append(user_input)
# Maintain and display the chat history
st.session_state['generated'].append(output)
# Maintain and display the chat history
if st.session_state['generated']:
with response_container:
# Maintain and display the chat history
for i in range(len(st.session_state['generated'])):
# Maintain and display the chat history
message(st.session_state["past"][i], is_user=True, key=str(i) + '_user', avatar_style="big-smile")
# Maintain and display the chat history
message(st.session_state["generated"][i], key=str(i), avatar_style="thumbs")
|