HemanthSai7 commited on
Commit
38ae2b6
·
1 Parent(s): cdca970

frontend with chatbot completion

Browse files
frontend/components/file_streaming.py CHANGED
@@ -7,11 +7,11 @@ from langchain.callbacks.base import BaseCallbackHandler
7
 
8
 
9
  @st.cache_resource(ttl="1h")
10
- def upload_data(uploaded_files):
11
  files = {"file": uploaded_files}
12
  with st.spinner("Uploading PDF..."):
13
  response = requests.post(
14
- "http://127.0.0.1:8000/api/upload", files=files
15
  )
16
 
17
  if response.status_code == 200:
 
7
 
8
 
9
  @st.cache_resource(ttl="1h")
10
+ def upload_data(uploaded_files, BASE_URL):
11
  files = {"file": uploaded_files}
12
  with st.spinner("Uploading PDF..."):
13
  response = requests.post(
14
+ f"{BASE_URL}/api/upload", files=files
15
  )
16
 
17
  if response.status_code == 200:
frontend/pages/2_🤖_bot.py CHANGED
@@ -3,7 +3,6 @@ import requests
3
  import streamlit as st
4
 
5
  from layouts.mainlayout import mainlayout
6
- from langchain.memory.chat_message_histories import StreamlitChatMessageHistory
7
 
8
  from components.file_streaming import *
9
  from components.display import *
@@ -32,13 +31,13 @@ def display():
32
 
33
  display()
34
 
35
- BASE_URL = "http://127.0.0.1:8000"
36
  uploaded_files = st.sidebar.file_uploader(label="Upload PDF files", type=["pdf"])
37
 
38
  if not uploaded_files:
39
  st.info("Please upload PDF documents to continue.")
40
  st.stop()
41
- upload_data(uploaded_files)
42
 
43
 
44
  if "messages" not in st.session_state.keys():
@@ -83,18 +82,14 @@ def generate_mistral_response(question: str):
83
 
84
  answer = response["result"]["answer"]
85
 
 
 
 
 
86
  except Exception as e:
87
  if response.json()=='exception.ModelDeployingException()':
88
  st.error("Model is deploying in the backend servers. Please try again after some time")
89
  st.stop()
90
-
91
-
92
- with st.expander("Source documents 🧐", expanded=True):
93
- source_documents = requests.post(
94
- f"{BASE_URL}/api/inference",
95
- json={"promptMessage": question}).json()["result"]["source_documents"]
96
- display_source_document(source_documents)
97
-
98
 
99
  return answer
100
 
 
3
  import streamlit as st
4
 
5
  from layouts.mainlayout import mainlayout
 
6
 
7
  from components.file_streaming import *
8
  from components.display import *
 
31
 
32
  display()
33
 
34
+ BASE_URL = "https://hemanthsai7-studybotapi.hf.space"
35
  uploaded_files = st.sidebar.file_uploader(label="Upload PDF files", type=["pdf"])
36
 
37
  if not uploaded_files:
38
  st.info("Please upload PDF documents to continue.")
39
  st.stop()
40
+ upload_data(uploaded_files, BASE_URL)
41
 
42
 
43
  if "messages" not in st.session_state.keys():
 
82
 
83
  answer = response["result"]["answer"]
84
 
85
+ with st.expander("Source documents 🧐", expanded=True):
86
+ source_documents = response["result"]["source_documents"]
87
+ display_source_document(source_documents)
88
+
89
  except Exception as e:
90
  if response.json()=='exception.ModelDeployingException()':
91
  st.error("Model is deploying in the backend servers. Please try again after some time")
92
  st.stop()
 
 
 
 
 
 
 
 
93
 
94
  return answer
95