Spaces:
Sleeping
Sleeping
update app.py
Browse files
app.py
CHANGED
@@ -51,6 +51,12 @@ def create_conversational_chain(vector_store):
|
|
51 |
replicate_api_token = "r8_AA3K1fhDykqLa5M74E5V0w5ss1z0P9S3foWJl" # Replace with your actual token
|
52 |
os.environ["REPLICATE_API_TOKEN"] = replicate_api_token
|
53 |
|
|
|
|
|
|
|
|
|
|
|
|
|
54 |
llm = Replicate(
|
55 |
streaming=True,
|
56 |
model="replicate/llama-2-70b-chat:58d078176e02c219e11eb4da5a02a7830a283b14cf8f94537af893ccff5ee781",
|
@@ -65,6 +71,7 @@ def create_conversational_chain(vector_store):
|
|
65 |
memory=memory)
|
66 |
return chain
|
67 |
|
|
|
68 |
def main():
|
69 |
initialize_session_state()
|
70 |
st.title("Chat With Your Doc")
|
|
|
51 |
replicate_api_token = "r8_AA3K1fhDykqLa5M74E5V0w5ss1z0P9S3foWJl" # Replace with your actual token
|
52 |
os.environ["REPLICATE_API_TOKEN"] = replicate_api_token
|
53 |
|
54 |
+
print("Length of text_chunks:", len(text_chunks))
|
55 |
+
print("Content of text_chunks:", text_chunks)
|
56 |
+
|
57 |
+
print("Length of embeddings:", len(embeddings))
|
58 |
+
print("Content of embeddings:", embeddings)
|
59 |
+
|
60 |
llm = Replicate(
|
61 |
streaming=True,
|
62 |
model="replicate/llama-2-70b-chat:58d078176e02c219e11eb4da5a02a7830a283b14cf8f94537af893ccff5ee781",
|
|
|
71 |
memory=memory)
|
72 |
return chain
|
73 |
|
74 |
+
|
75 |
def main():
|
76 |
initialize_session_state()
|
77 |
st.title("Chat With Your Doc")
|