Update app.py
Browse files
app.py
CHANGED
@@ -173,20 +173,24 @@ def invoke(openai_api_key, rag_option, prompt):
|
|
173 |
completion, chain = rag_chain(llm, prompt, db)
|
174 |
result = completion["result"]
|
175 |
print("###")
|
176 |
-
print(completion)
|
|
|
|
|
177 |
print("###")
|
178 |
-
print(
|
179 |
print("###")
|
180 |
-
|
|
|
|
|
181 |
print("###")
|
182 |
else:
|
183 |
completion, chain = llm_chain(llm, prompt)
|
184 |
result = completion.generations[0][0].text if (completion.generations[0] != None and
|
185 |
completion.generations[0][0] != None) else ""
|
186 |
-
print("###")
|
187 |
-
print(completion)
|
188 |
-
print("###")
|
189 |
-
print(str(chain))
|
190 |
print("###")
|
191 |
print(chain.llm.client)
|
192 |
print("###")
|
|
|
173 |
completion, chain = rag_chain(llm, prompt, db)
|
174 |
result = completion["result"]
|
175 |
print("###")
|
176 |
+
#print(completion)
|
177 |
+
#print("###")
|
178 |
+
#print(str(chain))
|
179 |
print("###")
|
180 |
+
print(combine_documents_chain.llm.client)
|
181 |
print("###")
|
182 |
+
print(combine_documents_chain.llm_chain)
|
183 |
+
print("###")
|
184 |
+
print(combine_documents_chain.retriever)
|
185 |
print("###")
|
186 |
else:
|
187 |
completion, chain = llm_chain(llm, prompt)
|
188 |
result = completion.generations[0][0].text if (completion.generations[0] != None and
|
189 |
completion.generations[0][0] != None) else ""
|
190 |
+
#print("###")
|
191 |
+
#print(completion)
|
192 |
+
#print("###")
|
193 |
+
#print(str(chain))
|
194 |
print("###")
|
195 |
print(chain.llm.client)
|
196 |
print("###")
|