Update app.py
Browse files
app.py
CHANGED
@@ -165,8 +165,15 @@ def create_vector_db(final_items):
|
|
165 |
|
166 |
Here is the antimony segment to summarize: {item}
|
167 |
"""
|
168 |
-
documents5 = llm(
|
169 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
170 |
documents.append(documents2)
|
171 |
|
172 |
if final_items:
|
@@ -208,8 +215,15 @@ def generate_response(db, query_text, previous_context):
|
|
208 |
{query_text}
|
209 |
|
210 |
"""
|
211 |
-
response = llm(
|
212 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
213 |
return final_response
|
214 |
|
215 |
def streamlit_app():
|
|
|
165 |
|
166 |
Here is the antimony segment to summarize: {item}
|
167 |
"""
|
168 |
+
documents5 = llm(
|
169 |
+
prompt,
|
170 |
+
max_tokens = 100000000,
|
171 |
+
temperature=0.0,
|
172 |
+
top_p=0.1,
|
173 |
+
echo=False,
|
174 |
+
stop = ["Q", "\n"]
|
175 |
+
)
|
176 |
+
documents2 = documents5["choices"][0]["text"].strip()
|
177 |
documents.append(documents2)
|
178 |
|
179 |
if final_items:
|
|
|
215 |
{query_text}
|
216 |
|
217 |
"""
|
218 |
+
response = llm(
|
219 |
+
prompt,
|
220 |
+
max_tokens = 100000000,
|
221 |
+
temperature=0.0,
|
222 |
+
top_p=0.1,
|
223 |
+
echo=False,
|
224 |
+
stop = ["Q", "\n"]
|
225 |
+
)
|
226 |
+
final_response = response["choices"][0]["text"].strip()
|
227 |
return final_response
|
228 |
|
229 |
def streamlit_app():
|