Update app.py
Browse files
app.py
CHANGED
@@ -77,7 +77,7 @@ qa_chat_prompt = ChatPromptTemplate.from_messages(
|
|
77 |
|
78 |
llm_model = "bigscience/bloom-560m"
|
79 |
print("check2")
|
80 |
-
pipe = pipeline(
|
81 |
print("check3")
|
82 |
|
83 |
#chain = qa_chat_prompt | pipe
|
|
|
77 |
|
78 |
llm_model = "bigscience/bloom-560m"
|
79 |
print("check2")
|
80 |
+
pipe = pipeline(model = llm_model, tokenizer = retriever, task="text-generation",chat = qa_chat_prompt)
|
81 |
print("check3")
|
82 |
|
83 |
#chain = qa_chat_prompt | pipe
|