Update app.py
Browse files
app.py
CHANGED
@@ -76,8 +76,13 @@ qa_chat_prompt = ChatPromptTemplate.from_messages(
|
|
76 |
)
|
77 |
|
78 |
llm_model = "bigscience/bloom-560m"
|
|
|
|
|
|
|
79 |
print("check2")
|
80 |
-
pipe = pipeline(model = llm_model, tokenizer =
|
|
|
|
|
81 |
print("check3")
|
82 |
|
83 |
#chain = qa_chat_prompt | pipe
|
|
|
76 |
)
|
77 |
|
78 |
llm_model = "bigscience/bloom-560m"
|
79 |
+
from transformers import AutoTokenizer
|
80 |
+
tokenizer = AutoTokenizer.from_pretrained(llm_model)
|
81 |
+
|
82 |
print("check2")
|
83 |
+
pipe = pipeline(model = llm_model, tokenizer = tokenizer, task="text-generation",chat = qa_chat_prompt)
|
84 |
+
#"question-answering", "conversational"
|
85 |
+
|
86 |
print("check3")
|
87 |
|
88 |
#chain = qa_chat_prompt | pipe
|