Update app.py
Browse files
app.py
CHANGED
@@ -92,6 +92,10 @@ from transformers import AutoTokenizer
|
|
92 |
tokenizer = AutoTokenizer.from_pretrained(llm_model,trust_remote_code=True)
|
93 |
from transformers import AutoModelForCausalLM
|
94 |
model = AutoModelForCausalLM.from_pretrained(llm_model,trust_remote_code=True)
|
|
|
|
|
|
|
|
|
95 |
|
96 |
#question = "How can I reverse diabetes?"
|
97 |
|
@@ -102,7 +106,7 @@ from langchain.chains.question_answering import load_qa_chain
|
|
102 |
#pipe = load_qa_chain(llm=llm_model,tokenizer =tokenizer, chain_type="map_reduce")
|
103 |
print("check2")
|
104 |
qa = ConversationalRetrievalChain.from_llm(
|
105 |
-
|
106 |
retriever=retriever,
|
107 |
memory=memory,
|
108 |
chain_type_kwargs={"prompt": QA_CHAIN_PROMPT}
|
|
|
92 |
tokenizer = AutoTokenizer.from_pretrained(llm_model,trust_remote_code=True)
|
93 |
from transformers import AutoModelForCausalLM
|
94 |
model = AutoModelForCausalLM.from_pretrained(llm_model,trust_remote_code=True)
|
95 |
+
llm = HuggingFaceLLM(
|
96 |
+
tokenizer_name="microsoft/Phi-3-mini-4k-instruct",
|
97 |
+
model_name="microsoft/Phi-3-mini-4k-instruct",
|
98 |
+
)
|
99 |
|
100 |
#question = "How can I reverse diabetes?"
|
101 |
|
|
|
106 |
#pipe = load_qa_chain(llm=llm_model,tokenizer =tokenizer, chain_type="map_reduce")
|
107 |
print("check2")
|
108 |
qa = ConversationalRetrievalChain.from_llm(
|
109 |
+
llm,
|
110 |
retriever=retriever,
|
111 |
memory=memory,
|
112 |
chain_type_kwargs={"prompt": QA_CHAIN_PROMPT}
|