Update app.py
Browse files
app.py
CHANGED
@@ -66,9 +66,10 @@ from langchain.chains import create_retrieval_chain
|
|
66 |
#from langchain_community.llms import Ollama
|
67 |
#READER_MODEL=Ollama(model="meta-llama/Meta-Llama-Guard-2-8B")
|
68 |
|
|
|
|
|
69 |
|
70 |
-
READER_MODEL = "HuggingFaceH4/zephyr-7b-beta"
|
71 |
-
#READER_MODEL=ChatMistralAI(model=""HuggingFaceH4/zephyr-7b-beta")
|
72 |
|
73 |
#qa = ConversationalRetrievalChain.from_llm(llm=READER_MODEL,retriever=retriever,memory=memory)
|
74 |
#qa = RetrievalQA.from_chain_type(llm=READER_MODEL,retriever=retriever)
|
@@ -89,7 +90,7 @@ qa_chat_prompt = ChatPromptTemplate.from_messages(
|
|
89 |
)
|
90 |
|
91 |
docs_chain = create_stuff_documents_chain(
|
92 |
-
|
93 |
)
|
94 |
retrieval_chain = create_retrieval_chain(retriever, docs_chain)
|
95 |
response = retrieval_chain.invoke({"context": "how can I reverse diabetes?"})
|
@@ -98,7 +99,7 @@ print(response["answer"])
|
|
98 |
|
99 |
#result = qa(question)
|
100 |
#import gradio as gr
|
101 |
-
#gr.load("
|
102 |
|
103 |
#result = ({"query": question})
|
104 |
#print("qa")
|
|
|
66 |
#from langchain_community.llms import Ollama
|
67 |
#READER_MODEL=Ollama(model="meta-llama/Meta-Llama-Guard-2-8B")
|
68 |
|
69 |
+
from transformers import AutoModel
|
70 |
+
lll_model = AutoModel.from_pretrained(""HuggingFaceH4/zephyr-7b-beta")
|
71 |
|
72 |
+
#READER_MODEL = "HuggingFaceH4/zephyr-7b-beta"
|
|
|
73 |
|
74 |
#qa = ConversationalRetrievalChain.from_llm(llm=READER_MODEL,retriever=retriever,memory=memory)
|
75 |
#qa = RetrievalQA.from_chain_type(llm=READER_MODEL,retriever=retriever)
|
|
|
90 |
)
|
91 |
|
92 |
docs_chain = create_stuff_documents_chain(
|
93 |
+
lll_model, qa_chat_prompt
|
94 |
)
|
95 |
retrieval_chain = create_retrieval_chain(retriever, docs_chain)
|
96 |
response = retrieval_chain.invoke({"context": "how can I reverse diabetes?"})
|
|
|
99 |
|
100 |
#result = qa(question)
|
101 |
#import gradio as gr
|
102 |
+
#gr.load("lll_model").launch()
|
103 |
|
104 |
#result = ({"query": question})
|
105 |
#print("qa")
|