Update app.py
Browse files
app.py
CHANGED
@@ -11,7 +11,7 @@ st.title("🩺 ChatDoctor - Medical Assistant")
|
|
11 |
# Load model and tokenizer
|
12 |
@st.cache_resource
|
13 |
def load_model():
|
14 |
-
model = AutoModelForCausalLM.from_pretrained("abhiyanta/chatDoctor").to("cpu")
|
15 |
tokenizer = AutoTokenizer.from_pretrained("abhiyanta/chatDoctor")
|
16 |
return model, tokenizer
|
17 |
|
@@ -20,7 +20,7 @@ model, tokenizer = load_model()
|
|
20 |
# Alpaca-style prompt template
|
21 |
alpaca_prompt = "### Instruction:\n{0}\n\n### Input:\n{1}\n\n### Output:\n{2}"
|
22 |
|
23 |
-
# Text input for the user
|
24 |
user_input = st.text_input("Ask your medical question:")
|
25 |
|
26 |
# Button to trigger response
|
|
|
11 |
# Load model and tokenizer
|
12 |
@st.cache_resource
|
13 |
def load_model():
|
14 |
+
model = AutoModelForCausalLM.from_pretrained("abhiyanta/chatDoctor", use_cache=True).to("cpu")
|
15 |
tokenizer = AutoTokenizer.from_pretrained("abhiyanta/chatDoctor")
|
16 |
return model, tokenizer
|
17 |
|
|
|
20 |
# Alpaca-style prompt template
|
21 |
alpaca_prompt = "### Instruction:\n{0}\n\n### Input:\n{1}\n\n### Output:\n{2}"
|
22 |
|
23 |
+
# Text input for the user
|
24 |
user_input = st.text_input("Ask your medical question:")
|
25 |
|
26 |
# Button to trigger response
|