Spaces:
Paused
Paused
Robin Genolet
commited on
Commit
·
2bfab33
1
Parent(s):
9730359
fix: prompt
Browse files- app.py +1 -0
- utils/epfl_meditron_utils.py +2 -2
app.py
CHANGED
@@ -513,6 +513,7 @@ def display_llm_output():
|
|
513 |
st.session_state["model_filename"],
|
514 |
st.session_state["model_type"],
|
515 |
st.session_state["gpu_layers"],
|
|
|
516 |
llm_message)
|
517 |
st.write(llm_response)
|
518 |
st.write('Done displaying LLM response')
|
|
|
513 |
st.session_state["model_filename"],
|
514 |
st.session_state["model_type"],
|
515 |
st.session_state["gpu_layers"],
|
516 |
+
"You are a medical assistant",
|
517 |
llm_message)
|
518 |
st.write(llm_response)
|
519 |
st.write('Done displaying LLM response')
|
utils/epfl_meditron_utils.py
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
|
2 |
|
3 |
|
4 |
-
def get_llm_response(repo, filename, model_type, gpu_layers, prompt):
|
5 |
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
6 |
|
7 |
model_name_or_path = "TheBloke/meditron-7B-GPTQ"
|
@@ -40,7 +40,7 @@ def get_llm_response(repo, filename, model_type, gpu_layers, prompt):
|
|
40 |
<|im_start|>user
|
41 |
{prompt}<|im_end|>
|
42 |
<|im_start|>assistant
|
43 |
-
'''
|
44 |
|
45 |
response = pipe(prompt_template)[0]['generated_text']
|
46 |
print(response)
|
|
|
1 |
|
2 |
|
3 |
|
4 |
+
def get_llm_response(repo, filename, model_type, gpu_layers, system_message, prompt):
|
5 |
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
6 |
|
7 |
model_name_or_path = "TheBloke/meditron-7B-GPTQ"
|
|
|
40 |
<|im_start|>user
|
41 |
{prompt}<|im_end|>
|
42 |
<|im_start|>assistant
|
43 |
+
'''
|
44 |
|
45 |
response = pipe(prompt_template)[0]['generated_text']
|
46 |
print(response)
|