moriire commited on
Commit
fae2893
·
verified ·
1 Parent(s): ba78b1c

Update app/llm.py

Browse files
Files changed (1) hide show
  1. app/llm.py +4 -4
app/llm.py CHANGED
@@ -30,8 +30,8 @@ class ChatModel(BaseModel):
30
  mirostat_tau: float=4.0
31
  mirostat_eta: float=1.1
32
  llm_chat = llama_cpp.Llama.from_pretrained(
33
- repo_id="moriire/healthcare-ai-adapter-merged",
34
- #filename="healthcare-GGUF-unsloth.Q4_K_M.gguf",
35
  #tokenizer=llama_cpp.llama_tokenizer.LlamaHFTokenizer.from_pretrained("moriire/healthcare-ai-adapter-merged"),
36
  verbose=False,
37
  n_ctx=1024,
@@ -39,8 +39,8 @@ llm_chat = llama_cpp.Llama.from_pretrained(
39
  chat_format="llama-2"
40
  )
41
  llm_generate = llama_cpp.Llama.from_pretrained(
42
- repo_id="moriire/healthcare-GGUF",
43
- filename="healthcare-GGUF-unsloth.Q4_K_M.gguf",
44
  #tokenizer=llama_cpp.llama_tokenizer.LlamaHFTokenizer.from_pretrained("moriire/healthcare-ai-adapter-merged"),
45
  verbose=False,
46
  n_ctx=4096,
 
30
  mirostat_tau: float=4.0
31
  mirostat_eta: float=1.1
32
  llm_chat = llama_cpp.Llama.from_pretrained(
33
+ repo_id="moriire/healthcare-ai-q8_0",
34
+ filename="healthcare-ai-q8_0-unsloth.Q8_0.gguf",
35
  #tokenizer=llama_cpp.llama_tokenizer.LlamaHFTokenizer.from_pretrained("moriire/healthcare-ai-adapter-merged"),
36
  verbose=False,
37
  n_ctx=1024,
 
39
  chat_format="llama-2"
40
  )
41
  llm_generate = llama_cpp.Llama.from_pretrained(
42
+ repo_id="moriire/healthcare-ai-q8_0",
43
+ filename="healthcare-ai-q8_0-unsloth.Q8_0.gguf",
44
  #tokenizer=llama_cpp.llama_tokenizer.LlamaHFTokenizer.from_pretrained("moriire/healthcare-ai-adapter-merged"),
45
  verbose=False,
46
  n_ctx=4096,