NicholasJohn commited on
Commit
65e6529
·
1 Parent(s): f0cd7b3
Files changed (3) hide show
  1. Dockerfile +6 -1
  2. app.py +1 -1
  3. requirements.txt +2 -2
Dockerfile CHANGED
@@ -23,7 +23,12 @@ ENV GRADIO_SERVER_NAME="0.0.0.0"
23
 
24
  # Download a checkpoint
25
  RUN mkdir content
26
- ADD --chown=user https://huggingface.co/LoneStriker/OpenBioLLM-Llama3-8B-GGUF/blob/main/OpenBioLLM-Llama3-8B-Q5_K_M.gguf content/OpenBioLLM-Llama3-8B-Q5_K_M.gguf
 
 
 
 
 
27
 
28
  RUN ls -l $HOME/app/content
29
 
 
23
 
24
  # Download a checkpoint
25
  RUN mkdir content
26
+ # Download the model using the Hugging Face CLI
27
+
28
+ # Install the Hugging Face CLI tool
29
+ RUN pip install huggingface_hub
30
+
31
+ RUN huggingface-cli download NicholasJohn/OpenBioLLM-Llama3-8B-Q5_K_M.gguf --local-dir ./models
32
 
33
  RUN ls -l $HOME/app/content
34
 
app.py CHANGED
@@ -11,7 +11,7 @@ from huggingface_hub.file_download import http_get
11
  llm = Llama(
12
  # model_path="./Phi-3-mini-4k-instruct-q4.gguf",
13
  # model_path="./llama3-gguf/OpenBioLLM-Llama3-8B-Q5_K_M.gguf",
14
- model_path = "./content/OpenBioLLM-Llama3-8B-Q5_K_M.gguf",
15
  n_ctx=2048,
16
  n_gpu_layers=50, # change n_gpu_layers if you have more or less VRAM
17
  )
 
11
  llm = Llama(
12
  # model_path="./Phi-3-mini-4k-instruct-q4.gguf",
13
  # model_path="./llama3-gguf/OpenBioLLM-Llama3-8B-Q5_K_M.gguf",
14
+ model_path = "./models/OpenBioLLM-Llama3-8B-Q5_K_M.gguf",
15
  n_ctx=2048,
16
  n_gpu_layers=50, # change n_gpu_layers if you have more or less VRAM
17
  )
requirements.txt CHANGED
@@ -1,3 +1,3 @@
1
- llama-cpp-python==0.2.69
2
- huggingface-hub==0.23.0
3
  gradio==4.28.3
 
1
+ # llama-cpp-python==0.2.69
2
+ # huggingface-hub==0.23.0
3
  gradio==4.28.3