Spaces:
Sleeping
Sleeping
NicholasJohn
commited on
Commit
·
65e6529
1
Parent(s):
f0cd7b3
tweak 8
Browse files- Dockerfile +6 -1
- app.py +1 -1
- requirements.txt +2 -2
Dockerfile
CHANGED
@@ -23,7 +23,12 @@ ENV GRADIO_SERVER_NAME="0.0.0.0"
|
|
23 |
|
24 |
# Download a checkpoint
|
25 |
RUN mkdir content
|
26 |
-
|
|
|
|
|
|
|
|
|
|
|
27 |
|
28 |
RUN ls -l $HOME/app/content
|
29 |
|
|
|
23 |
|
24 |
# Download a checkpoint
|
25 |
RUN mkdir content
|
26 |
+
# Download the model using the Hugging Face CLI
|
27 |
+
|
28 |
+
# Install the Hugging Face CLI tool
|
29 |
+
RUN pip install huggingface_hub
|
30 |
+
|
31 |
+
RUN huggingface-cli download NicholasJohn/OpenBioLLM-Llama3-8B-Q5_K_M.gguf --local-dir ./models
|
32 |
|
33 |
RUN ls -l $HOME/app/content
|
34 |
|
app.py
CHANGED
@@ -11,7 +11,7 @@ from huggingface_hub.file_download import http_get
|
|
11 |
llm = Llama(
|
12 |
# model_path="./Phi-3-mini-4k-instruct-q4.gguf",
|
13 |
# model_path="./llama3-gguf/OpenBioLLM-Llama3-8B-Q5_K_M.gguf",
|
14 |
-
model_path = "./
|
15 |
n_ctx=2048,
|
16 |
n_gpu_layers=50, # change n_gpu_layers if you have more or less VRAM
|
17 |
)
|
|
|
11 |
llm = Llama(
|
12 |
# model_path="./Phi-3-mini-4k-instruct-q4.gguf",
|
13 |
# model_path="./llama3-gguf/OpenBioLLM-Llama3-8B-Q5_K_M.gguf",
|
14 |
+
model_path = "./models/OpenBioLLM-Llama3-8B-Q5_K_M.gguf",
|
15 |
n_ctx=2048,
|
16 |
n_gpu_layers=50, # change n_gpu_layers if you have more or less VRAM
|
17 |
)
|
requirements.txt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
-
llama-cpp-python==0.2.69
|
2 |
-
huggingface-hub==0.23.0
|
3 |
gradio==4.28.3
|
|
|
1 |
+
# llama-cpp-python==0.2.69
|
2 |
+
# huggingface-hub==0.23.0
|
3 |
gradio==4.28.3
|