Spaces:
Running
on
Zero
Running
on
Zero
gokaygokay
commited on
Update app.py
Browse files
app.py
CHANGED
@@ -14,6 +14,9 @@ def run_command(command):
|
|
14 |
return output.decode('utf-8')
|
15 |
|
16 |
def install_packages():
|
|
|
|
|
|
|
17 |
# Clone the repository with submodules
|
18 |
run_command("git clone --recurse-submodules https://github.com/abetlen/llama-cpp-python.git")
|
19 |
|
@@ -32,10 +35,12 @@ def install_packages():
|
|
32 |
run_command("pip install -e .[all]")
|
33 |
|
34 |
# Clear the local build cache
|
|
|
|
|
35 |
run_command("make clean")
|
36 |
|
37 |
# Reinstall the package
|
38 |
-
run_command('
|
39 |
|
40 |
# Install llama-cpp-agent
|
41 |
run_command("pip install llama-cpp-agent")
|
|
|
14 |
return output.decode('utf-8')
|
15 |
|
16 |
def install_packages():
|
17 |
+
|
18 |
+
run_command("apt install nvidia-cuda-toolkit")
|
19 |
+
|
20 |
# Clone the repository with submodules
|
21 |
run_command("git clone --recurse-submodules https://github.com/abetlen/llama-cpp-python.git")
|
22 |
|
|
|
35 |
run_command("pip install -e .[all]")
|
36 |
|
37 |
# Clear the local build cache
|
38 |
+
run_command("make GGML_CUDA=1")
|
39 |
+
|
40 |
run_command("make clean")
|
41 |
|
42 |
# Reinstall the package
|
43 |
+
run_command('pip install -e .')
|
44 |
|
45 |
# Install llama-cpp-agent
|
46 |
run_command("pip install llama-cpp-agent")
|