Spaces:
Sleeping
Sleeping
Use 2 cores
Browse files
app.py
CHANGED
@@ -6,7 +6,9 @@ from llama_cpp import Llama
|
|
6 |
|
7 |
llm = Llama(
|
8 |
model_path="gemma-2b-uk.gguf",
|
9 |
-
chat_format="gemma"
|
|
|
|
|
10 |
)
|
11 |
|
12 |
|
|
|
6 |
|
7 |
llm = Llama(
|
8 |
model_path="gemma-2b-uk.gguf",
|
9 |
+
chat_format="gemma",
|
10 |
+
n_threads=2,
|
11 |
+
n_threads_batch=2,
|
12 |
)
|
13 |
|
14 |
|