Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -14,18 +14,13 @@ subprocess.run(
|
|
14 |
subprocess.run('pip install llama-cpp-agent==0.2.10', shell=True)
|
15 |
|
16 |
hf_hub_download(
|
17 |
-
repo_id="bartowski/
|
18 |
-
filename="
|
19 |
-
local_dir="./models"
|
20 |
-
)
|
21 |
-
hf_hub_download(
|
22 |
-
repo_id="bartowski/Llama-3-8B-Synthia-v3.5-GGUF",
|
23 |
-
filename="Llama-3-8B-Synthia-v3.5-f16.gguf",
|
24 |
local_dir="./models"
|
25 |
)
|
26 |
hf_hub_download(
|
27 |
-
repo_id="bartowski/
|
28 |
-
filename="
|
29 |
local_dir="./models"
|
30 |
)
|
31 |
|
@@ -215,9 +210,8 @@ demo = gr.ChatInterface(
|
|
215 |
label="Repetition penalty",
|
216 |
),
|
217 |
gr.Dropdown([
|
218 |
-
'
|
219 |
-
'
|
220 |
-
'Mistral-7B-Instruct-v0.3-f32.gguf'
|
221 |
],
|
222 |
value="Mistral-7B-Instruct-v0.3-f32.gguf",
|
223 |
label="Model"
|
|
|
14 |
subprocess.run('pip install llama-cpp-agent==0.2.10', shell=True)
|
15 |
|
16 |
hf_hub_download(
|
17 |
+
repo_id="bartowski/Mistral-7B-Instruct-v0.3-GGUF",
|
18 |
+
filename="Mistral-7B-Instruct-v0.3-f32.gguf",
|
|
|
|
|
|
|
|
|
|
|
19 |
local_dir="./models"
|
20 |
)
|
21 |
hf_hub_download(
|
22 |
+
repo_id="bartowski/Phi-3-medium-128k-instruct-GGUF",
|
23 |
+
filename="Phi-3-medium-128k-instruct-Q6_K.gguf",
|
24 |
local_dir="./models"
|
25 |
)
|
26 |
|
|
|
210 |
label="Repetition penalty",
|
211 |
),
|
212 |
gr.Dropdown([
|
213 |
+
'Mistral-7B-Instruct-v0.3-f32.gguf',
|
214 |
+
'Phi-3-medium-128k-instruct-Q6_K.gguf'
|
|
|
215 |
],
|
216 |
value="Mistral-7B-Instruct-v0.3-f32.gguf",
|
217 |
label="Model"
|