Testing of GGUF Llama3.2 3B
Browse files- README.md +1 -1
- app.py +2 -3
- models.lst +3 -0
README.md
CHANGED
@@ -5,7 +5,7 @@ colorFrom: green
|
|
5 |
colorTo: indigo
|
6 |
sdk: gradio
|
7 |
app_file: app.py
|
8 |
-
pinned:
|
9 |
license: cc-by-sa-4.0
|
10 |
short_description: SmallZOO runnigng SLMs directly on CPU with Llama.cpp&Python
|
11 |
---
|
|
|
5 |
colorTo: indigo
|
6 |
sdk: gradio
|
7 |
app_file: app.py
|
8 |
+
pinned: flase
|
9 |
license: cc-by-sa-4.0
|
10 |
short_description: SmallZOO runnigng SLMs directly on CPU with Llama.cpp&Python
|
11 |
---
|
app.py
CHANGED
@@ -2,10 +2,9 @@ import gradio as gr
|
|
2 |
from llama_cpp import Llama
|
3 |
import requests
|
4 |
|
5 |
-
|
6 |
llm = Llama.from_pretrained(
|
7 |
-
repo_id="
|
8 |
-
filename="*
|
9 |
verbose=True,
|
10 |
n_ctx=32768,
|
11 |
n_threads=2,
|
|
|
2 |
from llama_cpp import Llama
|
3 |
import requests
|
4 |
|
|
|
5 |
llm = Llama.from_pretrained(
|
6 |
+
repo_id="lmstudio-community/Llama-3.2-3B-Instruct-GGUF",
|
7 |
+
filename="*Q4_K_M.gguf",
|
8 |
verbose=True,
|
9 |
n_ctx=32768,
|
10 |
n_threads=2,
|
models.lst
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
Stack of modesl to try:
|
2 |
+
|
3 |
+
https://huggingface.co/lmstudio-community/Llama-3.2-3B-Instruct-GGUF
|