Sub 1B models ? - What wind.surf will do ?
Browse files- app.py +50 -12
- models.lst +11 -0
app.py
CHANGED
@@ -4,18 +4,14 @@ import requests
|
|
4 |
|
5 |
# Define available models
|
6 |
MODELS = {
|
7 |
-
|
8 |
-
|
9 |
-
"
|
10 |
-
"chat_format": "chatml"
|
11 |
-
},
|
12 |
-
"Llama-3.2-5B": {
|
13 |
-
"repo_id": "lmstudio-community/Llama-3.2-1B-Instruct-GGUF",
|
14 |
"filename": "*Q4_K_M.gguf",
|
15 |
"chat_format": "chatml"
|
16 |
-
},
|
17 |
-
"
|
18 |
-
"repo_id": "
|
19 |
"filename": "*Q4_K_M.gguf",
|
20 |
"chat_format": "chatml"
|
21 |
},
|
@@ -29,6 +25,11 @@ MODELS = {
|
|
29 |
"filename": "*Q4_K_M.gguf",
|
30 |
"chat_format": "chatml"
|
31 |
},
|
|
|
|
|
|
|
|
|
|
|
32 |
"SmolLM2-1.7B": {
|
33 |
"repo_id": "HuggingFaceTB/SmolLM2-1.7B-Instruct-GGUF",
|
34 |
"filename": "*Q4_K_M.gguf",
|
@@ -39,13 +40,50 @@ MODELS = {
|
|
39 |
"filename": "*Q4_K_M.gguf",
|
40 |
"chat_format": "chatml"
|
41 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
42 |
"Granite-1B": {
|
43 |
"repo_id": "lmstudio-community/granite-3.0-1b-a400m-instruct-GGUF",
|
44 |
"filename": "*Q4_K_M.gguf",
|
45 |
"chat_format": "chatml"
|
46 |
},
|
47 |
-
|
48 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
49 |
"filename": "*Q4_K_M.gguf",
|
50 |
"chat_format": "chatml"
|
51 |
}
|
|
|
4 |
|
5 |
# Define available models
|
6 |
MODELS = {
|
7 |
+
# 3B+ Models
|
8 |
+
"Phi-3.5-mini-4B": {
|
9 |
+
"repo_id": "bartowski/Phi-3.5-mini-instruct-GGUF",
|
|
|
|
|
|
|
|
|
10 |
"filename": "*Q4_K_M.gguf",
|
11 |
"chat_format": "chatml"
|
12 |
+
},
|
13 |
+
"Llama-3.2-3B": {
|
14 |
+
"repo_id": "lmstudio-community/Llama-3.2-3B-Instruct-GGUF",
|
15 |
"filename": "*Q4_K_M.gguf",
|
16 |
"chat_format": "chatml"
|
17 |
},
|
|
|
25 |
"filename": "*Q4_K_M.gguf",
|
26 |
"chat_format": "chatml"
|
27 |
},
|
28 |
+
"Gemma-2B": {
|
29 |
+
"repo_id": "lmstudio-community/gemma-2-2b-it-GGUF",
|
30 |
+
"filename": "*Q4_K_M.gguf",
|
31 |
+
"chat_format": "chatml"
|
32 |
+
},
|
33 |
"SmolLM2-1.7B": {
|
34 |
"repo_id": "HuggingFaceTB/SmolLM2-1.7B-Instruct-GGUF",
|
35 |
"filename": "*Q4_K_M.gguf",
|
|
|
40 |
"filename": "*Q4_K_M.gguf",
|
41 |
"chat_format": "chatml"
|
42 |
},
|
43 |
+
"Llama-3.2-1B": {
|
44 |
+
"repo_id": "lmstudio-community/Llama-3.2-1B-Instruct-GGUF",
|
45 |
+
"filename": "*Q4_K_M.gguf",
|
46 |
+
"chat_format": "chatml"
|
47 |
+
},
|
48 |
+
"AMD-OLMo-1B": {
|
49 |
+
"repo_id": "lmstudio-community/AMD-OLMo-1B-SFT-GGUF",
|
50 |
+
"filename": "*Q4_K_M.gguf",
|
51 |
+
"chat_format": "chatml"
|
52 |
+
},
|
53 |
+
"AMD-OLMo-1B-DPO": {
|
54 |
+
"repo_id": "lmstudio-community/AMD-OLMo-1B-SFT-DPO-GGUF",
|
55 |
+
"filename": "*Q4_K_M.gguf",
|
56 |
+
"chat_format": "chatml"
|
57 |
+
},
|
58 |
"Granite-1B": {
|
59 |
"repo_id": "lmstudio-community/granite-3.0-1b-a400m-instruct-GGUF",
|
60 |
"filename": "*Q4_K_M.gguf",
|
61 |
"chat_format": "chatml"
|
62 |
},
|
63 |
+
|
64 |
+
# Sub-1B Models
|
65 |
+
"MobileLLM-600M": {
|
66 |
+
"repo_id": "pjh64/MobileLLM-600M-GGUF",
|
67 |
+
"filename": "*Q4_K_M.gguf",
|
68 |
+
"chat_format": "chatml"
|
69 |
+
},
|
70 |
+
"Danube-500M": {
|
71 |
+
"repo_id": "BoscoTheDog/Danube_3-500M_Chat_GGUF",
|
72 |
+
"filename": "*Q4_K_M.gguf",
|
73 |
+
"chat_format": "chatml"
|
74 |
+
},
|
75 |
+
"Qwen2.5-500M": {
|
76 |
+
"repo_id": "Qwen/Qwen2.5-0.5B-Instruct-GGUF",
|
77 |
+
"filename": "*Q4_K_M.gguf",
|
78 |
+
"chat_format": "chatml"
|
79 |
+
},
|
80 |
+
"MobileLLM-350M": {
|
81 |
+
"repo_id": "pjh64/MobileLLM-350M-GGUF",
|
82 |
+
"filename": "*Q4_K_M.gguf",
|
83 |
+
"chat_format": "chatml"
|
84 |
+
},
|
85 |
+
"MobileLLM-125M": {
|
86 |
+
"repo_id": "pjh64/MobileLLM-125M-GGUF",
|
87 |
"filename": "*Q4_K_M.gguf",
|
88 |
"chat_format": "chatml"
|
89 |
}
|
models.lst
CHANGED
@@ -6,6 +6,8 @@ https://huggingface.co/lmstudio-community/granite-3.0-3b-a800m-instruct-GGUF
|
|
6 |
|
7 |
https://huggingface.co/lmstudio-community/Qwen2.5-3B-Instruct-GGUF
|
8 |
|
|
|
|
|
9 |
https://huggingface.co/HuggingFaceTB/SmolLM2-1.7B-Instruct-GGUF
|
10 |
|
11 |
https://huggingface.co/lmstudio-community/Qwen2.5-1.5B-Instruct-GGUF
|
@@ -14,4 +16,13 @@ https://huggingface.co/lmstudio-community/granite-3.0-1b-a400m-instruct-GGUF
|
|
14 |
|
15 |
https://huggingface.co/lmstudio-community/AMD-OLMo-1B-SFT-DPO-GGUF
|
16 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
17 |
|
|
|
6 |
|
7 |
https://huggingface.co/lmstudio-community/Qwen2.5-3B-Instruct-GGUF
|
8 |
|
9 |
+
https://huggingface.co/lmstudio-community/gemma-2-2b-it-GGUF
|
10 |
+
|
11 |
https://huggingface.co/HuggingFaceTB/SmolLM2-1.7B-Instruct-GGUF
|
12 |
|
13 |
https://huggingface.co/lmstudio-community/Qwen2.5-1.5B-Instruct-GGUF
|
|
|
16 |
|
17 |
https://huggingface.co/lmstudio-community/AMD-OLMo-1B-SFT-DPO-GGUF
|
18 |
|
19 |
+
https://huggingface.co/pjh64/MobileLLM-600M-GGUF
|
20 |
+
|
21 |
+
https://huggingface.co/BoscoTheDog/Danube_3-500M_Chat_GGUF
|
22 |
+
|
23 |
+
https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct-GGUF
|
24 |
+
|
25 |
+
https://huggingface.co/pjh64/MobileLLM-350M-GGUF
|
26 |
+
|
27 |
+
https://huggingface.co/pjh64/MobileLLM-125M-GGUF
|
28 |
|