expandme commited on
Commit
8ff9692
1 Parent(s): 82fe5ff

Sub 1B models ? - What wind.surf will do ?

Browse files
Files changed (2) hide show
  1. app.py +50 -12
  2. models.lst +11 -0
app.py CHANGED
@@ -4,18 +4,14 @@ import requests
4
 
5
  # Define available models
6
  MODELS = {
7
- "Llama-3.2-3B": {
8
- "repo_id": "lmstudio-community/Llama-3.2-3B-Instruct-GGUF",
9
- "filename": "*Q4_K_M.gguf",
10
- "chat_format": "chatml"
11
- },
12
- "Llama-3.2-5B": {
13
- "repo_id": "lmstudio-community/Llama-3.2-1B-Instruct-GGUF",
14
  "filename": "*Q4_K_M.gguf",
15
  "chat_format": "chatml"
16
- },
17
- "Phi-3.5-mini": {
18
- "repo_id": "bartowski/Phi-3.5-mini-instruct-GGUF",
19
  "filename": "*Q4_K_M.gguf",
20
  "chat_format": "chatml"
21
  },
@@ -29,6 +25,11 @@ MODELS = {
29
  "filename": "*Q4_K_M.gguf",
30
  "chat_format": "chatml"
31
  },
 
 
 
 
 
32
  "SmolLM2-1.7B": {
33
  "repo_id": "HuggingFaceTB/SmolLM2-1.7B-Instruct-GGUF",
34
  "filename": "*Q4_K_M.gguf",
@@ -39,13 +40,50 @@ MODELS = {
39
  "filename": "*Q4_K_M.gguf",
40
  "chat_format": "chatml"
41
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
42
  "Granite-1B": {
43
  "repo_id": "lmstudio-community/granite-3.0-1b-a400m-instruct-GGUF",
44
  "filename": "*Q4_K_M.gguf",
45
  "chat_format": "chatml"
46
  },
47
- "AMD-OLMo-1B": {
48
- "repo_id": "lmstudio-community/AMD-OLMo-1B-SFT-GGUF",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
49
  "filename": "*Q4_K_M.gguf",
50
  "chat_format": "chatml"
51
  }
 
4
 
5
  # Define available models
6
  MODELS = {
7
+ # 3B+ Models
8
+ "Phi-3.5-mini-4B": {
9
+ "repo_id": "bartowski/Phi-3.5-mini-instruct-GGUF",
 
 
 
 
10
  "filename": "*Q4_K_M.gguf",
11
  "chat_format": "chatml"
12
+ },
13
+ "Llama-3.2-3B": {
14
+ "repo_id": "lmstudio-community/Llama-3.2-3B-Instruct-GGUF",
15
  "filename": "*Q4_K_M.gguf",
16
  "chat_format": "chatml"
17
  },
 
25
  "filename": "*Q4_K_M.gguf",
26
  "chat_format": "chatml"
27
  },
28
+ "Gemma-2B": {
29
+ "repo_id": "lmstudio-community/gemma-2-2b-it-GGUF",
30
+ "filename": "*Q4_K_M.gguf",
31
+ "chat_format": "chatml"
32
+ },
33
  "SmolLM2-1.7B": {
34
  "repo_id": "HuggingFaceTB/SmolLM2-1.7B-Instruct-GGUF",
35
  "filename": "*Q4_K_M.gguf",
 
40
  "filename": "*Q4_K_M.gguf",
41
  "chat_format": "chatml"
42
  },
43
+ "Llama-3.2-1B": {
44
+ "repo_id": "lmstudio-community/Llama-3.2-1B-Instruct-GGUF",
45
+ "filename": "*Q4_K_M.gguf",
46
+ "chat_format": "chatml"
47
+ },
48
+ "AMD-OLMo-1B": {
49
+ "repo_id": "lmstudio-community/AMD-OLMo-1B-SFT-GGUF",
50
+ "filename": "*Q4_K_M.gguf",
51
+ "chat_format": "chatml"
52
+ },
53
+ "AMD-OLMo-1B-DPO": {
54
+ "repo_id": "lmstudio-community/AMD-OLMo-1B-SFT-DPO-GGUF",
55
+ "filename": "*Q4_K_M.gguf",
56
+ "chat_format": "chatml"
57
+ },
58
  "Granite-1B": {
59
  "repo_id": "lmstudio-community/granite-3.0-1b-a400m-instruct-GGUF",
60
  "filename": "*Q4_K_M.gguf",
61
  "chat_format": "chatml"
62
  },
63
+
64
+ # Sub-1B Models
65
+ "MobileLLM-600M": {
66
+ "repo_id": "pjh64/MobileLLM-600M-GGUF",
67
+ "filename": "*Q4_K_M.gguf",
68
+ "chat_format": "chatml"
69
+ },
70
+ "Danube-500M": {
71
+ "repo_id": "BoscoTheDog/Danube_3-500M_Chat_GGUF",
72
+ "filename": "*Q4_K_M.gguf",
73
+ "chat_format": "chatml"
74
+ },
75
+ "Qwen2.5-500M": {
76
+ "repo_id": "Qwen/Qwen2.5-0.5B-Instruct-GGUF",
77
+ "filename": "*Q4_K_M.gguf",
78
+ "chat_format": "chatml"
79
+ },
80
+ "MobileLLM-350M": {
81
+ "repo_id": "pjh64/MobileLLM-350M-GGUF",
82
+ "filename": "*Q4_K_M.gguf",
83
+ "chat_format": "chatml"
84
+ },
85
+ "MobileLLM-125M": {
86
+ "repo_id": "pjh64/MobileLLM-125M-GGUF",
87
  "filename": "*Q4_K_M.gguf",
88
  "chat_format": "chatml"
89
  }
models.lst CHANGED
@@ -6,6 +6,8 @@ https://huggingface.co/lmstudio-community/granite-3.0-3b-a800m-instruct-GGUF
6
 
7
  https://huggingface.co/lmstudio-community/Qwen2.5-3B-Instruct-GGUF
8
 
 
 
9
  https://huggingface.co/HuggingFaceTB/SmolLM2-1.7B-Instruct-GGUF
10
 
11
  https://huggingface.co/lmstudio-community/Qwen2.5-1.5B-Instruct-GGUF
@@ -14,4 +16,13 @@ https://huggingface.co/lmstudio-community/granite-3.0-1b-a400m-instruct-GGUF
14
 
15
  https://huggingface.co/lmstudio-community/AMD-OLMo-1B-SFT-DPO-GGUF
16
 
 
 
 
 
 
 
 
 
 
17
 
 
6
 
7
  https://huggingface.co/lmstudio-community/Qwen2.5-3B-Instruct-GGUF
8
 
9
+ https://huggingface.co/lmstudio-community/gemma-2-2b-it-GGUF
10
+
11
  https://huggingface.co/HuggingFaceTB/SmolLM2-1.7B-Instruct-GGUF
12
 
13
  https://huggingface.co/lmstudio-community/Qwen2.5-1.5B-Instruct-GGUF
 
16
 
17
  https://huggingface.co/lmstudio-community/AMD-OLMo-1B-SFT-DPO-GGUF
18
 
19
+ https://huggingface.co/pjh64/MobileLLM-600M-GGUF
20
+
21
+ https://huggingface.co/BoscoTheDog/Danube_3-500M_Chat_GGUF
22
+
23
+ https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct-GGUF
24
+
25
+ https://huggingface.co/pjh64/MobileLLM-350M-GGUF
26
+
27
+ https://huggingface.co/pjh64/MobileLLM-125M-GGUF
28