Spaces:
Running
on
Zero
Running
on
Zero
Upload llmdolphin.py
Browse files- llmdolphin.py +20 -0
llmdolphin.py
CHANGED
@@ -79,6 +79,26 @@ llm_models = {
|
|
79 |
"ChatWaifu_22B_v2.0_preview.Q4_K_S.gguf": ["mradermacher/ChatWaifu_22B_v2.0_preview-GGUF", MessagesFormatterType.MISTRAL],
|
80 |
"ChatWaifu_v1.4.Q5_K_M.gguf": ["mradermacher/ChatWaifu_v1.4-GGUF", MessagesFormatterType.MISTRAL],
|
81 |
"ChatWaifu_v1.3.1.Q4_K_M.gguf": ["mradermacher/ChatWaifu_v1.3.1-GGUF", MessagesFormatterType.MISTRAL],
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
82 |
"Llama-3.1-SuperNova-8B-Lite_TIES_with_Base.i1-Q5_K_M.gguf": ["mradermacher/Llama-3.1-SuperNova-8B-Lite_TIES_with_Base-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
83 |
"maverick-llama3-8B.i1-Q4_K_M.gguf": ["mradermacher/maverick-llama3-8B-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
84 |
"JSL-MedLlama-3-8B-v2.0.i1-Q5_K_M.gguf": ["mradermacher/JSL-MedLlama-3-8B-v2.0-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
|
|
79 |
"ChatWaifu_22B_v2.0_preview.Q4_K_S.gguf": ["mradermacher/ChatWaifu_22B_v2.0_preview-GGUF", MessagesFormatterType.MISTRAL],
|
80 |
"ChatWaifu_v1.4.Q5_K_M.gguf": ["mradermacher/ChatWaifu_v1.4-GGUF", MessagesFormatterType.MISTRAL],
|
81 |
"ChatWaifu_v1.3.1.Q4_K_M.gguf": ["mradermacher/ChatWaifu_v1.3.1-GGUF", MessagesFormatterType.MISTRAL],
|
82 |
+
"Infinirc-ArliAI-RPMax-v1.3-merge-8B.i1-Q5_K_M.gguf": ["mradermacher/Infinirc-ArliAI-RPMax-v1.3-merge-8B-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
83 |
+
"Verdandi-Qwen2.5-7B.Q5_K_M.gguf": ["mradermacher/Verdandi-Qwen2.5-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
84 |
+
"Qwen2-Wukong-7B.Q5_K_M.gguf": ["mradermacher/Qwen2-Wukong-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
85 |
+
"Lelantos-7B.Q5_K_M.gguf": ["mradermacher/Lelantos-7B-GGUF", MessagesFormatterType.MISTRAL],
|
86 |
+
"Kunoichi-7B.Q5_K_M.gguf": ["mradermacher/Kunoichi-7B-GGUF", MessagesFormatterType.MISTRAL],
|
87 |
+
"llama-3-cat-8b-instruct-v1.Q5_K_M.gguf": ["mradermacher/llama-3-cat-8b-instruct-v1-GGUF", MessagesFormatterType.LLAMA_3],
|
88 |
+
"DPOpenHermes-7B.i1-Q5_K_M.gguf": ["mradermacher/DPOpenHermes-7B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
89 |
+
"NeuralDaredevil-8B-abliterated.Q5_K_M.gguf": ["mradermacher/NeuralDaredevil-8B-abliterated-GGUF", MessagesFormatterType.LLAMA_3],
|
90 |
+
"AstroSage-8B.Q5_K_M.gguf": ["mradermacher/AstroSage-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
91 |
+
"Llama-3-8B-Theresa.Q5_K_M.gguf": ["mradermacher/Llama-3-8B-Theresa-GGUF", MessagesFormatterType.LLAMA_3],
|
92 |
+
"Sonya-7B.Q5_K_M.gguf": ["mradermacher/Sonya-7B-GGUF", MessagesFormatterType.MISTRAL],
|
93 |
+
"Hathor_Tahsin-L3-8B-v0.9.Q4_K_S.gguf": ["mradermacher/Hathor_Tahsin-L3-8B-v0.9-GGUF", MessagesFormatterType.LLAMA_3],
|
94 |
+
"Qwen-Qwen2.5-7B-Instruct-llamafied.Q5_K_M.gguf": ["mradermacher/Qwen-Qwen2.5-7B-Instruct-llamafied-GGUF", MessagesFormatterType.OPEN_CHAT],
|
95 |
+
"Ice0.34n-14.11-RP.i1-Q4_K_S.gguf": ["mradermacher/Ice0.34n-14.11-RP-i1-GGUF", MessagesFormatterType.MISTRAL],
|
96 |
+
"Qwen2.5-14B-Wernicke-SFT.Q4_K_S.gguf": ["mradermacher/Qwen2.5-14B-Wernicke-SFT-GGUF", MessagesFormatterType.OPEN_CHAT],
|
97 |
+
"Qwen2.5-7B-nerd-uncensored-v1.6.i1-Q4_K_S.gguf": ["mradermacher/Qwen2.5-7B-nerd-uncensored-v1.6-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
98 |
+
"ArliAI-RPMax-v1.3-merge-8B.Q4_K_S.gguf": ["mradermacher/ArliAI-RPMax-v1.3-merge-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
99 |
+
"MN-Violet-Lotus-12B.Q4_K_M.gguf": ["mradermacher/MN-Violet-Lotus-12B-GGUF", MessagesFormatterType.MISTRAL],
|
100 |
+
"Qwen2.5-7B-nerd-uncensored-v1.8.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-nerd-uncensored-v1.8-GGUF", MessagesFormatterType.OPEN_CHAT],
|
101 |
+
"ArliAI-RPMax-v1.3-merge-llama3-8B.i1-Q4_K_M.gguf": ["mradermacher/ArliAI-RPMax-v1.3-merge-llama3-8B-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
102 |
"Llama-3.1-SuperNova-8B-Lite_TIES_with_Base.i1-Q5_K_M.gguf": ["mradermacher/Llama-3.1-SuperNova-8B-Lite_TIES_with_Base-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
103 |
"maverick-llama3-8B.i1-Q4_K_M.gguf": ["mradermacher/maverick-llama3-8B-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
104 |
"JSL-MedLlama-3-8B-v2.0.i1-Q5_K_M.gguf": ["mradermacher/JSL-MedLlama-3-8B-v2.0-i1-GGUF", MessagesFormatterType.LLAMA_3],
|