Spaces:
Running
on
Zero
Running
on
Zero
Upload 6 files
Browse files- README.md +2 -0
- constants.py +17 -5
- llmdolphin.py +42 -0
- requirements.txt +1 -1
- utils.py +29 -15
README.md
CHANGED
@@ -9,6 +9,8 @@ app_file: app.py
|
|
9 |
license: mit
|
10 |
short_description: Text-to-Image
|
11 |
pinned: true
|
|
|
|
|
12 |
hf_oauth: true
|
13 |
---
|
14 |
|
|
|
9 |
license: mit
|
10 |
short_description: Text-to-Image
|
11 |
pinned: true
|
12 |
+
preload_from_hub:
|
13 |
+
- madebyollin/sdxl-vae-fp16-fix config.json,diffusion_pytorch_model.safetensors
|
14 |
hf_oauth: true
|
15 |
---
|
16 |
|
constants.py
CHANGED
@@ -7,10 +7,10 @@ from stablepy import (
|
|
7 |
)
|
8 |
|
9 |
# - **Download Models**
|
10 |
-
DOWNLOAD_MODEL = "https://
|
11 |
|
12 |
# - **Download VAEs**
|
13 |
-
DOWNLOAD_VAE = "https://huggingface.co/
|
14 |
|
15 |
# - **Download LoRAs**
|
16 |
DOWNLOAD_LORA = "https://huggingface.co/Leopain/color/resolve/main/Coloring_book_-_LineArt.safetensors, https://civitai.com/api/download/models/135867, https://huggingface.co/Linaqruf/anime-detailer-xl-lora/resolve/main/anime-detailer-xl.safetensors?download=true, https://huggingface.co/Linaqruf/style-enhancer-xl-lora/resolve/main/style-enhancer-xl.safetensors?download=true, https://huggingface.co/ByteDance/Hyper-SD/resolve/main/Hyper-SD15-8steps-CFG-lora.safetensors?download=true, https://huggingface.co/ByteDance/Hyper-SD/resolve/main/Hyper-SDXL-8steps-CFG-lora.safetensors?download=true"
|
@@ -24,7 +24,9 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
|
|
24 |
'John6666/lyh-anime-flux-v2a1-fp8-flux',
|
25 |
'John6666/carnival-unchained-v10-fp8-flux',
|
26 |
'Freepik/flux.1-lite-8B-alpha',
|
27 |
-
'
|
|
|
|
|
28 |
'John6666/noobai-xl-nai-xl-epsilonpred10version-sdxl',
|
29 |
'John6666/noobai-xl-nai-xl-epsilonpred075version-sdxl',
|
30 |
'John6666/noobai-xl-nai-xl-epsilonpred05version-sdxl',
|
@@ -38,6 +40,7 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
|
|
38 |
'John6666/illustriousxl-mmmix-v50-sdxl',
|
39 |
'John6666/illustrious-pencil-xl-v200-sdxl',
|
40 |
'John6666/obsession-illustriousxl-v21-sdxl',
|
|
|
41 |
'eienmojiki/Anything-XL',
|
42 |
'eienmojiki/Starry-XL-v5.2',
|
43 |
'John6666/meinaxl-v2-sdxl',
|
@@ -86,6 +89,8 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
|
|
86 |
'yodayo-ai/kivotos-xl-2.0',
|
87 |
'yodayo-ai/holodayo-xl-2.1',
|
88 |
'yodayo-ai/clandestine-xl-1.0',
|
|
|
|
|
89 |
'digiplay/majicMIX_sombre_v2',
|
90 |
'digiplay/majicMIX_realistic_v6',
|
91 |
'digiplay/majicMIX_realistic_v7',
|
@@ -120,8 +125,8 @@ DIFFUSERS_FORMAT_LORAS = [
|
|
120 |
|
121 |
DOWNLOAD_EMBEDS = [
|
122 |
'https://huggingface.co/datasets/Nerfgun3/bad_prompt/blob/main/bad_prompt_version2.pt',
|
123 |
-
'https://huggingface.co/embed/negative/resolve/main/EasyNegativeV2.safetensors',
|
124 |
-
'https://huggingface.co/embed/negative/resolve/main/bad-hands-5.pt',
|
125 |
]
|
126 |
|
127 |
CIVITAI_API_KEY = os.environ.get("CIVITAI_API_KEY")
|
@@ -291,6 +296,13 @@ MODEL_TYPE_CLASS = {
|
|
291 |
"diffusers:FluxPipeline": "FLUX",
|
292 |
}
|
293 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
294 |
POST_PROCESSING_SAMPLER = ["Use same sampler"] + [
|
295 |
name_s for name_s in scheduler_names if "Auto-Loader" not in name_s
|
296 |
]
|
|
|
7 |
)
|
8 |
|
9 |
# - **Download Models**
|
10 |
+
DOWNLOAD_MODEL = "https://huggingface.co/TechnoByte/MilkyWonderland/resolve/main/milkyWonderland_v40.safetensors"
|
11 |
|
12 |
# - **Download VAEs**
|
13 |
+
DOWNLOAD_VAE = "https://huggingface.co/fp16-guy/anything_kl-f8-anime2_vae-ft-mse-840000-ema-pruned_blessed_clearvae_fp16_cleaned/resolve/main/vae-ft-mse-840000-ema-pruned_fp16.safetensors?download=true"
|
14 |
|
15 |
# - **Download LoRAs**
|
16 |
DOWNLOAD_LORA = "https://huggingface.co/Leopain/color/resolve/main/Coloring_book_-_LineArt.safetensors, https://civitai.com/api/download/models/135867, https://huggingface.co/Linaqruf/anime-detailer-xl-lora/resolve/main/anime-detailer-xl.safetensors?download=true, https://huggingface.co/Linaqruf/style-enhancer-xl-lora/resolve/main/style-enhancer-xl.safetensors?download=true, https://huggingface.co/ByteDance/Hyper-SD/resolve/main/Hyper-SD15-8steps-CFG-lora.safetensors?download=true, https://huggingface.co/ByteDance/Hyper-SD/resolve/main/Hyper-SDXL-8steps-CFG-lora.safetensors?download=true"
|
|
|
24 |
'John6666/lyh-anime-flux-v2a1-fp8-flux',
|
25 |
'John6666/carnival-unchained-v10-fp8-flux',
|
26 |
'Freepik/flux.1-lite-8B-alpha',
|
27 |
+
'shauray/FluxDev-HyperSD-merged',
|
28 |
+
'mikeyandfriends/PixelWave_FLUX.1-dev_03',
|
29 |
+
'terminusresearch/FluxBooru-v0.3',
|
30 |
'John6666/noobai-xl-nai-xl-epsilonpred10version-sdxl',
|
31 |
'John6666/noobai-xl-nai-xl-epsilonpred075version-sdxl',
|
32 |
'John6666/noobai-xl-nai-xl-epsilonpred05version-sdxl',
|
|
|
40 |
'John6666/illustriousxl-mmmix-v50-sdxl',
|
41 |
'John6666/illustrious-pencil-xl-v200-sdxl',
|
42 |
'John6666/obsession-illustriousxl-v21-sdxl',
|
43 |
+
'John6666/obsession-illustriousxl-v30-sdxl',
|
44 |
'eienmojiki/Anything-XL',
|
45 |
'eienmojiki/Starry-XL-v5.2',
|
46 |
'John6666/meinaxl-v2-sdxl',
|
|
|
89 |
'yodayo-ai/kivotos-xl-2.0',
|
90 |
'yodayo-ai/holodayo-xl-2.1',
|
91 |
'yodayo-ai/clandestine-xl-1.0',
|
92 |
+
'ostris/OpenFLUX.1',
|
93 |
+
'shuttleai/shuttle-3-diffusion',
|
94 |
'digiplay/majicMIX_sombre_v2',
|
95 |
'digiplay/majicMIX_realistic_v6',
|
96 |
'digiplay/majicMIX_realistic_v7',
|
|
|
125 |
|
126 |
DOWNLOAD_EMBEDS = [
|
127 |
'https://huggingface.co/datasets/Nerfgun3/bad_prompt/blob/main/bad_prompt_version2.pt',
|
128 |
+
# 'https://huggingface.co/embed/negative/resolve/main/EasyNegativeV2.safetensors',
|
129 |
+
# 'https://huggingface.co/embed/negative/resolve/main/bad-hands-5.pt',
|
130 |
]
|
131 |
|
132 |
CIVITAI_API_KEY = os.environ.get("CIVITAI_API_KEY")
|
|
|
296 |
"diffusers:FluxPipeline": "FLUX",
|
297 |
}
|
298 |
|
299 |
+
DIFFUSECRAFT_CHECKPOINT_NAME = {
|
300 |
+
"sd1.5": "SD 1.5",
|
301 |
+
"sdxl": "SDXL",
|
302 |
+
"flux-dev": "FLUX",
|
303 |
+
"flux-schnell": "FLUX",
|
304 |
+
}
|
305 |
+
|
306 |
POST_PROCESSING_SAMPLER = ["Use same sampler"] + [
|
307 |
name_s for name_s in scheduler_names if "Auto-Loader" not in name_s
|
308 |
]
|
llmdolphin.py
CHANGED
@@ -27,6 +27,7 @@ llm_models = {
|
|
27 |
"mn-12b-lyra-v2a1-q5_k_m.gguf": ["HalleyStarbun/MN-12B-Lyra-v2a1-Q5_K_M-GGUF", MessagesFormatterType.CHATML],
|
28 |
"L3-8B-Tamamo-v1.i1-Q5_K_M.gguf": ["mradermacher/L3-8B-Tamamo-v1-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
29 |
"MN-Chinofun-12B-2.i1-Q4_K_M.gguf": ["mradermacher/MN-Chinofun-12B-2-i1-GGUF", MessagesFormatterType.MISTRAL],
|
|
|
30 |
"Mahou-1.5-mistral-nemo-12B.i1-Q4_K_M.gguf": ["mradermacher/Mahou-1.5-mistral-nemo-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
31 |
"MN-12B-Mag-Mell-Q4_K_M.gguf": ["inflatebot/MN-12B-Mag-Mell-R1-GGUF", MessagesFormatterType.MISTRAL],
|
32 |
"Qwen-modelstock-15B.i1-Q4_K_M.gguf": ["mradermacher/Qwen-modelstock-15B-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
@@ -51,6 +52,7 @@ llm_models = {
|
|
51 |
"Nemo-12B-Marlin-v7.Q4_K_M.gguf": ["mradermacher/Nemo-12B-Marlin-v7-GGUF", MessagesFormatterType.MISTRAL],
|
52 |
"Nemo-12B-Marlin-v8.Q4_K_S.gguf": ["mradermacher/Nemo-12B-Marlin-v8-GGUF", MessagesFormatterType.MISTRAL],
|
53 |
"NemoDori-v0.2-Upscaled.1-14B.Q4_K_M.gguf": ["mradermacher/NemoDori-v0.2-Upscaled.1-14B-GGUF", MessagesFormatterType.MISTRAL],
|
|
|
54 |
"Fireball-12B-v1.0.i1-Q4_K_M.gguf": ["mradermacher/Fireball-12B-v1.0-i1-GGUF", MessagesFormatterType.MISTRAL],
|
55 |
"Fireball-Mistral-Nemo-Base-2407-sft-v2.2a.Q4_K_M.gguf": ["mradermacher/Fireball-Mistral-Nemo-Base-2407-sft-v2.2a-GGUF", MessagesFormatterType.MISTRAL],
|
56 |
"T-III-12B.Q4_K_M.gguf": ["mradermacher/T-III-12B-GGUF", MessagesFormatterType.CHATML],
|
@@ -70,12 +72,52 @@ llm_models = {
|
|
70 |
"Rocinante-12B-v2h-Q4_K_M.gguf": ["BeaverAI/Rocinante-12B-v2h-GGUF", MessagesFormatterType.MISTRAL],
|
71 |
"Mistral-Nemo-12B-ArliAI-RPMax-v1.1.i1-Q4_K_M.gguf": ["mradermacher/Mistral-Nemo-12B-ArliAI-RPMax-v1.1-i1-GGUF", MessagesFormatterType.MISTRAL],
|
72 |
"Pans_Gutenbergum_V0.1.Q4_K_M.gguf": ["mradermacher/Pans_Gutenbergum_V0.1-GGUF", MessagesFormatterType.MISTRAL],
|
|
|
73 |
"Trinas_Nectar-8B-model_stock.i1-Q4_K_M.gguf": ["mradermacher/Trinas_Nectar-8B-model_stock-i1-GGUF", MessagesFormatterType.MISTRAL],
|
74 |
"ChatWaifu_Magnum_V0.2.Q4_K_M.gguf": ["mradermacher/ChatWaifu_Magnum_V0.2-GGUF", MessagesFormatterType.MISTRAL],
|
75 |
"ChatWaifu_12B_v2.0.Q5_K_M.gguf": ["mradermacher/ChatWaifu_12B_v2.0-GGUF", MessagesFormatterType.MISTRAL],
|
76 |
"ChatWaifu_22B_v2.0_preview.Q4_K_S.gguf": ["mradermacher/ChatWaifu_22B_v2.0_preview-GGUF", MessagesFormatterType.MISTRAL],
|
77 |
"ChatWaifu_v1.4.Q5_K_M.gguf": ["mradermacher/ChatWaifu_v1.4-GGUF", MessagesFormatterType.MISTRAL],
|
78 |
"ChatWaifu_v1.3.1.Q4_K_M.gguf": ["mradermacher/ChatWaifu_v1.3.1-GGUF", MessagesFormatterType.MISTRAL],
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
79 |
"dolphin-2.6-mistral-7b-dpo-laser.Q4_K_S.gguf": ["mradermacher/dolphin-2.6-mistral-7b-dpo-laser-GGUF", MessagesFormatterType.MISTRAL],
|
80 |
"Flowable-Docs-Llama-3.1-8B.Q5_K_M.gguf": ["mradermacher/Flowable-Docs-Llama-3.1-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
81 |
"slimorca-gemma2-9b-fft.Q4_K_M.gguf": ["mradermacher/slimorca-gemma2-9b-fft-GGUF", MessagesFormatterType.ALPACA],
|
|
|
27 |
"mn-12b-lyra-v2a1-q5_k_m.gguf": ["HalleyStarbun/MN-12B-Lyra-v2a1-Q5_K_M-GGUF", MessagesFormatterType.CHATML],
|
28 |
"L3-8B-Tamamo-v1.i1-Q5_K_M.gguf": ["mradermacher/L3-8B-Tamamo-v1-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
29 |
"MN-Chinofun-12B-2.i1-Q4_K_M.gguf": ["mradermacher/MN-Chinofun-12B-2-i1-GGUF", MessagesFormatterType.MISTRAL],
|
30 |
+
"Mistral-Nemo-Prism-12B-v2.i1-Q4_K_M.gguf": ["mradermacher/Mistral-Nemo-Prism-12B-v2-i1-GGUF", MessagesFormatterType.MISTRAL],
|
31 |
"Mahou-1.5-mistral-nemo-12B.i1-Q4_K_M.gguf": ["mradermacher/Mahou-1.5-mistral-nemo-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
32 |
"MN-12B-Mag-Mell-Q4_K_M.gguf": ["inflatebot/MN-12B-Mag-Mell-R1-GGUF", MessagesFormatterType.MISTRAL],
|
33 |
"Qwen-modelstock-15B.i1-Q4_K_M.gguf": ["mradermacher/Qwen-modelstock-15B-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
|
|
52 |
"Nemo-12B-Marlin-v7.Q4_K_M.gguf": ["mradermacher/Nemo-12B-Marlin-v7-GGUF", MessagesFormatterType.MISTRAL],
|
53 |
"Nemo-12B-Marlin-v8.Q4_K_S.gguf": ["mradermacher/Nemo-12B-Marlin-v8-GGUF", MessagesFormatterType.MISTRAL],
|
54 |
"NemoDori-v0.2-Upscaled.1-14B.Q4_K_M.gguf": ["mradermacher/NemoDori-v0.2-Upscaled.1-14B-GGUF", MessagesFormatterType.MISTRAL],
|
55 |
+
"MT-Gen2-GIMMMA-gemma-2-9B.Q4_K_M.gguf": ["mradermacher/MT-Gen2-GIMMMA-gemma-2-9B-GGUF", MessagesFormatterType.ALPACA],
|
56 |
"Fireball-12B-v1.0.i1-Q4_K_M.gguf": ["mradermacher/Fireball-12B-v1.0-i1-GGUF", MessagesFormatterType.MISTRAL],
|
57 |
"Fireball-Mistral-Nemo-Base-2407-sft-v2.2a.Q4_K_M.gguf": ["mradermacher/Fireball-Mistral-Nemo-Base-2407-sft-v2.2a-GGUF", MessagesFormatterType.MISTRAL],
|
58 |
"T-III-12B.Q4_K_M.gguf": ["mradermacher/T-III-12B-GGUF", MessagesFormatterType.CHATML],
|
|
|
72 |
"Rocinante-12B-v2h-Q4_K_M.gguf": ["BeaverAI/Rocinante-12B-v2h-GGUF", MessagesFormatterType.MISTRAL],
|
73 |
"Mistral-Nemo-12B-ArliAI-RPMax-v1.1.i1-Q4_K_M.gguf": ["mradermacher/Mistral-Nemo-12B-ArliAI-RPMax-v1.1-i1-GGUF", MessagesFormatterType.MISTRAL],
|
74 |
"Pans_Gutenbergum_V0.1.Q4_K_M.gguf": ["mradermacher/Pans_Gutenbergum_V0.1-GGUF", MessagesFormatterType.MISTRAL],
|
75 |
+
"ChronoStar-Unleashed-v0.1.i1-Q4_K_M.gguf": ["mradermacher/ChronoStar-Unleashed-v0.1-i1-GGUF", MessagesFormatterType.MISTRAL],
|
76 |
"Trinas_Nectar-8B-model_stock.i1-Q4_K_M.gguf": ["mradermacher/Trinas_Nectar-8B-model_stock-i1-GGUF", MessagesFormatterType.MISTRAL],
|
77 |
"ChatWaifu_Magnum_V0.2.Q4_K_M.gguf": ["mradermacher/ChatWaifu_Magnum_V0.2-GGUF", MessagesFormatterType.MISTRAL],
|
78 |
"ChatWaifu_12B_v2.0.Q5_K_M.gguf": ["mradermacher/ChatWaifu_12B_v2.0-GGUF", MessagesFormatterType.MISTRAL],
|
79 |
"ChatWaifu_22B_v2.0_preview.Q4_K_S.gguf": ["mradermacher/ChatWaifu_22B_v2.0_preview-GGUF", MessagesFormatterType.MISTRAL],
|
80 |
"ChatWaifu_v1.4.Q5_K_M.gguf": ["mradermacher/ChatWaifu_v1.4-GGUF", MessagesFormatterType.MISTRAL],
|
81 |
"ChatWaifu_v1.3.1.Q4_K_M.gguf": ["mradermacher/ChatWaifu_v1.3.1-GGUF", MessagesFormatterType.MISTRAL],
|
82 |
+
"magnum-12b-v2.5-kto.i1-Q4_K_M.gguf": ["mradermacher/magnum-12b-v2.5-kto-i1-GGUF", MessagesFormatterType.CHATML],
|
83 |
+
"L3.1-8B-Slush.i1-Q5_K_M.gguf": ["mradermacher/L3.1-8B-Slush-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
84 |
+
"QandoraExp-7B-Persona.i1-Q5_K_M.gguf": ["mradermacher/QandoraExp-7B-Persona-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
85 |
+
"Pantheon-RP-1.6-12b-Nemo-KTO.i1-Q4_K_M.gguf": ["mradermacher/Pantheon-RP-1.6-12b-Nemo-KTO-i1-GGUF", MessagesFormatterType.CHATML],
|
86 |
+
"Unaligned-Base-8b-1024K.i1-Q5_K_M.gguf": ["mradermacher/Unaligned-Base-8b-1024K-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
87 |
+
"Mistral-Nemo-Prism-12B-v6.Q4_K_M.gguf": ["mradermacher/Mistral-Nemo-Prism-12B-v6-GGUF", MessagesFormatterType.MISTRAL],
|
88 |
+
"hermes-llama3-roleplay-1000-v3.Q5_K_M.gguf": ["mradermacher/hermes-llama3-roleplay-1000-v3-GGUF", MessagesFormatterType.LLAMA_3],
|
89 |
+
"Prismatic-12b.Q4_K_M.gguf": ["mradermacher/Prismatic-12b-GGUF", MessagesFormatterType.MISTRAL],
|
90 |
+
"Qwen-2.5-Aether-SlerpFusion-7B.i1-Q5_K_M.gguf": ["mradermacher/Qwen-2.5-Aether-SlerpFusion-7B-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
91 |
+
"miscii-14b-1028.i1-Q4_K_M.gguf": ["mradermacher/miscii-14b-1028-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
92 |
+
"DarkAtom-12B-v3.i1-Q4_K_M.gguf": ["mradermacher/DarkAtom-12B-v3-i1-GGUF", MessagesFormatterType.MISTRAL],
|
93 |
+
"Mistral-Nemo-Prism-12B.i1-Q4_K_M.gguf": ["mradermacher/Mistral-Nemo-Prism-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
94 |
+
"QandoraExp-7B-v2.Q5_K_M.gguf": ["mradermacher/QandoraExp-7B-v2-GGUF", MessagesFormatterType.OPEN_CHAT],
|
95 |
+
"Kunocchini-7b-128k-test.Q5_K_M.gguf": ["mradermacher/Kunocchini-7b-128k-test-GGUF", MessagesFormatterType.MISTRAL],
|
96 |
+
"MT2-Gen2-BGMAMU-gemma-2-9B.Q4_K_M.gguf": ["mradermacher/MT2-Gen2-BGMAMU-gemma-2-9B-GGUF", MessagesFormatterType.ALPACA],
|
97 |
+
"dolphin-mixtral-2x7b.i1-Q4_K_M.gguf": ["mradermacher/dolphin-mixtral-2x7b-i1-GGUF", MessagesFormatterType.MISTRAL],
|
98 |
+
"NightyGurps-14b-v1.1.Q4_K_M.gguf": ["mradermacher/NightyGurps-14b-v1.1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
99 |
+
"MT2-Gen2-BB-gemma-2-MTMMT5-9B.Q4_K_M.gguf": ["mradermacher/MT2-Gen2-BB-gemma-2-MTMMT5-9B-GGUF", MessagesFormatterType.ALPACA],
|
100 |
+
"MT2-Gen2-IMM-gemma-2-9B.Q4_K_M.gguf": ["mradermacher/MT2-Gen2-IMM-gemma-2-9B-GGUF", MessagesFormatterType.ALPACA],
|
101 |
+
"Unaligned-RP-Base-8b-3.Q5_K_M.gguf": ["mradermacher/Unaligned-RP-Base-8b-3-GGUF", MessagesFormatterType.LLAMA_3],
|
102 |
+
"Unaligned-RP-Base-8b-2.Q5_K_M.gguf": ["mradermacher/Unaligned-RP-Base-8b-2-GGUF", MessagesFormatterType.LLAMA_3],
|
103 |
+
"LongWriter-Qwen2.5-7B-Instruct.i1-Q5_K_M.gguf": ["mradermacher/LongWriter-Qwen2.5-7B-Instruct-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
104 |
+
"Everyone-LLM-7b-Base.i1-Q5_K_M.gguf": ["mradermacher/Everyone-LLM-7b-Base-i1-GGUF", MessagesFormatterType.MISTRAL],
|
105 |
+
"QwenSlerp8-7B.Q5_K_M.gguf": ["mradermacher/QwenSlerp8-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
106 |
+
"Tess-10.7B-v1.5.Q4_K_M.gguf": ["mradermacher/Tess-10.7B-v1.5-GGUF", MessagesFormatterType.LLAMA_3],
|
107 |
+
"Rombos-Coder-V2.5-Qwen-7b.i1-Q5_K_M.gguf": ["mradermacher/Rombos-Coder-V2.5-Qwen-7b-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
108 |
+
"Qwen2.5-Gutenberg-Doppel-14B.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Gutenberg-Doppel-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
109 |
+
"MT-Gen2-MU-gemma-2-MT1RAv0.1t0.25-9B.Q4_K_M.gguf": ["mradermacher/MT-Gen2-MU-gemma-2-MT1RAv0.1t0.25-9B-GGUF", MessagesFormatterType.ALPACA],
|
110 |
+
"MT1-Gen2-GP-gemma-2-MT1DMv1-9B.Q4_K_M.gguf": ["mradermacher/MT1-Gen2-GP-gemma-2-MT1DMv1-9B-GGUF", MessagesFormatterType.ALPACA],
|
111 |
+
"MT1-Gen2-GMA-gemma-2-9B.Q4_K_M.gguf": ["mradermacher/MT1-Gen2-GMA-gemma-2-9B-GGUF", MessagesFormatterType.ALPACA],
|
112 |
+
"MT1-Gen2-MMMU-gemma-2-9B.Q4_K_M.gguf": ["mradermacher/MT1-Gen2-MMMU-gemma-2-9B-GGUF", MessagesFormatterType.ALPACA],
|
113 |
+
"LGRC-7B-slerp.Q5_K_M.gguf": ["mradermacher/LGRC-7B-slerp-GGUF", MessagesFormatterType.MISTRAL],
|
114 |
+
"Ice0.32-10.11-RP.Q5_K_M.gguf": ["mradermacher/Ice0.32-10.11-RP-GGUF", MessagesFormatterType.MISTRAL],
|
115 |
+
"Vecteus-v1.i1-Q5_K_M.gguf": ["mradermacher/Vecteus-v1-i1-GGUF", MessagesFormatterType.MISTRAL],
|
116 |
+
"L3.1-BaeZel-8B-Della.Q4_K_S.gguf": ["mradermacher/L3.1-BaeZel-8B-Della-GGUF", MessagesFormatterType.LLAMA_3],
|
117 |
+
"Qwen2.5-14B-Instruct-SEALONG.i1-Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Instruct-SEALONG-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
118 |
+
"Qwen2.5-7B-Instruct-SEALONG.i1-Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-SEALONG-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
119 |
+
"Fraken-Maid-TW-K-Slerp.i1-Q5_K_M.gguf": ["mradermacher/Fraken-Maid-TW-K-Slerp-i1-GGUF", MessagesFormatterType.MISTRAL],
|
120 |
+
"AutoTrain-Qwen-Rui-Elite.Q5_K_M.gguf": ["mradermacher/AutoTrain-Qwen-Rui-Elite-GGUF", MessagesFormatterType.OPEN_CHAT],
|
121 |
"dolphin-2.6-mistral-7b-dpo-laser.Q4_K_S.gguf": ["mradermacher/dolphin-2.6-mistral-7b-dpo-laser-GGUF", MessagesFormatterType.MISTRAL],
|
122 |
"Flowable-Docs-Llama-3.1-8B.Q5_K_M.gguf": ["mradermacher/Flowable-Docs-Llama-3.1-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
123 |
"slimorca-gemma2-9b-fft.Q4_K_M.gguf": ["mradermacher/slimorca-gemma2-9b-fft-GGUF", MessagesFormatterType.ALPACA],
|
requirements.txt
CHANGED
@@ -4,7 +4,7 @@ diffusers
|
|
4 |
invisible_watermark
|
5 |
transformers
|
6 |
xformers
|
7 |
-
git+https://github.com/R3gm/stablepy.git@
|
8 |
torch==2.2.0
|
9 |
gdown
|
10 |
opencv-python
|
|
|
4 |
invisible_watermark
|
5 |
transformers
|
6 |
xformers
|
7 |
+
git+https://github.com/R3gm/stablepy.git@f686946 # -b refactor_sampler_fix
|
8 |
torch==2.2.0
|
9 |
gdown
|
10 |
opencv-python
|
utils.py
CHANGED
@@ -7,11 +7,14 @@ from constants import (
|
|
7 |
HF_TOKEN,
|
8 |
MODEL_TYPE_CLASS,
|
9 |
DIRECTORY_LORAS,
|
|
|
10 |
)
|
11 |
from huggingface_hub import HfApi
|
|
|
12 |
from diffusers import DiffusionPipeline
|
13 |
from huggingface_hub import model_info as model_info_data
|
14 |
from diffusers.pipelines.pipeline_loading_utils import variant_compatible_siblings
|
|
|
15 |
from pathlib import PosixPath
|
16 |
from unidecode import unidecode
|
17 |
import urllib.parse
|
@@ -283,10 +286,15 @@ def get_model_type(repo_id: str):
|
|
283 |
api = HfApi(token=os.environ.get("HF_TOKEN")) # if use private or gated model
|
284 |
default = "SD 1.5"
|
285 |
try:
|
286 |
-
|
287 |
-
|
288 |
-
|
289 |
-
|
|
|
|
|
|
|
|
|
|
|
290 |
except Exception:
|
291 |
return default
|
292 |
return default
|
@@ -371,17 +379,23 @@ def download_diffuser_repo(repo_name: str, model_type: str, revision: str = "mai
|
|
371 |
if len(variant_filenames):
|
372 |
variant = "fp16"
|
373 |
|
374 |
-
|
375 |
-
|
376 |
-
|
377 |
-
|
378 |
-
|
379 |
-
|
380 |
-
|
381 |
-
|
382 |
-
|
383 |
-
|
384 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
385 |
|
386 |
if isinstance(cached_folder, PosixPath):
|
387 |
cached_folder = cached_folder.as_posix()
|
|
|
7 |
HF_TOKEN,
|
8 |
MODEL_TYPE_CLASS,
|
9 |
DIRECTORY_LORAS,
|
10 |
+
DIFFUSECRAFT_CHECKPOINT_NAME,
|
11 |
)
|
12 |
from huggingface_hub import HfApi
|
13 |
+
from huggingface_hub import snapshot_download
|
14 |
from diffusers import DiffusionPipeline
|
15 |
from huggingface_hub import model_info as model_info_data
|
16 |
from diffusers.pipelines.pipeline_loading_utils import variant_compatible_siblings
|
17 |
+
from stablepy.diffusers_vanilla.utils import checkpoint_model_type
|
18 |
from pathlib import PosixPath
|
19 |
from unidecode import unidecode
|
20 |
import urllib.parse
|
|
|
286 |
api = HfApi(token=os.environ.get("HF_TOKEN")) # if use private or gated model
|
287 |
default = "SD 1.5"
|
288 |
try:
|
289 |
+
if os.path.exists(repo_id):
|
290 |
+
tag = checkpoint_model_type(repo_id)
|
291 |
+
return DIFFUSECRAFT_CHECKPOINT_NAME[tag]
|
292 |
+
else:
|
293 |
+
model = api.model_info(repo_id=repo_id, timeout=5.0)
|
294 |
+
tags = model.tags
|
295 |
+
for tag in tags:
|
296 |
+
if tag in MODEL_TYPE_CLASS.keys(): return MODEL_TYPE_CLASS.get(tag, default)
|
297 |
+
|
298 |
except Exception:
|
299 |
return default
|
300 |
return default
|
|
|
379 |
if len(variant_filenames):
|
380 |
variant = "fp16"
|
381 |
|
382 |
+
if model_type == "FLUX":
|
383 |
+
cached_folder = snapshot_download(
|
384 |
+
repo_id=repo_name,
|
385 |
+
allow_patterns="transformer/*"
|
386 |
+
)
|
387 |
+
else:
|
388 |
+
cached_folder = DiffusionPipeline.download(
|
389 |
+
pretrained_model_name=repo_name,
|
390 |
+
force_download=False,
|
391 |
+
token=token,
|
392 |
+
revision=revision,
|
393 |
+
# mirror="https://hf-mirror.com",
|
394 |
+
variant=variant,
|
395 |
+
use_safetensors=True,
|
396 |
+
trust_remote_code=False,
|
397 |
+
timeout=5.0,
|
398 |
+
)
|
399 |
|
400 |
if isinstance(cached_folder, PosixPath):
|
401 |
cached_folder = cached_folder.as_posix()
|