John6666 commited on
Commit
2923686
1 Parent(s): 6f2a033

Upload 3 files

Browse files
Files changed (3) hide show
  1. env.py +2 -2
  2. llmdolphin.py +85 -6
  3. modutils.py +28 -24
env.py CHANGED
@@ -71,8 +71,8 @@ load_diffusers_format_model = [
71
  'KBlueLeaf/Kohaku-XL-Epsilon-rev2',
72
  'KBlueLeaf/Kohaku-XL-Epsilon-rev3',
73
  'KBlueLeaf/Kohaku-XL-Zeta',
74
- 'kayfahaarukku/UrangDiffusion-1.2',
75
- 'Eugeoter/artiwaifu-diffusion-1.0',
76
  'Raelina/Rae-Diffusion-XL-V2',
77
  'Raelina/Raemu-XL-V4',
78
  ]
 
71
  'KBlueLeaf/Kohaku-XL-Epsilon-rev2',
72
  'KBlueLeaf/Kohaku-XL-Epsilon-rev3',
73
  'KBlueLeaf/Kohaku-XL-Zeta',
74
+ 'kayfahaarukku/UrangDiffusion-1.3',
75
+ 'Eugeoter/artiwaifu-diffusion-2.0',
76
  'Raelina/Rae-Diffusion-XL-V2',
77
  'Raelina/Raemu-XL-V4',
78
  ]
llmdolphin.py CHANGED
@@ -25,6 +25,7 @@ llm_models = {
25
  "MN-12B-Starcannon-v3.i1-Q4_K_M.gguf": ["mradermacher/MN-12B-Starcannon-v3-i1-GGUF", MessagesFormatterType.CHATML],
26
  "MN-12B-Starcannon-v4-unofficial.i1-Q4_K_M.gguf": ["mradermacher/MN-12B-Starcannon-v4-unofficial-i1-GGUF", MessagesFormatterType.MISTRAL],
27
  "MN-12B-Starsong-v1.i1-Q4_K_M.gguf": ["mradermacher/MN-12B-Starsong-v1-i1-GGUF", MessagesFormatterType.CHATML],
 
28
  "Lumimaid-Magnum-12B.i1-Q4_K_M.gguf": ["mradermacher/Lumimaid-Magnum-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
29
  "Nemo-12B-Marlin-v1.i1-Q4_K_M.gguf": ["mradermacher/Nemo-12B-Marlin-v1-i1-GGUF", MessagesFormatterType.MISTRAL],
30
  "Nemo-12B-Marlin-v2.i1-Q4_K_M.gguf": ["mradermacher/Nemo-12B-Marlin-v2-i1-GGUF", MessagesFormatterType.MISTRAL],
@@ -38,10 +39,85 @@ llm_models = {
38
  "Fireball-Mistral-Nemo-Base-2407-sft-v2.2a.Q4_K_M.gguf": ["mradermacher/Fireball-Mistral-Nemo-Base-2407-sft-v2.2a-GGUF", MessagesFormatterType.MISTRAL],
39
  "T-III-12B.Q4_K_M.gguf": ["mradermacher/T-III-12B-GGUF", MessagesFormatterType.CHATML],
40
  "T-IIIa-12B.Q4_K_S.gguf": ["mradermacher/T-IIIa-12B-GGUF", MessagesFormatterType.MISTRAL],
 
41
  "StorieCreative.i1-Q4_K_S.gguf": ["mradermacher/StorieCreative-i1-GGUF", MessagesFormatterType.MISTRAL],
42
  "Deutscher-Pantheon-12B.Q4_K_M.gguf": ["mradermacher/Deutscher-Pantheon-12B-GGUF", MessagesFormatterType.MISTRAL],
43
  "guns-and-roses-r1-Q4_K_L-imat.gguf": ["Reiterate3680/guns-and-roses-r1-GGUF", MessagesFormatterType.MISTRAL],
 
 
44
  "Trinas_Nectar-8B-model_stock.i1-Q4_K_M.gguf": ["mradermacher/Trinas_Nectar-8B-model_stock-i1-GGUF", MessagesFormatterType.MISTRAL],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
45
  "nemo-12b-hiwaifu-Q4_K_L-imat.gguf": ["Reiterate3680/nemo-12b-hiwaifu-GGUF", MessagesFormatterType.MISTRAL],
46
  "Soliloquy-7B-v3-Q4_K_L-imat.gguf": ["Reiterate3680/Soliloquy-7B-v3-GGUF", MessagesFormatterType.OPEN_CHAT],
47
  "Lyra-Gutenberg-mistral-nemo-12B.Q4_K_M.gguf": ["mradermacher/Lyra-Gutenberg-mistral-nemo-12B-GGUF", MessagesFormatterType.MISTRAL],
@@ -314,7 +390,7 @@ llm_formats = {
314
  "Qwen2": MessagesFormatterType.OPEN_CHAT,
315
  }
316
  # https://github.com/Maximilian-Winter/llama-cpp-agent
317
- llm_languages = ["English", "Japanese", "Chinese"]
318
  llm_models_tupled_list = []
319
  default_llm_model_filename = list(llm_models.keys())[0]
320
  override_llm_format = None
@@ -380,7 +456,8 @@ def download_llm_model(filename):
380
  if not filename in llm_models.keys(): return default_llm_model_filename
381
  try:
382
  hf_hub_download(repo_id = llm_models[filename][0], filename = filename, local_dir = llm_models_dir)
383
- except Exception:
 
384
  return default_llm_model_filename
385
  update_llm_model_tupled_list()
386
  return filename
@@ -410,7 +487,6 @@ def select_dolphin_format(format_name):
410
  return gr.update(value=format_name)
411
 
412
 
413
- #download_llm_models()
414
  download_llm_model(default_llm_model_filename)
415
 
416
 
@@ -460,11 +536,14 @@ def add_dolphin_models(query, format_name):
460
  if not api.repo_exists(repo_id = repo) or not api.file_exists(repo_id = repo, filename = filename): return gr.update(visible=True)
461
  add_models[filename] = [repo, format]
462
  else: return gr.update(visible=True)
463
- except Exception:
 
464
  return gr.update(visible=True)
465
- print(add_models)
466
  llm_models = (llm_models | add_models).copy()
467
- return gr.update(choices=get_dolphin_models())
 
 
468
 
469
 
470
  dolphin_output_language = "English"
 
25
  "MN-12B-Starcannon-v3.i1-Q4_K_M.gguf": ["mradermacher/MN-12B-Starcannon-v3-i1-GGUF", MessagesFormatterType.CHATML],
26
  "MN-12B-Starcannon-v4-unofficial.i1-Q4_K_M.gguf": ["mradermacher/MN-12B-Starcannon-v4-unofficial-i1-GGUF", MessagesFormatterType.MISTRAL],
27
  "MN-12B-Starsong-v1.i1-Q4_K_M.gguf": ["mradermacher/MN-12B-Starsong-v1-i1-GGUF", MessagesFormatterType.CHATML],
28
+ "StarDust-12b-v1-Q4_K_M.gguf": ["Luni/StarDust-12b-v1-GGUF", MessagesFormatterType.MISTRAL],
29
  "Lumimaid-Magnum-12B.i1-Q4_K_M.gguf": ["mradermacher/Lumimaid-Magnum-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
30
  "Nemo-12B-Marlin-v1.i1-Q4_K_M.gguf": ["mradermacher/Nemo-12B-Marlin-v1-i1-GGUF", MessagesFormatterType.MISTRAL],
31
  "Nemo-12B-Marlin-v2.i1-Q4_K_M.gguf": ["mradermacher/Nemo-12B-Marlin-v2-i1-GGUF", MessagesFormatterType.MISTRAL],
 
39
  "Fireball-Mistral-Nemo-Base-2407-sft-v2.2a.Q4_K_M.gguf": ["mradermacher/Fireball-Mistral-Nemo-Base-2407-sft-v2.2a-GGUF", MessagesFormatterType.MISTRAL],
40
  "T-III-12B.Q4_K_M.gguf": ["mradermacher/T-III-12B-GGUF", MessagesFormatterType.CHATML],
41
  "T-IIIa-12B.Q4_K_S.gguf": ["mradermacher/T-IIIa-12B-GGUF", MessagesFormatterType.MISTRAL],
42
+ "Type-IV-Last-12B.Q4_K_M.gguf": ["mradermacher/Type-IV-Last-12B-GGUF", MessagesFormatterType.CHATML],
43
  "StorieCreative.i1-Q4_K_S.gguf": ["mradermacher/StorieCreative-i1-GGUF", MessagesFormatterType.MISTRAL],
44
  "Deutscher-Pantheon-12B.Q4_K_M.gguf": ["mradermacher/Deutscher-Pantheon-12B-GGUF", MessagesFormatterType.MISTRAL],
45
  "guns-and-roses-r1-Q4_K_L-imat.gguf": ["Reiterate3680/guns-and-roses-r1-GGUF", MessagesFormatterType.MISTRAL],
46
+ "MagnumChronos.Q4_K_M.gguf": ["mradermacher/MagnumChronos-GGUF", MessagesFormatterType.CHATML],
47
+ "StarDust-12b-v2.i1-Q5_K_M.gguf": ["mradermacher/StarDust-12b-v2-i1-GGUF", MessagesFormatterType.CHATML],
48
  "Trinas_Nectar-8B-model_stock.i1-Q4_K_M.gguf": ["mradermacher/Trinas_Nectar-8B-model_stock-i1-GGUF", MessagesFormatterType.MISTRAL],
49
+ "ChatWaifu_v1.3.1.Q4_K_M.gguf": ["mradermacher/ChatWaifu_v1.3.1-GGUF", MessagesFormatterType.MISTRAL],
50
+ "Lumimaid-Rocinate-v0.1.Q4_K_M.gguf": ["mradermacher/Lumimaid-Rocinate-v0.1-GGUF", MessagesFormatterType.MISTRAL],
51
+ "alfacentaury11b-q5_k_m.gguf": ["ClaudioItaly/AlfaCentaury11B-Q5_K_M-GGUF", MessagesFormatterType.MISTRAL],
52
+ "lit_red_storm-q4_k_m.gguf": ["Daemontatox/lit_red_storm-Q4_K_M-GGUF", MessagesFormatterType.LLAMA_3],
53
+ "llama-3-youko-8b-instruct.Q5_K_M.gguf": ["mradermacher/llama-3-youko-8b-instruct-GGUF", MessagesFormatterType.LLAMA_3],
54
+ "Fimbulvetr-11B-v2.Q5_K_M.gguf": ["QuantFactory/Fimbulvetr-11B-v2-GGUF", MessagesFormatterType.VICUNA],
55
+ "ogha-8b-model_stock-q4_k_m.gguf": ["DreadPoor/Ogha-8B-model_stock-Q4_K_M-GGUF", MessagesFormatterType.LLAMA_3],
56
+ "westicelemontearp-32k-7b-q5_k_s-imat.gguf": ["woofwolfy/WestIceLemonTeaRP-32k-7b-Q5_K_S-GGUF-Imatrix", MessagesFormatterType.ALPACA],
57
+ "ArliAI-RPMax-12B-v1.1-Q4_K_M.gguf": ["ArliAI/ArliAI-RPMax-12B-v1.1-GGUF", MessagesFormatterType.MISTRAL],
58
+ "l3-8b-serpentine-q4_k_m-imat.gguf": ["woofwolfy/L3-8B-Serpentine-Q4_K_M-GGUF", MessagesFormatterType.MISTRAL],
59
+ "llama-3-alpha-centauri-v0.1-q4_k_m-imat.gguf": ["woofwolfy/Llama-3-Alpha-Centauri-v0.1-Q4_K_M-GGUF-Imatrix", MessagesFormatterType.LLAMA_3],
60
+ "OpenCrystal-15B-L3-v2.i1-Q4_K_M.gguf": ["mradermacher/OpenCrystal-15B-L3-v2-i1-GGUF", MessagesFormatterType.LLAMA_3],
61
+ "Gukbap-Mistral-7B.i1-Q4_K_M.gguf": ["mradermacher/Gukbap-Mistral-7B-i1-GGUF", MessagesFormatterType.MISTRAL],
62
+ "Gukbap-Qwen2-7B.i1-Q4_K_M.gguf": ["mradermacher/Gukbap-Qwen2-7B-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
63
+ "Irina-8B-model_stock.Q5_K_M.gguf": ["mradermacher/Irina-8B-model_stock-GGUF", MessagesFormatterType.MISTRAL],
64
+ "silicon-alice-7b-q5_k_m-imat.gguf": ["woofwolfy/Silicon-Alice-7B-Q5_K_M-GGUF-Imatrix", MessagesFormatterType.MISTRAL],
65
+ "l3.1-8b-rpgrammax-q5_k_m.gguf": ["djuna/L3.1-8B-RPGramMax-Q5_K_M-GGUF", MessagesFormatterType.LLAMA_3],
66
+ "Instant-RP-Noodles-12B-v1.2.Q4_K_M.gguf": ["mradermacher/Instant-RP-Noodles-12B-v1.2-GGUF", MessagesFormatterType.MISTRAL],
67
+ "Ellaria-9B-Q4_K_M.gguf": ["bartowski/Ellaria-9B-GGUF", MessagesFormatterType.ALPACA],
68
+ "evolutionstory-7b-v2.2-q5_k_m.gguf": ["ClaudioItaly/Evolutionstory-7B-v2.2-Q5_K_M-GGUF", MessagesFormatterType.MISTRAL],
69
+ "NeuroNoise_v2.Q6_K.gguf": ["mradermacher/NeuroNoise_v2-GGUF", MessagesFormatterType.CHATML],
70
+ "daybreak-kunoichi-2dpo-7b.i1-Q5_K_M.gguf": ["mradermacher/daybreak-kunoichi-2dpo-7b-i1-GGUF", MessagesFormatterType.MISTRAL],
71
+ "l3.1-niitorm-8b-latcosx2-q8_0.gguf": ["v000000/L3.1-Niitorm-8B-LATCOSx2-Version-Q8_0-GGUF", MessagesFormatterType.MISTRAL],
72
+ "IceTea21EnergyDrinkRPV13-dpo240.Q4_K_M.gguf": ["mradermacher/IceTea21EnergyDrinkRPV13-dpo240-GGUF", MessagesFormatterType.MISTRAL],
73
+ "topneuralstory-q6_k.gguf": ["ClaudioItaly/TopNeuralStory-Q6_K-GGUF", MessagesFormatterType.MISTRAL],
74
+ #"L3.1-Ablaze-Vulca-v0.2h-8B.q8_0.gguf": ["kromquant/L3.1-Ablaze-Vulca-v0.2h-8B-GGUFs", MessagesFormatterType.MISTRAL],
75
+ "Fireball-12B-v1.13a-philosophers.Q4_K_M.gguf": ["mradermacher/Fireball-12B-v1.13a-philosophers-GGUF", MessagesFormatterType.MISTRAL],
76
+ "l3.1-niitorm-8b-q8_0.gguf": ["v000000/L3.1-Niitorm-8B-t0.0001-Q8_0-GGUF", MessagesFormatterType.MISTRAL],
77
+ "mn-12b-lyra-v3-q5_k_m.gguf": ["HalleyStarbun/MN-12B-Lyra-v3-Q5_K_M-GGUF", MessagesFormatterType.CHATML],
78
+ "mergekit-model_stock-ffibbcs-q4_k_s.gguf": ["DarwinAnim8or/mergekit-model_stock-ffibbcs-Q4_K_S-GGUF", MessagesFormatterType.ALPACA],
79
+ "experiment-wip-being-q4_k_m.gguf": ["DreadPoor/EXPERIMENT-WIP-BEING-Q4_K_M-GGUF", MessagesFormatterType.LLAMA_3],
80
+ "Sunmoy-9B-G2.Q4_K_M.gguf": ["backyardai/Sunmoy-9B-G2-GGUF", MessagesFormatterType.ALPACA],
81
+ "kainaticulous-rp-7b.Q5_K_M.gguf": ["mradermacher/kainaticulous-rp-7b-GGUF", MessagesFormatterType.MISTRAL],
82
+ "OpenCrystal-12B-L3.1-128K.Q4_K_S.gguf": ["mradermacher/OpenCrystal-12B-L3.1-128K-GGUF", MessagesFormatterType.MISTRAL],
83
+ "L3.1-12B-Niitama-v1.1.Q4_K_S.gguf": ["mradermacher/L3.1-12B-Niitama-v1.1-GGUF", MessagesFormatterType.MISTRAL],
84
+ "L3.1-12B-Celeste-V1.5.Q4_K_S.gguf": ["mradermacher/L3.1-12B-Celeste-V1.5-GGUF", MessagesFormatterType.MISTRAL],
85
+ "L3.1-Sthenorm-8B.Q4_K_S.gguf": ["mradermacher/L3.1-Sthenorm-8B-GGUF", MessagesFormatterType.MISTRAL],
86
+ "MN-12B-Estrella-v2.2.Q4_K_M.gguf": ["mradermacher/MN-12B-Estrella-v2.2-GGUF", MessagesFormatterType.MISTRAL],
87
+ "WoonaV1.2-9b.Q4_K_M.gguf": ["QuantFactory/WoonaV1.2-9b-GGUF", MessagesFormatterType.MISTRAL],
88
+ "kainaticulous-rp-7B-Q5_K_M.gguf": ["kainatq/kainaticulous-rp-7b-gguf", MessagesFormatterType.MISTRAL],
89
+ "experiment-wip-q4_k_m.gguf": ["DreadPoor/EXPERIMENT-WIP-Q4_K_M-GGUF", MessagesFormatterType.CHATML],
90
+ "eye_of_the_storm-q4_k_m.gguf": ["Daemontatox/Eye_of_the_Storm-Q4_K_M-GGUF", MessagesFormatterType.MISTRAL],
91
+ "Fireball-3.1-8B-ORPO.i1-Q5_K_M.gguf": ["mradermacher/Fireball-3.1-8B-ORPO-i1-GGUF", MessagesFormatterType.LLAMA_3],
92
+ "ggml-model-Q4_K_M.gguf": ["lightblue/suzume-llama-3-8B-multilingual-orpo-borda-half-gguf", MessagesFormatterType.LLAMA_3],
93
+ "kukulemon-v3-soul_mix-32k-7B.Q5_K_M.gguf": ["grimjim/kukulemon-v3-soul_mix-32k-7B-GGUF", MessagesFormatterType.MISTRAL],
94
+ #"CestLumi-v2.Q4_K_M.gguf": ["mradermacher/CestLumi-v2-GGUF", MessagesFormatterType.CHATML], # incomplete merges according to the author
95
+ "zephyr-wizard-kuno-royale-BF16-merge-7B.Q5_K_M.gguf": ["grimjim/zephyr-wizard-kuno-royale-BF16-merge-7B-GGUF", MessagesFormatterType.MISTRAL],
96
+ "Llama-3-Luminurse-v0.2-OAS-8B.Q5_K_M.gguf": ["grimjim/Llama-3-Luminurse-v0.2-OAS-8B-GGUF", MessagesFormatterType.LLAMA_3],
97
+ #"L3.1-Ablaze-Vulca-v0.2b-8B.q8_0.gguf": ["kromquant/L3.1-Ablaze-Vulca-v0.2b-8B-GGUFs", MessagesFormatterType.MISTRAL],
98
+ "MN-12B-Lyra-v2a1.Q4_K_M.gguf": ["mradermacher/MN-12B-Lyra-v2a1-GGUF", MessagesFormatterType.MISTRAL],
99
+ "OpenCrystal-8B-L3.Q5_K_M.gguf": ["mradermacher/OpenCrystal-8B-L3-GGUF", MessagesFormatterType.LLAMA_3],
100
+ "MN-CelesteGold-12B-Merge.Q4_K_M.gguf": ["mradermacher/MN-CelesteGold-12B-Merge-GGUF", MessagesFormatterType.CHATML],
101
+ "L3.1-Suze-Vume-calc.Q5_K_M.gguf": ["mradermacher/L3.1-Suze-Vume-calc-GGUF", MessagesFormatterType.MISTRAL],
102
+ "OpenCrystal-8B-L3.i1-Q4_K_M.gguf": ["mradermacher/OpenCrystal-8B-L3-i1-GGUF", MessagesFormatterType.LLAMA_3],
103
+ "cestlumi-v2-q4_k_m.gguf": ["TheDrunkenSnail/CestLumi-v2-Q4_K_M-GGUF", MessagesFormatterType.MISTRAL],
104
+ "cestlumi-v1-q4_k_m.gguf": ["TheDrunkenSnail/CestLumi-v1-Q4_K_M-GGUF", MessagesFormatterType.MISTRAL],
105
+ "Lyra_Gutenbergs-Twilight_Magnum-12B-Q4_K_M.gguf": ["bartowski/Lyra_Gutenbergs-Twilight_Magnum-12B-GGUF", MessagesFormatterType.MISTRAL],
106
+ "llama-3.1-8b-ultra-instruct-q4_k_s-imat.gguf": ["Dampfinchen/Llama-3.1-8B-Ultra-Instruct-Q4_K_S-GGUF", MessagesFormatterType.LLAMA_3],
107
+ "kunoichi-lemon-royale-v3-32K-7B.Q5_K_M.gguf": ["grimjim/kunoichi-lemon-royale-v3-32K-7B-GGUF", MessagesFormatterType.MISTRAL],
108
+ "kukulemon-spiked-9B.Q4_K_M.gguf": ["grimjim/kukulemon-spiked-9B-GGUF", MessagesFormatterType.MISTRAL],
109
+ "lemonade-rebase-32k-7B.Q8_0.gguf": ["grimjim/lemonade-rebase-32k-7B-GGUF", MessagesFormatterType.MISTRAL],
110
+ "infinite-lemonade-7B.Q8_0.gguf": ["grimjim/infinite-lemonade-SLERP-7B-GGUF", MessagesFormatterType.MISTRAL],
111
+ "llama-3-merge-pp-instruct-8B.Q8_0.gguf": ["grimjim/llama-3-merge-pp-instruct-8B-GGUF", MessagesFormatterType.LLAMA_3],
112
+ "L3-SthenoMaidBlackroot-12.2B-V1-INSTRUCT-16fp.i1-Q4_K_S.gguf": ["mradermacher/L3-SthenoMaidBlackroot-12.2B-V1-INSTRUCT-16fp-i1-GGUF", MessagesFormatterType.LLAMA_3],
113
+ "Fireball-Mistral-Nemo-Base-2407-Instruct-v1.Q4_K_M.gguf": ["mradermacher/Fireball-Mistral-Nemo-Base-2407-Instruct-v1-GGUF", MessagesFormatterType.MISTRAL],
114
+ "L3-12B-Niitama-v1.Q4_K_M.gguf": ["mradermacher/L3-12B-Niitama-v1-GGUF", MessagesFormatterType.LLAMA_3],
115
+ "L3-12B-Celeste-V1.2.Q4_K_M.gguf": ["mradermacher/L3-12B-Celeste-V1.2-GGUF", MessagesFormatterType.LLAMA_3],
116
+ "OpenCrystal-12B-L3.Q4_K_M.gguf": ["mradermacher/OpenCrystal-12B-L3-GGUF", MessagesFormatterType.LLAMA_3],
117
+ "L3-Stheno-v3.2-12.2B-Instruct.Q4_K_M.gguf": ["mradermacher/L3-Stheno-v3.2-12.2B-Instruct-GGUF", MessagesFormatterType.LLAMA_3],
118
+ "L3-SMB-Instruct-12.2B-F32.Q4_K_M.gguf": ["mradermacher/L3-SMB-Instruct-12.2B-F32-GGUF", MessagesFormatterType.LLAMA_3],
119
+ "OpenCrystal-12B-L3.i1-Q4_K_M.gguf": ["mradermacher/OpenCrystal-12B-L3-i1-GGUF", MessagesFormatterType.LLAMA_3],
120
+ "L3-Lumimaid-12.2B-v0.1-OAS-Instruct.Q4_K_M.gguf": ["mradermacher/L3-Lumimaid-12.2B-v0.1-OAS-Instruct-GGUF", MessagesFormatterType.LLAMA_3],
121
  "nemo-12b-hiwaifu-Q4_K_L-imat.gguf": ["Reiterate3680/nemo-12b-hiwaifu-GGUF", MessagesFormatterType.MISTRAL],
122
  "Soliloquy-7B-v3-Q4_K_L-imat.gguf": ["Reiterate3680/Soliloquy-7B-v3-GGUF", MessagesFormatterType.OPEN_CHAT],
123
  "Lyra-Gutenberg-mistral-nemo-12B.Q4_K_M.gguf": ["mradermacher/Lyra-Gutenberg-mistral-nemo-12B-GGUF", MessagesFormatterType.MISTRAL],
 
390
  "Qwen2": MessagesFormatterType.OPEN_CHAT,
391
  }
392
  # https://github.com/Maximilian-Winter/llama-cpp-agent
393
+ llm_languages = ["English", "Japanese", "Chinese", "Korean", "Spanish", "Portuguese", "German", "French", "Finnish", "Russian"]
394
  llm_models_tupled_list = []
395
  default_llm_model_filename = list(llm_models.keys())[0]
396
  override_llm_format = None
 
456
  if not filename in llm_models.keys(): return default_llm_model_filename
457
  try:
458
  hf_hub_download(repo_id = llm_models[filename][0], filename = filename, local_dir = llm_models_dir)
459
+ except Exception as e:
460
+ print(e)
461
  return default_llm_model_filename
462
  update_llm_model_tupled_list()
463
  return filename
 
487
  return gr.update(value=format_name)
488
 
489
 
 
490
  download_llm_model(default_llm_model_filename)
491
 
492
 
 
536
  if not api.repo_exists(repo_id = repo) or not api.file_exists(repo_id = repo, filename = filename): return gr.update(visible=True)
537
  add_models[filename] = [repo, format]
538
  else: return gr.update(visible=True)
539
+ except Exception as e:
540
+ print(e)
541
  return gr.update(visible=True)
542
+ #print(add_models)
543
  llm_models = (llm_models | add_models).copy()
544
+ update_llm_model_tupled_list()
545
+ choices = get_dolphin_models()
546
+ return gr.update(choices=choices, value=choices[-1][1])
547
 
548
 
549
  dolphin_output_language = "English"
modutils.py CHANGED
@@ -1,19 +1,14 @@
 
1
  import json
2
  import gradio as gr
3
  from huggingface_hub import HfApi
4
  import os
5
  from pathlib import Path
6
 
7
- from env import (
8
- HF_LORA_PRIVATE_REPOS1,
9
- HF_LORA_PRIVATE_REPOS2,
10
- HF_MODEL_USER_EX,
11
- HF_MODEL_USER_LIKES,
12
- directory_loras,
13
- hf_read_token,
14
- hf_token,
15
- CIVITAI_API_KEY,
16
- )
17
 
18
 
19
  def get_user_agent():
@@ -112,8 +107,8 @@ def save_gallery_images(images, progress=gr.Progress(track_tqdm=True)):
112
  try:
113
  if oldpath.exists():
114
  newpath = oldpath.resolve().rename(Path(filename).resolve())
115
- except Exception:
116
- pass
117
  finally:
118
  output_paths.append(str(newpath))
119
  output_images.append((str(newpath), str(filename)))
@@ -127,7 +122,8 @@ def download_private_repo(repo_id, dir_path, is_replace):
127
  try:
128
  snapshot_download(repo_id=repo_id, local_dir=dir_path, allow_patterns=['*.ckpt', '*.pt', '*.pth', '*.safetensors', '*.bin'], use_auth_token=hf_read_token)
129
  except Exception as e:
130
- print(f"Error: Failed to download {repo_id}. ")
 
131
  return
132
  if is_replace:
133
  for file in Path(dir_path).glob("*"):
@@ -146,7 +142,8 @@ def get_private_model_list(repo_id, dir_path):
146
  try:
147
  files = api.list_repo_files(repo_id, token=hf_read_token)
148
  except Exception as e:
149
- print(f"Error: Failed to list {repo_id}. ")
 
150
  return []
151
  model_list = []
152
  for file in files:
@@ -168,7 +165,8 @@ def download_private_file(repo_id, path, is_replace):
168
  try:
169
  hf_hub_download(repo_id=repo_id, filename=filename, local_dir=dirname, use_auth_token=hf_read_token)
170
  except Exception as e:
171
- print(f"Error: Failed to download {filename}. ")
 
172
  return
173
  if is_replace:
174
  file.resolve().rename(newpath.resolve())
@@ -194,7 +192,8 @@ def get_model_id_list():
194
  for author in HF_MODEL_USER_EX:
195
  models_ex = api.list_models(author=author, cardData=True, sort="last_modified")
196
  except Exception as e:
197
- print(f"Error: Failed to list {author}'s models. ")
 
198
  return model_ids
199
  for model in models_likes:
200
  model_ids.append(model.id) if not model.private else ""
@@ -218,17 +217,17 @@ def get_t2i_model_info(repo_id: str):
218
  if " " in repo_id or not api.repo_exists(repo_id): return ""
219
  model = api.model_info(repo_id=repo_id)
220
  except Exception as e:
221
- print(f"Error: Failed to get {repo_id}'s info. ")
 
222
  return ""
223
  if model.private or model.gated: return ""
224
  tags = model.tags
225
  info = []
226
  url = f"https://huggingface.co/{repo_id}/"
227
  if not 'diffusers' in tags: return ""
228
- if 'diffusers:StableDiffusionXLPipeline' in tags:
229
- info.append("SDXL")
230
- elif 'diffusers:StableDiffusionPipeline' in tags:
231
- info.append("SD1.5")
232
  if model.card_data and model.card_data.tags:
233
  info.extend(list_sub(model.card_data.tags, ['text-to-image', 'stable-diffusion', 'stable-diffusion-api', 'safetensors', 'stable-diffusion-xl']))
234
  info.append(f"DLs: {model.downloads}")
@@ -247,6 +246,7 @@ def get_tupled_model_list(model_list):
247
  if not api.repo_exists(repo_id): continue
248
  model = api.model_info(repo_id=repo_id)
249
  except Exception as e:
 
250
  continue
251
  if model.private or model.gated: continue
252
  tags = model.tags
@@ -273,8 +273,8 @@ try:
273
  d = json.load(f)
274
  for k, v in d.items():
275
  private_lora_dict[escape_lora_basename(k)] = v
276
- except Exception:
277
- pass
278
  loras_dict = {"None": ["", "", "", "", ""], "": ["", "", "", "", ""]} | private_lora_dict.copy()
279
  civitai_not_exists_list = []
280
  loras_url_to_path_dict = {} # {"URL to download": "local filepath", ...}
@@ -322,6 +322,7 @@ def get_civitai_info(path):
322
  try:
323
  r = session.get(url, params=params, headers=headers, stream=True, timeout=(3.0, 15))
324
  except Exception as e:
 
325
  return ["", "", "", "", ""]
326
  if not r.ok: return None
327
  json = r.json()
@@ -420,7 +421,8 @@ def copy_lora(path: str, new_path: str):
420
  if cpath.exists():
421
  try:
422
  shutil.copy(str(cpath.resolve()), str(npath.resolve()))
423
- except Exception:
 
424
  return None
425
  update_lora_dict(str(npath))
426
  return new_path
@@ -1215,6 +1217,8 @@ def get_model_pipeline(repo_id: str):
1215
  if model.private or model.gated: return default
1216
  tags = model.tags
1217
  if not 'diffusers' in tags: return default
 
 
1218
  if 'diffusers:StableDiffusionXLPipeline' in tags:
1219
  return "StableDiffusionXLPipeline"
1220
  elif 'diffusers:StableDiffusionPipeline' in tags:
 
1
+ import spaces
2
  import json
3
  import gradio as gr
4
  from huggingface_hub import HfApi
5
  import os
6
  from pathlib import Path
7
 
8
+
9
+ from env import (HF_LORA_PRIVATE_REPOS1, HF_LORA_PRIVATE_REPOS2,
10
+ HF_MODEL_USER_EX, HF_MODEL_USER_LIKES,
11
+ directory_loras, hf_read_token, hf_token, CIVITAI_API_KEY)
 
 
 
 
 
 
12
 
13
 
14
  def get_user_agent():
 
107
  try:
108
  if oldpath.exists():
109
  newpath = oldpath.resolve().rename(Path(filename).resolve())
110
+ except Exception as e:
111
+ print(e)
112
  finally:
113
  output_paths.append(str(newpath))
114
  output_images.append((str(newpath), str(filename)))
 
122
  try:
123
  snapshot_download(repo_id=repo_id, local_dir=dir_path, allow_patterns=['*.ckpt', '*.pt', '*.pth', '*.safetensors', '*.bin'], use_auth_token=hf_read_token)
124
  except Exception as e:
125
+ print(f"Error: Failed to download {repo_id}.")
126
+ print(e)
127
  return
128
  if is_replace:
129
  for file in Path(dir_path).glob("*"):
 
142
  try:
143
  files = api.list_repo_files(repo_id, token=hf_read_token)
144
  except Exception as e:
145
+ print(f"Error: Failed to list {repo_id}.")
146
+ print(e)
147
  return []
148
  model_list = []
149
  for file in files:
 
165
  try:
166
  hf_hub_download(repo_id=repo_id, filename=filename, local_dir=dirname, use_auth_token=hf_read_token)
167
  except Exception as e:
168
+ print(f"Error: Failed to download {filename}.")
169
+ print(e)
170
  return
171
  if is_replace:
172
  file.resolve().rename(newpath.resolve())
 
192
  for author in HF_MODEL_USER_EX:
193
  models_ex = api.list_models(author=author, cardData=True, sort="last_modified")
194
  except Exception as e:
195
+ print(f"Error: Failed to list {author}'s models.")
196
+ print(e)
197
  return model_ids
198
  for model in models_likes:
199
  model_ids.append(model.id) if not model.private else ""
 
217
  if " " in repo_id or not api.repo_exists(repo_id): return ""
218
  model = api.model_info(repo_id=repo_id)
219
  except Exception as e:
220
+ print(f"Error: Failed to get {repo_id}'s info.")
221
+ print(e)
222
  return ""
223
  if model.private or model.gated: return ""
224
  tags = model.tags
225
  info = []
226
  url = f"https://huggingface.co/{repo_id}/"
227
  if not 'diffusers' in tags: return ""
228
+ if 'diffusers:FluxPipeline' in tags: info.append("FLUX.1")
229
+ elif 'diffusers:StableDiffusionXLPipeline' in tags: info.append("SDXL")
230
+ elif 'diffusers:StableDiffusionPipeline' in tags: info.append("SD1.5")
 
231
  if model.card_data and model.card_data.tags:
232
  info.extend(list_sub(model.card_data.tags, ['text-to-image', 'stable-diffusion', 'stable-diffusion-api', 'safetensors', 'stable-diffusion-xl']))
233
  info.append(f"DLs: {model.downloads}")
 
246
  if not api.repo_exists(repo_id): continue
247
  model = api.model_info(repo_id=repo_id)
248
  except Exception as e:
249
+ print(e)
250
  continue
251
  if model.private or model.gated: continue
252
  tags = model.tags
 
273
  d = json.load(f)
274
  for k, v in d.items():
275
  private_lora_dict[escape_lora_basename(k)] = v
276
+ except Exception as e:
277
+ print(e)
278
  loras_dict = {"None": ["", "", "", "", ""], "": ["", "", "", "", ""]} | private_lora_dict.copy()
279
  civitai_not_exists_list = []
280
  loras_url_to_path_dict = {} # {"URL to download": "local filepath", ...}
 
322
  try:
323
  r = session.get(url, params=params, headers=headers, stream=True, timeout=(3.0, 15))
324
  except Exception as e:
325
+ print(e)
326
  return ["", "", "", "", ""]
327
  if not r.ok: return None
328
  json = r.json()
 
421
  if cpath.exists():
422
  try:
423
  shutil.copy(str(cpath.resolve()), str(npath.resolve()))
424
+ except Exception as e:
425
+ print(e)
426
  return None
427
  update_lora_dict(str(npath))
428
  return new_path
 
1217
  if model.private or model.gated: return default
1218
  tags = model.tags
1219
  if not 'diffusers' in tags: return default
1220
+ if 'diffusers:FluxPipeline' in tags:
1221
+ return "FluxPipeline"
1222
  if 'diffusers:StableDiffusionXLPipeline' in tags:
1223
  return "StableDiffusionXLPipeline"
1224
  elif 'diffusers:StableDiffusionPipeline' in tags: