John6666 commited on
Commit
501231e
·
verified ·
0 Parent(s):

Super-squash branch 'main' using huggingface_hub

Browse files
.gitattributes ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ title: Download and Convert FLUX.1 ComfyUI Safetensors To Diffusers (Give Up)
3
+ emoji: 🎨➡️🧨
4
+ colorFrom: indigo
5
+ colorTo: purple
6
+ sdk: gradio
7
+ sdk_version: 4.40.0
8
+ app_file: app.py
9
+ pinned: false
10
+ license: mit
11
+ ---
12
+
13
+ Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
app.py ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ import spaces
3
+ import os
4
+ from convert_url_to_diffusers_flux_gr import convert_url_to_diffusers_repo_flux
5
+ os.environ["GRADIO_ANALYTICS_ENABLED"] = "False"
6
+
7
+
8
+ css = """"""
9
+
10
+ with gr.Blocks(theme="NoCrypt/miku@>=1.2.2", fill_width=True, css=css) as demo:
11
+ gr.Markdown("# Download and convert FLUX.1 ComfyUI formatted safetensors to Diffusers and create your repo")
12
+ gr.Markdown(
13
+ f"""
14
+ **⚠️IMPORTANT NOTICE⚠️**<br>
15
+ # If the output setting was fp8, this space could be completed in about 10 minutes, but **the torch on HF's server apparently does not support fp8 input**, which makes no sense.
16
+ The conversion to bf16 does not seem feasible in any way at present. Even if the file is processed as shard, it still does not work due to lack of RAM. (P.S. But then the RAM was down to only 60% consumption. I don't know why anymore.)
17
+ I guess I'll have to freeze it until someone with more advanced technology realizes it, or until Diffusers, pytorch, or Quanto will be upgraded.<br><br>
18
+ From an information security standpoint, it is dangerous to expose your access token or key to others.
19
+ If you do use it, I recommend that you duplicate this space on your own account before doing so.
20
+ Keys and tokens could be set to SECRET (HF_TOKEN, CIVITAI_API_KEY) if it's placed in your own space.
21
+ It saves you the trouble of typing them in.<br>
22
+ <br>
23
+ **The steps are the following**:
24
+ - Paste a write-access token from [hf.co/settings/tokens](https://huggingface.co/settings/tokens).
25
+ - Input a model download url from the Hub or Civitai or other sites.
26
+ - If you want to download a model from Civitai, paste a Civitai API Key.
27
+ - Input your HF user ID. e.g. 'yourid'.
28
+ - Input your new repo name. If empty, auto-complete. e.g. 'newrepo'.
29
+ - Set the parameters. If not sure, just use the defaults.
30
+ - Click "Submit".
31
+ - Patiently wait until the output changes. It takes approximately ? minutes (downloading from HF).
32
+ """
33
+ )
34
+ with gr.Column():
35
+ dl_url = gr.Textbox(label="URL to download", placeholder="https://huggingface.co/marduk191/Flux.1_collection/blob/main/flux.1_dev_fp8_fp16t5-marduk191.safetensors", value="", max_lines=1)
36
+ hf_user = gr.Textbox(label="Your HF user ID", placeholder="username", value="", max_lines=1)
37
+ hf_repo = gr.Textbox(label="New repo name", placeholder="reponame", info="If empty, auto-complete", value="", max_lines=1)
38
+ hf_token = gr.Textbox(label="Your HF write token", placeholder="hf_...", value="", max_lines=1)
39
+ civitai_key = gr.Textbox(label="Your Civitai API Key (Optional)", info="If you download model from Civitai...", placeholder="", value="", max_lines=1)
40
+ is_upload_sf = gr.Checkbox(label="Upload single safetensors file into new repo", value=False, visible=False)
41
+ data_type = gr.Radio(label="Output data type", info="It only affects transformer and text encoder.", choices=["bf16", "fp8", "qfloat8"], value="fp8")
42
+ model_type = gr.Radio(label="Original model type", choices=["dev", "schnell"], value="dev")
43
+ is_dequat = gr.Checkbox(label="Dequantization", info="Deadly slow", value=False)
44
+ use_original = gr.CheckboxGroup(label="Use original version", choices=["vae", "text_encoder", "text_encoder_2"], value=["vae", "text_encoder"])
45
+ is_fix_only = gr.Checkbox(label="Only fixing", value=False)
46
+ run_button = gr.Button(value="Submit")
47
+ repo_urls = gr.CheckboxGroup(visible=False, choices=[], value=None)
48
+ output_md = gr.Markdown(label="Output")
49
+ gr.DuplicateButton(value="Duplicate Space")
50
+
51
+ gr.on(
52
+ triggers=[run_button.click],
53
+ fn=convert_url_to_diffusers_repo_flux,
54
+ inputs=[dl_url, hf_user, hf_repo, hf_token, civitai_key, is_upload_sf,
55
+ data_type, model_type, is_dequat, repo_urls, is_fix_only, use_original],
56
+ outputs=[repo_urls, output_md],
57
+ )
58
+
59
+ demo.queue(default_concurrency_limit=1, max_size=5).launch(debug=True, show_api=False)
convert_url_to_diffusers_flux_gr.py ADDED
@@ -0,0 +1,939 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import json
2
+ import torch
3
+ from safetensors.torch import load_file, save_file
4
+ from pathlib import Path
5
+ import gc
6
+ import gguf
7
+ from dequant import dequantize_tensor # https://github.com/city96/ComfyUI-GGUF
8
+
9
+ import os
10
+ import argparse
11
+ import gradio as gr
12
+ # also requires aria, gdown, peft, huggingface_hub, safetensors, transformers, accelerate, pytorch_lightning
13
+ import spaces
14
+
15
+ flux_dev_repo = "ChuckMcSneed/FLUX.1-dev"
16
+ flux_schnell_repo = "black-forest-labs/FLUX.1-schnell"
17
+ system_temp_dir = "temp"
18
+
19
+ device = "cuda" if torch.cuda.is_available() else "cpu"
20
+ torch.set_grad_enabled(False)
21
+
22
+ GGUF_QTYPE = [gguf.GGMLQuantizationType.Q8_0, gguf.GGMLQuantizationType.Q5_1,
23
+ gguf.GGMLQuantizationType.Q5_0, gguf.GGMLQuantizationType.Q4_1,
24
+ gguf.GGMLQuantizationType.Q4_0, gguf.GGMLQuantizationType.F32, gguf.GGMLQuantizationType.F16]
25
+
26
+ TORCH_DTYPE = [torch.float32, torch.float, torch.float64, torch.double, torch.float16, torch.half,
27
+ torch.bfloat16, torch.complex32, torch.chalf, torch.complex64, torch.cfloat,
28
+ torch.complex128, torch.cdouble, torch.uint8, torch.uint16, torch.uint32, torch.uint64,
29
+ torch.int8, torch.int16, torch.short, torch.int32, torch.int, torch.int64, torch.long,
30
+ torch.bool, torch.float8_e4m3fn, torch.float8_e5m2]
31
+
32
+ TORCH_QUANTIZED_DTYPE = [torch.quint8, torch.qint8, torch.qint32, torch.quint4x2]
33
+
34
+ def list_sub(a, b):
35
+ return [e for e in a if e not in b]
36
+
37
+ def is_repo_name(s):
38
+ import re
39
+ return re.fullmatch(r'^[^/,\s]+?/[^/,\s]+?$', s)
40
+
41
+ def print_resource_usage():
42
+ import psutil
43
+ cpu_usage = psutil.cpu_percent()
44
+ ram_usage = psutil.virtual_memory().used / psutil.virtual_memory().total * 100
45
+ print(f"CPU usage: {cpu_usage}% / RAM usage: {ram_usage}%")
46
+
47
+ def download_thing(directory, url, civitai_api_key="", progress=gr.Progress(track_tqdm=True)):
48
+ progress(0, desc="Start downloading...")
49
+ url = url.strip()
50
+ if "drive.google.com" in url:
51
+ original_dir = os.getcwd()
52
+ os.chdir(directory)
53
+ os.system(f"gdown --fuzzy {url}")
54
+ os.chdir(original_dir)
55
+ elif "huggingface.co" in url:
56
+ url = url.replace("?download=true", "")
57
+ if "/blob/" in url:
58
+ url = url.replace("/blob/", "/resolve/")
59
+ os.system(f"aria2c --console-log-level=error --summary-interval=10 -c -x 16 -k 1M -s 16 {url} -d {directory} -o {url.split('/')[-1]}")
60
+ else:
61
+ os.system (f"aria2c --optimize-concurrent-downloads --console-log-level=error --summary-interval=10 -c -x 16 -k 1M -s 16 {url} -d {directory} -o {url.split('/')[-1]}")
62
+ elif "civitai.com" in url:
63
+ if "?" in url:
64
+ url = url.split("?")[0]
65
+ if civitai_api_key:
66
+ url = url + f"?token={civitai_api_key}"
67
+ os.system(f"aria2c --console-log-level=error --summary-interval=10 -c -x 16 -k 1M -s 16 -d {directory} {url}")
68
+ else:
69
+ print("You need an API key to download Civitai models.")
70
+ else:
71
+ os.system(f"aria2c --console-log-level=error --summary-interval=10 -c -x 16 -k 1M -s 16 -d {directory} {url}")
72
+
73
+ def get_local_model_list(dir_path):
74
+ model_list = []
75
+ valid_extensions = ('.safetensors')
76
+ for file in Path(dir_path).glob("*"):
77
+ if file.suffix in valid_extensions:
78
+ file_path = str(Path(f"{dir_path}/{file.name}"))
79
+ model_list.append(file_path)
80
+ return model_list
81
+
82
+ def get_download_file(temp_dir, url, civitai_key, progress=gr.Progress(track_tqdm=True)):
83
+ if not "http" in url and is_repo_name(url) and not Path(url).exists():
84
+ print(f"Use HF Repo: {url}")
85
+ new_file = url
86
+ elif not "http" in url and Path(url).exists():
87
+ print(f"Use local file: {url}")
88
+ new_file = url
89
+ elif Path(f"{temp_dir}/{url.split('/')[-1]}").exists():
90
+ print(f"File to download alreday exists: {url}")
91
+ new_file = f"{temp_dir}/{url.split('/')[-1]}"
92
+ else:
93
+ print(f"Start downloading: {url}")
94
+ before = get_local_model_list(temp_dir)
95
+ try:
96
+ download_thing(temp_dir, url.strip(), civitai_key)
97
+ except Exception:
98
+ print(f"Download failed: {url}")
99
+ return ""
100
+ after = get_local_model_list(temp_dir)
101
+ new_file = list_sub(after, before)[0] if list_sub(after, before) else ""
102
+ if not new_file:
103
+ print(f"Download failed: {url}")
104
+ return ""
105
+ print(f"Download completed: {url}")
106
+ return new_file
107
+
108
+ def save_readme_md(dir, url):
109
+ orig_url = ""
110
+ if "http" in url:
111
+ orig_url = url
112
+ if orig_url:
113
+ md = f"""---
114
+ license: other
115
+ license_name: flux-1-dev-non-commercial-license
116
+ license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.
117
+ language:
118
+ - en
119
+ library_name: diffusers
120
+ pipeline_tag: text-to-image
121
+ tags:
122
+ - text-to-image
123
+ - Flux
124
+ ---
125
+ Converted from [{orig_url}]({orig_url}).
126
+ """
127
+ else:
128
+ md = f"""---
129
+ license: other
130
+ license_name: flux-1-dev-non-commercial-license
131
+ license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.
132
+ language:
133
+ - en
134
+ library_name: diffusers
135
+ pipeline_tag: text-to-image
136
+ tags:
137
+ - text-to-image
138
+ - Flux
139
+ ---
140
+ """
141
+ path = str(Path(dir, "README.md"))
142
+ with open(path, mode='w', encoding="utf-8") as f:
143
+ f.write(md)
144
+
145
+ def is_repo_exists(repo_id):
146
+ from huggingface_hub import HfApi
147
+ api = HfApi()
148
+ try:
149
+ if api.repo_exists(repo_id=repo_id): return True
150
+ else: return False
151
+ except Exception as e:
152
+ print(f"Error: Failed to connect {repo_id}. ")
153
+ return True # for safe
154
+
155
+ def create_diffusers_repo(new_repo_id, diffusers_folder, progress=gr.Progress(track_tqdm=True)):
156
+ from huggingface_hub import HfApi
157
+ import os
158
+ hf_token = os.environ.get("HF_TOKEN")
159
+ api = HfApi()
160
+ try:
161
+ progress(0, desc="Start uploading...")
162
+ api.create_repo(repo_id=new_repo_id, token=hf_token, private=True, exist_ok=True)
163
+ for path in Path(diffusers_folder).glob("*"):
164
+ if path.is_dir():
165
+ api.upload_folder(repo_id=new_repo_id, folder_path=str(path), path_in_repo=path.name, token=hf_token)
166
+ elif path.is_file():
167
+ api.upload_file(repo_id=new_repo_id, path_or_fileobj=str(path), path_in_repo=path.name, token=hf_token)
168
+ progress(1, desc="Uploaded.")
169
+ url = f"https://huggingface.co/{new_repo_id}"
170
+ except Exception as e:
171
+ print(f"Error: Failed to upload to {new_repo_id}. ")
172
+ print(e)
173
+ return ""
174
+ return url
175
+
176
+ # https://github.com/huggingface/diffusers/blob/main/scripts/convert_flux_to_diffusers.py
177
+ # in SD3 original implementation of AdaLayerNormContinuous, it split linear projection output into shift, scale;
178
+ # while in diffusers it split into scale, shift. Here we swap the linear projection weights in order to be able to use diffusers implementation
179
+ with torch.no_grad(), torch.autocast(device):
180
+ @torch.jit.script
181
+ def swap_scale_shift(weight):
182
+ shift, scale = weight.chunk(2, dim=0)
183
+ new_weight = torch.cat([scale, shift], dim=0)
184
+ return new_weight
185
+
186
+ with torch.no_grad(), torch.autocast(device):
187
+ def convert_flux_transformer_checkpoint_to_diffusers(
188
+ original_state_dict, num_layers, num_single_layers, inner_dim, mlp_ratio=4.0,
189
+ progress=gr.Progress(track_tqdm=True)):
190
+ def conv(cdict: dict, odict: dict, ckey: str, okey: str):
191
+ if okey in odict.keys():
192
+ progress(0, desc=f"Converting {okey} => {ckey}")
193
+ print(f"Converting {okey} => {ckey}")
194
+ cdict[ckey] = odict.pop(okey)
195
+ gc.collect()
196
+
197
+ def convswap(cdict: dict, odict: dict, ckey: str, okey: str):
198
+ if okey in odict.keys():
199
+ progress(0, desc=f"Converting (swap) {okey} => {ckey}")
200
+ print(f"Converting {okey} => {ckey} (swap)")
201
+ cdict[ckey] = swap_scale_shift(odict.pop(okey))
202
+ gc.collect()
203
+
204
+ def convqkv(cdict: dict, odict: dict, i: int):
205
+ keys = odict.keys()
206
+ if (f"double_blocks.{i}.img_attn.qkv.weight" in keys or f"double_blocks.{i}.txt_attn.qkv.weight" in keys\
207
+ or f"double_blocks.{i}.img_attn.qkv.bias" in keys or f"double_blocks.{i}.txt_attn.qkv.bias" in keys)\
208
+ and (f"double_blocks.{i}.img_attn.qkv.weight" not in keys or f"double_blocks.{i}.txt_attn.qkv.weight" not in keys\
209
+ or f"double_blocks.{i}.img_attn.qkv.bias" not in keys or f"double_blocks.{i}.txt_attn.qkv.bias" not in keys):
210
+ progress(0, desc=f"Key error in converting Q, K, V (double_blocks.{i}).")
211
+ print(f"Key error in converting Q, K, V (double_blocks.{i}).")
212
+ return
213
+ progress(0, desc=f"Converting Q, K, V (double_blocks.{i}).")
214
+ print(f"Converting Q, K, V (double_blocks.{i}).")
215
+ sample_q, sample_k, sample_v = torch.chunk(
216
+ odict.pop(f"double_blocks.{i}.img_attn.qkv.weight"), 3, dim=0
217
+ )
218
+ context_q, context_k, context_v = torch.chunk(
219
+ odict.pop(f"double_blocks.{i}.txt_attn.qkv.weight"), 3, dim=0
220
+ )
221
+ sample_q_bias, sample_k_bias, sample_v_bias = torch.chunk(
222
+ odict.pop(f"double_blocks.{i}.img_attn.qkv.bias"), 3, dim=0
223
+ )
224
+ context_q_bias, context_k_bias, context_v_bias = torch.chunk(
225
+ odict.pop(f"double_blocks.{i}.txt_attn.qkv.bias"), 3, dim=0
226
+ )
227
+ cdict[f"{block_prefix}attn.to_q.weight"] = torch.cat([sample_q])
228
+ cdict[f"{block_prefix}attn.to_q.bias"] = torch.cat([sample_q_bias])
229
+ cdict[f"{block_prefix}attn.to_k.weight"] = torch.cat([sample_k])
230
+ cdict[f"{block_prefix}attn.to_k.bias"] = torch.cat([sample_k_bias])
231
+ cdict[f"{block_prefix}attn.to_v.weight"] = torch.cat([sample_v])
232
+ cdict[f"{block_prefix}attn.to_v.bias"] = torch.cat([sample_v_bias])
233
+ cdict[f"{block_prefix}attn.add_q_proj.weight"] = torch.cat([context_q])
234
+ cdict[f"{block_prefix}attn.add_q_proj.bias"] = torch.cat([context_q_bias])
235
+ cdict[f"{block_prefix}attn.add_k_proj.weight"] = torch.cat([context_k])
236
+ cdict[f"{block_prefix}attn.add_k_proj.bias"] = torch.cat([context_k_bias])
237
+ cdict[f"{block_prefix}attn.add_v_proj.weight"] = torch.cat([context_v])
238
+ cdict[f"{block_prefix}attn.add_v_proj.bias"] = torch.cat([context_v_bias])
239
+ gc.collect()
240
+
241
+ def convqkvmlp(cdict: dict, odict: dict, i: int, inner_dim: int, mlp_ratio: float):
242
+ keys = odict.keys()
243
+ if (f"single_blocks.{i}.linear1.weight" in keys or f"single_blocks.{i}.linear1.bias" in keys)\
244
+ and (f"single_blocks.{i}.linear1.weight" not in keys or f"single_blocks.{i}.linear1.bias" not in keys):
245
+ progress(0, desc=f"Key error in converting Q, K, V, mlp (single_blocks.{i}).")
246
+ print(f"Key error in converting Q, K, V, mlp (single_blocks.{i}).")
247
+ return
248
+ progress(0, desc=f"Converting Q, K, V, mlp (single_blocks.{i}).")
249
+ print(f"Converting Q, K, V, mlp (single_blocks.{i}).")
250
+ mlp_hidden_dim = int(inner_dim * mlp_ratio)
251
+ split_size = (inner_dim, inner_dim, inner_dim, mlp_hidden_dim)
252
+ q, k, v, mlp = torch.split(odict.pop(f"single_blocks.{i}.linear1.weight"), split_size, dim=0)
253
+ q_bias, k_bias, v_bias, mlp_bias = torch.split(
254
+ odict.pop(f"single_blocks.{i}.linear1.bias"), split_size, dim=0
255
+ )
256
+ cdict[f"{block_prefix}attn.to_q.weight"] = torch.cat([q])
257
+ cdict[f"{block_prefix}attn.to_q.bias"] = torch.cat([q_bias])
258
+ cdict[f"{block_prefix}attn.to_k.weight"] = torch.cat([k])
259
+ cdict[f"{block_prefix}attn.to_k.bias"] = torch.cat([k_bias])
260
+ cdict[f"{block_prefix}attn.to_v.weight"] = torch.cat([v])
261
+ cdict[f"{block_prefix}attn.to_v.bias"] = torch.cat([v_bias])
262
+ cdict[f"{block_prefix}proj_mlp.weight"] = torch.cat([mlp])
263
+ cdict[f"{block_prefix}proj_mlp.bias"] = torch.cat([mlp_bias])
264
+ gc.collect()
265
+
266
+ converted_state_dict = {}
267
+ progress(0, desc="Converting FLUX.1 state dict to Diffusers format.")
268
+
269
+ ## time_text_embed.timestep_embedder <- time_in
270
+ conv(converted_state_dict, original_state_dict, "time_text_embed.timestep_embedder.linear_1.weight", "time_in.in_layer.weight")
271
+ conv(converted_state_dict, original_state_dict, "time_text_embed.timestep_embedder.linear_1.bias", "time_in.in_layer.bias")
272
+ conv(converted_state_dict, original_state_dict, "time_text_embed.timestep_embedder.linear_2.weight", "time_in.out_layer.weight")
273
+ conv(converted_state_dict, original_state_dict, "time_text_embed.timestep_embedder.linear_2.bias", "time_in.out_layer.bias")
274
+
275
+ ## time_text_embed.text_embedder <- vector_in
276
+ conv(converted_state_dict, original_state_dict, "time_text_embed.text_embedder.linear_1.weight", "vector_in.in_layer.weight")
277
+ conv(converted_state_dict, original_state_dict, "time_text_embed.text_embedder.linear_1.bias", "vector_in.in_layer.bias")
278
+ conv(converted_state_dict, original_state_dict, "time_text_embed.text_embedder.linear_2.weight", "vector_in.out_layer.weight")
279
+ conv(converted_state_dict, original_state_dict, "time_text_embed.text_embedder.linear_2.bias", "vector_in.out_layer.bias")
280
+
281
+ # guidance
282
+ has_guidance = any("guidance" in k for k in original_state_dict)
283
+ if has_guidance:
284
+ conv(converted_state_dict, original_state_dict, "time_text_embed.guidance_embedder.linear_1.weight", "guidance_in.in_layer.weight")
285
+ conv(converted_state_dict, original_state_dict, "time_text_embed.guidance_embedder.linear_1.bias", "guidance_in.in_layer.bias")
286
+ conv(converted_state_dict, original_state_dict, "time_text_embed.guidance_embedder.linear_2.weight", "guidance_in.out_layer.weight")
287
+ conv(converted_state_dict, original_state_dict, "time_text_embed.guidance_embedder.linear_2.bias", "guidance_in.out_layer.bias")
288
+
289
+ # context_embedder
290
+ conv(converted_state_dict, original_state_dict, "context_embedder.weight", "txt_in.weight")
291
+ conv(converted_state_dict, original_state_dict, "context_embedder.bias", "txt_in.bias")
292
+
293
+ # x_embedder
294
+ conv(converted_state_dict, original_state_dict, "x_embedder.weight", "img_in.weight")
295
+ conv(converted_state_dict, original_state_dict, "x_embedder.bias", "img_in.bias")
296
+
297
+ progress(0.25, desc="Converting FLUX.1 state dict to Diffusers format.")
298
+ # double transformer blocks
299
+ for i in range(num_layers):
300
+ block_prefix = f"transformer_blocks.{i}."
301
+ # norms.
302
+ ## norm1
303
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}norm1.linear.weight", f"double_blocks.{i}.img_mod.lin.weight")
304
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}norm1.linear.bias", f"double_blocks.{i}.img_mod.lin.bias")
305
+ ## norm1_context
306
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}norm1_context.linear.weight", f"double_blocks.{i}.txt_mod.lin.weight")
307
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}norm1_context.linear.bias", f"double_blocks.{i}.txt_mod.lin.bias")
308
+ # Q, K, V
309
+ convqkv(converted_state_dict, original_state_dict, i)
310
+ # qk_norm
311
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}attn.norm_q.weight", f"double_blocks.{i}.img_attn.norm.query_norm.scale")
312
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}attn.norm_k.weight", f"double_blocks.{i}.img_attn.norm.key_norm.scale")
313
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}attn.norm_added_q.weight", f"double_blocks.{i}.txt_attn.norm.query_norm.scale")
314
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}attn.norm_added_k.weight", f"double_blocks.{i}.txt_attn.norm.key_norm.scale")
315
+ # ff img_mlp
316
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}ff.net.0.proj.weight", f"double_blocks.{i}.img_mlp.0.weight")
317
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}ff.net.0.proj.bias", f"double_blocks.{i}.img_mlp.0.bias")
318
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}ff.net.2.weight", f"double_blocks.{i}.img_mlp.2.weight")
319
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}ff.net.2.bias", f"double_blocks.{i}.img_mlp.2.bias")
320
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}ff_context.net.0.proj.weight", f"double_blocks.{i}.txt_mlp.0.weight")
321
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}ff_context.net.0.proj.bias", f"double_blocks.{i}.txt_mlp.0.bias")
322
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}ff_context.net.2.weight", f"double_blocks.{i}.txt_mlp.2.weight")
323
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}ff_context.net.2.bias", f"double_blocks.{i}.txt_mlp.2.bias")
324
+ # output projections.
325
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}attn.to_out.0.weight", f"double_blocks.{i}.img_attn.proj.weight")
326
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}attn.to_out.0.bias", f"double_blocks.{i}.img_attn.proj.bias")
327
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}attn.to_add_out.weight", f"double_blocks.{i}.txt_attn.proj.weight")
328
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}attn.to_add_out.bias", f"double_blocks.{i}.txt_attn.proj.bias")
329
+
330
+ progress(0.5, desc="Converting FLUX.1 state dict to Diffusers format.")
331
+ # single transfomer blocks
332
+ for i in range(num_single_layers):
333
+ block_prefix = f"single_transformer_blocks.{i}."
334
+ # norm.linear <- single_blocks.0.modulation.lin
335
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}norm.linear.weight", f"single_blocks.{i}.modulation.lin.weight")
336
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}norm.linear.bias", f"single_blocks.{i}.modulation.lin.bias")
337
+ # Q, K, V, mlp
338
+ convqkvmlp(converted_state_dict, original_state_dict, i, inner_dim, mlp_ratio)
339
+ # qk norm
340
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}attn.norm_q.weight", f"single_blocks.{i}.norm.query_norm.scale")
341
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}attn.norm_k.weight", f"single_blocks.{i}.norm.key_norm.scale")
342
+ # output projections.
343
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}proj_out.weight", f"single_blocks.{i}.linear2.weight")
344
+ conv(converted_state_dict, original_state_dict, f"{block_prefix}proj_out.bias", f"single_blocks.{i}.linear2.bias")
345
+
346
+ progress(0.75, desc="Converting FLUX.1 state dict to Diffusers format.")
347
+ conv(converted_state_dict, original_state_dict, "proj_out.weight", "final_layer.linear.weight")
348
+ conv(converted_state_dict, original_state_dict, "proj_out.bias", "final_layer.linear.bias")
349
+ convswap(converted_state_dict, original_state_dict, "norm_out.linear.weight", "final_layer.adaLN_modulation.1.weight")
350
+ convswap(converted_state_dict, original_state_dict, "norm_out.linear.bias", "final_layer.adaLN_modulation.1.bias")
351
+
352
+ progress(1, desc="Converting FLUX.1 state dict to Diffusers format.")
353
+ return converted_state_dict
354
+
355
+ # read safetensors metadata
356
+ def read_safetensors_metadata(path):
357
+ with open(path, 'rb') as f:
358
+ header_size = int.from_bytes(f.read(8), 'little')
359
+ header_json = f.read(header_size).decode('utf-8')
360
+ header = json.loads(header_json)
361
+ metadata = header.get('__metadata__', {})
362
+ return metadata
363
+
364
+ def normalize_key(k: str):
365
+ return k.replace("vae.", "").replace("model.diffusion_model.", "")\
366
+ .replace("text_encoders.clip_l.transformer.text_model.", "")\
367
+ .replace("text_encoders.t5xxl.transformer.", "")
368
+
369
+ def load_json_list(path: str):
370
+ try:
371
+ with open(path, encoding='utf-8') as f:
372
+ return list(json.load(f))
373
+ except Exception as e:
374
+ print(e)
375
+ return []
376
+
377
+ # https://github.com/huggingface/diffusers/blob/main/src/diffusers/models/modeling_utils.py
378
+ # https://huggingface.co/docs/huggingface_hub/v0.24.5/package_reference/serialization
379
+ # https://huggingface.co/docs/huggingface_hub/index
380
+ with torch.no_grad():
381
+ def to_safetensors(sd: dict, path: str, pattern: str, size: str, progress=gr.Progress(track_tqdm=True)):
382
+ from huggingface_hub import save_torch_state_dict
383
+ print(f"Saving a temporary file to disk: {path}")
384
+ os.makedirs(path, exist_ok=True)
385
+ try:
386
+ for k, v in sd.items():
387
+ sd[k] = v.to(device="cpu")
388
+ save_torch_state_dict(sd, path, filename_pattern=pattern, max_shard_size=size)
389
+ except Exception as e:
390
+ print(e)
391
+
392
+ # https://discuss.huggingface.co/t/t5forconditionalgeneration-checkpoint-size-mismatch-19418/24119
393
+ # https://github.com/huggingface/transformers/issues/13769
394
+ # https://github.com/huggingface/optimum-quanto/issues/278
395
+ # https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/serialization/_torch.py
396
+ with torch.no_grad():
397
+ def to_safetensors_flux_module(sd: dict, path: str, pattern: str, size: str,
398
+ quantization: bool=False, name: str = "",
399
+ metadata: dict | None = None, progress=gr.Progress(track_tqdm=True)):
400
+ from huggingface_hub import save_torch_state_dict
401
+ try:
402
+ progress(0, desc=f"Preparing to save FLUX.1 {name} to Diffusers format.")
403
+ print(f"Preparing to save FLUX.1 {name} to Diffusers format.")
404
+ for k, v in sd.items():
405
+ sd[k] = v.to(device="cpu")
406
+ progress(0, desc=f"Loading FLUX.1 {name}.")
407
+ print(f"Loading FLUX.1 {name}.")
408
+ os.makedirs(path, exist_ok=True)
409
+ if quantization:
410
+ progress(0.5, desc=f"Saving quantized FLUX.1 {name} to {path}")
411
+ print(f"Saving quantized FLUX.1 {name} to {path}")
412
+ else:
413
+ progress(0.5, desc=f"Saving FLUX.1 {name} to: {path}")
414
+ print(f"Saving FLUX.1 {name} to: {path}")
415
+ if metadata is not None:
416
+ save_torch_state_dict(state_dict=sd, save_directory=path,
417
+ filename_pattern=pattern, max_shard_size=size, metadata=metadata)
418
+ else:
419
+ save_torch_state_dict(state_dict=sd, save_directory=path,
420
+ filename_pattern=pattern, max_shard_size=size)
421
+ progress(1, desc=f"Saved FLUX.1 {name} to: {path}")
422
+ print(f"Saved FLUX.1 {name} to: {path}")
423
+ except Exception as e:
424
+ print(e)
425
+ finally:
426
+ gc.collect()
427
+
428
+ flux_transformer_json = "flux_transformer_keys.json"
429
+ flux_t5xxl_json = "flux_t5xxl_keys.json"
430
+ flux_clip_json = "flux_clip_keys.json"
431
+ flux_vae_json = "flux_vae_keys.json"
432
+ keys_flux_t5xxl = set(load_json_list(flux_t5xxl_json))
433
+ keys_flux_transformer = set(load_json_list(flux_transformer_json))
434
+ keys_flux_clip = set(load_json_list(flux_clip_json))
435
+ keys_flux_vae = set(load_json_list(flux_vae_json))
436
+
437
+ with torch.no_grad():
438
+ def dequant_tensor(v: torch.Tensor, dtype: torch.dtype, dequant: bool):
439
+ try:
440
+ #print(f"shape: {v.shape} / dim: {v.ndim}")
441
+ if dequant:
442
+ qtype = v.tensor_type
443
+ if v.dtype in TORCH_DTYPE: return v.to(dtype) if v.dtype != dtype else v
444
+ elif qtype in GGUF_QTYPE: return dequantize_tensor(v, dtype)
445
+ elif torch.dtype in TORCH_QUANTIZED_DTYPE: return torch.dequantize(v).to(dtype)
446
+ else: return torch.dequantize(v).to(dtype)
447
+ else: return v.to(dtype) if v.dtype != dtype else v
448
+ except Exception as e:
449
+ print(e)
450
+
451
+ with torch.no_grad():
452
+ def normalize_flux_state_dict(path: str, savepath: str, dtype: torch.dtype = torch.bfloat16,
453
+ dequant: bool = False, progress=gr.Progress(track_tqdm=True)):
454
+ progress(0, desc=f"Loading and normalizing FLUX.1 safetensors: {path}")
455
+ print(f"Loading and normalizing FLUX.1 safetensors: {path}")
456
+ new_sd = dict()
457
+ state_dict = load_file(path, device="cpu")
458
+ try:
459
+ for k in list(state_dict.keys()):
460
+ v = state_dict.pop(k)
461
+ nk = normalize_key(k)
462
+ print(f"{k} => {nk}") #
463
+ new_sd[nk] = dequant_tensor(v, dtype, dequant)
464
+ except Exception as e:
465
+ print(e)
466
+ return
467
+ finally:
468
+ del state_dict
469
+ torch.cuda.empty_cache()
470
+ gc.collect()
471
+ new_path = str(Path(savepath, Path(path).stem + "_fixed" + Path(path).suffix))
472
+ metadata = read_safetensors_metadata(path)
473
+ progress(0.5, desc=f"Saving FLUX.1 safetensors: {new_path}")
474
+ print(f"Saving FLUX.1 safetensors: {new_path}")
475
+ os.makedirs(savepath, exist_ok=True)
476
+ save_file(new_sd, new_path, metadata={"format": "pt", **metadata})
477
+ progress(1, desc=f"Saved FLUX.1 safetensors: {new_path}")
478
+ print(f"Saved FLUX.1 safetensors: {new_path}")
479
+ del new_sd
480
+ torch.cuda.empty_cache()
481
+ gc.collect()
482
+
483
+ with torch.no_grad():
484
+ def extract_norm_flux_module_sd(path: str, dtype: torch.dtype = torch.bfloat16,
485
+ dequant: bool = False, name: str = "", keys: set = {},
486
+ progress=gr.Progress(track_tqdm=True)):
487
+ progress(0, desc=f"Loading and normalizing FLUX.1 {name} safetensors: {path}")
488
+ print(f"Loading and normalizing FLUX.1 {name} safetensors: {path}")
489
+ new_sd = dict()
490
+ state_dict = load_file(path, device="cpu")
491
+ try:
492
+ for k in list(state_dict.keys()):
493
+ if k not in keys: state_dict.pop(k)
494
+ gc.collect()
495
+ for k in list(state_dict.keys()):
496
+ v = state_dict.pop(k)
497
+ if k in keys:
498
+ nk = normalize_key(k)
499
+ progress(0.5, desc=f"{k} => {nk}") #
500
+ print(f"{k} => {nk}") #
501
+ new_sd[nk] = dequant_tensor(v, dtype, dequant)
502
+ #print_resource_usage() #
503
+ except Exception as e:
504
+ print(e)
505
+ return None
506
+ finally:
507
+ progress(1, desc=f"Normalized FLUX.1 {name} safetensors: {path}")
508
+ print(f"Normalized FLUX.1 {name} safetensors: {path}")
509
+ del state_dict
510
+ torch.cuda.empty_cache()
511
+ gc.collect()
512
+ return new_sd
513
+
514
+ with torch.no_grad():
515
+ def convert_flux_transformer_sd_to_diffusers(sd: dict, progress=gr.Progress(track_tqdm=True)):
516
+ progress(0, desc="Converting FLUX.1 state dict to Diffusers format.")
517
+ print("Converting FLUX.1 state dict to Diffusers format.")
518
+ num_layers = 19
519
+ num_single_layers = 38
520
+ inner_dim = 3072
521
+ mlp_ratio = 4.0
522
+ try:
523
+ sd = convert_flux_transformer_checkpoint_to_diffusers(
524
+ sd, num_layers, num_single_layers, inner_dim, mlp_ratio=mlp_ratio
525
+ )
526
+ except Exception as e:
527
+ print(e)
528
+ finally:
529
+ progress(1, desc="Converted FLUX.1 state dict to Diffusers format.")
530
+ print("Converted FLUX.1 state dict to Diffusers format.")
531
+ gc.collect()
532
+ return sd
533
+
534
+ with torch.no_grad():
535
+ def load_sharded_safetensors(path: str):
536
+ import glob
537
+ sd = {}
538
+ try:
539
+ for filepath in glob.glob(f"{path}/*.safetensors"):
540
+ sharded_sd = load_file(str(filepath), device="cpu")
541
+ for k, v in sharded_sd.items():
542
+ sharded_sd[k] = v.to(device="cpu")
543
+ sd = sd | sharded_sd.copy()
544
+ del sharded_sd
545
+ torch.cuda.empty_cache()
546
+ gc.collect()
547
+ except Exception as e:
548
+ print(e)
549
+ return sd
550
+
551
+ # https://huggingface.co/docs/safetensors/api/torch
552
+ with torch.no_grad():
553
+ def convert_flux_transformer_sd_to_diffusers_sharded(sd: dict, path: str, pattern: str,
554
+ size: str, progress=gr.Progress(track_tqdm=True)):
555
+ from huggingface_hub import save_torch_state_dict#, load_torch_model
556
+ import glob
557
+ try:
558
+ progress(0, desc=f"Saving temporary files to disk: {path}")
559
+ print(f"Saving temporary files to disk: {path}")
560
+ os.makedirs(path, exist_ok=True)
561
+ for k, v in sd.items():
562
+ if k in set(keys_flux_transformer): sd[k] = v.to(device="cpu")
563
+ save_torch_state_dict(sd, path, filename_pattern=pattern, max_shard_size=size)
564
+ del sd
565
+ torch.cuda.empty_cache()
566
+ gc.collect()
567
+ progress(0.25, desc=f"Saved temporary files to disk: {path}")
568
+ print(f"Saved temporary files to disk: {path}")
569
+ for filepath in glob.glob(f"{path}/*.safetensors"):
570
+ progress(0.25, desc=f"Processing temporary files: {str(filepath)}")
571
+ print(f"Processing temporary files: {str(filepath)}")
572
+ sharded_sd = load_file(str(filepath), device="cpu")
573
+ sharded_sd = convert_flux_transformer_sd_to_diffusers(sharded_sd)
574
+ for k, v in sharded_sd.items():
575
+ sharded_sd[k] = v.to(device="cpu")
576
+ save_file(sharded_sd, str(filepath))
577
+ del sharded_sd
578
+ torch.cuda.empty_cache()
579
+ gc.collect()
580
+ print(f"Loading temporary files from disk: {path}")
581
+ sd = load_sharded_safetensors(path)
582
+ print(f"Loaded temporary files from disk: {path}")
583
+ except Exception as e:
584
+ print(e)
585
+ return sd
586
+
587
+ with torch.no_grad():
588
+ def extract_normalized_flux_state_dict_sharded(loadpath: str, dtype: torch.dtype,
589
+ dequant: bool, path: str, pattern: str, size: str, progress=gr.Progress(track_tqdm=True)):
590
+ from huggingface_hub import save_torch_state_dict#, load_torch_model
591
+ import glob
592
+ try:
593
+ progress(0, desc=f"Loading model file: {loadpath}")
594
+ print(f"Loading model file: {loadpath}")
595
+ sd = load_file(loadpath, device="cpu")
596
+ progress(0, desc=f"Saving temporary files to disk: {path}")
597
+ print(f"Saving temporary files to disk: {path}")
598
+ os.makedirs(path, exist_ok=True)
599
+ for k, v in sd.items():
600
+ sd[k] = v.to(device="cpu")
601
+ save_torch_state_dict(sd, path, filename_pattern=pattern, max_shard_size=size)
602
+ del sd
603
+ torch.cuda.empty_cache()
604
+ gc.collect()
605
+ progress(0.25, desc=f"Saved temporary files to disk: {path}")
606
+ print(f"Saved temporary files to disk: {path}")
607
+ for filepath in glob.glob(f"{path}/*.safetensors"):
608
+ progress(0.25, desc=f"Processing temporary files: {str(filepath)}")
609
+ print(f"Processing temporary files: {str(filepath)}")
610
+ sharded_sd = extract_normalized_flux_state_dict_unet(str(filepath), dtype, dequant)
611
+ for k, v in sharded_sd.items():
612
+ sharded_sd[k] = v.to(device="cpu")
613
+ save_file(sharded_sd, str(filepath))
614
+ del sharded_sd
615
+ torch.cuda.empty_cache()
616
+ gc.collect()
617
+ print(f"Processed temporary files: {str(filepath)}")
618
+ print(f"Loading temporary files from disk: {path}")
619
+ sd = load_sharded_safetensors(path)
620
+ print(f"Loaded temporary files from disk: {path}")
621
+ except Exception as e:
622
+ print(e)
623
+ return sd
624
+
625
+ def download_repo(repo_name, path, download_sf=False, progress=gr.Progress(track_tqdm=True)):
626
+ from huggingface_hub import snapshot_download
627
+ print(f"Downloading {repo_name}.")
628
+ try:
629
+ if download_sf:
630
+ snapshot_download(repo_id=repo_name, local_dir=path, ignore_patterns=["transformer/", "*.sft", ".*", "README*", "*.md", "*.index", "*.jpg", "*.png", "*.webp"])
631
+ else:
632
+ snapshot_download(repo_id=repo_name, local_dir=path, ignore_patterns=["transformer/", "text_encoder_2/", "*.sft", ".*", "README*", "*.md", "*.index", "*.jpg", "*.png", "*.webp"])
633
+ except Exception as e:
634
+ print(e)
635
+
636
+ def copy_nontensor_files(from_path, to_path, copy_sf=False):
637
+ import shutil
638
+ if copy_sf:
639
+ print(f"Copying non-tensor files {from_path} to {to_path}")
640
+ shutil.copytree(from_path, to_path, ignore=shutil.ignore_patterns("*.safetensors", "*.bin", "*.sft", ".*", "README*", "*.md", "*.index", "*.jpg", "*.png", "*.webp"), dirs_exist_ok=True)
641
+ te_from = str(Path(from_path, "text_encoder_2"))
642
+ te_to = str(Path(to_path, "text_encoder_2"))
643
+ print(f"Copying Text Encoder 2 files {te_from} to {te_to}")
644
+ shutil.copytree(te_from, te_to, ignore=shutil.ignore_patterns(".*", "README*", "*.md", "*.jpg", "*.png", "*.webp"), dirs_exist_ok=True)
645
+ te1_from = str(Path(from_path, "text_encoder"))
646
+ te1_to = str(Path(to_path, "text_encoder"))
647
+ print(f"Copying Text Encoder 1 files {te1_from} to {te1_to}")
648
+ shutil.copytree(te1_from, te1_to, ignore=shutil.ignore_patterns(".*", "README*", "*.md", "*.jpg", "*.png", "*.webp"), dirs_exist_ok=True)
649
+ tn2_from = str(Path(from_path, "tokenizer_2"))
650
+ tn2_to = str(Path(to_path, "tokenizer_2"))
651
+ print(f"Copying Tokenizer 2 files {tn2_from} to {tn2_to}")
652
+ shutil.copytree(tn2_from, tn2_to, ignore=shutil.ignore_patterns(".*", "README*", "*.md", "*.jpg", "*.png", "*.webp"), dirs_exist_ok=True)
653
+ vae_from = str(Path(from_path, "vae"))
654
+ vae_to = str(Path(to_path, "vae"))
655
+ print(f"Copying VAE files {vae_from} to {vae_to}")
656
+ shutil.copytree(vae_from, vae_to, ignore=shutil.ignore_patterns(".*", "README*", "*.md", "*.jpg", "*.png", "*.webp"), dirs_exist_ok=True)
657
+ else:
658
+ print(f"Copying non-tensor files {from_path} to {to_path}")
659
+ shutil.copytree(from_path, to_path, ignore=shutil.ignore_patterns("*.safetensors", "*.bin", "*.sft", ".*", "README*", "*.md", "*.index", "*.jpg", "*.png", "*.webp"), dirs_exist_ok=True)
660
+ te1_from = str(Path(from_path, "text_encoder"))
661
+ te1_to = str(Path(to_path, "text_encoder"))
662
+ print(f"Copying Text Encoder 1 files {te1_from} to {te1_to}")
663
+ shutil.copytree(te1_from, te1_to, ignore=shutil.ignore_patterns(".*", "README*", "*.md", "*.jpg", "*.png", "*.webp"), dirs_exist_ok=True)
664
+ tn2_from = str(Path(from_path, "tokenizer_2"))
665
+ tn2_to = str(Path(to_path, "tokenizer_2"))
666
+ print(f"Copying Tokenizer 2 files {tn2_from} to {tn2_to}")
667
+ shutil.copytree(tn2_from, tn2_to, ignore=shutil.ignore_patterns(".*", "README*", "*.md", "*.jpg", "*.png", "*.webp"), dirs_exist_ok=True)
668
+ vae_from = str(Path(from_path, "vae"))
669
+ vae_to = str(Path(to_path, "vae"))
670
+ print(f"Copying VAE files {vae_from} to {vae_to}")
671
+ shutil.copytree(vae_from, vae_to, ignore=shutil.ignore_patterns(".*", "README*", "*.md", "*.jpg", "*.png", "*.webp"), dirs_exist_ok=True)
672
+
673
+ def save_flux_other_diffusers(path: str, model_type: str = "dev", copy_te: bool = False, progress=gr.Progress(track_tqdm=True)):
674
+ import shutil
675
+ progress(0, desc="Loading FLUX.1 Components.")
676
+ print("Loading FLUX.1 Components.")
677
+ temppath = system_temp_dir
678
+ if model_type == "schnell": repo = flux_schnell_repo
679
+ else: repo = flux_dev_repo
680
+ os.makedirs(temppath, exist_ok=True)
681
+ os.makedirs(path, exist_ok=True)
682
+ download_repo(repo, temppath, copy_te)
683
+ progress(0.5, desc="Saving FLUX.1 Components.")
684
+ print("Saving FLUX.1 Components.")
685
+ copy_nontensor_files(temppath, path, copy_te)
686
+ shutil.rmtree(temppath)
687
+
688
+ with torch.no_grad():
689
+ def fix_flux_safetensors(loadpath: str, savepath: str, dtype: torch.dtype = torch.bfloat16,
690
+ quantization: bool = False, model_type: str = "dev", dequant: bool = False):
691
+ save_flux_other_diffusers(savepath, model_type)
692
+ normalize_flux_state_dict(loadpath, savepath, dtype, dequant)
693
+ torch.cuda.empty_cache()
694
+ gc.collect()
695
+
696
+ with torch.no_grad(): # Much lower memory consumption, but higher disk load
697
+ def flux_to_diffusers_lowmem(loadpath: str, savepath: str, dtype: torch.dtype = torch.bfloat16,
698
+ quantization: bool = False, model_type: str = "dev",
699
+ dequant: bool = False, use_original: list = ["vae", "text_encoder"],
700
+ new_repo_id: str = "", local: bool = False, progress=gr.Progress(track_tqdm=True)):
701
+ unet_sd_path = savepath.removesuffix("/") + "/transformer"
702
+ unet_sd_pattern = "diffusion_pytorch_model{suffix}.safetensors"
703
+ unet_sd_size = "10GB"
704
+ te_sd_path = savepath.removesuffix("/") + "/text_encoder_2"
705
+ te_sd_pattern = "model{suffix}.safetensors"
706
+ te_sd_size = "5GB"
707
+ clip_sd_path = savepath.removesuffix("/") + "/text_encoder"
708
+ clip_sd_pattern = "model{suffix}.safetensors"
709
+ clip_sd_size = "10GB"
710
+ vae_sd_path = savepath.removesuffix("/") + "/vae"
711
+ vae_sd_pattern = "diffusion_pytorch_model{suffix}.safetensors"
712
+ vae_sd_size = "10GB"
713
+ metadata = {"format": "pt", **read_safetensors_metadata(loadpath)}
714
+ save_flux_other_diffusers(savepath, model_type, "text_encoder_2" in use_original)
715
+ if "vae" not in use_original:
716
+ vae_sd = extract_norm_flux_module_sd(loadpath, torch.bfloat16, dequant, "VAE",
717
+ keys_flux_vae)
718
+ to_safetensors_flux_module(vae_sd, vae_sd_path, vae_sd_pattern, vae_sd_size,
719
+ quantization, "VAE", None)
720
+ del vae_sd
721
+ torch.cuda.empty_cache()
722
+ gc.collect()
723
+ if "text_encoder" not in use_original:
724
+ clip_sd = extract_norm_flux_module_sd(loadpath, torch.bfloat16, dequant, "Text Encoder",
725
+ keys_flux_clip)
726
+ to_safetensors_flux_module(clip_sd, clip_sd_path, clip_sd_pattern, clip_sd_size,
727
+ quantization, "Text Encoder", None)
728
+ del clip_sd
729
+ torch.cuda.empty_cache()
730
+ gc.collect()
731
+ if "text_encoder_2" not in use_original:
732
+ te_sd = extract_norm_flux_module_sd(loadpath, dtype, dequant, "Text Encoder 2",
733
+ keys_flux_t5xxl)
734
+ to_safetensors_flux_module(te_sd, te_sd_path, te_sd_pattern, te_sd_size,
735
+ quantization, "Text Encoder 2", None)
736
+ del te_sd
737
+ torch.cuda.empty_cache()
738
+ gc.collect()
739
+ unet_sd = extract_norm_flux_module_sd(loadpath, dtype, dequant, "Transformer",
740
+ keys_flux_transformer)
741
+ if not local: os.remove(loadpath)
742
+ to_safetensors_flux_module(unet_sd, unet_sd_path, unet_sd_pattern, unet_sd_size,
743
+ quantization, "Transformer", metadata)
744
+ del unet_sd
745
+ torch.cuda.empty_cache()
746
+ gc.collect()
747
+
748
+ with torch.no_grad(): # lowest memory consumption, but higheest disk load
749
+ def flux_to_diffusers_lowmem2(loadpath: str, savepath: str, dtype: torch.dtype = torch.bfloat16,
750
+ quantization: bool = False, model_type: str = "dev",
751
+ dequant: bool = False, use_original: list = ["vae", "text_encoder"],
752
+ new_repo_id: str = "", progress=gr.Progress(track_tqdm=True)):
753
+ unet_sd_path = savepath.removesuffix("/") + "/transformer"
754
+ unet_temp_path = system_temp_dir.removesuffix("/") + "/sharded"
755
+ unet_sd_pattern = "diffusion_pytorch_model{suffix}.safetensors"
756
+ unet_sd_size = "10GB"
757
+ unet_temp_size = "5GB"
758
+ te_sd_path = savepath.removesuffix("/") + "/text_encoder_2"
759
+ te_sd_pattern = "model{suffix}.safetensors"
760
+ te_sd_size = "5GB"
761
+ clip_sd_path = savepath.removesuffix("/") + "/text_encoder"
762
+ clip_sd_pattern = "model{suffix}.safetensors"
763
+ clip_sd_size = "10GB"
764
+ vae_sd_path = savepath.removesuffix("/") + "/vae"
765
+ vae_sd_pattern = "diffusion_pytorch_model{suffix}.safetensors"
766
+ vae_sd_size = "10GB"
767
+ metadata = {"format": "pt", **read_safetensors_metadata(loadpath)}
768
+ save_flux_other_diffusers(savepath, model_type, "text_encoder_2" in use_original)
769
+ if "vae" not in use_original:
770
+ vae_sd = extract_norm_flux_module_sd(loadpath, torch.bfloat16, dequant, "VAE",
771
+ keys_flux_vae)
772
+ to_safetensors_flux_module(vae_sd, vae_sd_path, vae_sd_pattern, vae_sd_size,
773
+ quantization, "VAE", None)
774
+ del vae_sd
775
+ torch.cuda.empty_cache()
776
+ gc.collect()
777
+ if "text_encoder" not in use_original:
778
+ clip_sd = extract_norm_flux_module_sd(loadpath, torch.bfloat16, dequant, "Text Encoder",
779
+ keys_flux_clip)
780
+ to_safetensors_flux_module(clip_sd, clip_sd_path, clip_sd_pattern, clip_sd_size,
781
+ quantization, "Text Encoder", None)
782
+ del clip_sd
783
+ torch.cuda.empty_cache()
784
+ gc.collect()
785
+ if "text_encoder_2" not in use_original:
786
+ te_sd = extract_norm_flux_module_sd(loadpath, dtype, dequant, "Text Encoder 2",
787
+ keys_flux_t5xxl)
788
+ to_safetensors_flux_module(te_sd, te_sd_path, te_sd_pattern, te_sd_size,
789
+ quantization, "Text Encoder 2", None)
790
+ del te_sd
791
+ torch.cuda.empty_cache()
792
+ gc.collect()
793
+ unet_sd = extract_normalized_flux_state_dict_sharded(loadpath, dtype, dequant,
794
+ unet_temp_path, unet_sd_pattern, unet_temp_size)
795
+ unet_sd = convert_flux_transformer_sd_to_diffusers_sharded(unet_sd, unet_temp_path,
796
+ unet_sd_pattern, unet_temp_size)
797
+ to_safetensors_flux_module(unet_sd, unet_sd_path, unet_sd_pattern, unet_sd_size,
798
+ quantization, "Transformer", metadata)
799
+ del unet_sd
800
+ torch.cuda.empty_cache()
801
+ gc.collect()
802
+
803
+ def convert_url_to_diffusers_flux(url, civitai_key="", is_upload_sf=False, data_type="bf16",
804
+ model_type="dev", dequant=False, use_original=["vae", "text_encoder"],
805
+ hf_user="", hf_repo="", q=None, progress=gr.Progress(track_tqdm=True)):
806
+ progress(0, desc="Start converting...")
807
+ temp_dir = "."
808
+ new_file = get_download_file(temp_dir, url, civitai_key)
809
+ if not new_file:
810
+ print(f"Not found: {url}")
811
+ return ""
812
+ new_repo_name = Path(new_file).stem.replace(" ", "_").replace(",", "_").replace(".", "_") #
813
+
814
+ dtype = torch.bfloat16
815
+ quantization = False
816
+ if data_type == "fp8": dtype = torch.float8_e4m3fn
817
+ elif data_type == "fp16": dtype = torch.float16
818
+ elif data_type == "qfloat8":
819
+ dtype = torch.bfloat16
820
+ quantization = True
821
+ else: dtype = torch.bfloat16
822
+
823
+ new_repo_id = f"{hf_user}/{Path(new_repo_name).stem}"
824
+ if hf_repo != "": new_repo_id = f"{hf_user}/{hf_repo}"
825
+ flux_to_diffusers_lowmem(new_file, new_repo_name, dtype, quantization, model_type, dequant, use_original, new_repo_id)
826
+
827
+ """if is_upload_sf:
828
+ import shutil
829
+ shutil.move(str(Path(new_file).resolve()), str(Path(new_repo_name, Path(new_file).name).resolve()))
830
+ else: os.remove(new_file)"""
831
+
832
+ progress(1, desc="Converted.")
833
+ q.put(new_repo_name)
834
+ return new_repo_name
835
+
836
+ def convert_url_to_fixed_flux_safetensors(url, civitai_key="", is_upload_sf=False, data_type="bf16",
837
+ model_type="dev", dequant=False, q=None, progress=gr.Progress(track_tqdm=True)):
838
+ progress(0, desc="Start converting...")
839
+ temp_dir = "."
840
+ new_file = get_download_file(temp_dir, url, civitai_key)
841
+ if not new_file:
842
+ print(f"Not found: {url}")
843
+ return ""
844
+ new_repo_name = Path(new_file).stem.replace(" ", "_").replace(",", "_").replace(".", "_") #
845
+
846
+ dtype = torch.bfloat16
847
+ quantization = False
848
+ if data_type == "fp8": dtype = torch.float8_e4m3fn
849
+ elif data_type == "fp16": dtype = torch.float16
850
+ elif data_type == "qfloat8":
851
+ dtype = torch.bfloat16
852
+ quantization = True
853
+ else: dtype = torch.bfloat16
854
+
855
+ fix_flux_safetensors(new_file, new_repo_name, dtype, model_type, dequant)
856
+
857
+ os.remove(new_file)
858
+
859
+ progress(1, desc="Converted.")
860
+ q.put(new_repo_name)
861
+ return new_repo_name
862
+
863
+ def convert_url_to_diffusers_repo_flux(dl_url, hf_user, hf_repo, hf_token, civitai_key="",
864
+ is_upload_sf=False, data_type="bf16", model_type="dev", dequant=False,
865
+ repo_urls=[], fix_only=False, use_original=["vae", "text_encoder"],
866
+ progress=gr.Progress(track_tqdm=True)):
867
+ import multiprocessing as mp
868
+ import shutil
869
+ if not hf_user:
870
+ print(f"Invalid user name: {hf_user}")
871
+ progress(1, desc=f"Invalid user name: {hf_user}")
872
+ return gr.update(value=repo_urls, choices=repo_urls), gr.update(value="")
873
+ if hf_token and not os.environ.get("HF_TOKEN"): os.environ['HF_TOKEN'] = hf_token
874
+ if not civitai_key and os.environ.get("CIVITAI_API_KEY"): civitai_key = os.environ.get("CIVITAI_API_KEY")
875
+ q = mp.Queue()
876
+ if fix_only:
877
+ p = mp.Process(target=convert_url_to_fixed_flux_safetensors, args=(dl_url, civitai_key,
878
+ is_upload_sf, data_type, model_type, dequant, q))
879
+ #new_path = convert_url_to_fixed_flux_safetensors(dl_url, civitai_key, is_upload_sf, data_type, model_type, dequant)
880
+ else:
881
+ p = mp.Process(target=convert_url_to_diffusers_flux, args=(dl_url, civitai_key,
882
+ is_upload_sf, data_type, model_type, dequant, use_original, hf_user, hf_repo, q))
883
+ #new_path = convert_url_to_diffusers_flux(dl_url, civitai_key, is_upload_sf, data_type, model_type, dequant)
884
+ p.start()
885
+ new_path = q.get()
886
+ p.join()
887
+ if not new_path: return ""
888
+ new_repo_id = f"{hf_user}/{Path(new_path).stem}"
889
+ if hf_repo != "": new_repo_id = f"{hf_user}/{hf_repo}"
890
+ if not is_repo_name(new_repo_id):
891
+ print(f"Invalid repo name: {new_repo_id}")
892
+ progress(1, desc=f"Invalid repo name: {new_repo_id}")
893
+ return gr.update(value=repo_urls, choices=repo_urls), gr.update(value="")
894
+ if is_repo_exists(new_repo_id):
895
+ print(f"Repo already exists: {new_repo_id}")
896
+ progress(1, desc=f"Repo already exists: {new_repo_id}")
897
+ return gr.update(value=repo_urls, choices=repo_urls), gr.update(value="")
898
+ save_readme_md(new_path, dl_url)
899
+ repo_url = create_diffusers_repo(new_repo_id, new_path)
900
+ shutil.rmtree(new_path)
901
+ if not repo_urls: repo_urls = []
902
+ repo_urls.append(repo_url)
903
+ md = "Your new repo:<br>"
904
+ for u in repo_urls:
905
+ md += f"[{str(u).split('/')[-2]}/{str(u).split('/')[-1]}]({str(u)})<br>"
906
+ return gr.update(value=repo_urls, choices=repo_urls), gr.update(value=md)
907
+
908
+ if __name__ == "__main__":
909
+ parser = argparse.ArgumentParser()
910
+ parser.add_argument("--url", default=None, type=str, required=False, help="URL of the model to convert.")
911
+ parser.add_argument("--file", default=None, type=str, required=False, help="Filename of the model to convert.")
912
+ parser.add_argument("--fix", action="store_true", help="Only fix the keys of the local model.")
913
+ parser.add_argument("--civitai_key", default=None, type=str, required=False, help="Civitai API Key (If you want to download file from Civitai).")
914
+ parser.add_argument("--dtype", type=str, default="fp8")
915
+ parser.add_argument("--model", type=str, default="dev")
916
+ parser.add_argument("--dequant", action="store_true", help="Dequantize model.")
917
+ args = parser.parse_args()
918
+ assert (args.url, args.file) != (None, None), "Must provide --url or --file!"
919
+
920
+ dtype = torch.bfloat16
921
+ quantization = False
922
+ if args.dtype == "fp8": dtype = torch.float8_e4m3fn
923
+ elif args.dtype == "fp16": dtype = torch.float16
924
+ elif args.dtype == "qfloat8":
925
+ dtype = torch.bfloat16
926
+ quantization = True
927
+ else: dtype = torch.bfloat16
928
+
929
+ use_original = ["vae", "text_encoder"]
930
+ new_repo_id = ""
931
+ use_local = True
932
+
933
+ if args.file is not None and Path(args.file).exists():
934
+ if args.fix: normalize_flux_state_dict(args.file, ".", dtype, args.dequant)
935
+ else: flux_to_diffusers_lowmem(args.file, Path(args.file).stem, dtype, quantization,
936
+ args.model, args.dequant, use_original, new_repo_id, use_local)
937
+ elif args.url is not None:
938
+ convert_url_to_diffusers_flux(args.url, args.civitai_key, False, args.dtype, args.model,
939
+ args.dequant)
dequant.py ADDED
@@ -0,0 +1,114 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # (c) City96 || Apache-2.0 (apache.org/licenses/LICENSE-2.0)
2
+ import gguf
3
+ import torch
4
+ import numpy as np
5
+
6
+ def dequantize_tensor(tensor, dtype=torch.float16):
7
+ data = torch.tensor(tensor.data)
8
+ qtype = tensor.tensor_type
9
+ oshape = tensor.tensor_shape
10
+
11
+ if qtype == gguf.GGMLQuantizationType.F32:
12
+ return data.to(dtype)
13
+ elif qtype == gguf.GGMLQuantizationType.F16:
14
+ return data.to(dtype)
15
+ elif qtype in dequantize_functions:
16
+ # dequantize in fp16 then convert instead of keeping FP32
17
+ out = dequantize(data, qtype, oshape, dtype=None)
18
+ return out.to(dtype) if out.dtype != dtype else out # why is .to() not a no-op?
19
+ else:
20
+ # this is incredibly slow
21
+ new = gguf.quants.dequantize(data.cpu().numpy(), qtype)
22
+ return torch.from_numpy(new).to(data.device, dtype=dtype)
23
+
24
+ def dequantize(data, qtype, oshape, dtype=None):
25
+ """
26
+ Dequantize tensor back to usable shape/dtype
27
+ """
28
+ block_size, type_size = gguf.GGML_QUANT_SIZES[qtype]
29
+ dequantize_blocks = dequantize_functions[qtype]
30
+
31
+ rows = data.reshape(
32
+ (-1, data.shape[-1])
33
+ ).view(torch.uint8)
34
+
35
+ n_blocks = rows.numel() // type_size
36
+ blocks = rows.reshape((n_blocks, type_size))
37
+ blocks = dequantize_blocks(blocks, block_size, type_size, dtype)
38
+ return blocks.reshape(oshape)
39
+
40
+ def to_uint32(x):
41
+ # no uint32 :(
42
+ x = x.view(torch.uint8).to(torch.int32)
43
+ return (x[:, 0] | x[:, 1] << 8 | x[:, 2] << 16 | x[:, 3] << 24).unsqueeze(1)
44
+
45
+ def dequantize_blocks_Q8_0(blocks, block_size, type_size, dtype=None):
46
+ d = blocks[:, :2].view(torch.float16).to(dtype)
47
+ x = blocks[:, 2:].view(torch.int8)
48
+ return (d * x)
49
+
50
+ def dequantize_blocks_Q5_1(blocks, block_size, type_size, dtype=None):
51
+ n_blocks = blocks.shape[0]
52
+
53
+ d = blocks[:, :2].view(torch.float16).to(dtype)
54
+ m = blocks[:, 2:4].view(torch.float16).to(dtype)
55
+ qh = blocks[:, 4:8]
56
+ qs = blocks[:, 8: ]
57
+
58
+ qh = to_uint32(qh)
59
+
60
+ qh = qh.reshape((n_blocks, 1)) >> torch.arange(32, device=d.device, dtype=torch.int32).reshape(1, 32)
61
+ ql = qs.reshape((n_blocks, -1, 1, block_size // 2)) >> torch.tensor([0, 4], device=d.device, dtype=torch.uint8).reshape(1, 1, 2, 1)
62
+ qh = (qh & 1).to(torch.uint8)
63
+ ql = (ql & 0x0F).reshape((n_blocks, -1))
64
+
65
+ qs = (ql | (qh << 4))
66
+ return (d * qs) + m
67
+
68
+ def dequantize_blocks_Q5_0(blocks, block_size, type_size, dtype=None):
69
+ n_blocks = blocks.shape[0]
70
+
71
+ d = blocks[:, :2].view(torch.float16).to(dtype)
72
+ qh = blocks[:, 2:6]
73
+ qs = blocks[:, 6: ]
74
+
75
+ qh = to_uint32(qh)
76
+
77
+ qh = qh.reshape(n_blocks, 1) >> torch.arange(32, device=d.device, dtype=torch.int32).reshape(1, 32)
78
+ ql = qs.reshape(n_blocks, -1, 1, block_size // 2) >> torch.tensor([0, 4], device=d.device, dtype=torch.uint8).reshape(1, 1, 2, 1)
79
+
80
+ qh = (qh & 1).to(torch.uint8)
81
+ ql = (ql & 0x0F).reshape(n_blocks, -1)
82
+
83
+ qs = (ql | (qh << 4)).to(torch.int8) - 16
84
+ return (d * qs)
85
+
86
+ def dequantize_blocks_Q4_1(blocks, block_size, type_size, dtype=None):
87
+ n_blocks = blocks.shape[0]
88
+
89
+ d = blocks[:, :2].view(torch.float16).to(dtype)
90
+ m = blocks[:, 2:4].view(torch.float16).to(dtype)
91
+ qs = blocks[:, 4: ]
92
+
93
+ qs = qs.reshape((n_blocks, -1, 1, block_size // 2)) >> torch.tensor([0, 4], device=d.device, dtype=torch.uint8).reshape(1, 1, 2, 1)
94
+ qs = (qs & 0x0F).reshape(n_blocks, -1)
95
+
96
+ return (d * qs) + m
97
+
98
+ def dequantize_blocks_Q4_0(blocks, block_size, type_size, dtype=None):
99
+ n_blocks = blocks.shape[0]
100
+
101
+ d = blocks[:, :2].view(torch.float16).to(dtype)
102
+ qs = blocks[:, 2:]
103
+
104
+ qs = qs.reshape((n_blocks, -1, 1, block_size // 2)) >> torch.tensor([0, 4], device=d.device, dtype=torch.uint8).reshape((1, 1, 2, 1))
105
+ qs = (qs & 0x0F).reshape((n_blocks, -1)).to(torch.int8) - 8
106
+ return (d * qs)
107
+
108
+ dequantize_functions = {
109
+ gguf.GGMLQuantizationType.Q8_0: dequantize_blocks_Q8_0,
110
+ gguf.GGMLQuantizationType.Q5_1: dequantize_blocks_Q5_1,
111
+ gguf.GGMLQuantizationType.Q5_0: dequantize_blocks_Q5_0,
112
+ gguf.GGMLQuantizationType.Q4_1: dequantize_blocks_Q4_1,
113
+ gguf.GGMLQuantizationType.Q4_0: dequantize_blocks_Q4_0,
114
+ }
flux_clip_keys.json ADDED
@@ -0,0 +1,396 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ "embeddings.position_embedding.weight",
3
+ "embeddings.token_embedding.weight",
4
+ "encoder.layers.0.layer_norm1.bias",
5
+ "encoder.layers.0.layer_norm1.weight",
6
+ "encoder.layers.0.layer_norm2.bias",
7
+ "encoder.layers.0.layer_norm2.weight",
8
+ "encoder.layers.0.mlp.fc1.bias",
9
+ "encoder.layers.0.mlp.fc1.weight",
10
+ "encoder.layers.0.mlp.fc2.bias",
11
+ "encoder.layers.0.mlp.fc2.weight",
12
+ "encoder.layers.0.self_attn.k_proj.bias",
13
+ "encoder.layers.0.self_attn.k_proj.weight",
14
+ "encoder.layers.0.self_attn.out_proj.bias",
15
+ "encoder.layers.0.self_attn.out_proj.weight",
16
+ "encoder.layers.0.self_attn.q_proj.bias",
17
+ "encoder.layers.0.self_attn.q_proj.weight",
18
+ "encoder.layers.0.self_attn.v_proj.bias",
19
+ "encoder.layers.0.self_attn.v_proj.weight",
20
+ "encoder.layers.1.layer_norm1.bias",
21
+ "encoder.layers.1.layer_norm1.weight",
22
+ "encoder.layers.1.layer_norm2.bias",
23
+ "encoder.layers.1.layer_norm2.weight",
24
+ "encoder.layers.1.mlp.fc1.bias",
25
+ "encoder.layers.1.mlp.fc1.weight",
26
+ "encoder.layers.1.mlp.fc2.bias",
27
+ "encoder.layers.1.mlp.fc2.weight",
28
+ "encoder.layers.1.self_attn.k_proj.bias",
29
+ "encoder.layers.1.self_attn.k_proj.weight",
30
+ "encoder.layers.1.self_attn.out_proj.bias",
31
+ "encoder.layers.1.self_attn.out_proj.weight",
32
+ "encoder.layers.1.self_attn.q_proj.bias",
33
+ "encoder.layers.1.self_attn.q_proj.weight",
34
+ "encoder.layers.1.self_attn.v_proj.bias",
35
+ "encoder.layers.1.self_attn.v_proj.weight",
36
+ "encoder.layers.10.layer_norm1.bias",
37
+ "encoder.layers.10.layer_norm1.weight",
38
+ "encoder.layers.10.layer_norm2.bias",
39
+ "encoder.layers.10.layer_norm2.weight",
40
+ "encoder.layers.10.mlp.fc1.bias",
41
+ "encoder.layers.10.mlp.fc1.weight",
42
+ "encoder.layers.10.mlp.fc2.bias",
43
+ "encoder.layers.10.mlp.fc2.weight",
44
+ "encoder.layers.10.self_attn.k_proj.bias",
45
+ "encoder.layers.10.self_attn.k_proj.weight",
46
+ "encoder.layers.10.self_attn.out_proj.bias",
47
+ "encoder.layers.10.self_attn.out_proj.weight",
48
+ "encoder.layers.10.self_attn.q_proj.bias",
49
+ "encoder.layers.10.self_attn.q_proj.weight",
50
+ "encoder.layers.10.self_attn.v_proj.bias",
51
+ "encoder.layers.10.self_attn.v_proj.weight",
52
+ "encoder.layers.11.layer_norm1.bias",
53
+ "encoder.layers.11.layer_norm1.weight",
54
+ "encoder.layers.11.layer_norm2.bias",
55
+ "encoder.layers.11.layer_norm2.weight",
56
+ "encoder.layers.11.mlp.fc1.bias",
57
+ "encoder.layers.11.mlp.fc1.weight",
58
+ "encoder.layers.11.mlp.fc2.bias",
59
+ "encoder.layers.11.mlp.fc2.weight",
60
+ "encoder.layers.11.self_attn.k_proj.bias",
61
+ "encoder.layers.11.self_attn.k_proj.weight",
62
+ "encoder.layers.11.self_attn.out_proj.bias",
63
+ "encoder.layers.11.self_attn.out_proj.weight",
64
+ "encoder.layers.11.self_attn.q_proj.bias",
65
+ "encoder.layers.11.self_attn.q_proj.weight",
66
+ "encoder.layers.11.self_attn.v_proj.bias",
67
+ "encoder.layers.11.self_attn.v_proj.weight",
68
+ "encoder.layers.2.layer_norm1.bias",
69
+ "encoder.layers.2.layer_norm1.weight",
70
+ "encoder.layers.2.layer_norm2.bias",
71
+ "encoder.layers.2.layer_norm2.weight",
72
+ "encoder.layers.2.mlp.fc1.bias",
73
+ "encoder.layers.2.mlp.fc1.weight",
74
+ "encoder.layers.2.mlp.fc2.bias",
75
+ "encoder.layers.2.mlp.fc2.weight",
76
+ "encoder.layers.2.self_attn.k_proj.bias",
77
+ "encoder.layers.2.self_attn.k_proj.weight",
78
+ "encoder.layers.2.self_attn.out_proj.bias",
79
+ "encoder.layers.2.self_attn.out_proj.weight",
80
+ "encoder.layers.2.self_attn.q_proj.bias",
81
+ "encoder.layers.2.self_attn.q_proj.weight",
82
+ "encoder.layers.2.self_attn.v_proj.bias",
83
+ "encoder.layers.2.self_attn.v_proj.weight",
84
+ "encoder.layers.3.layer_norm1.bias",
85
+ "encoder.layers.3.layer_norm1.weight",
86
+ "encoder.layers.3.layer_norm2.bias",
87
+ "encoder.layers.3.layer_norm2.weight",
88
+ "encoder.layers.3.mlp.fc1.bias",
89
+ "encoder.layers.3.mlp.fc1.weight",
90
+ "encoder.layers.3.mlp.fc2.bias",
91
+ "encoder.layers.3.mlp.fc2.weight",
92
+ "encoder.layers.3.self_attn.k_proj.bias",
93
+ "encoder.layers.3.self_attn.k_proj.weight",
94
+ "encoder.layers.3.self_attn.out_proj.bias",
95
+ "encoder.layers.3.self_attn.out_proj.weight",
96
+ "encoder.layers.3.self_attn.q_proj.bias",
97
+ "encoder.layers.3.self_attn.q_proj.weight",
98
+ "encoder.layers.3.self_attn.v_proj.bias",
99
+ "encoder.layers.3.self_attn.v_proj.weight",
100
+ "encoder.layers.4.layer_norm1.bias",
101
+ "encoder.layers.4.layer_norm1.weight",
102
+ "encoder.layers.4.layer_norm2.bias",
103
+ "encoder.layers.4.layer_norm2.weight",
104
+ "encoder.layers.4.mlp.fc1.bias",
105
+ "encoder.layers.4.mlp.fc1.weight",
106
+ "encoder.layers.4.mlp.fc2.bias",
107
+ "encoder.layers.4.mlp.fc2.weight",
108
+ "encoder.layers.4.self_attn.k_proj.bias",
109
+ "encoder.layers.4.self_attn.k_proj.weight",
110
+ "encoder.layers.4.self_attn.out_proj.bias",
111
+ "encoder.layers.4.self_attn.out_proj.weight",
112
+ "encoder.layers.4.self_attn.q_proj.bias",
113
+ "encoder.layers.4.self_attn.q_proj.weight",
114
+ "encoder.layers.4.self_attn.v_proj.bias",
115
+ "encoder.layers.4.self_attn.v_proj.weight",
116
+ "encoder.layers.5.layer_norm1.bias",
117
+ "encoder.layers.5.layer_norm1.weight",
118
+ "encoder.layers.5.layer_norm2.bias",
119
+ "encoder.layers.5.layer_norm2.weight",
120
+ "encoder.layers.5.mlp.fc1.bias",
121
+ "encoder.layers.5.mlp.fc1.weight",
122
+ "encoder.layers.5.mlp.fc2.bias",
123
+ "encoder.layers.5.mlp.fc2.weight",
124
+ "encoder.layers.5.self_attn.k_proj.bias",
125
+ "encoder.layers.5.self_attn.k_proj.weight",
126
+ "encoder.layers.5.self_attn.out_proj.bias",
127
+ "encoder.layers.5.self_attn.out_proj.weight",
128
+ "encoder.layers.5.self_attn.q_proj.bias",
129
+ "encoder.layers.5.self_attn.q_proj.weight",
130
+ "encoder.layers.5.self_attn.v_proj.bias",
131
+ "encoder.layers.5.self_attn.v_proj.weight",
132
+ "encoder.layers.6.layer_norm1.bias",
133
+ "encoder.layers.6.layer_norm1.weight",
134
+ "encoder.layers.6.layer_norm2.bias",
135
+ "encoder.layers.6.layer_norm2.weight",
136
+ "encoder.layers.6.mlp.fc1.bias",
137
+ "encoder.layers.6.mlp.fc1.weight",
138
+ "encoder.layers.6.mlp.fc2.bias",
139
+ "encoder.layers.6.mlp.fc2.weight",
140
+ "encoder.layers.6.self_attn.k_proj.bias",
141
+ "encoder.layers.6.self_attn.k_proj.weight",
142
+ "encoder.layers.6.self_attn.out_proj.bias",
143
+ "encoder.layers.6.self_attn.out_proj.weight",
144
+ "encoder.layers.6.self_attn.q_proj.bias",
145
+ "encoder.layers.6.self_attn.q_proj.weight",
146
+ "encoder.layers.6.self_attn.v_proj.bias",
147
+ "encoder.layers.6.self_attn.v_proj.weight",
148
+ "encoder.layers.7.layer_norm1.bias",
149
+ "encoder.layers.7.layer_norm1.weight",
150
+ "encoder.layers.7.layer_norm2.bias",
151
+ "encoder.layers.7.layer_norm2.weight",
152
+ "encoder.layers.7.mlp.fc1.bias",
153
+ "encoder.layers.7.mlp.fc1.weight",
154
+ "encoder.layers.7.mlp.fc2.bias",
155
+ "encoder.layers.7.mlp.fc2.weight",
156
+ "encoder.layers.7.self_attn.k_proj.bias",
157
+ "encoder.layers.7.self_attn.k_proj.weight",
158
+ "encoder.layers.7.self_attn.out_proj.bias",
159
+ "encoder.layers.7.self_attn.out_proj.weight",
160
+ "encoder.layers.7.self_attn.q_proj.bias",
161
+ "encoder.layers.7.self_attn.q_proj.weight",
162
+ "encoder.layers.7.self_attn.v_proj.bias",
163
+ "encoder.layers.7.self_attn.v_proj.weight",
164
+ "encoder.layers.8.layer_norm1.bias",
165
+ "encoder.layers.8.layer_norm1.weight",
166
+ "encoder.layers.8.layer_norm2.bias",
167
+ "encoder.layers.8.layer_norm2.weight",
168
+ "encoder.layers.8.mlp.fc1.bias",
169
+ "encoder.layers.8.mlp.fc1.weight",
170
+ "encoder.layers.8.mlp.fc2.bias",
171
+ "encoder.layers.8.mlp.fc2.weight",
172
+ "encoder.layers.8.self_attn.k_proj.bias",
173
+ "encoder.layers.8.self_attn.k_proj.weight",
174
+ "encoder.layers.8.self_attn.out_proj.bias",
175
+ "encoder.layers.8.self_attn.out_proj.weight",
176
+ "encoder.layers.8.self_attn.q_proj.bias",
177
+ "encoder.layers.8.self_attn.q_proj.weight",
178
+ "encoder.layers.8.self_attn.v_proj.bias",
179
+ "encoder.layers.8.self_attn.v_proj.weight",
180
+ "encoder.layers.9.layer_norm1.bias",
181
+ "encoder.layers.9.layer_norm1.weight",
182
+ "encoder.layers.9.layer_norm2.bias",
183
+ "encoder.layers.9.layer_norm2.weight",
184
+ "encoder.layers.9.mlp.fc1.bias",
185
+ "encoder.layers.9.mlp.fc1.weight",
186
+ "encoder.layers.9.mlp.fc2.bias",
187
+ "encoder.layers.9.mlp.fc2.weight",
188
+ "encoder.layers.9.self_attn.k_proj.bias",
189
+ "encoder.layers.9.self_attn.k_proj.weight",
190
+ "encoder.layers.9.self_attn.out_proj.bias",
191
+ "encoder.layers.9.self_attn.out_proj.weight",
192
+ "encoder.layers.9.self_attn.q_proj.bias",
193
+ "encoder.layers.9.self_attn.q_proj.weight",
194
+ "encoder.layers.9.self_attn.v_proj.bias",
195
+ "encoder.layers.9.self_attn.v_proj.weight",
196
+ "final_layer_norm.bias",
197
+ "final_layer_norm.weight",
198
+ "text_encoders.clip_l.transformer.text_projection.weight",
199
+ "text_encoders.clip_l.transformer.text_model.embeddings.position_embedding.weight",
200
+ "text_encoders.clip_l.transformer.text_model.embeddings.token_embedding.weight",
201
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.0.layer_norm1.bias",
202
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.0.layer_norm1.weight",
203
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.0.layer_norm2.bias",
204
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.0.layer_norm2.weight",
205
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.0.mlp.fc1.bias",
206
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.0.mlp.fc1.weight",
207
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.0.mlp.fc2.bias",
208
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.0.mlp.fc2.weight",
209
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.0.self_attn.k_proj.bias",
210
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.0.self_attn.k_proj.weight",
211
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.0.self_attn.out_proj.bias",
212
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.0.self_attn.out_proj.weight",
213
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.0.self_attn.q_proj.bias",
214
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.0.self_attn.q_proj.weight",
215
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.0.self_attn.v_proj.bias",
216
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.0.self_attn.v_proj.weight",
217
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.1.layer_norm1.bias",
218
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.1.layer_norm1.weight",
219
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.1.layer_norm2.bias",
220
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.1.layer_norm2.weight",
221
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.1.mlp.fc1.bias",
222
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.1.mlp.fc1.weight",
223
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.1.mlp.fc2.bias",
224
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.1.mlp.fc2.weight",
225
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.1.self_attn.k_proj.bias",
226
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.1.self_attn.k_proj.weight",
227
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.1.self_attn.out_proj.bias",
228
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.1.self_attn.out_proj.weight",
229
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.1.self_attn.q_proj.bias",
230
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.1.self_attn.q_proj.weight",
231
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.1.self_attn.v_proj.bias",
232
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.1.self_attn.v_proj.weight",
233
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.10.layer_norm1.bias",
234
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.10.layer_norm1.weight",
235
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.10.layer_norm2.bias",
236
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.10.layer_norm2.weight",
237
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.10.mlp.fc1.bias",
238
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.10.mlp.fc1.weight",
239
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.10.mlp.fc2.bias",
240
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.10.mlp.fc2.weight",
241
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.10.self_attn.k_proj.bias",
242
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.10.self_attn.k_proj.weight",
243
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.10.self_attn.out_proj.bias",
244
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.10.self_attn.out_proj.weight",
245
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.10.self_attn.q_proj.bias",
246
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.10.self_attn.q_proj.weight",
247
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.10.self_attn.v_proj.bias",
248
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.10.self_attn.v_proj.weight",
249
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.11.layer_norm1.bias",
250
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.11.layer_norm1.weight",
251
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.11.layer_norm2.bias",
252
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.11.layer_norm2.weight",
253
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.11.mlp.fc1.bias",
254
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.11.mlp.fc1.weight",
255
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.11.mlp.fc2.bias",
256
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.11.mlp.fc2.weight",
257
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.11.self_attn.k_proj.bias",
258
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.11.self_attn.k_proj.weight",
259
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.11.self_attn.out_proj.bias",
260
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.11.self_attn.out_proj.weight",
261
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.11.self_attn.q_proj.bias",
262
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.11.self_attn.q_proj.weight",
263
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.11.self_attn.v_proj.bias",
264
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.11.self_attn.v_proj.weight",
265
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.2.layer_norm1.bias",
266
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.2.layer_norm1.weight",
267
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.2.layer_norm2.bias",
268
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.2.layer_norm2.weight",
269
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.2.mlp.fc1.bias",
270
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.2.mlp.fc1.weight",
271
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.2.mlp.fc2.bias",
272
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.2.mlp.fc2.weight",
273
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.2.self_attn.k_proj.bias",
274
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.2.self_attn.k_proj.weight",
275
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.2.self_attn.out_proj.bias",
276
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.2.self_attn.out_proj.weight",
277
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.2.self_attn.q_proj.bias",
278
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.2.self_attn.q_proj.weight",
279
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.2.self_attn.v_proj.bias",
280
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.2.self_attn.v_proj.weight",
281
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.3.layer_norm1.bias",
282
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.3.layer_norm1.weight",
283
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.3.layer_norm2.bias",
284
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.3.layer_norm2.weight",
285
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.3.mlp.fc1.bias",
286
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.3.mlp.fc1.weight",
287
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.3.mlp.fc2.bias",
288
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.3.mlp.fc2.weight",
289
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.3.self_attn.k_proj.bias",
290
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.3.self_attn.k_proj.weight",
291
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.3.self_attn.out_proj.bias",
292
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.3.self_attn.out_proj.weight",
293
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.3.self_attn.q_proj.bias",
294
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.3.self_attn.q_proj.weight",
295
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.3.self_attn.v_proj.bias",
296
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.3.self_attn.v_proj.weight",
297
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.4.layer_norm1.bias",
298
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.4.layer_norm1.weight",
299
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.4.layer_norm2.bias",
300
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.4.layer_norm2.weight",
301
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.4.mlp.fc1.bias",
302
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.4.mlp.fc1.weight",
303
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.4.mlp.fc2.bias",
304
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.4.mlp.fc2.weight",
305
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.4.self_attn.k_proj.bias",
306
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.4.self_attn.k_proj.weight",
307
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.4.self_attn.out_proj.bias",
308
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.4.self_attn.out_proj.weight",
309
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.4.self_attn.q_proj.bias",
310
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.4.self_attn.q_proj.weight",
311
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.4.self_attn.v_proj.bias",
312
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.4.self_attn.v_proj.weight",
313
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.5.layer_norm1.bias",
314
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.5.layer_norm1.weight",
315
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.5.layer_norm2.bias",
316
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.5.layer_norm2.weight",
317
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.5.mlp.fc1.bias",
318
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.5.mlp.fc1.weight",
319
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.5.mlp.fc2.bias",
320
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.5.mlp.fc2.weight",
321
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.5.self_attn.k_proj.bias",
322
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.5.self_attn.k_proj.weight",
323
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.5.self_attn.out_proj.bias",
324
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.5.self_attn.out_proj.weight",
325
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.5.self_attn.q_proj.bias",
326
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.5.self_attn.q_proj.weight",
327
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.5.self_attn.v_proj.bias",
328
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.5.self_attn.v_proj.weight",
329
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.6.layer_norm1.bias",
330
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.6.layer_norm1.weight",
331
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.6.layer_norm2.bias",
332
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.6.layer_norm2.weight",
333
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.6.mlp.fc1.bias",
334
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.6.mlp.fc1.weight",
335
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.6.mlp.fc2.bias",
336
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.6.mlp.fc2.weight",
337
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.6.self_attn.k_proj.bias",
338
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.6.self_attn.k_proj.weight",
339
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.6.self_attn.out_proj.bias",
340
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.6.self_attn.out_proj.weight",
341
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.6.self_attn.q_proj.bias",
342
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.6.self_attn.q_proj.weight",
343
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.6.self_attn.v_proj.bias",
344
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.6.self_attn.v_proj.weight",
345
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.7.layer_norm1.bias",
346
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.7.layer_norm1.weight",
347
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.7.layer_norm2.bias",
348
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.7.layer_norm2.weight",
349
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.7.mlp.fc1.bias",
350
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.7.mlp.fc1.weight",
351
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.7.mlp.fc2.bias",
352
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.7.mlp.fc2.weight",
353
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.7.self_attn.k_proj.bias",
354
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.7.self_attn.k_proj.weight",
355
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.7.self_attn.out_proj.bias",
356
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.7.self_attn.out_proj.weight",
357
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.7.self_attn.q_proj.bias",
358
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.7.self_attn.q_proj.weight",
359
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.7.self_attn.v_proj.bias",
360
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.7.self_attn.v_proj.weight",
361
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.8.layer_norm1.bias",
362
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.8.layer_norm1.weight",
363
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.8.layer_norm2.bias",
364
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.8.layer_norm2.weight",
365
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.8.mlp.fc1.bias",
366
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.8.mlp.fc1.weight",
367
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.8.mlp.fc2.bias",
368
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.8.mlp.fc2.weight",
369
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.8.self_attn.k_proj.bias",
370
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.8.self_attn.k_proj.weight",
371
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.8.self_attn.out_proj.bias",
372
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.8.self_attn.out_proj.weight",
373
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.8.self_attn.q_proj.bias",
374
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.8.self_attn.q_proj.weight",
375
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.8.self_attn.v_proj.bias",
376
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.8.self_attn.v_proj.weight",
377
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.9.layer_norm1.bias",
378
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.9.layer_norm1.weight",
379
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.9.layer_norm2.bias",
380
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.9.layer_norm2.weight",
381
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.9.mlp.fc1.bias",
382
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.9.mlp.fc1.weight",
383
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.9.mlp.fc2.bias",
384
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.9.mlp.fc2.weight",
385
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.9.self_attn.k_proj.bias",
386
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.9.self_attn.k_proj.weight",
387
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.9.self_attn.out_proj.bias",
388
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.9.self_attn.out_proj.weight",
389
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.9.self_attn.q_proj.bias",
390
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.9.self_attn.q_proj.weight",
391
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.9.self_attn.v_proj.bias",
392
+ "text_encoders.clip_l.transformer.text_model.encoder.layers.9.self_attn.v_proj.weight",
393
+ "text_encoders.clip_l.transformer.text_model.final_layer_norm.bias",
394
+ "text_encoders.clip_l.transformer.text_model.final_layer_norm.weight",
395
+ "text_encoders.clip_l.transformer.text_projection.weight"
396
+ ]
flux_t5xxl_keys.json ADDED
@@ -0,0 +1,440 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ "encoder.block.0.layer.0.SelfAttention.k.weight",
3
+ "encoder.block.0.layer.0.SelfAttention.o.weight",
4
+ "encoder.block.0.layer.0.SelfAttention.q.weight",
5
+ "encoder.block.0.layer.0.SelfAttention.relative_attention_bias.weight",
6
+ "encoder.block.0.layer.0.SelfAttention.v.weight",
7
+ "encoder.block.0.layer.0.layer_norm.weight",
8
+ "encoder.block.0.layer.1.DenseReluDense.wi_0.weight",
9
+ "encoder.block.0.layer.1.DenseReluDense.wi_1.weight",
10
+ "encoder.block.0.layer.1.DenseReluDense.wo.weight",
11
+ "encoder.block.0.layer.1.layer_norm.weight",
12
+ "encoder.block.1.layer.0.SelfAttention.k.weight",
13
+ "encoder.block.1.layer.0.SelfAttention.o.weight",
14
+ "encoder.block.1.layer.0.SelfAttention.q.weight",
15
+ "encoder.block.1.layer.0.SelfAttention.v.weight",
16
+ "encoder.block.1.layer.0.layer_norm.weight",
17
+ "encoder.block.1.layer.1.DenseReluDense.wi_0.weight",
18
+ "encoder.block.1.layer.1.DenseReluDense.wi_1.weight",
19
+ "encoder.block.1.layer.1.DenseReluDense.wo.weight",
20
+ "encoder.block.1.layer.1.layer_norm.weight",
21
+ "encoder.block.10.layer.0.SelfAttention.k.weight",
22
+ "encoder.block.10.layer.0.SelfAttention.o.weight",
23
+ "encoder.block.10.layer.0.SelfAttention.q.weight",
24
+ "encoder.block.10.layer.0.SelfAttention.v.weight",
25
+ "encoder.block.10.layer.0.layer_norm.weight",
26
+ "encoder.block.10.layer.1.DenseReluDense.wi_0.weight",
27
+ "encoder.block.10.layer.1.DenseReluDense.wi_1.weight",
28
+ "encoder.block.10.layer.1.DenseReluDense.wo.weight",
29
+ "encoder.block.10.layer.1.layer_norm.weight",
30
+ "encoder.block.11.layer.0.SelfAttention.k.weight",
31
+ "encoder.block.11.layer.0.SelfAttention.o.weight",
32
+ "encoder.block.11.layer.0.SelfAttention.q.weight",
33
+ "encoder.block.11.layer.0.SelfAttention.v.weight",
34
+ "encoder.block.11.layer.0.layer_norm.weight",
35
+ "encoder.block.11.layer.1.DenseReluDense.wi_0.weight",
36
+ "encoder.block.11.layer.1.DenseReluDense.wi_1.weight",
37
+ "encoder.block.11.layer.1.DenseReluDense.wo.weight",
38
+ "encoder.block.11.layer.1.layer_norm.weight",
39
+ "encoder.block.12.layer.0.SelfAttention.k.weight",
40
+ "encoder.block.12.layer.0.SelfAttention.o.weight",
41
+ "encoder.block.12.layer.0.SelfAttention.q.weight",
42
+ "encoder.block.12.layer.0.SelfAttention.v.weight",
43
+ "encoder.block.12.layer.0.layer_norm.weight",
44
+ "encoder.block.12.layer.1.DenseReluDense.wi_0.weight",
45
+ "encoder.block.12.layer.1.DenseReluDense.wi_1.weight",
46
+ "encoder.block.12.layer.1.DenseReluDense.wo.weight",
47
+ "encoder.block.12.layer.1.layer_norm.weight",
48
+ "encoder.block.13.layer.0.SelfAttention.k.weight",
49
+ "encoder.block.13.layer.0.SelfAttention.o.weight",
50
+ "encoder.block.13.layer.0.SelfAttention.q.weight",
51
+ "encoder.block.13.layer.0.SelfAttention.v.weight",
52
+ "encoder.block.13.layer.0.layer_norm.weight",
53
+ "encoder.block.13.layer.1.DenseReluDense.wi_0.weight",
54
+ "encoder.block.13.layer.1.DenseReluDense.wi_1.weight",
55
+ "encoder.block.13.layer.1.DenseReluDense.wo.weight",
56
+ "encoder.block.13.layer.1.layer_norm.weight",
57
+ "encoder.block.14.layer.0.SelfAttention.k.weight",
58
+ "encoder.block.14.layer.0.SelfAttention.o.weight",
59
+ "encoder.block.14.layer.0.SelfAttention.q.weight",
60
+ "encoder.block.14.layer.0.SelfAttention.v.weight",
61
+ "encoder.block.14.layer.0.layer_norm.weight",
62
+ "encoder.block.14.layer.1.DenseReluDense.wi_0.weight",
63
+ "encoder.block.14.layer.1.DenseReluDense.wi_1.weight",
64
+ "encoder.block.14.layer.1.DenseReluDense.wo.weight",
65
+ "encoder.block.14.layer.1.layer_norm.weight",
66
+ "encoder.block.15.layer.0.SelfAttention.k.weight",
67
+ "encoder.block.15.layer.0.SelfAttention.o.weight",
68
+ "encoder.block.15.layer.0.SelfAttention.q.weight",
69
+ "encoder.block.15.layer.0.SelfAttention.v.weight",
70
+ "encoder.block.15.layer.0.layer_norm.weight",
71
+ "encoder.block.15.layer.1.DenseReluDense.wi_0.weight",
72
+ "encoder.block.15.layer.1.DenseReluDense.wi_1.weight",
73
+ "encoder.block.15.layer.1.DenseReluDense.wo.weight",
74
+ "encoder.block.15.layer.1.layer_norm.weight",
75
+ "encoder.block.16.layer.0.SelfAttention.k.weight",
76
+ "encoder.block.16.layer.0.SelfAttention.o.weight",
77
+ "encoder.block.16.layer.0.SelfAttention.q.weight",
78
+ "encoder.block.16.layer.0.SelfAttention.v.weight",
79
+ "encoder.block.16.layer.0.layer_norm.weight",
80
+ "encoder.block.16.layer.1.DenseReluDense.wi_0.weight",
81
+ "encoder.block.16.layer.1.DenseReluDense.wi_1.weight",
82
+ "encoder.block.16.layer.1.DenseReluDense.wo.weight",
83
+ "encoder.block.16.layer.1.layer_norm.weight",
84
+ "encoder.block.17.layer.0.SelfAttention.k.weight",
85
+ "encoder.block.17.layer.0.SelfAttention.o.weight",
86
+ "encoder.block.17.layer.0.SelfAttention.q.weight",
87
+ "encoder.block.17.layer.0.SelfAttention.v.weight",
88
+ "encoder.block.17.layer.0.layer_norm.weight",
89
+ "encoder.block.17.layer.1.DenseReluDense.wi_0.weight",
90
+ "encoder.block.17.layer.1.DenseReluDense.wi_1.weight",
91
+ "encoder.block.17.layer.1.DenseReluDense.wo.weight",
92
+ "encoder.block.17.layer.1.layer_norm.weight",
93
+ "encoder.block.18.layer.0.SelfAttention.k.weight",
94
+ "encoder.block.18.layer.0.SelfAttention.o.weight",
95
+ "encoder.block.18.layer.0.SelfAttention.q.weight",
96
+ "encoder.block.18.layer.0.SelfAttention.v.weight",
97
+ "encoder.block.18.layer.0.layer_norm.weight",
98
+ "encoder.block.18.layer.1.DenseReluDense.wi_0.weight",
99
+ "encoder.block.18.layer.1.DenseReluDense.wi_1.weight",
100
+ "encoder.block.18.layer.1.DenseReluDense.wo.weight",
101
+ "encoder.block.18.layer.1.layer_norm.weight",
102
+ "encoder.block.19.layer.0.SelfAttention.k.weight",
103
+ "encoder.block.19.layer.0.SelfAttention.o.weight",
104
+ "encoder.block.19.layer.0.SelfAttention.q.weight",
105
+ "encoder.block.19.layer.0.SelfAttention.v.weight",
106
+ "encoder.block.19.layer.0.layer_norm.weight",
107
+ "encoder.block.19.layer.1.DenseReluDense.wi_0.weight",
108
+ "encoder.block.19.layer.1.DenseReluDense.wi_1.weight",
109
+ "encoder.block.19.layer.1.DenseReluDense.wo.weight",
110
+ "encoder.block.19.layer.1.layer_norm.weight",
111
+ "encoder.block.2.layer.0.SelfAttention.k.weight",
112
+ "encoder.block.2.layer.0.SelfAttention.o.weight",
113
+ "encoder.block.2.layer.0.SelfAttention.q.weight",
114
+ "encoder.block.2.layer.0.SelfAttention.v.weight",
115
+ "encoder.block.2.layer.0.layer_norm.weight",
116
+ "encoder.block.2.layer.1.DenseReluDense.wi_0.weight",
117
+ "encoder.block.2.layer.1.DenseReluDense.wi_1.weight",
118
+ "encoder.block.2.layer.1.DenseReluDense.wo.weight",
119
+ "encoder.block.2.layer.1.layer_norm.weight",
120
+ "encoder.block.20.layer.0.SelfAttention.k.weight",
121
+ "encoder.block.20.layer.0.SelfAttention.o.weight",
122
+ "encoder.block.20.layer.0.SelfAttention.q.weight",
123
+ "encoder.block.20.layer.0.SelfAttention.v.weight",
124
+ "encoder.block.20.layer.0.layer_norm.weight",
125
+ "encoder.block.20.layer.1.DenseReluDense.wi_0.weight",
126
+ "encoder.block.20.layer.1.DenseReluDense.wi_1.weight",
127
+ "encoder.block.20.layer.1.DenseReluDense.wo.weight",
128
+ "encoder.block.20.layer.1.layer_norm.weight",
129
+ "encoder.block.21.layer.0.SelfAttention.k.weight",
130
+ "encoder.block.21.layer.0.SelfAttention.o.weight",
131
+ "encoder.block.21.layer.0.SelfAttention.q.weight",
132
+ "encoder.block.21.layer.0.SelfAttention.v.weight",
133
+ "encoder.block.21.layer.0.layer_norm.weight",
134
+ "encoder.block.21.layer.1.DenseReluDense.wi_0.weight",
135
+ "encoder.block.21.layer.1.DenseReluDense.wi_1.weight",
136
+ "encoder.block.21.layer.1.DenseReluDense.wo.weight",
137
+ "encoder.block.21.layer.1.layer_norm.weight",
138
+ "encoder.block.22.layer.0.SelfAttention.k.weight",
139
+ "encoder.block.22.layer.0.SelfAttention.o.weight",
140
+ "encoder.block.22.layer.0.SelfAttention.q.weight",
141
+ "encoder.block.22.layer.0.SelfAttention.v.weight",
142
+ "encoder.block.22.layer.0.layer_norm.weight",
143
+ "encoder.block.22.layer.1.DenseReluDense.wi_0.weight",
144
+ "encoder.block.22.layer.1.DenseReluDense.wi_1.weight",
145
+ "encoder.block.22.layer.1.DenseReluDense.wo.weight",
146
+ "encoder.block.22.layer.1.layer_norm.weight",
147
+ "encoder.block.23.layer.0.SelfAttention.k.weight",
148
+ "encoder.block.23.layer.0.SelfAttention.o.weight",
149
+ "encoder.block.23.layer.0.SelfAttention.q.weight",
150
+ "encoder.block.23.layer.0.SelfAttention.v.weight",
151
+ "encoder.block.23.layer.0.layer_norm.weight",
152
+ "encoder.block.23.layer.1.DenseReluDense.wi_0.weight",
153
+ "encoder.block.23.layer.1.DenseReluDense.wi_1.weight",
154
+ "encoder.block.23.layer.1.DenseReluDense.wo.weight",
155
+ "encoder.block.23.layer.1.layer_norm.weight",
156
+ "encoder.block.3.layer.0.SelfAttention.k.weight",
157
+ "encoder.block.3.layer.0.SelfAttention.o.weight",
158
+ "encoder.block.3.layer.0.SelfAttention.q.weight",
159
+ "encoder.block.3.layer.0.SelfAttention.v.weight",
160
+ "encoder.block.3.layer.0.layer_norm.weight",
161
+ "encoder.block.3.layer.1.DenseReluDense.wi_0.weight",
162
+ "encoder.block.3.layer.1.DenseReluDense.wi_1.weight",
163
+ "encoder.block.3.layer.1.DenseReluDense.wo.weight",
164
+ "encoder.block.3.layer.1.layer_norm.weight",
165
+ "encoder.block.4.layer.0.SelfAttention.k.weight",
166
+ "encoder.block.4.layer.0.SelfAttention.o.weight",
167
+ "encoder.block.4.layer.0.SelfAttention.q.weight",
168
+ "encoder.block.4.layer.0.SelfAttention.v.weight",
169
+ "encoder.block.4.layer.0.layer_norm.weight",
170
+ "encoder.block.4.layer.1.DenseReluDense.wi_0.weight",
171
+ "encoder.block.4.layer.1.DenseReluDense.wi_1.weight",
172
+ "encoder.block.4.layer.1.DenseReluDense.wo.weight",
173
+ "encoder.block.4.layer.1.layer_norm.weight",
174
+ "encoder.block.5.layer.0.SelfAttention.k.weight",
175
+ "encoder.block.5.layer.0.SelfAttention.o.weight",
176
+ "encoder.block.5.layer.0.SelfAttention.q.weight",
177
+ "encoder.block.5.layer.0.SelfAttention.v.weight",
178
+ "encoder.block.5.layer.0.layer_norm.weight",
179
+ "encoder.block.5.layer.1.DenseReluDense.wi_0.weight",
180
+ "encoder.block.5.layer.1.DenseReluDense.wi_1.weight",
181
+ "encoder.block.5.layer.1.DenseReluDense.wo.weight",
182
+ "encoder.block.5.layer.1.layer_norm.weight",
183
+ "encoder.block.6.layer.0.SelfAttention.k.weight",
184
+ "encoder.block.6.layer.0.SelfAttention.o.weight",
185
+ "encoder.block.6.layer.0.SelfAttention.q.weight",
186
+ "encoder.block.6.layer.0.SelfAttention.v.weight",
187
+ "encoder.block.6.layer.0.layer_norm.weight",
188
+ "encoder.block.6.layer.1.DenseReluDense.wi_0.weight",
189
+ "encoder.block.6.layer.1.DenseReluDense.wi_1.weight",
190
+ "encoder.block.6.layer.1.DenseReluDense.wo.weight",
191
+ "encoder.block.6.layer.1.layer_norm.weight",
192
+ "encoder.block.7.layer.0.SelfAttention.k.weight",
193
+ "encoder.block.7.layer.0.SelfAttention.o.weight",
194
+ "encoder.block.7.layer.0.SelfAttention.q.weight",
195
+ "encoder.block.7.layer.0.SelfAttention.v.weight",
196
+ "encoder.block.7.layer.0.layer_norm.weight",
197
+ "encoder.block.7.layer.1.DenseReluDense.wi_0.weight",
198
+ "encoder.block.7.layer.1.DenseReluDense.wi_1.weight",
199
+ "encoder.block.7.layer.1.DenseReluDense.wo.weight",
200
+ "encoder.block.7.layer.1.layer_norm.weight",
201
+ "encoder.block.8.layer.0.SelfAttention.k.weight",
202
+ "encoder.block.8.layer.0.SelfAttention.o.weight",
203
+ "encoder.block.8.layer.0.SelfAttention.q.weight",
204
+ "encoder.block.8.layer.0.SelfAttention.v.weight",
205
+ "encoder.block.8.layer.0.layer_norm.weight",
206
+ "encoder.block.8.layer.1.DenseReluDense.wi_0.weight",
207
+ "encoder.block.8.layer.1.DenseReluDense.wi_1.weight",
208
+ "encoder.block.8.layer.1.DenseReluDense.wo.weight",
209
+ "encoder.block.8.layer.1.layer_norm.weight",
210
+ "encoder.block.9.layer.0.SelfAttention.k.weight",
211
+ "encoder.block.9.layer.0.SelfAttention.o.weight",
212
+ "encoder.block.9.layer.0.SelfAttention.q.weight",
213
+ "encoder.block.9.layer.0.SelfAttention.v.weight",
214
+ "encoder.block.9.layer.0.layer_norm.weight",
215
+ "encoder.block.9.layer.1.DenseReluDense.wi_0.weight",
216
+ "encoder.block.9.layer.1.DenseReluDense.wi_1.weight",
217
+ "encoder.block.9.layer.1.DenseReluDense.wo.weight",
218
+ "encoder.block.9.layer.1.layer_norm.weight",
219
+ "encoder.final_layer_norm.weight",
220
+ "shared.weight",
221
+ "text_encoders.t5xxl.transformer.encoder.block.0.layer.0.SelfAttention.k.weight",
222
+ "text_encoders.t5xxl.transformer.encoder.block.0.layer.0.SelfAttention.o.weight",
223
+ "text_encoders.t5xxl.transformer.encoder.block.0.layer.0.SelfAttention.q.weight",
224
+ "text_encoders.t5xxl.transformer.encoder.block.0.layer.0.SelfAttention.relative_attention_bias.weight",
225
+ "text_encoders.t5xxl.transformer.encoder.block.0.layer.0.SelfAttention.v.weight",
226
+ "text_encoders.t5xxl.transformer.encoder.block.0.layer.0.layer_norm.weight",
227
+ "text_encoders.t5xxl.transformer.encoder.block.0.layer.1.DenseReluDense.wi_0.weight",
228
+ "text_encoders.t5xxl.transformer.encoder.block.0.layer.1.DenseReluDense.wi_1.weight",
229
+ "text_encoders.t5xxl.transformer.encoder.block.0.layer.1.DenseReluDense.wo.weight",
230
+ "text_encoders.t5xxl.transformer.encoder.block.0.layer.1.layer_norm.weight",
231
+ "text_encoders.t5xxl.transformer.encoder.block.1.layer.0.SelfAttention.k.weight",
232
+ "text_encoders.t5xxl.transformer.encoder.block.1.layer.0.SelfAttention.o.weight",
233
+ "text_encoders.t5xxl.transformer.encoder.block.1.layer.0.SelfAttention.q.weight",
234
+ "text_encoders.t5xxl.transformer.encoder.block.1.layer.0.SelfAttention.v.weight",
235
+ "text_encoders.t5xxl.transformer.encoder.block.1.layer.0.layer_norm.weight",
236
+ "text_encoders.t5xxl.transformer.encoder.block.1.layer.1.DenseReluDense.wi_0.weight",
237
+ "text_encoders.t5xxl.transformer.encoder.block.1.layer.1.DenseReluDense.wi_1.weight",
238
+ "text_encoders.t5xxl.transformer.encoder.block.1.layer.1.DenseReluDense.wo.weight",
239
+ "text_encoders.t5xxl.transformer.encoder.block.1.layer.1.layer_norm.weight",
240
+ "text_encoders.t5xxl.transformer.encoder.block.10.layer.0.SelfAttention.k.weight",
241
+ "text_encoders.t5xxl.transformer.encoder.block.10.layer.0.SelfAttention.o.weight",
242
+ "text_encoders.t5xxl.transformer.encoder.block.10.layer.0.SelfAttention.q.weight",
243
+ "text_encoders.t5xxl.transformer.encoder.block.10.layer.0.SelfAttention.v.weight",
244
+ "text_encoders.t5xxl.transformer.encoder.block.10.layer.0.layer_norm.weight",
245
+ "text_encoders.t5xxl.transformer.encoder.block.10.layer.1.DenseReluDense.wi_0.weight",
246
+ "text_encoders.t5xxl.transformer.encoder.block.10.layer.1.DenseReluDense.wi_1.weight",
247
+ "text_encoders.t5xxl.transformer.encoder.block.10.layer.1.DenseReluDense.wo.weight",
248
+ "text_encoders.t5xxl.transformer.encoder.block.10.layer.1.layer_norm.weight",
249
+ "text_encoders.t5xxl.transformer.encoder.block.11.layer.0.SelfAttention.k.weight",
250
+ "text_encoders.t5xxl.transformer.encoder.block.11.layer.0.SelfAttention.o.weight",
251
+ "text_encoders.t5xxl.transformer.encoder.block.11.layer.0.SelfAttention.q.weight",
252
+ "text_encoders.t5xxl.transformer.encoder.block.11.layer.0.SelfAttention.v.weight",
253
+ "text_encoders.t5xxl.transformer.encoder.block.11.layer.0.layer_norm.weight",
254
+ "text_encoders.t5xxl.transformer.encoder.block.11.layer.1.DenseReluDense.wi_0.weight",
255
+ "text_encoders.t5xxl.transformer.encoder.block.11.layer.1.DenseReluDense.wi_1.weight",
256
+ "text_encoders.t5xxl.transformer.encoder.block.11.layer.1.DenseReluDense.wo.weight",
257
+ "text_encoders.t5xxl.transformer.encoder.block.11.layer.1.layer_norm.weight",
258
+ "text_encoders.t5xxl.transformer.encoder.block.12.layer.0.SelfAttention.k.weight",
259
+ "text_encoders.t5xxl.transformer.encoder.block.12.layer.0.SelfAttention.o.weight",
260
+ "text_encoders.t5xxl.transformer.encoder.block.12.layer.0.SelfAttention.q.weight",
261
+ "text_encoders.t5xxl.transformer.encoder.block.12.layer.0.SelfAttention.v.weight",
262
+ "text_encoders.t5xxl.transformer.encoder.block.12.layer.0.layer_norm.weight",
263
+ "text_encoders.t5xxl.transformer.encoder.block.12.layer.1.DenseReluDense.wi_0.weight",
264
+ "text_encoders.t5xxl.transformer.encoder.block.12.layer.1.DenseReluDense.wi_1.weight",
265
+ "text_encoders.t5xxl.transformer.encoder.block.12.layer.1.DenseReluDense.wo.weight",
266
+ "text_encoders.t5xxl.transformer.encoder.block.12.layer.1.layer_norm.weight",
267
+ "text_encoders.t5xxl.transformer.encoder.block.13.layer.0.SelfAttention.k.weight",
268
+ "text_encoders.t5xxl.transformer.encoder.block.13.layer.0.SelfAttention.o.weight",
269
+ "text_encoders.t5xxl.transformer.encoder.block.13.layer.0.SelfAttention.q.weight",
270
+ "text_encoders.t5xxl.transformer.encoder.block.13.layer.0.SelfAttention.v.weight",
271
+ "text_encoders.t5xxl.transformer.encoder.block.13.layer.0.layer_norm.weight",
272
+ "text_encoders.t5xxl.transformer.encoder.block.13.layer.1.DenseReluDense.wi_0.weight",
273
+ "text_encoders.t5xxl.transformer.encoder.block.13.layer.1.DenseReluDense.wi_1.weight",
274
+ "text_encoders.t5xxl.transformer.encoder.block.13.layer.1.DenseReluDense.wo.weight",
275
+ "text_encoders.t5xxl.transformer.encoder.block.13.layer.1.layer_norm.weight",
276
+ "text_encoders.t5xxl.transformer.encoder.block.14.layer.0.SelfAttention.k.weight",
277
+ "text_encoders.t5xxl.transformer.encoder.block.14.layer.0.SelfAttention.o.weight",
278
+ "text_encoders.t5xxl.transformer.encoder.block.14.layer.0.SelfAttention.q.weight",
279
+ "text_encoders.t5xxl.transformer.encoder.block.14.layer.0.SelfAttention.v.weight",
280
+ "text_encoders.t5xxl.transformer.encoder.block.14.layer.0.layer_norm.weight",
281
+ "text_encoders.t5xxl.transformer.encoder.block.14.layer.1.DenseReluDense.wi_0.weight",
282
+ "text_encoders.t5xxl.transformer.encoder.block.14.layer.1.DenseReluDense.wi_1.weight",
283
+ "text_encoders.t5xxl.transformer.encoder.block.14.layer.1.DenseReluDense.wo.weight",
284
+ "text_encoders.t5xxl.transformer.encoder.block.14.layer.1.layer_norm.weight",
285
+ "text_encoders.t5xxl.transformer.encoder.block.15.layer.0.SelfAttention.k.weight",
286
+ "text_encoders.t5xxl.transformer.encoder.block.15.layer.0.SelfAttention.o.weight",
287
+ "text_encoders.t5xxl.transformer.encoder.block.15.layer.0.SelfAttention.q.weight",
288
+ "text_encoders.t5xxl.transformer.encoder.block.15.layer.0.SelfAttention.v.weight",
289
+ "text_encoders.t5xxl.transformer.encoder.block.15.layer.0.layer_norm.weight",
290
+ "text_encoders.t5xxl.transformer.encoder.block.15.layer.1.DenseReluDense.wi_0.weight",
291
+ "text_encoders.t5xxl.transformer.encoder.block.15.layer.1.DenseReluDense.wi_1.weight",
292
+ "text_encoders.t5xxl.transformer.encoder.block.15.layer.1.DenseReluDense.wo.weight",
293
+ "text_encoders.t5xxl.transformer.encoder.block.15.layer.1.layer_norm.weight",
294
+ "text_encoders.t5xxl.transformer.encoder.block.16.layer.0.SelfAttention.k.weight",
295
+ "text_encoders.t5xxl.transformer.encoder.block.16.layer.0.SelfAttention.o.weight",
296
+ "text_encoders.t5xxl.transformer.encoder.block.16.layer.0.SelfAttention.q.weight",
297
+ "text_encoders.t5xxl.transformer.encoder.block.16.layer.0.SelfAttention.v.weight",
298
+ "text_encoders.t5xxl.transformer.encoder.block.16.layer.0.layer_norm.weight",
299
+ "text_encoders.t5xxl.transformer.encoder.block.16.layer.1.DenseReluDense.wi_0.weight",
300
+ "text_encoders.t5xxl.transformer.encoder.block.16.layer.1.DenseReluDense.wi_1.weight",
301
+ "text_encoders.t5xxl.transformer.encoder.block.16.layer.1.DenseReluDense.wo.weight",
302
+ "text_encoders.t5xxl.transformer.encoder.block.16.layer.1.layer_norm.weight",
303
+ "text_encoders.t5xxl.transformer.encoder.block.17.layer.0.SelfAttention.k.weight",
304
+ "text_encoders.t5xxl.transformer.encoder.block.17.layer.0.SelfAttention.o.weight",
305
+ "text_encoders.t5xxl.transformer.encoder.block.17.layer.0.SelfAttention.q.weight",
306
+ "text_encoders.t5xxl.transformer.encoder.block.17.layer.0.SelfAttention.v.weight",
307
+ "text_encoders.t5xxl.transformer.encoder.block.17.layer.0.layer_norm.weight",
308
+ "text_encoders.t5xxl.transformer.encoder.block.17.layer.1.DenseReluDense.wi_0.weight",
309
+ "text_encoders.t5xxl.transformer.encoder.block.17.layer.1.DenseReluDense.wi_1.weight",
310
+ "text_encoders.t5xxl.transformer.encoder.block.17.layer.1.DenseReluDense.wo.weight",
311
+ "text_encoders.t5xxl.transformer.encoder.block.17.layer.1.layer_norm.weight",
312
+ "text_encoders.t5xxl.transformer.encoder.block.18.layer.0.SelfAttention.k.weight",
313
+ "text_encoders.t5xxl.transformer.encoder.block.18.layer.0.SelfAttention.o.weight",
314
+ "text_encoders.t5xxl.transformer.encoder.block.18.layer.0.SelfAttention.q.weight",
315
+ "text_encoders.t5xxl.transformer.encoder.block.18.layer.0.SelfAttention.v.weight",
316
+ "text_encoders.t5xxl.transformer.encoder.block.18.layer.0.layer_norm.weight",
317
+ "text_encoders.t5xxl.transformer.encoder.block.18.layer.1.DenseReluDense.wi_0.weight",
318
+ "text_encoders.t5xxl.transformer.encoder.block.18.layer.1.DenseReluDense.wi_1.weight",
319
+ "text_encoders.t5xxl.transformer.encoder.block.18.layer.1.DenseReluDense.wo.weight",
320
+ "text_encoders.t5xxl.transformer.encoder.block.18.layer.1.layer_norm.weight",
321
+ "text_encoders.t5xxl.transformer.encoder.block.19.layer.0.SelfAttention.k.weight",
322
+ "text_encoders.t5xxl.transformer.encoder.block.19.layer.0.SelfAttention.o.weight",
323
+ "text_encoders.t5xxl.transformer.encoder.block.19.layer.0.SelfAttention.q.weight",
324
+ "text_encoders.t5xxl.transformer.encoder.block.19.layer.0.SelfAttention.v.weight",
325
+ "text_encoders.t5xxl.transformer.encoder.block.19.layer.0.layer_norm.weight",
326
+ "text_encoders.t5xxl.transformer.encoder.block.19.layer.1.DenseReluDense.wi_0.weight",
327
+ "text_encoders.t5xxl.transformer.encoder.block.19.layer.1.DenseReluDense.wi_1.weight",
328
+ "text_encoders.t5xxl.transformer.encoder.block.19.layer.1.DenseReluDense.wo.weight",
329
+ "text_encoders.t5xxl.transformer.encoder.block.19.layer.1.layer_norm.weight",
330
+ "text_encoders.t5xxl.transformer.encoder.block.2.layer.0.SelfAttention.k.weight",
331
+ "text_encoders.t5xxl.transformer.encoder.block.2.layer.0.SelfAttention.o.weight",
332
+ "text_encoders.t5xxl.transformer.encoder.block.2.layer.0.SelfAttention.q.weight",
333
+ "text_encoders.t5xxl.transformer.encoder.block.2.layer.0.SelfAttention.v.weight",
334
+ "text_encoders.t5xxl.transformer.encoder.block.2.layer.0.layer_norm.weight",
335
+ "text_encoders.t5xxl.transformer.encoder.block.2.layer.1.DenseReluDense.wi_0.weight",
336
+ "text_encoders.t5xxl.transformer.encoder.block.2.layer.1.DenseReluDense.wi_1.weight",
337
+ "text_encoders.t5xxl.transformer.encoder.block.2.layer.1.DenseReluDense.wo.weight",
338
+ "text_encoders.t5xxl.transformer.encoder.block.2.layer.1.layer_norm.weight",
339
+ "text_encoders.t5xxl.transformer.encoder.block.20.layer.0.SelfAttention.k.weight",
340
+ "text_encoders.t5xxl.transformer.encoder.block.20.layer.0.SelfAttention.o.weight",
341
+ "text_encoders.t5xxl.transformer.encoder.block.20.layer.0.SelfAttention.q.weight",
342
+ "text_encoders.t5xxl.transformer.encoder.block.20.layer.0.SelfAttention.v.weight",
343
+ "text_encoders.t5xxl.transformer.encoder.block.20.layer.0.layer_norm.weight",
344
+ "text_encoders.t5xxl.transformer.encoder.block.20.layer.1.DenseReluDense.wi_0.weight",
345
+ "text_encoders.t5xxl.transformer.encoder.block.20.layer.1.DenseReluDense.wi_1.weight",
346
+ "text_encoders.t5xxl.transformer.encoder.block.20.layer.1.DenseReluDense.wo.weight",
347
+ "text_encoders.t5xxl.transformer.encoder.block.20.layer.1.layer_norm.weight",
348
+ "text_encoders.t5xxl.transformer.encoder.block.21.layer.0.SelfAttention.k.weight",
349
+ "text_encoders.t5xxl.transformer.encoder.block.21.layer.0.SelfAttention.o.weight",
350
+ "text_encoders.t5xxl.transformer.encoder.block.21.layer.0.SelfAttention.q.weight",
351
+ "text_encoders.t5xxl.transformer.encoder.block.21.layer.0.SelfAttention.v.weight",
352
+ "text_encoders.t5xxl.transformer.encoder.block.21.layer.0.layer_norm.weight",
353
+ "text_encoders.t5xxl.transformer.encoder.block.21.layer.1.DenseReluDense.wi_0.weight",
354
+ "text_encoders.t5xxl.transformer.encoder.block.21.layer.1.DenseReluDense.wi_1.weight",
355
+ "text_encoders.t5xxl.transformer.encoder.block.21.layer.1.DenseReluDense.wo.weight",
356
+ "text_encoders.t5xxl.transformer.encoder.block.21.layer.1.layer_norm.weight",
357
+ "text_encoders.t5xxl.transformer.encoder.block.22.layer.0.SelfAttention.k.weight",
358
+ "text_encoders.t5xxl.transformer.encoder.block.22.layer.0.SelfAttention.o.weight",
359
+ "text_encoders.t5xxl.transformer.encoder.block.22.layer.0.SelfAttention.q.weight",
360
+ "text_encoders.t5xxl.transformer.encoder.block.22.layer.0.SelfAttention.v.weight",
361
+ "text_encoders.t5xxl.transformer.encoder.block.22.layer.0.layer_norm.weight",
362
+ "text_encoders.t5xxl.transformer.encoder.block.22.layer.1.DenseReluDense.wi_0.weight",
363
+ "text_encoders.t5xxl.transformer.encoder.block.22.layer.1.DenseReluDense.wi_1.weight",
364
+ "text_encoders.t5xxl.transformer.encoder.block.22.layer.1.DenseReluDense.wo.weight",
365
+ "text_encoders.t5xxl.transformer.encoder.block.22.layer.1.layer_norm.weight",
366
+ "text_encoders.t5xxl.transformer.encoder.block.23.layer.0.SelfAttention.k.weight",
367
+ "text_encoders.t5xxl.transformer.encoder.block.23.layer.0.SelfAttention.o.weight",
368
+ "text_encoders.t5xxl.transformer.encoder.block.23.layer.0.SelfAttention.q.weight",
369
+ "text_encoders.t5xxl.transformer.encoder.block.23.layer.0.SelfAttention.v.weight",
370
+ "text_encoders.t5xxl.transformer.encoder.block.23.layer.0.layer_norm.weight",
371
+ "text_encoders.t5xxl.transformer.encoder.block.23.layer.1.DenseReluDense.wi_0.weight",
372
+ "text_encoders.t5xxl.transformer.encoder.block.23.layer.1.DenseReluDense.wi_1.weight",
373
+ "text_encoders.t5xxl.transformer.encoder.block.23.layer.1.DenseReluDense.wo.weight",
374
+ "text_encoders.t5xxl.transformer.encoder.block.23.layer.1.layer_norm.weight",
375
+ "text_encoders.t5xxl.transformer.encoder.block.3.layer.0.SelfAttention.k.weight",
376
+ "text_encoders.t5xxl.transformer.encoder.block.3.layer.0.SelfAttention.o.weight",
377
+ "text_encoders.t5xxl.transformer.encoder.block.3.layer.0.SelfAttention.q.weight",
378
+ "text_encoders.t5xxl.transformer.encoder.block.3.layer.0.SelfAttention.v.weight",
379
+ "text_encoders.t5xxl.transformer.encoder.block.3.layer.0.layer_norm.weight",
380
+ "text_encoders.t5xxl.transformer.encoder.block.3.layer.1.DenseReluDense.wi_0.weight",
381
+ "text_encoders.t5xxl.transformer.encoder.block.3.layer.1.DenseReluDense.wi_1.weight",
382
+ "text_encoders.t5xxl.transformer.encoder.block.3.layer.1.DenseReluDense.wo.weight",
383
+ "text_encoders.t5xxl.transformer.encoder.block.3.layer.1.layer_norm.weight",
384
+ "text_encoders.t5xxl.transformer.encoder.block.4.layer.0.SelfAttention.k.weight",
385
+ "text_encoders.t5xxl.transformer.encoder.block.4.layer.0.SelfAttention.o.weight",
386
+ "text_encoders.t5xxl.transformer.encoder.block.4.layer.0.SelfAttention.q.weight",
387
+ "text_encoders.t5xxl.transformer.encoder.block.4.layer.0.SelfAttention.v.weight",
388
+ "text_encoders.t5xxl.transformer.encoder.block.4.layer.0.layer_norm.weight",
389
+ "text_encoders.t5xxl.transformer.encoder.block.4.layer.1.DenseReluDense.wi_0.weight",
390
+ "text_encoders.t5xxl.transformer.encoder.block.4.layer.1.DenseReluDense.wi_1.weight",
391
+ "text_encoders.t5xxl.transformer.encoder.block.4.layer.1.DenseReluDense.wo.weight",
392
+ "text_encoders.t5xxl.transformer.encoder.block.4.layer.1.layer_norm.weight",
393
+ "text_encoders.t5xxl.transformer.encoder.block.5.layer.0.SelfAttention.k.weight",
394
+ "text_encoders.t5xxl.transformer.encoder.block.5.layer.0.SelfAttention.o.weight",
395
+ "text_encoders.t5xxl.transformer.encoder.block.5.layer.0.SelfAttention.q.weight",
396
+ "text_encoders.t5xxl.transformer.encoder.block.5.layer.0.SelfAttention.v.weight",
397
+ "text_encoders.t5xxl.transformer.encoder.block.5.layer.0.layer_norm.weight",
398
+ "text_encoders.t5xxl.transformer.encoder.block.5.layer.1.DenseReluDense.wi_0.weight",
399
+ "text_encoders.t5xxl.transformer.encoder.block.5.layer.1.DenseReluDense.wi_1.weight",
400
+ "text_encoders.t5xxl.transformer.encoder.block.5.layer.1.DenseReluDense.wo.weight",
401
+ "text_encoders.t5xxl.transformer.encoder.block.5.layer.1.layer_norm.weight",
402
+ "text_encoders.t5xxl.transformer.encoder.block.6.layer.0.SelfAttention.k.weight",
403
+ "text_encoders.t5xxl.transformer.encoder.block.6.layer.0.SelfAttention.o.weight",
404
+ "text_encoders.t5xxl.transformer.encoder.block.6.layer.0.SelfAttention.q.weight",
405
+ "text_encoders.t5xxl.transformer.encoder.block.6.layer.0.SelfAttention.v.weight",
406
+ "text_encoders.t5xxl.transformer.encoder.block.6.layer.0.layer_norm.weight",
407
+ "text_encoders.t5xxl.transformer.encoder.block.6.layer.1.DenseReluDense.wi_0.weight",
408
+ "text_encoders.t5xxl.transformer.encoder.block.6.layer.1.DenseReluDense.wi_1.weight",
409
+ "text_encoders.t5xxl.transformer.encoder.block.6.layer.1.DenseReluDense.wo.weight",
410
+ "text_encoders.t5xxl.transformer.encoder.block.6.layer.1.layer_norm.weight",
411
+ "text_encoders.t5xxl.transformer.encoder.block.7.layer.0.SelfAttention.k.weight",
412
+ "text_encoders.t5xxl.transformer.encoder.block.7.layer.0.SelfAttention.o.weight",
413
+ "text_encoders.t5xxl.transformer.encoder.block.7.layer.0.SelfAttention.q.weight",
414
+ "text_encoders.t5xxl.transformer.encoder.block.7.layer.0.SelfAttention.v.weight",
415
+ "text_encoders.t5xxl.transformer.encoder.block.7.layer.0.layer_norm.weight",
416
+ "text_encoders.t5xxl.transformer.encoder.block.7.layer.1.DenseReluDense.wi_0.weight",
417
+ "text_encoders.t5xxl.transformer.encoder.block.7.layer.1.DenseReluDense.wi_1.weight",
418
+ "text_encoders.t5xxl.transformer.encoder.block.7.layer.1.DenseReluDense.wo.weight",
419
+ "text_encoders.t5xxl.transformer.encoder.block.7.layer.1.layer_norm.weight",
420
+ "text_encoders.t5xxl.transformer.encoder.block.8.layer.0.SelfAttention.k.weight",
421
+ "text_encoders.t5xxl.transformer.encoder.block.8.layer.0.SelfAttention.o.weight",
422
+ "text_encoders.t5xxl.transformer.encoder.block.8.layer.0.SelfAttention.q.weight",
423
+ "text_encoders.t5xxl.transformer.encoder.block.8.layer.0.SelfAttention.v.weight",
424
+ "text_encoders.t5xxl.transformer.encoder.block.8.layer.0.layer_norm.weight",
425
+ "text_encoders.t5xxl.transformer.encoder.block.8.layer.1.DenseReluDense.wi_0.weight",
426
+ "text_encoders.t5xxl.transformer.encoder.block.8.layer.1.DenseReluDense.wi_1.weight",
427
+ "text_encoders.t5xxl.transformer.encoder.block.8.layer.1.DenseReluDense.wo.weight",
428
+ "text_encoders.t5xxl.transformer.encoder.block.8.layer.1.layer_norm.weight",
429
+ "text_encoders.t5xxl.transformer.encoder.block.9.layer.0.SelfAttention.k.weight",
430
+ "text_encoders.t5xxl.transformer.encoder.block.9.layer.0.SelfAttention.o.weight",
431
+ "text_encoders.t5xxl.transformer.encoder.block.9.layer.0.SelfAttention.q.weight",
432
+ "text_encoders.t5xxl.transformer.encoder.block.9.layer.0.SelfAttention.v.weight",
433
+ "text_encoders.t5xxl.transformer.encoder.block.9.layer.0.layer_norm.weight",
434
+ "text_encoders.t5xxl.transformer.encoder.block.9.layer.1.DenseReluDense.wi_0.weight",
435
+ "text_encoders.t5xxl.transformer.encoder.block.9.layer.1.DenseReluDense.wi_1.weight",
436
+ "text_encoders.t5xxl.transformer.encoder.block.9.layer.1.DenseReluDense.wo.weight",
437
+ "text_encoders.t5xxl.transformer.encoder.block.9.layer.1.layer_norm.weight",
438
+ "text_encoders.t5xxl.transformer.encoder.final_layer_norm.weight",
439
+ "text_encoders.t5xxl.transformer.shared.weight"
440
+ ]
flux_transformer_keys.json ADDED
@@ -0,0 +1,1562 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ "double_blocks.0.img_attn.norm.key_norm.scale",
3
+ "double_blocks.0.img_attn.norm.query_norm.scale",
4
+ "double_blocks.0.img_attn.proj.bias",
5
+ "double_blocks.0.img_attn.proj.weight",
6
+ "double_blocks.0.img_attn.qkv.bias",
7
+ "double_blocks.0.img_attn.qkv.weight",
8
+ "double_blocks.0.img_mlp.0.bias",
9
+ "double_blocks.0.img_mlp.0.weight",
10
+ "double_blocks.0.img_mlp.2.bias",
11
+ "double_blocks.0.img_mlp.2.weight",
12
+ "double_blocks.0.img_mod.lin.bias",
13
+ "double_blocks.0.img_mod.lin.weight",
14
+ "double_blocks.0.txt_attn.norm.key_norm.scale",
15
+ "double_blocks.0.txt_attn.norm.query_norm.scale",
16
+ "double_blocks.0.txt_attn.proj.bias",
17
+ "double_blocks.0.txt_attn.proj.weight",
18
+ "double_blocks.0.txt_attn.qkv.bias",
19
+ "double_blocks.0.txt_attn.qkv.weight",
20
+ "double_blocks.0.txt_mlp.0.bias",
21
+ "double_blocks.0.txt_mlp.0.weight",
22
+ "double_blocks.0.txt_mlp.2.bias",
23
+ "double_blocks.0.txt_mlp.2.weight",
24
+ "double_blocks.0.txt_mod.lin.bias",
25
+ "double_blocks.0.txt_mod.lin.weight",
26
+ "double_blocks.1.img_attn.norm.key_norm.scale",
27
+ "double_blocks.1.img_attn.norm.query_norm.scale",
28
+ "double_blocks.1.img_attn.proj.bias",
29
+ "double_blocks.1.img_attn.proj.weight",
30
+ "double_blocks.1.img_attn.qkv.bias",
31
+ "double_blocks.1.img_attn.qkv.weight",
32
+ "double_blocks.1.img_mlp.0.bias",
33
+ "double_blocks.1.img_mlp.0.weight",
34
+ "double_blocks.1.img_mlp.2.bias",
35
+ "double_blocks.1.img_mlp.2.weight",
36
+ "double_blocks.1.img_mod.lin.bias",
37
+ "double_blocks.1.img_mod.lin.weight",
38
+ "double_blocks.1.txt_attn.norm.key_norm.scale",
39
+ "double_blocks.1.txt_attn.norm.query_norm.scale",
40
+ "double_blocks.1.txt_attn.proj.bias",
41
+ "double_blocks.1.txt_attn.proj.weight",
42
+ "double_blocks.1.txt_attn.qkv.bias",
43
+ "double_blocks.1.txt_attn.qkv.weight",
44
+ "double_blocks.1.txt_mlp.0.bias",
45
+ "double_blocks.1.txt_mlp.0.weight",
46
+ "double_blocks.1.txt_mlp.2.bias",
47
+ "double_blocks.1.txt_mlp.2.weight",
48
+ "double_blocks.1.txt_mod.lin.bias",
49
+ "double_blocks.1.txt_mod.lin.weight",
50
+ "double_blocks.10.img_attn.norm.key_norm.scale",
51
+ "double_blocks.10.img_attn.norm.query_norm.scale",
52
+ "double_blocks.10.img_attn.proj.bias",
53
+ "double_blocks.10.img_attn.proj.weight",
54
+ "double_blocks.10.img_attn.qkv.bias",
55
+ "double_blocks.10.img_attn.qkv.weight",
56
+ "double_blocks.10.img_mlp.0.bias",
57
+ "double_blocks.10.img_mlp.0.weight",
58
+ "double_blocks.10.img_mlp.2.bias",
59
+ "double_blocks.10.img_mlp.2.weight",
60
+ "double_blocks.10.img_mod.lin.bias",
61
+ "double_blocks.10.img_mod.lin.weight",
62
+ "double_blocks.10.txt_attn.norm.key_norm.scale",
63
+ "double_blocks.10.txt_attn.norm.query_norm.scale",
64
+ "double_blocks.10.txt_attn.proj.bias",
65
+ "double_blocks.10.txt_attn.proj.weight",
66
+ "double_blocks.10.txt_attn.qkv.bias",
67
+ "double_blocks.10.txt_attn.qkv.weight",
68
+ "double_blocks.10.txt_mlp.0.bias",
69
+ "double_blocks.10.txt_mlp.0.weight",
70
+ "double_blocks.10.txt_mlp.2.bias",
71
+ "double_blocks.10.txt_mlp.2.weight",
72
+ "double_blocks.10.txt_mod.lin.bias",
73
+ "double_blocks.10.txt_mod.lin.weight",
74
+ "double_blocks.11.img_attn.norm.key_norm.scale",
75
+ "double_blocks.11.img_attn.norm.query_norm.scale",
76
+ "double_blocks.11.img_attn.proj.bias",
77
+ "double_blocks.11.img_attn.proj.weight",
78
+ "double_blocks.11.img_attn.qkv.bias",
79
+ "double_blocks.11.img_attn.qkv.weight",
80
+ "double_blocks.11.img_mlp.0.bias",
81
+ "double_blocks.11.img_mlp.0.weight",
82
+ "double_blocks.11.img_mlp.2.bias",
83
+ "double_blocks.11.img_mlp.2.weight",
84
+ "double_blocks.11.img_mod.lin.bias",
85
+ "double_blocks.11.img_mod.lin.weight",
86
+ "double_blocks.11.txt_attn.norm.key_norm.scale",
87
+ "double_blocks.11.txt_attn.norm.query_norm.scale",
88
+ "double_blocks.11.txt_attn.proj.bias",
89
+ "double_blocks.11.txt_attn.proj.weight",
90
+ "double_blocks.11.txt_attn.qkv.bias",
91
+ "double_blocks.11.txt_attn.qkv.weight",
92
+ "double_blocks.11.txt_mlp.0.bias",
93
+ "double_blocks.11.txt_mlp.0.weight",
94
+ "double_blocks.11.txt_mlp.2.bias",
95
+ "double_blocks.11.txt_mlp.2.weight",
96
+ "double_blocks.11.txt_mod.lin.bias",
97
+ "double_blocks.11.txt_mod.lin.weight",
98
+ "double_blocks.12.img_attn.norm.key_norm.scale",
99
+ "double_blocks.12.img_attn.norm.query_norm.scale",
100
+ "double_blocks.12.img_attn.proj.bias",
101
+ "double_blocks.12.img_attn.proj.weight",
102
+ "double_blocks.12.img_attn.qkv.bias",
103
+ "double_blocks.12.img_attn.qkv.weight",
104
+ "double_blocks.12.img_mlp.0.bias",
105
+ "double_blocks.12.img_mlp.0.weight",
106
+ "double_blocks.12.img_mlp.2.bias",
107
+ "double_blocks.12.img_mlp.2.weight",
108
+ "double_blocks.12.img_mod.lin.bias",
109
+ "double_blocks.12.img_mod.lin.weight",
110
+ "double_blocks.12.txt_attn.norm.key_norm.scale",
111
+ "double_blocks.12.txt_attn.norm.query_norm.scale",
112
+ "double_blocks.12.txt_attn.proj.bias",
113
+ "double_blocks.12.txt_attn.proj.weight",
114
+ "double_blocks.12.txt_attn.qkv.bias",
115
+ "double_blocks.12.txt_attn.qkv.weight",
116
+ "double_blocks.12.txt_mlp.0.bias",
117
+ "double_blocks.12.txt_mlp.0.weight",
118
+ "double_blocks.12.txt_mlp.2.bias",
119
+ "double_blocks.12.txt_mlp.2.weight",
120
+ "double_blocks.12.txt_mod.lin.bias",
121
+ "double_blocks.12.txt_mod.lin.weight",
122
+ "double_blocks.13.img_attn.norm.key_norm.scale",
123
+ "double_blocks.13.img_attn.norm.query_norm.scale",
124
+ "double_blocks.13.img_attn.proj.bias",
125
+ "double_blocks.13.img_attn.proj.weight",
126
+ "double_blocks.13.img_attn.qkv.bias",
127
+ "double_blocks.13.img_attn.qkv.weight",
128
+ "double_blocks.13.img_mlp.0.bias",
129
+ "double_blocks.13.img_mlp.0.weight",
130
+ "double_blocks.13.img_mlp.2.bias",
131
+ "double_blocks.13.img_mlp.2.weight",
132
+ "double_blocks.13.img_mod.lin.bias",
133
+ "double_blocks.13.img_mod.lin.weight",
134
+ "double_blocks.13.txt_attn.norm.key_norm.scale",
135
+ "double_blocks.13.txt_attn.norm.query_norm.scale",
136
+ "double_blocks.13.txt_attn.proj.bias",
137
+ "double_blocks.13.txt_attn.proj.weight",
138
+ "double_blocks.13.txt_attn.qkv.bias",
139
+ "double_blocks.13.txt_attn.qkv.weight",
140
+ "double_blocks.13.txt_mlp.0.bias",
141
+ "double_blocks.13.txt_mlp.0.weight",
142
+ "double_blocks.13.txt_mlp.2.bias",
143
+ "double_blocks.13.txt_mlp.2.weight",
144
+ "double_blocks.13.txt_mod.lin.bias",
145
+ "double_blocks.13.txt_mod.lin.weight",
146
+ "double_blocks.14.img_attn.norm.key_norm.scale",
147
+ "double_blocks.14.img_attn.norm.query_norm.scale",
148
+ "double_blocks.14.img_attn.proj.bias",
149
+ "double_blocks.14.img_attn.proj.weight",
150
+ "double_blocks.14.img_attn.qkv.bias",
151
+ "double_blocks.14.img_attn.qkv.weight",
152
+ "double_blocks.14.img_mlp.0.bias",
153
+ "double_blocks.14.img_mlp.0.weight",
154
+ "double_blocks.14.img_mlp.2.bias",
155
+ "double_blocks.14.img_mlp.2.weight",
156
+ "double_blocks.14.img_mod.lin.bias",
157
+ "double_blocks.14.img_mod.lin.weight",
158
+ "double_blocks.14.txt_attn.norm.key_norm.scale",
159
+ "double_blocks.14.txt_attn.norm.query_norm.scale",
160
+ "double_blocks.14.txt_attn.proj.bias",
161
+ "double_blocks.14.txt_attn.proj.weight",
162
+ "double_blocks.14.txt_attn.qkv.bias",
163
+ "double_blocks.14.txt_attn.qkv.weight",
164
+ "double_blocks.14.txt_mlp.0.bias",
165
+ "double_blocks.14.txt_mlp.0.weight",
166
+ "double_blocks.14.txt_mlp.2.bias",
167
+ "double_blocks.14.txt_mlp.2.weight",
168
+ "double_blocks.14.txt_mod.lin.bias",
169
+ "double_blocks.14.txt_mod.lin.weight",
170
+ "double_blocks.15.img_attn.norm.key_norm.scale",
171
+ "double_blocks.15.img_attn.norm.query_norm.scale",
172
+ "double_blocks.15.img_attn.proj.bias",
173
+ "double_blocks.15.img_attn.proj.weight",
174
+ "double_blocks.15.img_attn.qkv.bias",
175
+ "double_blocks.15.img_attn.qkv.weight",
176
+ "double_blocks.15.img_mlp.0.bias",
177
+ "double_blocks.15.img_mlp.0.weight",
178
+ "double_blocks.15.img_mlp.2.bias",
179
+ "double_blocks.15.img_mlp.2.weight",
180
+ "double_blocks.15.img_mod.lin.bias",
181
+ "double_blocks.15.img_mod.lin.weight",
182
+ "double_blocks.15.txt_attn.norm.key_norm.scale",
183
+ "double_blocks.15.txt_attn.norm.query_norm.scale",
184
+ "double_blocks.15.txt_attn.proj.bias",
185
+ "double_blocks.15.txt_attn.proj.weight",
186
+ "double_blocks.15.txt_attn.qkv.bias",
187
+ "double_blocks.15.txt_attn.qkv.weight",
188
+ "double_blocks.15.txt_mlp.0.bias",
189
+ "double_blocks.15.txt_mlp.0.weight",
190
+ "double_blocks.15.txt_mlp.2.bias",
191
+ "double_blocks.15.txt_mlp.2.weight",
192
+ "double_blocks.15.txt_mod.lin.bias",
193
+ "double_blocks.15.txt_mod.lin.weight",
194
+ "double_blocks.16.img_attn.norm.key_norm.scale",
195
+ "double_blocks.16.img_attn.norm.query_norm.scale",
196
+ "double_blocks.16.img_attn.proj.bias",
197
+ "double_blocks.16.img_attn.proj.weight",
198
+ "double_blocks.16.img_attn.qkv.bias",
199
+ "double_blocks.16.img_attn.qkv.weight",
200
+ "double_blocks.16.img_mlp.0.bias",
201
+ "double_blocks.16.img_mlp.0.weight",
202
+ "double_blocks.16.img_mlp.2.bias",
203
+ "double_blocks.16.img_mlp.2.weight",
204
+ "double_blocks.16.img_mod.lin.bias",
205
+ "double_blocks.16.img_mod.lin.weight",
206
+ "double_blocks.16.txt_attn.norm.key_norm.scale",
207
+ "double_blocks.16.txt_attn.norm.query_norm.scale",
208
+ "double_blocks.16.txt_attn.proj.bias",
209
+ "double_blocks.16.txt_attn.proj.weight",
210
+ "double_blocks.16.txt_attn.qkv.bias",
211
+ "double_blocks.16.txt_attn.qkv.weight",
212
+ "double_blocks.16.txt_mlp.0.bias",
213
+ "double_blocks.16.txt_mlp.0.weight",
214
+ "double_blocks.16.txt_mlp.2.bias",
215
+ "double_blocks.16.txt_mlp.2.weight",
216
+ "double_blocks.16.txt_mod.lin.bias",
217
+ "double_blocks.16.txt_mod.lin.weight",
218
+ "double_blocks.17.img_attn.norm.key_norm.scale",
219
+ "double_blocks.17.img_attn.norm.query_norm.scale",
220
+ "double_blocks.17.img_attn.proj.bias",
221
+ "double_blocks.17.img_attn.proj.weight",
222
+ "double_blocks.17.img_attn.qkv.bias",
223
+ "double_blocks.17.img_attn.qkv.weight",
224
+ "double_blocks.17.img_mlp.0.bias",
225
+ "double_blocks.17.img_mlp.0.weight",
226
+ "double_blocks.17.img_mlp.2.bias",
227
+ "double_blocks.17.img_mlp.2.weight",
228
+ "double_blocks.17.img_mod.lin.bias",
229
+ "double_blocks.17.img_mod.lin.weight",
230
+ "double_blocks.17.txt_attn.norm.key_norm.scale",
231
+ "double_blocks.17.txt_attn.norm.query_norm.scale",
232
+ "double_blocks.17.txt_attn.proj.bias",
233
+ "double_blocks.17.txt_attn.proj.weight",
234
+ "double_blocks.17.txt_attn.qkv.bias",
235
+ "double_blocks.17.txt_attn.qkv.weight",
236
+ "double_blocks.17.txt_mlp.0.bias",
237
+ "double_blocks.17.txt_mlp.0.weight",
238
+ "double_blocks.17.txt_mlp.2.bias",
239
+ "double_blocks.17.txt_mlp.2.weight",
240
+ "double_blocks.17.txt_mod.lin.bias",
241
+ "double_blocks.17.txt_mod.lin.weight",
242
+ "double_blocks.18.img_attn.norm.key_norm.scale",
243
+ "double_blocks.18.img_attn.norm.query_norm.scale",
244
+ "double_blocks.18.img_attn.proj.bias",
245
+ "double_blocks.18.img_attn.proj.weight",
246
+ "double_blocks.18.img_attn.qkv.bias",
247
+ "double_blocks.18.img_attn.qkv.weight",
248
+ "double_blocks.18.img_mlp.0.bias",
249
+ "double_blocks.18.img_mlp.0.weight",
250
+ "double_blocks.18.img_mlp.2.bias",
251
+ "double_blocks.18.img_mlp.2.weight",
252
+ "double_blocks.18.img_mod.lin.bias",
253
+ "double_blocks.18.img_mod.lin.weight",
254
+ "double_blocks.18.txt_attn.norm.key_norm.scale",
255
+ "double_blocks.18.txt_attn.norm.query_norm.scale",
256
+ "double_blocks.18.txt_attn.proj.bias",
257
+ "double_blocks.18.txt_attn.proj.weight",
258
+ "double_blocks.18.txt_attn.qkv.bias",
259
+ "double_blocks.18.txt_attn.qkv.weight",
260
+ "double_blocks.18.txt_mlp.0.bias",
261
+ "double_blocks.18.txt_mlp.0.weight",
262
+ "double_blocks.18.txt_mlp.2.bias",
263
+ "double_blocks.18.txt_mlp.2.weight",
264
+ "double_blocks.18.txt_mod.lin.bias",
265
+ "double_blocks.18.txt_mod.lin.weight",
266
+ "double_blocks.2.img_attn.norm.key_norm.scale",
267
+ "double_blocks.2.img_attn.norm.query_norm.scale",
268
+ "double_blocks.2.img_attn.proj.bias",
269
+ "double_blocks.2.img_attn.proj.weight",
270
+ "double_blocks.2.img_attn.qkv.bias",
271
+ "double_blocks.2.img_attn.qkv.weight",
272
+ "double_blocks.2.img_mlp.0.bias",
273
+ "double_blocks.2.img_mlp.0.weight",
274
+ "double_blocks.2.img_mlp.2.bias",
275
+ "double_blocks.2.img_mlp.2.weight",
276
+ "double_blocks.2.img_mod.lin.bias",
277
+ "double_blocks.2.img_mod.lin.weight",
278
+ "double_blocks.2.txt_attn.norm.key_norm.scale",
279
+ "double_blocks.2.txt_attn.norm.query_norm.scale",
280
+ "double_blocks.2.txt_attn.proj.bias",
281
+ "double_blocks.2.txt_attn.proj.weight",
282
+ "double_blocks.2.txt_attn.qkv.bias",
283
+ "double_blocks.2.txt_attn.qkv.weight",
284
+ "double_blocks.2.txt_mlp.0.bias",
285
+ "double_blocks.2.txt_mlp.0.weight",
286
+ "double_blocks.2.txt_mlp.2.bias",
287
+ "double_blocks.2.txt_mlp.2.weight",
288
+ "double_blocks.2.txt_mod.lin.bias",
289
+ "double_blocks.2.txt_mod.lin.weight",
290
+ "double_blocks.3.img_attn.norm.key_norm.scale",
291
+ "double_blocks.3.img_attn.norm.query_norm.scale",
292
+ "double_blocks.3.img_attn.proj.bias",
293
+ "double_blocks.3.img_attn.proj.weight",
294
+ "double_blocks.3.img_attn.qkv.bias",
295
+ "double_blocks.3.img_attn.qkv.weight",
296
+ "double_blocks.3.img_mlp.0.bias",
297
+ "double_blocks.3.img_mlp.0.weight",
298
+ "double_blocks.3.img_mlp.2.bias",
299
+ "double_blocks.3.img_mlp.2.weight",
300
+ "double_blocks.3.img_mod.lin.bias",
301
+ "double_blocks.3.img_mod.lin.weight",
302
+ "double_blocks.3.txt_attn.norm.key_norm.scale",
303
+ "double_blocks.3.txt_attn.norm.query_norm.scale",
304
+ "double_blocks.3.txt_attn.proj.bias",
305
+ "double_blocks.3.txt_attn.proj.weight",
306
+ "double_blocks.3.txt_attn.qkv.bias",
307
+ "double_blocks.3.txt_attn.qkv.weight",
308
+ "double_blocks.3.txt_mlp.0.bias",
309
+ "double_blocks.3.txt_mlp.0.weight",
310
+ "double_blocks.3.txt_mlp.2.bias",
311
+ "double_blocks.3.txt_mlp.2.weight",
312
+ "double_blocks.3.txt_mod.lin.bias",
313
+ "double_blocks.3.txt_mod.lin.weight",
314
+ "double_blocks.4.img_attn.norm.key_norm.scale",
315
+ "double_blocks.4.img_attn.norm.query_norm.scale",
316
+ "double_blocks.4.img_attn.proj.bias",
317
+ "double_blocks.4.img_attn.proj.weight",
318
+ "double_blocks.4.img_attn.qkv.bias",
319
+ "double_blocks.4.img_attn.qkv.weight",
320
+ "double_blocks.4.img_mlp.0.bias",
321
+ "double_blocks.4.img_mlp.0.weight",
322
+ "double_blocks.4.img_mlp.2.bias",
323
+ "double_blocks.4.img_mlp.2.weight",
324
+ "double_blocks.4.img_mod.lin.bias",
325
+ "double_blocks.4.img_mod.lin.weight",
326
+ "double_blocks.4.txt_attn.norm.key_norm.scale",
327
+ "double_blocks.4.txt_attn.norm.query_norm.scale",
328
+ "double_blocks.4.txt_attn.proj.bias",
329
+ "double_blocks.4.txt_attn.proj.weight",
330
+ "double_blocks.4.txt_attn.qkv.bias",
331
+ "double_blocks.4.txt_attn.qkv.weight",
332
+ "double_blocks.4.txt_mlp.0.bias",
333
+ "double_blocks.4.txt_mlp.0.weight",
334
+ "double_blocks.4.txt_mlp.2.bias",
335
+ "double_blocks.4.txt_mlp.2.weight",
336
+ "double_blocks.4.txt_mod.lin.bias",
337
+ "double_blocks.4.txt_mod.lin.weight",
338
+ "double_blocks.5.img_attn.norm.key_norm.scale",
339
+ "double_blocks.5.img_attn.norm.query_norm.scale",
340
+ "double_blocks.5.img_attn.proj.bias",
341
+ "double_blocks.5.img_attn.proj.weight",
342
+ "double_blocks.5.img_attn.qkv.bias",
343
+ "double_blocks.5.img_attn.qkv.weight",
344
+ "double_blocks.5.img_mlp.0.bias",
345
+ "double_blocks.5.img_mlp.0.weight",
346
+ "double_blocks.5.img_mlp.2.bias",
347
+ "double_blocks.5.img_mlp.2.weight",
348
+ "double_blocks.5.img_mod.lin.bias",
349
+ "double_blocks.5.img_mod.lin.weight",
350
+ "double_blocks.5.txt_attn.norm.key_norm.scale",
351
+ "double_blocks.5.txt_attn.norm.query_norm.scale",
352
+ "double_blocks.5.txt_attn.proj.bias",
353
+ "double_blocks.5.txt_attn.proj.weight",
354
+ "double_blocks.5.txt_attn.qkv.bias",
355
+ "double_blocks.5.txt_attn.qkv.weight",
356
+ "double_blocks.5.txt_mlp.0.bias",
357
+ "double_blocks.5.txt_mlp.0.weight",
358
+ "double_blocks.5.txt_mlp.2.bias",
359
+ "double_blocks.5.txt_mlp.2.weight",
360
+ "double_blocks.5.txt_mod.lin.bias",
361
+ "double_blocks.5.txt_mod.lin.weight",
362
+ "double_blocks.6.img_attn.norm.key_norm.scale",
363
+ "double_blocks.6.img_attn.norm.query_norm.scale",
364
+ "double_blocks.6.img_attn.proj.bias",
365
+ "double_blocks.6.img_attn.proj.weight",
366
+ "double_blocks.6.img_attn.qkv.bias",
367
+ "double_blocks.6.img_attn.qkv.weight",
368
+ "double_blocks.6.img_mlp.0.bias",
369
+ "double_blocks.6.img_mlp.0.weight",
370
+ "double_blocks.6.img_mlp.2.bias",
371
+ "double_blocks.6.img_mlp.2.weight",
372
+ "double_blocks.6.img_mod.lin.bias",
373
+ "double_blocks.6.img_mod.lin.weight",
374
+ "double_blocks.6.txt_attn.norm.key_norm.scale",
375
+ "double_blocks.6.txt_attn.norm.query_norm.scale",
376
+ "double_blocks.6.txt_attn.proj.bias",
377
+ "double_blocks.6.txt_attn.proj.weight",
378
+ "double_blocks.6.txt_attn.qkv.bias",
379
+ "double_blocks.6.txt_attn.qkv.weight",
380
+ "double_blocks.6.txt_mlp.0.bias",
381
+ "double_blocks.6.txt_mlp.0.weight",
382
+ "double_blocks.6.txt_mlp.2.bias",
383
+ "double_blocks.6.txt_mlp.2.weight",
384
+ "double_blocks.6.txt_mod.lin.bias",
385
+ "double_blocks.6.txt_mod.lin.weight",
386
+ "double_blocks.7.img_attn.norm.key_norm.scale",
387
+ "double_blocks.7.img_attn.norm.query_norm.scale",
388
+ "double_blocks.7.img_attn.proj.bias",
389
+ "double_blocks.7.img_attn.proj.weight",
390
+ "double_blocks.7.img_attn.qkv.bias",
391
+ "double_blocks.7.img_attn.qkv.weight",
392
+ "double_blocks.7.img_mlp.0.bias",
393
+ "double_blocks.7.img_mlp.0.weight",
394
+ "double_blocks.7.img_mlp.2.bias",
395
+ "double_blocks.7.img_mlp.2.weight",
396
+ "double_blocks.7.img_mod.lin.bias",
397
+ "double_blocks.7.img_mod.lin.weight",
398
+ "double_blocks.7.txt_attn.norm.key_norm.scale",
399
+ "double_blocks.7.txt_attn.norm.query_norm.scale",
400
+ "double_blocks.7.txt_attn.proj.bias",
401
+ "double_blocks.7.txt_attn.proj.weight",
402
+ "double_blocks.7.txt_attn.qkv.bias",
403
+ "double_blocks.7.txt_attn.qkv.weight",
404
+ "double_blocks.7.txt_mlp.0.bias",
405
+ "double_blocks.7.txt_mlp.0.weight",
406
+ "double_blocks.7.txt_mlp.2.bias",
407
+ "double_blocks.7.txt_mlp.2.weight",
408
+ "double_blocks.7.txt_mod.lin.bias",
409
+ "double_blocks.7.txt_mod.lin.weight",
410
+ "double_blocks.8.img_attn.norm.key_norm.scale",
411
+ "double_blocks.8.img_attn.norm.query_norm.scale",
412
+ "double_blocks.8.img_attn.proj.bias",
413
+ "double_blocks.8.img_attn.proj.weight",
414
+ "double_blocks.8.img_attn.qkv.bias",
415
+ "double_blocks.8.img_attn.qkv.weight",
416
+ "double_blocks.8.img_mlp.0.bias",
417
+ "double_blocks.8.img_mlp.0.weight",
418
+ "double_blocks.8.img_mlp.2.bias",
419
+ "double_blocks.8.img_mlp.2.weight",
420
+ "double_blocks.8.img_mod.lin.bias",
421
+ "double_blocks.8.img_mod.lin.weight",
422
+ "double_blocks.8.txt_attn.norm.key_norm.scale",
423
+ "double_blocks.8.txt_attn.norm.query_norm.scale",
424
+ "double_blocks.8.txt_attn.proj.bias",
425
+ "double_blocks.8.txt_attn.proj.weight",
426
+ "double_blocks.8.txt_attn.qkv.bias",
427
+ "double_blocks.8.txt_attn.qkv.weight",
428
+ "double_blocks.8.txt_mlp.0.bias",
429
+ "double_blocks.8.txt_mlp.0.weight",
430
+ "double_blocks.8.txt_mlp.2.bias",
431
+ "double_blocks.8.txt_mlp.2.weight",
432
+ "double_blocks.8.txt_mod.lin.bias",
433
+ "double_blocks.8.txt_mod.lin.weight",
434
+ "double_blocks.9.img_attn.norm.key_norm.scale",
435
+ "double_blocks.9.img_attn.norm.query_norm.scale",
436
+ "double_blocks.9.img_attn.proj.bias",
437
+ "double_blocks.9.img_attn.proj.weight",
438
+ "double_blocks.9.img_attn.qkv.bias",
439
+ "double_blocks.9.img_attn.qkv.weight",
440
+ "double_blocks.9.img_mlp.0.bias",
441
+ "double_blocks.9.img_mlp.0.weight",
442
+ "double_blocks.9.img_mlp.2.bias",
443
+ "double_blocks.9.img_mlp.2.weight",
444
+ "double_blocks.9.img_mod.lin.bias",
445
+ "double_blocks.9.img_mod.lin.weight",
446
+ "double_blocks.9.txt_attn.norm.key_norm.scale",
447
+ "double_blocks.9.txt_attn.norm.query_norm.scale",
448
+ "double_blocks.9.txt_attn.proj.bias",
449
+ "double_blocks.9.txt_attn.proj.weight",
450
+ "double_blocks.9.txt_attn.qkv.bias",
451
+ "double_blocks.9.txt_attn.qkv.weight",
452
+ "double_blocks.9.txt_mlp.0.bias",
453
+ "double_blocks.9.txt_mlp.0.weight",
454
+ "double_blocks.9.txt_mlp.2.bias",
455
+ "double_blocks.9.txt_mlp.2.weight",
456
+ "double_blocks.9.txt_mod.lin.bias",
457
+ "double_blocks.9.txt_mod.lin.weight",
458
+ "final_layer.adaLN_modulation.1.bias",
459
+ "final_layer.adaLN_modulation.1.weight",
460
+ "final_layer.linear.bias",
461
+ "final_layer.linear.weight",
462
+ "guidance_in.in_layer.bias",
463
+ "guidance_in.in_layer.weight",
464
+ "guidance_in.out_layer.bias",
465
+ "guidance_in.out_layer.weight",
466
+ "img_in.bias",
467
+ "img_in.weight",
468
+ "single_blocks.0.linear1.bias",
469
+ "single_blocks.0.linear1.weight",
470
+ "single_blocks.0.linear2.bias",
471
+ "single_blocks.0.linear2.weight",
472
+ "single_blocks.0.modulation.lin.bias",
473
+ "single_blocks.0.modulation.lin.weight",
474
+ "single_blocks.0.norm.key_norm.scale",
475
+ "single_blocks.0.norm.query_norm.scale",
476
+ "single_blocks.1.linear1.bias",
477
+ "single_blocks.1.linear1.weight",
478
+ "single_blocks.1.linear2.bias",
479
+ "single_blocks.1.linear2.weight",
480
+ "single_blocks.1.modulation.lin.bias",
481
+ "single_blocks.1.modulation.lin.weight",
482
+ "single_blocks.1.norm.key_norm.scale",
483
+ "single_blocks.1.norm.query_norm.scale",
484
+ "single_blocks.10.linear1.bias",
485
+ "single_blocks.10.linear1.weight",
486
+ "single_blocks.10.linear2.bias",
487
+ "single_blocks.10.linear2.weight",
488
+ "single_blocks.10.modulation.lin.bias",
489
+ "single_blocks.10.modulation.lin.weight",
490
+ "single_blocks.10.norm.key_norm.scale",
491
+ "single_blocks.10.norm.query_norm.scale",
492
+ "single_blocks.11.linear1.bias",
493
+ "single_blocks.11.linear1.weight",
494
+ "single_blocks.11.linear2.bias",
495
+ "single_blocks.11.linear2.weight",
496
+ "single_blocks.11.modulation.lin.bias",
497
+ "single_blocks.11.modulation.lin.weight",
498
+ "single_blocks.11.norm.key_norm.scale",
499
+ "single_blocks.11.norm.query_norm.scale",
500
+ "single_blocks.12.linear1.bias",
501
+ "single_blocks.12.linear1.weight",
502
+ "single_blocks.12.linear2.bias",
503
+ "single_blocks.12.linear2.weight",
504
+ "single_blocks.12.modulation.lin.bias",
505
+ "single_blocks.12.modulation.lin.weight",
506
+ "single_blocks.12.norm.key_norm.scale",
507
+ "single_blocks.12.norm.query_norm.scale",
508
+ "single_blocks.13.linear1.bias",
509
+ "single_blocks.13.linear1.weight",
510
+ "single_blocks.13.linear2.bias",
511
+ "single_blocks.13.linear2.weight",
512
+ "single_blocks.13.modulation.lin.bias",
513
+ "single_blocks.13.modulation.lin.weight",
514
+ "single_blocks.13.norm.key_norm.scale",
515
+ "single_blocks.13.norm.query_norm.scale",
516
+ "single_blocks.14.linear1.bias",
517
+ "single_blocks.14.linear1.weight",
518
+ "single_blocks.14.linear2.bias",
519
+ "single_blocks.14.linear2.weight",
520
+ "single_blocks.14.modulation.lin.bias",
521
+ "single_blocks.14.modulation.lin.weight",
522
+ "single_blocks.14.norm.key_norm.scale",
523
+ "single_blocks.14.norm.query_norm.scale",
524
+ "single_blocks.15.linear1.bias",
525
+ "single_blocks.15.linear1.weight",
526
+ "single_blocks.15.linear2.bias",
527
+ "single_blocks.15.linear2.weight",
528
+ "single_blocks.15.modulation.lin.bias",
529
+ "single_blocks.15.modulation.lin.weight",
530
+ "single_blocks.15.norm.key_norm.scale",
531
+ "single_blocks.15.norm.query_norm.scale",
532
+ "single_blocks.16.linear1.bias",
533
+ "single_blocks.16.linear1.weight",
534
+ "single_blocks.16.linear2.bias",
535
+ "single_blocks.16.linear2.weight",
536
+ "single_blocks.16.modulation.lin.bias",
537
+ "single_blocks.16.modulation.lin.weight",
538
+ "single_blocks.16.norm.key_norm.scale",
539
+ "single_blocks.16.norm.query_norm.scale",
540
+ "single_blocks.17.linear1.bias",
541
+ "single_blocks.17.linear1.weight",
542
+ "single_blocks.17.linear2.bias",
543
+ "single_blocks.17.linear2.weight",
544
+ "single_blocks.17.modulation.lin.bias",
545
+ "single_blocks.17.modulation.lin.weight",
546
+ "single_blocks.17.norm.key_norm.scale",
547
+ "single_blocks.17.norm.query_norm.scale",
548
+ "single_blocks.18.linear1.bias",
549
+ "single_blocks.18.linear1.weight",
550
+ "single_blocks.18.linear2.bias",
551
+ "single_blocks.18.linear2.weight",
552
+ "single_blocks.18.modulation.lin.bias",
553
+ "single_blocks.18.modulation.lin.weight",
554
+ "single_blocks.18.norm.key_norm.scale",
555
+ "single_blocks.18.norm.query_norm.scale",
556
+ "single_blocks.19.linear1.bias",
557
+ "single_blocks.19.linear1.weight",
558
+ "single_blocks.19.linear2.bias",
559
+ "single_blocks.19.linear2.weight",
560
+ "single_blocks.19.modulation.lin.bias",
561
+ "single_blocks.19.modulation.lin.weight",
562
+ "single_blocks.19.norm.key_norm.scale",
563
+ "single_blocks.19.norm.query_norm.scale",
564
+ "single_blocks.2.linear1.bias",
565
+ "single_blocks.2.linear1.weight",
566
+ "single_blocks.2.linear2.bias",
567
+ "single_blocks.2.linear2.weight",
568
+ "single_blocks.2.modulation.lin.bias",
569
+ "single_blocks.2.modulation.lin.weight",
570
+ "single_blocks.2.norm.key_norm.scale",
571
+ "single_blocks.2.norm.query_norm.scale",
572
+ "single_blocks.20.linear1.bias",
573
+ "single_blocks.20.linear1.weight",
574
+ "single_blocks.20.linear2.bias",
575
+ "single_blocks.20.linear2.weight",
576
+ "single_blocks.20.modulation.lin.bias",
577
+ "single_blocks.20.modulation.lin.weight",
578
+ "single_blocks.20.norm.key_norm.scale",
579
+ "single_blocks.20.norm.query_norm.scale",
580
+ "single_blocks.21.linear1.bias",
581
+ "single_blocks.21.linear1.weight",
582
+ "single_blocks.21.linear2.bias",
583
+ "single_blocks.21.linear2.weight",
584
+ "single_blocks.21.modulation.lin.bias",
585
+ "single_blocks.21.modulation.lin.weight",
586
+ "single_blocks.21.norm.key_norm.scale",
587
+ "single_blocks.21.norm.query_norm.scale",
588
+ "single_blocks.22.linear1.bias",
589
+ "single_blocks.22.linear1.weight",
590
+ "single_blocks.22.linear2.bias",
591
+ "single_blocks.22.linear2.weight",
592
+ "single_blocks.22.modulation.lin.bias",
593
+ "single_blocks.22.modulation.lin.weight",
594
+ "single_blocks.22.norm.key_norm.scale",
595
+ "single_blocks.22.norm.query_norm.scale",
596
+ "single_blocks.23.linear1.bias",
597
+ "single_blocks.23.linear1.weight",
598
+ "single_blocks.23.linear2.bias",
599
+ "single_blocks.23.linear2.weight",
600
+ "single_blocks.23.modulation.lin.bias",
601
+ "single_blocks.23.modulation.lin.weight",
602
+ "single_blocks.23.norm.key_norm.scale",
603
+ "single_blocks.23.norm.query_norm.scale",
604
+ "single_blocks.24.linear1.bias",
605
+ "single_blocks.24.linear1.weight",
606
+ "single_blocks.24.linear2.bias",
607
+ "single_blocks.24.linear2.weight",
608
+ "single_blocks.24.modulation.lin.bias",
609
+ "single_blocks.24.modulation.lin.weight",
610
+ "single_blocks.24.norm.key_norm.scale",
611
+ "single_blocks.24.norm.query_norm.scale",
612
+ "single_blocks.25.linear1.bias",
613
+ "single_blocks.25.linear1.weight",
614
+ "single_blocks.25.linear2.bias",
615
+ "single_blocks.25.linear2.weight",
616
+ "single_blocks.25.modulation.lin.bias",
617
+ "single_blocks.25.modulation.lin.weight",
618
+ "single_blocks.25.norm.key_norm.scale",
619
+ "single_blocks.25.norm.query_norm.scale",
620
+ "single_blocks.26.linear1.bias",
621
+ "single_blocks.26.linear1.weight",
622
+ "single_blocks.26.linear2.bias",
623
+ "single_blocks.26.linear2.weight",
624
+ "single_blocks.26.modulation.lin.bias",
625
+ "single_blocks.26.modulation.lin.weight",
626
+ "single_blocks.26.norm.key_norm.scale",
627
+ "single_blocks.26.norm.query_norm.scale",
628
+ "single_blocks.27.linear1.bias",
629
+ "single_blocks.27.linear1.weight",
630
+ "single_blocks.27.linear2.bias",
631
+ "single_blocks.27.linear2.weight",
632
+ "single_blocks.27.modulation.lin.bias",
633
+ "single_blocks.27.modulation.lin.weight",
634
+ "single_blocks.27.norm.key_norm.scale",
635
+ "single_blocks.27.norm.query_norm.scale",
636
+ "single_blocks.28.linear1.bias",
637
+ "single_blocks.28.linear1.weight",
638
+ "single_blocks.28.linear2.bias",
639
+ "single_blocks.28.linear2.weight",
640
+ "single_blocks.28.modulation.lin.bias",
641
+ "single_blocks.28.modulation.lin.weight",
642
+ "single_blocks.28.norm.key_norm.scale",
643
+ "single_blocks.28.norm.query_norm.scale",
644
+ "single_blocks.29.linear1.bias",
645
+ "single_blocks.29.linear1.weight",
646
+ "single_blocks.29.linear2.bias",
647
+ "single_blocks.29.linear2.weight",
648
+ "single_blocks.29.modulation.lin.bias",
649
+ "single_blocks.29.modulation.lin.weight",
650
+ "single_blocks.29.norm.key_norm.scale",
651
+ "single_blocks.29.norm.query_norm.scale",
652
+ "single_blocks.3.linear1.bias",
653
+ "single_blocks.3.linear1.weight",
654
+ "single_blocks.3.linear2.bias",
655
+ "single_blocks.3.linear2.weight",
656
+ "single_blocks.3.modulation.lin.bias",
657
+ "single_blocks.3.modulation.lin.weight",
658
+ "single_blocks.3.norm.key_norm.scale",
659
+ "single_blocks.3.norm.query_norm.scale",
660
+ "single_blocks.30.linear1.bias",
661
+ "single_blocks.30.linear1.weight",
662
+ "single_blocks.30.linear2.bias",
663
+ "single_blocks.30.linear2.weight",
664
+ "single_blocks.30.modulation.lin.bias",
665
+ "single_blocks.30.modulation.lin.weight",
666
+ "single_blocks.30.norm.key_norm.scale",
667
+ "single_blocks.30.norm.query_norm.scale",
668
+ "single_blocks.31.linear1.bias",
669
+ "single_blocks.31.linear1.weight",
670
+ "single_blocks.31.linear2.bias",
671
+ "single_blocks.31.linear2.weight",
672
+ "single_blocks.31.modulation.lin.bias",
673
+ "single_blocks.31.modulation.lin.weight",
674
+ "single_blocks.31.norm.key_norm.scale",
675
+ "single_blocks.31.norm.query_norm.scale",
676
+ "single_blocks.32.linear1.bias",
677
+ "single_blocks.32.linear1.weight",
678
+ "single_blocks.32.linear2.bias",
679
+ "single_blocks.32.linear2.weight",
680
+ "single_blocks.32.modulation.lin.bias",
681
+ "single_blocks.32.modulation.lin.weight",
682
+ "single_blocks.32.norm.key_norm.scale",
683
+ "single_blocks.32.norm.query_norm.scale",
684
+ "single_blocks.33.linear1.bias",
685
+ "single_blocks.33.linear1.weight",
686
+ "single_blocks.33.linear2.bias",
687
+ "single_blocks.33.linear2.weight",
688
+ "single_blocks.33.modulation.lin.bias",
689
+ "single_blocks.33.modulation.lin.weight",
690
+ "single_blocks.33.norm.key_norm.scale",
691
+ "single_blocks.33.norm.query_norm.scale",
692
+ "single_blocks.34.linear1.bias",
693
+ "single_blocks.34.linear1.weight",
694
+ "single_blocks.34.linear2.bias",
695
+ "single_blocks.34.linear2.weight",
696
+ "single_blocks.34.modulation.lin.bias",
697
+ "single_blocks.34.modulation.lin.weight",
698
+ "single_blocks.34.norm.key_norm.scale",
699
+ "single_blocks.34.norm.query_norm.scale",
700
+ "single_blocks.35.linear1.bias",
701
+ "single_blocks.35.linear1.weight",
702
+ "single_blocks.35.linear2.bias",
703
+ "single_blocks.35.linear2.weight",
704
+ "single_blocks.35.modulation.lin.bias",
705
+ "single_blocks.35.modulation.lin.weight",
706
+ "single_blocks.35.norm.key_norm.scale",
707
+ "single_blocks.35.norm.query_norm.scale",
708
+ "single_blocks.36.linear1.bias",
709
+ "single_blocks.36.linear1.weight",
710
+ "single_blocks.36.linear2.bias",
711
+ "single_blocks.36.linear2.weight",
712
+ "single_blocks.36.modulation.lin.bias",
713
+ "single_blocks.36.modulation.lin.weight",
714
+ "single_blocks.36.norm.key_norm.scale",
715
+ "single_blocks.36.norm.query_norm.scale",
716
+ "single_blocks.37.linear1.bias",
717
+ "single_blocks.37.linear1.weight",
718
+ "single_blocks.37.linear2.bias",
719
+ "single_blocks.37.linear2.weight",
720
+ "single_blocks.37.modulation.lin.bias",
721
+ "single_blocks.37.modulation.lin.weight",
722
+ "single_blocks.37.norm.key_norm.scale",
723
+ "single_blocks.37.norm.query_norm.scale",
724
+ "single_blocks.4.linear1.bias",
725
+ "single_blocks.4.linear1.weight",
726
+ "single_blocks.4.linear2.bias",
727
+ "single_blocks.4.linear2.weight",
728
+ "single_blocks.4.modulation.lin.bias",
729
+ "single_blocks.4.modulation.lin.weight",
730
+ "single_blocks.4.norm.key_norm.scale",
731
+ "single_blocks.4.norm.query_norm.scale",
732
+ "single_blocks.5.linear1.bias",
733
+ "single_blocks.5.linear1.weight",
734
+ "single_blocks.5.linear2.bias",
735
+ "single_blocks.5.linear2.weight",
736
+ "single_blocks.5.modulation.lin.bias",
737
+ "single_blocks.5.modulation.lin.weight",
738
+ "single_blocks.5.norm.key_norm.scale",
739
+ "single_blocks.5.norm.query_norm.scale",
740
+ "single_blocks.6.linear1.bias",
741
+ "single_blocks.6.linear1.weight",
742
+ "single_blocks.6.linear2.bias",
743
+ "single_blocks.6.linear2.weight",
744
+ "single_blocks.6.modulation.lin.bias",
745
+ "single_blocks.6.modulation.lin.weight",
746
+ "single_blocks.6.norm.key_norm.scale",
747
+ "single_blocks.6.norm.query_norm.scale",
748
+ "single_blocks.7.linear1.bias",
749
+ "single_blocks.7.linear1.weight",
750
+ "single_blocks.7.linear2.bias",
751
+ "single_blocks.7.linear2.weight",
752
+ "single_blocks.7.modulation.lin.bias",
753
+ "single_blocks.7.modulation.lin.weight",
754
+ "single_blocks.7.norm.key_norm.scale",
755
+ "single_blocks.7.norm.query_norm.scale",
756
+ "single_blocks.8.linear1.bias",
757
+ "single_blocks.8.linear1.weight",
758
+ "single_blocks.8.linear2.bias",
759
+ "single_blocks.8.linear2.weight",
760
+ "single_blocks.8.modulation.lin.bias",
761
+ "single_blocks.8.modulation.lin.weight",
762
+ "single_blocks.8.norm.key_norm.scale",
763
+ "single_blocks.8.norm.query_norm.scale",
764
+ "single_blocks.9.linear1.bias",
765
+ "single_blocks.9.linear1.weight",
766
+ "single_blocks.9.linear2.bias",
767
+ "single_blocks.9.linear2.weight",
768
+ "single_blocks.9.modulation.lin.bias",
769
+ "single_blocks.9.modulation.lin.weight",
770
+ "single_blocks.9.norm.key_norm.scale",
771
+ "single_blocks.9.norm.query_norm.scale",
772
+ "time_in.in_layer.bias",
773
+ "time_in.in_layer.weight",
774
+ "time_in.out_layer.bias",
775
+ "time_in.out_layer.weight",
776
+ "txt_in.bias",
777
+ "txt_in.weight",
778
+ "vector_in.in_layer.bias",
779
+ "vector_in.in_layer.weight",
780
+ "vector_in.out_layer.bias",
781
+ "vector_in.out_layer.weight",
782
+ "model.diffusion_model.double_blocks.0.img_attn.norm.key_norm.scale",
783
+ "model.diffusion_model.double_blocks.0.img_attn.norm.query_norm.scale",
784
+ "model.diffusion_model.double_blocks.0.img_attn.proj.bias",
785
+ "model.diffusion_model.double_blocks.0.img_attn.proj.weight",
786
+ "model.diffusion_model.double_blocks.0.img_attn.qkv.bias",
787
+ "model.diffusion_model.double_blocks.0.img_attn.qkv.weight",
788
+ "model.diffusion_model.double_blocks.0.img_mlp.0.bias",
789
+ "model.diffusion_model.double_blocks.0.img_mlp.0.weight",
790
+ "model.diffusion_model.double_blocks.0.img_mlp.2.bias",
791
+ "model.diffusion_model.double_blocks.0.img_mlp.2.weight",
792
+ "model.diffusion_model.double_blocks.0.img_mod.lin.bias",
793
+ "model.diffusion_model.double_blocks.0.img_mod.lin.weight",
794
+ "model.diffusion_model.double_blocks.0.txt_attn.norm.key_norm.scale",
795
+ "model.diffusion_model.double_blocks.0.txt_attn.norm.query_norm.scale",
796
+ "model.diffusion_model.double_blocks.0.txt_attn.proj.bias",
797
+ "model.diffusion_model.double_blocks.0.txt_attn.proj.weight",
798
+ "model.diffusion_model.double_blocks.0.txt_attn.qkv.bias",
799
+ "model.diffusion_model.double_blocks.0.txt_attn.qkv.weight",
800
+ "model.diffusion_model.double_blocks.0.txt_mlp.0.bias",
801
+ "model.diffusion_model.double_blocks.0.txt_mlp.0.weight",
802
+ "model.diffusion_model.double_blocks.0.txt_mlp.2.bias",
803
+ "model.diffusion_model.double_blocks.0.txt_mlp.2.weight",
804
+ "model.diffusion_model.double_blocks.0.txt_mod.lin.bias",
805
+ "model.diffusion_model.double_blocks.0.txt_mod.lin.weight",
806
+ "model.diffusion_model.double_blocks.1.img_attn.norm.key_norm.scale",
807
+ "model.diffusion_model.double_blocks.1.img_attn.norm.query_norm.scale",
808
+ "model.diffusion_model.double_blocks.1.img_attn.proj.bias",
809
+ "model.diffusion_model.double_blocks.1.img_attn.proj.weight",
810
+ "model.diffusion_model.double_blocks.1.img_attn.qkv.bias",
811
+ "model.diffusion_model.double_blocks.1.img_attn.qkv.weight",
812
+ "model.diffusion_model.double_blocks.1.img_mlp.0.bias",
813
+ "model.diffusion_model.double_blocks.1.img_mlp.0.weight",
814
+ "model.diffusion_model.double_blocks.1.img_mlp.2.bias",
815
+ "model.diffusion_model.double_blocks.1.img_mlp.2.weight",
816
+ "model.diffusion_model.double_blocks.1.img_mod.lin.bias",
817
+ "model.diffusion_model.double_blocks.1.img_mod.lin.weight",
818
+ "model.diffusion_model.double_blocks.1.txt_attn.norm.key_norm.scale",
819
+ "model.diffusion_model.double_blocks.1.txt_attn.norm.query_norm.scale",
820
+ "model.diffusion_model.double_blocks.1.txt_attn.proj.bias",
821
+ "model.diffusion_model.double_blocks.1.txt_attn.proj.weight",
822
+ "model.diffusion_model.double_blocks.1.txt_attn.qkv.bias",
823
+ "model.diffusion_model.double_blocks.1.txt_attn.qkv.weight",
824
+ "model.diffusion_model.double_blocks.1.txt_mlp.0.bias",
825
+ "model.diffusion_model.double_blocks.1.txt_mlp.0.weight",
826
+ "model.diffusion_model.double_blocks.1.txt_mlp.2.bias",
827
+ "model.diffusion_model.double_blocks.1.txt_mlp.2.weight",
828
+ "model.diffusion_model.double_blocks.1.txt_mod.lin.bias",
829
+ "model.diffusion_model.double_blocks.1.txt_mod.lin.weight",
830
+ "model.diffusion_model.double_blocks.10.img_attn.norm.key_norm.scale",
831
+ "model.diffusion_model.double_blocks.10.img_attn.norm.query_norm.scale",
832
+ "model.diffusion_model.double_blocks.10.img_attn.proj.bias",
833
+ "model.diffusion_model.double_blocks.10.img_attn.proj.weight",
834
+ "model.diffusion_model.double_blocks.10.img_attn.qkv.bias",
835
+ "model.diffusion_model.double_blocks.10.img_attn.qkv.weight",
836
+ "model.diffusion_model.double_blocks.10.img_mlp.0.bias",
837
+ "model.diffusion_model.double_blocks.10.img_mlp.0.weight",
838
+ "model.diffusion_model.double_blocks.10.img_mlp.2.bias",
839
+ "model.diffusion_model.double_blocks.10.img_mlp.2.weight",
840
+ "model.diffusion_model.double_blocks.10.img_mod.lin.bias",
841
+ "model.diffusion_model.double_blocks.10.img_mod.lin.weight",
842
+ "model.diffusion_model.double_blocks.10.txt_attn.norm.key_norm.scale",
843
+ "model.diffusion_model.double_blocks.10.txt_attn.norm.query_norm.scale",
844
+ "model.diffusion_model.double_blocks.10.txt_attn.proj.bias",
845
+ "model.diffusion_model.double_blocks.10.txt_attn.proj.weight",
846
+ "model.diffusion_model.double_blocks.10.txt_attn.qkv.bias",
847
+ "model.diffusion_model.double_blocks.10.txt_attn.qkv.weight",
848
+ "model.diffusion_model.double_blocks.10.txt_mlp.0.bias",
849
+ "model.diffusion_model.double_blocks.10.txt_mlp.0.weight",
850
+ "model.diffusion_model.double_blocks.10.txt_mlp.2.bias",
851
+ "model.diffusion_model.double_blocks.10.txt_mlp.2.weight",
852
+ "model.diffusion_model.double_blocks.10.txt_mod.lin.bias",
853
+ "model.diffusion_model.double_blocks.10.txt_mod.lin.weight",
854
+ "model.diffusion_model.double_blocks.11.img_attn.norm.key_norm.scale",
855
+ "model.diffusion_model.double_blocks.11.img_attn.norm.query_norm.scale",
856
+ "model.diffusion_model.double_blocks.11.img_attn.proj.bias",
857
+ "model.diffusion_model.double_blocks.11.img_attn.proj.weight",
858
+ "model.diffusion_model.double_blocks.11.img_attn.qkv.bias",
859
+ "model.diffusion_model.double_blocks.11.img_attn.qkv.weight",
860
+ "model.diffusion_model.double_blocks.11.img_mlp.0.bias",
861
+ "model.diffusion_model.double_blocks.11.img_mlp.0.weight",
862
+ "model.diffusion_model.double_blocks.11.img_mlp.2.bias",
863
+ "model.diffusion_model.double_blocks.11.img_mlp.2.weight",
864
+ "model.diffusion_model.double_blocks.11.img_mod.lin.bias",
865
+ "model.diffusion_model.double_blocks.11.img_mod.lin.weight",
866
+ "model.diffusion_model.double_blocks.11.txt_attn.norm.key_norm.scale",
867
+ "model.diffusion_model.double_blocks.11.txt_attn.norm.query_norm.scale",
868
+ "model.diffusion_model.double_blocks.11.txt_attn.proj.bias",
869
+ "model.diffusion_model.double_blocks.11.txt_attn.proj.weight",
870
+ "model.diffusion_model.double_blocks.11.txt_attn.qkv.bias",
871
+ "model.diffusion_model.double_blocks.11.txt_attn.qkv.weight",
872
+ "model.diffusion_model.double_blocks.11.txt_mlp.0.bias",
873
+ "model.diffusion_model.double_blocks.11.txt_mlp.0.weight",
874
+ "model.diffusion_model.double_blocks.11.txt_mlp.2.bias",
875
+ "model.diffusion_model.double_blocks.11.txt_mlp.2.weight",
876
+ "model.diffusion_model.double_blocks.11.txt_mod.lin.bias",
877
+ "model.diffusion_model.double_blocks.11.txt_mod.lin.weight",
878
+ "model.diffusion_model.double_blocks.12.img_attn.norm.key_norm.scale",
879
+ "model.diffusion_model.double_blocks.12.img_attn.norm.query_norm.scale",
880
+ "model.diffusion_model.double_blocks.12.img_attn.proj.bias",
881
+ "model.diffusion_model.double_blocks.12.img_attn.proj.weight",
882
+ "model.diffusion_model.double_blocks.12.img_attn.qkv.bias",
883
+ "model.diffusion_model.double_blocks.12.img_attn.qkv.weight",
884
+ "model.diffusion_model.double_blocks.12.img_mlp.0.bias",
885
+ "model.diffusion_model.double_blocks.12.img_mlp.0.weight",
886
+ "model.diffusion_model.double_blocks.12.img_mlp.2.bias",
887
+ "model.diffusion_model.double_blocks.12.img_mlp.2.weight",
888
+ "model.diffusion_model.double_blocks.12.img_mod.lin.bias",
889
+ "model.diffusion_model.double_blocks.12.img_mod.lin.weight",
890
+ "model.diffusion_model.double_blocks.12.txt_attn.norm.key_norm.scale",
891
+ "model.diffusion_model.double_blocks.12.txt_attn.norm.query_norm.scale",
892
+ "model.diffusion_model.double_blocks.12.txt_attn.proj.bias",
893
+ "model.diffusion_model.double_blocks.12.txt_attn.proj.weight",
894
+ "model.diffusion_model.double_blocks.12.txt_attn.qkv.bias",
895
+ "model.diffusion_model.double_blocks.12.txt_attn.qkv.weight",
896
+ "model.diffusion_model.double_blocks.12.txt_mlp.0.bias",
897
+ "model.diffusion_model.double_blocks.12.txt_mlp.0.weight",
898
+ "model.diffusion_model.double_blocks.12.txt_mlp.2.bias",
899
+ "model.diffusion_model.double_blocks.12.txt_mlp.2.weight",
900
+ "model.diffusion_model.double_blocks.12.txt_mod.lin.bias",
901
+ "model.diffusion_model.double_blocks.12.txt_mod.lin.weight",
902
+ "model.diffusion_model.double_blocks.13.img_attn.norm.key_norm.scale",
903
+ "model.diffusion_model.double_blocks.13.img_attn.norm.query_norm.scale",
904
+ "model.diffusion_model.double_blocks.13.img_attn.proj.bias",
905
+ "model.diffusion_model.double_blocks.13.img_attn.proj.weight",
906
+ "model.diffusion_model.double_blocks.13.img_attn.qkv.bias",
907
+ "model.diffusion_model.double_blocks.13.img_attn.qkv.weight",
908
+ "model.diffusion_model.double_blocks.13.img_mlp.0.bias",
909
+ "model.diffusion_model.double_blocks.13.img_mlp.0.weight",
910
+ "model.diffusion_model.double_blocks.13.img_mlp.2.bias",
911
+ "model.diffusion_model.double_blocks.13.img_mlp.2.weight",
912
+ "model.diffusion_model.double_blocks.13.img_mod.lin.bias",
913
+ "model.diffusion_model.double_blocks.13.img_mod.lin.weight",
914
+ "model.diffusion_model.double_blocks.13.txt_attn.norm.key_norm.scale",
915
+ "model.diffusion_model.double_blocks.13.txt_attn.norm.query_norm.scale",
916
+ "model.diffusion_model.double_blocks.13.txt_attn.proj.bias",
917
+ "model.diffusion_model.double_blocks.13.txt_attn.proj.weight",
918
+ "model.diffusion_model.double_blocks.13.txt_attn.qkv.bias",
919
+ "model.diffusion_model.double_blocks.13.txt_attn.qkv.weight",
920
+ "model.diffusion_model.double_blocks.13.txt_mlp.0.bias",
921
+ "model.diffusion_model.double_blocks.13.txt_mlp.0.weight",
922
+ "model.diffusion_model.double_blocks.13.txt_mlp.2.bias",
923
+ "model.diffusion_model.double_blocks.13.txt_mlp.2.weight",
924
+ "model.diffusion_model.double_blocks.13.txt_mod.lin.bias",
925
+ "model.diffusion_model.double_blocks.13.txt_mod.lin.weight",
926
+ "model.diffusion_model.double_blocks.14.img_attn.norm.key_norm.scale",
927
+ "model.diffusion_model.double_blocks.14.img_attn.norm.query_norm.scale",
928
+ "model.diffusion_model.double_blocks.14.img_attn.proj.bias",
929
+ "model.diffusion_model.double_blocks.14.img_attn.proj.weight",
930
+ "model.diffusion_model.double_blocks.14.img_attn.qkv.bias",
931
+ "model.diffusion_model.double_blocks.14.img_attn.qkv.weight",
932
+ "model.diffusion_model.double_blocks.14.img_mlp.0.bias",
933
+ "model.diffusion_model.double_blocks.14.img_mlp.0.weight",
934
+ "model.diffusion_model.double_blocks.14.img_mlp.2.bias",
935
+ "model.diffusion_model.double_blocks.14.img_mlp.2.weight",
936
+ "model.diffusion_model.double_blocks.14.img_mod.lin.bias",
937
+ "model.diffusion_model.double_blocks.14.img_mod.lin.weight",
938
+ "model.diffusion_model.double_blocks.14.txt_attn.norm.key_norm.scale",
939
+ "model.diffusion_model.double_blocks.14.txt_attn.norm.query_norm.scale",
940
+ "model.diffusion_model.double_blocks.14.txt_attn.proj.bias",
941
+ "model.diffusion_model.double_blocks.14.txt_attn.proj.weight",
942
+ "model.diffusion_model.double_blocks.14.txt_attn.qkv.bias",
943
+ "model.diffusion_model.double_blocks.14.txt_attn.qkv.weight",
944
+ "model.diffusion_model.double_blocks.14.txt_mlp.0.bias",
945
+ "model.diffusion_model.double_blocks.14.txt_mlp.0.weight",
946
+ "model.diffusion_model.double_blocks.14.txt_mlp.2.bias",
947
+ "model.diffusion_model.double_blocks.14.txt_mlp.2.weight",
948
+ "model.diffusion_model.double_blocks.14.txt_mod.lin.bias",
949
+ "model.diffusion_model.double_blocks.14.txt_mod.lin.weight",
950
+ "model.diffusion_model.double_blocks.15.img_attn.norm.key_norm.scale",
951
+ "model.diffusion_model.double_blocks.15.img_attn.norm.query_norm.scale",
952
+ "model.diffusion_model.double_blocks.15.img_attn.proj.bias",
953
+ "model.diffusion_model.double_blocks.15.img_attn.proj.weight",
954
+ "model.diffusion_model.double_blocks.15.img_attn.qkv.bias",
955
+ "model.diffusion_model.double_blocks.15.img_attn.qkv.weight",
956
+ "model.diffusion_model.double_blocks.15.img_mlp.0.bias",
957
+ "model.diffusion_model.double_blocks.15.img_mlp.0.weight",
958
+ "model.diffusion_model.double_blocks.15.img_mlp.2.bias",
959
+ "model.diffusion_model.double_blocks.15.img_mlp.2.weight",
960
+ "model.diffusion_model.double_blocks.15.img_mod.lin.bias",
961
+ "model.diffusion_model.double_blocks.15.img_mod.lin.weight",
962
+ "model.diffusion_model.double_blocks.15.txt_attn.norm.key_norm.scale",
963
+ "model.diffusion_model.double_blocks.15.txt_attn.norm.query_norm.scale",
964
+ "model.diffusion_model.double_blocks.15.txt_attn.proj.bias",
965
+ "model.diffusion_model.double_blocks.15.txt_attn.proj.weight",
966
+ "model.diffusion_model.double_blocks.15.txt_attn.qkv.bias",
967
+ "model.diffusion_model.double_blocks.15.txt_attn.qkv.weight",
968
+ "model.diffusion_model.double_blocks.15.txt_mlp.0.bias",
969
+ "model.diffusion_model.double_blocks.15.txt_mlp.0.weight",
970
+ "model.diffusion_model.double_blocks.15.txt_mlp.2.bias",
971
+ "model.diffusion_model.double_blocks.15.txt_mlp.2.weight",
972
+ "model.diffusion_model.double_blocks.15.txt_mod.lin.bias",
973
+ "model.diffusion_model.double_blocks.15.txt_mod.lin.weight",
974
+ "model.diffusion_model.double_blocks.16.img_attn.norm.key_norm.scale",
975
+ "model.diffusion_model.double_blocks.16.img_attn.norm.query_norm.scale",
976
+ "model.diffusion_model.double_blocks.16.img_attn.proj.bias",
977
+ "model.diffusion_model.double_blocks.16.img_attn.proj.weight",
978
+ "model.diffusion_model.double_blocks.16.img_attn.qkv.bias",
979
+ "model.diffusion_model.double_blocks.16.img_attn.qkv.weight",
980
+ "model.diffusion_model.double_blocks.16.img_mlp.0.bias",
981
+ "model.diffusion_model.double_blocks.16.img_mlp.0.weight",
982
+ "model.diffusion_model.double_blocks.16.img_mlp.2.bias",
983
+ "model.diffusion_model.double_blocks.16.img_mlp.2.weight",
984
+ "model.diffusion_model.double_blocks.16.img_mod.lin.bias",
985
+ "model.diffusion_model.double_blocks.16.img_mod.lin.weight",
986
+ "model.diffusion_model.double_blocks.16.txt_attn.norm.key_norm.scale",
987
+ "model.diffusion_model.double_blocks.16.txt_attn.norm.query_norm.scale",
988
+ "model.diffusion_model.double_blocks.16.txt_attn.proj.bias",
989
+ "model.diffusion_model.double_blocks.16.txt_attn.proj.weight",
990
+ "model.diffusion_model.double_blocks.16.txt_attn.qkv.bias",
991
+ "model.diffusion_model.double_blocks.16.txt_attn.qkv.weight",
992
+ "model.diffusion_model.double_blocks.16.txt_mlp.0.bias",
993
+ "model.diffusion_model.double_blocks.16.txt_mlp.0.weight",
994
+ "model.diffusion_model.double_blocks.16.txt_mlp.2.bias",
995
+ "model.diffusion_model.double_blocks.16.txt_mlp.2.weight",
996
+ "model.diffusion_model.double_blocks.16.txt_mod.lin.bias",
997
+ "model.diffusion_model.double_blocks.16.txt_mod.lin.weight",
998
+ "model.diffusion_model.double_blocks.17.img_attn.norm.key_norm.scale",
999
+ "model.diffusion_model.double_blocks.17.img_attn.norm.query_norm.scale",
1000
+ "model.diffusion_model.double_blocks.17.img_attn.proj.bias",
1001
+ "model.diffusion_model.double_blocks.17.img_attn.proj.weight",
1002
+ "model.diffusion_model.double_blocks.17.img_attn.qkv.bias",
1003
+ "model.diffusion_model.double_blocks.17.img_attn.qkv.weight",
1004
+ "model.diffusion_model.double_blocks.17.img_mlp.0.bias",
1005
+ "model.diffusion_model.double_blocks.17.img_mlp.0.weight",
1006
+ "model.diffusion_model.double_blocks.17.img_mlp.2.bias",
1007
+ "model.diffusion_model.double_blocks.17.img_mlp.2.weight",
1008
+ "model.diffusion_model.double_blocks.17.img_mod.lin.bias",
1009
+ "model.diffusion_model.double_blocks.17.img_mod.lin.weight",
1010
+ "model.diffusion_model.double_blocks.17.txt_attn.norm.key_norm.scale",
1011
+ "model.diffusion_model.double_blocks.17.txt_attn.norm.query_norm.scale",
1012
+ "model.diffusion_model.double_blocks.17.txt_attn.proj.bias",
1013
+ "model.diffusion_model.double_blocks.17.txt_attn.proj.weight",
1014
+ "model.diffusion_model.double_blocks.17.txt_attn.qkv.bias",
1015
+ "model.diffusion_model.double_blocks.17.txt_attn.qkv.weight",
1016
+ "model.diffusion_model.double_blocks.17.txt_mlp.0.bias",
1017
+ "model.diffusion_model.double_blocks.17.txt_mlp.0.weight",
1018
+ "model.diffusion_model.double_blocks.17.txt_mlp.2.bias",
1019
+ "model.diffusion_model.double_blocks.17.txt_mlp.2.weight",
1020
+ "model.diffusion_model.double_blocks.17.txt_mod.lin.bias",
1021
+ "model.diffusion_model.double_blocks.17.txt_mod.lin.weight",
1022
+ "model.diffusion_model.double_blocks.18.img_attn.norm.key_norm.scale",
1023
+ "model.diffusion_model.double_blocks.18.img_attn.norm.query_norm.scale",
1024
+ "model.diffusion_model.double_blocks.18.img_attn.proj.bias",
1025
+ "model.diffusion_model.double_blocks.18.img_attn.proj.weight",
1026
+ "model.diffusion_model.double_blocks.18.img_attn.qkv.bias",
1027
+ "model.diffusion_model.double_blocks.18.img_attn.qkv.weight",
1028
+ "model.diffusion_model.double_blocks.18.img_mlp.0.bias",
1029
+ "model.diffusion_model.double_blocks.18.img_mlp.0.weight",
1030
+ "model.diffusion_model.double_blocks.18.img_mlp.2.bias",
1031
+ "model.diffusion_model.double_blocks.18.img_mlp.2.weight",
1032
+ "model.diffusion_model.double_blocks.18.img_mod.lin.bias",
1033
+ "model.diffusion_model.double_blocks.18.img_mod.lin.weight",
1034
+ "model.diffusion_model.double_blocks.18.txt_attn.norm.key_norm.scale",
1035
+ "model.diffusion_model.double_blocks.18.txt_attn.norm.query_norm.scale",
1036
+ "model.diffusion_model.double_blocks.18.txt_attn.proj.bias",
1037
+ "model.diffusion_model.double_blocks.18.txt_attn.proj.weight",
1038
+ "model.diffusion_model.double_blocks.18.txt_attn.qkv.bias",
1039
+ "model.diffusion_model.double_blocks.18.txt_attn.qkv.weight",
1040
+ "model.diffusion_model.double_blocks.18.txt_mlp.0.bias",
1041
+ "model.diffusion_model.double_blocks.18.txt_mlp.0.weight",
1042
+ "model.diffusion_model.double_blocks.18.txt_mlp.2.bias",
1043
+ "model.diffusion_model.double_blocks.18.txt_mlp.2.weight",
1044
+ "model.diffusion_model.double_blocks.18.txt_mod.lin.bias",
1045
+ "model.diffusion_model.double_blocks.18.txt_mod.lin.weight",
1046
+ "model.diffusion_model.double_blocks.2.img_attn.norm.key_norm.scale",
1047
+ "model.diffusion_model.double_blocks.2.img_attn.norm.query_norm.scale",
1048
+ "model.diffusion_model.double_blocks.2.img_attn.proj.bias",
1049
+ "model.diffusion_model.double_blocks.2.img_attn.proj.weight",
1050
+ "model.diffusion_model.double_blocks.2.img_attn.qkv.bias",
1051
+ "model.diffusion_model.double_blocks.2.img_attn.qkv.weight",
1052
+ "model.diffusion_model.double_blocks.2.img_mlp.0.bias",
1053
+ "model.diffusion_model.double_blocks.2.img_mlp.0.weight",
1054
+ "model.diffusion_model.double_blocks.2.img_mlp.2.bias",
1055
+ "model.diffusion_model.double_blocks.2.img_mlp.2.weight",
1056
+ "model.diffusion_model.double_blocks.2.img_mod.lin.bias",
1057
+ "model.diffusion_model.double_blocks.2.img_mod.lin.weight",
1058
+ "model.diffusion_model.double_blocks.2.txt_attn.norm.key_norm.scale",
1059
+ "model.diffusion_model.double_blocks.2.txt_attn.norm.query_norm.scale",
1060
+ "model.diffusion_model.double_blocks.2.txt_attn.proj.bias",
1061
+ "model.diffusion_model.double_blocks.2.txt_attn.proj.weight",
1062
+ "model.diffusion_model.double_blocks.2.txt_attn.qkv.bias",
1063
+ "model.diffusion_model.double_blocks.2.txt_attn.qkv.weight",
1064
+ "model.diffusion_model.double_blocks.2.txt_mlp.0.bias",
1065
+ "model.diffusion_model.double_blocks.2.txt_mlp.0.weight",
1066
+ "model.diffusion_model.double_blocks.2.txt_mlp.2.bias",
1067
+ "model.diffusion_model.double_blocks.2.txt_mlp.2.weight",
1068
+ "model.diffusion_model.double_blocks.2.txt_mod.lin.bias",
1069
+ "model.diffusion_model.double_blocks.2.txt_mod.lin.weight",
1070
+ "model.diffusion_model.double_blocks.3.img_attn.norm.key_norm.scale",
1071
+ "model.diffusion_model.double_blocks.3.img_attn.norm.query_norm.scale",
1072
+ "model.diffusion_model.double_blocks.3.img_attn.proj.bias",
1073
+ "model.diffusion_model.double_blocks.3.img_attn.proj.weight",
1074
+ "model.diffusion_model.double_blocks.3.img_attn.qkv.bias",
1075
+ "model.diffusion_model.double_blocks.3.img_attn.qkv.weight",
1076
+ "model.diffusion_model.double_blocks.3.img_mlp.0.bias",
1077
+ "model.diffusion_model.double_blocks.3.img_mlp.0.weight",
1078
+ "model.diffusion_model.double_blocks.3.img_mlp.2.bias",
1079
+ "model.diffusion_model.double_blocks.3.img_mlp.2.weight",
1080
+ "model.diffusion_model.double_blocks.3.img_mod.lin.bias",
1081
+ "model.diffusion_model.double_blocks.3.img_mod.lin.weight",
1082
+ "model.diffusion_model.double_blocks.3.txt_attn.norm.key_norm.scale",
1083
+ "model.diffusion_model.double_blocks.3.txt_attn.norm.query_norm.scale",
1084
+ "model.diffusion_model.double_blocks.3.txt_attn.proj.bias",
1085
+ "model.diffusion_model.double_blocks.3.txt_attn.proj.weight",
1086
+ "model.diffusion_model.double_blocks.3.txt_attn.qkv.bias",
1087
+ "model.diffusion_model.double_blocks.3.txt_attn.qkv.weight",
1088
+ "model.diffusion_model.double_blocks.3.txt_mlp.0.bias",
1089
+ "model.diffusion_model.double_blocks.3.txt_mlp.0.weight",
1090
+ "model.diffusion_model.double_blocks.3.txt_mlp.2.bias",
1091
+ "model.diffusion_model.double_blocks.3.txt_mlp.2.weight",
1092
+ "model.diffusion_model.double_blocks.3.txt_mod.lin.bias",
1093
+ "model.diffusion_model.double_blocks.3.txt_mod.lin.weight",
1094
+ "model.diffusion_model.double_blocks.4.img_attn.norm.key_norm.scale",
1095
+ "model.diffusion_model.double_blocks.4.img_attn.norm.query_norm.scale",
1096
+ "model.diffusion_model.double_blocks.4.img_attn.proj.bias",
1097
+ "model.diffusion_model.double_blocks.4.img_attn.proj.weight",
1098
+ "model.diffusion_model.double_blocks.4.img_attn.qkv.bias",
1099
+ "model.diffusion_model.double_blocks.4.img_attn.qkv.weight",
1100
+ "model.diffusion_model.double_blocks.4.img_mlp.0.bias",
1101
+ "model.diffusion_model.double_blocks.4.img_mlp.0.weight",
1102
+ "model.diffusion_model.double_blocks.4.img_mlp.2.bias",
1103
+ "model.diffusion_model.double_blocks.4.img_mlp.2.weight",
1104
+ "model.diffusion_model.double_blocks.4.img_mod.lin.bias",
1105
+ "model.diffusion_model.double_blocks.4.img_mod.lin.weight",
1106
+ "model.diffusion_model.double_blocks.4.txt_attn.norm.key_norm.scale",
1107
+ "model.diffusion_model.double_blocks.4.txt_attn.norm.query_norm.scale",
1108
+ "model.diffusion_model.double_blocks.4.txt_attn.proj.bias",
1109
+ "model.diffusion_model.double_blocks.4.txt_attn.proj.weight",
1110
+ "model.diffusion_model.double_blocks.4.txt_attn.qkv.bias",
1111
+ "model.diffusion_model.double_blocks.4.txt_attn.qkv.weight",
1112
+ "model.diffusion_model.double_blocks.4.txt_mlp.0.bias",
1113
+ "model.diffusion_model.double_blocks.4.txt_mlp.0.weight",
1114
+ "model.diffusion_model.double_blocks.4.txt_mlp.2.bias",
1115
+ "model.diffusion_model.double_blocks.4.txt_mlp.2.weight",
1116
+ "model.diffusion_model.double_blocks.4.txt_mod.lin.bias",
1117
+ "model.diffusion_model.double_blocks.4.txt_mod.lin.weight",
1118
+ "model.diffusion_model.double_blocks.5.img_attn.norm.key_norm.scale",
1119
+ "model.diffusion_model.double_blocks.5.img_attn.norm.query_norm.scale",
1120
+ "model.diffusion_model.double_blocks.5.img_attn.proj.bias",
1121
+ "model.diffusion_model.double_blocks.5.img_attn.proj.weight",
1122
+ "model.diffusion_model.double_blocks.5.img_attn.qkv.bias",
1123
+ "model.diffusion_model.double_blocks.5.img_attn.qkv.weight",
1124
+ "model.diffusion_model.double_blocks.5.img_mlp.0.bias",
1125
+ "model.diffusion_model.double_blocks.5.img_mlp.0.weight",
1126
+ "model.diffusion_model.double_blocks.5.img_mlp.2.bias",
1127
+ "model.diffusion_model.double_blocks.5.img_mlp.2.weight",
1128
+ "model.diffusion_model.double_blocks.5.img_mod.lin.bias",
1129
+ "model.diffusion_model.double_blocks.5.img_mod.lin.weight",
1130
+ "model.diffusion_model.double_blocks.5.txt_attn.norm.key_norm.scale",
1131
+ "model.diffusion_model.double_blocks.5.txt_attn.norm.query_norm.scale",
1132
+ "model.diffusion_model.double_blocks.5.txt_attn.proj.bias",
1133
+ "model.diffusion_model.double_blocks.5.txt_attn.proj.weight",
1134
+ "model.diffusion_model.double_blocks.5.txt_attn.qkv.bias",
1135
+ "model.diffusion_model.double_blocks.5.txt_attn.qkv.weight",
1136
+ "model.diffusion_model.double_blocks.5.txt_mlp.0.bias",
1137
+ "model.diffusion_model.double_blocks.5.txt_mlp.0.weight",
1138
+ "model.diffusion_model.double_blocks.5.txt_mlp.2.bias",
1139
+ "model.diffusion_model.double_blocks.5.txt_mlp.2.weight",
1140
+ "model.diffusion_model.double_blocks.5.txt_mod.lin.bias",
1141
+ "model.diffusion_model.double_blocks.5.txt_mod.lin.weight",
1142
+ "model.diffusion_model.double_blocks.6.img_attn.norm.key_norm.scale",
1143
+ "model.diffusion_model.double_blocks.6.img_attn.norm.query_norm.scale",
1144
+ "model.diffusion_model.double_blocks.6.img_attn.proj.bias",
1145
+ "model.diffusion_model.double_blocks.6.img_attn.proj.weight",
1146
+ "model.diffusion_model.double_blocks.6.img_attn.qkv.bias",
1147
+ "model.diffusion_model.double_blocks.6.img_attn.qkv.weight",
1148
+ "model.diffusion_model.double_blocks.6.img_mlp.0.bias",
1149
+ "model.diffusion_model.double_blocks.6.img_mlp.0.weight",
1150
+ "model.diffusion_model.double_blocks.6.img_mlp.2.bias",
1151
+ "model.diffusion_model.double_blocks.6.img_mlp.2.weight",
1152
+ "model.diffusion_model.double_blocks.6.img_mod.lin.bias",
1153
+ "model.diffusion_model.double_blocks.6.img_mod.lin.weight",
1154
+ "model.diffusion_model.double_blocks.6.txt_attn.norm.key_norm.scale",
1155
+ "model.diffusion_model.double_blocks.6.txt_attn.norm.query_norm.scale",
1156
+ "model.diffusion_model.double_blocks.6.txt_attn.proj.bias",
1157
+ "model.diffusion_model.double_blocks.6.txt_attn.proj.weight",
1158
+ "model.diffusion_model.double_blocks.6.txt_attn.qkv.bias",
1159
+ "model.diffusion_model.double_blocks.6.txt_attn.qkv.weight",
1160
+ "model.diffusion_model.double_blocks.6.txt_mlp.0.bias",
1161
+ "model.diffusion_model.double_blocks.6.txt_mlp.0.weight",
1162
+ "model.diffusion_model.double_blocks.6.txt_mlp.2.bias",
1163
+ "model.diffusion_model.double_blocks.6.txt_mlp.2.weight",
1164
+ "model.diffusion_model.double_blocks.6.txt_mod.lin.bias",
1165
+ "model.diffusion_model.double_blocks.6.txt_mod.lin.weight",
1166
+ "model.diffusion_model.double_blocks.7.img_attn.norm.key_norm.scale",
1167
+ "model.diffusion_model.double_blocks.7.img_attn.norm.query_norm.scale",
1168
+ "model.diffusion_model.double_blocks.7.img_attn.proj.bias",
1169
+ "model.diffusion_model.double_blocks.7.img_attn.proj.weight",
1170
+ "model.diffusion_model.double_blocks.7.img_attn.qkv.bias",
1171
+ "model.diffusion_model.double_blocks.7.img_attn.qkv.weight",
1172
+ "model.diffusion_model.double_blocks.7.img_mlp.0.bias",
1173
+ "model.diffusion_model.double_blocks.7.img_mlp.0.weight",
1174
+ "model.diffusion_model.double_blocks.7.img_mlp.2.bias",
1175
+ "model.diffusion_model.double_blocks.7.img_mlp.2.weight",
1176
+ "model.diffusion_model.double_blocks.7.img_mod.lin.bias",
1177
+ "model.diffusion_model.double_blocks.7.img_mod.lin.weight",
1178
+ "model.diffusion_model.double_blocks.7.txt_attn.norm.key_norm.scale",
1179
+ "model.diffusion_model.double_blocks.7.txt_attn.norm.query_norm.scale",
1180
+ "model.diffusion_model.double_blocks.7.txt_attn.proj.bias",
1181
+ "model.diffusion_model.double_blocks.7.txt_attn.proj.weight",
1182
+ "model.diffusion_model.double_blocks.7.txt_attn.qkv.bias",
1183
+ "model.diffusion_model.double_blocks.7.txt_attn.qkv.weight",
1184
+ "model.diffusion_model.double_blocks.7.txt_mlp.0.bias",
1185
+ "model.diffusion_model.double_blocks.7.txt_mlp.0.weight",
1186
+ "model.diffusion_model.double_blocks.7.txt_mlp.2.bias",
1187
+ "model.diffusion_model.double_blocks.7.txt_mlp.2.weight",
1188
+ "model.diffusion_model.double_blocks.7.txt_mod.lin.bias",
1189
+ "model.diffusion_model.double_blocks.7.txt_mod.lin.weight",
1190
+ "model.diffusion_model.double_blocks.8.img_attn.norm.key_norm.scale",
1191
+ "model.diffusion_model.double_blocks.8.img_attn.norm.query_norm.scale",
1192
+ "model.diffusion_model.double_blocks.8.img_attn.proj.bias",
1193
+ "model.diffusion_model.double_blocks.8.img_attn.proj.weight",
1194
+ "model.diffusion_model.double_blocks.8.img_attn.qkv.bias",
1195
+ "model.diffusion_model.double_blocks.8.img_attn.qkv.weight",
1196
+ "model.diffusion_model.double_blocks.8.img_mlp.0.bias",
1197
+ "model.diffusion_model.double_blocks.8.img_mlp.0.weight",
1198
+ "model.diffusion_model.double_blocks.8.img_mlp.2.bias",
1199
+ "model.diffusion_model.double_blocks.8.img_mlp.2.weight",
1200
+ "model.diffusion_model.double_blocks.8.img_mod.lin.bias",
1201
+ "model.diffusion_model.double_blocks.8.img_mod.lin.weight",
1202
+ "model.diffusion_model.double_blocks.8.txt_attn.norm.key_norm.scale",
1203
+ "model.diffusion_model.double_blocks.8.txt_attn.norm.query_norm.scale",
1204
+ "model.diffusion_model.double_blocks.8.txt_attn.proj.bias",
1205
+ "model.diffusion_model.double_blocks.8.txt_attn.proj.weight",
1206
+ "model.diffusion_model.double_blocks.8.txt_attn.qkv.bias",
1207
+ "model.diffusion_model.double_blocks.8.txt_attn.qkv.weight",
1208
+ "model.diffusion_model.double_blocks.8.txt_mlp.0.bias",
1209
+ "model.diffusion_model.double_blocks.8.txt_mlp.0.weight",
1210
+ "model.diffusion_model.double_blocks.8.txt_mlp.2.bias",
1211
+ "model.diffusion_model.double_blocks.8.txt_mlp.2.weight",
1212
+ "model.diffusion_model.double_blocks.8.txt_mod.lin.bias",
1213
+ "model.diffusion_model.double_blocks.8.txt_mod.lin.weight",
1214
+ "model.diffusion_model.double_blocks.9.img_attn.norm.key_norm.scale",
1215
+ "model.diffusion_model.double_blocks.9.img_attn.norm.query_norm.scale",
1216
+ "model.diffusion_model.double_blocks.9.img_attn.proj.bias",
1217
+ "model.diffusion_model.double_blocks.9.img_attn.proj.weight",
1218
+ "model.diffusion_model.double_blocks.9.img_attn.qkv.bias",
1219
+ "model.diffusion_model.double_blocks.9.img_attn.qkv.weight",
1220
+ "model.diffusion_model.double_blocks.9.img_mlp.0.bias",
1221
+ "model.diffusion_model.double_blocks.9.img_mlp.0.weight",
1222
+ "model.diffusion_model.double_blocks.9.img_mlp.2.bias",
1223
+ "model.diffusion_model.double_blocks.9.img_mlp.2.weight",
1224
+ "model.diffusion_model.double_blocks.9.img_mod.lin.bias",
1225
+ "model.diffusion_model.double_blocks.9.img_mod.lin.weight",
1226
+ "model.diffusion_model.double_blocks.9.txt_attn.norm.key_norm.scale",
1227
+ "model.diffusion_model.double_blocks.9.txt_attn.norm.query_norm.scale",
1228
+ "model.diffusion_model.double_blocks.9.txt_attn.proj.bias",
1229
+ "model.diffusion_model.double_blocks.9.txt_attn.proj.weight",
1230
+ "model.diffusion_model.double_blocks.9.txt_attn.qkv.bias",
1231
+ "model.diffusion_model.double_blocks.9.txt_attn.qkv.weight",
1232
+ "model.diffusion_model.double_blocks.9.txt_mlp.0.bias",
1233
+ "model.diffusion_model.double_blocks.9.txt_mlp.0.weight",
1234
+ "model.diffusion_model.double_blocks.9.txt_mlp.2.bias",
1235
+ "model.diffusion_model.double_blocks.9.txt_mlp.2.weight",
1236
+ "model.diffusion_model.double_blocks.9.txt_mod.lin.bias",
1237
+ "model.diffusion_model.double_blocks.9.txt_mod.lin.weight",
1238
+ "model.diffusion_model.final_layer.adaLN_modulation.1.bias",
1239
+ "model.diffusion_model.final_layer.adaLN_modulation.1.weight",
1240
+ "model.diffusion_model.final_layer.linear.bias",
1241
+ "model.diffusion_model.final_layer.linear.weight",
1242
+ "model.diffusion_model.guidance_in.in_layer.bias",
1243
+ "model.diffusion_model.guidance_in.in_layer.weight",
1244
+ "model.diffusion_model.guidance_in.out_layer.bias",
1245
+ "model.diffusion_model.guidance_in.out_layer.weight",
1246
+ "model.diffusion_model.img_in.bias",
1247
+ "model.diffusion_model.img_in.weight",
1248
+ "model.diffusion_model.single_blocks.0.linear1.bias",
1249
+ "model.diffusion_model.single_blocks.0.linear1.weight",
1250
+ "model.diffusion_model.single_blocks.0.linear2.bias",
1251
+ "model.diffusion_model.single_blocks.0.linear2.weight",
1252
+ "model.diffusion_model.single_blocks.0.modulation.lin.bias",
1253
+ "model.diffusion_model.single_blocks.0.modulation.lin.weight",
1254
+ "model.diffusion_model.single_blocks.0.norm.key_norm.scale",
1255
+ "model.diffusion_model.single_blocks.0.norm.query_norm.scale",
1256
+ "model.diffusion_model.single_blocks.1.linear1.bias",
1257
+ "model.diffusion_model.single_blocks.1.linear1.weight",
1258
+ "model.diffusion_model.single_blocks.1.linear2.bias",
1259
+ "model.diffusion_model.single_blocks.1.linear2.weight",
1260
+ "model.diffusion_model.single_blocks.1.modulation.lin.bias",
1261
+ "model.diffusion_model.single_blocks.1.modulation.lin.weight",
1262
+ "model.diffusion_model.single_blocks.1.norm.key_norm.scale",
1263
+ "model.diffusion_model.single_blocks.1.norm.query_norm.scale",
1264
+ "model.diffusion_model.single_blocks.10.linear1.bias",
1265
+ "model.diffusion_model.single_blocks.10.linear1.weight",
1266
+ "model.diffusion_model.single_blocks.10.linear2.bias",
1267
+ "model.diffusion_model.single_blocks.10.linear2.weight",
1268
+ "model.diffusion_model.single_blocks.10.modulation.lin.bias",
1269
+ "model.diffusion_model.single_blocks.10.modulation.lin.weight",
1270
+ "model.diffusion_model.single_blocks.10.norm.key_norm.scale",
1271
+ "model.diffusion_model.single_blocks.10.norm.query_norm.scale",
1272
+ "model.diffusion_model.single_blocks.11.linear1.bias",
1273
+ "model.diffusion_model.single_blocks.11.linear1.weight",
1274
+ "model.diffusion_model.single_blocks.11.linear2.bias",
1275
+ "model.diffusion_model.single_blocks.11.linear2.weight",
1276
+ "model.diffusion_model.single_blocks.11.modulation.lin.bias",
1277
+ "model.diffusion_model.single_blocks.11.modulation.lin.weight",
1278
+ "model.diffusion_model.single_blocks.11.norm.key_norm.scale",
1279
+ "model.diffusion_model.single_blocks.11.norm.query_norm.scale",
1280
+ "model.diffusion_model.single_blocks.12.linear1.bias",
1281
+ "model.diffusion_model.single_blocks.12.linear1.weight",
1282
+ "model.diffusion_model.single_blocks.12.linear2.bias",
1283
+ "model.diffusion_model.single_blocks.12.linear2.weight",
1284
+ "model.diffusion_model.single_blocks.12.modulation.lin.bias",
1285
+ "model.diffusion_model.single_blocks.12.modulation.lin.weight",
1286
+ "model.diffusion_model.single_blocks.12.norm.key_norm.scale",
1287
+ "model.diffusion_model.single_blocks.12.norm.query_norm.scale",
1288
+ "model.diffusion_model.single_blocks.13.linear1.bias",
1289
+ "model.diffusion_model.single_blocks.13.linear1.weight",
1290
+ "model.diffusion_model.single_blocks.13.linear2.bias",
1291
+ "model.diffusion_model.single_blocks.13.linear2.weight",
1292
+ "model.diffusion_model.single_blocks.13.modulation.lin.bias",
1293
+ "model.diffusion_model.single_blocks.13.modulation.lin.weight",
1294
+ "model.diffusion_model.single_blocks.13.norm.key_norm.scale",
1295
+ "model.diffusion_model.single_blocks.13.norm.query_norm.scale",
1296
+ "model.diffusion_model.single_blocks.14.linear1.bias",
1297
+ "model.diffusion_model.single_blocks.14.linear1.weight",
1298
+ "model.diffusion_model.single_blocks.14.linear2.bias",
1299
+ "model.diffusion_model.single_blocks.14.linear2.weight",
1300
+ "model.diffusion_model.single_blocks.14.modulation.lin.bias",
1301
+ "model.diffusion_model.single_blocks.14.modulation.lin.weight",
1302
+ "model.diffusion_model.single_blocks.14.norm.key_norm.scale",
1303
+ "model.diffusion_model.single_blocks.14.norm.query_norm.scale",
1304
+ "model.diffusion_model.single_blocks.15.linear1.bias",
1305
+ "model.diffusion_model.single_blocks.15.linear1.weight",
1306
+ "model.diffusion_model.single_blocks.15.linear2.bias",
1307
+ "model.diffusion_model.single_blocks.15.linear2.weight",
1308
+ "model.diffusion_model.single_blocks.15.modulation.lin.bias",
1309
+ "model.diffusion_model.single_blocks.15.modulation.lin.weight",
1310
+ "model.diffusion_model.single_blocks.15.norm.key_norm.scale",
1311
+ "model.diffusion_model.single_blocks.15.norm.query_norm.scale",
1312
+ "model.diffusion_model.single_blocks.16.linear1.bias",
1313
+ "model.diffusion_model.single_blocks.16.linear1.weight",
1314
+ "model.diffusion_model.single_blocks.16.linear2.bias",
1315
+ "model.diffusion_model.single_blocks.16.linear2.weight",
1316
+ "model.diffusion_model.single_blocks.16.modulation.lin.bias",
1317
+ "model.diffusion_model.single_blocks.16.modulation.lin.weight",
1318
+ "model.diffusion_model.single_blocks.16.norm.key_norm.scale",
1319
+ "model.diffusion_model.single_blocks.16.norm.query_norm.scale",
1320
+ "model.diffusion_model.single_blocks.17.linear1.bias",
1321
+ "model.diffusion_model.single_blocks.17.linear1.weight",
1322
+ "model.diffusion_model.single_blocks.17.linear2.bias",
1323
+ "model.diffusion_model.single_blocks.17.linear2.weight",
1324
+ "model.diffusion_model.single_blocks.17.modulation.lin.bias",
1325
+ "model.diffusion_model.single_blocks.17.modulation.lin.weight",
1326
+ "model.diffusion_model.single_blocks.17.norm.key_norm.scale",
1327
+ "model.diffusion_model.single_blocks.17.norm.query_norm.scale",
1328
+ "model.diffusion_model.single_blocks.18.linear1.bias",
1329
+ "model.diffusion_model.single_blocks.18.linear1.weight",
1330
+ "model.diffusion_model.single_blocks.18.linear2.bias",
1331
+ "model.diffusion_model.single_blocks.18.linear2.weight",
1332
+ "model.diffusion_model.single_blocks.18.modulation.lin.bias",
1333
+ "model.diffusion_model.single_blocks.18.modulation.lin.weight",
1334
+ "model.diffusion_model.single_blocks.18.norm.key_norm.scale",
1335
+ "model.diffusion_model.single_blocks.18.norm.query_norm.scale",
1336
+ "model.diffusion_model.single_blocks.19.linear1.bias",
1337
+ "model.diffusion_model.single_blocks.19.linear1.weight",
1338
+ "model.diffusion_model.single_blocks.19.linear2.bias",
1339
+ "model.diffusion_model.single_blocks.19.linear2.weight",
1340
+ "model.diffusion_model.single_blocks.19.modulation.lin.bias",
1341
+ "model.diffusion_model.single_blocks.19.modulation.lin.weight",
1342
+ "model.diffusion_model.single_blocks.19.norm.key_norm.scale",
1343
+ "model.diffusion_model.single_blocks.19.norm.query_norm.scale",
1344
+ "model.diffusion_model.single_blocks.2.linear1.bias",
1345
+ "model.diffusion_model.single_blocks.2.linear1.weight",
1346
+ "model.diffusion_model.single_blocks.2.linear2.bias",
1347
+ "model.diffusion_model.single_blocks.2.linear2.weight",
1348
+ "model.diffusion_model.single_blocks.2.modulation.lin.bias",
1349
+ "model.diffusion_model.single_blocks.2.modulation.lin.weight",
1350
+ "model.diffusion_model.single_blocks.2.norm.key_norm.scale",
1351
+ "model.diffusion_model.single_blocks.2.norm.query_norm.scale",
1352
+ "model.diffusion_model.single_blocks.20.linear1.bias",
1353
+ "model.diffusion_model.single_blocks.20.linear1.weight",
1354
+ "model.diffusion_model.single_blocks.20.linear2.bias",
1355
+ "model.diffusion_model.single_blocks.20.linear2.weight",
1356
+ "model.diffusion_model.single_blocks.20.modulation.lin.bias",
1357
+ "model.diffusion_model.single_blocks.20.modulation.lin.weight",
1358
+ "model.diffusion_model.single_blocks.20.norm.key_norm.scale",
1359
+ "model.diffusion_model.single_blocks.20.norm.query_norm.scale",
1360
+ "model.diffusion_model.single_blocks.21.linear1.bias",
1361
+ "model.diffusion_model.single_blocks.21.linear1.weight",
1362
+ "model.diffusion_model.single_blocks.21.linear2.bias",
1363
+ "model.diffusion_model.single_blocks.21.linear2.weight",
1364
+ "model.diffusion_model.single_blocks.21.modulation.lin.bias",
1365
+ "model.diffusion_model.single_blocks.21.modulation.lin.weight",
1366
+ "model.diffusion_model.single_blocks.21.norm.key_norm.scale",
1367
+ "model.diffusion_model.single_blocks.21.norm.query_norm.scale",
1368
+ "model.diffusion_model.single_blocks.22.linear1.bias",
1369
+ "model.diffusion_model.single_blocks.22.linear1.weight",
1370
+ "model.diffusion_model.single_blocks.22.linear2.bias",
1371
+ "model.diffusion_model.single_blocks.22.linear2.weight",
1372
+ "model.diffusion_model.single_blocks.22.modulation.lin.bias",
1373
+ "model.diffusion_model.single_blocks.22.modulation.lin.weight",
1374
+ "model.diffusion_model.single_blocks.22.norm.key_norm.scale",
1375
+ "model.diffusion_model.single_blocks.22.norm.query_norm.scale",
1376
+ "model.diffusion_model.single_blocks.23.linear1.bias",
1377
+ "model.diffusion_model.single_blocks.23.linear1.weight",
1378
+ "model.diffusion_model.single_blocks.23.linear2.bias",
1379
+ "model.diffusion_model.single_blocks.23.linear2.weight",
1380
+ "model.diffusion_model.single_blocks.23.modulation.lin.bias",
1381
+ "model.diffusion_model.single_blocks.23.modulation.lin.weight",
1382
+ "model.diffusion_model.single_blocks.23.norm.key_norm.scale",
1383
+ "model.diffusion_model.single_blocks.23.norm.query_norm.scale",
1384
+ "model.diffusion_model.single_blocks.24.linear1.bias",
1385
+ "model.diffusion_model.single_blocks.24.linear1.weight",
1386
+ "model.diffusion_model.single_blocks.24.linear2.bias",
1387
+ "model.diffusion_model.single_blocks.24.linear2.weight",
1388
+ "model.diffusion_model.single_blocks.24.modulation.lin.bias",
1389
+ "model.diffusion_model.single_blocks.24.modulation.lin.weight",
1390
+ "model.diffusion_model.single_blocks.24.norm.key_norm.scale",
1391
+ "model.diffusion_model.single_blocks.24.norm.query_norm.scale",
1392
+ "model.diffusion_model.single_blocks.25.linear1.bias",
1393
+ "model.diffusion_model.single_blocks.25.linear1.weight",
1394
+ "model.diffusion_model.single_blocks.25.linear2.bias",
1395
+ "model.diffusion_model.single_blocks.25.linear2.weight",
1396
+ "model.diffusion_model.single_blocks.25.modulation.lin.bias",
1397
+ "model.diffusion_model.single_blocks.25.modulation.lin.weight",
1398
+ "model.diffusion_model.single_blocks.25.norm.key_norm.scale",
1399
+ "model.diffusion_model.single_blocks.25.norm.query_norm.scale",
1400
+ "model.diffusion_model.single_blocks.26.linear1.bias",
1401
+ "model.diffusion_model.single_blocks.26.linear1.weight",
1402
+ "model.diffusion_model.single_blocks.26.linear2.bias",
1403
+ "model.diffusion_model.single_blocks.26.linear2.weight",
1404
+ "model.diffusion_model.single_blocks.26.modulation.lin.bias",
1405
+ "model.diffusion_model.single_blocks.26.modulation.lin.weight",
1406
+ "model.diffusion_model.single_blocks.26.norm.key_norm.scale",
1407
+ "model.diffusion_model.single_blocks.26.norm.query_norm.scale",
1408
+ "model.diffusion_model.single_blocks.27.linear1.bias",
1409
+ "model.diffusion_model.single_blocks.27.linear1.weight",
1410
+ "model.diffusion_model.single_blocks.27.linear2.bias",
1411
+ "model.diffusion_model.single_blocks.27.linear2.weight",
1412
+ "model.diffusion_model.single_blocks.27.modulation.lin.bias",
1413
+ "model.diffusion_model.single_blocks.27.modulation.lin.weight",
1414
+ "model.diffusion_model.single_blocks.27.norm.key_norm.scale",
1415
+ "model.diffusion_model.single_blocks.27.norm.query_norm.scale",
1416
+ "model.diffusion_model.single_blocks.28.linear1.bias",
1417
+ "model.diffusion_model.single_blocks.28.linear1.weight",
1418
+ "model.diffusion_model.single_blocks.28.linear2.bias",
1419
+ "model.diffusion_model.single_blocks.28.linear2.weight",
1420
+ "model.diffusion_model.single_blocks.28.modulation.lin.bias",
1421
+ "model.diffusion_model.single_blocks.28.modulation.lin.weight",
1422
+ "model.diffusion_model.single_blocks.28.norm.key_norm.scale",
1423
+ "model.diffusion_model.single_blocks.28.norm.query_norm.scale",
1424
+ "model.diffusion_model.single_blocks.29.linear1.bias",
1425
+ "model.diffusion_model.single_blocks.29.linear1.weight",
1426
+ "model.diffusion_model.single_blocks.29.linear2.bias",
1427
+ "model.diffusion_model.single_blocks.29.linear2.weight",
1428
+ "model.diffusion_model.single_blocks.29.modulation.lin.bias",
1429
+ "model.diffusion_model.single_blocks.29.modulation.lin.weight",
1430
+ "model.diffusion_model.single_blocks.29.norm.key_norm.scale",
1431
+ "model.diffusion_model.single_blocks.29.norm.query_norm.scale",
1432
+ "model.diffusion_model.single_blocks.3.linear1.bias",
1433
+ "model.diffusion_model.single_blocks.3.linear1.weight",
1434
+ "model.diffusion_model.single_blocks.3.linear2.bias",
1435
+ "model.diffusion_model.single_blocks.3.linear2.weight",
1436
+ "model.diffusion_model.single_blocks.3.modulation.lin.bias",
1437
+ "model.diffusion_model.single_blocks.3.modulation.lin.weight",
1438
+ "model.diffusion_model.single_blocks.3.norm.key_norm.scale",
1439
+ "model.diffusion_model.single_blocks.3.norm.query_norm.scale",
1440
+ "model.diffusion_model.single_blocks.30.linear1.bias",
1441
+ "model.diffusion_model.single_blocks.30.linear1.weight",
1442
+ "model.diffusion_model.single_blocks.30.linear2.bias",
1443
+ "model.diffusion_model.single_blocks.30.linear2.weight",
1444
+ "model.diffusion_model.single_blocks.30.modulation.lin.bias",
1445
+ "model.diffusion_model.single_blocks.30.modulation.lin.weight",
1446
+ "model.diffusion_model.single_blocks.30.norm.key_norm.scale",
1447
+ "model.diffusion_model.single_blocks.30.norm.query_norm.scale",
1448
+ "model.diffusion_model.single_blocks.31.linear1.bias",
1449
+ "model.diffusion_model.single_blocks.31.linear1.weight",
1450
+ "model.diffusion_model.single_blocks.31.linear2.bias",
1451
+ "model.diffusion_model.single_blocks.31.linear2.weight",
1452
+ "model.diffusion_model.single_blocks.31.modulation.lin.bias",
1453
+ "model.diffusion_model.single_blocks.31.modulation.lin.weight",
1454
+ "model.diffusion_model.single_blocks.31.norm.key_norm.scale",
1455
+ "model.diffusion_model.single_blocks.31.norm.query_norm.scale",
1456
+ "model.diffusion_model.single_blocks.32.linear1.bias",
1457
+ "model.diffusion_model.single_blocks.32.linear1.weight",
1458
+ "model.diffusion_model.single_blocks.32.linear2.bias",
1459
+ "model.diffusion_model.single_blocks.32.linear2.weight",
1460
+ "model.diffusion_model.single_blocks.32.modulation.lin.bias",
1461
+ "model.diffusion_model.single_blocks.32.modulation.lin.weight",
1462
+ "model.diffusion_model.single_blocks.32.norm.key_norm.scale",
1463
+ "model.diffusion_model.single_blocks.32.norm.query_norm.scale",
1464
+ "model.diffusion_model.single_blocks.33.linear1.bias",
1465
+ "model.diffusion_model.single_blocks.33.linear1.weight",
1466
+ "model.diffusion_model.single_blocks.33.linear2.bias",
1467
+ "model.diffusion_model.single_blocks.33.linear2.weight",
1468
+ "model.diffusion_model.single_blocks.33.modulation.lin.bias",
1469
+ "model.diffusion_model.single_blocks.33.modulation.lin.weight",
1470
+ "model.diffusion_model.single_blocks.33.norm.key_norm.scale",
1471
+ "model.diffusion_model.single_blocks.33.norm.query_norm.scale",
1472
+ "model.diffusion_model.single_blocks.34.linear1.bias",
1473
+ "model.diffusion_model.single_blocks.34.linear1.weight",
1474
+ "model.diffusion_model.single_blocks.34.linear2.bias",
1475
+ "model.diffusion_model.single_blocks.34.linear2.weight",
1476
+ "model.diffusion_model.single_blocks.34.modulation.lin.bias",
1477
+ "model.diffusion_model.single_blocks.34.modulation.lin.weight",
1478
+ "model.diffusion_model.single_blocks.34.norm.key_norm.scale",
1479
+ "model.diffusion_model.single_blocks.34.norm.query_norm.scale",
1480
+ "model.diffusion_model.single_blocks.35.linear1.bias",
1481
+ "model.diffusion_model.single_blocks.35.linear1.weight",
1482
+ "model.diffusion_model.single_blocks.35.linear2.bias",
1483
+ "model.diffusion_model.single_blocks.35.linear2.weight",
1484
+ "model.diffusion_model.single_blocks.35.modulation.lin.bias",
1485
+ "model.diffusion_model.single_blocks.35.modulation.lin.weight",
1486
+ "model.diffusion_model.single_blocks.35.norm.key_norm.scale",
1487
+ "model.diffusion_model.single_blocks.35.norm.query_norm.scale",
1488
+ "model.diffusion_model.single_blocks.36.linear1.bias",
1489
+ "model.diffusion_model.single_blocks.36.linear1.weight",
1490
+ "model.diffusion_model.single_blocks.36.linear2.bias",
1491
+ "model.diffusion_model.single_blocks.36.linear2.weight",
1492
+ "model.diffusion_model.single_blocks.36.modulation.lin.bias",
1493
+ "model.diffusion_model.single_blocks.36.modulation.lin.weight",
1494
+ "model.diffusion_model.single_blocks.36.norm.key_norm.scale",
1495
+ "model.diffusion_model.single_blocks.36.norm.query_norm.scale",
1496
+ "model.diffusion_model.single_blocks.37.linear1.bias",
1497
+ "model.diffusion_model.single_blocks.37.linear1.weight",
1498
+ "model.diffusion_model.single_blocks.37.linear2.bias",
1499
+ "model.diffusion_model.single_blocks.37.linear2.weight",
1500
+ "model.diffusion_model.single_blocks.37.modulation.lin.bias",
1501
+ "model.diffusion_model.single_blocks.37.modulation.lin.weight",
1502
+ "model.diffusion_model.single_blocks.37.norm.key_norm.scale",
1503
+ "model.diffusion_model.single_blocks.37.norm.query_norm.scale",
1504
+ "model.diffusion_model.single_blocks.4.linear1.bias",
1505
+ "model.diffusion_model.single_blocks.4.linear1.weight",
1506
+ "model.diffusion_model.single_blocks.4.linear2.bias",
1507
+ "model.diffusion_model.single_blocks.4.linear2.weight",
1508
+ "model.diffusion_model.single_blocks.4.modulation.lin.bias",
1509
+ "model.diffusion_model.single_blocks.4.modulation.lin.weight",
1510
+ "model.diffusion_model.single_blocks.4.norm.key_norm.scale",
1511
+ "model.diffusion_model.single_blocks.4.norm.query_norm.scale",
1512
+ "model.diffusion_model.single_blocks.5.linear1.bias",
1513
+ "model.diffusion_model.single_blocks.5.linear1.weight",
1514
+ "model.diffusion_model.single_blocks.5.linear2.bias",
1515
+ "model.diffusion_model.single_blocks.5.linear2.weight",
1516
+ "model.diffusion_model.single_blocks.5.modulation.lin.bias",
1517
+ "model.diffusion_model.single_blocks.5.modulation.lin.weight",
1518
+ "model.diffusion_model.single_blocks.5.norm.key_norm.scale",
1519
+ "model.diffusion_model.single_blocks.5.norm.query_norm.scale",
1520
+ "model.diffusion_model.single_blocks.6.linear1.bias",
1521
+ "model.diffusion_model.single_blocks.6.linear1.weight",
1522
+ "model.diffusion_model.single_blocks.6.linear2.bias",
1523
+ "model.diffusion_model.single_blocks.6.linear2.weight",
1524
+ "model.diffusion_model.single_blocks.6.modulation.lin.bias",
1525
+ "model.diffusion_model.single_blocks.6.modulation.lin.weight",
1526
+ "model.diffusion_model.single_blocks.6.norm.key_norm.scale",
1527
+ "model.diffusion_model.single_blocks.6.norm.query_norm.scale",
1528
+ "model.diffusion_model.single_blocks.7.linear1.bias",
1529
+ "model.diffusion_model.single_blocks.7.linear1.weight",
1530
+ "model.diffusion_model.single_blocks.7.linear2.bias",
1531
+ "model.diffusion_model.single_blocks.7.linear2.weight",
1532
+ "model.diffusion_model.single_blocks.7.modulation.lin.bias",
1533
+ "model.diffusion_model.single_blocks.7.modulation.lin.weight",
1534
+ "model.diffusion_model.single_blocks.7.norm.key_norm.scale",
1535
+ "model.diffusion_model.single_blocks.7.norm.query_norm.scale",
1536
+ "model.diffusion_model.single_blocks.8.linear1.bias",
1537
+ "model.diffusion_model.single_blocks.8.linear1.weight",
1538
+ "model.diffusion_model.single_blocks.8.linear2.bias",
1539
+ "model.diffusion_model.single_blocks.8.linear2.weight",
1540
+ "model.diffusion_model.single_blocks.8.modulation.lin.bias",
1541
+ "model.diffusion_model.single_blocks.8.modulation.lin.weight",
1542
+ "model.diffusion_model.single_blocks.8.norm.key_norm.scale",
1543
+ "model.diffusion_model.single_blocks.8.norm.query_norm.scale",
1544
+ "model.diffusion_model.single_blocks.9.linear1.bias",
1545
+ "model.diffusion_model.single_blocks.9.linear1.weight",
1546
+ "model.diffusion_model.single_blocks.9.linear2.bias",
1547
+ "model.diffusion_model.single_blocks.9.linear2.weight",
1548
+ "model.diffusion_model.single_blocks.9.modulation.lin.bias",
1549
+ "model.diffusion_model.single_blocks.9.modulation.lin.weight",
1550
+ "model.diffusion_model.single_blocks.9.norm.key_norm.scale",
1551
+ "model.diffusion_model.single_blocks.9.norm.query_norm.scale",
1552
+ "model.diffusion_model.time_in.in_layer.bias",
1553
+ "model.diffusion_model.time_in.in_layer.weight",
1554
+ "model.diffusion_model.time_in.out_layer.bias",
1555
+ "model.diffusion_model.time_in.out_layer.weight",
1556
+ "model.diffusion_model.txt_in.bias",
1557
+ "model.diffusion_model.txt_in.weight",
1558
+ "model.diffusion_model.vector_in.in_layer.bias",
1559
+ "model.diffusion_model.vector_in.in_layer.weight",
1560
+ "model.diffusion_model.vector_in.out_layer.bias",
1561
+ "model.diffusion_model.vector_in.out_layer.weight"
1562
+ ]
flux_vae_keys.json ADDED
@@ -0,0 +1,490 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ "decoder.conv_in.bias",
3
+ "decoder.conv_in.weight",
4
+ "decoder.conv_out.bias",
5
+ "decoder.conv_out.weight",
6
+ "decoder.mid.attn_1.k.bias",
7
+ "decoder.mid.attn_1.k.weight",
8
+ "decoder.mid.attn_1.norm.bias",
9
+ "decoder.mid.attn_1.norm.weight",
10
+ "decoder.mid.attn_1.proj_out.bias",
11
+ "decoder.mid.attn_1.proj_out.weight",
12
+ "decoder.mid.attn_1.q.bias",
13
+ "decoder.mid.attn_1.q.weight",
14
+ "decoder.mid.attn_1.v.bias",
15
+ "decoder.mid.attn_1.v.weight",
16
+ "decoder.mid.block_1.conv1.bias",
17
+ "decoder.mid.block_1.conv1.weight",
18
+ "decoder.mid.block_1.conv2.bias",
19
+ "decoder.mid.block_1.conv2.weight",
20
+ "decoder.mid.block_1.norm1.bias",
21
+ "decoder.mid.block_1.norm1.weight",
22
+ "decoder.mid.block_1.norm2.bias",
23
+ "decoder.mid.block_1.norm2.weight",
24
+ "decoder.mid.block_2.conv1.bias",
25
+ "decoder.mid.block_2.conv1.weight",
26
+ "decoder.mid.block_2.conv2.bias",
27
+ "decoder.mid.block_2.conv2.weight",
28
+ "decoder.mid.block_2.norm1.bias",
29
+ "decoder.mid.block_2.norm1.weight",
30
+ "decoder.mid.block_2.norm2.bias",
31
+ "decoder.mid.block_2.norm2.weight",
32
+ "decoder.norm_out.bias",
33
+ "decoder.norm_out.weight",
34
+ "decoder.up.0.block.0.conv1.bias",
35
+ "decoder.up.0.block.0.conv1.weight",
36
+ "decoder.up.0.block.0.conv2.bias",
37
+ "decoder.up.0.block.0.conv2.weight",
38
+ "decoder.up.0.block.0.nin_shortcut.bias",
39
+ "decoder.up.0.block.0.nin_shortcut.weight",
40
+ "decoder.up.0.block.0.norm1.bias",
41
+ "decoder.up.0.block.0.norm1.weight",
42
+ "decoder.up.0.block.0.norm2.bias",
43
+ "decoder.up.0.block.0.norm2.weight",
44
+ "decoder.up.0.block.1.conv1.bias",
45
+ "decoder.up.0.block.1.conv1.weight",
46
+ "decoder.up.0.block.1.conv2.bias",
47
+ "decoder.up.0.block.1.conv2.weight",
48
+ "decoder.up.0.block.1.norm1.bias",
49
+ "decoder.up.0.block.1.norm1.weight",
50
+ "decoder.up.0.block.1.norm2.bias",
51
+ "decoder.up.0.block.1.norm2.weight",
52
+ "decoder.up.0.block.2.conv1.bias",
53
+ "decoder.up.0.block.2.conv1.weight",
54
+ "decoder.up.0.block.2.conv2.bias",
55
+ "decoder.up.0.block.2.conv2.weight",
56
+ "decoder.up.0.block.2.norm1.bias",
57
+ "decoder.up.0.block.2.norm1.weight",
58
+ "decoder.up.0.block.2.norm2.bias",
59
+ "decoder.up.0.block.2.norm2.weight",
60
+ "decoder.up.1.block.0.conv1.bias",
61
+ "decoder.up.1.block.0.conv1.weight",
62
+ "decoder.up.1.block.0.conv2.bias",
63
+ "decoder.up.1.block.0.conv2.weight",
64
+ "decoder.up.1.block.0.nin_shortcut.bias",
65
+ "decoder.up.1.block.0.nin_shortcut.weight",
66
+ "decoder.up.1.block.0.norm1.bias",
67
+ "decoder.up.1.block.0.norm1.weight",
68
+ "decoder.up.1.block.0.norm2.bias",
69
+ "decoder.up.1.block.0.norm2.weight",
70
+ "decoder.up.1.block.1.conv1.bias",
71
+ "decoder.up.1.block.1.conv1.weight",
72
+ "decoder.up.1.block.1.conv2.bias",
73
+ "decoder.up.1.block.1.conv2.weight",
74
+ "decoder.up.1.block.1.norm1.bias",
75
+ "decoder.up.1.block.1.norm1.weight",
76
+ "decoder.up.1.block.1.norm2.bias",
77
+ "decoder.up.1.block.1.norm2.weight",
78
+ "decoder.up.1.block.2.conv1.bias",
79
+ "decoder.up.1.block.2.conv1.weight",
80
+ "decoder.up.1.block.2.conv2.bias",
81
+ "decoder.up.1.block.2.conv2.weight",
82
+ "decoder.up.1.block.2.norm1.bias",
83
+ "decoder.up.1.block.2.norm1.weight",
84
+ "decoder.up.1.block.2.norm2.bias",
85
+ "decoder.up.1.block.2.norm2.weight",
86
+ "decoder.up.1.upsample.conv.bias",
87
+ "decoder.up.1.upsample.conv.weight",
88
+ "decoder.up.2.block.0.conv1.bias",
89
+ "decoder.up.2.block.0.conv1.weight",
90
+ "decoder.up.2.block.0.conv2.bias",
91
+ "decoder.up.2.block.0.conv2.weight",
92
+ "decoder.up.2.block.0.norm1.bias",
93
+ "decoder.up.2.block.0.norm1.weight",
94
+ "decoder.up.2.block.0.norm2.bias",
95
+ "decoder.up.2.block.0.norm2.weight",
96
+ "decoder.up.2.block.1.conv1.bias",
97
+ "decoder.up.2.block.1.conv1.weight",
98
+ "decoder.up.2.block.1.conv2.bias",
99
+ "decoder.up.2.block.1.conv2.weight",
100
+ "decoder.up.2.block.1.norm1.bias",
101
+ "decoder.up.2.block.1.norm1.weight",
102
+ "decoder.up.2.block.1.norm2.bias",
103
+ "decoder.up.2.block.1.norm2.weight",
104
+ "decoder.up.2.block.2.conv1.bias",
105
+ "decoder.up.2.block.2.conv1.weight",
106
+ "decoder.up.2.block.2.conv2.bias",
107
+ "decoder.up.2.block.2.conv2.weight",
108
+ "decoder.up.2.block.2.norm1.bias",
109
+ "decoder.up.2.block.2.norm1.weight",
110
+ "decoder.up.2.block.2.norm2.bias",
111
+ "decoder.up.2.block.2.norm2.weight",
112
+ "decoder.up.2.upsample.conv.bias",
113
+ "decoder.up.2.upsample.conv.weight",
114
+ "decoder.up.3.block.0.conv1.bias",
115
+ "decoder.up.3.block.0.conv1.weight",
116
+ "decoder.up.3.block.0.conv2.bias",
117
+ "decoder.up.3.block.0.conv2.weight",
118
+ "decoder.up.3.block.0.norm1.bias",
119
+ "decoder.up.3.block.0.norm1.weight",
120
+ "decoder.up.3.block.0.norm2.bias",
121
+ "decoder.up.3.block.0.norm2.weight",
122
+ "decoder.up.3.block.1.conv1.bias",
123
+ "decoder.up.3.block.1.conv1.weight",
124
+ "decoder.up.3.block.1.conv2.bias",
125
+ "decoder.up.3.block.1.conv2.weight",
126
+ "decoder.up.3.block.1.norm1.bias",
127
+ "decoder.up.3.block.1.norm1.weight",
128
+ "decoder.up.3.block.1.norm2.bias",
129
+ "decoder.up.3.block.1.norm2.weight",
130
+ "decoder.up.3.block.2.conv1.bias",
131
+ "decoder.up.3.block.2.conv1.weight",
132
+ "decoder.up.3.block.2.conv2.bias",
133
+ "decoder.up.3.block.2.conv2.weight",
134
+ "decoder.up.3.block.2.norm1.bias",
135
+ "decoder.up.3.block.2.norm1.weight",
136
+ "decoder.up.3.block.2.norm2.bias",
137
+ "decoder.up.3.block.2.norm2.weight",
138
+ "decoder.up.3.upsample.conv.bias",
139
+ "decoder.up.3.upsample.conv.weight",
140
+ "encoder.conv_in.bias",
141
+ "encoder.conv_in.weight",
142
+ "encoder.conv_out.bias",
143
+ "encoder.conv_out.weight",
144
+ "encoder.down.0.block.0.conv1.bias",
145
+ "encoder.down.0.block.0.conv1.weight",
146
+ "encoder.down.0.block.0.conv2.bias",
147
+ "encoder.down.0.block.0.conv2.weight",
148
+ "encoder.down.0.block.0.norm1.bias",
149
+ "encoder.down.0.block.0.norm1.weight",
150
+ "encoder.down.0.block.0.norm2.bias",
151
+ "encoder.down.0.block.0.norm2.weight",
152
+ "encoder.down.0.block.1.conv1.bias",
153
+ "encoder.down.0.block.1.conv1.weight",
154
+ "encoder.down.0.block.1.conv2.bias",
155
+ "encoder.down.0.block.1.conv2.weight",
156
+ "encoder.down.0.block.1.norm1.bias",
157
+ "encoder.down.0.block.1.norm1.weight",
158
+ "encoder.down.0.block.1.norm2.bias",
159
+ "encoder.down.0.block.1.norm2.weight",
160
+ "encoder.down.0.downsample.conv.bias",
161
+ "encoder.down.0.downsample.conv.weight",
162
+ "encoder.down.1.block.0.conv1.bias",
163
+ "encoder.down.1.block.0.conv1.weight",
164
+ "encoder.down.1.block.0.conv2.bias",
165
+ "encoder.down.1.block.0.conv2.weight",
166
+ "encoder.down.1.block.0.nin_shortcut.bias",
167
+ "encoder.down.1.block.0.nin_shortcut.weight",
168
+ "encoder.down.1.block.0.norm1.bias",
169
+ "encoder.down.1.block.0.norm1.weight",
170
+ "encoder.down.1.block.0.norm2.bias",
171
+ "encoder.down.1.block.0.norm2.weight",
172
+ "encoder.down.1.block.1.conv1.bias",
173
+ "encoder.down.1.block.1.conv1.weight",
174
+ "encoder.down.1.block.1.conv2.bias",
175
+ "encoder.down.1.block.1.conv2.weight",
176
+ "encoder.down.1.block.1.norm1.bias",
177
+ "encoder.down.1.block.1.norm1.weight",
178
+ "encoder.down.1.block.1.norm2.bias",
179
+ "encoder.down.1.block.1.norm2.weight",
180
+ "encoder.down.1.downsample.conv.bias",
181
+ "encoder.down.1.downsample.conv.weight",
182
+ "encoder.down.2.block.0.conv1.bias",
183
+ "encoder.down.2.block.0.conv1.weight",
184
+ "encoder.down.2.block.0.conv2.bias",
185
+ "encoder.down.2.block.0.conv2.weight",
186
+ "encoder.down.2.block.0.nin_shortcut.bias",
187
+ "encoder.down.2.block.0.nin_shortcut.weight",
188
+ "encoder.down.2.block.0.norm1.bias",
189
+ "encoder.down.2.block.0.norm1.weight",
190
+ "encoder.down.2.block.0.norm2.bias",
191
+ "encoder.down.2.block.0.norm2.weight",
192
+ "encoder.down.2.block.1.conv1.bias",
193
+ "encoder.down.2.block.1.conv1.weight",
194
+ "encoder.down.2.block.1.conv2.bias",
195
+ "encoder.down.2.block.1.conv2.weight",
196
+ "encoder.down.2.block.1.norm1.bias",
197
+ "encoder.down.2.block.1.norm1.weight",
198
+ "encoder.down.2.block.1.norm2.bias",
199
+ "encoder.down.2.block.1.norm2.weight",
200
+ "encoder.down.2.downsample.conv.bias",
201
+ "encoder.down.2.downsample.conv.weight",
202
+ "encoder.down.3.block.0.conv1.bias",
203
+ "encoder.down.3.block.0.conv1.weight",
204
+ "encoder.down.3.block.0.conv2.bias",
205
+ "encoder.down.3.block.0.conv2.weight",
206
+ "encoder.down.3.block.0.norm1.bias",
207
+ "encoder.down.3.block.0.norm1.weight",
208
+ "encoder.down.3.block.0.norm2.bias",
209
+ "encoder.down.3.block.0.norm2.weight",
210
+ "encoder.down.3.block.1.conv1.bias",
211
+ "encoder.down.3.block.1.conv1.weight",
212
+ "encoder.down.3.block.1.conv2.bias",
213
+ "encoder.down.3.block.1.conv2.weight",
214
+ "encoder.down.3.block.1.norm1.bias",
215
+ "encoder.down.3.block.1.norm1.weight",
216
+ "encoder.down.3.block.1.norm2.bias",
217
+ "encoder.down.3.block.1.norm2.weight",
218
+ "encoder.mid.attn_1.k.bias",
219
+ "encoder.mid.attn_1.k.weight",
220
+ "encoder.mid.attn_1.norm.bias",
221
+ "encoder.mid.attn_1.norm.weight",
222
+ "encoder.mid.attn_1.proj_out.bias",
223
+ "encoder.mid.attn_1.proj_out.weight",
224
+ "encoder.mid.attn_1.q.bias",
225
+ "encoder.mid.attn_1.q.weight",
226
+ "encoder.mid.attn_1.v.bias",
227
+ "encoder.mid.attn_1.v.weight",
228
+ "encoder.mid.block_1.conv1.bias",
229
+ "encoder.mid.block_1.conv1.weight",
230
+ "encoder.mid.block_1.conv2.bias",
231
+ "encoder.mid.block_1.conv2.weight",
232
+ "encoder.mid.block_1.norm1.bias",
233
+ "encoder.mid.block_1.norm1.weight",
234
+ "encoder.mid.block_1.norm2.bias",
235
+ "encoder.mid.block_1.norm2.weight",
236
+ "encoder.mid.block_2.conv1.bias",
237
+ "encoder.mid.block_2.conv1.weight",
238
+ "encoder.mid.block_2.conv2.bias",
239
+ "encoder.mid.block_2.conv2.weight",
240
+ "encoder.mid.block_2.norm1.bias",
241
+ "encoder.mid.block_2.norm1.weight",
242
+ "encoder.mid.block_2.norm2.bias",
243
+ "encoder.mid.block_2.norm2.weight",
244
+ "encoder.norm_out.bias",
245
+ "encoder.norm_out.weight",
246
+ "vae.decoder.conv_in.bias",
247
+ "vae.decoder.conv_in.weight",
248
+ "vae.decoder.conv_out.bias",
249
+ "vae.decoder.conv_out.weight",
250
+ "vae.decoder.mid.attn_1.k.bias",
251
+ "vae.decoder.mid.attn_1.k.weight",
252
+ "vae.decoder.mid.attn_1.norm.bias",
253
+ "vae.decoder.mid.attn_1.norm.weight",
254
+ "vae.decoder.mid.attn_1.proj_out.bias",
255
+ "vae.decoder.mid.attn_1.proj_out.weight",
256
+ "vae.decoder.mid.attn_1.q.bias",
257
+ "vae.decoder.mid.attn_1.q.weight",
258
+ "vae.decoder.mid.attn_1.v.bias",
259
+ "vae.decoder.mid.attn_1.v.weight",
260
+ "vae.decoder.mid.block_1.conv1.bias",
261
+ "vae.decoder.mid.block_1.conv1.weight",
262
+ "vae.decoder.mid.block_1.conv2.bias",
263
+ "vae.decoder.mid.block_1.conv2.weight",
264
+ "vae.decoder.mid.block_1.norm1.bias",
265
+ "vae.decoder.mid.block_1.norm1.weight",
266
+ "vae.decoder.mid.block_1.norm2.bias",
267
+ "vae.decoder.mid.block_1.norm2.weight",
268
+ "vae.decoder.mid.block_2.conv1.bias",
269
+ "vae.decoder.mid.block_2.conv1.weight",
270
+ "vae.decoder.mid.block_2.conv2.bias",
271
+ "vae.decoder.mid.block_2.conv2.weight",
272
+ "vae.decoder.mid.block_2.norm1.bias",
273
+ "vae.decoder.mid.block_2.norm1.weight",
274
+ "vae.decoder.mid.block_2.norm2.bias",
275
+ "vae.decoder.mid.block_2.norm2.weight",
276
+ "vae.decoder.norm_out.bias",
277
+ "vae.decoder.norm_out.weight",
278
+ "vae.decoder.up.0.block.0.conv1.bias",
279
+ "vae.decoder.up.0.block.0.conv1.weight",
280
+ "vae.decoder.up.0.block.0.conv2.bias",
281
+ "vae.decoder.up.0.block.0.conv2.weight",
282
+ "vae.decoder.up.0.block.0.nin_shortcut.bias",
283
+ "vae.decoder.up.0.block.0.nin_shortcut.weight",
284
+ "vae.decoder.up.0.block.0.norm1.bias",
285
+ "vae.decoder.up.0.block.0.norm1.weight",
286
+ "vae.decoder.up.0.block.0.norm2.bias",
287
+ "vae.decoder.up.0.block.0.norm2.weight",
288
+ "vae.decoder.up.0.block.1.conv1.bias",
289
+ "vae.decoder.up.0.block.1.conv1.weight",
290
+ "vae.decoder.up.0.block.1.conv2.bias",
291
+ "vae.decoder.up.0.block.1.conv2.weight",
292
+ "vae.decoder.up.0.block.1.norm1.bias",
293
+ "vae.decoder.up.0.block.1.norm1.weight",
294
+ "vae.decoder.up.0.block.1.norm2.bias",
295
+ "vae.decoder.up.0.block.1.norm2.weight",
296
+ "vae.decoder.up.0.block.2.conv1.bias",
297
+ "vae.decoder.up.0.block.2.conv1.weight",
298
+ "vae.decoder.up.0.block.2.conv2.bias",
299
+ "vae.decoder.up.0.block.2.conv2.weight",
300
+ "vae.decoder.up.0.block.2.norm1.bias",
301
+ "vae.decoder.up.0.block.2.norm1.weight",
302
+ "vae.decoder.up.0.block.2.norm2.bias",
303
+ "vae.decoder.up.0.block.2.norm2.weight",
304
+ "vae.decoder.up.1.block.0.conv1.bias",
305
+ "vae.decoder.up.1.block.0.conv1.weight",
306
+ "vae.decoder.up.1.block.0.conv2.bias",
307
+ "vae.decoder.up.1.block.0.conv2.weight",
308
+ "vae.decoder.up.1.block.0.nin_shortcut.bias",
309
+ "vae.decoder.up.1.block.0.nin_shortcut.weight",
310
+ "vae.decoder.up.1.block.0.norm1.bias",
311
+ "vae.decoder.up.1.block.0.norm1.weight",
312
+ "vae.decoder.up.1.block.0.norm2.bias",
313
+ "vae.decoder.up.1.block.0.norm2.weight",
314
+ "vae.decoder.up.1.block.1.conv1.bias",
315
+ "vae.decoder.up.1.block.1.conv1.weight",
316
+ "vae.decoder.up.1.block.1.conv2.bias",
317
+ "vae.decoder.up.1.block.1.conv2.weight",
318
+ "vae.decoder.up.1.block.1.norm1.bias",
319
+ "vae.decoder.up.1.block.1.norm1.weight",
320
+ "vae.decoder.up.1.block.1.norm2.bias",
321
+ "vae.decoder.up.1.block.1.norm2.weight",
322
+ "vae.decoder.up.1.block.2.conv1.bias",
323
+ "vae.decoder.up.1.block.2.conv1.weight",
324
+ "vae.decoder.up.1.block.2.conv2.bias",
325
+ "vae.decoder.up.1.block.2.conv2.weight",
326
+ "vae.decoder.up.1.block.2.norm1.bias",
327
+ "vae.decoder.up.1.block.2.norm1.weight",
328
+ "vae.decoder.up.1.block.2.norm2.bias",
329
+ "vae.decoder.up.1.block.2.norm2.weight",
330
+ "vae.decoder.up.1.upsample.conv.bias",
331
+ "vae.decoder.up.1.upsample.conv.weight",
332
+ "vae.decoder.up.2.block.0.conv1.bias",
333
+ "vae.decoder.up.2.block.0.conv1.weight",
334
+ "vae.decoder.up.2.block.0.conv2.bias",
335
+ "vae.decoder.up.2.block.0.conv2.weight",
336
+ "vae.decoder.up.2.block.0.norm1.bias",
337
+ "vae.decoder.up.2.block.0.norm1.weight",
338
+ "vae.decoder.up.2.block.0.norm2.bias",
339
+ "vae.decoder.up.2.block.0.norm2.weight",
340
+ "vae.decoder.up.2.block.1.conv1.bias",
341
+ "vae.decoder.up.2.block.1.conv1.weight",
342
+ "vae.decoder.up.2.block.1.conv2.bias",
343
+ "vae.decoder.up.2.block.1.conv2.weight",
344
+ "vae.decoder.up.2.block.1.norm1.bias",
345
+ "vae.decoder.up.2.block.1.norm1.weight",
346
+ "vae.decoder.up.2.block.1.norm2.bias",
347
+ "vae.decoder.up.2.block.1.norm2.weight",
348
+ "vae.decoder.up.2.block.2.conv1.bias",
349
+ "vae.decoder.up.2.block.2.conv1.weight",
350
+ "vae.decoder.up.2.block.2.conv2.bias",
351
+ "vae.decoder.up.2.block.2.conv2.weight",
352
+ "vae.decoder.up.2.block.2.norm1.bias",
353
+ "vae.decoder.up.2.block.2.norm1.weight",
354
+ "vae.decoder.up.2.block.2.norm2.bias",
355
+ "vae.decoder.up.2.block.2.norm2.weight",
356
+ "vae.decoder.up.2.upsample.conv.bias",
357
+ "vae.decoder.up.2.upsample.conv.weight",
358
+ "vae.decoder.up.3.block.0.conv1.bias",
359
+ "vae.decoder.up.3.block.0.conv1.weight",
360
+ "vae.decoder.up.3.block.0.conv2.bias",
361
+ "vae.decoder.up.3.block.0.conv2.weight",
362
+ "vae.decoder.up.3.block.0.norm1.bias",
363
+ "vae.decoder.up.3.block.0.norm1.weight",
364
+ "vae.decoder.up.3.block.0.norm2.bias",
365
+ "vae.decoder.up.3.block.0.norm2.weight",
366
+ "vae.decoder.up.3.block.1.conv1.bias",
367
+ "vae.decoder.up.3.block.1.conv1.weight",
368
+ "vae.decoder.up.3.block.1.conv2.bias",
369
+ "vae.decoder.up.3.block.1.conv2.weight",
370
+ "vae.decoder.up.3.block.1.norm1.bias",
371
+ "vae.decoder.up.3.block.1.norm1.weight",
372
+ "vae.decoder.up.3.block.1.norm2.bias",
373
+ "vae.decoder.up.3.block.1.norm2.weight",
374
+ "vae.decoder.up.3.block.2.conv1.bias",
375
+ "vae.decoder.up.3.block.2.conv1.weight",
376
+ "vae.decoder.up.3.block.2.conv2.bias",
377
+ "vae.decoder.up.3.block.2.conv2.weight",
378
+ "vae.decoder.up.3.block.2.norm1.bias",
379
+ "vae.decoder.up.3.block.2.norm1.weight",
380
+ "vae.decoder.up.3.block.2.norm2.bias",
381
+ "vae.decoder.up.3.block.2.norm2.weight",
382
+ "vae.decoder.up.3.upsample.conv.bias",
383
+ "vae.decoder.up.3.upsample.conv.weight",
384
+ "vae.encoder.conv_in.bias",
385
+ "vae.encoder.conv_in.weight",
386
+ "vae.encoder.conv_out.bias",
387
+ "vae.encoder.conv_out.weight",
388
+ "vae.encoder.down.0.block.0.conv1.bias",
389
+ "vae.encoder.down.0.block.0.conv1.weight",
390
+ "vae.encoder.down.0.block.0.conv2.bias",
391
+ "vae.encoder.down.0.block.0.conv2.weight",
392
+ "vae.encoder.down.0.block.0.norm1.bias",
393
+ "vae.encoder.down.0.block.0.norm1.weight",
394
+ "vae.encoder.down.0.block.0.norm2.bias",
395
+ "vae.encoder.down.0.block.0.norm2.weight",
396
+ "vae.encoder.down.0.block.1.conv1.bias",
397
+ "vae.encoder.down.0.block.1.conv1.weight",
398
+ "vae.encoder.down.0.block.1.conv2.bias",
399
+ "vae.encoder.down.0.block.1.conv2.weight",
400
+ "vae.encoder.down.0.block.1.norm1.bias",
401
+ "vae.encoder.down.0.block.1.norm1.weight",
402
+ "vae.encoder.down.0.block.1.norm2.bias",
403
+ "vae.encoder.down.0.block.1.norm2.weight",
404
+ "vae.encoder.down.0.downsample.conv.bias",
405
+ "vae.encoder.down.0.downsample.conv.weight",
406
+ "vae.encoder.down.1.block.0.conv1.bias",
407
+ "vae.encoder.down.1.block.0.conv1.weight",
408
+ "vae.encoder.down.1.block.0.conv2.bias",
409
+ "vae.encoder.down.1.block.0.conv2.weight",
410
+ "vae.encoder.down.1.block.0.nin_shortcut.bias",
411
+ "vae.encoder.down.1.block.0.nin_shortcut.weight",
412
+ "vae.encoder.down.1.block.0.norm1.bias",
413
+ "vae.encoder.down.1.block.0.norm1.weight",
414
+ "vae.encoder.down.1.block.0.norm2.bias",
415
+ "vae.encoder.down.1.block.0.norm2.weight",
416
+ "vae.encoder.down.1.block.1.conv1.bias",
417
+ "vae.encoder.down.1.block.1.conv1.weight",
418
+ "vae.encoder.down.1.block.1.conv2.bias",
419
+ "vae.encoder.down.1.block.1.conv2.weight",
420
+ "vae.encoder.down.1.block.1.norm1.bias",
421
+ "vae.encoder.down.1.block.1.norm1.weight",
422
+ "vae.encoder.down.1.block.1.norm2.bias",
423
+ "vae.encoder.down.1.block.1.norm2.weight",
424
+ "vae.encoder.down.1.downsample.conv.bias",
425
+ "vae.encoder.down.1.downsample.conv.weight",
426
+ "vae.encoder.down.2.block.0.conv1.bias",
427
+ "vae.encoder.down.2.block.0.conv1.weight",
428
+ "vae.encoder.down.2.block.0.conv2.bias",
429
+ "vae.encoder.down.2.block.0.conv2.weight",
430
+ "vae.encoder.down.2.block.0.nin_shortcut.bias",
431
+ "vae.encoder.down.2.block.0.nin_shortcut.weight",
432
+ "vae.encoder.down.2.block.0.norm1.bias",
433
+ "vae.encoder.down.2.block.0.norm1.weight",
434
+ "vae.encoder.down.2.block.0.norm2.bias",
435
+ "vae.encoder.down.2.block.0.norm2.weight",
436
+ "vae.encoder.down.2.block.1.conv1.bias",
437
+ "vae.encoder.down.2.block.1.conv1.weight",
438
+ "vae.encoder.down.2.block.1.conv2.bias",
439
+ "vae.encoder.down.2.block.1.conv2.weight",
440
+ "vae.encoder.down.2.block.1.norm1.bias",
441
+ "vae.encoder.down.2.block.1.norm1.weight",
442
+ "vae.encoder.down.2.block.1.norm2.bias",
443
+ "vae.encoder.down.2.block.1.norm2.weight",
444
+ "vae.encoder.down.2.downsample.conv.bias",
445
+ "vae.encoder.down.2.downsample.conv.weight",
446
+ "vae.encoder.down.3.block.0.conv1.bias",
447
+ "vae.encoder.down.3.block.0.conv1.weight",
448
+ "vae.encoder.down.3.block.0.conv2.bias",
449
+ "vae.encoder.down.3.block.0.conv2.weight",
450
+ "vae.encoder.down.3.block.0.norm1.bias",
451
+ "vae.encoder.down.3.block.0.norm1.weight",
452
+ "vae.encoder.down.3.block.0.norm2.bias",
453
+ "vae.encoder.down.3.block.0.norm2.weight",
454
+ "vae.encoder.down.3.block.1.conv1.bias",
455
+ "vae.encoder.down.3.block.1.conv1.weight",
456
+ "vae.encoder.down.3.block.1.conv2.bias",
457
+ "vae.encoder.down.3.block.1.conv2.weight",
458
+ "vae.encoder.down.3.block.1.norm1.bias",
459
+ "vae.encoder.down.3.block.1.norm1.weight",
460
+ "vae.encoder.down.3.block.1.norm2.bias",
461
+ "vae.encoder.down.3.block.1.norm2.weight",
462
+ "vae.encoder.mid.attn_1.k.bias",
463
+ "vae.encoder.mid.attn_1.k.weight",
464
+ "vae.encoder.mid.attn_1.norm.bias",
465
+ "vae.encoder.mid.attn_1.norm.weight",
466
+ "vae.encoder.mid.attn_1.proj_out.bias",
467
+ "vae.encoder.mid.attn_1.proj_out.weight",
468
+ "vae.encoder.mid.attn_1.q.bias",
469
+ "vae.encoder.mid.attn_1.q.weight",
470
+ "vae.encoder.mid.attn_1.v.bias",
471
+ "vae.encoder.mid.attn_1.v.weight",
472
+ "vae.encoder.mid.block_1.conv1.bias",
473
+ "vae.encoder.mid.block_1.conv1.weight",
474
+ "vae.encoder.mid.block_1.conv2.bias",
475
+ "vae.encoder.mid.block_1.conv2.weight",
476
+ "vae.encoder.mid.block_1.norm1.bias",
477
+ "vae.encoder.mid.block_1.norm1.weight",
478
+ "vae.encoder.mid.block_1.norm2.bias",
479
+ "vae.encoder.mid.block_1.norm2.weight",
480
+ "vae.encoder.mid.block_2.conv1.bias",
481
+ "vae.encoder.mid.block_2.conv1.weight",
482
+ "vae.encoder.mid.block_2.conv2.bias",
483
+ "vae.encoder.mid.block_2.conv2.weight",
484
+ "vae.encoder.mid.block_2.norm1.bias",
485
+ "vae.encoder.mid.block_2.norm1.weight",
486
+ "vae.encoder.mid.block_2.norm2.bias",
487
+ "vae.encoder.mid.block_2.norm2.weight",
488
+ "vae.encoder.norm_out.bias",
489
+ "vae.encoder.norm_out.weight"
490
+ ]
fluxunchainedArtfulNSFW_fuT516xfp8E4m3fnV11_fixed.safetensors.new.txt.txt ADDED
@@ -0,0 +1,1442 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ double_blocks.0.img_attn.norm.key_norm.scale
2
+ double_blocks.0.img_attn.norm.query_norm.scale
3
+ double_blocks.0.img_attn.proj.bias
4
+ double_blocks.0.img_attn.proj.weight
5
+ double_blocks.0.img_attn.qkv.bias
6
+ double_blocks.0.img_attn.qkv.weight
7
+ double_blocks.0.img_mlp.0.bias
8
+ double_blocks.0.img_mlp.0.weight
9
+ double_blocks.0.img_mlp.2.bias
10
+ double_blocks.0.img_mlp.2.weight
11
+ double_blocks.0.img_mod.lin.bias
12
+ double_blocks.0.img_mod.lin.weight
13
+ double_blocks.0.txt_attn.norm.key_norm.scale
14
+ double_blocks.0.txt_attn.norm.query_norm.scale
15
+ double_blocks.0.txt_attn.proj.bias
16
+ double_blocks.0.txt_attn.proj.weight
17
+ double_blocks.0.txt_attn.qkv.bias
18
+ double_blocks.0.txt_attn.qkv.weight
19
+ double_blocks.0.txt_mlp.0.bias
20
+ double_blocks.0.txt_mlp.0.weight
21
+ double_blocks.0.txt_mlp.2.bias
22
+ double_blocks.0.txt_mlp.2.weight
23
+ double_blocks.0.txt_mod.lin.bias
24
+ double_blocks.0.txt_mod.lin.weight
25
+ double_blocks.1.img_attn.norm.key_norm.scale
26
+ double_blocks.1.img_attn.norm.query_norm.scale
27
+ double_blocks.1.img_attn.proj.bias
28
+ double_blocks.1.img_attn.proj.weight
29
+ double_blocks.1.img_attn.qkv.bias
30
+ double_blocks.1.img_attn.qkv.weight
31
+ double_blocks.1.img_mlp.0.bias
32
+ double_blocks.1.img_mlp.0.weight
33
+ double_blocks.1.img_mlp.2.bias
34
+ double_blocks.1.img_mlp.2.weight
35
+ double_blocks.1.img_mod.lin.bias
36
+ double_blocks.1.img_mod.lin.weight
37
+ double_blocks.1.txt_attn.norm.key_norm.scale
38
+ double_blocks.1.txt_attn.norm.query_norm.scale
39
+ double_blocks.1.txt_attn.proj.bias
40
+ double_blocks.1.txt_attn.proj.weight
41
+ double_blocks.1.txt_attn.qkv.bias
42
+ double_blocks.1.txt_attn.qkv.weight
43
+ double_blocks.1.txt_mlp.0.bias
44
+ double_blocks.1.txt_mlp.0.weight
45
+ double_blocks.1.txt_mlp.2.bias
46
+ double_blocks.1.txt_mlp.2.weight
47
+ double_blocks.1.txt_mod.lin.bias
48
+ double_blocks.1.txt_mod.lin.weight
49
+ double_blocks.10.img_attn.norm.key_norm.scale
50
+ double_blocks.10.img_attn.norm.query_norm.scale
51
+ double_blocks.10.img_attn.proj.bias
52
+ double_blocks.10.img_attn.proj.weight
53
+ double_blocks.10.img_attn.qkv.bias
54
+ double_blocks.10.img_attn.qkv.weight
55
+ double_blocks.10.img_mlp.0.bias
56
+ double_blocks.10.img_mlp.0.weight
57
+ double_blocks.10.img_mlp.2.bias
58
+ double_blocks.10.img_mlp.2.weight
59
+ double_blocks.10.img_mod.lin.bias
60
+ double_blocks.10.img_mod.lin.weight
61
+ double_blocks.10.txt_attn.norm.key_norm.scale
62
+ double_blocks.10.txt_attn.norm.query_norm.scale
63
+ double_blocks.10.txt_attn.proj.bias
64
+ double_blocks.10.txt_attn.proj.weight
65
+ double_blocks.10.txt_attn.qkv.bias
66
+ double_blocks.10.txt_attn.qkv.weight
67
+ double_blocks.10.txt_mlp.0.bias
68
+ double_blocks.10.txt_mlp.0.weight
69
+ double_blocks.10.txt_mlp.2.bias
70
+ double_blocks.10.txt_mlp.2.weight
71
+ double_blocks.10.txt_mod.lin.bias
72
+ double_blocks.10.txt_mod.lin.weight
73
+ double_blocks.11.img_attn.norm.key_norm.scale
74
+ double_blocks.11.img_attn.norm.query_norm.scale
75
+ double_blocks.11.img_attn.proj.bias
76
+ double_blocks.11.img_attn.proj.weight
77
+ double_blocks.11.img_attn.qkv.bias
78
+ double_blocks.11.img_attn.qkv.weight
79
+ double_blocks.11.img_mlp.0.bias
80
+ double_blocks.11.img_mlp.0.weight
81
+ double_blocks.11.img_mlp.2.bias
82
+ double_blocks.11.img_mlp.2.weight
83
+ double_blocks.11.img_mod.lin.bias
84
+ double_blocks.11.img_mod.lin.weight
85
+ double_blocks.11.txt_attn.norm.key_norm.scale
86
+ double_blocks.11.txt_attn.norm.query_norm.scale
87
+ double_blocks.11.txt_attn.proj.bias
88
+ double_blocks.11.txt_attn.proj.weight
89
+ double_blocks.11.txt_attn.qkv.bias
90
+ double_blocks.11.txt_attn.qkv.weight
91
+ double_blocks.11.txt_mlp.0.bias
92
+ double_blocks.11.txt_mlp.0.weight
93
+ double_blocks.11.txt_mlp.2.bias
94
+ double_blocks.11.txt_mlp.2.weight
95
+ double_blocks.11.txt_mod.lin.bias
96
+ double_blocks.11.txt_mod.lin.weight
97
+ double_blocks.12.img_attn.norm.key_norm.scale
98
+ double_blocks.12.img_attn.norm.query_norm.scale
99
+ double_blocks.12.img_attn.proj.bias
100
+ double_blocks.12.img_attn.proj.weight
101
+ double_blocks.12.img_attn.qkv.bias
102
+ double_blocks.12.img_attn.qkv.weight
103
+ double_blocks.12.img_mlp.0.bias
104
+ double_blocks.12.img_mlp.0.weight
105
+ double_blocks.12.img_mlp.2.bias
106
+ double_blocks.12.img_mlp.2.weight
107
+ double_blocks.12.img_mod.lin.bias
108
+ double_blocks.12.img_mod.lin.weight
109
+ double_blocks.12.txt_attn.norm.key_norm.scale
110
+ double_blocks.12.txt_attn.norm.query_norm.scale
111
+ double_blocks.12.txt_attn.proj.bias
112
+ double_blocks.12.txt_attn.proj.weight
113
+ double_blocks.12.txt_attn.qkv.bias
114
+ double_blocks.12.txt_attn.qkv.weight
115
+ double_blocks.12.txt_mlp.0.bias
116
+ double_blocks.12.txt_mlp.0.weight
117
+ double_blocks.12.txt_mlp.2.bias
118
+ double_blocks.12.txt_mlp.2.weight
119
+ double_blocks.12.txt_mod.lin.bias
120
+ double_blocks.12.txt_mod.lin.weight
121
+ double_blocks.13.img_attn.norm.key_norm.scale
122
+ double_blocks.13.img_attn.norm.query_norm.scale
123
+ double_blocks.13.img_attn.proj.bias
124
+ double_blocks.13.img_attn.proj.weight
125
+ double_blocks.13.img_attn.qkv.bias
126
+ double_blocks.13.img_attn.qkv.weight
127
+ double_blocks.13.img_mlp.0.bias
128
+ double_blocks.13.img_mlp.0.weight
129
+ double_blocks.13.img_mlp.2.bias
130
+ double_blocks.13.img_mlp.2.weight
131
+ double_blocks.13.img_mod.lin.bias
132
+ double_blocks.13.img_mod.lin.weight
133
+ double_blocks.13.txt_attn.norm.key_norm.scale
134
+ double_blocks.13.txt_attn.norm.query_norm.scale
135
+ double_blocks.13.txt_attn.proj.bias
136
+ double_blocks.13.txt_attn.proj.weight
137
+ double_blocks.13.txt_attn.qkv.bias
138
+ double_blocks.13.txt_attn.qkv.weight
139
+ double_blocks.13.txt_mlp.0.bias
140
+ double_blocks.13.txt_mlp.0.weight
141
+ double_blocks.13.txt_mlp.2.bias
142
+ double_blocks.13.txt_mlp.2.weight
143
+ double_blocks.13.txt_mod.lin.bias
144
+ double_blocks.13.txt_mod.lin.weight
145
+ double_blocks.14.img_attn.norm.key_norm.scale
146
+ double_blocks.14.img_attn.norm.query_norm.scale
147
+ double_blocks.14.img_attn.proj.bias
148
+ double_blocks.14.img_attn.proj.weight
149
+ double_blocks.14.img_attn.qkv.bias
150
+ double_blocks.14.img_attn.qkv.weight
151
+ double_blocks.14.img_mlp.0.bias
152
+ double_blocks.14.img_mlp.0.weight
153
+ double_blocks.14.img_mlp.2.bias
154
+ double_blocks.14.img_mlp.2.weight
155
+ double_blocks.14.img_mod.lin.bias
156
+ double_blocks.14.img_mod.lin.weight
157
+ double_blocks.14.txt_attn.norm.key_norm.scale
158
+ double_blocks.14.txt_attn.norm.query_norm.scale
159
+ double_blocks.14.txt_attn.proj.bias
160
+ double_blocks.14.txt_attn.proj.weight
161
+ double_blocks.14.txt_attn.qkv.bias
162
+ double_blocks.14.txt_attn.qkv.weight
163
+ double_blocks.14.txt_mlp.0.bias
164
+ double_blocks.14.txt_mlp.0.weight
165
+ double_blocks.14.txt_mlp.2.bias
166
+ double_blocks.14.txt_mlp.2.weight
167
+ double_blocks.14.txt_mod.lin.bias
168
+ double_blocks.14.txt_mod.lin.weight
169
+ double_blocks.15.img_attn.norm.key_norm.scale
170
+ double_blocks.15.img_attn.norm.query_norm.scale
171
+ double_blocks.15.img_attn.proj.bias
172
+ double_blocks.15.img_attn.proj.weight
173
+ double_blocks.15.img_attn.qkv.bias
174
+ double_blocks.15.img_attn.qkv.weight
175
+ double_blocks.15.img_mlp.0.bias
176
+ double_blocks.15.img_mlp.0.weight
177
+ double_blocks.15.img_mlp.2.bias
178
+ double_blocks.15.img_mlp.2.weight
179
+ double_blocks.15.img_mod.lin.bias
180
+ double_blocks.15.img_mod.lin.weight
181
+ double_blocks.15.txt_attn.norm.key_norm.scale
182
+ double_blocks.15.txt_attn.norm.query_norm.scale
183
+ double_blocks.15.txt_attn.proj.bias
184
+ double_blocks.15.txt_attn.proj.weight
185
+ double_blocks.15.txt_attn.qkv.bias
186
+ double_blocks.15.txt_attn.qkv.weight
187
+ double_blocks.15.txt_mlp.0.bias
188
+ double_blocks.15.txt_mlp.0.weight
189
+ double_blocks.15.txt_mlp.2.bias
190
+ double_blocks.15.txt_mlp.2.weight
191
+ double_blocks.15.txt_mod.lin.bias
192
+ double_blocks.15.txt_mod.lin.weight
193
+ double_blocks.16.img_attn.norm.key_norm.scale
194
+ double_blocks.16.img_attn.norm.query_norm.scale
195
+ double_blocks.16.img_attn.proj.bias
196
+ double_blocks.16.img_attn.proj.weight
197
+ double_blocks.16.img_attn.qkv.bias
198
+ double_blocks.16.img_attn.qkv.weight
199
+ double_blocks.16.img_mlp.0.bias
200
+ double_blocks.16.img_mlp.0.weight
201
+ double_blocks.16.img_mlp.2.bias
202
+ double_blocks.16.img_mlp.2.weight
203
+ double_blocks.16.img_mod.lin.bias
204
+ double_blocks.16.img_mod.lin.weight
205
+ double_blocks.16.txt_attn.norm.key_norm.scale
206
+ double_blocks.16.txt_attn.norm.query_norm.scale
207
+ double_blocks.16.txt_attn.proj.bias
208
+ double_blocks.16.txt_attn.proj.weight
209
+ double_blocks.16.txt_attn.qkv.bias
210
+ double_blocks.16.txt_attn.qkv.weight
211
+ double_blocks.16.txt_mlp.0.bias
212
+ double_blocks.16.txt_mlp.0.weight
213
+ double_blocks.16.txt_mlp.2.bias
214
+ double_blocks.16.txt_mlp.2.weight
215
+ double_blocks.16.txt_mod.lin.bias
216
+ double_blocks.16.txt_mod.lin.weight
217
+ double_blocks.17.img_attn.norm.key_norm.scale
218
+ double_blocks.17.img_attn.norm.query_norm.scale
219
+ double_blocks.17.img_attn.proj.bias
220
+ double_blocks.17.img_attn.proj.weight
221
+ double_blocks.17.img_attn.qkv.bias
222
+ double_blocks.17.img_attn.qkv.weight
223
+ double_blocks.17.img_mlp.0.bias
224
+ double_blocks.17.img_mlp.0.weight
225
+ double_blocks.17.img_mlp.2.bias
226
+ double_blocks.17.img_mlp.2.weight
227
+ double_blocks.17.img_mod.lin.bias
228
+ double_blocks.17.img_mod.lin.weight
229
+ double_blocks.17.txt_attn.norm.key_norm.scale
230
+ double_blocks.17.txt_attn.norm.query_norm.scale
231
+ double_blocks.17.txt_attn.proj.bias
232
+ double_blocks.17.txt_attn.proj.weight
233
+ double_blocks.17.txt_attn.qkv.bias
234
+ double_blocks.17.txt_attn.qkv.weight
235
+ double_blocks.17.txt_mlp.0.bias
236
+ double_blocks.17.txt_mlp.0.weight
237
+ double_blocks.17.txt_mlp.2.bias
238
+ double_blocks.17.txt_mlp.2.weight
239
+ double_blocks.17.txt_mod.lin.bias
240
+ double_blocks.17.txt_mod.lin.weight
241
+ double_blocks.18.img_attn.norm.key_norm.scale
242
+ double_blocks.18.img_attn.norm.query_norm.scale
243
+ double_blocks.18.img_attn.proj.bias
244
+ double_blocks.18.img_attn.proj.weight
245
+ double_blocks.18.img_attn.qkv.bias
246
+ double_blocks.18.img_attn.qkv.weight
247
+ double_blocks.18.img_mlp.0.bias
248
+ double_blocks.18.img_mlp.0.weight
249
+ double_blocks.18.img_mlp.2.bias
250
+ double_blocks.18.img_mlp.2.weight
251
+ double_blocks.18.img_mod.lin.bias
252
+ double_blocks.18.img_mod.lin.weight
253
+ double_blocks.18.txt_attn.norm.key_norm.scale
254
+ double_blocks.18.txt_attn.norm.query_norm.scale
255
+ double_blocks.18.txt_attn.proj.bias
256
+ double_blocks.18.txt_attn.proj.weight
257
+ double_blocks.18.txt_attn.qkv.bias
258
+ double_blocks.18.txt_attn.qkv.weight
259
+ double_blocks.18.txt_mlp.0.bias
260
+ double_blocks.18.txt_mlp.0.weight
261
+ double_blocks.18.txt_mlp.2.bias
262
+ double_blocks.18.txt_mlp.2.weight
263
+ double_blocks.18.txt_mod.lin.bias
264
+ double_blocks.18.txt_mod.lin.weight
265
+ double_blocks.2.img_attn.norm.key_norm.scale
266
+ double_blocks.2.img_attn.norm.query_norm.scale
267
+ double_blocks.2.img_attn.proj.bias
268
+ double_blocks.2.img_attn.proj.weight
269
+ double_blocks.2.img_attn.qkv.bias
270
+ double_blocks.2.img_attn.qkv.weight
271
+ double_blocks.2.img_mlp.0.bias
272
+ double_blocks.2.img_mlp.0.weight
273
+ double_blocks.2.img_mlp.2.bias
274
+ double_blocks.2.img_mlp.2.weight
275
+ double_blocks.2.img_mod.lin.bias
276
+ double_blocks.2.img_mod.lin.weight
277
+ double_blocks.2.txt_attn.norm.key_norm.scale
278
+ double_blocks.2.txt_attn.norm.query_norm.scale
279
+ double_blocks.2.txt_attn.proj.bias
280
+ double_blocks.2.txt_attn.proj.weight
281
+ double_blocks.2.txt_attn.qkv.bias
282
+ double_blocks.2.txt_attn.qkv.weight
283
+ double_blocks.2.txt_mlp.0.bias
284
+ double_blocks.2.txt_mlp.0.weight
285
+ double_blocks.2.txt_mlp.2.bias
286
+ double_blocks.2.txt_mlp.2.weight
287
+ double_blocks.2.txt_mod.lin.bias
288
+ double_blocks.2.txt_mod.lin.weight
289
+ double_blocks.3.img_attn.norm.key_norm.scale
290
+ double_blocks.3.img_attn.norm.query_norm.scale
291
+ double_blocks.3.img_attn.proj.bias
292
+ double_blocks.3.img_attn.proj.weight
293
+ double_blocks.3.img_attn.qkv.bias
294
+ double_blocks.3.img_attn.qkv.weight
295
+ double_blocks.3.img_mlp.0.bias
296
+ double_blocks.3.img_mlp.0.weight
297
+ double_blocks.3.img_mlp.2.bias
298
+ double_blocks.3.img_mlp.2.weight
299
+ double_blocks.3.img_mod.lin.bias
300
+ double_blocks.3.img_mod.lin.weight
301
+ double_blocks.3.txt_attn.norm.key_norm.scale
302
+ double_blocks.3.txt_attn.norm.query_norm.scale
303
+ double_blocks.3.txt_attn.proj.bias
304
+ double_blocks.3.txt_attn.proj.weight
305
+ double_blocks.3.txt_attn.qkv.bias
306
+ double_blocks.3.txt_attn.qkv.weight
307
+ double_blocks.3.txt_mlp.0.bias
308
+ double_blocks.3.txt_mlp.0.weight
309
+ double_blocks.3.txt_mlp.2.bias
310
+ double_blocks.3.txt_mlp.2.weight
311
+ double_blocks.3.txt_mod.lin.bias
312
+ double_blocks.3.txt_mod.lin.weight
313
+ double_blocks.4.img_attn.norm.key_norm.scale
314
+ double_blocks.4.img_attn.norm.query_norm.scale
315
+ double_blocks.4.img_attn.proj.bias
316
+ double_blocks.4.img_attn.proj.weight
317
+ double_blocks.4.img_attn.qkv.bias
318
+ double_blocks.4.img_attn.qkv.weight
319
+ double_blocks.4.img_mlp.0.bias
320
+ double_blocks.4.img_mlp.0.weight
321
+ double_blocks.4.img_mlp.2.bias
322
+ double_blocks.4.img_mlp.2.weight
323
+ double_blocks.4.img_mod.lin.bias
324
+ double_blocks.4.img_mod.lin.weight
325
+ double_blocks.4.txt_attn.norm.key_norm.scale
326
+ double_blocks.4.txt_attn.norm.query_norm.scale
327
+ double_blocks.4.txt_attn.proj.bias
328
+ double_blocks.4.txt_attn.proj.weight
329
+ double_blocks.4.txt_attn.qkv.bias
330
+ double_blocks.4.txt_attn.qkv.weight
331
+ double_blocks.4.txt_mlp.0.bias
332
+ double_blocks.4.txt_mlp.0.weight
333
+ double_blocks.4.txt_mlp.2.bias
334
+ double_blocks.4.txt_mlp.2.weight
335
+ double_blocks.4.txt_mod.lin.bias
336
+ double_blocks.4.txt_mod.lin.weight
337
+ double_blocks.5.img_attn.norm.key_norm.scale
338
+ double_blocks.5.img_attn.norm.query_norm.scale
339
+ double_blocks.5.img_attn.proj.bias
340
+ double_blocks.5.img_attn.proj.weight
341
+ double_blocks.5.img_attn.qkv.bias
342
+ double_blocks.5.img_attn.qkv.weight
343
+ double_blocks.5.img_mlp.0.bias
344
+ double_blocks.5.img_mlp.0.weight
345
+ double_blocks.5.img_mlp.2.bias
346
+ double_blocks.5.img_mlp.2.weight
347
+ double_blocks.5.img_mod.lin.bias
348
+ double_blocks.5.img_mod.lin.weight
349
+ double_blocks.5.txt_attn.norm.key_norm.scale
350
+ double_blocks.5.txt_attn.norm.query_norm.scale
351
+ double_blocks.5.txt_attn.proj.bias
352
+ double_blocks.5.txt_attn.proj.weight
353
+ double_blocks.5.txt_attn.qkv.bias
354
+ double_blocks.5.txt_attn.qkv.weight
355
+ double_blocks.5.txt_mlp.0.bias
356
+ double_blocks.5.txt_mlp.0.weight
357
+ double_blocks.5.txt_mlp.2.bias
358
+ double_blocks.5.txt_mlp.2.weight
359
+ double_blocks.5.txt_mod.lin.bias
360
+ double_blocks.5.txt_mod.lin.weight
361
+ double_blocks.6.img_attn.norm.key_norm.scale
362
+ double_blocks.6.img_attn.norm.query_norm.scale
363
+ double_blocks.6.img_attn.proj.bias
364
+ double_blocks.6.img_attn.proj.weight
365
+ double_blocks.6.img_attn.qkv.bias
366
+ double_blocks.6.img_attn.qkv.weight
367
+ double_blocks.6.img_mlp.0.bias
368
+ double_blocks.6.img_mlp.0.weight
369
+ double_blocks.6.img_mlp.2.bias
370
+ double_blocks.6.img_mlp.2.weight
371
+ double_blocks.6.img_mod.lin.bias
372
+ double_blocks.6.img_mod.lin.weight
373
+ double_blocks.6.txt_attn.norm.key_norm.scale
374
+ double_blocks.6.txt_attn.norm.query_norm.scale
375
+ double_blocks.6.txt_attn.proj.bias
376
+ double_blocks.6.txt_attn.proj.weight
377
+ double_blocks.6.txt_attn.qkv.bias
378
+ double_blocks.6.txt_attn.qkv.weight
379
+ double_blocks.6.txt_mlp.0.bias
380
+ double_blocks.6.txt_mlp.0.weight
381
+ double_blocks.6.txt_mlp.2.bias
382
+ double_blocks.6.txt_mlp.2.weight
383
+ double_blocks.6.txt_mod.lin.bias
384
+ double_blocks.6.txt_mod.lin.weight
385
+ double_blocks.7.img_attn.norm.key_norm.scale
386
+ double_blocks.7.img_attn.norm.query_norm.scale
387
+ double_blocks.7.img_attn.proj.bias
388
+ double_blocks.7.img_attn.proj.weight
389
+ double_blocks.7.img_attn.qkv.bias
390
+ double_blocks.7.img_attn.qkv.weight
391
+ double_blocks.7.img_mlp.0.bias
392
+ double_blocks.7.img_mlp.0.weight
393
+ double_blocks.7.img_mlp.2.bias
394
+ double_blocks.7.img_mlp.2.weight
395
+ double_blocks.7.img_mod.lin.bias
396
+ double_blocks.7.img_mod.lin.weight
397
+ double_blocks.7.txt_attn.norm.key_norm.scale
398
+ double_blocks.7.txt_attn.norm.query_norm.scale
399
+ double_blocks.7.txt_attn.proj.bias
400
+ double_blocks.7.txt_attn.proj.weight
401
+ double_blocks.7.txt_attn.qkv.bias
402
+ double_blocks.7.txt_attn.qkv.weight
403
+ double_blocks.7.txt_mlp.0.bias
404
+ double_blocks.7.txt_mlp.0.weight
405
+ double_blocks.7.txt_mlp.2.bias
406
+ double_blocks.7.txt_mlp.2.weight
407
+ double_blocks.7.txt_mod.lin.bias
408
+ double_blocks.7.txt_mod.lin.weight
409
+ double_blocks.8.img_attn.norm.key_norm.scale
410
+ double_blocks.8.img_attn.norm.query_norm.scale
411
+ double_blocks.8.img_attn.proj.bias
412
+ double_blocks.8.img_attn.proj.weight
413
+ double_blocks.8.img_attn.qkv.bias
414
+ double_blocks.8.img_attn.qkv.weight
415
+ double_blocks.8.img_mlp.0.bias
416
+ double_blocks.8.img_mlp.0.weight
417
+ double_blocks.8.img_mlp.2.bias
418
+ double_blocks.8.img_mlp.2.weight
419
+ double_blocks.8.img_mod.lin.bias
420
+ double_blocks.8.img_mod.lin.weight
421
+ double_blocks.8.txt_attn.norm.key_norm.scale
422
+ double_blocks.8.txt_attn.norm.query_norm.scale
423
+ double_blocks.8.txt_attn.proj.bias
424
+ double_blocks.8.txt_attn.proj.weight
425
+ double_blocks.8.txt_attn.qkv.bias
426
+ double_blocks.8.txt_attn.qkv.weight
427
+ double_blocks.8.txt_mlp.0.bias
428
+ double_blocks.8.txt_mlp.0.weight
429
+ double_blocks.8.txt_mlp.2.bias
430
+ double_blocks.8.txt_mlp.2.weight
431
+ double_blocks.8.txt_mod.lin.bias
432
+ double_blocks.8.txt_mod.lin.weight
433
+ double_blocks.9.img_attn.norm.key_norm.scale
434
+ double_blocks.9.img_attn.norm.query_norm.scale
435
+ double_blocks.9.img_attn.proj.bias
436
+ double_blocks.9.img_attn.proj.weight
437
+ double_blocks.9.img_attn.qkv.bias
438
+ double_blocks.9.img_attn.qkv.weight
439
+ double_blocks.9.img_mlp.0.bias
440
+ double_blocks.9.img_mlp.0.weight
441
+ double_blocks.9.img_mlp.2.bias
442
+ double_blocks.9.img_mlp.2.weight
443
+ double_blocks.9.img_mod.lin.bias
444
+ double_blocks.9.img_mod.lin.weight
445
+ double_blocks.9.txt_attn.norm.key_norm.scale
446
+ double_blocks.9.txt_attn.norm.query_norm.scale
447
+ double_blocks.9.txt_attn.proj.bias
448
+ double_blocks.9.txt_attn.proj.weight
449
+ double_blocks.9.txt_attn.qkv.bias
450
+ double_blocks.9.txt_attn.qkv.weight
451
+ double_blocks.9.txt_mlp.0.bias
452
+ double_blocks.9.txt_mlp.0.weight
453
+ double_blocks.9.txt_mlp.2.bias
454
+ double_blocks.9.txt_mlp.2.weight
455
+ double_blocks.9.txt_mod.lin.bias
456
+ double_blocks.9.txt_mod.lin.weight
457
+ final_layer.adaLN_modulation.1.bias
458
+ final_layer.adaLN_modulation.1.weight
459
+ final_layer.linear.bias
460
+ final_layer.linear.weight
461
+ guidance_in.in_layer.bias
462
+ guidance_in.in_layer.weight
463
+ guidance_in.out_layer.bias
464
+ guidance_in.out_layer.weight
465
+ img_in.bias
466
+ img_in.weight
467
+ single_blocks.0.linear1.bias
468
+ single_blocks.0.linear1.weight
469
+ single_blocks.0.linear2.bias
470
+ single_blocks.0.linear2.weight
471
+ single_blocks.0.modulation.lin.bias
472
+ single_blocks.0.modulation.lin.weight
473
+ single_blocks.0.norm.key_norm.scale
474
+ single_blocks.0.norm.query_norm.scale
475
+ single_blocks.1.linear1.bias
476
+ single_blocks.1.linear1.weight
477
+ single_blocks.1.linear2.bias
478
+ single_blocks.1.linear2.weight
479
+ single_blocks.1.modulation.lin.bias
480
+ single_blocks.1.modulation.lin.weight
481
+ single_blocks.1.norm.key_norm.scale
482
+ single_blocks.1.norm.query_norm.scale
483
+ single_blocks.10.linear1.bias
484
+ single_blocks.10.linear1.weight
485
+ single_blocks.10.linear2.bias
486
+ single_blocks.10.linear2.weight
487
+ single_blocks.10.modulation.lin.bias
488
+ single_blocks.10.modulation.lin.weight
489
+ single_blocks.10.norm.key_norm.scale
490
+ single_blocks.10.norm.query_norm.scale
491
+ single_blocks.11.linear1.bias
492
+ single_blocks.11.linear1.weight
493
+ single_blocks.11.linear2.bias
494
+ single_blocks.11.linear2.weight
495
+ single_blocks.11.modulation.lin.bias
496
+ single_blocks.11.modulation.lin.weight
497
+ single_blocks.11.norm.key_norm.scale
498
+ single_blocks.11.norm.query_norm.scale
499
+ single_blocks.12.linear1.bias
500
+ single_blocks.12.linear1.weight
501
+ single_blocks.12.linear2.bias
502
+ single_blocks.12.linear2.weight
503
+ single_blocks.12.modulation.lin.bias
504
+ single_blocks.12.modulation.lin.weight
505
+ single_blocks.12.norm.key_norm.scale
506
+ single_blocks.12.norm.query_norm.scale
507
+ single_blocks.13.linear1.bias
508
+ single_blocks.13.linear1.weight
509
+ single_blocks.13.linear2.bias
510
+ single_blocks.13.linear2.weight
511
+ single_blocks.13.modulation.lin.bias
512
+ single_blocks.13.modulation.lin.weight
513
+ single_blocks.13.norm.key_norm.scale
514
+ single_blocks.13.norm.query_norm.scale
515
+ single_blocks.14.linear1.bias
516
+ single_blocks.14.linear1.weight
517
+ single_blocks.14.linear2.bias
518
+ single_blocks.14.linear2.weight
519
+ single_blocks.14.modulation.lin.bias
520
+ single_blocks.14.modulation.lin.weight
521
+ single_blocks.14.norm.key_norm.scale
522
+ single_blocks.14.norm.query_norm.scale
523
+ single_blocks.15.linear1.bias
524
+ single_blocks.15.linear1.weight
525
+ single_blocks.15.linear2.bias
526
+ single_blocks.15.linear2.weight
527
+ single_blocks.15.modulation.lin.bias
528
+ single_blocks.15.modulation.lin.weight
529
+ single_blocks.15.norm.key_norm.scale
530
+ single_blocks.15.norm.query_norm.scale
531
+ single_blocks.16.linear1.bias
532
+ single_blocks.16.linear1.weight
533
+ single_blocks.16.linear2.bias
534
+ single_blocks.16.linear2.weight
535
+ single_blocks.16.modulation.lin.bias
536
+ single_blocks.16.modulation.lin.weight
537
+ single_blocks.16.norm.key_norm.scale
538
+ single_blocks.16.norm.query_norm.scale
539
+ single_blocks.17.linear1.bias
540
+ single_blocks.17.linear1.weight
541
+ single_blocks.17.linear2.bias
542
+ single_blocks.17.linear2.weight
543
+ single_blocks.17.modulation.lin.bias
544
+ single_blocks.17.modulation.lin.weight
545
+ single_blocks.17.norm.key_norm.scale
546
+ single_blocks.17.norm.query_norm.scale
547
+ single_blocks.18.linear1.bias
548
+ single_blocks.18.linear1.weight
549
+ single_blocks.18.linear2.bias
550
+ single_blocks.18.linear2.weight
551
+ single_blocks.18.modulation.lin.bias
552
+ single_blocks.18.modulation.lin.weight
553
+ single_blocks.18.norm.key_norm.scale
554
+ single_blocks.18.norm.query_norm.scale
555
+ single_blocks.19.linear1.bias
556
+ single_blocks.19.linear1.weight
557
+ single_blocks.19.linear2.bias
558
+ single_blocks.19.linear2.weight
559
+ single_blocks.19.modulation.lin.bias
560
+ single_blocks.19.modulation.lin.weight
561
+ single_blocks.19.norm.key_norm.scale
562
+ single_blocks.19.norm.query_norm.scale
563
+ single_blocks.2.linear1.bias
564
+ single_blocks.2.linear1.weight
565
+ single_blocks.2.linear2.bias
566
+ single_blocks.2.linear2.weight
567
+ single_blocks.2.modulation.lin.bias
568
+ single_blocks.2.modulation.lin.weight
569
+ single_blocks.2.norm.key_norm.scale
570
+ single_blocks.2.norm.query_norm.scale
571
+ single_blocks.20.linear1.bias
572
+ single_blocks.20.linear1.weight
573
+ single_blocks.20.linear2.bias
574
+ single_blocks.20.linear2.weight
575
+ single_blocks.20.modulation.lin.bias
576
+ single_blocks.20.modulation.lin.weight
577
+ single_blocks.20.norm.key_norm.scale
578
+ single_blocks.20.norm.query_norm.scale
579
+ single_blocks.21.linear1.bias
580
+ single_blocks.21.linear1.weight
581
+ single_blocks.21.linear2.bias
582
+ single_blocks.21.linear2.weight
583
+ single_blocks.21.modulation.lin.bias
584
+ single_blocks.21.modulation.lin.weight
585
+ single_blocks.21.norm.key_norm.scale
586
+ single_blocks.21.norm.query_norm.scale
587
+ single_blocks.22.linear1.bias
588
+ single_blocks.22.linear1.weight
589
+ single_blocks.22.linear2.bias
590
+ single_blocks.22.linear2.weight
591
+ single_blocks.22.modulation.lin.bias
592
+ single_blocks.22.modulation.lin.weight
593
+ single_blocks.22.norm.key_norm.scale
594
+ single_blocks.22.norm.query_norm.scale
595
+ single_blocks.23.linear1.bias
596
+ single_blocks.23.linear1.weight
597
+ single_blocks.23.linear2.bias
598
+ single_blocks.23.linear2.weight
599
+ single_blocks.23.modulation.lin.bias
600
+ single_blocks.23.modulation.lin.weight
601
+ single_blocks.23.norm.key_norm.scale
602
+ single_blocks.23.norm.query_norm.scale
603
+ single_blocks.24.linear1.bias
604
+ single_blocks.24.linear1.weight
605
+ single_blocks.24.linear2.bias
606
+ single_blocks.24.linear2.weight
607
+ single_blocks.24.modulation.lin.bias
608
+ single_blocks.24.modulation.lin.weight
609
+ single_blocks.24.norm.key_norm.scale
610
+ single_blocks.24.norm.query_norm.scale
611
+ single_blocks.25.linear1.bias
612
+ single_blocks.25.linear1.weight
613
+ single_blocks.25.linear2.bias
614
+ single_blocks.25.linear2.weight
615
+ single_blocks.25.modulation.lin.bias
616
+ single_blocks.25.modulation.lin.weight
617
+ single_blocks.25.norm.key_norm.scale
618
+ single_blocks.25.norm.query_norm.scale
619
+ single_blocks.26.linear1.bias
620
+ single_blocks.26.linear1.weight
621
+ single_blocks.26.linear2.bias
622
+ single_blocks.26.linear2.weight
623
+ single_blocks.26.modulation.lin.bias
624
+ single_blocks.26.modulation.lin.weight
625
+ single_blocks.26.norm.key_norm.scale
626
+ single_blocks.26.norm.query_norm.scale
627
+ single_blocks.27.linear1.bias
628
+ single_blocks.27.linear1.weight
629
+ single_blocks.27.linear2.bias
630
+ single_blocks.27.linear2.weight
631
+ single_blocks.27.modulation.lin.bias
632
+ single_blocks.27.modulation.lin.weight
633
+ single_blocks.27.norm.key_norm.scale
634
+ single_blocks.27.norm.query_norm.scale
635
+ single_blocks.28.linear1.bias
636
+ single_blocks.28.linear1.weight
637
+ single_blocks.28.linear2.bias
638
+ single_blocks.28.linear2.weight
639
+ single_blocks.28.modulation.lin.bias
640
+ single_blocks.28.modulation.lin.weight
641
+ single_blocks.28.norm.key_norm.scale
642
+ single_blocks.28.norm.query_norm.scale
643
+ single_blocks.29.linear1.bias
644
+ single_blocks.29.linear1.weight
645
+ single_blocks.29.linear2.bias
646
+ single_blocks.29.linear2.weight
647
+ single_blocks.29.modulation.lin.bias
648
+ single_blocks.29.modulation.lin.weight
649
+ single_blocks.29.norm.key_norm.scale
650
+ single_blocks.29.norm.query_norm.scale
651
+ single_blocks.3.linear1.bias
652
+ single_blocks.3.linear1.weight
653
+ single_blocks.3.linear2.bias
654
+ single_blocks.3.linear2.weight
655
+ single_blocks.3.modulation.lin.bias
656
+ single_blocks.3.modulation.lin.weight
657
+ single_blocks.3.norm.key_norm.scale
658
+ single_blocks.3.norm.query_norm.scale
659
+ single_blocks.30.linear1.bias
660
+ single_blocks.30.linear1.weight
661
+ single_blocks.30.linear2.bias
662
+ single_blocks.30.linear2.weight
663
+ single_blocks.30.modulation.lin.bias
664
+ single_blocks.30.modulation.lin.weight
665
+ single_blocks.30.norm.key_norm.scale
666
+ single_blocks.30.norm.query_norm.scale
667
+ single_blocks.31.linear1.bias
668
+ single_blocks.31.linear1.weight
669
+ single_blocks.31.linear2.bias
670
+ single_blocks.31.linear2.weight
671
+ single_blocks.31.modulation.lin.bias
672
+ single_blocks.31.modulation.lin.weight
673
+ single_blocks.31.norm.key_norm.scale
674
+ single_blocks.31.norm.query_norm.scale
675
+ single_blocks.32.linear1.bias
676
+ single_blocks.32.linear1.weight
677
+ single_blocks.32.linear2.bias
678
+ single_blocks.32.linear2.weight
679
+ single_blocks.32.modulation.lin.bias
680
+ single_blocks.32.modulation.lin.weight
681
+ single_blocks.32.norm.key_norm.scale
682
+ single_blocks.32.norm.query_norm.scale
683
+ single_blocks.33.linear1.bias
684
+ single_blocks.33.linear1.weight
685
+ single_blocks.33.linear2.bias
686
+ single_blocks.33.linear2.weight
687
+ single_blocks.33.modulation.lin.bias
688
+ single_blocks.33.modulation.lin.weight
689
+ single_blocks.33.norm.key_norm.scale
690
+ single_blocks.33.norm.query_norm.scale
691
+ single_blocks.34.linear1.bias
692
+ single_blocks.34.linear1.weight
693
+ single_blocks.34.linear2.bias
694
+ single_blocks.34.linear2.weight
695
+ single_blocks.34.modulation.lin.bias
696
+ single_blocks.34.modulation.lin.weight
697
+ single_blocks.34.norm.key_norm.scale
698
+ single_blocks.34.norm.query_norm.scale
699
+ single_blocks.35.linear1.bias
700
+ single_blocks.35.linear1.weight
701
+ single_blocks.35.linear2.bias
702
+ single_blocks.35.linear2.weight
703
+ single_blocks.35.modulation.lin.bias
704
+ single_blocks.35.modulation.lin.weight
705
+ single_blocks.35.norm.key_norm.scale
706
+ single_blocks.35.norm.query_norm.scale
707
+ single_blocks.36.linear1.bias
708
+ single_blocks.36.linear1.weight
709
+ single_blocks.36.linear2.bias
710
+ single_blocks.36.linear2.weight
711
+ single_blocks.36.modulation.lin.bias
712
+ single_blocks.36.modulation.lin.weight
713
+ single_blocks.36.norm.key_norm.scale
714
+ single_blocks.36.norm.query_norm.scale
715
+ single_blocks.37.linear1.bias
716
+ single_blocks.37.linear1.weight
717
+ single_blocks.37.linear2.bias
718
+ single_blocks.37.linear2.weight
719
+ single_blocks.37.modulation.lin.bias
720
+ single_blocks.37.modulation.lin.weight
721
+ single_blocks.37.norm.key_norm.scale
722
+ single_blocks.37.norm.query_norm.scale
723
+ single_blocks.4.linear1.bias
724
+ single_blocks.4.linear1.weight
725
+ single_blocks.4.linear2.bias
726
+ single_blocks.4.linear2.weight
727
+ single_blocks.4.modulation.lin.bias
728
+ single_blocks.4.modulation.lin.weight
729
+ single_blocks.4.norm.key_norm.scale
730
+ single_blocks.4.norm.query_norm.scale
731
+ single_blocks.5.linear1.bias
732
+ single_blocks.5.linear1.weight
733
+ single_blocks.5.linear2.bias
734
+ single_blocks.5.linear2.weight
735
+ single_blocks.5.modulation.lin.bias
736
+ single_blocks.5.modulation.lin.weight
737
+ single_blocks.5.norm.key_norm.scale
738
+ single_blocks.5.norm.query_norm.scale
739
+ single_blocks.6.linear1.bias
740
+ single_blocks.6.linear1.weight
741
+ single_blocks.6.linear2.bias
742
+ single_blocks.6.linear2.weight
743
+ single_blocks.6.modulation.lin.bias
744
+ single_blocks.6.modulation.lin.weight
745
+ single_blocks.6.norm.key_norm.scale
746
+ single_blocks.6.norm.query_norm.scale
747
+ single_blocks.7.linear1.bias
748
+ single_blocks.7.linear1.weight
749
+ single_blocks.7.linear2.bias
750
+ single_blocks.7.linear2.weight
751
+ single_blocks.7.modulation.lin.bias
752
+ single_blocks.7.modulation.lin.weight
753
+ single_blocks.7.norm.key_norm.scale
754
+ single_blocks.7.norm.query_norm.scale
755
+ single_blocks.8.linear1.bias
756
+ single_blocks.8.linear1.weight
757
+ single_blocks.8.linear2.bias
758
+ single_blocks.8.linear2.weight
759
+ single_blocks.8.modulation.lin.bias
760
+ single_blocks.8.modulation.lin.weight
761
+ single_blocks.8.norm.key_norm.scale
762
+ single_blocks.8.norm.query_norm.scale
763
+ single_blocks.9.linear1.bias
764
+ single_blocks.9.linear1.weight
765
+ single_blocks.9.linear2.bias
766
+ single_blocks.9.linear2.weight
767
+ single_blocks.9.modulation.lin.bias
768
+ single_blocks.9.modulation.lin.weight
769
+ single_blocks.9.norm.key_norm.scale
770
+ single_blocks.9.norm.query_norm.scale
771
+ time_in.in_layer.bias
772
+ time_in.in_layer.weight
773
+ time_in.out_layer.bias
774
+ time_in.out_layer.weight
775
+ txt_in.bias
776
+ txt_in.weight
777
+ vector_in.in_layer.bias
778
+ vector_in.in_layer.weight
779
+ vector_in.out_layer.bias
780
+ vector_in.out_layer.weight
781
+ text_encoders.clip_l.logit_scale
782
+ embeddings.position_embedding.weight
783
+ embeddings.token_embedding.weight
784
+ encoder.layers.0.layer_norm1.bias
785
+ encoder.layers.0.layer_norm1.weight
786
+ encoder.layers.0.layer_norm2.bias
787
+ encoder.layers.0.layer_norm2.weight
788
+ encoder.layers.0.mlp.fc1.bias
789
+ encoder.layers.0.mlp.fc1.weight
790
+ encoder.layers.0.mlp.fc2.bias
791
+ encoder.layers.0.mlp.fc2.weight
792
+ encoder.layers.0.self_attn.k_proj.bias
793
+ encoder.layers.0.self_attn.k_proj.weight
794
+ encoder.layers.0.self_attn.out_proj.bias
795
+ encoder.layers.0.self_attn.out_proj.weight
796
+ encoder.layers.0.self_attn.q_proj.bias
797
+ encoder.layers.0.self_attn.q_proj.weight
798
+ encoder.layers.0.self_attn.v_proj.bias
799
+ encoder.layers.0.self_attn.v_proj.weight
800
+ encoder.layers.1.layer_norm1.bias
801
+ encoder.layers.1.layer_norm1.weight
802
+ encoder.layers.1.layer_norm2.bias
803
+ encoder.layers.1.layer_norm2.weight
804
+ encoder.layers.1.mlp.fc1.bias
805
+ encoder.layers.1.mlp.fc1.weight
806
+ encoder.layers.1.mlp.fc2.bias
807
+ encoder.layers.1.mlp.fc2.weight
808
+ encoder.layers.1.self_attn.k_proj.bias
809
+ encoder.layers.1.self_attn.k_proj.weight
810
+ encoder.layers.1.self_attn.out_proj.bias
811
+ encoder.layers.1.self_attn.out_proj.weight
812
+ encoder.layers.1.self_attn.q_proj.bias
813
+ encoder.layers.1.self_attn.q_proj.weight
814
+ encoder.layers.1.self_attn.v_proj.bias
815
+ encoder.layers.1.self_attn.v_proj.weight
816
+ encoder.layers.10.layer_norm1.bias
817
+ encoder.layers.10.layer_norm1.weight
818
+ encoder.layers.10.layer_norm2.bias
819
+ encoder.layers.10.layer_norm2.weight
820
+ encoder.layers.10.mlp.fc1.bias
821
+ encoder.layers.10.mlp.fc1.weight
822
+ encoder.layers.10.mlp.fc2.bias
823
+ encoder.layers.10.mlp.fc2.weight
824
+ encoder.layers.10.self_attn.k_proj.bias
825
+ encoder.layers.10.self_attn.k_proj.weight
826
+ encoder.layers.10.self_attn.out_proj.bias
827
+ encoder.layers.10.self_attn.out_proj.weight
828
+ encoder.layers.10.self_attn.q_proj.bias
829
+ encoder.layers.10.self_attn.q_proj.weight
830
+ encoder.layers.10.self_attn.v_proj.bias
831
+ encoder.layers.10.self_attn.v_proj.weight
832
+ encoder.layers.11.layer_norm1.bias
833
+ encoder.layers.11.layer_norm1.weight
834
+ encoder.layers.11.layer_norm2.bias
835
+ encoder.layers.11.layer_norm2.weight
836
+ encoder.layers.11.mlp.fc1.bias
837
+ encoder.layers.11.mlp.fc1.weight
838
+ encoder.layers.11.mlp.fc2.bias
839
+ encoder.layers.11.mlp.fc2.weight
840
+ encoder.layers.11.self_attn.k_proj.bias
841
+ encoder.layers.11.self_attn.k_proj.weight
842
+ encoder.layers.11.self_attn.out_proj.bias
843
+ encoder.layers.11.self_attn.out_proj.weight
844
+ encoder.layers.11.self_attn.q_proj.bias
845
+ encoder.layers.11.self_attn.q_proj.weight
846
+ encoder.layers.11.self_attn.v_proj.bias
847
+ encoder.layers.11.self_attn.v_proj.weight
848
+ encoder.layers.2.layer_norm1.bias
849
+ encoder.layers.2.layer_norm1.weight
850
+ encoder.layers.2.layer_norm2.bias
851
+ encoder.layers.2.layer_norm2.weight
852
+ encoder.layers.2.mlp.fc1.bias
853
+ encoder.layers.2.mlp.fc1.weight
854
+ encoder.layers.2.mlp.fc2.bias
855
+ encoder.layers.2.mlp.fc2.weight
856
+ encoder.layers.2.self_attn.k_proj.bias
857
+ encoder.layers.2.self_attn.k_proj.weight
858
+ encoder.layers.2.self_attn.out_proj.bias
859
+ encoder.layers.2.self_attn.out_proj.weight
860
+ encoder.layers.2.self_attn.q_proj.bias
861
+ encoder.layers.2.self_attn.q_proj.weight
862
+ encoder.layers.2.self_attn.v_proj.bias
863
+ encoder.layers.2.self_attn.v_proj.weight
864
+ encoder.layers.3.layer_norm1.bias
865
+ encoder.layers.3.layer_norm1.weight
866
+ encoder.layers.3.layer_norm2.bias
867
+ encoder.layers.3.layer_norm2.weight
868
+ encoder.layers.3.mlp.fc1.bias
869
+ encoder.layers.3.mlp.fc1.weight
870
+ encoder.layers.3.mlp.fc2.bias
871
+ encoder.layers.3.mlp.fc2.weight
872
+ encoder.layers.3.self_attn.k_proj.bias
873
+ encoder.layers.3.self_attn.k_proj.weight
874
+ encoder.layers.3.self_attn.out_proj.bias
875
+ encoder.layers.3.self_attn.out_proj.weight
876
+ encoder.layers.3.self_attn.q_proj.bias
877
+ encoder.layers.3.self_attn.q_proj.weight
878
+ encoder.layers.3.self_attn.v_proj.bias
879
+ encoder.layers.3.self_attn.v_proj.weight
880
+ encoder.layers.4.layer_norm1.bias
881
+ encoder.layers.4.layer_norm1.weight
882
+ encoder.layers.4.layer_norm2.bias
883
+ encoder.layers.4.layer_norm2.weight
884
+ encoder.layers.4.mlp.fc1.bias
885
+ encoder.layers.4.mlp.fc1.weight
886
+ encoder.layers.4.mlp.fc2.bias
887
+ encoder.layers.4.mlp.fc2.weight
888
+ encoder.layers.4.self_attn.k_proj.bias
889
+ encoder.layers.4.self_attn.k_proj.weight
890
+ encoder.layers.4.self_attn.out_proj.bias
891
+ encoder.layers.4.self_attn.out_proj.weight
892
+ encoder.layers.4.self_attn.q_proj.bias
893
+ encoder.layers.4.self_attn.q_proj.weight
894
+ encoder.layers.4.self_attn.v_proj.bias
895
+ encoder.layers.4.self_attn.v_proj.weight
896
+ encoder.layers.5.layer_norm1.bias
897
+ encoder.layers.5.layer_norm1.weight
898
+ encoder.layers.5.layer_norm2.bias
899
+ encoder.layers.5.layer_norm2.weight
900
+ encoder.layers.5.mlp.fc1.bias
901
+ encoder.layers.5.mlp.fc1.weight
902
+ encoder.layers.5.mlp.fc2.bias
903
+ encoder.layers.5.mlp.fc2.weight
904
+ encoder.layers.5.self_attn.k_proj.bias
905
+ encoder.layers.5.self_attn.k_proj.weight
906
+ encoder.layers.5.self_attn.out_proj.bias
907
+ encoder.layers.5.self_attn.out_proj.weight
908
+ encoder.layers.5.self_attn.q_proj.bias
909
+ encoder.layers.5.self_attn.q_proj.weight
910
+ encoder.layers.5.self_attn.v_proj.bias
911
+ encoder.layers.5.self_attn.v_proj.weight
912
+ encoder.layers.6.layer_norm1.bias
913
+ encoder.layers.6.layer_norm1.weight
914
+ encoder.layers.6.layer_norm2.bias
915
+ encoder.layers.6.layer_norm2.weight
916
+ encoder.layers.6.mlp.fc1.bias
917
+ encoder.layers.6.mlp.fc1.weight
918
+ encoder.layers.6.mlp.fc2.bias
919
+ encoder.layers.6.mlp.fc2.weight
920
+ encoder.layers.6.self_attn.k_proj.bias
921
+ encoder.layers.6.self_attn.k_proj.weight
922
+ encoder.layers.6.self_attn.out_proj.bias
923
+ encoder.layers.6.self_attn.out_proj.weight
924
+ encoder.layers.6.self_attn.q_proj.bias
925
+ encoder.layers.6.self_attn.q_proj.weight
926
+ encoder.layers.6.self_attn.v_proj.bias
927
+ encoder.layers.6.self_attn.v_proj.weight
928
+ encoder.layers.7.layer_norm1.bias
929
+ encoder.layers.7.layer_norm1.weight
930
+ encoder.layers.7.layer_norm2.bias
931
+ encoder.layers.7.layer_norm2.weight
932
+ encoder.layers.7.mlp.fc1.bias
933
+ encoder.layers.7.mlp.fc1.weight
934
+ encoder.layers.7.mlp.fc2.bias
935
+ encoder.layers.7.mlp.fc2.weight
936
+ encoder.layers.7.self_attn.k_proj.bias
937
+ encoder.layers.7.self_attn.k_proj.weight
938
+ encoder.layers.7.self_attn.out_proj.bias
939
+ encoder.layers.7.self_attn.out_proj.weight
940
+ encoder.layers.7.self_attn.q_proj.bias
941
+ encoder.layers.7.self_attn.q_proj.weight
942
+ encoder.layers.7.self_attn.v_proj.bias
943
+ encoder.layers.7.self_attn.v_proj.weight
944
+ encoder.layers.8.layer_norm1.bias
945
+ encoder.layers.8.layer_norm1.weight
946
+ encoder.layers.8.layer_norm2.bias
947
+ encoder.layers.8.layer_norm2.weight
948
+ encoder.layers.8.mlp.fc1.bias
949
+ encoder.layers.8.mlp.fc1.weight
950
+ encoder.layers.8.mlp.fc2.bias
951
+ encoder.layers.8.mlp.fc2.weight
952
+ encoder.layers.8.self_attn.k_proj.bias
953
+ encoder.layers.8.self_attn.k_proj.weight
954
+ encoder.layers.8.self_attn.out_proj.bias
955
+ encoder.layers.8.self_attn.out_proj.weight
956
+ encoder.layers.8.self_attn.q_proj.bias
957
+ encoder.layers.8.self_attn.q_proj.weight
958
+ encoder.layers.8.self_attn.v_proj.bias
959
+ encoder.layers.8.self_attn.v_proj.weight
960
+ encoder.layers.9.layer_norm1.bias
961
+ encoder.layers.9.layer_norm1.weight
962
+ encoder.layers.9.layer_norm2.bias
963
+ encoder.layers.9.layer_norm2.weight
964
+ encoder.layers.9.mlp.fc1.bias
965
+ encoder.layers.9.mlp.fc1.weight
966
+ encoder.layers.9.mlp.fc2.bias
967
+ encoder.layers.9.mlp.fc2.weight
968
+ encoder.layers.9.self_attn.k_proj.bias
969
+ encoder.layers.9.self_attn.k_proj.weight
970
+ encoder.layers.9.self_attn.out_proj.bias
971
+ encoder.layers.9.self_attn.out_proj.weight
972
+ encoder.layers.9.self_attn.q_proj.bias
973
+ encoder.layers.9.self_attn.q_proj.weight
974
+ encoder.layers.9.self_attn.v_proj.bias
975
+ encoder.layers.9.self_attn.v_proj.weight
976
+ final_layer_norm.bias
977
+ final_layer_norm.weight
978
+ text_encoders.clip_l.transformer.text_projection.weight
979
+ text_encoders.t5xxl.logit_scale
980
+ encoder.block.0.layer.0.SelfAttention.k.weight
981
+ encoder.block.0.layer.0.SelfAttention.o.weight
982
+ encoder.block.0.layer.0.SelfAttention.q.weight
983
+ encoder.block.0.layer.0.SelfAttention.relative_attention_bias.weight
984
+ encoder.block.0.layer.0.SelfAttention.v.weight
985
+ encoder.block.0.layer.0.layer_norm.weight
986
+ encoder.block.0.layer.1.DenseReluDense.wi_0.weight
987
+ encoder.block.0.layer.1.DenseReluDense.wi_1.weight
988
+ encoder.block.0.layer.1.DenseReluDense.wo.weight
989
+ encoder.block.0.layer.1.layer_norm.weight
990
+ encoder.block.1.layer.0.SelfAttention.k.weight
991
+ encoder.block.1.layer.0.SelfAttention.o.weight
992
+ encoder.block.1.layer.0.SelfAttention.q.weight
993
+ encoder.block.1.layer.0.SelfAttention.v.weight
994
+ encoder.block.1.layer.0.layer_norm.weight
995
+ encoder.block.1.layer.1.DenseReluDense.wi_0.weight
996
+ encoder.block.1.layer.1.DenseReluDense.wi_1.weight
997
+ encoder.block.1.layer.1.DenseReluDense.wo.weight
998
+ encoder.block.1.layer.1.layer_norm.weight
999
+ encoder.block.10.layer.0.SelfAttention.k.weight
1000
+ encoder.block.10.layer.0.SelfAttention.o.weight
1001
+ encoder.block.10.layer.0.SelfAttention.q.weight
1002
+ encoder.block.10.layer.0.SelfAttention.v.weight
1003
+ encoder.block.10.layer.0.layer_norm.weight
1004
+ encoder.block.10.layer.1.DenseReluDense.wi_0.weight
1005
+ encoder.block.10.layer.1.DenseReluDense.wi_1.weight
1006
+ encoder.block.10.layer.1.DenseReluDense.wo.weight
1007
+ encoder.block.10.layer.1.layer_norm.weight
1008
+ encoder.block.11.layer.0.SelfAttention.k.weight
1009
+ encoder.block.11.layer.0.SelfAttention.o.weight
1010
+ encoder.block.11.layer.0.SelfAttention.q.weight
1011
+ encoder.block.11.layer.0.SelfAttention.v.weight
1012
+ encoder.block.11.layer.0.layer_norm.weight
1013
+ encoder.block.11.layer.1.DenseReluDense.wi_0.weight
1014
+ encoder.block.11.layer.1.DenseReluDense.wi_1.weight
1015
+ encoder.block.11.layer.1.DenseReluDense.wo.weight
1016
+ encoder.block.11.layer.1.layer_norm.weight
1017
+ encoder.block.12.layer.0.SelfAttention.k.weight
1018
+ encoder.block.12.layer.0.SelfAttention.o.weight
1019
+ encoder.block.12.layer.0.SelfAttention.q.weight
1020
+ encoder.block.12.layer.0.SelfAttention.v.weight
1021
+ encoder.block.12.layer.0.layer_norm.weight
1022
+ encoder.block.12.layer.1.DenseReluDense.wi_0.weight
1023
+ encoder.block.12.layer.1.DenseReluDense.wi_1.weight
1024
+ encoder.block.12.layer.1.DenseReluDense.wo.weight
1025
+ encoder.block.12.layer.1.layer_norm.weight
1026
+ encoder.block.13.layer.0.SelfAttention.k.weight
1027
+ encoder.block.13.layer.0.SelfAttention.o.weight
1028
+ encoder.block.13.layer.0.SelfAttention.q.weight
1029
+ encoder.block.13.layer.0.SelfAttention.v.weight
1030
+ encoder.block.13.layer.0.layer_norm.weight
1031
+ encoder.block.13.layer.1.DenseReluDense.wi_0.weight
1032
+ encoder.block.13.layer.1.DenseReluDense.wi_1.weight
1033
+ encoder.block.13.layer.1.DenseReluDense.wo.weight
1034
+ encoder.block.13.layer.1.layer_norm.weight
1035
+ encoder.block.14.layer.0.SelfAttention.k.weight
1036
+ encoder.block.14.layer.0.SelfAttention.o.weight
1037
+ encoder.block.14.layer.0.SelfAttention.q.weight
1038
+ encoder.block.14.layer.0.SelfAttention.v.weight
1039
+ encoder.block.14.layer.0.layer_norm.weight
1040
+ encoder.block.14.layer.1.DenseReluDense.wi_0.weight
1041
+ encoder.block.14.layer.1.DenseReluDense.wi_1.weight
1042
+ encoder.block.14.layer.1.DenseReluDense.wo.weight
1043
+ encoder.block.14.layer.1.layer_norm.weight
1044
+ encoder.block.15.layer.0.SelfAttention.k.weight
1045
+ encoder.block.15.layer.0.SelfAttention.o.weight
1046
+ encoder.block.15.layer.0.SelfAttention.q.weight
1047
+ encoder.block.15.layer.0.SelfAttention.v.weight
1048
+ encoder.block.15.layer.0.layer_norm.weight
1049
+ encoder.block.15.layer.1.DenseReluDense.wi_0.weight
1050
+ encoder.block.15.layer.1.DenseReluDense.wi_1.weight
1051
+ encoder.block.15.layer.1.DenseReluDense.wo.weight
1052
+ encoder.block.15.layer.1.layer_norm.weight
1053
+ encoder.block.16.layer.0.SelfAttention.k.weight
1054
+ encoder.block.16.layer.0.SelfAttention.o.weight
1055
+ encoder.block.16.layer.0.SelfAttention.q.weight
1056
+ encoder.block.16.layer.0.SelfAttention.v.weight
1057
+ encoder.block.16.layer.0.layer_norm.weight
1058
+ encoder.block.16.layer.1.DenseReluDense.wi_0.weight
1059
+ encoder.block.16.layer.1.DenseReluDense.wi_1.weight
1060
+ encoder.block.16.layer.1.DenseReluDense.wo.weight
1061
+ encoder.block.16.layer.1.layer_norm.weight
1062
+ encoder.block.17.layer.0.SelfAttention.k.weight
1063
+ encoder.block.17.layer.0.SelfAttention.o.weight
1064
+ encoder.block.17.layer.0.SelfAttention.q.weight
1065
+ encoder.block.17.layer.0.SelfAttention.v.weight
1066
+ encoder.block.17.layer.0.layer_norm.weight
1067
+ encoder.block.17.layer.1.DenseReluDense.wi_0.weight
1068
+ encoder.block.17.layer.1.DenseReluDense.wi_1.weight
1069
+ encoder.block.17.layer.1.DenseReluDense.wo.weight
1070
+ encoder.block.17.layer.1.layer_norm.weight
1071
+ encoder.block.18.layer.0.SelfAttention.k.weight
1072
+ encoder.block.18.layer.0.SelfAttention.o.weight
1073
+ encoder.block.18.layer.0.SelfAttention.q.weight
1074
+ encoder.block.18.layer.0.SelfAttention.v.weight
1075
+ encoder.block.18.layer.0.layer_norm.weight
1076
+ encoder.block.18.layer.1.DenseReluDense.wi_0.weight
1077
+ encoder.block.18.layer.1.DenseReluDense.wi_1.weight
1078
+ encoder.block.18.layer.1.DenseReluDense.wo.weight
1079
+ encoder.block.18.layer.1.layer_norm.weight
1080
+ encoder.block.19.layer.0.SelfAttention.k.weight
1081
+ encoder.block.19.layer.0.SelfAttention.o.weight
1082
+ encoder.block.19.layer.0.SelfAttention.q.weight
1083
+ encoder.block.19.layer.0.SelfAttention.v.weight
1084
+ encoder.block.19.layer.0.layer_norm.weight
1085
+ encoder.block.19.layer.1.DenseReluDense.wi_0.weight
1086
+ encoder.block.19.layer.1.DenseReluDense.wi_1.weight
1087
+ encoder.block.19.layer.1.DenseReluDense.wo.weight
1088
+ encoder.block.19.layer.1.layer_norm.weight
1089
+ encoder.block.2.layer.0.SelfAttention.k.weight
1090
+ encoder.block.2.layer.0.SelfAttention.o.weight
1091
+ encoder.block.2.layer.0.SelfAttention.q.weight
1092
+ encoder.block.2.layer.0.SelfAttention.v.weight
1093
+ encoder.block.2.layer.0.layer_norm.weight
1094
+ encoder.block.2.layer.1.DenseReluDense.wi_0.weight
1095
+ encoder.block.2.layer.1.DenseReluDense.wi_1.weight
1096
+ encoder.block.2.layer.1.DenseReluDense.wo.weight
1097
+ encoder.block.2.layer.1.layer_norm.weight
1098
+ encoder.block.20.layer.0.SelfAttention.k.weight
1099
+ encoder.block.20.layer.0.SelfAttention.o.weight
1100
+ encoder.block.20.layer.0.SelfAttention.q.weight
1101
+ encoder.block.20.layer.0.SelfAttention.v.weight
1102
+ encoder.block.20.layer.0.layer_norm.weight
1103
+ encoder.block.20.layer.1.DenseReluDense.wi_0.weight
1104
+ encoder.block.20.layer.1.DenseReluDense.wi_1.weight
1105
+ encoder.block.20.layer.1.DenseReluDense.wo.weight
1106
+ encoder.block.20.layer.1.layer_norm.weight
1107
+ encoder.block.21.layer.0.SelfAttention.k.weight
1108
+ encoder.block.21.layer.0.SelfAttention.o.weight
1109
+ encoder.block.21.layer.0.SelfAttention.q.weight
1110
+ encoder.block.21.layer.0.SelfAttention.v.weight
1111
+ encoder.block.21.layer.0.layer_norm.weight
1112
+ encoder.block.21.layer.1.DenseReluDense.wi_0.weight
1113
+ encoder.block.21.layer.1.DenseReluDense.wi_1.weight
1114
+ encoder.block.21.layer.1.DenseReluDense.wo.weight
1115
+ encoder.block.21.layer.1.layer_norm.weight
1116
+ encoder.block.22.layer.0.SelfAttention.k.weight
1117
+ encoder.block.22.layer.0.SelfAttention.o.weight
1118
+ encoder.block.22.layer.0.SelfAttention.q.weight
1119
+ encoder.block.22.layer.0.SelfAttention.v.weight
1120
+ encoder.block.22.layer.0.layer_norm.weight
1121
+ encoder.block.22.layer.1.DenseReluDense.wi_0.weight
1122
+ encoder.block.22.layer.1.DenseReluDense.wi_1.weight
1123
+ encoder.block.22.layer.1.DenseReluDense.wo.weight
1124
+ encoder.block.22.layer.1.layer_norm.weight
1125
+ encoder.block.23.layer.0.SelfAttention.k.weight
1126
+ encoder.block.23.layer.0.SelfAttention.o.weight
1127
+ encoder.block.23.layer.0.SelfAttention.q.weight
1128
+ encoder.block.23.layer.0.SelfAttention.v.weight
1129
+ encoder.block.23.layer.0.layer_norm.weight
1130
+ encoder.block.23.layer.1.DenseReluDense.wi_0.weight
1131
+ encoder.block.23.layer.1.DenseReluDense.wi_1.weight
1132
+ encoder.block.23.layer.1.DenseReluDense.wo.weight
1133
+ encoder.block.23.layer.1.layer_norm.weight
1134
+ encoder.block.3.layer.0.SelfAttention.k.weight
1135
+ encoder.block.3.layer.0.SelfAttention.o.weight
1136
+ encoder.block.3.layer.0.SelfAttention.q.weight
1137
+ encoder.block.3.layer.0.SelfAttention.v.weight
1138
+ encoder.block.3.layer.0.layer_norm.weight
1139
+ encoder.block.3.layer.1.DenseReluDense.wi_0.weight
1140
+ encoder.block.3.layer.1.DenseReluDense.wi_1.weight
1141
+ encoder.block.3.layer.1.DenseReluDense.wo.weight
1142
+ encoder.block.3.layer.1.layer_norm.weight
1143
+ encoder.block.4.layer.0.SelfAttention.k.weight
1144
+ encoder.block.4.layer.0.SelfAttention.o.weight
1145
+ encoder.block.4.layer.0.SelfAttention.q.weight
1146
+ encoder.block.4.layer.0.SelfAttention.v.weight
1147
+ encoder.block.4.layer.0.layer_norm.weight
1148
+ encoder.block.4.layer.1.DenseReluDense.wi_0.weight
1149
+ encoder.block.4.layer.1.DenseReluDense.wi_1.weight
1150
+ encoder.block.4.layer.1.DenseReluDense.wo.weight
1151
+ encoder.block.4.layer.1.layer_norm.weight
1152
+ encoder.block.5.layer.0.SelfAttention.k.weight
1153
+ encoder.block.5.layer.0.SelfAttention.o.weight
1154
+ encoder.block.5.layer.0.SelfAttention.q.weight
1155
+ encoder.block.5.layer.0.SelfAttention.v.weight
1156
+ encoder.block.5.layer.0.layer_norm.weight
1157
+ encoder.block.5.layer.1.DenseReluDense.wi_0.weight
1158
+ encoder.block.5.layer.1.DenseReluDense.wi_1.weight
1159
+ encoder.block.5.layer.1.DenseReluDense.wo.weight
1160
+ encoder.block.5.layer.1.layer_norm.weight
1161
+ encoder.block.6.layer.0.SelfAttention.k.weight
1162
+ encoder.block.6.layer.0.SelfAttention.o.weight
1163
+ encoder.block.6.layer.0.SelfAttention.q.weight
1164
+ encoder.block.6.layer.0.SelfAttention.v.weight
1165
+ encoder.block.6.layer.0.layer_norm.weight
1166
+ encoder.block.6.layer.1.DenseReluDense.wi_0.weight
1167
+ encoder.block.6.layer.1.DenseReluDense.wi_1.weight
1168
+ encoder.block.6.layer.1.DenseReluDense.wo.weight
1169
+ encoder.block.6.layer.1.layer_norm.weight
1170
+ encoder.block.7.layer.0.SelfAttention.k.weight
1171
+ encoder.block.7.layer.0.SelfAttention.o.weight
1172
+ encoder.block.7.layer.0.SelfAttention.q.weight
1173
+ encoder.block.7.layer.0.SelfAttention.v.weight
1174
+ encoder.block.7.layer.0.layer_norm.weight
1175
+ encoder.block.7.layer.1.DenseReluDense.wi_0.weight
1176
+ encoder.block.7.layer.1.DenseReluDense.wi_1.weight
1177
+ encoder.block.7.layer.1.DenseReluDense.wo.weight
1178
+ encoder.block.7.layer.1.layer_norm.weight
1179
+ encoder.block.8.layer.0.SelfAttention.k.weight
1180
+ encoder.block.8.layer.0.SelfAttention.o.weight
1181
+ encoder.block.8.layer.0.SelfAttention.q.weight
1182
+ encoder.block.8.layer.0.SelfAttention.v.weight
1183
+ encoder.block.8.layer.0.layer_norm.weight
1184
+ encoder.block.8.layer.1.DenseReluDense.wi_0.weight
1185
+ encoder.block.8.layer.1.DenseReluDense.wi_1.weight
1186
+ encoder.block.8.layer.1.DenseReluDense.wo.weight
1187
+ encoder.block.8.layer.1.layer_norm.weight
1188
+ encoder.block.9.layer.0.SelfAttention.k.weight
1189
+ encoder.block.9.layer.0.SelfAttention.o.weight
1190
+ encoder.block.9.layer.0.SelfAttention.q.weight
1191
+ encoder.block.9.layer.0.SelfAttention.v.weight
1192
+ encoder.block.9.layer.0.layer_norm.weight
1193
+ encoder.block.9.layer.1.DenseReluDense.wi_0.weight
1194
+ encoder.block.9.layer.1.DenseReluDense.wi_1.weight
1195
+ encoder.block.9.layer.1.DenseReluDense.wo.weight
1196
+ encoder.block.9.layer.1.layer_norm.weight
1197
+ encoder.final_layer_norm.weight
1198
+ shared.weight
1199
+ decoder.conv_in.bias
1200
+ decoder.conv_in.weight
1201
+ decoder.conv_out.bias
1202
+ decoder.conv_out.weight
1203
+ decoder.mid.attn_1.k.bias
1204
+ decoder.mid.attn_1.k.weight
1205
+ decoder.mid.attn_1.norm.bias
1206
+ decoder.mid.attn_1.norm.weight
1207
+ decoder.mid.attn_1.proj_out.bias
1208
+ decoder.mid.attn_1.proj_out.weight
1209
+ decoder.mid.attn_1.q.bias
1210
+ decoder.mid.attn_1.q.weight
1211
+ decoder.mid.attn_1.v.bias
1212
+ decoder.mid.attn_1.v.weight
1213
+ decoder.mid.block_1.conv1.bias
1214
+ decoder.mid.block_1.conv1.weight
1215
+ decoder.mid.block_1.conv2.bias
1216
+ decoder.mid.block_1.conv2.weight
1217
+ decoder.mid.block_1.norm1.bias
1218
+ decoder.mid.block_1.norm1.weight
1219
+ decoder.mid.block_1.norm2.bias
1220
+ decoder.mid.block_1.norm2.weight
1221
+ decoder.mid.block_2.conv1.bias
1222
+ decoder.mid.block_2.conv1.weight
1223
+ decoder.mid.block_2.conv2.bias
1224
+ decoder.mid.block_2.conv2.weight
1225
+ decoder.mid.block_2.norm1.bias
1226
+ decoder.mid.block_2.norm1.weight
1227
+ decoder.mid.block_2.norm2.bias
1228
+ decoder.mid.block_2.norm2.weight
1229
+ decoder.norm_out.bias
1230
+ decoder.norm_out.weight
1231
+ decoder.up.0.block.0.conv1.bias
1232
+ decoder.up.0.block.0.conv1.weight
1233
+ decoder.up.0.block.0.conv2.bias
1234
+ decoder.up.0.block.0.conv2.weight
1235
+ decoder.up.0.block.0.nin_shortcut.bias
1236
+ decoder.up.0.block.0.nin_shortcut.weight
1237
+ decoder.up.0.block.0.norm1.bias
1238
+ decoder.up.0.block.0.norm1.weight
1239
+ decoder.up.0.block.0.norm2.bias
1240
+ decoder.up.0.block.0.norm2.weight
1241
+ decoder.up.0.block.1.conv1.bias
1242
+ decoder.up.0.block.1.conv1.weight
1243
+ decoder.up.0.block.1.conv2.bias
1244
+ decoder.up.0.block.1.conv2.weight
1245
+ decoder.up.0.block.1.norm1.bias
1246
+ decoder.up.0.block.1.norm1.weight
1247
+ decoder.up.0.block.1.norm2.bias
1248
+ decoder.up.0.block.1.norm2.weight
1249
+ decoder.up.0.block.2.conv1.bias
1250
+ decoder.up.0.block.2.conv1.weight
1251
+ decoder.up.0.block.2.conv2.bias
1252
+ decoder.up.0.block.2.conv2.weight
1253
+ decoder.up.0.block.2.norm1.bias
1254
+ decoder.up.0.block.2.norm1.weight
1255
+ decoder.up.0.block.2.norm2.bias
1256
+ decoder.up.0.block.2.norm2.weight
1257
+ decoder.up.1.block.0.conv1.bias
1258
+ decoder.up.1.block.0.conv1.weight
1259
+ decoder.up.1.block.0.conv2.bias
1260
+ decoder.up.1.block.0.conv2.weight
1261
+ decoder.up.1.block.0.nin_shortcut.bias
1262
+ decoder.up.1.block.0.nin_shortcut.weight
1263
+ decoder.up.1.block.0.norm1.bias
1264
+ decoder.up.1.block.0.norm1.weight
1265
+ decoder.up.1.block.0.norm2.bias
1266
+ decoder.up.1.block.0.norm2.weight
1267
+ decoder.up.1.block.1.conv1.bias
1268
+ decoder.up.1.block.1.conv1.weight
1269
+ decoder.up.1.block.1.conv2.bias
1270
+ decoder.up.1.block.1.conv2.weight
1271
+ decoder.up.1.block.1.norm1.bias
1272
+ decoder.up.1.block.1.norm1.weight
1273
+ decoder.up.1.block.1.norm2.bias
1274
+ decoder.up.1.block.1.norm2.weight
1275
+ decoder.up.1.block.2.conv1.bias
1276
+ decoder.up.1.block.2.conv1.weight
1277
+ decoder.up.1.block.2.conv2.bias
1278
+ decoder.up.1.block.2.conv2.weight
1279
+ decoder.up.1.block.2.norm1.bias
1280
+ decoder.up.1.block.2.norm1.weight
1281
+ decoder.up.1.block.2.norm2.bias
1282
+ decoder.up.1.block.2.norm2.weight
1283
+ decoder.up.1.upsample.conv.bias
1284
+ decoder.up.1.upsample.conv.weight
1285
+ decoder.up.2.block.0.conv1.bias
1286
+ decoder.up.2.block.0.conv1.weight
1287
+ decoder.up.2.block.0.conv2.bias
1288
+ decoder.up.2.block.0.conv2.weight
1289
+ decoder.up.2.block.0.norm1.bias
1290
+ decoder.up.2.block.0.norm1.weight
1291
+ decoder.up.2.block.0.norm2.bias
1292
+ decoder.up.2.block.0.norm2.weight
1293
+ decoder.up.2.block.1.conv1.bias
1294
+ decoder.up.2.block.1.conv1.weight
1295
+ decoder.up.2.block.1.conv2.bias
1296
+ decoder.up.2.block.1.conv2.weight
1297
+ decoder.up.2.block.1.norm1.bias
1298
+ decoder.up.2.block.1.norm1.weight
1299
+ decoder.up.2.block.1.norm2.bias
1300
+ decoder.up.2.block.1.norm2.weight
1301
+ decoder.up.2.block.2.conv1.bias
1302
+ decoder.up.2.block.2.conv1.weight
1303
+ decoder.up.2.block.2.conv2.bias
1304
+ decoder.up.2.block.2.conv2.weight
1305
+ decoder.up.2.block.2.norm1.bias
1306
+ decoder.up.2.block.2.norm1.weight
1307
+ decoder.up.2.block.2.norm2.bias
1308
+ decoder.up.2.block.2.norm2.weight
1309
+ decoder.up.2.upsample.conv.bias
1310
+ decoder.up.2.upsample.conv.weight
1311
+ decoder.up.3.block.0.conv1.bias
1312
+ decoder.up.3.block.0.conv1.weight
1313
+ decoder.up.3.block.0.conv2.bias
1314
+ decoder.up.3.block.0.conv2.weight
1315
+ decoder.up.3.block.0.norm1.bias
1316
+ decoder.up.3.block.0.norm1.weight
1317
+ decoder.up.3.block.0.norm2.bias
1318
+ decoder.up.3.block.0.norm2.weight
1319
+ decoder.up.3.block.1.conv1.bias
1320
+ decoder.up.3.block.1.conv1.weight
1321
+ decoder.up.3.block.1.conv2.bias
1322
+ decoder.up.3.block.1.conv2.weight
1323
+ decoder.up.3.block.1.norm1.bias
1324
+ decoder.up.3.block.1.norm1.weight
1325
+ decoder.up.3.block.1.norm2.bias
1326
+ decoder.up.3.block.1.norm2.weight
1327
+ decoder.up.3.block.2.conv1.bias
1328
+ decoder.up.3.block.2.conv1.weight
1329
+ decoder.up.3.block.2.conv2.bias
1330
+ decoder.up.3.block.2.conv2.weight
1331
+ decoder.up.3.block.2.norm1.bias
1332
+ decoder.up.3.block.2.norm1.weight
1333
+ decoder.up.3.block.2.norm2.bias
1334
+ decoder.up.3.block.2.norm2.weight
1335
+ decoder.up.3.upsample.conv.bias
1336
+ decoder.up.3.upsample.conv.weight
1337
+ encoder.conv_in.bias
1338
+ encoder.conv_in.weight
1339
+ encoder.conv_out.bias
1340
+ encoder.conv_out.weight
1341
+ encoder.down.0.block.0.conv1.bias
1342
+ encoder.down.0.block.0.conv1.weight
1343
+ encoder.down.0.block.0.conv2.bias
1344
+ encoder.down.0.block.0.conv2.weight
1345
+ encoder.down.0.block.0.norm1.bias
1346
+ encoder.down.0.block.0.norm1.weight
1347
+ encoder.down.0.block.0.norm2.bias
1348
+ encoder.down.0.block.0.norm2.weight
1349
+ encoder.down.0.block.1.conv1.bias
1350
+ encoder.down.0.block.1.conv1.weight
1351
+ encoder.down.0.block.1.conv2.bias
1352
+ encoder.down.0.block.1.conv2.weight
1353
+ encoder.down.0.block.1.norm1.bias
1354
+ encoder.down.0.block.1.norm1.weight
1355
+ encoder.down.0.block.1.norm2.bias
1356
+ encoder.down.0.block.1.norm2.weight
1357
+ encoder.down.0.downsample.conv.bias
1358
+ encoder.down.0.downsample.conv.weight
1359
+ encoder.down.1.block.0.conv1.bias
1360
+ encoder.down.1.block.0.conv1.weight
1361
+ encoder.down.1.block.0.conv2.bias
1362
+ encoder.down.1.block.0.conv2.weight
1363
+ encoder.down.1.block.0.nin_shortcut.bias
1364
+ encoder.down.1.block.0.nin_shortcut.weight
1365
+ encoder.down.1.block.0.norm1.bias
1366
+ encoder.down.1.block.0.norm1.weight
1367
+ encoder.down.1.block.0.norm2.bias
1368
+ encoder.down.1.block.0.norm2.weight
1369
+ encoder.down.1.block.1.conv1.bias
1370
+ encoder.down.1.block.1.conv1.weight
1371
+ encoder.down.1.block.1.conv2.bias
1372
+ encoder.down.1.block.1.conv2.weight
1373
+ encoder.down.1.block.1.norm1.bias
1374
+ encoder.down.1.block.1.norm1.weight
1375
+ encoder.down.1.block.1.norm2.bias
1376
+ encoder.down.1.block.1.norm2.weight
1377
+ encoder.down.1.downsample.conv.bias
1378
+ encoder.down.1.downsample.conv.weight
1379
+ encoder.down.2.block.0.conv1.bias
1380
+ encoder.down.2.block.0.conv1.weight
1381
+ encoder.down.2.block.0.conv2.bias
1382
+ encoder.down.2.block.0.conv2.weight
1383
+ encoder.down.2.block.0.nin_shortcut.bias
1384
+ encoder.down.2.block.0.nin_shortcut.weight
1385
+ encoder.down.2.block.0.norm1.bias
1386
+ encoder.down.2.block.0.norm1.weight
1387
+ encoder.down.2.block.0.norm2.bias
1388
+ encoder.down.2.block.0.norm2.weight
1389
+ encoder.down.2.block.1.conv1.bias
1390
+ encoder.down.2.block.1.conv1.weight
1391
+ encoder.down.2.block.1.conv2.bias
1392
+ encoder.down.2.block.1.conv2.weight
1393
+ encoder.down.2.block.1.norm1.bias
1394
+ encoder.down.2.block.1.norm1.weight
1395
+ encoder.down.2.block.1.norm2.bias
1396
+ encoder.down.2.block.1.norm2.weight
1397
+ encoder.down.2.downsample.conv.bias
1398
+ encoder.down.2.downsample.conv.weight
1399
+ encoder.down.3.block.0.conv1.bias
1400
+ encoder.down.3.block.0.conv1.weight
1401
+ encoder.down.3.block.0.conv2.bias
1402
+ encoder.down.3.block.0.conv2.weight
1403
+ encoder.down.3.block.0.norm1.bias
1404
+ encoder.down.3.block.0.norm1.weight
1405
+ encoder.down.3.block.0.norm2.bias
1406
+ encoder.down.3.block.0.norm2.weight
1407
+ encoder.down.3.block.1.conv1.bias
1408
+ encoder.down.3.block.1.conv1.weight
1409
+ encoder.down.3.block.1.conv2.bias
1410
+ encoder.down.3.block.1.conv2.weight
1411
+ encoder.down.3.block.1.norm1.bias
1412
+ encoder.down.3.block.1.norm1.weight
1413
+ encoder.down.3.block.1.norm2.bias
1414
+ encoder.down.3.block.1.norm2.weight
1415
+ encoder.mid.attn_1.k.bias
1416
+ encoder.mid.attn_1.k.weight
1417
+ encoder.mid.attn_1.norm.bias
1418
+ encoder.mid.attn_1.norm.weight
1419
+ encoder.mid.attn_1.proj_out.bias
1420
+ encoder.mid.attn_1.proj_out.weight
1421
+ encoder.mid.attn_1.q.bias
1422
+ encoder.mid.attn_1.q.weight
1423
+ encoder.mid.attn_1.v.bias
1424
+ encoder.mid.attn_1.v.weight
1425
+ encoder.mid.block_1.conv1.bias
1426
+ encoder.mid.block_1.conv1.weight
1427
+ encoder.mid.block_1.conv2.bias
1428
+ encoder.mid.block_1.conv2.weight
1429
+ encoder.mid.block_1.norm1.bias
1430
+ encoder.mid.block_1.norm1.weight
1431
+ encoder.mid.block_1.norm2.bias
1432
+ encoder.mid.block_1.norm2.weight
1433
+ encoder.mid.block_2.conv1.bias
1434
+ encoder.mid.block_2.conv1.weight
1435
+ encoder.mid.block_2.conv2.bias
1436
+ encoder.mid.block_2.conv2.weight
1437
+ encoder.mid.block_2.norm1.bias
1438
+ encoder.mid.block_2.norm1.weight
1439
+ encoder.mid.block_2.norm2.bias
1440
+ encoder.mid.block_2.norm2.weight
1441
+ encoder.norm_out.bias
1442
+ encoder.norm_out.weight
fluxunchainedArtfulNSFW_fuT516xfp8E4m3fnV11_fixed.safetensors.old.txt.txt ADDED
@@ -0,0 +1,1442 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model.diffusion_model.double_blocks.0.img_attn.norm.key_norm.scale
2
+ model.diffusion_model.double_blocks.0.img_attn.norm.query_norm.scale
3
+ model.diffusion_model.double_blocks.0.img_attn.proj.bias
4
+ model.diffusion_model.double_blocks.0.img_attn.proj.weight
5
+ model.diffusion_model.double_blocks.0.img_attn.qkv.bias
6
+ model.diffusion_model.double_blocks.0.img_attn.qkv.weight
7
+ model.diffusion_model.double_blocks.0.img_mlp.0.bias
8
+ model.diffusion_model.double_blocks.0.img_mlp.0.weight
9
+ model.diffusion_model.double_blocks.0.img_mlp.2.bias
10
+ model.diffusion_model.double_blocks.0.img_mlp.2.weight
11
+ model.diffusion_model.double_blocks.0.img_mod.lin.bias
12
+ model.diffusion_model.double_blocks.0.img_mod.lin.weight
13
+ model.diffusion_model.double_blocks.0.txt_attn.norm.key_norm.scale
14
+ model.diffusion_model.double_blocks.0.txt_attn.norm.query_norm.scale
15
+ model.diffusion_model.double_blocks.0.txt_attn.proj.bias
16
+ model.diffusion_model.double_blocks.0.txt_attn.proj.weight
17
+ model.diffusion_model.double_blocks.0.txt_attn.qkv.bias
18
+ model.diffusion_model.double_blocks.0.txt_attn.qkv.weight
19
+ model.diffusion_model.double_blocks.0.txt_mlp.0.bias
20
+ model.diffusion_model.double_blocks.0.txt_mlp.0.weight
21
+ model.diffusion_model.double_blocks.0.txt_mlp.2.bias
22
+ model.diffusion_model.double_blocks.0.txt_mlp.2.weight
23
+ model.diffusion_model.double_blocks.0.txt_mod.lin.bias
24
+ model.diffusion_model.double_blocks.0.txt_mod.lin.weight
25
+ model.diffusion_model.double_blocks.1.img_attn.norm.key_norm.scale
26
+ model.diffusion_model.double_blocks.1.img_attn.norm.query_norm.scale
27
+ model.diffusion_model.double_blocks.1.img_attn.proj.bias
28
+ model.diffusion_model.double_blocks.1.img_attn.proj.weight
29
+ model.diffusion_model.double_blocks.1.img_attn.qkv.bias
30
+ model.diffusion_model.double_blocks.1.img_attn.qkv.weight
31
+ model.diffusion_model.double_blocks.1.img_mlp.0.bias
32
+ model.diffusion_model.double_blocks.1.img_mlp.0.weight
33
+ model.diffusion_model.double_blocks.1.img_mlp.2.bias
34
+ model.diffusion_model.double_blocks.1.img_mlp.2.weight
35
+ model.diffusion_model.double_blocks.1.img_mod.lin.bias
36
+ model.diffusion_model.double_blocks.1.img_mod.lin.weight
37
+ model.diffusion_model.double_blocks.1.txt_attn.norm.key_norm.scale
38
+ model.diffusion_model.double_blocks.1.txt_attn.norm.query_norm.scale
39
+ model.diffusion_model.double_blocks.1.txt_attn.proj.bias
40
+ model.diffusion_model.double_blocks.1.txt_attn.proj.weight
41
+ model.diffusion_model.double_blocks.1.txt_attn.qkv.bias
42
+ model.diffusion_model.double_blocks.1.txt_attn.qkv.weight
43
+ model.diffusion_model.double_blocks.1.txt_mlp.0.bias
44
+ model.diffusion_model.double_blocks.1.txt_mlp.0.weight
45
+ model.diffusion_model.double_blocks.1.txt_mlp.2.bias
46
+ model.diffusion_model.double_blocks.1.txt_mlp.2.weight
47
+ model.diffusion_model.double_blocks.1.txt_mod.lin.bias
48
+ model.diffusion_model.double_blocks.1.txt_mod.lin.weight
49
+ model.diffusion_model.double_blocks.10.img_attn.norm.key_norm.scale
50
+ model.diffusion_model.double_blocks.10.img_attn.norm.query_norm.scale
51
+ model.diffusion_model.double_blocks.10.img_attn.proj.bias
52
+ model.diffusion_model.double_blocks.10.img_attn.proj.weight
53
+ model.diffusion_model.double_blocks.10.img_attn.qkv.bias
54
+ model.diffusion_model.double_blocks.10.img_attn.qkv.weight
55
+ model.diffusion_model.double_blocks.10.img_mlp.0.bias
56
+ model.diffusion_model.double_blocks.10.img_mlp.0.weight
57
+ model.diffusion_model.double_blocks.10.img_mlp.2.bias
58
+ model.diffusion_model.double_blocks.10.img_mlp.2.weight
59
+ model.diffusion_model.double_blocks.10.img_mod.lin.bias
60
+ model.diffusion_model.double_blocks.10.img_mod.lin.weight
61
+ model.diffusion_model.double_blocks.10.txt_attn.norm.key_norm.scale
62
+ model.diffusion_model.double_blocks.10.txt_attn.norm.query_norm.scale
63
+ model.diffusion_model.double_blocks.10.txt_attn.proj.bias
64
+ model.diffusion_model.double_blocks.10.txt_attn.proj.weight
65
+ model.diffusion_model.double_blocks.10.txt_attn.qkv.bias
66
+ model.diffusion_model.double_blocks.10.txt_attn.qkv.weight
67
+ model.diffusion_model.double_blocks.10.txt_mlp.0.bias
68
+ model.diffusion_model.double_blocks.10.txt_mlp.0.weight
69
+ model.diffusion_model.double_blocks.10.txt_mlp.2.bias
70
+ model.diffusion_model.double_blocks.10.txt_mlp.2.weight
71
+ model.diffusion_model.double_blocks.10.txt_mod.lin.bias
72
+ model.diffusion_model.double_blocks.10.txt_mod.lin.weight
73
+ model.diffusion_model.double_blocks.11.img_attn.norm.key_norm.scale
74
+ model.diffusion_model.double_blocks.11.img_attn.norm.query_norm.scale
75
+ model.diffusion_model.double_blocks.11.img_attn.proj.bias
76
+ model.diffusion_model.double_blocks.11.img_attn.proj.weight
77
+ model.diffusion_model.double_blocks.11.img_attn.qkv.bias
78
+ model.diffusion_model.double_blocks.11.img_attn.qkv.weight
79
+ model.diffusion_model.double_blocks.11.img_mlp.0.bias
80
+ model.diffusion_model.double_blocks.11.img_mlp.0.weight
81
+ model.diffusion_model.double_blocks.11.img_mlp.2.bias
82
+ model.diffusion_model.double_blocks.11.img_mlp.2.weight
83
+ model.diffusion_model.double_blocks.11.img_mod.lin.bias
84
+ model.diffusion_model.double_blocks.11.img_mod.lin.weight
85
+ model.diffusion_model.double_blocks.11.txt_attn.norm.key_norm.scale
86
+ model.diffusion_model.double_blocks.11.txt_attn.norm.query_norm.scale
87
+ model.diffusion_model.double_blocks.11.txt_attn.proj.bias
88
+ model.diffusion_model.double_blocks.11.txt_attn.proj.weight
89
+ model.diffusion_model.double_blocks.11.txt_attn.qkv.bias
90
+ model.diffusion_model.double_blocks.11.txt_attn.qkv.weight
91
+ model.diffusion_model.double_blocks.11.txt_mlp.0.bias
92
+ model.diffusion_model.double_blocks.11.txt_mlp.0.weight
93
+ model.diffusion_model.double_blocks.11.txt_mlp.2.bias
94
+ model.diffusion_model.double_blocks.11.txt_mlp.2.weight
95
+ model.diffusion_model.double_blocks.11.txt_mod.lin.bias
96
+ model.diffusion_model.double_blocks.11.txt_mod.lin.weight
97
+ model.diffusion_model.double_blocks.12.img_attn.norm.key_norm.scale
98
+ model.diffusion_model.double_blocks.12.img_attn.norm.query_norm.scale
99
+ model.diffusion_model.double_blocks.12.img_attn.proj.bias
100
+ model.diffusion_model.double_blocks.12.img_attn.proj.weight
101
+ model.diffusion_model.double_blocks.12.img_attn.qkv.bias
102
+ model.diffusion_model.double_blocks.12.img_attn.qkv.weight
103
+ model.diffusion_model.double_blocks.12.img_mlp.0.bias
104
+ model.diffusion_model.double_blocks.12.img_mlp.0.weight
105
+ model.diffusion_model.double_blocks.12.img_mlp.2.bias
106
+ model.diffusion_model.double_blocks.12.img_mlp.2.weight
107
+ model.diffusion_model.double_blocks.12.img_mod.lin.bias
108
+ model.diffusion_model.double_blocks.12.img_mod.lin.weight
109
+ model.diffusion_model.double_blocks.12.txt_attn.norm.key_norm.scale
110
+ model.diffusion_model.double_blocks.12.txt_attn.norm.query_norm.scale
111
+ model.diffusion_model.double_blocks.12.txt_attn.proj.bias
112
+ model.diffusion_model.double_blocks.12.txt_attn.proj.weight
113
+ model.diffusion_model.double_blocks.12.txt_attn.qkv.bias
114
+ model.diffusion_model.double_blocks.12.txt_attn.qkv.weight
115
+ model.diffusion_model.double_blocks.12.txt_mlp.0.bias
116
+ model.diffusion_model.double_blocks.12.txt_mlp.0.weight
117
+ model.diffusion_model.double_blocks.12.txt_mlp.2.bias
118
+ model.diffusion_model.double_blocks.12.txt_mlp.2.weight
119
+ model.diffusion_model.double_blocks.12.txt_mod.lin.bias
120
+ model.diffusion_model.double_blocks.12.txt_mod.lin.weight
121
+ model.diffusion_model.double_blocks.13.img_attn.norm.key_norm.scale
122
+ model.diffusion_model.double_blocks.13.img_attn.norm.query_norm.scale
123
+ model.diffusion_model.double_blocks.13.img_attn.proj.bias
124
+ model.diffusion_model.double_blocks.13.img_attn.proj.weight
125
+ model.diffusion_model.double_blocks.13.img_attn.qkv.bias
126
+ model.diffusion_model.double_blocks.13.img_attn.qkv.weight
127
+ model.diffusion_model.double_blocks.13.img_mlp.0.bias
128
+ model.diffusion_model.double_blocks.13.img_mlp.0.weight
129
+ model.diffusion_model.double_blocks.13.img_mlp.2.bias
130
+ model.diffusion_model.double_blocks.13.img_mlp.2.weight
131
+ model.diffusion_model.double_blocks.13.img_mod.lin.bias
132
+ model.diffusion_model.double_blocks.13.img_mod.lin.weight
133
+ model.diffusion_model.double_blocks.13.txt_attn.norm.key_norm.scale
134
+ model.diffusion_model.double_blocks.13.txt_attn.norm.query_norm.scale
135
+ model.diffusion_model.double_blocks.13.txt_attn.proj.bias
136
+ model.diffusion_model.double_blocks.13.txt_attn.proj.weight
137
+ model.diffusion_model.double_blocks.13.txt_attn.qkv.bias
138
+ model.diffusion_model.double_blocks.13.txt_attn.qkv.weight
139
+ model.diffusion_model.double_blocks.13.txt_mlp.0.bias
140
+ model.diffusion_model.double_blocks.13.txt_mlp.0.weight
141
+ model.diffusion_model.double_blocks.13.txt_mlp.2.bias
142
+ model.diffusion_model.double_blocks.13.txt_mlp.2.weight
143
+ model.diffusion_model.double_blocks.13.txt_mod.lin.bias
144
+ model.diffusion_model.double_blocks.13.txt_mod.lin.weight
145
+ model.diffusion_model.double_blocks.14.img_attn.norm.key_norm.scale
146
+ model.diffusion_model.double_blocks.14.img_attn.norm.query_norm.scale
147
+ model.diffusion_model.double_blocks.14.img_attn.proj.bias
148
+ model.diffusion_model.double_blocks.14.img_attn.proj.weight
149
+ model.diffusion_model.double_blocks.14.img_attn.qkv.bias
150
+ model.diffusion_model.double_blocks.14.img_attn.qkv.weight
151
+ model.diffusion_model.double_blocks.14.img_mlp.0.bias
152
+ model.diffusion_model.double_blocks.14.img_mlp.0.weight
153
+ model.diffusion_model.double_blocks.14.img_mlp.2.bias
154
+ model.diffusion_model.double_blocks.14.img_mlp.2.weight
155
+ model.diffusion_model.double_blocks.14.img_mod.lin.bias
156
+ model.diffusion_model.double_blocks.14.img_mod.lin.weight
157
+ model.diffusion_model.double_blocks.14.txt_attn.norm.key_norm.scale
158
+ model.diffusion_model.double_blocks.14.txt_attn.norm.query_norm.scale
159
+ model.diffusion_model.double_blocks.14.txt_attn.proj.bias
160
+ model.diffusion_model.double_blocks.14.txt_attn.proj.weight
161
+ model.diffusion_model.double_blocks.14.txt_attn.qkv.bias
162
+ model.diffusion_model.double_blocks.14.txt_attn.qkv.weight
163
+ model.diffusion_model.double_blocks.14.txt_mlp.0.bias
164
+ model.diffusion_model.double_blocks.14.txt_mlp.0.weight
165
+ model.diffusion_model.double_blocks.14.txt_mlp.2.bias
166
+ model.diffusion_model.double_blocks.14.txt_mlp.2.weight
167
+ model.diffusion_model.double_blocks.14.txt_mod.lin.bias
168
+ model.diffusion_model.double_blocks.14.txt_mod.lin.weight
169
+ model.diffusion_model.double_blocks.15.img_attn.norm.key_norm.scale
170
+ model.diffusion_model.double_blocks.15.img_attn.norm.query_norm.scale
171
+ model.diffusion_model.double_blocks.15.img_attn.proj.bias
172
+ model.diffusion_model.double_blocks.15.img_attn.proj.weight
173
+ model.diffusion_model.double_blocks.15.img_attn.qkv.bias
174
+ model.diffusion_model.double_blocks.15.img_attn.qkv.weight
175
+ model.diffusion_model.double_blocks.15.img_mlp.0.bias
176
+ model.diffusion_model.double_blocks.15.img_mlp.0.weight
177
+ model.diffusion_model.double_blocks.15.img_mlp.2.bias
178
+ model.diffusion_model.double_blocks.15.img_mlp.2.weight
179
+ model.diffusion_model.double_blocks.15.img_mod.lin.bias
180
+ model.diffusion_model.double_blocks.15.img_mod.lin.weight
181
+ model.diffusion_model.double_blocks.15.txt_attn.norm.key_norm.scale
182
+ model.diffusion_model.double_blocks.15.txt_attn.norm.query_norm.scale
183
+ model.diffusion_model.double_blocks.15.txt_attn.proj.bias
184
+ model.diffusion_model.double_blocks.15.txt_attn.proj.weight
185
+ model.diffusion_model.double_blocks.15.txt_attn.qkv.bias
186
+ model.diffusion_model.double_blocks.15.txt_attn.qkv.weight
187
+ model.diffusion_model.double_blocks.15.txt_mlp.0.bias
188
+ model.diffusion_model.double_blocks.15.txt_mlp.0.weight
189
+ model.diffusion_model.double_blocks.15.txt_mlp.2.bias
190
+ model.diffusion_model.double_blocks.15.txt_mlp.2.weight
191
+ model.diffusion_model.double_blocks.15.txt_mod.lin.bias
192
+ model.diffusion_model.double_blocks.15.txt_mod.lin.weight
193
+ model.diffusion_model.double_blocks.16.img_attn.norm.key_norm.scale
194
+ model.diffusion_model.double_blocks.16.img_attn.norm.query_norm.scale
195
+ model.diffusion_model.double_blocks.16.img_attn.proj.bias
196
+ model.diffusion_model.double_blocks.16.img_attn.proj.weight
197
+ model.diffusion_model.double_blocks.16.img_attn.qkv.bias
198
+ model.diffusion_model.double_blocks.16.img_attn.qkv.weight
199
+ model.diffusion_model.double_blocks.16.img_mlp.0.bias
200
+ model.diffusion_model.double_blocks.16.img_mlp.0.weight
201
+ model.diffusion_model.double_blocks.16.img_mlp.2.bias
202
+ model.diffusion_model.double_blocks.16.img_mlp.2.weight
203
+ model.diffusion_model.double_blocks.16.img_mod.lin.bias
204
+ model.diffusion_model.double_blocks.16.img_mod.lin.weight
205
+ model.diffusion_model.double_blocks.16.txt_attn.norm.key_norm.scale
206
+ model.diffusion_model.double_blocks.16.txt_attn.norm.query_norm.scale
207
+ model.diffusion_model.double_blocks.16.txt_attn.proj.bias
208
+ model.diffusion_model.double_blocks.16.txt_attn.proj.weight
209
+ model.diffusion_model.double_blocks.16.txt_attn.qkv.bias
210
+ model.diffusion_model.double_blocks.16.txt_attn.qkv.weight
211
+ model.diffusion_model.double_blocks.16.txt_mlp.0.bias
212
+ model.diffusion_model.double_blocks.16.txt_mlp.0.weight
213
+ model.diffusion_model.double_blocks.16.txt_mlp.2.bias
214
+ model.diffusion_model.double_blocks.16.txt_mlp.2.weight
215
+ model.diffusion_model.double_blocks.16.txt_mod.lin.bias
216
+ model.diffusion_model.double_blocks.16.txt_mod.lin.weight
217
+ model.diffusion_model.double_blocks.17.img_attn.norm.key_norm.scale
218
+ model.diffusion_model.double_blocks.17.img_attn.norm.query_norm.scale
219
+ model.diffusion_model.double_blocks.17.img_attn.proj.bias
220
+ model.diffusion_model.double_blocks.17.img_attn.proj.weight
221
+ model.diffusion_model.double_blocks.17.img_attn.qkv.bias
222
+ model.diffusion_model.double_blocks.17.img_attn.qkv.weight
223
+ model.diffusion_model.double_blocks.17.img_mlp.0.bias
224
+ model.diffusion_model.double_blocks.17.img_mlp.0.weight
225
+ model.diffusion_model.double_blocks.17.img_mlp.2.bias
226
+ model.diffusion_model.double_blocks.17.img_mlp.2.weight
227
+ model.diffusion_model.double_blocks.17.img_mod.lin.bias
228
+ model.diffusion_model.double_blocks.17.img_mod.lin.weight
229
+ model.diffusion_model.double_blocks.17.txt_attn.norm.key_norm.scale
230
+ model.diffusion_model.double_blocks.17.txt_attn.norm.query_norm.scale
231
+ model.diffusion_model.double_blocks.17.txt_attn.proj.bias
232
+ model.diffusion_model.double_blocks.17.txt_attn.proj.weight
233
+ model.diffusion_model.double_blocks.17.txt_attn.qkv.bias
234
+ model.diffusion_model.double_blocks.17.txt_attn.qkv.weight
235
+ model.diffusion_model.double_blocks.17.txt_mlp.0.bias
236
+ model.diffusion_model.double_blocks.17.txt_mlp.0.weight
237
+ model.diffusion_model.double_blocks.17.txt_mlp.2.bias
238
+ model.diffusion_model.double_blocks.17.txt_mlp.2.weight
239
+ model.diffusion_model.double_blocks.17.txt_mod.lin.bias
240
+ model.diffusion_model.double_blocks.17.txt_mod.lin.weight
241
+ model.diffusion_model.double_blocks.18.img_attn.norm.key_norm.scale
242
+ model.diffusion_model.double_blocks.18.img_attn.norm.query_norm.scale
243
+ model.diffusion_model.double_blocks.18.img_attn.proj.bias
244
+ model.diffusion_model.double_blocks.18.img_attn.proj.weight
245
+ model.diffusion_model.double_blocks.18.img_attn.qkv.bias
246
+ model.diffusion_model.double_blocks.18.img_attn.qkv.weight
247
+ model.diffusion_model.double_blocks.18.img_mlp.0.bias
248
+ model.diffusion_model.double_blocks.18.img_mlp.0.weight
249
+ model.diffusion_model.double_blocks.18.img_mlp.2.bias
250
+ model.diffusion_model.double_blocks.18.img_mlp.2.weight
251
+ model.diffusion_model.double_blocks.18.img_mod.lin.bias
252
+ model.diffusion_model.double_blocks.18.img_mod.lin.weight
253
+ model.diffusion_model.double_blocks.18.txt_attn.norm.key_norm.scale
254
+ model.diffusion_model.double_blocks.18.txt_attn.norm.query_norm.scale
255
+ model.diffusion_model.double_blocks.18.txt_attn.proj.bias
256
+ model.diffusion_model.double_blocks.18.txt_attn.proj.weight
257
+ model.diffusion_model.double_blocks.18.txt_attn.qkv.bias
258
+ model.diffusion_model.double_blocks.18.txt_attn.qkv.weight
259
+ model.diffusion_model.double_blocks.18.txt_mlp.0.bias
260
+ model.diffusion_model.double_blocks.18.txt_mlp.0.weight
261
+ model.diffusion_model.double_blocks.18.txt_mlp.2.bias
262
+ model.diffusion_model.double_blocks.18.txt_mlp.2.weight
263
+ model.diffusion_model.double_blocks.18.txt_mod.lin.bias
264
+ model.diffusion_model.double_blocks.18.txt_mod.lin.weight
265
+ model.diffusion_model.double_blocks.2.img_attn.norm.key_norm.scale
266
+ model.diffusion_model.double_blocks.2.img_attn.norm.query_norm.scale
267
+ model.diffusion_model.double_blocks.2.img_attn.proj.bias
268
+ model.diffusion_model.double_blocks.2.img_attn.proj.weight
269
+ model.diffusion_model.double_blocks.2.img_attn.qkv.bias
270
+ model.diffusion_model.double_blocks.2.img_attn.qkv.weight
271
+ model.diffusion_model.double_blocks.2.img_mlp.0.bias
272
+ model.diffusion_model.double_blocks.2.img_mlp.0.weight
273
+ model.diffusion_model.double_blocks.2.img_mlp.2.bias
274
+ model.diffusion_model.double_blocks.2.img_mlp.2.weight
275
+ model.diffusion_model.double_blocks.2.img_mod.lin.bias
276
+ model.diffusion_model.double_blocks.2.img_mod.lin.weight
277
+ model.diffusion_model.double_blocks.2.txt_attn.norm.key_norm.scale
278
+ model.diffusion_model.double_blocks.2.txt_attn.norm.query_norm.scale
279
+ model.diffusion_model.double_blocks.2.txt_attn.proj.bias
280
+ model.diffusion_model.double_blocks.2.txt_attn.proj.weight
281
+ model.diffusion_model.double_blocks.2.txt_attn.qkv.bias
282
+ model.diffusion_model.double_blocks.2.txt_attn.qkv.weight
283
+ model.diffusion_model.double_blocks.2.txt_mlp.0.bias
284
+ model.diffusion_model.double_blocks.2.txt_mlp.0.weight
285
+ model.diffusion_model.double_blocks.2.txt_mlp.2.bias
286
+ model.diffusion_model.double_blocks.2.txt_mlp.2.weight
287
+ model.diffusion_model.double_blocks.2.txt_mod.lin.bias
288
+ model.diffusion_model.double_blocks.2.txt_mod.lin.weight
289
+ model.diffusion_model.double_blocks.3.img_attn.norm.key_norm.scale
290
+ model.diffusion_model.double_blocks.3.img_attn.norm.query_norm.scale
291
+ model.diffusion_model.double_blocks.3.img_attn.proj.bias
292
+ model.diffusion_model.double_blocks.3.img_attn.proj.weight
293
+ model.diffusion_model.double_blocks.3.img_attn.qkv.bias
294
+ model.diffusion_model.double_blocks.3.img_attn.qkv.weight
295
+ model.diffusion_model.double_blocks.3.img_mlp.0.bias
296
+ model.diffusion_model.double_blocks.3.img_mlp.0.weight
297
+ model.diffusion_model.double_blocks.3.img_mlp.2.bias
298
+ model.diffusion_model.double_blocks.3.img_mlp.2.weight
299
+ model.diffusion_model.double_blocks.3.img_mod.lin.bias
300
+ model.diffusion_model.double_blocks.3.img_mod.lin.weight
301
+ model.diffusion_model.double_blocks.3.txt_attn.norm.key_norm.scale
302
+ model.diffusion_model.double_blocks.3.txt_attn.norm.query_norm.scale
303
+ model.diffusion_model.double_blocks.3.txt_attn.proj.bias
304
+ model.diffusion_model.double_blocks.3.txt_attn.proj.weight
305
+ model.diffusion_model.double_blocks.3.txt_attn.qkv.bias
306
+ model.diffusion_model.double_blocks.3.txt_attn.qkv.weight
307
+ model.diffusion_model.double_blocks.3.txt_mlp.0.bias
308
+ model.diffusion_model.double_blocks.3.txt_mlp.0.weight
309
+ model.diffusion_model.double_blocks.3.txt_mlp.2.bias
310
+ model.diffusion_model.double_blocks.3.txt_mlp.2.weight
311
+ model.diffusion_model.double_blocks.3.txt_mod.lin.bias
312
+ model.diffusion_model.double_blocks.3.txt_mod.lin.weight
313
+ model.diffusion_model.double_blocks.4.img_attn.norm.key_norm.scale
314
+ model.diffusion_model.double_blocks.4.img_attn.norm.query_norm.scale
315
+ model.diffusion_model.double_blocks.4.img_attn.proj.bias
316
+ model.diffusion_model.double_blocks.4.img_attn.proj.weight
317
+ model.diffusion_model.double_blocks.4.img_attn.qkv.bias
318
+ model.diffusion_model.double_blocks.4.img_attn.qkv.weight
319
+ model.diffusion_model.double_blocks.4.img_mlp.0.bias
320
+ model.diffusion_model.double_blocks.4.img_mlp.0.weight
321
+ model.diffusion_model.double_blocks.4.img_mlp.2.bias
322
+ model.diffusion_model.double_blocks.4.img_mlp.2.weight
323
+ model.diffusion_model.double_blocks.4.img_mod.lin.bias
324
+ model.diffusion_model.double_blocks.4.img_mod.lin.weight
325
+ model.diffusion_model.double_blocks.4.txt_attn.norm.key_norm.scale
326
+ model.diffusion_model.double_blocks.4.txt_attn.norm.query_norm.scale
327
+ model.diffusion_model.double_blocks.4.txt_attn.proj.bias
328
+ model.diffusion_model.double_blocks.4.txt_attn.proj.weight
329
+ model.diffusion_model.double_blocks.4.txt_attn.qkv.bias
330
+ model.diffusion_model.double_blocks.4.txt_attn.qkv.weight
331
+ model.diffusion_model.double_blocks.4.txt_mlp.0.bias
332
+ model.diffusion_model.double_blocks.4.txt_mlp.0.weight
333
+ model.diffusion_model.double_blocks.4.txt_mlp.2.bias
334
+ model.diffusion_model.double_blocks.4.txt_mlp.2.weight
335
+ model.diffusion_model.double_blocks.4.txt_mod.lin.bias
336
+ model.diffusion_model.double_blocks.4.txt_mod.lin.weight
337
+ model.diffusion_model.double_blocks.5.img_attn.norm.key_norm.scale
338
+ model.diffusion_model.double_blocks.5.img_attn.norm.query_norm.scale
339
+ model.diffusion_model.double_blocks.5.img_attn.proj.bias
340
+ model.diffusion_model.double_blocks.5.img_attn.proj.weight
341
+ model.diffusion_model.double_blocks.5.img_attn.qkv.bias
342
+ model.diffusion_model.double_blocks.5.img_attn.qkv.weight
343
+ model.diffusion_model.double_blocks.5.img_mlp.0.bias
344
+ model.diffusion_model.double_blocks.5.img_mlp.0.weight
345
+ model.diffusion_model.double_blocks.5.img_mlp.2.bias
346
+ model.diffusion_model.double_blocks.5.img_mlp.2.weight
347
+ model.diffusion_model.double_blocks.5.img_mod.lin.bias
348
+ model.diffusion_model.double_blocks.5.img_mod.lin.weight
349
+ model.diffusion_model.double_blocks.5.txt_attn.norm.key_norm.scale
350
+ model.diffusion_model.double_blocks.5.txt_attn.norm.query_norm.scale
351
+ model.diffusion_model.double_blocks.5.txt_attn.proj.bias
352
+ model.diffusion_model.double_blocks.5.txt_attn.proj.weight
353
+ model.diffusion_model.double_blocks.5.txt_attn.qkv.bias
354
+ model.diffusion_model.double_blocks.5.txt_attn.qkv.weight
355
+ model.diffusion_model.double_blocks.5.txt_mlp.0.bias
356
+ model.diffusion_model.double_blocks.5.txt_mlp.0.weight
357
+ model.diffusion_model.double_blocks.5.txt_mlp.2.bias
358
+ model.diffusion_model.double_blocks.5.txt_mlp.2.weight
359
+ model.diffusion_model.double_blocks.5.txt_mod.lin.bias
360
+ model.diffusion_model.double_blocks.5.txt_mod.lin.weight
361
+ model.diffusion_model.double_blocks.6.img_attn.norm.key_norm.scale
362
+ model.diffusion_model.double_blocks.6.img_attn.norm.query_norm.scale
363
+ model.diffusion_model.double_blocks.6.img_attn.proj.bias
364
+ model.diffusion_model.double_blocks.6.img_attn.proj.weight
365
+ model.diffusion_model.double_blocks.6.img_attn.qkv.bias
366
+ model.diffusion_model.double_blocks.6.img_attn.qkv.weight
367
+ model.diffusion_model.double_blocks.6.img_mlp.0.bias
368
+ model.diffusion_model.double_blocks.6.img_mlp.0.weight
369
+ model.diffusion_model.double_blocks.6.img_mlp.2.bias
370
+ model.diffusion_model.double_blocks.6.img_mlp.2.weight
371
+ model.diffusion_model.double_blocks.6.img_mod.lin.bias
372
+ model.diffusion_model.double_blocks.6.img_mod.lin.weight
373
+ model.diffusion_model.double_blocks.6.txt_attn.norm.key_norm.scale
374
+ model.diffusion_model.double_blocks.6.txt_attn.norm.query_norm.scale
375
+ model.diffusion_model.double_blocks.6.txt_attn.proj.bias
376
+ model.diffusion_model.double_blocks.6.txt_attn.proj.weight
377
+ model.diffusion_model.double_blocks.6.txt_attn.qkv.bias
378
+ model.diffusion_model.double_blocks.6.txt_attn.qkv.weight
379
+ model.diffusion_model.double_blocks.6.txt_mlp.0.bias
380
+ model.diffusion_model.double_blocks.6.txt_mlp.0.weight
381
+ model.diffusion_model.double_blocks.6.txt_mlp.2.bias
382
+ model.diffusion_model.double_blocks.6.txt_mlp.2.weight
383
+ model.diffusion_model.double_blocks.6.txt_mod.lin.bias
384
+ model.diffusion_model.double_blocks.6.txt_mod.lin.weight
385
+ model.diffusion_model.double_blocks.7.img_attn.norm.key_norm.scale
386
+ model.diffusion_model.double_blocks.7.img_attn.norm.query_norm.scale
387
+ model.diffusion_model.double_blocks.7.img_attn.proj.bias
388
+ model.diffusion_model.double_blocks.7.img_attn.proj.weight
389
+ model.diffusion_model.double_blocks.7.img_attn.qkv.bias
390
+ model.diffusion_model.double_blocks.7.img_attn.qkv.weight
391
+ model.diffusion_model.double_blocks.7.img_mlp.0.bias
392
+ model.diffusion_model.double_blocks.7.img_mlp.0.weight
393
+ model.diffusion_model.double_blocks.7.img_mlp.2.bias
394
+ model.diffusion_model.double_blocks.7.img_mlp.2.weight
395
+ model.diffusion_model.double_blocks.7.img_mod.lin.bias
396
+ model.diffusion_model.double_blocks.7.img_mod.lin.weight
397
+ model.diffusion_model.double_blocks.7.txt_attn.norm.key_norm.scale
398
+ model.diffusion_model.double_blocks.7.txt_attn.norm.query_norm.scale
399
+ model.diffusion_model.double_blocks.7.txt_attn.proj.bias
400
+ model.diffusion_model.double_blocks.7.txt_attn.proj.weight
401
+ model.diffusion_model.double_blocks.7.txt_attn.qkv.bias
402
+ model.diffusion_model.double_blocks.7.txt_attn.qkv.weight
403
+ model.diffusion_model.double_blocks.7.txt_mlp.0.bias
404
+ model.diffusion_model.double_blocks.7.txt_mlp.0.weight
405
+ model.diffusion_model.double_blocks.7.txt_mlp.2.bias
406
+ model.diffusion_model.double_blocks.7.txt_mlp.2.weight
407
+ model.diffusion_model.double_blocks.7.txt_mod.lin.bias
408
+ model.diffusion_model.double_blocks.7.txt_mod.lin.weight
409
+ model.diffusion_model.double_blocks.8.img_attn.norm.key_norm.scale
410
+ model.diffusion_model.double_blocks.8.img_attn.norm.query_norm.scale
411
+ model.diffusion_model.double_blocks.8.img_attn.proj.bias
412
+ model.diffusion_model.double_blocks.8.img_attn.proj.weight
413
+ model.diffusion_model.double_blocks.8.img_attn.qkv.bias
414
+ model.diffusion_model.double_blocks.8.img_attn.qkv.weight
415
+ model.diffusion_model.double_blocks.8.img_mlp.0.bias
416
+ model.diffusion_model.double_blocks.8.img_mlp.0.weight
417
+ model.diffusion_model.double_blocks.8.img_mlp.2.bias
418
+ model.diffusion_model.double_blocks.8.img_mlp.2.weight
419
+ model.diffusion_model.double_blocks.8.img_mod.lin.bias
420
+ model.diffusion_model.double_blocks.8.img_mod.lin.weight
421
+ model.diffusion_model.double_blocks.8.txt_attn.norm.key_norm.scale
422
+ model.diffusion_model.double_blocks.8.txt_attn.norm.query_norm.scale
423
+ model.diffusion_model.double_blocks.8.txt_attn.proj.bias
424
+ model.diffusion_model.double_blocks.8.txt_attn.proj.weight
425
+ model.diffusion_model.double_blocks.8.txt_attn.qkv.bias
426
+ model.diffusion_model.double_blocks.8.txt_attn.qkv.weight
427
+ model.diffusion_model.double_blocks.8.txt_mlp.0.bias
428
+ model.diffusion_model.double_blocks.8.txt_mlp.0.weight
429
+ model.diffusion_model.double_blocks.8.txt_mlp.2.bias
430
+ model.diffusion_model.double_blocks.8.txt_mlp.2.weight
431
+ model.diffusion_model.double_blocks.8.txt_mod.lin.bias
432
+ model.diffusion_model.double_blocks.8.txt_mod.lin.weight
433
+ model.diffusion_model.double_blocks.9.img_attn.norm.key_norm.scale
434
+ model.diffusion_model.double_blocks.9.img_attn.norm.query_norm.scale
435
+ model.diffusion_model.double_blocks.9.img_attn.proj.bias
436
+ model.diffusion_model.double_blocks.9.img_attn.proj.weight
437
+ model.diffusion_model.double_blocks.9.img_attn.qkv.bias
438
+ model.diffusion_model.double_blocks.9.img_attn.qkv.weight
439
+ model.diffusion_model.double_blocks.9.img_mlp.0.bias
440
+ model.diffusion_model.double_blocks.9.img_mlp.0.weight
441
+ model.diffusion_model.double_blocks.9.img_mlp.2.bias
442
+ model.diffusion_model.double_blocks.9.img_mlp.2.weight
443
+ model.diffusion_model.double_blocks.9.img_mod.lin.bias
444
+ model.diffusion_model.double_blocks.9.img_mod.lin.weight
445
+ model.diffusion_model.double_blocks.9.txt_attn.norm.key_norm.scale
446
+ model.diffusion_model.double_blocks.9.txt_attn.norm.query_norm.scale
447
+ model.diffusion_model.double_blocks.9.txt_attn.proj.bias
448
+ model.diffusion_model.double_blocks.9.txt_attn.proj.weight
449
+ model.diffusion_model.double_blocks.9.txt_attn.qkv.bias
450
+ model.diffusion_model.double_blocks.9.txt_attn.qkv.weight
451
+ model.diffusion_model.double_blocks.9.txt_mlp.0.bias
452
+ model.diffusion_model.double_blocks.9.txt_mlp.0.weight
453
+ model.diffusion_model.double_blocks.9.txt_mlp.2.bias
454
+ model.diffusion_model.double_blocks.9.txt_mlp.2.weight
455
+ model.diffusion_model.double_blocks.9.txt_mod.lin.bias
456
+ model.diffusion_model.double_blocks.9.txt_mod.lin.weight
457
+ model.diffusion_model.final_layer.adaLN_modulation.1.bias
458
+ model.diffusion_model.final_layer.adaLN_modulation.1.weight
459
+ model.diffusion_model.final_layer.linear.bias
460
+ model.diffusion_model.final_layer.linear.weight
461
+ model.diffusion_model.guidance_in.in_layer.bias
462
+ model.diffusion_model.guidance_in.in_layer.weight
463
+ model.diffusion_model.guidance_in.out_layer.bias
464
+ model.diffusion_model.guidance_in.out_layer.weight
465
+ model.diffusion_model.img_in.bias
466
+ model.diffusion_model.img_in.weight
467
+ model.diffusion_model.single_blocks.0.linear1.bias
468
+ model.diffusion_model.single_blocks.0.linear1.weight
469
+ model.diffusion_model.single_blocks.0.linear2.bias
470
+ model.diffusion_model.single_blocks.0.linear2.weight
471
+ model.diffusion_model.single_blocks.0.modulation.lin.bias
472
+ model.diffusion_model.single_blocks.0.modulation.lin.weight
473
+ model.diffusion_model.single_blocks.0.norm.key_norm.scale
474
+ model.diffusion_model.single_blocks.0.norm.query_norm.scale
475
+ model.diffusion_model.single_blocks.1.linear1.bias
476
+ model.diffusion_model.single_blocks.1.linear1.weight
477
+ model.diffusion_model.single_blocks.1.linear2.bias
478
+ model.diffusion_model.single_blocks.1.linear2.weight
479
+ model.diffusion_model.single_blocks.1.modulation.lin.bias
480
+ model.diffusion_model.single_blocks.1.modulation.lin.weight
481
+ model.diffusion_model.single_blocks.1.norm.key_norm.scale
482
+ model.diffusion_model.single_blocks.1.norm.query_norm.scale
483
+ model.diffusion_model.single_blocks.10.linear1.bias
484
+ model.diffusion_model.single_blocks.10.linear1.weight
485
+ model.diffusion_model.single_blocks.10.linear2.bias
486
+ model.diffusion_model.single_blocks.10.linear2.weight
487
+ model.diffusion_model.single_blocks.10.modulation.lin.bias
488
+ model.diffusion_model.single_blocks.10.modulation.lin.weight
489
+ model.diffusion_model.single_blocks.10.norm.key_norm.scale
490
+ model.diffusion_model.single_blocks.10.norm.query_norm.scale
491
+ model.diffusion_model.single_blocks.11.linear1.bias
492
+ model.diffusion_model.single_blocks.11.linear1.weight
493
+ model.diffusion_model.single_blocks.11.linear2.bias
494
+ model.diffusion_model.single_blocks.11.linear2.weight
495
+ model.diffusion_model.single_blocks.11.modulation.lin.bias
496
+ model.diffusion_model.single_blocks.11.modulation.lin.weight
497
+ model.diffusion_model.single_blocks.11.norm.key_norm.scale
498
+ model.diffusion_model.single_blocks.11.norm.query_norm.scale
499
+ model.diffusion_model.single_blocks.12.linear1.bias
500
+ model.diffusion_model.single_blocks.12.linear1.weight
501
+ model.diffusion_model.single_blocks.12.linear2.bias
502
+ model.diffusion_model.single_blocks.12.linear2.weight
503
+ model.diffusion_model.single_blocks.12.modulation.lin.bias
504
+ model.diffusion_model.single_blocks.12.modulation.lin.weight
505
+ model.diffusion_model.single_blocks.12.norm.key_norm.scale
506
+ model.diffusion_model.single_blocks.12.norm.query_norm.scale
507
+ model.diffusion_model.single_blocks.13.linear1.bias
508
+ model.diffusion_model.single_blocks.13.linear1.weight
509
+ model.diffusion_model.single_blocks.13.linear2.bias
510
+ model.diffusion_model.single_blocks.13.linear2.weight
511
+ model.diffusion_model.single_blocks.13.modulation.lin.bias
512
+ model.diffusion_model.single_blocks.13.modulation.lin.weight
513
+ model.diffusion_model.single_blocks.13.norm.key_norm.scale
514
+ model.diffusion_model.single_blocks.13.norm.query_norm.scale
515
+ model.diffusion_model.single_blocks.14.linear1.bias
516
+ model.diffusion_model.single_blocks.14.linear1.weight
517
+ model.diffusion_model.single_blocks.14.linear2.bias
518
+ model.diffusion_model.single_blocks.14.linear2.weight
519
+ model.diffusion_model.single_blocks.14.modulation.lin.bias
520
+ model.diffusion_model.single_blocks.14.modulation.lin.weight
521
+ model.diffusion_model.single_blocks.14.norm.key_norm.scale
522
+ model.diffusion_model.single_blocks.14.norm.query_norm.scale
523
+ model.diffusion_model.single_blocks.15.linear1.bias
524
+ model.diffusion_model.single_blocks.15.linear1.weight
525
+ model.diffusion_model.single_blocks.15.linear2.bias
526
+ model.diffusion_model.single_blocks.15.linear2.weight
527
+ model.diffusion_model.single_blocks.15.modulation.lin.bias
528
+ model.diffusion_model.single_blocks.15.modulation.lin.weight
529
+ model.diffusion_model.single_blocks.15.norm.key_norm.scale
530
+ model.diffusion_model.single_blocks.15.norm.query_norm.scale
531
+ model.diffusion_model.single_blocks.16.linear1.bias
532
+ model.diffusion_model.single_blocks.16.linear1.weight
533
+ model.diffusion_model.single_blocks.16.linear2.bias
534
+ model.diffusion_model.single_blocks.16.linear2.weight
535
+ model.diffusion_model.single_blocks.16.modulation.lin.bias
536
+ model.diffusion_model.single_blocks.16.modulation.lin.weight
537
+ model.diffusion_model.single_blocks.16.norm.key_norm.scale
538
+ model.diffusion_model.single_blocks.16.norm.query_norm.scale
539
+ model.diffusion_model.single_blocks.17.linear1.bias
540
+ model.diffusion_model.single_blocks.17.linear1.weight
541
+ model.diffusion_model.single_blocks.17.linear2.bias
542
+ model.diffusion_model.single_blocks.17.linear2.weight
543
+ model.diffusion_model.single_blocks.17.modulation.lin.bias
544
+ model.diffusion_model.single_blocks.17.modulation.lin.weight
545
+ model.diffusion_model.single_blocks.17.norm.key_norm.scale
546
+ model.diffusion_model.single_blocks.17.norm.query_norm.scale
547
+ model.diffusion_model.single_blocks.18.linear1.bias
548
+ model.diffusion_model.single_blocks.18.linear1.weight
549
+ model.diffusion_model.single_blocks.18.linear2.bias
550
+ model.diffusion_model.single_blocks.18.linear2.weight
551
+ model.diffusion_model.single_blocks.18.modulation.lin.bias
552
+ model.diffusion_model.single_blocks.18.modulation.lin.weight
553
+ model.diffusion_model.single_blocks.18.norm.key_norm.scale
554
+ model.diffusion_model.single_blocks.18.norm.query_norm.scale
555
+ model.diffusion_model.single_blocks.19.linear1.bias
556
+ model.diffusion_model.single_blocks.19.linear1.weight
557
+ model.diffusion_model.single_blocks.19.linear2.bias
558
+ model.diffusion_model.single_blocks.19.linear2.weight
559
+ model.diffusion_model.single_blocks.19.modulation.lin.bias
560
+ model.diffusion_model.single_blocks.19.modulation.lin.weight
561
+ model.diffusion_model.single_blocks.19.norm.key_norm.scale
562
+ model.diffusion_model.single_blocks.19.norm.query_norm.scale
563
+ model.diffusion_model.single_blocks.2.linear1.bias
564
+ model.diffusion_model.single_blocks.2.linear1.weight
565
+ model.diffusion_model.single_blocks.2.linear2.bias
566
+ model.diffusion_model.single_blocks.2.linear2.weight
567
+ model.diffusion_model.single_blocks.2.modulation.lin.bias
568
+ model.diffusion_model.single_blocks.2.modulation.lin.weight
569
+ model.diffusion_model.single_blocks.2.norm.key_norm.scale
570
+ model.diffusion_model.single_blocks.2.norm.query_norm.scale
571
+ model.diffusion_model.single_blocks.20.linear1.bias
572
+ model.diffusion_model.single_blocks.20.linear1.weight
573
+ model.diffusion_model.single_blocks.20.linear2.bias
574
+ model.diffusion_model.single_blocks.20.linear2.weight
575
+ model.diffusion_model.single_blocks.20.modulation.lin.bias
576
+ model.diffusion_model.single_blocks.20.modulation.lin.weight
577
+ model.diffusion_model.single_blocks.20.norm.key_norm.scale
578
+ model.diffusion_model.single_blocks.20.norm.query_norm.scale
579
+ model.diffusion_model.single_blocks.21.linear1.bias
580
+ model.diffusion_model.single_blocks.21.linear1.weight
581
+ model.diffusion_model.single_blocks.21.linear2.bias
582
+ model.diffusion_model.single_blocks.21.linear2.weight
583
+ model.diffusion_model.single_blocks.21.modulation.lin.bias
584
+ model.diffusion_model.single_blocks.21.modulation.lin.weight
585
+ model.diffusion_model.single_blocks.21.norm.key_norm.scale
586
+ model.diffusion_model.single_blocks.21.norm.query_norm.scale
587
+ model.diffusion_model.single_blocks.22.linear1.bias
588
+ model.diffusion_model.single_blocks.22.linear1.weight
589
+ model.diffusion_model.single_blocks.22.linear2.bias
590
+ model.diffusion_model.single_blocks.22.linear2.weight
591
+ model.diffusion_model.single_blocks.22.modulation.lin.bias
592
+ model.diffusion_model.single_blocks.22.modulation.lin.weight
593
+ model.diffusion_model.single_blocks.22.norm.key_norm.scale
594
+ model.diffusion_model.single_blocks.22.norm.query_norm.scale
595
+ model.diffusion_model.single_blocks.23.linear1.bias
596
+ model.diffusion_model.single_blocks.23.linear1.weight
597
+ model.diffusion_model.single_blocks.23.linear2.bias
598
+ model.diffusion_model.single_blocks.23.linear2.weight
599
+ model.diffusion_model.single_blocks.23.modulation.lin.bias
600
+ model.diffusion_model.single_blocks.23.modulation.lin.weight
601
+ model.diffusion_model.single_blocks.23.norm.key_norm.scale
602
+ model.diffusion_model.single_blocks.23.norm.query_norm.scale
603
+ model.diffusion_model.single_blocks.24.linear1.bias
604
+ model.diffusion_model.single_blocks.24.linear1.weight
605
+ model.diffusion_model.single_blocks.24.linear2.bias
606
+ model.diffusion_model.single_blocks.24.linear2.weight
607
+ model.diffusion_model.single_blocks.24.modulation.lin.bias
608
+ model.diffusion_model.single_blocks.24.modulation.lin.weight
609
+ model.diffusion_model.single_blocks.24.norm.key_norm.scale
610
+ model.diffusion_model.single_blocks.24.norm.query_norm.scale
611
+ model.diffusion_model.single_blocks.25.linear1.bias
612
+ model.diffusion_model.single_blocks.25.linear1.weight
613
+ model.diffusion_model.single_blocks.25.linear2.bias
614
+ model.diffusion_model.single_blocks.25.linear2.weight
615
+ model.diffusion_model.single_blocks.25.modulation.lin.bias
616
+ model.diffusion_model.single_blocks.25.modulation.lin.weight
617
+ model.diffusion_model.single_blocks.25.norm.key_norm.scale
618
+ model.diffusion_model.single_blocks.25.norm.query_norm.scale
619
+ model.diffusion_model.single_blocks.26.linear1.bias
620
+ model.diffusion_model.single_blocks.26.linear1.weight
621
+ model.diffusion_model.single_blocks.26.linear2.bias
622
+ model.diffusion_model.single_blocks.26.linear2.weight
623
+ model.diffusion_model.single_blocks.26.modulation.lin.bias
624
+ model.diffusion_model.single_blocks.26.modulation.lin.weight
625
+ model.diffusion_model.single_blocks.26.norm.key_norm.scale
626
+ model.diffusion_model.single_blocks.26.norm.query_norm.scale
627
+ model.diffusion_model.single_blocks.27.linear1.bias
628
+ model.diffusion_model.single_blocks.27.linear1.weight
629
+ model.diffusion_model.single_blocks.27.linear2.bias
630
+ model.diffusion_model.single_blocks.27.linear2.weight
631
+ model.diffusion_model.single_blocks.27.modulation.lin.bias
632
+ model.diffusion_model.single_blocks.27.modulation.lin.weight
633
+ model.diffusion_model.single_blocks.27.norm.key_norm.scale
634
+ model.diffusion_model.single_blocks.27.norm.query_norm.scale
635
+ model.diffusion_model.single_blocks.28.linear1.bias
636
+ model.diffusion_model.single_blocks.28.linear1.weight
637
+ model.diffusion_model.single_blocks.28.linear2.bias
638
+ model.diffusion_model.single_blocks.28.linear2.weight
639
+ model.diffusion_model.single_blocks.28.modulation.lin.bias
640
+ model.diffusion_model.single_blocks.28.modulation.lin.weight
641
+ model.diffusion_model.single_blocks.28.norm.key_norm.scale
642
+ model.diffusion_model.single_blocks.28.norm.query_norm.scale
643
+ model.diffusion_model.single_blocks.29.linear1.bias
644
+ model.diffusion_model.single_blocks.29.linear1.weight
645
+ model.diffusion_model.single_blocks.29.linear2.bias
646
+ model.diffusion_model.single_blocks.29.linear2.weight
647
+ model.diffusion_model.single_blocks.29.modulation.lin.bias
648
+ model.diffusion_model.single_blocks.29.modulation.lin.weight
649
+ model.diffusion_model.single_blocks.29.norm.key_norm.scale
650
+ model.diffusion_model.single_blocks.29.norm.query_norm.scale
651
+ model.diffusion_model.single_blocks.3.linear1.bias
652
+ model.diffusion_model.single_blocks.3.linear1.weight
653
+ model.diffusion_model.single_blocks.3.linear2.bias
654
+ model.diffusion_model.single_blocks.3.linear2.weight
655
+ model.diffusion_model.single_blocks.3.modulation.lin.bias
656
+ model.diffusion_model.single_blocks.3.modulation.lin.weight
657
+ model.diffusion_model.single_blocks.3.norm.key_norm.scale
658
+ model.diffusion_model.single_blocks.3.norm.query_norm.scale
659
+ model.diffusion_model.single_blocks.30.linear1.bias
660
+ model.diffusion_model.single_blocks.30.linear1.weight
661
+ model.diffusion_model.single_blocks.30.linear2.bias
662
+ model.diffusion_model.single_blocks.30.linear2.weight
663
+ model.diffusion_model.single_blocks.30.modulation.lin.bias
664
+ model.diffusion_model.single_blocks.30.modulation.lin.weight
665
+ model.diffusion_model.single_blocks.30.norm.key_norm.scale
666
+ model.diffusion_model.single_blocks.30.norm.query_norm.scale
667
+ model.diffusion_model.single_blocks.31.linear1.bias
668
+ model.diffusion_model.single_blocks.31.linear1.weight
669
+ model.diffusion_model.single_blocks.31.linear2.bias
670
+ model.diffusion_model.single_blocks.31.linear2.weight
671
+ model.diffusion_model.single_blocks.31.modulation.lin.bias
672
+ model.diffusion_model.single_blocks.31.modulation.lin.weight
673
+ model.diffusion_model.single_blocks.31.norm.key_norm.scale
674
+ model.diffusion_model.single_blocks.31.norm.query_norm.scale
675
+ model.diffusion_model.single_blocks.32.linear1.bias
676
+ model.diffusion_model.single_blocks.32.linear1.weight
677
+ model.diffusion_model.single_blocks.32.linear2.bias
678
+ model.diffusion_model.single_blocks.32.linear2.weight
679
+ model.diffusion_model.single_blocks.32.modulation.lin.bias
680
+ model.diffusion_model.single_blocks.32.modulation.lin.weight
681
+ model.diffusion_model.single_blocks.32.norm.key_norm.scale
682
+ model.diffusion_model.single_blocks.32.norm.query_norm.scale
683
+ model.diffusion_model.single_blocks.33.linear1.bias
684
+ model.diffusion_model.single_blocks.33.linear1.weight
685
+ model.diffusion_model.single_blocks.33.linear2.bias
686
+ model.diffusion_model.single_blocks.33.linear2.weight
687
+ model.diffusion_model.single_blocks.33.modulation.lin.bias
688
+ model.diffusion_model.single_blocks.33.modulation.lin.weight
689
+ model.diffusion_model.single_blocks.33.norm.key_norm.scale
690
+ model.diffusion_model.single_blocks.33.norm.query_norm.scale
691
+ model.diffusion_model.single_blocks.34.linear1.bias
692
+ model.diffusion_model.single_blocks.34.linear1.weight
693
+ model.diffusion_model.single_blocks.34.linear2.bias
694
+ model.diffusion_model.single_blocks.34.linear2.weight
695
+ model.diffusion_model.single_blocks.34.modulation.lin.bias
696
+ model.diffusion_model.single_blocks.34.modulation.lin.weight
697
+ model.diffusion_model.single_blocks.34.norm.key_norm.scale
698
+ model.diffusion_model.single_blocks.34.norm.query_norm.scale
699
+ model.diffusion_model.single_blocks.35.linear1.bias
700
+ model.diffusion_model.single_blocks.35.linear1.weight
701
+ model.diffusion_model.single_blocks.35.linear2.bias
702
+ model.diffusion_model.single_blocks.35.linear2.weight
703
+ model.diffusion_model.single_blocks.35.modulation.lin.bias
704
+ model.diffusion_model.single_blocks.35.modulation.lin.weight
705
+ model.diffusion_model.single_blocks.35.norm.key_norm.scale
706
+ model.diffusion_model.single_blocks.35.norm.query_norm.scale
707
+ model.diffusion_model.single_blocks.36.linear1.bias
708
+ model.diffusion_model.single_blocks.36.linear1.weight
709
+ model.diffusion_model.single_blocks.36.linear2.bias
710
+ model.diffusion_model.single_blocks.36.linear2.weight
711
+ model.diffusion_model.single_blocks.36.modulation.lin.bias
712
+ model.diffusion_model.single_blocks.36.modulation.lin.weight
713
+ model.diffusion_model.single_blocks.36.norm.key_norm.scale
714
+ model.diffusion_model.single_blocks.36.norm.query_norm.scale
715
+ model.diffusion_model.single_blocks.37.linear1.bias
716
+ model.diffusion_model.single_blocks.37.linear1.weight
717
+ model.diffusion_model.single_blocks.37.linear2.bias
718
+ model.diffusion_model.single_blocks.37.linear2.weight
719
+ model.diffusion_model.single_blocks.37.modulation.lin.bias
720
+ model.diffusion_model.single_blocks.37.modulation.lin.weight
721
+ model.diffusion_model.single_blocks.37.norm.key_norm.scale
722
+ model.diffusion_model.single_blocks.37.norm.query_norm.scale
723
+ model.diffusion_model.single_blocks.4.linear1.bias
724
+ model.diffusion_model.single_blocks.4.linear1.weight
725
+ model.diffusion_model.single_blocks.4.linear2.bias
726
+ model.diffusion_model.single_blocks.4.linear2.weight
727
+ model.diffusion_model.single_blocks.4.modulation.lin.bias
728
+ model.diffusion_model.single_blocks.4.modulation.lin.weight
729
+ model.diffusion_model.single_blocks.4.norm.key_norm.scale
730
+ model.diffusion_model.single_blocks.4.norm.query_norm.scale
731
+ model.diffusion_model.single_blocks.5.linear1.bias
732
+ model.diffusion_model.single_blocks.5.linear1.weight
733
+ model.diffusion_model.single_blocks.5.linear2.bias
734
+ model.diffusion_model.single_blocks.5.linear2.weight
735
+ model.diffusion_model.single_blocks.5.modulation.lin.bias
736
+ model.diffusion_model.single_blocks.5.modulation.lin.weight
737
+ model.diffusion_model.single_blocks.5.norm.key_norm.scale
738
+ model.diffusion_model.single_blocks.5.norm.query_norm.scale
739
+ model.diffusion_model.single_blocks.6.linear1.bias
740
+ model.diffusion_model.single_blocks.6.linear1.weight
741
+ model.diffusion_model.single_blocks.6.linear2.bias
742
+ model.diffusion_model.single_blocks.6.linear2.weight
743
+ model.diffusion_model.single_blocks.6.modulation.lin.bias
744
+ model.diffusion_model.single_blocks.6.modulation.lin.weight
745
+ model.diffusion_model.single_blocks.6.norm.key_norm.scale
746
+ model.diffusion_model.single_blocks.6.norm.query_norm.scale
747
+ model.diffusion_model.single_blocks.7.linear1.bias
748
+ model.diffusion_model.single_blocks.7.linear1.weight
749
+ model.diffusion_model.single_blocks.7.linear2.bias
750
+ model.diffusion_model.single_blocks.7.linear2.weight
751
+ model.diffusion_model.single_blocks.7.modulation.lin.bias
752
+ model.diffusion_model.single_blocks.7.modulation.lin.weight
753
+ model.diffusion_model.single_blocks.7.norm.key_norm.scale
754
+ model.diffusion_model.single_blocks.7.norm.query_norm.scale
755
+ model.diffusion_model.single_blocks.8.linear1.bias
756
+ model.diffusion_model.single_blocks.8.linear1.weight
757
+ model.diffusion_model.single_blocks.8.linear2.bias
758
+ model.diffusion_model.single_blocks.8.linear2.weight
759
+ model.diffusion_model.single_blocks.8.modulation.lin.bias
760
+ model.diffusion_model.single_blocks.8.modulation.lin.weight
761
+ model.diffusion_model.single_blocks.8.norm.key_norm.scale
762
+ model.diffusion_model.single_blocks.8.norm.query_norm.scale
763
+ model.diffusion_model.single_blocks.9.linear1.bias
764
+ model.diffusion_model.single_blocks.9.linear1.weight
765
+ model.diffusion_model.single_blocks.9.linear2.bias
766
+ model.diffusion_model.single_blocks.9.linear2.weight
767
+ model.diffusion_model.single_blocks.9.modulation.lin.bias
768
+ model.diffusion_model.single_blocks.9.modulation.lin.weight
769
+ model.diffusion_model.single_blocks.9.norm.key_norm.scale
770
+ model.diffusion_model.single_blocks.9.norm.query_norm.scale
771
+ model.diffusion_model.time_in.in_layer.bias
772
+ model.diffusion_model.time_in.in_layer.weight
773
+ model.diffusion_model.time_in.out_layer.bias
774
+ model.diffusion_model.time_in.out_layer.weight
775
+ model.diffusion_model.txt_in.bias
776
+ model.diffusion_model.txt_in.weight
777
+ model.diffusion_model.vector_in.in_layer.bias
778
+ model.diffusion_model.vector_in.in_layer.weight
779
+ model.diffusion_model.vector_in.out_layer.bias
780
+ model.diffusion_model.vector_in.out_layer.weight
781
+ text_encoders.clip_l.logit_scale
782
+ text_encoders.clip_l.transformer.text_model.embeddings.position_embedding.weight
783
+ text_encoders.clip_l.transformer.text_model.embeddings.token_embedding.weight
784
+ text_encoders.clip_l.transformer.text_model.encoder.layers.0.layer_norm1.bias
785
+ text_encoders.clip_l.transformer.text_model.encoder.layers.0.layer_norm1.weight
786
+ text_encoders.clip_l.transformer.text_model.encoder.layers.0.layer_norm2.bias
787
+ text_encoders.clip_l.transformer.text_model.encoder.layers.0.layer_norm2.weight
788
+ text_encoders.clip_l.transformer.text_model.encoder.layers.0.mlp.fc1.bias
789
+ text_encoders.clip_l.transformer.text_model.encoder.layers.0.mlp.fc1.weight
790
+ text_encoders.clip_l.transformer.text_model.encoder.layers.0.mlp.fc2.bias
791
+ text_encoders.clip_l.transformer.text_model.encoder.layers.0.mlp.fc2.weight
792
+ text_encoders.clip_l.transformer.text_model.encoder.layers.0.self_attn.k_proj.bias
793
+ text_encoders.clip_l.transformer.text_model.encoder.layers.0.self_attn.k_proj.weight
794
+ text_encoders.clip_l.transformer.text_model.encoder.layers.0.self_attn.out_proj.bias
795
+ text_encoders.clip_l.transformer.text_model.encoder.layers.0.self_attn.out_proj.weight
796
+ text_encoders.clip_l.transformer.text_model.encoder.layers.0.self_attn.q_proj.bias
797
+ text_encoders.clip_l.transformer.text_model.encoder.layers.0.self_attn.q_proj.weight
798
+ text_encoders.clip_l.transformer.text_model.encoder.layers.0.self_attn.v_proj.bias
799
+ text_encoders.clip_l.transformer.text_model.encoder.layers.0.self_attn.v_proj.weight
800
+ text_encoders.clip_l.transformer.text_model.encoder.layers.1.layer_norm1.bias
801
+ text_encoders.clip_l.transformer.text_model.encoder.layers.1.layer_norm1.weight
802
+ text_encoders.clip_l.transformer.text_model.encoder.layers.1.layer_norm2.bias
803
+ text_encoders.clip_l.transformer.text_model.encoder.layers.1.layer_norm2.weight
804
+ text_encoders.clip_l.transformer.text_model.encoder.layers.1.mlp.fc1.bias
805
+ text_encoders.clip_l.transformer.text_model.encoder.layers.1.mlp.fc1.weight
806
+ text_encoders.clip_l.transformer.text_model.encoder.layers.1.mlp.fc2.bias
807
+ text_encoders.clip_l.transformer.text_model.encoder.layers.1.mlp.fc2.weight
808
+ text_encoders.clip_l.transformer.text_model.encoder.layers.1.self_attn.k_proj.bias
809
+ text_encoders.clip_l.transformer.text_model.encoder.layers.1.self_attn.k_proj.weight
810
+ text_encoders.clip_l.transformer.text_model.encoder.layers.1.self_attn.out_proj.bias
811
+ text_encoders.clip_l.transformer.text_model.encoder.layers.1.self_attn.out_proj.weight
812
+ text_encoders.clip_l.transformer.text_model.encoder.layers.1.self_attn.q_proj.bias
813
+ text_encoders.clip_l.transformer.text_model.encoder.layers.1.self_attn.q_proj.weight
814
+ text_encoders.clip_l.transformer.text_model.encoder.layers.1.self_attn.v_proj.bias
815
+ text_encoders.clip_l.transformer.text_model.encoder.layers.1.self_attn.v_proj.weight
816
+ text_encoders.clip_l.transformer.text_model.encoder.layers.10.layer_norm1.bias
817
+ text_encoders.clip_l.transformer.text_model.encoder.layers.10.layer_norm1.weight
818
+ text_encoders.clip_l.transformer.text_model.encoder.layers.10.layer_norm2.bias
819
+ text_encoders.clip_l.transformer.text_model.encoder.layers.10.layer_norm2.weight
820
+ text_encoders.clip_l.transformer.text_model.encoder.layers.10.mlp.fc1.bias
821
+ text_encoders.clip_l.transformer.text_model.encoder.layers.10.mlp.fc1.weight
822
+ text_encoders.clip_l.transformer.text_model.encoder.layers.10.mlp.fc2.bias
823
+ text_encoders.clip_l.transformer.text_model.encoder.layers.10.mlp.fc2.weight
824
+ text_encoders.clip_l.transformer.text_model.encoder.layers.10.self_attn.k_proj.bias
825
+ text_encoders.clip_l.transformer.text_model.encoder.layers.10.self_attn.k_proj.weight
826
+ text_encoders.clip_l.transformer.text_model.encoder.layers.10.self_attn.out_proj.bias
827
+ text_encoders.clip_l.transformer.text_model.encoder.layers.10.self_attn.out_proj.weight
828
+ text_encoders.clip_l.transformer.text_model.encoder.layers.10.self_attn.q_proj.bias
829
+ text_encoders.clip_l.transformer.text_model.encoder.layers.10.self_attn.q_proj.weight
830
+ text_encoders.clip_l.transformer.text_model.encoder.layers.10.self_attn.v_proj.bias
831
+ text_encoders.clip_l.transformer.text_model.encoder.layers.10.self_attn.v_proj.weight
832
+ text_encoders.clip_l.transformer.text_model.encoder.layers.11.layer_norm1.bias
833
+ text_encoders.clip_l.transformer.text_model.encoder.layers.11.layer_norm1.weight
834
+ text_encoders.clip_l.transformer.text_model.encoder.layers.11.layer_norm2.bias
835
+ text_encoders.clip_l.transformer.text_model.encoder.layers.11.layer_norm2.weight
836
+ text_encoders.clip_l.transformer.text_model.encoder.layers.11.mlp.fc1.bias
837
+ text_encoders.clip_l.transformer.text_model.encoder.layers.11.mlp.fc1.weight
838
+ text_encoders.clip_l.transformer.text_model.encoder.layers.11.mlp.fc2.bias
839
+ text_encoders.clip_l.transformer.text_model.encoder.layers.11.mlp.fc2.weight
840
+ text_encoders.clip_l.transformer.text_model.encoder.layers.11.self_attn.k_proj.bias
841
+ text_encoders.clip_l.transformer.text_model.encoder.layers.11.self_attn.k_proj.weight
842
+ text_encoders.clip_l.transformer.text_model.encoder.layers.11.self_attn.out_proj.bias
843
+ text_encoders.clip_l.transformer.text_model.encoder.layers.11.self_attn.out_proj.weight
844
+ text_encoders.clip_l.transformer.text_model.encoder.layers.11.self_attn.q_proj.bias
845
+ text_encoders.clip_l.transformer.text_model.encoder.layers.11.self_attn.q_proj.weight
846
+ text_encoders.clip_l.transformer.text_model.encoder.layers.11.self_attn.v_proj.bias
847
+ text_encoders.clip_l.transformer.text_model.encoder.layers.11.self_attn.v_proj.weight
848
+ text_encoders.clip_l.transformer.text_model.encoder.layers.2.layer_norm1.bias
849
+ text_encoders.clip_l.transformer.text_model.encoder.layers.2.layer_norm1.weight
850
+ text_encoders.clip_l.transformer.text_model.encoder.layers.2.layer_norm2.bias
851
+ text_encoders.clip_l.transformer.text_model.encoder.layers.2.layer_norm2.weight
852
+ text_encoders.clip_l.transformer.text_model.encoder.layers.2.mlp.fc1.bias
853
+ text_encoders.clip_l.transformer.text_model.encoder.layers.2.mlp.fc1.weight
854
+ text_encoders.clip_l.transformer.text_model.encoder.layers.2.mlp.fc2.bias
855
+ text_encoders.clip_l.transformer.text_model.encoder.layers.2.mlp.fc2.weight
856
+ text_encoders.clip_l.transformer.text_model.encoder.layers.2.self_attn.k_proj.bias
857
+ text_encoders.clip_l.transformer.text_model.encoder.layers.2.self_attn.k_proj.weight
858
+ text_encoders.clip_l.transformer.text_model.encoder.layers.2.self_attn.out_proj.bias
859
+ text_encoders.clip_l.transformer.text_model.encoder.layers.2.self_attn.out_proj.weight
860
+ text_encoders.clip_l.transformer.text_model.encoder.layers.2.self_attn.q_proj.bias
861
+ text_encoders.clip_l.transformer.text_model.encoder.layers.2.self_attn.q_proj.weight
862
+ text_encoders.clip_l.transformer.text_model.encoder.layers.2.self_attn.v_proj.bias
863
+ text_encoders.clip_l.transformer.text_model.encoder.layers.2.self_attn.v_proj.weight
864
+ text_encoders.clip_l.transformer.text_model.encoder.layers.3.layer_norm1.bias
865
+ text_encoders.clip_l.transformer.text_model.encoder.layers.3.layer_norm1.weight
866
+ text_encoders.clip_l.transformer.text_model.encoder.layers.3.layer_norm2.bias
867
+ text_encoders.clip_l.transformer.text_model.encoder.layers.3.layer_norm2.weight
868
+ text_encoders.clip_l.transformer.text_model.encoder.layers.3.mlp.fc1.bias
869
+ text_encoders.clip_l.transformer.text_model.encoder.layers.3.mlp.fc1.weight
870
+ text_encoders.clip_l.transformer.text_model.encoder.layers.3.mlp.fc2.bias
871
+ text_encoders.clip_l.transformer.text_model.encoder.layers.3.mlp.fc2.weight
872
+ text_encoders.clip_l.transformer.text_model.encoder.layers.3.self_attn.k_proj.bias
873
+ text_encoders.clip_l.transformer.text_model.encoder.layers.3.self_attn.k_proj.weight
874
+ text_encoders.clip_l.transformer.text_model.encoder.layers.3.self_attn.out_proj.bias
875
+ text_encoders.clip_l.transformer.text_model.encoder.layers.3.self_attn.out_proj.weight
876
+ text_encoders.clip_l.transformer.text_model.encoder.layers.3.self_attn.q_proj.bias
877
+ text_encoders.clip_l.transformer.text_model.encoder.layers.3.self_attn.q_proj.weight
878
+ text_encoders.clip_l.transformer.text_model.encoder.layers.3.self_attn.v_proj.bias
879
+ text_encoders.clip_l.transformer.text_model.encoder.layers.3.self_attn.v_proj.weight
880
+ text_encoders.clip_l.transformer.text_model.encoder.layers.4.layer_norm1.bias
881
+ text_encoders.clip_l.transformer.text_model.encoder.layers.4.layer_norm1.weight
882
+ text_encoders.clip_l.transformer.text_model.encoder.layers.4.layer_norm2.bias
883
+ text_encoders.clip_l.transformer.text_model.encoder.layers.4.layer_norm2.weight
884
+ text_encoders.clip_l.transformer.text_model.encoder.layers.4.mlp.fc1.bias
885
+ text_encoders.clip_l.transformer.text_model.encoder.layers.4.mlp.fc1.weight
886
+ text_encoders.clip_l.transformer.text_model.encoder.layers.4.mlp.fc2.bias
887
+ text_encoders.clip_l.transformer.text_model.encoder.layers.4.mlp.fc2.weight
888
+ text_encoders.clip_l.transformer.text_model.encoder.layers.4.self_attn.k_proj.bias
889
+ text_encoders.clip_l.transformer.text_model.encoder.layers.4.self_attn.k_proj.weight
890
+ text_encoders.clip_l.transformer.text_model.encoder.layers.4.self_attn.out_proj.bias
891
+ text_encoders.clip_l.transformer.text_model.encoder.layers.4.self_attn.out_proj.weight
892
+ text_encoders.clip_l.transformer.text_model.encoder.layers.4.self_attn.q_proj.bias
893
+ text_encoders.clip_l.transformer.text_model.encoder.layers.4.self_attn.q_proj.weight
894
+ text_encoders.clip_l.transformer.text_model.encoder.layers.4.self_attn.v_proj.bias
895
+ text_encoders.clip_l.transformer.text_model.encoder.layers.4.self_attn.v_proj.weight
896
+ text_encoders.clip_l.transformer.text_model.encoder.layers.5.layer_norm1.bias
897
+ text_encoders.clip_l.transformer.text_model.encoder.layers.5.layer_norm1.weight
898
+ text_encoders.clip_l.transformer.text_model.encoder.layers.5.layer_norm2.bias
899
+ text_encoders.clip_l.transformer.text_model.encoder.layers.5.layer_norm2.weight
900
+ text_encoders.clip_l.transformer.text_model.encoder.layers.5.mlp.fc1.bias
901
+ text_encoders.clip_l.transformer.text_model.encoder.layers.5.mlp.fc1.weight
902
+ text_encoders.clip_l.transformer.text_model.encoder.layers.5.mlp.fc2.bias
903
+ text_encoders.clip_l.transformer.text_model.encoder.layers.5.mlp.fc2.weight
904
+ text_encoders.clip_l.transformer.text_model.encoder.layers.5.self_attn.k_proj.bias
905
+ text_encoders.clip_l.transformer.text_model.encoder.layers.5.self_attn.k_proj.weight
906
+ text_encoders.clip_l.transformer.text_model.encoder.layers.5.self_attn.out_proj.bias
907
+ text_encoders.clip_l.transformer.text_model.encoder.layers.5.self_attn.out_proj.weight
908
+ text_encoders.clip_l.transformer.text_model.encoder.layers.5.self_attn.q_proj.bias
909
+ text_encoders.clip_l.transformer.text_model.encoder.layers.5.self_attn.q_proj.weight
910
+ text_encoders.clip_l.transformer.text_model.encoder.layers.5.self_attn.v_proj.bias
911
+ text_encoders.clip_l.transformer.text_model.encoder.layers.5.self_attn.v_proj.weight
912
+ text_encoders.clip_l.transformer.text_model.encoder.layers.6.layer_norm1.bias
913
+ text_encoders.clip_l.transformer.text_model.encoder.layers.6.layer_norm1.weight
914
+ text_encoders.clip_l.transformer.text_model.encoder.layers.6.layer_norm2.bias
915
+ text_encoders.clip_l.transformer.text_model.encoder.layers.6.layer_norm2.weight
916
+ text_encoders.clip_l.transformer.text_model.encoder.layers.6.mlp.fc1.bias
917
+ text_encoders.clip_l.transformer.text_model.encoder.layers.6.mlp.fc1.weight
918
+ text_encoders.clip_l.transformer.text_model.encoder.layers.6.mlp.fc2.bias
919
+ text_encoders.clip_l.transformer.text_model.encoder.layers.6.mlp.fc2.weight
920
+ text_encoders.clip_l.transformer.text_model.encoder.layers.6.self_attn.k_proj.bias
921
+ text_encoders.clip_l.transformer.text_model.encoder.layers.6.self_attn.k_proj.weight
922
+ text_encoders.clip_l.transformer.text_model.encoder.layers.6.self_attn.out_proj.bias
923
+ text_encoders.clip_l.transformer.text_model.encoder.layers.6.self_attn.out_proj.weight
924
+ text_encoders.clip_l.transformer.text_model.encoder.layers.6.self_attn.q_proj.bias
925
+ text_encoders.clip_l.transformer.text_model.encoder.layers.6.self_attn.q_proj.weight
926
+ text_encoders.clip_l.transformer.text_model.encoder.layers.6.self_attn.v_proj.bias
927
+ text_encoders.clip_l.transformer.text_model.encoder.layers.6.self_attn.v_proj.weight
928
+ text_encoders.clip_l.transformer.text_model.encoder.layers.7.layer_norm1.bias
929
+ text_encoders.clip_l.transformer.text_model.encoder.layers.7.layer_norm1.weight
930
+ text_encoders.clip_l.transformer.text_model.encoder.layers.7.layer_norm2.bias
931
+ text_encoders.clip_l.transformer.text_model.encoder.layers.7.layer_norm2.weight
932
+ text_encoders.clip_l.transformer.text_model.encoder.layers.7.mlp.fc1.bias
933
+ text_encoders.clip_l.transformer.text_model.encoder.layers.7.mlp.fc1.weight
934
+ text_encoders.clip_l.transformer.text_model.encoder.layers.7.mlp.fc2.bias
935
+ text_encoders.clip_l.transformer.text_model.encoder.layers.7.mlp.fc2.weight
936
+ text_encoders.clip_l.transformer.text_model.encoder.layers.7.self_attn.k_proj.bias
937
+ text_encoders.clip_l.transformer.text_model.encoder.layers.7.self_attn.k_proj.weight
938
+ text_encoders.clip_l.transformer.text_model.encoder.layers.7.self_attn.out_proj.bias
939
+ text_encoders.clip_l.transformer.text_model.encoder.layers.7.self_attn.out_proj.weight
940
+ text_encoders.clip_l.transformer.text_model.encoder.layers.7.self_attn.q_proj.bias
941
+ text_encoders.clip_l.transformer.text_model.encoder.layers.7.self_attn.q_proj.weight
942
+ text_encoders.clip_l.transformer.text_model.encoder.layers.7.self_attn.v_proj.bias
943
+ text_encoders.clip_l.transformer.text_model.encoder.layers.7.self_attn.v_proj.weight
944
+ text_encoders.clip_l.transformer.text_model.encoder.layers.8.layer_norm1.bias
945
+ text_encoders.clip_l.transformer.text_model.encoder.layers.8.layer_norm1.weight
946
+ text_encoders.clip_l.transformer.text_model.encoder.layers.8.layer_norm2.bias
947
+ text_encoders.clip_l.transformer.text_model.encoder.layers.8.layer_norm2.weight
948
+ text_encoders.clip_l.transformer.text_model.encoder.layers.8.mlp.fc1.bias
949
+ text_encoders.clip_l.transformer.text_model.encoder.layers.8.mlp.fc1.weight
950
+ text_encoders.clip_l.transformer.text_model.encoder.layers.8.mlp.fc2.bias
951
+ text_encoders.clip_l.transformer.text_model.encoder.layers.8.mlp.fc2.weight
952
+ text_encoders.clip_l.transformer.text_model.encoder.layers.8.self_attn.k_proj.bias
953
+ text_encoders.clip_l.transformer.text_model.encoder.layers.8.self_attn.k_proj.weight
954
+ text_encoders.clip_l.transformer.text_model.encoder.layers.8.self_attn.out_proj.bias
955
+ text_encoders.clip_l.transformer.text_model.encoder.layers.8.self_attn.out_proj.weight
956
+ text_encoders.clip_l.transformer.text_model.encoder.layers.8.self_attn.q_proj.bias
957
+ text_encoders.clip_l.transformer.text_model.encoder.layers.8.self_attn.q_proj.weight
958
+ text_encoders.clip_l.transformer.text_model.encoder.layers.8.self_attn.v_proj.bias
959
+ text_encoders.clip_l.transformer.text_model.encoder.layers.8.self_attn.v_proj.weight
960
+ text_encoders.clip_l.transformer.text_model.encoder.layers.9.layer_norm1.bias
961
+ text_encoders.clip_l.transformer.text_model.encoder.layers.9.layer_norm1.weight
962
+ text_encoders.clip_l.transformer.text_model.encoder.layers.9.layer_norm2.bias
963
+ text_encoders.clip_l.transformer.text_model.encoder.layers.9.layer_norm2.weight
964
+ text_encoders.clip_l.transformer.text_model.encoder.layers.9.mlp.fc1.bias
965
+ text_encoders.clip_l.transformer.text_model.encoder.layers.9.mlp.fc1.weight
966
+ text_encoders.clip_l.transformer.text_model.encoder.layers.9.mlp.fc2.bias
967
+ text_encoders.clip_l.transformer.text_model.encoder.layers.9.mlp.fc2.weight
968
+ text_encoders.clip_l.transformer.text_model.encoder.layers.9.self_attn.k_proj.bias
969
+ text_encoders.clip_l.transformer.text_model.encoder.layers.9.self_attn.k_proj.weight
970
+ text_encoders.clip_l.transformer.text_model.encoder.layers.9.self_attn.out_proj.bias
971
+ text_encoders.clip_l.transformer.text_model.encoder.layers.9.self_attn.out_proj.weight
972
+ text_encoders.clip_l.transformer.text_model.encoder.layers.9.self_attn.q_proj.bias
973
+ text_encoders.clip_l.transformer.text_model.encoder.layers.9.self_attn.q_proj.weight
974
+ text_encoders.clip_l.transformer.text_model.encoder.layers.9.self_attn.v_proj.bias
975
+ text_encoders.clip_l.transformer.text_model.encoder.layers.9.self_attn.v_proj.weight
976
+ text_encoders.clip_l.transformer.text_model.final_layer_norm.bias
977
+ text_encoders.clip_l.transformer.text_model.final_layer_norm.weight
978
+ text_encoders.clip_l.transformer.text_projection.weight
979
+ text_encoders.t5xxl.logit_scale
980
+ text_encoders.t5xxl.transformer.encoder.block.0.layer.0.SelfAttention.k.weight
981
+ text_encoders.t5xxl.transformer.encoder.block.0.layer.0.SelfAttention.o.weight
982
+ text_encoders.t5xxl.transformer.encoder.block.0.layer.0.SelfAttention.q.weight
983
+ text_encoders.t5xxl.transformer.encoder.block.0.layer.0.SelfAttention.relative_attention_bias.weight
984
+ text_encoders.t5xxl.transformer.encoder.block.0.layer.0.SelfAttention.v.weight
985
+ text_encoders.t5xxl.transformer.encoder.block.0.layer.0.layer_norm.weight
986
+ text_encoders.t5xxl.transformer.encoder.block.0.layer.1.DenseReluDense.wi_0.weight
987
+ text_encoders.t5xxl.transformer.encoder.block.0.layer.1.DenseReluDense.wi_1.weight
988
+ text_encoders.t5xxl.transformer.encoder.block.0.layer.1.DenseReluDense.wo.weight
989
+ text_encoders.t5xxl.transformer.encoder.block.0.layer.1.layer_norm.weight
990
+ text_encoders.t5xxl.transformer.encoder.block.1.layer.0.SelfAttention.k.weight
991
+ text_encoders.t5xxl.transformer.encoder.block.1.layer.0.SelfAttention.o.weight
992
+ text_encoders.t5xxl.transformer.encoder.block.1.layer.0.SelfAttention.q.weight
993
+ text_encoders.t5xxl.transformer.encoder.block.1.layer.0.SelfAttention.v.weight
994
+ text_encoders.t5xxl.transformer.encoder.block.1.layer.0.layer_norm.weight
995
+ text_encoders.t5xxl.transformer.encoder.block.1.layer.1.DenseReluDense.wi_0.weight
996
+ text_encoders.t5xxl.transformer.encoder.block.1.layer.1.DenseReluDense.wi_1.weight
997
+ text_encoders.t5xxl.transformer.encoder.block.1.layer.1.DenseReluDense.wo.weight
998
+ text_encoders.t5xxl.transformer.encoder.block.1.layer.1.layer_norm.weight
999
+ text_encoders.t5xxl.transformer.encoder.block.10.layer.0.SelfAttention.k.weight
1000
+ text_encoders.t5xxl.transformer.encoder.block.10.layer.0.SelfAttention.o.weight
1001
+ text_encoders.t5xxl.transformer.encoder.block.10.layer.0.SelfAttention.q.weight
1002
+ text_encoders.t5xxl.transformer.encoder.block.10.layer.0.SelfAttention.v.weight
1003
+ text_encoders.t5xxl.transformer.encoder.block.10.layer.0.layer_norm.weight
1004
+ text_encoders.t5xxl.transformer.encoder.block.10.layer.1.DenseReluDense.wi_0.weight
1005
+ text_encoders.t5xxl.transformer.encoder.block.10.layer.1.DenseReluDense.wi_1.weight
1006
+ text_encoders.t5xxl.transformer.encoder.block.10.layer.1.DenseReluDense.wo.weight
1007
+ text_encoders.t5xxl.transformer.encoder.block.10.layer.1.layer_norm.weight
1008
+ text_encoders.t5xxl.transformer.encoder.block.11.layer.0.SelfAttention.k.weight
1009
+ text_encoders.t5xxl.transformer.encoder.block.11.layer.0.SelfAttention.o.weight
1010
+ text_encoders.t5xxl.transformer.encoder.block.11.layer.0.SelfAttention.q.weight
1011
+ text_encoders.t5xxl.transformer.encoder.block.11.layer.0.SelfAttention.v.weight
1012
+ text_encoders.t5xxl.transformer.encoder.block.11.layer.0.layer_norm.weight
1013
+ text_encoders.t5xxl.transformer.encoder.block.11.layer.1.DenseReluDense.wi_0.weight
1014
+ text_encoders.t5xxl.transformer.encoder.block.11.layer.1.DenseReluDense.wi_1.weight
1015
+ text_encoders.t5xxl.transformer.encoder.block.11.layer.1.DenseReluDense.wo.weight
1016
+ text_encoders.t5xxl.transformer.encoder.block.11.layer.1.layer_norm.weight
1017
+ text_encoders.t5xxl.transformer.encoder.block.12.layer.0.SelfAttention.k.weight
1018
+ text_encoders.t5xxl.transformer.encoder.block.12.layer.0.SelfAttention.o.weight
1019
+ text_encoders.t5xxl.transformer.encoder.block.12.layer.0.SelfAttention.q.weight
1020
+ text_encoders.t5xxl.transformer.encoder.block.12.layer.0.SelfAttention.v.weight
1021
+ text_encoders.t5xxl.transformer.encoder.block.12.layer.0.layer_norm.weight
1022
+ text_encoders.t5xxl.transformer.encoder.block.12.layer.1.DenseReluDense.wi_0.weight
1023
+ text_encoders.t5xxl.transformer.encoder.block.12.layer.1.DenseReluDense.wi_1.weight
1024
+ text_encoders.t5xxl.transformer.encoder.block.12.layer.1.DenseReluDense.wo.weight
1025
+ text_encoders.t5xxl.transformer.encoder.block.12.layer.1.layer_norm.weight
1026
+ text_encoders.t5xxl.transformer.encoder.block.13.layer.0.SelfAttention.k.weight
1027
+ text_encoders.t5xxl.transformer.encoder.block.13.layer.0.SelfAttention.o.weight
1028
+ text_encoders.t5xxl.transformer.encoder.block.13.layer.0.SelfAttention.q.weight
1029
+ text_encoders.t5xxl.transformer.encoder.block.13.layer.0.SelfAttention.v.weight
1030
+ text_encoders.t5xxl.transformer.encoder.block.13.layer.0.layer_norm.weight
1031
+ text_encoders.t5xxl.transformer.encoder.block.13.layer.1.DenseReluDense.wi_0.weight
1032
+ text_encoders.t5xxl.transformer.encoder.block.13.layer.1.DenseReluDense.wi_1.weight
1033
+ text_encoders.t5xxl.transformer.encoder.block.13.layer.1.DenseReluDense.wo.weight
1034
+ text_encoders.t5xxl.transformer.encoder.block.13.layer.1.layer_norm.weight
1035
+ text_encoders.t5xxl.transformer.encoder.block.14.layer.0.SelfAttention.k.weight
1036
+ text_encoders.t5xxl.transformer.encoder.block.14.layer.0.SelfAttention.o.weight
1037
+ text_encoders.t5xxl.transformer.encoder.block.14.layer.0.SelfAttention.q.weight
1038
+ text_encoders.t5xxl.transformer.encoder.block.14.layer.0.SelfAttention.v.weight
1039
+ text_encoders.t5xxl.transformer.encoder.block.14.layer.0.layer_norm.weight
1040
+ text_encoders.t5xxl.transformer.encoder.block.14.layer.1.DenseReluDense.wi_0.weight
1041
+ text_encoders.t5xxl.transformer.encoder.block.14.layer.1.DenseReluDense.wi_1.weight
1042
+ text_encoders.t5xxl.transformer.encoder.block.14.layer.1.DenseReluDense.wo.weight
1043
+ text_encoders.t5xxl.transformer.encoder.block.14.layer.1.layer_norm.weight
1044
+ text_encoders.t5xxl.transformer.encoder.block.15.layer.0.SelfAttention.k.weight
1045
+ text_encoders.t5xxl.transformer.encoder.block.15.layer.0.SelfAttention.o.weight
1046
+ text_encoders.t5xxl.transformer.encoder.block.15.layer.0.SelfAttention.q.weight
1047
+ text_encoders.t5xxl.transformer.encoder.block.15.layer.0.SelfAttention.v.weight
1048
+ text_encoders.t5xxl.transformer.encoder.block.15.layer.0.layer_norm.weight
1049
+ text_encoders.t5xxl.transformer.encoder.block.15.layer.1.DenseReluDense.wi_0.weight
1050
+ text_encoders.t5xxl.transformer.encoder.block.15.layer.1.DenseReluDense.wi_1.weight
1051
+ text_encoders.t5xxl.transformer.encoder.block.15.layer.1.DenseReluDense.wo.weight
1052
+ text_encoders.t5xxl.transformer.encoder.block.15.layer.1.layer_norm.weight
1053
+ text_encoders.t5xxl.transformer.encoder.block.16.layer.0.SelfAttention.k.weight
1054
+ text_encoders.t5xxl.transformer.encoder.block.16.layer.0.SelfAttention.o.weight
1055
+ text_encoders.t5xxl.transformer.encoder.block.16.layer.0.SelfAttention.q.weight
1056
+ text_encoders.t5xxl.transformer.encoder.block.16.layer.0.SelfAttention.v.weight
1057
+ text_encoders.t5xxl.transformer.encoder.block.16.layer.0.layer_norm.weight
1058
+ text_encoders.t5xxl.transformer.encoder.block.16.layer.1.DenseReluDense.wi_0.weight
1059
+ text_encoders.t5xxl.transformer.encoder.block.16.layer.1.DenseReluDense.wi_1.weight
1060
+ text_encoders.t5xxl.transformer.encoder.block.16.layer.1.DenseReluDense.wo.weight
1061
+ text_encoders.t5xxl.transformer.encoder.block.16.layer.1.layer_norm.weight
1062
+ text_encoders.t5xxl.transformer.encoder.block.17.layer.0.SelfAttention.k.weight
1063
+ text_encoders.t5xxl.transformer.encoder.block.17.layer.0.SelfAttention.o.weight
1064
+ text_encoders.t5xxl.transformer.encoder.block.17.layer.0.SelfAttention.q.weight
1065
+ text_encoders.t5xxl.transformer.encoder.block.17.layer.0.SelfAttention.v.weight
1066
+ text_encoders.t5xxl.transformer.encoder.block.17.layer.0.layer_norm.weight
1067
+ text_encoders.t5xxl.transformer.encoder.block.17.layer.1.DenseReluDense.wi_0.weight
1068
+ text_encoders.t5xxl.transformer.encoder.block.17.layer.1.DenseReluDense.wi_1.weight
1069
+ text_encoders.t5xxl.transformer.encoder.block.17.layer.1.DenseReluDense.wo.weight
1070
+ text_encoders.t5xxl.transformer.encoder.block.17.layer.1.layer_norm.weight
1071
+ text_encoders.t5xxl.transformer.encoder.block.18.layer.0.SelfAttention.k.weight
1072
+ text_encoders.t5xxl.transformer.encoder.block.18.layer.0.SelfAttention.o.weight
1073
+ text_encoders.t5xxl.transformer.encoder.block.18.layer.0.SelfAttention.q.weight
1074
+ text_encoders.t5xxl.transformer.encoder.block.18.layer.0.SelfAttention.v.weight
1075
+ text_encoders.t5xxl.transformer.encoder.block.18.layer.0.layer_norm.weight
1076
+ text_encoders.t5xxl.transformer.encoder.block.18.layer.1.DenseReluDense.wi_0.weight
1077
+ text_encoders.t5xxl.transformer.encoder.block.18.layer.1.DenseReluDense.wi_1.weight
1078
+ text_encoders.t5xxl.transformer.encoder.block.18.layer.1.DenseReluDense.wo.weight
1079
+ text_encoders.t5xxl.transformer.encoder.block.18.layer.1.layer_norm.weight
1080
+ text_encoders.t5xxl.transformer.encoder.block.19.layer.0.SelfAttention.k.weight
1081
+ text_encoders.t5xxl.transformer.encoder.block.19.layer.0.SelfAttention.o.weight
1082
+ text_encoders.t5xxl.transformer.encoder.block.19.layer.0.SelfAttention.q.weight
1083
+ text_encoders.t5xxl.transformer.encoder.block.19.layer.0.SelfAttention.v.weight
1084
+ text_encoders.t5xxl.transformer.encoder.block.19.layer.0.layer_norm.weight
1085
+ text_encoders.t5xxl.transformer.encoder.block.19.layer.1.DenseReluDense.wi_0.weight
1086
+ text_encoders.t5xxl.transformer.encoder.block.19.layer.1.DenseReluDense.wi_1.weight
1087
+ text_encoders.t5xxl.transformer.encoder.block.19.layer.1.DenseReluDense.wo.weight
1088
+ text_encoders.t5xxl.transformer.encoder.block.19.layer.1.layer_norm.weight
1089
+ text_encoders.t5xxl.transformer.encoder.block.2.layer.0.SelfAttention.k.weight
1090
+ text_encoders.t5xxl.transformer.encoder.block.2.layer.0.SelfAttention.o.weight
1091
+ text_encoders.t5xxl.transformer.encoder.block.2.layer.0.SelfAttention.q.weight
1092
+ text_encoders.t5xxl.transformer.encoder.block.2.layer.0.SelfAttention.v.weight
1093
+ text_encoders.t5xxl.transformer.encoder.block.2.layer.0.layer_norm.weight
1094
+ text_encoders.t5xxl.transformer.encoder.block.2.layer.1.DenseReluDense.wi_0.weight
1095
+ text_encoders.t5xxl.transformer.encoder.block.2.layer.1.DenseReluDense.wi_1.weight
1096
+ text_encoders.t5xxl.transformer.encoder.block.2.layer.1.DenseReluDense.wo.weight
1097
+ text_encoders.t5xxl.transformer.encoder.block.2.layer.1.layer_norm.weight
1098
+ text_encoders.t5xxl.transformer.encoder.block.20.layer.0.SelfAttention.k.weight
1099
+ text_encoders.t5xxl.transformer.encoder.block.20.layer.0.SelfAttention.o.weight
1100
+ text_encoders.t5xxl.transformer.encoder.block.20.layer.0.SelfAttention.q.weight
1101
+ text_encoders.t5xxl.transformer.encoder.block.20.layer.0.SelfAttention.v.weight
1102
+ text_encoders.t5xxl.transformer.encoder.block.20.layer.0.layer_norm.weight
1103
+ text_encoders.t5xxl.transformer.encoder.block.20.layer.1.DenseReluDense.wi_0.weight
1104
+ text_encoders.t5xxl.transformer.encoder.block.20.layer.1.DenseReluDense.wi_1.weight
1105
+ text_encoders.t5xxl.transformer.encoder.block.20.layer.1.DenseReluDense.wo.weight
1106
+ text_encoders.t5xxl.transformer.encoder.block.20.layer.1.layer_norm.weight
1107
+ text_encoders.t5xxl.transformer.encoder.block.21.layer.0.SelfAttention.k.weight
1108
+ text_encoders.t5xxl.transformer.encoder.block.21.layer.0.SelfAttention.o.weight
1109
+ text_encoders.t5xxl.transformer.encoder.block.21.layer.0.SelfAttention.q.weight
1110
+ text_encoders.t5xxl.transformer.encoder.block.21.layer.0.SelfAttention.v.weight
1111
+ text_encoders.t5xxl.transformer.encoder.block.21.layer.0.layer_norm.weight
1112
+ text_encoders.t5xxl.transformer.encoder.block.21.layer.1.DenseReluDense.wi_0.weight
1113
+ text_encoders.t5xxl.transformer.encoder.block.21.layer.1.DenseReluDense.wi_1.weight
1114
+ text_encoders.t5xxl.transformer.encoder.block.21.layer.1.DenseReluDense.wo.weight
1115
+ text_encoders.t5xxl.transformer.encoder.block.21.layer.1.layer_norm.weight
1116
+ text_encoders.t5xxl.transformer.encoder.block.22.layer.0.SelfAttention.k.weight
1117
+ text_encoders.t5xxl.transformer.encoder.block.22.layer.0.SelfAttention.o.weight
1118
+ text_encoders.t5xxl.transformer.encoder.block.22.layer.0.SelfAttention.q.weight
1119
+ text_encoders.t5xxl.transformer.encoder.block.22.layer.0.SelfAttention.v.weight
1120
+ text_encoders.t5xxl.transformer.encoder.block.22.layer.0.layer_norm.weight
1121
+ text_encoders.t5xxl.transformer.encoder.block.22.layer.1.DenseReluDense.wi_0.weight
1122
+ text_encoders.t5xxl.transformer.encoder.block.22.layer.1.DenseReluDense.wi_1.weight
1123
+ text_encoders.t5xxl.transformer.encoder.block.22.layer.1.DenseReluDense.wo.weight
1124
+ text_encoders.t5xxl.transformer.encoder.block.22.layer.1.layer_norm.weight
1125
+ text_encoders.t5xxl.transformer.encoder.block.23.layer.0.SelfAttention.k.weight
1126
+ text_encoders.t5xxl.transformer.encoder.block.23.layer.0.SelfAttention.o.weight
1127
+ text_encoders.t5xxl.transformer.encoder.block.23.layer.0.SelfAttention.q.weight
1128
+ text_encoders.t5xxl.transformer.encoder.block.23.layer.0.SelfAttention.v.weight
1129
+ text_encoders.t5xxl.transformer.encoder.block.23.layer.0.layer_norm.weight
1130
+ text_encoders.t5xxl.transformer.encoder.block.23.layer.1.DenseReluDense.wi_0.weight
1131
+ text_encoders.t5xxl.transformer.encoder.block.23.layer.1.DenseReluDense.wi_1.weight
1132
+ text_encoders.t5xxl.transformer.encoder.block.23.layer.1.DenseReluDense.wo.weight
1133
+ text_encoders.t5xxl.transformer.encoder.block.23.layer.1.layer_norm.weight
1134
+ text_encoders.t5xxl.transformer.encoder.block.3.layer.0.SelfAttention.k.weight
1135
+ text_encoders.t5xxl.transformer.encoder.block.3.layer.0.SelfAttention.o.weight
1136
+ text_encoders.t5xxl.transformer.encoder.block.3.layer.0.SelfAttention.q.weight
1137
+ text_encoders.t5xxl.transformer.encoder.block.3.layer.0.SelfAttention.v.weight
1138
+ text_encoders.t5xxl.transformer.encoder.block.3.layer.0.layer_norm.weight
1139
+ text_encoders.t5xxl.transformer.encoder.block.3.layer.1.DenseReluDense.wi_0.weight
1140
+ text_encoders.t5xxl.transformer.encoder.block.3.layer.1.DenseReluDense.wi_1.weight
1141
+ text_encoders.t5xxl.transformer.encoder.block.3.layer.1.DenseReluDense.wo.weight
1142
+ text_encoders.t5xxl.transformer.encoder.block.3.layer.1.layer_norm.weight
1143
+ text_encoders.t5xxl.transformer.encoder.block.4.layer.0.SelfAttention.k.weight
1144
+ text_encoders.t5xxl.transformer.encoder.block.4.layer.0.SelfAttention.o.weight
1145
+ text_encoders.t5xxl.transformer.encoder.block.4.layer.0.SelfAttention.q.weight
1146
+ text_encoders.t5xxl.transformer.encoder.block.4.layer.0.SelfAttention.v.weight
1147
+ text_encoders.t5xxl.transformer.encoder.block.4.layer.0.layer_norm.weight
1148
+ text_encoders.t5xxl.transformer.encoder.block.4.layer.1.DenseReluDense.wi_0.weight
1149
+ text_encoders.t5xxl.transformer.encoder.block.4.layer.1.DenseReluDense.wi_1.weight
1150
+ text_encoders.t5xxl.transformer.encoder.block.4.layer.1.DenseReluDense.wo.weight
1151
+ text_encoders.t5xxl.transformer.encoder.block.4.layer.1.layer_norm.weight
1152
+ text_encoders.t5xxl.transformer.encoder.block.5.layer.0.SelfAttention.k.weight
1153
+ text_encoders.t5xxl.transformer.encoder.block.5.layer.0.SelfAttention.o.weight
1154
+ text_encoders.t5xxl.transformer.encoder.block.5.layer.0.SelfAttention.q.weight
1155
+ text_encoders.t5xxl.transformer.encoder.block.5.layer.0.SelfAttention.v.weight
1156
+ text_encoders.t5xxl.transformer.encoder.block.5.layer.0.layer_norm.weight
1157
+ text_encoders.t5xxl.transformer.encoder.block.5.layer.1.DenseReluDense.wi_0.weight
1158
+ text_encoders.t5xxl.transformer.encoder.block.5.layer.1.DenseReluDense.wi_1.weight
1159
+ text_encoders.t5xxl.transformer.encoder.block.5.layer.1.DenseReluDense.wo.weight
1160
+ text_encoders.t5xxl.transformer.encoder.block.5.layer.1.layer_norm.weight
1161
+ text_encoders.t5xxl.transformer.encoder.block.6.layer.0.SelfAttention.k.weight
1162
+ text_encoders.t5xxl.transformer.encoder.block.6.layer.0.SelfAttention.o.weight
1163
+ text_encoders.t5xxl.transformer.encoder.block.6.layer.0.SelfAttention.q.weight
1164
+ text_encoders.t5xxl.transformer.encoder.block.6.layer.0.SelfAttention.v.weight
1165
+ text_encoders.t5xxl.transformer.encoder.block.6.layer.0.layer_norm.weight
1166
+ text_encoders.t5xxl.transformer.encoder.block.6.layer.1.DenseReluDense.wi_0.weight
1167
+ text_encoders.t5xxl.transformer.encoder.block.6.layer.1.DenseReluDense.wi_1.weight
1168
+ text_encoders.t5xxl.transformer.encoder.block.6.layer.1.DenseReluDense.wo.weight
1169
+ text_encoders.t5xxl.transformer.encoder.block.6.layer.1.layer_norm.weight
1170
+ text_encoders.t5xxl.transformer.encoder.block.7.layer.0.SelfAttention.k.weight
1171
+ text_encoders.t5xxl.transformer.encoder.block.7.layer.0.SelfAttention.o.weight
1172
+ text_encoders.t5xxl.transformer.encoder.block.7.layer.0.SelfAttention.q.weight
1173
+ text_encoders.t5xxl.transformer.encoder.block.7.layer.0.SelfAttention.v.weight
1174
+ text_encoders.t5xxl.transformer.encoder.block.7.layer.0.layer_norm.weight
1175
+ text_encoders.t5xxl.transformer.encoder.block.7.layer.1.DenseReluDense.wi_0.weight
1176
+ text_encoders.t5xxl.transformer.encoder.block.7.layer.1.DenseReluDense.wi_1.weight
1177
+ text_encoders.t5xxl.transformer.encoder.block.7.layer.1.DenseReluDense.wo.weight
1178
+ text_encoders.t5xxl.transformer.encoder.block.7.layer.1.layer_norm.weight
1179
+ text_encoders.t5xxl.transformer.encoder.block.8.layer.0.SelfAttention.k.weight
1180
+ text_encoders.t5xxl.transformer.encoder.block.8.layer.0.SelfAttention.o.weight
1181
+ text_encoders.t5xxl.transformer.encoder.block.8.layer.0.SelfAttention.q.weight
1182
+ text_encoders.t5xxl.transformer.encoder.block.8.layer.0.SelfAttention.v.weight
1183
+ text_encoders.t5xxl.transformer.encoder.block.8.layer.0.layer_norm.weight
1184
+ text_encoders.t5xxl.transformer.encoder.block.8.layer.1.DenseReluDense.wi_0.weight
1185
+ text_encoders.t5xxl.transformer.encoder.block.8.layer.1.DenseReluDense.wi_1.weight
1186
+ text_encoders.t5xxl.transformer.encoder.block.8.layer.1.DenseReluDense.wo.weight
1187
+ text_encoders.t5xxl.transformer.encoder.block.8.layer.1.layer_norm.weight
1188
+ text_encoders.t5xxl.transformer.encoder.block.9.layer.0.SelfAttention.k.weight
1189
+ text_encoders.t5xxl.transformer.encoder.block.9.layer.0.SelfAttention.o.weight
1190
+ text_encoders.t5xxl.transformer.encoder.block.9.layer.0.SelfAttention.q.weight
1191
+ text_encoders.t5xxl.transformer.encoder.block.9.layer.0.SelfAttention.v.weight
1192
+ text_encoders.t5xxl.transformer.encoder.block.9.layer.0.layer_norm.weight
1193
+ text_encoders.t5xxl.transformer.encoder.block.9.layer.1.DenseReluDense.wi_0.weight
1194
+ text_encoders.t5xxl.transformer.encoder.block.9.layer.1.DenseReluDense.wi_1.weight
1195
+ text_encoders.t5xxl.transformer.encoder.block.9.layer.1.DenseReluDense.wo.weight
1196
+ text_encoders.t5xxl.transformer.encoder.block.9.layer.1.layer_norm.weight
1197
+ text_encoders.t5xxl.transformer.encoder.final_layer_norm.weight
1198
+ text_encoders.t5xxl.transformer.shared.weight
1199
+ vae.decoder.conv_in.bias
1200
+ vae.decoder.conv_in.weight
1201
+ vae.decoder.conv_out.bias
1202
+ vae.decoder.conv_out.weight
1203
+ vae.decoder.mid.attn_1.k.bias
1204
+ vae.decoder.mid.attn_1.k.weight
1205
+ vae.decoder.mid.attn_1.norm.bias
1206
+ vae.decoder.mid.attn_1.norm.weight
1207
+ vae.decoder.mid.attn_1.proj_out.bias
1208
+ vae.decoder.mid.attn_1.proj_out.weight
1209
+ vae.decoder.mid.attn_1.q.bias
1210
+ vae.decoder.mid.attn_1.q.weight
1211
+ vae.decoder.mid.attn_1.v.bias
1212
+ vae.decoder.mid.attn_1.v.weight
1213
+ vae.decoder.mid.block_1.conv1.bias
1214
+ vae.decoder.mid.block_1.conv1.weight
1215
+ vae.decoder.mid.block_1.conv2.bias
1216
+ vae.decoder.mid.block_1.conv2.weight
1217
+ vae.decoder.mid.block_1.norm1.bias
1218
+ vae.decoder.mid.block_1.norm1.weight
1219
+ vae.decoder.mid.block_1.norm2.bias
1220
+ vae.decoder.mid.block_1.norm2.weight
1221
+ vae.decoder.mid.block_2.conv1.bias
1222
+ vae.decoder.mid.block_2.conv1.weight
1223
+ vae.decoder.mid.block_2.conv2.bias
1224
+ vae.decoder.mid.block_2.conv2.weight
1225
+ vae.decoder.mid.block_2.norm1.bias
1226
+ vae.decoder.mid.block_2.norm1.weight
1227
+ vae.decoder.mid.block_2.norm2.bias
1228
+ vae.decoder.mid.block_2.norm2.weight
1229
+ vae.decoder.norm_out.bias
1230
+ vae.decoder.norm_out.weight
1231
+ vae.decoder.up.0.block.0.conv1.bias
1232
+ vae.decoder.up.0.block.0.conv1.weight
1233
+ vae.decoder.up.0.block.0.conv2.bias
1234
+ vae.decoder.up.0.block.0.conv2.weight
1235
+ vae.decoder.up.0.block.0.nin_shortcut.bias
1236
+ vae.decoder.up.0.block.0.nin_shortcut.weight
1237
+ vae.decoder.up.0.block.0.norm1.bias
1238
+ vae.decoder.up.0.block.0.norm1.weight
1239
+ vae.decoder.up.0.block.0.norm2.bias
1240
+ vae.decoder.up.0.block.0.norm2.weight
1241
+ vae.decoder.up.0.block.1.conv1.bias
1242
+ vae.decoder.up.0.block.1.conv1.weight
1243
+ vae.decoder.up.0.block.1.conv2.bias
1244
+ vae.decoder.up.0.block.1.conv2.weight
1245
+ vae.decoder.up.0.block.1.norm1.bias
1246
+ vae.decoder.up.0.block.1.norm1.weight
1247
+ vae.decoder.up.0.block.1.norm2.bias
1248
+ vae.decoder.up.0.block.1.norm2.weight
1249
+ vae.decoder.up.0.block.2.conv1.bias
1250
+ vae.decoder.up.0.block.2.conv1.weight
1251
+ vae.decoder.up.0.block.2.conv2.bias
1252
+ vae.decoder.up.0.block.2.conv2.weight
1253
+ vae.decoder.up.0.block.2.norm1.bias
1254
+ vae.decoder.up.0.block.2.norm1.weight
1255
+ vae.decoder.up.0.block.2.norm2.bias
1256
+ vae.decoder.up.0.block.2.norm2.weight
1257
+ vae.decoder.up.1.block.0.conv1.bias
1258
+ vae.decoder.up.1.block.0.conv1.weight
1259
+ vae.decoder.up.1.block.0.conv2.bias
1260
+ vae.decoder.up.1.block.0.conv2.weight
1261
+ vae.decoder.up.1.block.0.nin_shortcut.bias
1262
+ vae.decoder.up.1.block.0.nin_shortcut.weight
1263
+ vae.decoder.up.1.block.0.norm1.bias
1264
+ vae.decoder.up.1.block.0.norm1.weight
1265
+ vae.decoder.up.1.block.0.norm2.bias
1266
+ vae.decoder.up.1.block.0.norm2.weight
1267
+ vae.decoder.up.1.block.1.conv1.bias
1268
+ vae.decoder.up.1.block.1.conv1.weight
1269
+ vae.decoder.up.1.block.1.conv2.bias
1270
+ vae.decoder.up.1.block.1.conv2.weight
1271
+ vae.decoder.up.1.block.1.norm1.bias
1272
+ vae.decoder.up.1.block.1.norm1.weight
1273
+ vae.decoder.up.1.block.1.norm2.bias
1274
+ vae.decoder.up.1.block.1.norm2.weight
1275
+ vae.decoder.up.1.block.2.conv1.bias
1276
+ vae.decoder.up.1.block.2.conv1.weight
1277
+ vae.decoder.up.1.block.2.conv2.bias
1278
+ vae.decoder.up.1.block.2.conv2.weight
1279
+ vae.decoder.up.1.block.2.norm1.bias
1280
+ vae.decoder.up.1.block.2.norm1.weight
1281
+ vae.decoder.up.1.block.2.norm2.bias
1282
+ vae.decoder.up.1.block.2.norm2.weight
1283
+ vae.decoder.up.1.upsample.conv.bias
1284
+ vae.decoder.up.1.upsample.conv.weight
1285
+ vae.decoder.up.2.block.0.conv1.bias
1286
+ vae.decoder.up.2.block.0.conv1.weight
1287
+ vae.decoder.up.2.block.0.conv2.bias
1288
+ vae.decoder.up.2.block.0.conv2.weight
1289
+ vae.decoder.up.2.block.0.norm1.bias
1290
+ vae.decoder.up.2.block.0.norm1.weight
1291
+ vae.decoder.up.2.block.0.norm2.bias
1292
+ vae.decoder.up.2.block.0.norm2.weight
1293
+ vae.decoder.up.2.block.1.conv1.bias
1294
+ vae.decoder.up.2.block.1.conv1.weight
1295
+ vae.decoder.up.2.block.1.conv2.bias
1296
+ vae.decoder.up.2.block.1.conv2.weight
1297
+ vae.decoder.up.2.block.1.norm1.bias
1298
+ vae.decoder.up.2.block.1.norm1.weight
1299
+ vae.decoder.up.2.block.1.norm2.bias
1300
+ vae.decoder.up.2.block.1.norm2.weight
1301
+ vae.decoder.up.2.block.2.conv1.bias
1302
+ vae.decoder.up.2.block.2.conv1.weight
1303
+ vae.decoder.up.2.block.2.conv2.bias
1304
+ vae.decoder.up.2.block.2.conv2.weight
1305
+ vae.decoder.up.2.block.2.norm1.bias
1306
+ vae.decoder.up.2.block.2.norm1.weight
1307
+ vae.decoder.up.2.block.2.norm2.bias
1308
+ vae.decoder.up.2.block.2.norm2.weight
1309
+ vae.decoder.up.2.upsample.conv.bias
1310
+ vae.decoder.up.2.upsample.conv.weight
1311
+ vae.decoder.up.3.block.0.conv1.bias
1312
+ vae.decoder.up.3.block.0.conv1.weight
1313
+ vae.decoder.up.3.block.0.conv2.bias
1314
+ vae.decoder.up.3.block.0.conv2.weight
1315
+ vae.decoder.up.3.block.0.norm1.bias
1316
+ vae.decoder.up.3.block.0.norm1.weight
1317
+ vae.decoder.up.3.block.0.norm2.bias
1318
+ vae.decoder.up.3.block.0.norm2.weight
1319
+ vae.decoder.up.3.block.1.conv1.bias
1320
+ vae.decoder.up.3.block.1.conv1.weight
1321
+ vae.decoder.up.3.block.1.conv2.bias
1322
+ vae.decoder.up.3.block.1.conv2.weight
1323
+ vae.decoder.up.3.block.1.norm1.bias
1324
+ vae.decoder.up.3.block.1.norm1.weight
1325
+ vae.decoder.up.3.block.1.norm2.bias
1326
+ vae.decoder.up.3.block.1.norm2.weight
1327
+ vae.decoder.up.3.block.2.conv1.bias
1328
+ vae.decoder.up.3.block.2.conv1.weight
1329
+ vae.decoder.up.3.block.2.conv2.bias
1330
+ vae.decoder.up.3.block.2.conv2.weight
1331
+ vae.decoder.up.3.block.2.norm1.bias
1332
+ vae.decoder.up.3.block.2.norm1.weight
1333
+ vae.decoder.up.3.block.2.norm2.bias
1334
+ vae.decoder.up.3.block.2.norm2.weight
1335
+ vae.decoder.up.3.upsample.conv.bias
1336
+ vae.decoder.up.3.upsample.conv.weight
1337
+ vae.encoder.conv_in.bias
1338
+ vae.encoder.conv_in.weight
1339
+ vae.encoder.conv_out.bias
1340
+ vae.encoder.conv_out.weight
1341
+ vae.encoder.down.0.block.0.conv1.bias
1342
+ vae.encoder.down.0.block.0.conv1.weight
1343
+ vae.encoder.down.0.block.0.conv2.bias
1344
+ vae.encoder.down.0.block.0.conv2.weight
1345
+ vae.encoder.down.0.block.0.norm1.bias
1346
+ vae.encoder.down.0.block.0.norm1.weight
1347
+ vae.encoder.down.0.block.0.norm2.bias
1348
+ vae.encoder.down.0.block.0.norm2.weight
1349
+ vae.encoder.down.0.block.1.conv1.bias
1350
+ vae.encoder.down.0.block.1.conv1.weight
1351
+ vae.encoder.down.0.block.1.conv2.bias
1352
+ vae.encoder.down.0.block.1.conv2.weight
1353
+ vae.encoder.down.0.block.1.norm1.bias
1354
+ vae.encoder.down.0.block.1.norm1.weight
1355
+ vae.encoder.down.0.block.1.norm2.bias
1356
+ vae.encoder.down.0.block.1.norm2.weight
1357
+ vae.encoder.down.0.downsample.conv.bias
1358
+ vae.encoder.down.0.downsample.conv.weight
1359
+ vae.encoder.down.1.block.0.conv1.bias
1360
+ vae.encoder.down.1.block.0.conv1.weight
1361
+ vae.encoder.down.1.block.0.conv2.bias
1362
+ vae.encoder.down.1.block.0.conv2.weight
1363
+ vae.encoder.down.1.block.0.nin_shortcut.bias
1364
+ vae.encoder.down.1.block.0.nin_shortcut.weight
1365
+ vae.encoder.down.1.block.0.norm1.bias
1366
+ vae.encoder.down.1.block.0.norm1.weight
1367
+ vae.encoder.down.1.block.0.norm2.bias
1368
+ vae.encoder.down.1.block.0.norm2.weight
1369
+ vae.encoder.down.1.block.1.conv1.bias
1370
+ vae.encoder.down.1.block.1.conv1.weight
1371
+ vae.encoder.down.1.block.1.conv2.bias
1372
+ vae.encoder.down.1.block.1.conv2.weight
1373
+ vae.encoder.down.1.block.1.norm1.bias
1374
+ vae.encoder.down.1.block.1.norm1.weight
1375
+ vae.encoder.down.1.block.1.norm2.bias
1376
+ vae.encoder.down.1.block.1.norm2.weight
1377
+ vae.encoder.down.1.downsample.conv.bias
1378
+ vae.encoder.down.1.downsample.conv.weight
1379
+ vae.encoder.down.2.block.0.conv1.bias
1380
+ vae.encoder.down.2.block.0.conv1.weight
1381
+ vae.encoder.down.2.block.0.conv2.bias
1382
+ vae.encoder.down.2.block.0.conv2.weight
1383
+ vae.encoder.down.2.block.0.nin_shortcut.bias
1384
+ vae.encoder.down.2.block.0.nin_shortcut.weight
1385
+ vae.encoder.down.2.block.0.norm1.bias
1386
+ vae.encoder.down.2.block.0.norm1.weight
1387
+ vae.encoder.down.2.block.0.norm2.bias
1388
+ vae.encoder.down.2.block.0.norm2.weight
1389
+ vae.encoder.down.2.block.1.conv1.bias
1390
+ vae.encoder.down.2.block.1.conv1.weight
1391
+ vae.encoder.down.2.block.1.conv2.bias
1392
+ vae.encoder.down.2.block.1.conv2.weight
1393
+ vae.encoder.down.2.block.1.norm1.bias
1394
+ vae.encoder.down.2.block.1.norm1.weight
1395
+ vae.encoder.down.2.block.1.norm2.bias
1396
+ vae.encoder.down.2.block.1.norm2.weight
1397
+ vae.encoder.down.2.downsample.conv.bias
1398
+ vae.encoder.down.2.downsample.conv.weight
1399
+ vae.encoder.down.3.block.0.conv1.bias
1400
+ vae.encoder.down.3.block.0.conv1.weight
1401
+ vae.encoder.down.3.block.0.conv2.bias
1402
+ vae.encoder.down.3.block.0.conv2.weight
1403
+ vae.encoder.down.3.block.0.norm1.bias
1404
+ vae.encoder.down.3.block.0.norm1.weight
1405
+ vae.encoder.down.3.block.0.norm2.bias
1406
+ vae.encoder.down.3.block.0.norm2.weight
1407
+ vae.encoder.down.3.block.1.conv1.bias
1408
+ vae.encoder.down.3.block.1.conv1.weight
1409
+ vae.encoder.down.3.block.1.conv2.bias
1410
+ vae.encoder.down.3.block.1.conv2.weight
1411
+ vae.encoder.down.3.block.1.norm1.bias
1412
+ vae.encoder.down.3.block.1.norm1.weight
1413
+ vae.encoder.down.3.block.1.norm2.bias
1414
+ vae.encoder.down.3.block.1.norm2.weight
1415
+ vae.encoder.mid.attn_1.k.bias
1416
+ vae.encoder.mid.attn_1.k.weight
1417
+ vae.encoder.mid.attn_1.norm.bias
1418
+ vae.encoder.mid.attn_1.norm.weight
1419
+ vae.encoder.mid.attn_1.proj_out.bias
1420
+ vae.encoder.mid.attn_1.proj_out.weight
1421
+ vae.encoder.mid.attn_1.q.bias
1422
+ vae.encoder.mid.attn_1.q.weight
1423
+ vae.encoder.mid.attn_1.v.bias
1424
+ vae.encoder.mid.attn_1.v.weight
1425
+ vae.encoder.mid.block_1.conv1.bias
1426
+ vae.encoder.mid.block_1.conv1.weight
1427
+ vae.encoder.mid.block_1.conv2.bias
1428
+ vae.encoder.mid.block_1.conv2.weight
1429
+ vae.encoder.mid.block_1.norm1.bias
1430
+ vae.encoder.mid.block_1.norm1.weight
1431
+ vae.encoder.mid.block_1.norm2.bias
1432
+ vae.encoder.mid.block_1.norm2.weight
1433
+ vae.encoder.mid.block_2.conv1.bias
1434
+ vae.encoder.mid.block_2.conv1.weight
1435
+ vae.encoder.mid.block_2.conv2.bias
1436
+ vae.encoder.mid.block_2.conv2.weight
1437
+ vae.encoder.mid.block_2.norm1.bias
1438
+ vae.encoder.mid.block_2.norm1.weight
1439
+ vae.encoder.mid.block_2.norm2.bias
1440
+ vae.encoder.mid.block_2.norm2.weight
1441
+ vae.encoder.norm_out.bias
1442
+ vae.encoder.norm_out.weight
pre-requirements.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ pip>=23.0.0
requirements.txt ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ huggingface_hub
2
+ safetensors
3
+ git+https://github.com/huggingface/transformers
4
+ git+https://github.com/huggingface/accelerate
5
+ git+https://github.com/huggingface/diffusers
6
+ git+https://github.com/huggingface/peft
7
+ optimum-quanto
8
+ sentencepiece
9
+ torch
10
+ torchaudio
11
+ torchvision
12
+ pytorch_lightning
13
+ aria2
14
+ gdown
15
+ gguf>=0.9.1
16
+ numpy
17
+ psutil