dtype: bfloat16 | |
flux_path: black-forest-labs/FLUX.1-Fill-dev | |
model: | |
add_cond_attn: false | |
attn_mask: false | |
latent_lora: false | |
mask_para: | |
base_ratio: 0.6 | |
union_cond_attn: true | |
use_byt5_mask: false | |
model_type: flux_fill | |
train: | |
accumulate_grad_batches: 4 | |
bucket_config: | |
1024px: | |
1: | |
- 1.0 | |
- 2 | |
512px: | |
1: | |
- 1.0 | |
- 8 | |
768px: | |
1: | |
- 1.0 | |
- 4 | |
condition_type: word_fill | |
dataloader_workers: 5 | |
dataset: | |
drop_image_prob: 0.1 | |
drop_text_prob: 0.1 | |
glyph_scale: 1 | |
random_select: true | |
type: word | |
gradient_checkpointing: true | |
lora_config: | |
init_lora_weights: gaussian | |
lora_alpha: 256 | |
r: 256 | |
target_modules: (.*x_embedder|.*(?<!single_)transformer_blocks\.[0-9]+\.norm1\.linear|.*(?<!single_)transformer_blocks\.[0-9]+\.attn\.to_k|.*(?<!single_)transformer_blocks\.[0-9]+\.attn\.to_q|.*(?<!single_)transformer_blocks\.[0-9]+\.attn\.to_v|.*(?<!single_)transformer_blocks\.[0-9]+\.attn\.to_out\.0|.*(?<!single_)transformer_blocks\.[0-9]+\.ff\.net\.2|.*single_transformer_blocks\.[0-9]+\.norm\.linear|.*single_transformer_blocks\.[0-9]+\.proj_mlp|.*single_transformer_blocks\.[0-9]+\.proj_out|.*single_transformer_blocks\.[0-9]+\.attn.to_k|.*single_transformer_blocks\.[0-9]+\.attn.to_q|.*single_transformer_blocks\.[0-9]+\.attn.to_v|.*single_transformer_blocks\.[0-9]+\.attn.to_out) | |
max_steps: -1 | |
odm_loss: | |
modelpath: epoch_100.pt | |
w_loss_1: 20 | |
w_loss_2: 20 | |
w_loss_3: 20 | |
w_loss_4: 20 | |
w_loss_f: 1 | |
optimizer: | |
params: | |
lr: 1 | |
safeguard_warmup: true | |
use_bias_correction: true | |
weight_decay: 0.01 | |
type: Prodigy | |
reuse_lora_path: reuse.safetensors | |
sample_interval: 2000 | |
save_interval: 1000 | |
save_path: output_dir | |