RedbeardNZ's picture
Duplicate from GD-ML/FLUX-Text
58d8d48 verified
dtype: bfloat16
flux_path: black-forest-labs/FLUX.1-Fill-dev
model:
add_cond_attn: false
attn_mask: false
latent_lora: false
mask_para:
base_ratio: 0.6
union_cond_attn: true
use_byt5_mask: false
model_type: flux_fill
train:
accumulate_grad_batches: 4
bucket_config:
1024px:
1:
- 1.0
- 2
512px:
1:
- 1.0
- 8
768px:
1:
- 1.0
- 4
condition_type: word_fill
dataloader_workers: 5
dataset:
drop_image_prob: 0.1
drop_text_prob: 0.1
glyph_scale: 1
random_select: true
type: word
gradient_checkpointing: true
lora_config:
init_lora_weights: gaussian
lora_alpha: 256
r: 256
target_modules: (.*x_embedder|.*(?<!single_)transformer_blocks\.[0-9]+\.norm1\.linear|.*(?<!single_)transformer_blocks\.[0-9]+\.attn\.to_k|.*(?<!single_)transformer_blocks\.[0-9]+\.attn\.to_q|.*(?<!single_)transformer_blocks\.[0-9]+\.attn\.to_v|.*(?<!single_)transformer_blocks\.[0-9]+\.attn\.to_out\.0|.*(?<!single_)transformer_blocks\.[0-9]+\.ff\.net\.2|.*single_transformer_blocks\.[0-9]+\.norm\.linear|.*single_transformer_blocks\.[0-9]+\.proj_mlp|.*single_transformer_blocks\.[0-9]+\.proj_out|.*single_transformer_blocks\.[0-9]+\.attn.to_k|.*single_transformer_blocks\.[0-9]+\.attn.to_q|.*single_transformer_blocks\.[0-9]+\.attn.to_v|.*single_transformer_blocks\.[0-9]+\.attn.to_out)
max_steps: -1
odm_loss:
modelpath: epoch_100.pt
w_loss_1: 20
w_loss_2: 20
w_loss_3: 20
w_loss_4: 20
w_loss_f: 1
optimizer:
params:
lr: 1
safeguard_warmup: true
use_bias_correction: true
weight_decay: 0.01
type: Prodigy
reuse_lora_path: reuse.safetensors
sample_interval: 2000
save_interval: 1000
save_path: output_dir