p1atdev commited on
Commit
1e0c958
·
verified ·
1 Parent(s): 0aa6656

Upload rope_migration.yml

Browse files
Files changed (1) hide show
  1. rope-12/rope_migration.yml +92 -0
rope-12/rope_migration.yml ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ checkpoint_path: "./models/aura_flow_0.3.bnb_nf4.safetensors"
3
+ pretrained_model_name_or_path: fal/AuraFlow-v0.3
4
+
5
+ dtype: bfloat16
6
+
7
+ denoiser:
8
+ use_flash_attn: true
9
+
10
+ use_rope: True
11
+ rope_theta: 10000
12
+ rope_dim_sizes: [32, 112, 112]
13
+
14
+ noise_prediction_loss: true
15
+ migration_loss: true
16
+ # prior_preservation_loss: true
17
+
18
+ peft:
19
+ type: lora
20
+ rank: 4
21
+ alpha: 1.0
22
+ dropout: 0.0
23
+
24
+ dtype: bfloat16
25
+
26
+ # include_keys: [".attn.", ".mlpC.", ".mlpX.", ".mlp.", ".modC.", ".modCX.", ".modX."]
27
+ include_keys:
28
+ - regex: .*\.attn\.w2[qkvo] # only latent path
29
+ - regex: .*\.mlp[X]?\. # only latent or (cond+latent) path
30
+ - regex: .*single_layers\.\d+\.attn\.w1[qkvo] # (cond+latent) path
31
+
32
+ exclude_keys:
33
+ - "text_encoder"
34
+ - "vae"
35
+ - "t_embedder"
36
+ - "final_linear"
37
+ - regex: \.mod[CX]{1,2} # exclude modulation layers (modC, modCX, modX)
38
+
39
+ dataset:
40
+ folder: "data/pexels-1k-random"
41
+ num_repeats: 2
42
+ batch_size: 2
43
+
44
+ bucket_base_size: 1024
45
+ step: 128
46
+ min_size: 384
47
+ do_upscale: false
48
+
49
+ caption_processors: []
50
+
51
+ optimizer:
52
+ name: "schedulefree.RAdamScheduleFree"
53
+ # name: "bitsandbytes.optim.AdamW8bit"
54
+ args:
55
+ lr: 0.005
56
+
57
+ scheduler:
58
+ # name: "torch.optim.lr_scheduler.ConstantLR"
59
+ # args: {}
60
+
61
+ tracker:
62
+ project_name: "auraflow-rope-1"
63
+ loggers:
64
+ - wandb
65
+
66
+ saving:
67
+ strategy:
68
+ per_epochs: 1
69
+ per_steps: null
70
+ save_last: true
71
+
72
+ callbacks:
73
+ - type: "hf_hub" # or "hf_hub" to push to hub
74
+ name: "rope-12"
75
+ save_dir: "./output/rope-12"
76
+
77
+ hub_id: "p1atdev/afv03-lora"
78
+ dir_in_repo: "rope-12"
79
+
80
+ seed: 42
81
+ num_train_epochs: 20
82
+
83
+ trainer:
84
+ # debug_mode: "1step"
85
+
86
+ gradient_checkpointing: true
87
+
88
+ torch_compile: true
89
+ torch_compile_args:
90
+ mode: max-autotune
91
+ fullgraph: true
92
+ fp32_matmul_precision: "medium"