p1atdev commited on
Commit
4318878
·
verified ·
1 Parent(s): c80dccb

Upload rope_migration_ppl.yml

Browse files
Files changed (1) hide show
  1. rope-4/rope_migration_ppl.yml +77 -0
rope-4/rope_migration_ppl.yml ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ checkpoint_path: "./models/aura_flow_0.3.bnb_nf4.safetensors"
3
+ pretrained_model_name_or_path: fal/AuraFlow-v0.3
4
+
5
+ dtype: bfloat16
6
+
7
+ denoiser:
8
+ use_flash_attn: true
9
+
10
+ use_rope: True
11
+ rope_theta: 10000
12
+ rope_dim_sizes: [32, 112, 112]
13
+
14
+ peft:
15
+ type: lora
16
+ rank: 4
17
+ alpha: 1.0
18
+ dropout: 0.0
19
+
20
+ dtype: bfloat16
21
+
22
+ include_keys: [".attn."]
23
+ exclude_keys: ["text_encoder", "vae", "t_embedder", "final_linear"]
24
+
25
+ dataset:
26
+ folder: "data/pexels-1k-random"
27
+ num_repeats: 2
28
+ batch_size: 2
29
+
30
+ bucket_base_size: 1024
31
+ step: 128
32
+ min_size: 384
33
+
34
+ caption_processors: []
35
+
36
+ optimizer:
37
+ name: "schedulefree.RAdamScheduleFree"
38
+ # name: "bitsandbytes.optim.AdamW8bit"
39
+ args:
40
+ lr: 0.005
41
+
42
+ scheduler:
43
+ # name: "torch.optim.lr_scheduler.ConstantLR"
44
+ # args: {}
45
+
46
+ tracker:
47
+ project_name: "auraflow-rope-1"
48
+ loggers:
49
+ - wandb
50
+
51
+ saving:
52
+ strategy:
53
+ per_epochs: 1
54
+ per_steps: null
55
+ save_last: true
56
+
57
+ callbacks:
58
+ - type: "hf_hub" # or "hf_hub" to push to hub
59
+ name: "rope-4"
60
+ save_dir: "./output/rope-4"
61
+
62
+ hub_id: "p1atdev/afv03-lora"
63
+ dir_in_repo: "rope-4"
64
+
65
+ seed: 42
66
+ num_train_epochs: 5
67
+
68
+ trainer:
69
+ # debug_mode: "1step"
70
+
71
+ gradient_checkpointing: true
72
+
73
+ torch_compile: true
74
+ torch_compile_args:
75
+ mode: max-autotune
76
+ fullgraph: true
77
+ fp32_matmul_precision: "medium"