File size: 2,412 Bytes
3ce2981 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 |
# GENERATE TIME: Sat Jun 15 14:15:49 2024
# CMD:
# train_edlora.py -opt single-concept/train_configs/7657_hiro_ortho.yml
name: 7657_hiro_ortho
manual_seed: 7657
mixed_precision: fp16
gradient_accumulation_steps: 1
datasets:
train:
name: LoraDataset
concept_list: single-concept/data_configs/hiro.json
use_caption: true
use_mask: true
instance_transform:
- type: HumanResizeCropFinalV3
size: 512
crop_p: 0.5
- type: ToTensor
- type: Normalize
mean:
- 0.5
std:
- 0.5
- type: ShuffleCaption
keep_token_num: 1
- type: EnhanceText
enhance_type: human
replace_mapping:
<TOK>: <hiro1> <hiro2>
batch_size_per_gpu: 2
dataset_enlarge_ratio: 500
val_vis:
name: PromptDataset
prompts: single-concept/validation_prompts/characters/test_boy_disney.txt
num_samples_per_prompt: 8
latent_size:
- 4
- 64
- 64
replace_mapping:
<TOK>: <hiro1> <hiro2>
batch_size_per_gpu: 4
models:
pretrained_path: nitrosocke/mo-di-diffusion
enable_edlora: true
finetune_cfg:
text_embedding:
enable_tuning: true
lr: 0.001
text_encoder:
enable_tuning: true
lora_cfg:
rank: 5
alpha: 1.0
where: CLIPAttention
lr: 1.0e-05
unet:
enable_tuning: true
lora_cfg:
rank: 5
alpha: 1.0
where: Attention
lr: 0.0001
new_concept_token: <hiro1>+<hiro2>
initializer_token: <rand-0.013>+boy
noise_offset: 0.01
attn_reg_weight: 0.01
reg_full_identity: false
use_mask_loss: true
gradient_checkpoint: false
enable_xformers: true
path:
pretrain_network: null
train:
optim_g:
type: AdamW
lr: 0.0
weight_decay: 0.01
betas:
- 0.9
- 0.999
unet_kv_drop_rate: 0
scheduler: linear
emb_norm_threshold: 0.55
val:
val_during_save: false
compose_visualize: false
alpha_list:
- 0.7
- 1.0
sample:
num_inference_steps: 50
guidance_scale: 7.5
logger:
print_freq: 10
save_checkpoint_freq: 10000.0
|