Upload config.yaml with huggingface_hub
Browse files- config.yaml +10 -10
config.yaml
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
logging:
|
2 |
project: titok_video
|
3 |
-
run_name: BaseAll-CB16k-
|
4 |
|
5 |
logging_interval: 50
|
6 |
|
@@ -9,7 +9,7 @@ logging:
|
|
9 |
keep_prior_checkpoints: -1 # -1 to keep all
|
10 |
|
11 |
resume_from_checkpoint:
|
12 |
-
init_from_checkpoint:
|
13 |
|
14 |
model:
|
15 |
titok:
|
@@ -17,7 +17,7 @@ model:
|
|
17 |
spatial_patch_size: 4
|
18 |
|
19 |
fsq_levels: [8, 8, 8, 6, 5] # [7, 5, 5, 5, 5]
|
20 |
-
num_latent_tokens:
|
21 |
|
22 |
encoder_size: base
|
23 |
decoder_size: base
|
@@ -49,10 +49,10 @@ model:
|
|
49 |
final_gamma: 0.1
|
50 |
|
51 |
dataset:
|
52 |
-
train_dataset: "/workspace/
|
53 |
-
eval_dataset: "/workspace/
|
54 |
-
resolution:
|
55 |
-
num_frames:
|
56 |
frames_per_second: 8
|
57 |
workers: 8
|
58 |
|
@@ -62,7 +62,7 @@ optimizer:
|
|
62 |
beta1: 0.9
|
63 |
beta2: 0.99
|
64 |
weight_decay: 1e-4
|
65 |
-
warmup_steps:
|
66 |
end_lr: 1e-5
|
67 |
|
68 |
disc: # not used
|
@@ -76,9 +76,9 @@ optimizer:
|
|
76 |
training:
|
77 |
torch_compile: True
|
78 |
seed: 42
|
79 |
-
max_grad_norm: 1.0
|
80 |
|
81 |
-
batch_size:
|
82 |
# strategy: # ddp
|
83 |
|
84 |
enable_tf32: True
|
|
|
1 |
logging:
|
2 |
project: titok_video
|
3 |
+
run_name: BaseAll-CB16k-TL64-128x17-BS256
|
4 |
|
5 |
logging_interval: 50
|
6 |
|
|
|
9 |
keep_prior_checkpoints: -1 # -1 to keep all
|
10 |
|
11 |
resume_from_checkpoint:
|
12 |
+
init_from_checkpoint:
|
13 |
|
14 |
model:
|
15 |
titok:
|
|
|
17 |
spatial_patch_size: 4
|
18 |
|
19 |
fsq_levels: [8, 8, 8, 6, 5] # [7, 5, 5, 5, 5]
|
20 |
+
num_latent_tokens: 64
|
21 |
|
22 |
encoder_size: base
|
23 |
decoder_size: base
|
|
|
49 |
final_gamma: 0.1
|
50 |
|
51 |
dataset:
|
52 |
+
train_dataset: "/workspace/out_enc_128_17/**/*.pt"
|
53 |
+
eval_dataset: "/workspace/out_enc_128_17_eval/*.pt"
|
54 |
+
resolution: 128
|
55 |
+
num_frames: 17
|
56 |
frames_per_second: 8
|
57 |
workers: 8
|
58 |
|
|
|
62 |
beta1: 0.9
|
63 |
beta2: 0.99
|
64 |
weight_decay: 1e-4
|
65 |
+
warmup_steps: 10000
|
66 |
end_lr: 1e-5
|
67 |
|
68 |
disc: # not used
|
|
|
76 |
training:
|
77 |
torch_compile: True
|
78 |
seed: 42
|
79 |
+
max_grad_norm: # 1.0
|
80 |
|
81 |
+
batch_size: 256
|
82 |
# strategy: # ddp
|
83 |
|
84 |
enable_tf32: True
|