CoderCowMoo commited on
Commit
950d2dd
·
1 Parent(s): d015833

Upload model, config, vocab and training log

Browse files
Files changed (4) hide show
  1. best_model.pth +3 -0
  2. config.json +159 -0
  3. trainer_0_log.txt +211 -0
  4. vocab.json +0 -0
best_model.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd02035aa2815b805f1db26f4cce42a2cfbe30ca0c4825f432961bc74aa4b420
3
+ size 5607926869
config.json ADDED
@@ -0,0 +1,159 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "output_path": "output",
3
+ "logger_uri": null,
4
+ "run_name": "run",
5
+ "project_name": null,
6
+ "run_description": "\ud83d\udc38Coqui trainer run.",
7
+ "print_step": 25,
8
+ "plot_step": 100,
9
+ "model_param_stats": false,
10
+ "wandb_entity": null,
11
+ "dashboard_logger": "tensorboard",
12
+ "save_on_interrupt": true,
13
+ "log_model_step": null,
14
+ "save_step": 10000,
15
+ "save_n_checkpoints": 5,
16
+ "save_checkpoints": true,
17
+ "save_all_best": false,
18
+ "save_best_after": 10000,
19
+ "target_loss": null,
20
+ "print_eval": false,
21
+ "test_delay_epochs": 0,
22
+ "run_eval": true,
23
+ "run_eval_steps": null,
24
+ "distributed_backend": "nccl",
25
+ "distributed_url": "tcp://localhost:54321",
26
+ "mixed_precision": false,
27
+ "precision": "fp16",
28
+ "epochs": 1000,
29
+ "batch_size": 32,
30
+ "eval_batch_size": 16,
31
+ "grad_clip": 0.0,
32
+ "scheduler_after_epoch": true,
33
+ "lr": 0.001,
34
+ "optimizer": "radam",
35
+ "optimizer_params": null,
36
+ "lr_scheduler": null,
37
+ "lr_scheduler_params": {},
38
+ "use_grad_scaler": false,
39
+ "allow_tf32": false,
40
+ "cudnn_enable": true,
41
+ "cudnn_deterministic": false,
42
+ "cudnn_benchmark": false,
43
+ "training_seed": 54321,
44
+ "model": "xtts",
45
+ "num_loader_workers": 0,
46
+ "num_eval_loader_workers": 0,
47
+ "use_noise_augment": false,
48
+ "audio": {
49
+ "sample_rate": 22050,
50
+ "output_sample_rate": 24000
51
+ },
52
+ "use_phonemes": false,
53
+ "phonemizer": null,
54
+ "phoneme_language": null,
55
+ "compute_input_seq_cache": false,
56
+ "text_cleaner": null,
57
+ "enable_eos_bos_chars": false,
58
+ "test_sentences_file": "",
59
+ "phoneme_cache_path": null,
60
+ "characters": null,
61
+ "add_blank": false,
62
+ "batch_group_size": 0,
63
+ "loss_masking": null,
64
+ "min_audio_len": 1,
65
+ "max_audio_len": Infinity,
66
+ "min_text_len": 1,
67
+ "max_text_len": Infinity,
68
+ "compute_f0": false,
69
+ "compute_energy": false,
70
+ "compute_linear_spec": false,
71
+ "precompute_num_workers": 0,
72
+ "start_by_longest": false,
73
+ "shuffle": false,
74
+ "drop_last": false,
75
+ "datasets": [
76
+ {
77
+ "formatter": "",
78
+ "dataset_name": "",
79
+ "path": "",
80
+ "meta_file_train": "",
81
+ "ignored_speakers": null,
82
+ "language": "",
83
+ "phonemizer": "",
84
+ "meta_file_val": "",
85
+ "meta_file_attn_mask": ""
86
+ }
87
+ ],
88
+ "test_sentences": [],
89
+ "eval_split_max_size": null,
90
+ "eval_split_size": 0.01,
91
+ "use_speaker_weighted_sampler": false,
92
+ "speaker_weighted_sampler_alpha": 1.0,
93
+ "use_language_weighted_sampler": false,
94
+ "language_weighted_sampler_alpha": 1.0,
95
+ "use_length_weighted_sampler": false,
96
+ "length_weighted_sampler_alpha": 1.0,
97
+ "model_args": {
98
+ "gpt_batch_size": 1,
99
+ "enable_redaction": false,
100
+ "kv_cache": true,
101
+ "gpt_checkpoint": null,
102
+ "clvp_checkpoint": null,
103
+ "decoder_checkpoint": null,
104
+ "num_chars": 255,
105
+ "tokenizer_file": "",
106
+ "gpt_max_audio_tokens": 605,
107
+ "gpt_max_text_tokens": 402,
108
+ "gpt_max_prompt_tokens": 70,
109
+ "gpt_layers": 30,
110
+ "gpt_n_model_channels": 1024,
111
+ "gpt_n_heads": 16,
112
+ "gpt_number_text_tokens": 6681,
113
+ "gpt_start_text_token": null,
114
+ "gpt_stop_text_token": null,
115
+ "gpt_num_audio_tokens": 1026,
116
+ "gpt_start_audio_token": 1024,
117
+ "gpt_stop_audio_token": 1025,
118
+ "gpt_code_stride_len": 1024,
119
+ "gpt_use_masking_gt_prompt_approach": true,
120
+ "gpt_use_perceiver_resampler": true,
121
+ "input_sample_rate": 22050,
122
+ "output_sample_rate": 24000,
123
+ "output_hop_length": 256,
124
+ "decoder_input_dim": 1024,
125
+ "d_vector_dim": 512,
126
+ "cond_d_vector_in_each_upsampling_layer": true,
127
+ "duration_const": 102400
128
+ },
129
+ "model_dir": null,
130
+ "languages": [
131
+ "en",
132
+ "es",
133
+ "fr",
134
+ "de",
135
+ "it",
136
+ "pt",
137
+ "pl",
138
+ "tr",
139
+ "ru",
140
+ "nl",
141
+ "cs",
142
+ "ar",
143
+ "zh-cn",
144
+ "hu",
145
+ "ko",
146
+ "ja",
147
+ "hi"
148
+ ],
149
+ "temperature": 0.75,
150
+ "length_penalty": 1.0,
151
+ "repetition_penalty": 5.0,
152
+ "top_k": 50,
153
+ "top_p": 0.85,
154
+ "num_gpt_outputs": 1,
155
+ "gpt_cond_len": 30,
156
+ "gpt_cond_chunk_len": 4,
157
+ "max_ref_len": 30,
158
+ "sound_norm_refs": false
159
+ }
trainer_0_log.txt ADDED
@@ -0,0 +1,211 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ > Training Environment:
2
+ | > Backend: Torch
3
+ | > Mixed precision: False
4
+ | > Precision: float32
5
+ | > Current device: 0
6
+ | > Num. of GPUs: 1
7
+ | > Num. of CPUs: 2
8
+ | > Num. of Torch Threads: 1
9
+ | > Torch seed: 1
10
+ | > Torch CUDNN: True
11
+ | > Torch CUDNN deterministic: False
12
+ | > Torch CUDNN benchmark: False
13
+ | > Torch TF32 MatMul: False
14
+ > Start Tensorboard: tensorboard --logdir=/tmp/xtts_ft/run/training/GPT_XTTS_FT-December-04-2023_03+00PM-0000000
15
+
16
+ > Model has 518442047 parameters
17
+
18
+  > EPOCH: 0/8
19
+ --> /tmp/xtts_ft/run/training/GPT_XTTS_FT-December-04-2023_03+00PM-0000000
20
+
21
+  > TRAINING (2023-12-04 15:00:11) 
22
+
23
+  --> TIME: 2023-12-04 15:00:19 -- STEP: 0/40 -- GLOBAL_STEP: 0
24
+ | > loss_text_ce: 0.021964602172374725 (0.021964602172374725)
25
+ | > loss_mel_ce: 4.907757759094238 (4.907757759094238)
26
+ | > loss: 4.929722309112549 (4.929722309112549)
27
+ | > grad_norm: 0 (0)
28
+ | > current_lr: 5e-06
29
+ | > step_time: 1.4868 (1.4868354797363281)
30
+ | > loader_time: 6.5441 (6.544092655181885)
31
+
32
+
33
+  > EVALUATION 
34
+
35
+
36
+ --> EVAL PERFORMANCE
37
+ | > avg_loader_time: 0.07195439338684081 (+0)
38
+ | > avg_loss_text_ce: 0.021994752064347266 (+0)
39
+ | > avg_loss_mel_ce: 3.3762893676757812 (+0)
40
+ | > avg_loss: 3.398284101486206 (+0)
41
+
42
+ > BEST MODEL : /tmp/xtts_ft/run/training/GPT_XTTS_FT-December-04-2023_03+00PM-0000000/best_model_40.pth
43
+
44
+  > EPOCH: 1/8
45
+ --> /tmp/xtts_ft/run/training/GPT_XTTS_FT-December-04-2023_03+00PM-0000000
46
+
47
+  > TRAINING (2023-12-04 15:02:56) 
48
+
49
+  --> TIME: 2023-12-04 15:03:03 -- STEP: 10/40 -- GLOBAL_STEP: 50
50
+ | > loss_text_ce: 0.023661285638809204 (0.023023789189755915)
51
+ | > loss_mel_ce: 3.8315794467926025 (3.305363488197327)
52
+ | > loss: 3.855240821838379 (3.328387236595154)
53
+ | > grad_norm: 0 (0.0)
54
+ | > current_lr: 5e-06
55
+ | > step_time: 0.2479 (0.20201985836029052)
56
+ | > loader_time: 0.0221 (0.017214274406433104)
57
+
58
+
59
+  > EVALUATION 
60
+
61
+
62
+ --> EVAL PERFORMANCE
63
+ | > avg_loader_time: 0.1108086109161377 (+0.03885421752929688)
64
+ | > avg_loss_text_ce: 0.021816403046250342 (-0.00017834901809692452)
65
+ | > avg_loss_mel_ce: 3.2666261196136475 (-0.10966324806213379)
66
+ | > avg_loss: 3.288442516326904 (-0.10984158515930176)
67
+
68
+ > BEST MODEL : /tmp/xtts_ft/run/training/GPT_XTTS_FT-December-04-2023_03+00PM-0000000/best_model_80.pth
69
+
70
+  > EPOCH: 2/8
71
+ --> /tmp/xtts_ft/run/training/GPT_XTTS_FT-December-04-2023_03+00PM-0000000
72
+
73
+  > TRAINING (2023-12-04 15:06:18) 
74
+
75
+  --> TIME: 2023-12-04 15:06:33 -- STEP: 20/40 -- GLOBAL_STEP: 100
76
+ | > loss_text_ce: 0.017979152500629425 (0.02109259101562202)
77
+ | > loss_mel_ce: 3.1569440364837646 (2.9896609008312227)
78
+ | > loss: 3.1749231815338135 (3.0107534766197204)
79
+ | > grad_norm: 0 (0.0)
80
+ | > current_lr: 5e-06
81
+ | > step_time: 0.1982 (0.25715711116790774)
82
+ | > loader_time: 0.0208 (0.015216124057769776)
83
+
84
+
85
+  > EVALUATION 
86
+
87
+
88
+ --> EVAL PERFORMANCE
89
+ | > avg_loader_time: 0.06513414382934571 (-0.04567446708679199)
90
+ | > avg_loss_text_ce: 0.021773791685700417 (-4.2611360549924676e-05)
91
+ | > avg_loss_mel_ce: 3.2272082805633544 (-0.03941783905029306)
92
+ | > avg_loss: 3.248982048034668 (-0.03946046829223615)
93
+
94
+ > BEST MODEL : /tmp/xtts_ft/run/training/GPT_XTTS_FT-December-04-2023_03+00PM-0000000/best_model_120.pth
95
+
96
+  > EPOCH: 3/8
97
+ --> /tmp/xtts_ft/run/training/GPT_XTTS_FT-December-04-2023_03+00PM-0000000
98
+
99
+  > TRAINING (2023-12-04 15:10:08) 
100
+
101
+  --> TIME: 2023-12-04 15:10:26 -- STEP: 30/40 -- GLOBAL_STEP: 150
102
+ | > loss_text_ce: 0.021550316363573074 (0.020829477223257224)
103
+ | > loss_mel_ce: 3.5817322731018066 (2.8207703987757364)
104
+ | > loss: 3.6032826900482178 (2.8415998578071595)
105
+ | > grad_norm: 0 (0.0)
106
+ | > current_lr: 5e-06
107
+ | > step_time: 0.2316 (0.26780527432759604)
108
+ | > loader_time: 0.0088 (0.015432175000508625)
109
+
110
+
111
+  > EVALUATION 
112
+
113
+
114
+ --> EVAL PERFORMANCE
115
+ | > avg_loader_time: 0.12182736396789551 (+0.0566932201385498)
116
+ | > avg_loss_text_ce: 0.021623440831899644 (-0.00015035085380077362)
117
+ | > avg_loss_mel_ce: 3.205338716506958 (-0.021869564056396396)
118
+ | > avg_loss: 3.226962184906006 (-0.02201986312866211)
119
+
120
+ > BEST MODEL : /tmp/xtts_ft/run/training/GPT_XTTS_FT-December-04-2023_03+00PM-0000000/best_model_160.pth
121
+
122
+  > EPOCH: 4/8
123
+ --> /tmp/xtts_ft/run/training/GPT_XTTS_FT-December-04-2023_03+00PM-0000000
124
+
125
+  > TRAINING (2023-12-04 15:13:43) 
126
+
127
+  > EVALUATION 
128
+
129
+
130
+ --> EVAL PERFORMANCE
131
+ | > avg_loader_time: 0.06858286857604981 (-0.0532444953918457)
132
+ | > avg_loss_text_ce: 0.021538139879703523 (-8.530095219612052e-05)
133
+ | > avg_loss_mel_ce: 3.19382529258728 (-0.011513423919677912)
134
+ | > avg_loss: 3.2153634548187258 (-0.011598730087280185)
135
+
136
+ > BEST MODEL : /tmp/xtts_ft/run/training/GPT_XTTS_FT-December-04-2023_03+00PM-0000000/best_model_200.pth
137
+
138
+  > EPOCH: 5/8
139
+ --> /tmp/xtts_ft/run/training/GPT_XTTS_FT-December-04-2023_03+00PM-0000000
140
+
141
+  > TRAINING (2023-12-04 15:17:08) 
142
+
143
+  --> TIME: 2023-12-04 15:17:10 -- STEP: 0/40 -- GLOBAL_STEP: 200
144
+ | > loss_text_ce: 0.023776765912771225 (0.023776765912771225)
145
+ | > loss_mel_ce: 2.0250589847564697 (2.0250589847564697)
146
+ | > loss: 2.0488357543945312 (2.0488357543945312)
147
+ | > grad_norm: 0 (0)
148
+ | > current_lr: 5e-06
149
+ | > step_time: 0.9392 (0.9391729831695557)
150
+ | > loader_time: 1.0643 (1.0642502307891846)
151
+
152
+
153
+  > EVALUATION 
154
+
155
+
156
+ --> EVAL PERFORMANCE
157
+ | > avg_loader_time: 0.06770339012145996 (-0.0008794784545898549)
158
+ | > avg_loss_text_ce: 0.02152172140777111 (-1.641847193241397e-05)
159
+ | > avg_loss_mel_ce: 3.21250319480896 (+0.018677902221679865)
160
+ | > avg_loss: 3.234024906158447 (+0.018661451339721413)
161
+
162
+
163
+  > EPOCH: 6/8
164
+ --> /tmp/xtts_ft/run/training/GPT_XTTS_FT-December-04-2023_03+00PM-0000000
165
+
166
+  > TRAINING (2023-12-04 15:17:31) 
167
+
168
+  --> TIME: 2023-12-04 15:17:40 -- STEP: 10/40 -- GLOBAL_STEP: 250
169
+ | > loss_text_ce: 0.018674146384000778 (0.021112211793661118)
170
+ | > loss_mel_ce: 2.7348833084106445 (2.436060166358948)
171
+ | > loss: 2.7535574436187744 (2.457172393798828)
172
+ | > grad_norm: 0 (0.0)
173
+ | > current_lr: 5e-06
174
+ | > step_time: 0.1765 (0.2836411952972412)
175
+ | > loader_time: 0.0086 (0.01687464714050293)
176
+
177
+
178
+  > EVALUATION 
179
+
180
+
181
+ --> EVAL PERFORMANCE
182
+ | > avg_loader_time: 0.06701068878173828 (-0.0006927013397216714)
183
+ | > avg_loss_text_ce: 0.021476111933588983 (-4.560947418212613e-05)
184
+ | > avg_loss_mel_ce: 3.206305408477783 (-0.006197786331176847)
185
+ | > avg_loss: 3.2277815341949463 (-0.006243371963500888)
186
+
187
+
188
+  > EPOCH: 7/8
189
+ --> /tmp/xtts_ft/run/training/GPT_XTTS_FT-December-04-2023_03+00PM-0000000
190
+
191
+  > TRAINING (2023-12-04 15:17:57) 
192
+
193
+  --> TIME: 2023-12-04 15:18:11 -- STEP: 20/40 -- GLOBAL_STEP: 300
194
+ | > loss_text_ce: 0.023174753412604332 (0.019879171112552285)
195
+ | > loss_mel_ce: 3.1051435470581055 (2.4108093440532685)
196
+ | > loss: 3.1283183097839355 (2.430688518285751)
197
+ | > grad_norm: 0 (0.0)
198
+ | > current_lr: 5e-06
199
+ | > step_time: 0.27 (0.29051125049591064)
200
+ | > loader_time: 0.0138 (0.01517837047576904)
201
+
202
+
203
+  > EVALUATION 
204
+
205
+
206
+ --> EVAL PERFORMANCE
207
+ | > avg_loader_time: 0.11030998229980468 (+0.0432992935180664)
208
+ | > avg_loss_text_ce: 0.02148539908230305 (+9.287148714065552e-06)
209
+ | > avg_loss_mel_ce: 3.223378849029541 (+0.01707344055175808)
210
+ | > avg_loss: 3.244864273071289 (+0.017082738876342596)
211
+
vocab.json ADDED
The diff for this file is too large to render. See raw diff