darshanmakwana commited on
Commit
855d8c4
·
verified ·
1 Parent(s): e0c2d04

Upload folder using huggingface_hub

Browse files
models/.ipynb_checkpoints/config-checkpoint.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "resblock": "1",
3
+ "num_gpus": 3,
4
+ "batch_size": 60,
5
+ "learning_rate": 0.0001,
6
+ "adam_b1": 0.5,
7
+ "adam_b2": 0.9,
8
+ "lr_decay": 0.98,
9
+ "seed": 1234,
10
+ "lambda_distill": 0.15,
11
+
12
+ "n_filters": 64,
13
+ "strides": [8,5,4,2],
14
+ "dimension": 1024,
15
+ "semantic_dimension": 768,
16
+ "bidirectional": true,
17
+ "dilation_base": 2,
18
+ "residual_kernel_size": 3,
19
+ "n_residual_layers": 1,
20
+ "lstm_layers": 2,
21
+ "activation": "ELU",
22
+
23
+
24
+ "segment_size": 48000,
25
+ "num_mels": 80,
26
+ "num_freq": 1025,
27
+ "n_fft": 1024,
28
+ "hop_size": 240,
29
+ "win_size": 1024,
30
+
31
+ "sampling_rate": 16000,
32
+ "sample_rate": 16000,
33
+
34
+ "codebook_size": 1024,
35
+ "n_q": 8,
36
+
37
+ "fmin": 0,
38
+ "fmax": 8000,
39
+ "fmax_for_loss": null,
40
+
41
+ "num_workers": 12,
42
+
43
+ "dist_config": {
44
+ "dist_backend": "nccl",
45
+ "dist_url": "tcp://localhost:54322",
46
+ "world_size": 1
47
+ }
48
+ }
models/checkpoint-10000/config.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "gpt2",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPT2LMHeadModel"
6
+ ],
7
+ "attn_pdrop": 0.1,
8
+ "bos_token_id": 50256,
9
+ "embd_pdrop": 0.1,
10
+ "eos_token_id": 50256,
11
+ "initializer_range": 0.02,
12
+ "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_ctx": 1024,
15
+ "n_embd": 768,
16
+ "n_head": 12,
17
+ "n_inner": null,
18
+ "n_layer": 12,
19
+ "n_positions": 1877,
20
+ "pad_token_id": 1026,
21
+ "reorder_and_upcast_attn": false,
22
+ "resid_pdrop": 0.1,
23
+ "scale_attn_by_inverse_layer_idx": false,
24
+ "scale_attn_weights": true,
25
+ "summary_activation": null,
26
+ "summary_first_dropout": 0.1,
27
+ "summary_proj_to_labels": true,
28
+ "summary_type": "cls_index",
29
+ "summary_use_proj": true,
30
+ "task_specific_params": {
31
+ "text-generation": {
32
+ "do_sample": true,
33
+ "max_length": 50
34
+ }
35
+ },
36
+ "torch_dtype": "float32",
37
+ "transformers_version": "4.41.2",
38
+ "use_cache": true,
39
+ "vocab_size": 6027
40
+ }
models/checkpoint-10000/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 50256,
4
+ "eos_token_id": 50256,
5
+ "pad_token_id": 1026,
6
+ "transformers_version": "4.41.2"
7
+ }
models/checkpoint-10000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71210d9cf05a4a71355518f1db65691da071bfe0ad4915c62287c0179f769778
3
+ size 364520064
models/checkpoint-10000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a40385edf3632f109407116c8abd0d09997740d5df389f052283035aed5a10b
3
+ size 729134010
models/checkpoint-10000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f4ba3793dc249607794ec829c5ea04db1de200de17944eece563c95795e2ee8
3
+ size 14244
models/checkpoint-10000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:082303d814e0ecbb2362009f05712441bc4dd59741b38d940cb9613ca211fb09
3
+ size 1064
models/checkpoint-10000/special_tokens_map.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "pad_token": {
3
+ "content": "<|padding|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ }
9
+ }
models/checkpoint-10000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
models/checkpoint-10000/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
models/checkpoint-10000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
models/checkpoint-10000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab476b2a193a8bb001f6e0c70ae3640ae41d93adf15e6543544ac2aba8e1587f
3
+ size 5112
models/ckpt_librispeech_clean.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11edfb7f9c4fc84d7b2930855efe94aeabeaae3cd99d849f4015f2241b5e6fbc
3
+ size 1170787339
models/ckpt_librispeech_other.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4bcf9d781f9157bff3342fe77381230869ce30832fce779972d6036f0740e65
3
+ size 1509396715
models/config.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "resblock": "1",
3
+ "num_gpus": 3,
4
+ "batch_size": 60,
5
+ "learning_rate": 0.0001,
6
+ "adam_b1": 0.5,
7
+ "adam_b2": 0.9,
8
+ "lr_decay": 0.98,
9
+ "seed": 1234,
10
+ "lambda_distill": 0.15,
11
+
12
+ "n_filters": 64,
13
+ "strides": [8,5,4,2],
14
+ "dimension": 1024,
15
+ "semantic_dimension": 768,
16
+ "bidirectional": true,
17
+ "dilation_base": 2,
18
+ "residual_kernel_size": 3,
19
+ "n_residual_layers": 1,
20
+ "lstm_layers": 2,
21
+ "activation": "ELU",
22
+
23
+
24
+ "segment_size": 48000,
25
+ "num_mels": 80,
26
+ "num_freq": 1025,
27
+ "n_fft": 1024,
28
+ "hop_size": 240,
29
+ "win_size": 1024,
30
+
31
+ "sampling_rate": 16000,
32
+ "sample_rate": 16000,
33
+
34
+ "codebook_size": 1024,
35
+ "n_q": 8,
36
+
37
+ "fmin": 0,
38
+ "fmax": 8000,
39
+ "fmax_for_loss": null,
40
+
41
+ "num_workers": 12,
42
+
43
+ "dist_config": {
44
+ "dist_backend": "nccl",
45
+ "dist_url": "tcp://localhost:54322",
46
+ "world_size": 1
47
+ }
48
+ }
models/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f987b64ae39991d8cb84e736a3a3d8813d8139809bbdcfdf8b2990ed4d6b440b
3
+ size 970027718
models/tokenizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d04593b6c9a4b475f91ca481141a6ef5b23e6ac112f347dd2b2717f193c1c728
3
+ size 481906997
models/vocoder.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d1f7188b95b06304bc05e524fddf93c7fe682fdd93acff022685663a5e26b97
3
+ size 54051213
models/vocoder_config.json ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "input_wavs_dir": "/private/home/adampolyak/datasets/LJ/LJSpeech-1.1/wavs_16khz_padded",
3
+ "input_training_file": "/large_experiments/ust/annl/datasets/tts/LJSpeech/filelist/mhubert_vp_en_es_fr_it3_400k/lj_train_layer11_hubert1000_filelist.txt",
4
+ "input_validation_file": "/large_experiments/ust/annl/datasets/tts/LJSpeech/filelist/mhubert_vp_en_es_fr_it3_400k/lj_dev_layer11_hubert1000_filelist.txt",
5
+
6
+ "resblock": "1",
7
+ "num_gpus": 0,
8
+ "batch_size": 16,
9
+ "learning_rate": 0.0002,
10
+ "adam_b1": 0.8,
11
+ "adam_b2": 0.99,
12
+ "lr_decay": 0.999,
13
+ "seed": 1234,
14
+
15
+ "upsample_rates": [5,4,4,2,2],
16
+ "upsample_kernel_sizes": [11,8,8,4,4],
17
+ "upsample_initial_channel": 512,
18
+ "resblock_kernel_sizes": [3,7,11],
19
+ "resblock_dilation_sizes": [[1,3,5], [1,3,5], [1,3,5]],
20
+ "num_embeddings": 1000,
21
+ "embedding_dim": 128,
22
+ "model_in_dim": 128,
23
+
24
+ "segment_size": 8960,
25
+ "code_hop_size": 320,
26
+ "f0": false,
27
+ "num_mels": 80,
28
+ "num_freq": 1025,
29
+ "n_fft": 1024,
30
+ "hop_size": 256,
31
+ "win_size": 1024,
32
+
33
+ "dur_prediction_weight": 1.0,
34
+ "dur_predictor_params": {
35
+ "encoder_embed_dim": 128,
36
+ "var_pred_hidden_dim": 128,
37
+ "var_pred_kernel_size": 3,
38
+ "var_pred_dropout": 0.5
39
+ },
40
+
41
+ "sampling_rate": 16000,
42
+
43
+ "fmin": 0,
44
+ "fmax": 8000,
45
+ "fmax_for_loss": null,
46
+
47
+ "num_workers": 4,
48
+
49
+ "dist_config": {
50
+ "dist_backend": "nccl",
51
+ "dist_url": "env://"
52
+ }
53
+ }