senstella commited on
Commit
485c2ed
·
verified ·
1 Parent(s): d14f0a4

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ pipeline_tag: text-to-speech
4
+ library_name: index-tts
5
+ tags:
6
+ - mlx
7
+ ---
8
+
9
+ # mlx-community/IndexTTS
10
+ This model was converted to MLX format from [`IndexTeam/Index-TTS`](https://huggingface.co/IndexTeam/Index-TTS) using mlx-audio version **0.2.3**.
11
+ Refer to the [original model card](https://huggingface.co/IndexTeam/Index-TTS) for more details on the model.
12
+ ## Use with mlx
13
+
14
+ ```bash
15
+ pip install -U mlx-audio
16
+ ```
17
+
18
+ ```bash
19
+ python -m mlx_audio.tts.generate --model mlx-community/IndexTTS --text "Describe this image."
20
+ ```
bigvgan_generator.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84d41b095e801f1cf01a56feabe9e5381cc97792f43d9791a31b0bf83da2c136
3
+ size 524895512
config.json ADDED
@@ -0,0 +1,178 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bigvgan": {
3
+ "adam_b1": 0.8,
4
+ "adam_b2": 0.99,
5
+ "lr_decay": 0.999998,
6
+ "seed": 1234,
7
+ "resblock": "1",
8
+ "upsample_rates": [
9
+ 4,
10
+ 4,
11
+ 4,
12
+ 4,
13
+ 2,
14
+ 2
15
+ ],
16
+ "upsample_kernel_sizes": [
17
+ 8,
18
+ 8,
19
+ 4,
20
+ 4,
21
+ 4,
22
+ 4
23
+ ],
24
+ "upsample_initial_channel": 1536,
25
+ "resblock_kernel_sizes": [
26
+ 3,
27
+ 7,
28
+ 11
29
+ ],
30
+ "resblock_dilation_sizes": [
31
+ [
32
+ 1,
33
+ 3,
34
+ 5
35
+ ],
36
+ [
37
+ 1,
38
+ 3,
39
+ 5
40
+ ],
41
+ [
42
+ 1,
43
+ 3,
44
+ 5
45
+ ]
46
+ ],
47
+ "feat_upsample": false,
48
+ "speaker_embedding_dim": 512,
49
+ "cond_d_vector_in_each_upsampling_layer": true,
50
+ "gpt_dim": 1024,
51
+ "activation": "snakebeta",
52
+ "snake_logscale": true,
53
+ "use_cqtd_instead_of_mrd": true,
54
+ "cqtd_filters": 128,
55
+ "cqtd_max_filters": 1024,
56
+ "cqtd_filters_scale": 1,
57
+ "cqtd_dilations": [
58
+ 1,
59
+ 2,
60
+ 4
61
+ ],
62
+ "cqtd_hop_lengths": [
63
+ 512,
64
+ 256,
65
+ 256
66
+ ],
67
+ "cqtd_n_octaves": [
68
+ 9,
69
+ 9,
70
+ 9
71
+ ],
72
+ "cqtd_bins_per_octaves": [
73
+ 24,
74
+ 36,
75
+ 48
76
+ ],
77
+ "resolutions": [
78
+ [
79
+ 1024,
80
+ 120,
81
+ 600
82
+ ],
83
+ [
84
+ 2048,
85
+ 240,
86
+ 1200
87
+ ],
88
+ [
89
+ 512,
90
+ 50,
91
+ 240
92
+ ]
93
+ ],
94
+ "mpd_reshapes": [
95
+ 2,
96
+ 3,
97
+ 5,
98
+ 7,
99
+ 11
100
+ ],
101
+ "use_spectral_norm": false,
102
+ "discriminator_channel_mult": 1,
103
+ "use_multiscale_melloss": true,
104
+ "lambda_melloss": 15,
105
+ "clip_grad_norm": 1000,
106
+ "segment_size": 16384,
107
+ "num_mels": 100,
108
+ "num_freq": 1025,
109
+ "n_fft": 1024,
110
+ "hop_size": 256,
111
+ "win_size": 1024,
112
+ "sampling_rate": 24000,
113
+ "fmin": 0,
114
+ "fmax": null,
115
+ "fmax_for_loss": null,
116
+ "mel_type": "pytorch",
117
+ "num_workers": 2,
118
+ "dist_config": {
119
+ "dist_backend": "nccl",
120
+ "dist_url": "tcp://localhost:54321",
121
+ "world_size": 1
122
+ }
123
+ },
124
+ "bigvgan_checkpoint": "bigvgan_generator.pth",
125
+ "dataset": {
126
+ "bpe_model": "checkpoints/bpe.model",
127
+ "sample_rate": 24000,
128
+ "squeeze": false,
129
+ "mel": {
130
+ "sample_rate": 24000,
131
+ "n_fft": 1024,
132
+ "hop_length": 256,
133
+ "win_length": 1024,
134
+ "n_mels": 100,
135
+ "mel_fmin": 0,
136
+ "normalize": false
137
+ }
138
+ },
139
+ "dvae_checkpoint": "dvae.pth",
140
+ "gpt": {
141
+ "model_dim": 1024,
142
+ "max_mel_tokens": 605,
143
+ "max_text_tokens": 402,
144
+ "heads": 16,
145
+ "use_mel_codes_as_input": true,
146
+ "mel_length_compression": 1024,
147
+ "layers": 20,
148
+ "number_text_tokens": 12000,
149
+ "number_mel_codes": 8194,
150
+ "start_mel_token": 8192,
151
+ "stop_mel_token": 8193,
152
+ "start_text_token": 0,
153
+ "stop_text_token": 1,
154
+ "train_solo_embeddings": false,
155
+ "condition_type": "conformer_perceiver",
156
+ "condition_module": {
157
+ "output_size": 512,
158
+ "linear_units": 2048,
159
+ "attention_heads": 8,
160
+ "num_blocks": 6,
161
+ "input_layer": "conv2d2",
162
+ "perceiver_mult": 2
163
+ }
164
+ },
165
+ "gpt_checkpoint": "gpt.pth",
166
+ "vqvae": {
167
+ "channels": 100,
168
+ "num_tokens": 8192,
169
+ "hidden_dim": 512,
170
+ "num_resnet_blocks": 3,
171
+ "codebook_dim": 512,
172
+ "num_layers": 2,
173
+ "positional_dims": 1,
174
+ "kernel_size": 3,
175
+ "smooth_l1_loss": true,
176
+ "use_transposed_convs": false
177
+ }
178
+ }
gpt.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2921cef3d321d4759d6f85c0df42c1a55d84c6adc0a12637d8e7b2c424fa6bc
3
+ size 696434870
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a4b1b52c4078878b010ba47224c93ed857261d59b9af38acca5e5c4744046ec
3
+ size 953746416
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff