mrq commited on
Commit
1ea585e
1 Parent(s): 64c7ab6

replaced ar+nar-tts+stt-llama-8 as the base ar+nar-llama-8 since it pretty much surpasses it now

Browse files
models/ckpt/ar+nar-llama-8/fp32.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:aca5e72710bcff66244d823b8bb887ebb48621bd74fbe8aacdac7e21cb2e7823
3
- size 455756877
 
 
 
 
models/ckpt/ar+nar-llama-8/fp32.sft CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2bbafd8afb5403c206c28f51ea3e872769dab8de99b5f441825ff31c893b0911
3
- size 455745602
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0920e7eef0884631f00513b700924538db0853d662530e4bcf7ac1d8666430b6
3
+ size 456274402
models/ckpt/ar+nar-tts+stt-llama-8/fp32.sft DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:0920e7eef0884631f00513b700924538db0853d662530e4bcf7ac1d8666430b6
3
- size 456274402
 
 
 
 
models/config.llama.tts+stt.yaml DELETED
@@ -1,130 +0,0 @@
1
- sample_rate: 24_000
2
- audio_backend: "vocos"
3
-
4
- models:
5
- - name: "ar+nar-tts+stt"
6
- size: "full"
7
- resp_levels: 8
8
- prom_levels: 8
9
- tasks: 9
10
- langs: 4
11
- tones: 1
12
- arch_type: llama
13
- training: False
14
- version: 5
15
- attention: auto
16
- dropout: 0.1
17
- #loss_factors:
18
- # text: 0.01
19
- # prom: 0.5
20
- # resp: 1.0
21
- capabilities: ["ar", "nar"]
22
- experimental:
23
- p_rvq_levels: "auto"
24
- audio_embedding_sums: True
25
- unified_position_ids: False
26
- split_classifiers: True
27
- #
28
- causal_size: 1
29
- interleave: False
30
- rvq_level_range: []
31
- tie_classifier_to_embedding: False
32
-
33
- #loras:
34
- #- name : "lora"
35
- # rank: 128
36
- # alpha: 128
37
- # training: True
38
- # rvq_levels: []
39
-
40
- hyperparameters:
41
- batch_size: 32
42
- gradient_accumulation_steps: 8
43
- gradient_clipping: 1.0
44
- warmup_steps: 10
45
-
46
- optimizer: Prodigy
47
- learning_rate: 1.0
48
- torch_optimizer: True
49
-
50
- scheduler: "" # ScheduleFree
51
- torch_scheduler: True
52
-
53
- evaluation:
54
- batch_size: 4
55
- frequency: 250
56
- size: 4
57
-
58
- steps: 500
59
- ar_temperature: 1.0
60
- nar_temperature: 0.0
61
-
62
- trainer:
63
- iterations: 1_000_000
64
- save_frequency: 250
65
- keep_last_checkpoints: 4
66
-
67
- resize_modules: True
68
-
69
- check_for_oom: False
70
- gradient_checkpointing: True
71
-
72
- weight_dtype: bfloat16
73
- amp: True
74
-
75
- backend: deepspeed
76
- deepspeed:
77
- inferencing: False
78
- amp: False
79
-
80
- load_webui: False
81
-
82
- inference:
83
- backend: local
84
- normalize: False
85
-
86
- weight_dtype: bfloat16
87
- amp: True
88
-
89
- optimizations:
90
- injects: False
91
- replace: True
92
-
93
- linear: False
94
- embedding: False
95
- optimizers: True
96
-
97
- bitsandbytes: False
98
- dadaptation: False
99
- bitnet: False
100
- fp8: False
101
-
102
- dataset:
103
- use_hdf5: True
104
- hdf5_flag: r
105
-
106
- use_metadata: True
107
- validate: True
108
-
109
- workers: 1
110
- cache: True
111
-
112
- duration_range: [3.0, 12.0]
113
-
114
- random_utterance: 1.0
115
- max_prompts: 1
116
- prompt_duration_range: [3.0, 3.0]
117
-
118
- max_resps: 1
119
- p_resp_append: 0.25
120
-
121
- sample_type: path # path # speaker
122
- sample_order: duration
123
- sample_max_duration_batch: 300
124
- sample_shuffle: False
125
-
126
- tasks_list: [ "tts", "stt" ]
127
-
128
- training: []
129
- validation: []
130
- noise: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/config.llama.yaml CHANGED
@@ -1,145 +1,130 @@
1
- sample_rate: 24_000
2
- audio_backend: "vocos"
3
-
4
- models:
5
- - name: "ar+nar"
6
- size: "full"
7
- resp_levels: 8
8
- prom_levels: 8
9
- tasks: 8
10
- langs: 2
11
- tones: 1
12
- arch_type: llama
13
- training: False
14
- version: 5
15
- attention: auto
16
- dropout: 0.1
17
- loss_factors:
18
- text: 0.01
19
- prom: 0.5
20
- resp: 1.0
21
- capabilities: ["ar", "nar"]
22
- experimental:
23
- audio_embedding_sums: False
24
- unified_position_ids: False
25
- split_classifiers: True
26
-
27
- hyperparameters:
28
- autotune: False
29
- autotune_params:
30
- start_profile_step: 1
31
- end_profile_step: 50
32
- num_tuning_micro_batch_sizes: 8
33
-
34
- batch_size: 16
35
- gradient_accumulation_steps: 8
36
- gradient_clipping: 1.0
37
- warmup_steps: 250
38
-
39
- optimizer: Prodigy
40
- learning_rate: 1.0
41
- torch_optimizer: True
42
-
43
- scheduler: "" # ScheduleFree
44
- torch_scheduler: True
45
-
46
- evaluation:
47
- batch_size: 16
48
- frequency: 1000
49
- size: 16
50
-
51
- steps: 500
52
- ar_temperature: 0.95
53
- nar_temperature: 0.25
54
- load_disabled_engines: True
55
-
56
- trainer:
57
- #no_logger: True
58
- ddp: False
59
- check_for_oom: False
60
- iterations: 1_000_000
61
-
62
- save_tag: step
63
- save_on_oom: True
64
- save_on_quit: True
65
- save_frequency: 500
66
- export_on_save: True
67
-
68
- keep_last_checkpoints: 8
69
-
70
- aggressive_optimizations: False
71
- load_disabled_engines: False
72
- gradient_checkpointing: True
73
-
74
- #load_state_dict: True
75
- strict_loading: False
76
- #load_tag: "9500"
77
- #load_states: False
78
- #restart_step_count: True
79
-
80
- gc_mode: None # "global_step"
81
-
82
- weight_dtype: bfloat16
83
- amp: True
84
-
85
- backend: deepspeed
86
- deepspeed:
87
- inferencing: True
88
- zero_optimization_level: 0
89
- use_compression_training: False
90
-
91
- amp: False
92
-
93
- load_webui: False
94
-
95
- inference:
96
- backend: deepspeed
97
- audio_backend: "vocos"
98
- normalize: False
99
-
100
- weight_dtype: bfloat16
101
- amp: True
102
-
103
- optimizations:
104
- injects: False
105
- replace: True
106
-
107
- linear: False
108
- embedding: False
109
- optimizers: True
110
-
111
- bitsandbytes: False
112
- dadaptation: False
113
- bitnet: False
114
- fp8: False
115
-
116
- dataset:
117
- speaker_name_getter: "lambda p: f'{p.parts[-3]}_{p.parts[-2]}'"
118
- speaker_group_getter: "lambda p: f'{p.parts[-3]}'"
119
- speaker_languages:
120
- ja: []
121
-
122
- use_hdf5: True
123
- use_metadata: True
124
- hdf5_flag: r
125
- validate: True
126
-
127
- workers: 6
128
- cache: True
129
-
130
- duration_range: [24.0, 32.0]
131
-
132
- random_utterance: 1.0
133
- max_prompts: 1
134
- prompt_duration_range: [3.0, 9.0]
135
-
136
- max_resps: 1
137
- p_resp_append: 0.25
138
-
139
- sample_type: path # path # speaker
140
-
141
- tasks_list: [ "tts" ] # , [ "tts", "tts-c", "ns", "sr", "tse", "cse", "nse", "tts"]
142
-
143
- training: []
144
- validation: []
145
- noise: []
 
1
+ sample_rate: 24_000
2
+ audio_backend: "vocos"
3
+
4
+ models:
5
+ - name: "ar+nar-tts+stt"
6
+ size: "full"
7
+ resp_levels: 8
8
+ prom_levels: 8
9
+ tasks: 9
10
+ langs: 4
11
+ tones: 1
12
+ arch_type: llama
13
+ training: False
14
+ version: 5
15
+ attention: auto
16
+ dropout: 0.1
17
+ #loss_factors:
18
+ # text: 0.01
19
+ # prom: 0.5
20
+ # resp: 1.0
21
+ capabilities: ["ar", "nar"]
22
+ experimental:
23
+ p_rvq_levels: "auto"
24
+ audio_embedding_sums: True
25
+ unified_position_ids: False
26
+ split_classifiers: True
27
+ #
28
+ causal_size: 1
29
+ interleave: False
30
+ rvq_level_range: []
31
+ tie_classifier_to_embedding: False
32
+
33
+ #loras:
34
+ #- name : "lora"
35
+ # rank: 128
36
+ # alpha: 128
37
+ # training: True
38
+ # rvq_levels: []
39
+
40
+ hyperparameters:
41
+ batch_size: 32
42
+ gradient_accumulation_steps: 8
43
+ gradient_clipping: 1.0
44
+ warmup_steps: 10
45
+
46
+ optimizer: Prodigy
47
+ learning_rate: 1.0
48
+ torch_optimizer: True
49
+
50
+ scheduler: "" # ScheduleFree
51
+ torch_scheduler: True
52
+
53
+ evaluation:
54
+ batch_size: 4
55
+ frequency: 250
56
+ size: 4
57
+
58
+ steps: 500
59
+ ar_temperature: 1.0
60
+ nar_temperature: 0.0
61
+
62
+ trainer:
63
+ iterations: 1_000_000
64
+ save_frequency: 250
65
+ keep_last_checkpoints: 4
66
+
67
+ resize_modules: True
68
+
69
+ check_for_oom: False
70
+ gradient_checkpointing: True
71
+
72
+ weight_dtype: bfloat16
73
+ amp: True
74
+
75
+ backend: deepspeed
76
+ deepspeed:
77
+ inferencing: False
78
+ amp: False
79
+
80
+ load_webui: False
81
+
82
+ inference:
83
+ backend: local
84
+ normalize: False
85
+
86
+ weight_dtype: bfloat16
87
+ amp: True
88
+
89
+ optimizations:
90
+ injects: False
91
+ replace: True
92
+
93
+ linear: False
94
+ embedding: False
95
+ optimizers: True
96
+
97
+ bitsandbytes: False
98
+ dadaptation: False
99
+ bitnet: False
100
+ fp8: False
101
+
102
+ dataset:
103
+ use_hdf5: True
104
+ hdf5_flag: r
105
+
106
+ use_metadata: True
107
+ validate: True
108
+
109
+ workers: 1
110
+ cache: True
111
+
112
+ duration_range: [3.0, 12.0]
113
+
114
+ random_utterance: 1.0
115
+ max_prompts: 1
116
+ prompt_duration_range: [3.0, 3.0]
117
+
118
+ max_resps: 1
119
+ p_resp_append: 0.25
120
+
121
+ sample_type: path # path # speaker
122
+ sample_order: duration
123
+ sample_max_duration_batch: 300
124
+ sample_shuffle: False
125
+
126
+ tasks_list: [ "tts", "stt" ]
127
+
128
+ training: []
129
+ validation: []
130
+ noise: []