akxier commited on
Commit
599eef8
·
verified ·
1 Parent(s): e42165d

Commit inicial de Whisper medium dataset asfeme

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. added_tokens.json +1609 -0
  2. checkpoint-10000/config.json +144 -0
  3. checkpoint-10000/generation_config.json +104 -0
  4. checkpoint-10000/optimizer.pt +3 -0
  5. checkpoint-10000/preprocessor_config.json +0 -0
  6. checkpoint-10000/pytorch_model.bin +3 -0
  7. checkpoint-10000/rng_state.pth +3 -0
  8. checkpoint-10000/scaler.pt +3 -0
  9. checkpoint-10000/scheduler.pt +3 -0
  10. checkpoint-10000/trainer_state.json +2641 -0
  11. checkpoint-10000/training_args.bin +3 -0
  12. checkpoint-1200/config.json +144 -0
  13. checkpoint-1200/generation_config.json +104 -0
  14. checkpoint-1200/optimizer.pt +3 -0
  15. checkpoint-1200/preprocessor_config.json +0 -0
  16. checkpoint-1200/pytorch_model.bin +3 -0
  17. checkpoint-1200/rng_state.pth +3 -0
  18. checkpoint-1200/scaler.pt +3 -0
  19. checkpoint-1200/scheduler.pt +3 -0
  20. checkpoint-1200/trainer_state.json +331 -0
  21. checkpoint-1200/training_args.bin +3 -0
  22. checkpoint-1600/config.json +144 -0
  23. checkpoint-1600/generation_config.json +104 -0
  24. checkpoint-1600/optimizer.pt +3 -0
  25. checkpoint-1600/preprocessor_config.json +0 -0
  26. checkpoint-1600/pytorch_model.bin +3 -0
  27. checkpoint-1600/rng_state.pth +3 -0
  28. checkpoint-1600/scaler.pt +3 -0
  29. checkpoint-1600/scheduler.pt +3 -0
  30. checkpoint-1600/trainer_state.json +436 -0
  31. checkpoint-1600/training_args.bin +3 -0
  32. checkpoint-2000/config.json +144 -0
  33. checkpoint-2000/generation_config.json +104 -0
  34. checkpoint-2000/optimizer.pt +3 -0
  35. checkpoint-2000/preprocessor_config.json +0 -0
  36. checkpoint-2000/pytorch_model.bin +3 -0
  37. checkpoint-2000/rng_state.pth +3 -0
  38. checkpoint-2000/scaler.pt +3 -0
  39. checkpoint-2000/scheduler.pt +3 -0
  40. checkpoint-2000/trainer_state.json +541 -0
  41. checkpoint-2000/training_args.bin +3 -0
  42. checkpoint-2400/config.json +144 -0
  43. checkpoint-2400/generation_config.json +104 -0
  44. checkpoint-2400/optimizer.pt +3 -0
  45. checkpoint-2400/preprocessor_config.json +0 -0
  46. checkpoint-2400/pytorch_model.bin +3 -0
  47. checkpoint-2400/rng_state.pth +3 -0
  48. checkpoint-2400/scaler.pt +3 -0
  49. checkpoint-2400/scheduler.pt +3 -0
  50. checkpoint-2400/trainer_state.json +646 -0
added_tokens.json ADDED
@@ -0,0 +1,1609 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "<|0.00|>": 50364,
3
+ "<|0.02|>": 50365,
4
+ "<|0.04|>": 50366,
5
+ "<|0.06|>": 50367,
6
+ "<|0.08|>": 50368,
7
+ "<|0.10|>": 50369,
8
+ "<|0.12|>": 50370,
9
+ "<|0.14|>": 50371,
10
+ "<|0.16|>": 50372,
11
+ "<|0.18|>": 50373,
12
+ "<|0.20|>": 50374,
13
+ "<|0.22|>": 50375,
14
+ "<|0.24|>": 50376,
15
+ "<|0.26|>": 50377,
16
+ "<|0.28|>": 50378,
17
+ "<|0.30|>": 50379,
18
+ "<|0.32|>": 50380,
19
+ "<|0.34|>": 50381,
20
+ "<|0.36|>": 50382,
21
+ "<|0.38|>": 50383,
22
+ "<|0.40|>": 50384,
23
+ "<|0.42|>": 50385,
24
+ "<|0.44|>": 50386,
25
+ "<|0.46|>": 50387,
26
+ "<|0.48|>": 50388,
27
+ "<|0.50|>": 50389,
28
+ "<|0.52|>": 50390,
29
+ "<|0.54|>": 50391,
30
+ "<|0.56|>": 50392,
31
+ "<|0.58|>": 50393,
32
+ "<|0.60|>": 50394,
33
+ "<|0.62|>": 50395,
34
+ "<|0.64|>": 50396,
35
+ "<|0.66|>": 50397,
36
+ "<|0.68|>": 50398,
37
+ "<|0.70|>": 50399,
38
+ "<|0.72|>": 50400,
39
+ "<|0.74|>": 50401,
40
+ "<|0.76|>": 50402,
41
+ "<|0.78|>": 50403,
42
+ "<|0.80|>": 50404,
43
+ "<|0.82|>": 50405,
44
+ "<|0.84|>": 50406,
45
+ "<|0.86|>": 50407,
46
+ "<|0.88|>": 50408,
47
+ "<|0.90|>": 50409,
48
+ "<|0.92|>": 50410,
49
+ "<|0.94|>": 50411,
50
+ "<|0.96|>": 50412,
51
+ "<|0.98|>": 50413,
52
+ "<|1.00|>": 50414,
53
+ "<|1.02|>": 50415,
54
+ "<|1.04|>": 50416,
55
+ "<|1.06|>": 50417,
56
+ "<|1.08|>": 50418,
57
+ "<|1.10|>": 50419,
58
+ "<|1.12|>": 50420,
59
+ "<|1.14|>": 50421,
60
+ "<|1.16|>": 50422,
61
+ "<|1.18|>": 50423,
62
+ "<|1.20|>": 50424,
63
+ "<|1.22|>": 50425,
64
+ "<|1.24|>": 50426,
65
+ "<|1.26|>": 50427,
66
+ "<|1.28|>": 50428,
67
+ "<|1.30|>": 50429,
68
+ "<|1.32|>": 50430,
69
+ "<|1.34|>": 50431,
70
+ "<|1.36|>": 50432,
71
+ "<|1.38|>": 50433,
72
+ "<|1.40|>": 50434,
73
+ "<|1.42|>": 50435,
74
+ "<|1.44|>": 50436,
75
+ "<|1.46|>": 50437,
76
+ "<|1.48|>": 50438,
77
+ "<|1.50|>": 50439,
78
+ "<|1.52|>": 50440,
79
+ "<|1.54|>": 50441,
80
+ "<|1.56|>": 50442,
81
+ "<|1.58|>": 50443,
82
+ "<|1.60|>": 50444,
83
+ "<|1.62|>": 50445,
84
+ "<|1.64|>": 50446,
85
+ "<|1.66|>": 50447,
86
+ "<|1.68|>": 50448,
87
+ "<|1.70|>": 50449,
88
+ "<|1.72|>": 50450,
89
+ "<|1.74|>": 50451,
90
+ "<|1.76|>": 50452,
91
+ "<|1.78|>": 50453,
92
+ "<|1.80|>": 50454,
93
+ "<|1.82|>": 50455,
94
+ "<|1.84|>": 50456,
95
+ "<|1.86|>": 50457,
96
+ "<|1.88|>": 50458,
97
+ "<|1.90|>": 50459,
98
+ "<|1.92|>": 50460,
99
+ "<|1.94|>": 50461,
100
+ "<|1.96|>": 50462,
101
+ "<|1.98|>": 50463,
102
+ "<|10.00|>": 50864,
103
+ "<|10.02|>": 50865,
104
+ "<|10.04|>": 50866,
105
+ "<|10.06|>": 50867,
106
+ "<|10.08|>": 50868,
107
+ "<|10.10|>": 50869,
108
+ "<|10.12|>": 50870,
109
+ "<|10.14|>": 50871,
110
+ "<|10.16|>": 50872,
111
+ "<|10.18|>": 50873,
112
+ "<|10.20|>": 50874,
113
+ "<|10.22|>": 50875,
114
+ "<|10.24|>": 50876,
115
+ "<|10.26|>": 50877,
116
+ "<|10.28|>": 50878,
117
+ "<|10.30|>": 50879,
118
+ "<|10.32|>": 50880,
119
+ "<|10.34|>": 50881,
120
+ "<|10.36|>": 50882,
121
+ "<|10.38|>": 50883,
122
+ "<|10.40|>": 50884,
123
+ "<|10.42|>": 50885,
124
+ "<|10.44|>": 50886,
125
+ "<|10.46|>": 50887,
126
+ "<|10.48|>": 50888,
127
+ "<|10.50|>": 50889,
128
+ "<|10.52|>": 50890,
129
+ "<|10.54|>": 50891,
130
+ "<|10.56|>": 50892,
131
+ "<|10.58|>": 50893,
132
+ "<|10.60|>": 50894,
133
+ "<|10.62|>": 50895,
134
+ "<|10.64|>": 50896,
135
+ "<|10.66|>": 50897,
136
+ "<|10.68|>": 50898,
137
+ "<|10.70|>": 50899,
138
+ "<|10.72|>": 50900,
139
+ "<|10.74|>": 50901,
140
+ "<|10.76|>": 50902,
141
+ "<|10.78|>": 50903,
142
+ "<|10.80|>": 50904,
143
+ "<|10.82|>": 50905,
144
+ "<|10.84|>": 50906,
145
+ "<|10.86|>": 50907,
146
+ "<|10.88|>": 50908,
147
+ "<|10.90|>": 50909,
148
+ "<|10.92|>": 50910,
149
+ "<|10.94|>": 50911,
150
+ "<|10.96|>": 50912,
151
+ "<|10.98|>": 50913,
152
+ "<|11.00|>": 50914,
153
+ "<|11.02|>": 50915,
154
+ "<|11.04|>": 50916,
155
+ "<|11.06|>": 50917,
156
+ "<|11.08|>": 50918,
157
+ "<|11.10|>": 50919,
158
+ "<|11.12|>": 50920,
159
+ "<|11.14|>": 50921,
160
+ "<|11.16|>": 50922,
161
+ "<|11.18|>": 50923,
162
+ "<|11.20|>": 50924,
163
+ "<|11.22|>": 50925,
164
+ "<|11.24|>": 50926,
165
+ "<|11.26|>": 50927,
166
+ "<|11.28|>": 50928,
167
+ "<|11.30|>": 50929,
168
+ "<|11.32|>": 50930,
169
+ "<|11.34|>": 50931,
170
+ "<|11.36|>": 50932,
171
+ "<|11.38|>": 50933,
172
+ "<|11.40|>": 50934,
173
+ "<|11.42|>": 50935,
174
+ "<|11.44|>": 50936,
175
+ "<|11.46|>": 50937,
176
+ "<|11.48|>": 50938,
177
+ "<|11.50|>": 50939,
178
+ "<|11.52|>": 50940,
179
+ "<|11.54|>": 50941,
180
+ "<|11.56|>": 50942,
181
+ "<|11.58|>": 50943,
182
+ "<|11.60|>": 50944,
183
+ "<|11.62|>": 50945,
184
+ "<|11.64|>": 50946,
185
+ "<|11.66|>": 50947,
186
+ "<|11.68|>": 50948,
187
+ "<|11.70|>": 50949,
188
+ "<|11.72|>": 50950,
189
+ "<|11.74|>": 50951,
190
+ "<|11.76|>": 50952,
191
+ "<|11.78|>": 50953,
192
+ "<|11.80|>": 50954,
193
+ "<|11.82|>": 50955,
194
+ "<|11.84|>": 50956,
195
+ "<|11.86|>": 50957,
196
+ "<|11.88|>": 50958,
197
+ "<|11.90|>": 50959,
198
+ "<|11.92|>": 50960,
199
+ "<|11.94|>": 50961,
200
+ "<|11.96|>": 50962,
201
+ "<|11.98|>": 50963,
202
+ "<|12.00|>": 50964,
203
+ "<|12.02|>": 50965,
204
+ "<|12.04|>": 50966,
205
+ "<|12.06|>": 50967,
206
+ "<|12.08|>": 50968,
207
+ "<|12.10|>": 50969,
208
+ "<|12.12|>": 50970,
209
+ "<|12.14|>": 50971,
210
+ "<|12.16|>": 50972,
211
+ "<|12.18|>": 50973,
212
+ "<|12.20|>": 50974,
213
+ "<|12.22|>": 50975,
214
+ "<|12.24|>": 50976,
215
+ "<|12.26|>": 50977,
216
+ "<|12.28|>": 50978,
217
+ "<|12.30|>": 50979,
218
+ "<|12.32|>": 50980,
219
+ "<|12.34|>": 50981,
220
+ "<|12.36|>": 50982,
221
+ "<|12.38|>": 50983,
222
+ "<|12.40|>": 50984,
223
+ "<|12.42|>": 50985,
224
+ "<|12.44|>": 50986,
225
+ "<|12.46|>": 50987,
226
+ "<|12.48|>": 50988,
227
+ "<|12.50|>": 50989,
228
+ "<|12.52|>": 50990,
229
+ "<|12.54|>": 50991,
230
+ "<|12.56|>": 50992,
231
+ "<|12.58|>": 50993,
232
+ "<|12.60|>": 50994,
233
+ "<|12.62|>": 50995,
234
+ "<|12.64|>": 50996,
235
+ "<|12.66|>": 50997,
236
+ "<|12.68|>": 50998,
237
+ "<|12.70|>": 50999,
238
+ "<|12.72|>": 51000,
239
+ "<|12.74|>": 51001,
240
+ "<|12.76|>": 51002,
241
+ "<|12.78|>": 51003,
242
+ "<|12.80|>": 51004,
243
+ "<|12.82|>": 51005,
244
+ "<|12.84|>": 51006,
245
+ "<|12.86|>": 51007,
246
+ "<|12.88|>": 51008,
247
+ "<|12.90|>": 51009,
248
+ "<|12.92|>": 51010,
249
+ "<|12.94|>": 51011,
250
+ "<|12.96|>": 51012,
251
+ "<|12.98|>": 51013,
252
+ "<|13.00|>": 51014,
253
+ "<|13.02|>": 51015,
254
+ "<|13.04|>": 51016,
255
+ "<|13.06|>": 51017,
256
+ "<|13.08|>": 51018,
257
+ "<|13.10|>": 51019,
258
+ "<|13.12|>": 51020,
259
+ "<|13.14|>": 51021,
260
+ "<|13.16|>": 51022,
261
+ "<|13.18|>": 51023,
262
+ "<|13.20|>": 51024,
263
+ "<|13.22|>": 51025,
264
+ "<|13.24|>": 51026,
265
+ "<|13.26|>": 51027,
266
+ "<|13.28|>": 51028,
267
+ "<|13.30|>": 51029,
268
+ "<|13.32|>": 51030,
269
+ "<|13.34|>": 51031,
270
+ "<|13.36|>": 51032,
271
+ "<|13.38|>": 51033,
272
+ "<|13.40|>": 51034,
273
+ "<|13.42|>": 51035,
274
+ "<|13.44|>": 51036,
275
+ "<|13.46|>": 51037,
276
+ "<|13.48|>": 51038,
277
+ "<|13.50|>": 51039,
278
+ "<|13.52|>": 51040,
279
+ "<|13.54|>": 51041,
280
+ "<|13.56|>": 51042,
281
+ "<|13.58|>": 51043,
282
+ "<|13.60|>": 51044,
283
+ "<|13.62|>": 51045,
284
+ "<|13.64|>": 51046,
285
+ "<|13.66|>": 51047,
286
+ "<|13.68|>": 51048,
287
+ "<|13.70|>": 51049,
288
+ "<|13.72|>": 51050,
289
+ "<|13.74|>": 51051,
290
+ "<|13.76|>": 51052,
291
+ "<|13.78|>": 51053,
292
+ "<|13.80|>": 51054,
293
+ "<|13.82|>": 51055,
294
+ "<|13.84|>": 51056,
295
+ "<|13.86|>": 51057,
296
+ "<|13.88|>": 51058,
297
+ "<|13.90|>": 51059,
298
+ "<|13.92|>": 51060,
299
+ "<|13.94|>": 51061,
300
+ "<|13.96|>": 51062,
301
+ "<|13.98|>": 51063,
302
+ "<|14.00|>": 51064,
303
+ "<|14.02|>": 51065,
304
+ "<|14.04|>": 51066,
305
+ "<|14.06|>": 51067,
306
+ "<|14.08|>": 51068,
307
+ "<|14.10|>": 51069,
308
+ "<|14.12|>": 51070,
309
+ "<|14.14|>": 51071,
310
+ "<|14.16|>": 51072,
311
+ "<|14.18|>": 51073,
312
+ "<|14.20|>": 51074,
313
+ "<|14.22|>": 51075,
314
+ "<|14.24|>": 51076,
315
+ "<|14.26|>": 51077,
316
+ "<|14.28|>": 51078,
317
+ "<|14.30|>": 51079,
318
+ "<|14.32|>": 51080,
319
+ "<|14.34|>": 51081,
320
+ "<|14.36|>": 51082,
321
+ "<|14.38|>": 51083,
322
+ "<|14.40|>": 51084,
323
+ "<|14.42|>": 51085,
324
+ "<|14.44|>": 51086,
325
+ "<|14.46|>": 51087,
326
+ "<|14.48|>": 51088,
327
+ "<|14.50|>": 51089,
328
+ "<|14.52|>": 51090,
329
+ "<|14.54|>": 51091,
330
+ "<|14.56|>": 51092,
331
+ "<|14.58|>": 51093,
332
+ "<|14.60|>": 51094,
333
+ "<|14.62|>": 51095,
334
+ "<|14.64|>": 51096,
335
+ "<|14.66|>": 51097,
336
+ "<|14.68|>": 51098,
337
+ "<|14.70|>": 51099,
338
+ "<|14.72|>": 51100,
339
+ "<|14.74|>": 51101,
340
+ "<|14.76|>": 51102,
341
+ "<|14.78|>": 51103,
342
+ "<|14.80|>": 51104,
343
+ "<|14.82|>": 51105,
344
+ "<|14.84|>": 51106,
345
+ "<|14.86|>": 51107,
346
+ "<|14.88|>": 51108,
347
+ "<|14.90|>": 51109,
348
+ "<|14.92|>": 51110,
349
+ "<|14.94|>": 51111,
350
+ "<|14.96|>": 51112,
351
+ "<|14.98|>": 51113,
352
+ "<|15.00|>": 51114,
353
+ "<|15.02|>": 51115,
354
+ "<|15.04|>": 51116,
355
+ "<|15.06|>": 51117,
356
+ "<|15.08|>": 51118,
357
+ "<|15.10|>": 51119,
358
+ "<|15.12|>": 51120,
359
+ "<|15.14|>": 51121,
360
+ "<|15.16|>": 51122,
361
+ "<|15.18|>": 51123,
362
+ "<|15.20|>": 51124,
363
+ "<|15.22|>": 51125,
364
+ "<|15.24|>": 51126,
365
+ "<|15.26|>": 51127,
366
+ "<|15.28|>": 51128,
367
+ "<|15.30|>": 51129,
368
+ "<|15.32|>": 51130,
369
+ "<|15.34|>": 51131,
370
+ "<|15.36|>": 51132,
371
+ "<|15.38|>": 51133,
372
+ "<|15.40|>": 51134,
373
+ "<|15.42|>": 51135,
374
+ "<|15.44|>": 51136,
375
+ "<|15.46|>": 51137,
376
+ "<|15.48|>": 51138,
377
+ "<|15.50|>": 51139,
378
+ "<|15.52|>": 51140,
379
+ "<|15.54|>": 51141,
380
+ "<|15.56|>": 51142,
381
+ "<|15.58|>": 51143,
382
+ "<|15.60|>": 51144,
383
+ "<|15.62|>": 51145,
384
+ "<|15.64|>": 51146,
385
+ "<|15.66|>": 51147,
386
+ "<|15.68|>": 51148,
387
+ "<|15.70|>": 51149,
388
+ "<|15.72|>": 51150,
389
+ "<|15.74|>": 51151,
390
+ "<|15.76|>": 51152,
391
+ "<|15.78|>": 51153,
392
+ "<|15.80|>": 51154,
393
+ "<|15.82|>": 51155,
394
+ "<|15.84|>": 51156,
395
+ "<|15.86|>": 51157,
396
+ "<|15.88|>": 51158,
397
+ "<|15.90|>": 51159,
398
+ "<|15.92|>": 51160,
399
+ "<|15.94|>": 51161,
400
+ "<|15.96|>": 51162,
401
+ "<|15.98|>": 51163,
402
+ "<|16.00|>": 51164,
403
+ "<|16.02|>": 51165,
404
+ "<|16.04|>": 51166,
405
+ "<|16.06|>": 51167,
406
+ "<|16.08|>": 51168,
407
+ "<|16.10|>": 51169,
408
+ "<|16.12|>": 51170,
409
+ "<|16.14|>": 51171,
410
+ "<|16.16|>": 51172,
411
+ "<|16.18|>": 51173,
412
+ "<|16.20|>": 51174,
413
+ "<|16.22|>": 51175,
414
+ "<|16.24|>": 51176,
415
+ "<|16.26|>": 51177,
416
+ "<|16.28|>": 51178,
417
+ "<|16.30|>": 51179,
418
+ "<|16.32|>": 51180,
419
+ "<|16.34|>": 51181,
420
+ "<|16.36|>": 51182,
421
+ "<|16.38|>": 51183,
422
+ "<|16.40|>": 51184,
423
+ "<|16.42|>": 51185,
424
+ "<|16.44|>": 51186,
425
+ "<|16.46|>": 51187,
426
+ "<|16.48|>": 51188,
427
+ "<|16.50|>": 51189,
428
+ "<|16.52|>": 51190,
429
+ "<|16.54|>": 51191,
430
+ "<|16.56|>": 51192,
431
+ "<|16.58|>": 51193,
432
+ "<|16.60|>": 51194,
433
+ "<|16.62|>": 51195,
434
+ "<|16.64|>": 51196,
435
+ "<|16.66|>": 51197,
436
+ "<|16.68|>": 51198,
437
+ "<|16.70|>": 51199,
438
+ "<|16.72|>": 51200,
439
+ "<|16.74|>": 51201,
440
+ "<|16.76|>": 51202,
441
+ "<|16.78|>": 51203,
442
+ "<|16.80|>": 51204,
443
+ "<|16.82|>": 51205,
444
+ "<|16.84|>": 51206,
445
+ "<|16.86|>": 51207,
446
+ "<|16.88|>": 51208,
447
+ "<|16.90|>": 51209,
448
+ "<|16.92|>": 51210,
449
+ "<|16.94|>": 51211,
450
+ "<|16.96|>": 51212,
451
+ "<|16.98|>": 51213,
452
+ "<|17.00|>": 51214,
453
+ "<|17.02|>": 51215,
454
+ "<|17.04|>": 51216,
455
+ "<|17.06|>": 51217,
456
+ "<|17.08|>": 51218,
457
+ "<|17.10|>": 51219,
458
+ "<|17.12|>": 51220,
459
+ "<|17.14|>": 51221,
460
+ "<|17.16|>": 51222,
461
+ "<|17.18|>": 51223,
462
+ "<|17.20|>": 51224,
463
+ "<|17.22|>": 51225,
464
+ "<|17.24|>": 51226,
465
+ "<|17.26|>": 51227,
466
+ "<|17.28|>": 51228,
467
+ "<|17.30|>": 51229,
468
+ "<|17.32|>": 51230,
469
+ "<|17.34|>": 51231,
470
+ "<|17.36|>": 51232,
471
+ "<|17.38|>": 51233,
472
+ "<|17.40|>": 51234,
473
+ "<|17.42|>": 51235,
474
+ "<|17.44|>": 51236,
475
+ "<|17.46|>": 51237,
476
+ "<|17.48|>": 51238,
477
+ "<|17.50|>": 51239,
478
+ "<|17.52|>": 51240,
479
+ "<|17.54|>": 51241,
480
+ "<|17.56|>": 51242,
481
+ "<|17.58|>": 51243,
482
+ "<|17.60|>": 51244,
483
+ "<|17.62|>": 51245,
484
+ "<|17.64|>": 51246,
485
+ "<|17.66|>": 51247,
486
+ "<|17.68|>": 51248,
487
+ "<|17.70|>": 51249,
488
+ "<|17.72|>": 51250,
489
+ "<|17.74|>": 51251,
490
+ "<|17.76|>": 51252,
491
+ "<|17.78|>": 51253,
492
+ "<|17.80|>": 51254,
493
+ "<|17.82|>": 51255,
494
+ "<|17.84|>": 51256,
495
+ "<|17.86|>": 51257,
496
+ "<|17.88|>": 51258,
497
+ "<|17.90|>": 51259,
498
+ "<|17.92|>": 51260,
499
+ "<|17.94|>": 51261,
500
+ "<|17.96|>": 51262,
501
+ "<|17.98|>": 51263,
502
+ "<|18.00|>": 51264,
503
+ "<|18.02|>": 51265,
504
+ "<|18.04|>": 51266,
505
+ "<|18.06|>": 51267,
506
+ "<|18.08|>": 51268,
507
+ "<|18.10|>": 51269,
508
+ "<|18.12|>": 51270,
509
+ "<|18.14|>": 51271,
510
+ "<|18.16|>": 51272,
511
+ "<|18.18|>": 51273,
512
+ "<|18.20|>": 51274,
513
+ "<|18.22|>": 51275,
514
+ "<|18.24|>": 51276,
515
+ "<|18.26|>": 51277,
516
+ "<|18.28|>": 51278,
517
+ "<|18.30|>": 51279,
518
+ "<|18.32|>": 51280,
519
+ "<|18.34|>": 51281,
520
+ "<|18.36|>": 51282,
521
+ "<|18.38|>": 51283,
522
+ "<|18.40|>": 51284,
523
+ "<|18.42|>": 51285,
524
+ "<|18.44|>": 51286,
525
+ "<|18.46|>": 51287,
526
+ "<|18.48|>": 51288,
527
+ "<|18.50|>": 51289,
528
+ "<|18.52|>": 51290,
529
+ "<|18.54|>": 51291,
530
+ "<|18.56|>": 51292,
531
+ "<|18.58|>": 51293,
532
+ "<|18.60|>": 51294,
533
+ "<|18.62|>": 51295,
534
+ "<|18.64|>": 51296,
535
+ "<|18.66|>": 51297,
536
+ "<|18.68|>": 51298,
537
+ "<|18.70|>": 51299,
538
+ "<|18.72|>": 51300,
539
+ "<|18.74|>": 51301,
540
+ "<|18.76|>": 51302,
541
+ "<|18.78|>": 51303,
542
+ "<|18.80|>": 51304,
543
+ "<|18.82|>": 51305,
544
+ "<|18.84|>": 51306,
545
+ "<|18.86|>": 51307,
546
+ "<|18.88|>": 51308,
547
+ "<|18.90|>": 51309,
548
+ "<|18.92|>": 51310,
549
+ "<|18.94|>": 51311,
550
+ "<|18.96|>": 51312,
551
+ "<|18.98|>": 51313,
552
+ "<|19.00|>": 51314,
553
+ "<|19.02|>": 51315,
554
+ "<|19.04|>": 51316,
555
+ "<|19.06|>": 51317,
556
+ "<|19.08|>": 51318,
557
+ "<|19.10|>": 51319,
558
+ "<|19.12|>": 51320,
559
+ "<|19.14|>": 51321,
560
+ "<|19.16|>": 51322,
561
+ "<|19.18|>": 51323,
562
+ "<|19.20|>": 51324,
563
+ "<|19.22|>": 51325,
564
+ "<|19.24|>": 51326,
565
+ "<|19.26|>": 51327,
566
+ "<|19.28|>": 51328,
567
+ "<|19.30|>": 51329,
568
+ "<|19.32|>": 51330,
569
+ "<|19.34|>": 51331,
570
+ "<|19.36|>": 51332,
571
+ "<|19.38|>": 51333,
572
+ "<|19.40|>": 51334,
573
+ "<|19.42|>": 51335,
574
+ "<|19.44|>": 51336,
575
+ "<|19.46|>": 51337,
576
+ "<|19.48|>": 51338,
577
+ "<|19.50|>": 51339,
578
+ "<|19.52|>": 51340,
579
+ "<|19.54|>": 51341,
580
+ "<|19.56|>": 51342,
581
+ "<|19.58|>": 51343,
582
+ "<|19.60|>": 51344,
583
+ "<|19.62|>": 51345,
584
+ "<|19.64|>": 51346,
585
+ "<|19.66|>": 51347,
586
+ "<|19.68|>": 51348,
587
+ "<|19.70|>": 51349,
588
+ "<|19.72|>": 51350,
589
+ "<|19.74|>": 51351,
590
+ "<|19.76|>": 51352,
591
+ "<|19.78|>": 51353,
592
+ "<|19.80|>": 51354,
593
+ "<|19.82|>": 51355,
594
+ "<|19.84|>": 51356,
595
+ "<|19.86|>": 51357,
596
+ "<|19.88|>": 51358,
597
+ "<|19.90|>": 51359,
598
+ "<|19.92|>": 51360,
599
+ "<|19.94|>": 51361,
600
+ "<|19.96|>": 51362,
601
+ "<|19.98|>": 51363,
602
+ "<|2.00|>": 50464,
603
+ "<|2.02|>": 50465,
604
+ "<|2.04|>": 50466,
605
+ "<|2.06|>": 50467,
606
+ "<|2.08|>": 50468,
607
+ "<|2.10|>": 50469,
608
+ "<|2.12|>": 50470,
609
+ "<|2.14|>": 50471,
610
+ "<|2.16|>": 50472,
611
+ "<|2.18|>": 50473,
612
+ "<|2.20|>": 50474,
613
+ "<|2.22|>": 50475,
614
+ "<|2.24|>": 50476,
615
+ "<|2.26|>": 50477,
616
+ "<|2.28|>": 50478,
617
+ "<|2.30|>": 50479,
618
+ "<|2.32|>": 50480,
619
+ "<|2.34|>": 50481,
620
+ "<|2.36|>": 50482,
621
+ "<|2.38|>": 50483,
622
+ "<|2.40|>": 50484,
623
+ "<|2.42|>": 50485,
624
+ "<|2.44|>": 50486,
625
+ "<|2.46|>": 50487,
626
+ "<|2.48|>": 50488,
627
+ "<|2.50|>": 50489,
628
+ "<|2.52|>": 50490,
629
+ "<|2.54|>": 50491,
630
+ "<|2.56|>": 50492,
631
+ "<|2.58|>": 50493,
632
+ "<|2.60|>": 50494,
633
+ "<|2.62|>": 50495,
634
+ "<|2.64|>": 50496,
635
+ "<|2.66|>": 50497,
636
+ "<|2.68|>": 50498,
637
+ "<|2.70|>": 50499,
638
+ "<|2.72|>": 50500,
639
+ "<|2.74|>": 50501,
640
+ "<|2.76|>": 50502,
641
+ "<|2.78|>": 50503,
642
+ "<|2.80|>": 50504,
643
+ "<|2.82|>": 50505,
644
+ "<|2.84|>": 50506,
645
+ "<|2.86|>": 50507,
646
+ "<|2.88|>": 50508,
647
+ "<|2.90|>": 50509,
648
+ "<|2.92|>": 50510,
649
+ "<|2.94|>": 50511,
650
+ "<|2.96|>": 50512,
651
+ "<|2.98|>": 50513,
652
+ "<|20.00|>": 51364,
653
+ "<|20.02|>": 51365,
654
+ "<|20.04|>": 51366,
655
+ "<|20.06|>": 51367,
656
+ "<|20.08|>": 51368,
657
+ "<|20.10|>": 51369,
658
+ "<|20.12|>": 51370,
659
+ "<|20.14|>": 51371,
660
+ "<|20.16|>": 51372,
661
+ "<|20.18|>": 51373,
662
+ "<|20.20|>": 51374,
663
+ "<|20.22|>": 51375,
664
+ "<|20.24|>": 51376,
665
+ "<|20.26|>": 51377,
666
+ "<|20.28|>": 51378,
667
+ "<|20.30|>": 51379,
668
+ "<|20.32|>": 51380,
669
+ "<|20.34|>": 51381,
670
+ "<|20.36|>": 51382,
671
+ "<|20.38|>": 51383,
672
+ "<|20.40|>": 51384,
673
+ "<|20.42|>": 51385,
674
+ "<|20.44|>": 51386,
675
+ "<|20.46|>": 51387,
676
+ "<|20.48|>": 51388,
677
+ "<|20.50|>": 51389,
678
+ "<|20.52|>": 51390,
679
+ "<|20.54|>": 51391,
680
+ "<|20.56|>": 51392,
681
+ "<|20.58|>": 51393,
682
+ "<|20.60|>": 51394,
683
+ "<|20.62|>": 51395,
684
+ "<|20.64|>": 51396,
685
+ "<|20.66|>": 51397,
686
+ "<|20.68|>": 51398,
687
+ "<|20.70|>": 51399,
688
+ "<|20.72|>": 51400,
689
+ "<|20.74|>": 51401,
690
+ "<|20.76|>": 51402,
691
+ "<|20.78|>": 51403,
692
+ "<|20.80|>": 51404,
693
+ "<|20.82|>": 51405,
694
+ "<|20.84|>": 51406,
695
+ "<|20.86|>": 51407,
696
+ "<|20.88|>": 51408,
697
+ "<|20.90|>": 51409,
698
+ "<|20.92|>": 51410,
699
+ "<|20.94|>": 51411,
700
+ "<|20.96|>": 51412,
701
+ "<|20.98|>": 51413,
702
+ "<|21.00|>": 51414,
703
+ "<|21.02|>": 51415,
704
+ "<|21.04|>": 51416,
705
+ "<|21.06|>": 51417,
706
+ "<|21.08|>": 51418,
707
+ "<|21.10|>": 51419,
708
+ "<|21.12|>": 51420,
709
+ "<|21.14|>": 51421,
710
+ "<|21.16|>": 51422,
711
+ "<|21.18|>": 51423,
712
+ "<|21.20|>": 51424,
713
+ "<|21.22|>": 51425,
714
+ "<|21.24|>": 51426,
715
+ "<|21.26|>": 51427,
716
+ "<|21.28|>": 51428,
717
+ "<|21.30|>": 51429,
718
+ "<|21.32|>": 51430,
719
+ "<|21.34|>": 51431,
720
+ "<|21.36|>": 51432,
721
+ "<|21.38|>": 51433,
722
+ "<|21.40|>": 51434,
723
+ "<|21.42|>": 51435,
724
+ "<|21.44|>": 51436,
725
+ "<|21.46|>": 51437,
726
+ "<|21.48|>": 51438,
727
+ "<|21.50|>": 51439,
728
+ "<|21.52|>": 51440,
729
+ "<|21.54|>": 51441,
730
+ "<|21.56|>": 51442,
731
+ "<|21.58|>": 51443,
732
+ "<|21.60|>": 51444,
733
+ "<|21.62|>": 51445,
734
+ "<|21.64|>": 51446,
735
+ "<|21.66|>": 51447,
736
+ "<|21.68|>": 51448,
737
+ "<|21.70|>": 51449,
738
+ "<|21.72|>": 51450,
739
+ "<|21.74|>": 51451,
740
+ "<|21.76|>": 51452,
741
+ "<|21.78|>": 51453,
742
+ "<|21.80|>": 51454,
743
+ "<|21.82|>": 51455,
744
+ "<|21.84|>": 51456,
745
+ "<|21.86|>": 51457,
746
+ "<|21.88|>": 51458,
747
+ "<|21.90|>": 51459,
748
+ "<|21.92|>": 51460,
749
+ "<|21.94|>": 51461,
750
+ "<|21.96|>": 51462,
751
+ "<|21.98|>": 51463,
752
+ "<|22.00|>": 51464,
753
+ "<|22.02|>": 51465,
754
+ "<|22.04|>": 51466,
755
+ "<|22.06|>": 51467,
756
+ "<|22.08|>": 51468,
757
+ "<|22.10|>": 51469,
758
+ "<|22.12|>": 51470,
759
+ "<|22.14|>": 51471,
760
+ "<|22.16|>": 51472,
761
+ "<|22.18|>": 51473,
762
+ "<|22.20|>": 51474,
763
+ "<|22.22|>": 51475,
764
+ "<|22.24|>": 51476,
765
+ "<|22.26|>": 51477,
766
+ "<|22.28|>": 51478,
767
+ "<|22.30|>": 51479,
768
+ "<|22.32|>": 51480,
769
+ "<|22.34|>": 51481,
770
+ "<|22.36|>": 51482,
771
+ "<|22.38|>": 51483,
772
+ "<|22.40|>": 51484,
773
+ "<|22.42|>": 51485,
774
+ "<|22.44|>": 51486,
775
+ "<|22.46|>": 51487,
776
+ "<|22.48|>": 51488,
777
+ "<|22.50|>": 51489,
778
+ "<|22.52|>": 51490,
779
+ "<|22.54|>": 51491,
780
+ "<|22.56|>": 51492,
781
+ "<|22.58|>": 51493,
782
+ "<|22.60|>": 51494,
783
+ "<|22.62|>": 51495,
784
+ "<|22.64|>": 51496,
785
+ "<|22.66|>": 51497,
786
+ "<|22.68|>": 51498,
787
+ "<|22.70|>": 51499,
788
+ "<|22.72|>": 51500,
789
+ "<|22.74|>": 51501,
790
+ "<|22.76|>": 51502,
791
+ "<|22.78|>": 51503,
792
+ "<|22.80|>": 51504,
793
+ "<|22.82|>": 51505,
794
+ "<|22.84|>": 51506,
795
+ "<|22.86|>": 51507,
796
+ "<|22.88|>": 51508,
797
+ "<|22.90|>": 51509,
798
+ "<|22.92|>": 51510,
799
+ "<|22.94|>": 51511,
800
+ "<|22.96|>": 51512,
801
+ "<|22.98|>": 51513,
802
+ "<|23.00|>": 51514,
803
+ "<|23.02|>": 51515,
804
+ "<|23.04|>": 51516,
805
+ "<|23.06|>": 51517,
806
+ "<|23.08|>": 51518,
807
+ "<|23.10|>": 51519,
808
+ "<|23.12|>": 51520,
809
+ "<|23.14|>": 51521,
810
+ "<|23.16|>": 51522,
811
+ "<|23.18|>": 51523,
812
+ "<|23.20|>": 51524,
813
+ "<|23.22|>": 51525,
814
+ "<|23.24|>": 51526,
815
+ "<|23.26|>": 51527,
816
+ "<|23.28|>": 51528,
817
+ "<|23.30|>": 51529,
818
+ "<|23.32|>": 51530,
819
+ "<|23.34|>": 51531,
820
+ "<|23.36|>": 51532,
821
+ "<|23.38|>": 51533,
822
+ "<|23.40|>": 51534,
823
+ "<|23.42|>": 51535,
824
+ "<|23.44|>": 51536,
825
+ "<|23.46|>": 51537,
826
+ "<|23.48|>": 51538,
827
+ "<|23.50|>": 51539,
828
+ "<|23.52|>": 51540,
829
+ "<|23.54|>": 51541,
830
+ "<|23.56|>": 51542,
831
+ "<|23.58|>": 51543,
832
+ "<|23.60|>": 51544,
833
+ "<|23.62|>": 51545,
834
+ "<|23.64|>": 51546,
835
+ "<|23.66|>": 51547,
836
+ "<|23.68|>": 51548,
837
+ "<|23.70|>": 51549,
838
+ "<|23.72|>": 51550,
839
+ "<|23.74|>": 51551,
840
+ "<|23.76|>": 51552,
841
+ "<|23.78|>": 51553,
842
+ "<|23.80|>": 51554,
843
+ "<|23.82|>": 51555,
844
+ "<|23.84|>": 51556,
845
+ "<|23.86|>": 51557,
846
+ "<|23.88|>": 51558,
847
+ "<|23.90|>": 51559,
848
+ "<|23.92|>": 51560,
849
+ "<|23.94|>": 51561,
850
+ "<|23.96|>": 51562,
851
+ "<|23.98|>": 51563,
852
+ "<|24.00|>": 51564,
853
+ "<|24.02|>": 51565,
854
+ "<|24.04|>": 51566,
855
+ "<|24.06|>": 51567,
856
+ "<|24.08|>": 51568,
857
+ "<|24.10|>": 51569,
858
+ "<|24.12|>": 51570,
859
+ "<|24.14|>": 51571,
860
+ "<|24.16|>": 51572,
861
+ "<|24.18|>": 51573,
862
+ "<|24.20|>": 51574,
863
+ "<|24.22|>": 51575,
864
+ "<|24.24|>": 51576,
865
+ "<|24.26|>": 51577,
866
+ "<|24.28|>": 51578,
867
+ "<|24.30|>": 51579,
868
+ "<|24.32|>": 51580,
869
+ "<|24.34|>": 51581,
870
+ "<|24.36|>": 51582,
871
+ "<|24.38|>": 51583,
872
+ "<|24.40|>": 51584,
873
+ "<|24.42|>": 51585,
874
+ "<|24.44|>": 51586,
875
+ "<|24.46|>": 51587,
876
+ "<|24.48|>": 51588,
877
+ "<|24.50|>": 51589,
878
+ "<|24.52|>": 51590,
879
+ "<|24.54|>": 51591,
880
+ "<|24.56|>": 51592,
881
+ "<|24.58|>": 51593,
882
+ "<|24.60|>": 51594,
883
+ "<|24.62|>": 51595,
884
+ "<|24.64|>": 51596,
885
+ "<|24.66|>": 51597,
886
+ "<|24.68|>": 51598,
887
+ "<|24.70|>": 51599,
888
+ "<|24.72|>": 51600,
889
+ "<|24.74|>": 51601,
890
+ "<|24.76|>": 51602,
891
+ "<|24.78|>": 51603,
892
+ "<|24.80|>": 51604,
893
+ "<|24.82|>": 51605,
894
+ "<|24.84|>": 51606,
895
+ "<|24.86|>": 51607,
896
+ "<|24.88|>": 51608,
897
+ "<|24.90|>": 51609,
898
+ "<|24.92|>": 51610,
899
+ "<|24.94|>": 51611,
900
+ "<|24.96|>": 51612,
901
+ "<|24.98|>": 51613,
902
+ "<|25.00|>": 51614,
903
+ "<|25.02|>": 51615,
904
+ "<|25.04|>": 51616,
905
+ "<|25.06|>": 51617,
906
+ "<|25.08|>": 51618,
907
+ "<|25.10|>": 51619,
908
+ "<|25.12|>": 51620,
909
+ "<|25.14|>": 51621,
910
+ "<|25.16|>": 51622,
911
+ "<|25.18|>": 51623,
912
+ "<|25.20|>": 51624,
913
+ "<|25.22|>": 51625,
914
+ "<|25.24|>": 51626,
915
+ "<|25.26|>": 51627,
916
+ "<|25.28|>": 51628,
917
+ "<|25.30|>": 51629,
918
+ "<|25.32|>": 51630,
919
+ "<|25.34|>": 51631,
920
+ "<|25.36|>": 51632,
921
+ "<|25.38|>": 51633,
922
+ "<|25.40|>": 51634,
923
+ "<|25.42|>": 51635,
924
+ "<|25.44|>": 51636,
925
+ "<|25.46|>": 51637,
926
+ "<|25.48|>": 51638,
927
+ "<|25.50|>": 51639,
928
+ "<|25.52|>": 51640,
929
+ "<|25.54|>": 51641,
930
+ "<|25.56|>": 51642,
931
+ "<|25.58|>": 51643,
932
+ "<|25.60|>": 51644,
933
+ "<|25.62|>": 51645,
934
+ "<|25.64|>": 51646,
935
+ "<|25.66|>": 51647,
936
+ "<|25.68|>": 51648,
937
+ "<|25.70|>": 51649,
938
+ "<|25.72|>": 51650,
939
+ "<|25.74|>": 51651,
940
+ "<|25.76|>": 51652,
941
+ "<|25.78|>": 51653,
942
+ "<|25.80|>": 51654,
943
+ "<|25.82|>": 51655,
944
+ "<|25.84|>": 51656,
945
+ "<|25.86|>": 51657,
946
+ "<|25.88|>": 51658,
947
+ "<|25.90|>": 51659,
948
+ "<|25.92|>": 51660,
949
+ "<|25.94|>": 51661,
950
+ "<|25.96|>": 51662,
951
+ "<|25.98|>": 51663,
952
+ "<|26.00|>": 51664,
953
+ "<|26.02|>": 51665,
954
+ "<|26.04|>": 51666,
955
+ "<|26.06|>": 51667,
956
+ "<|26.08|>": 51668,
957
+ "<|26.10|>": 51669,
958
+ "<|26.12|>": 51670,
959
+ "<|26.14|>": 51671,
960
+ "<|26.16|>": 51672,
961
+ "<|26.18|>": 51673,
962
+ "<|26.20|>": 51674,
963
+ "<|26.22|>": 51675,
964
+ "<|26.24|>": 51676,
965
+ "<|26.26|>": 51677,
966
+ "<|26.28|>": 51678,
967
+ "<|26.30|>": 51679,
968
+ "<|26.32|>": 51680,
969
+ "<|26.34|>": 51681,
970
+ "<|26.36|>": 51682,
971
+ "<|26.38|>": 51683,
972
+ "<|26.40|>": 51684,
973
+ "<|26.42|>": 51685,
974
+ "<|26.44|>": 51686,
975
+ "<|26.46|>": 51687,
976
+ "<|26.48|>": 51688,
977
+ "<|26.50|>": 51689,
978
+ "<|26.52|>": 51690,
979
+ "<|26.54|>": 51691,
980
+ "<|26.56|>": 51692,
981
+ "<|26.58|>": 51693,
982
+ "<|26.60|>": 51694,
983
+ "<|26.62|>": 51695,
984
+ "<|26.64|>": 51696,
985
+ "<|26.66|>": 51697,
986
+ "<|26.68|>": 51698,
987
+ "<|26.70|>": 51699,
988
+ "<|26.72|>": 51700,
989
+ "<|26.74|>": 51701,
990
+ "<|26.76|>": 51702,
991
+ "<|26.78|>": 51703,
992
+ "<|26.80|>": 51704,
993
+ "<|26.82|>": 51705,
994
+ "<|26.84|>": 51706,
995
+ "<|26.86|>": 51707,
996
+ "<|26.88|>": 51708,
997
+ "<|26.90|>": 51709,
998
+ "<|26.92|>": 51710,
999
+ "<|26.94|>": 51711,
1000
+ "<|26.96|>": 51712,
1001
+ "<|26.98|>": 51713,
1002
+ "<|27.00|>": 51714,
1003
+ "<|27.02|>": 51715,
1004
+ "<|27.04|>": 51716,
1005
+ "<|27.06|>": 51717,
1006
+ "<|27.08|>": 51718,
1007
+ "<|27.10|>": 51719,
1008
+ "<|27.12|>": 51720,
1009
+ "<|27.14|>": 51721,
1010
+ "<|27.16|>": 51722,
1011
+ "<|27.18|>": 51723,
1012
+ "<|27.20|>": 51724,
1013
+ "<|27.22|>": 51725,
1014
+ "<|27.24|>": 51726,
1015
+ "<|27.26|>": 51727,
1016
+ "<|27.28|>": 51728,
1017
+ "<|27.30|>": 51729,
1018
+ "<|27.32|>": 51730,
1019
+ "<|27.34|>": 51731,
1020
+ "<|27.36|>": 51732,
1021
+ "<|27.38|>": 51733,
1022
+ "<|27.40|>": 51734,
1023
+ "<|27.42|>": 51735,
1024
+ "<|27.44|>": 51736,
1025
+ "<|27.46|>": 51737,
1026
+ "<|27.48|>": 51738,
1027
+ "<|27.50|>": 51739,
1028
+ "<|27.52|>": 51740,
1029
+ "<|27.54|>": 51741,
1030
+ "<|27.56|>": 51742,
1031
+ "<|27.58|>": 51743,
1032
+ "<|27.60|>": 51744,
1033
+ "<|27.62|>": 51745,
1034
+ "<|27.64|>": 51746,
1035
+ "<|27.66|>": 51747,
1036
+ "<|27.68|>": 51748,
1037
+ "<|27.70|>": 51749,
1038
+ "<|27.72|>": 51750,
1039
+ "<|27.74|>": 51751,
1040
+ "<|27.76|>": 51752,
1041
+ "<|27.78|>": 51753,
1042
+ "<|27.80|>": 51754,
1043
+ "<|27.82|>": 51755,
1044
+ "<|27.84|>": 51756,
1045
+ "<|27.86|>": 51757,
1046
+ "<|27.88|>": 51758,
1047
+ "<|27.90|>": 51759,
1048
+ "<|27.92|>": 51760,
1049
+ "<|27.94|>": 51761,
1050
+ "<|27.96|>": 51762,
1051
+ "<|27.98|>": 51763,
1052
+ "<|28.00|>": 51764,
1053
+ "<|28.02|>": 51765,
1054
+ "<|28.04|>": 51766,
1055
+ "<|28.06|>": 51767,
1056
+ "<|28.08|>": 51768,
1057
+ "<|28.10|>": 51769,
1058
+ "<|28.12|>": 51770,
1059
+ "<|28.14|>": 51771,
1060
+ "<|28.16|>": 51772,
1061
+ "<|28.18|>": 51773,
1062
+ "<|28.20|>": 51774,
1063
+ "<|28.22|>": 51775,
1064
+ "<|28.24|>": 51776,
1065
+ "<|28.26|>": 51777,
1066
+ "<|28.28|>": 51778,
1067
+ "<|28.30|>": 51779,
1068
+ "<|28.32|>": 51780,
1069
+ "<|28.34|>": 51781,
1070
+ "<|28.36|>": 51782,
1071
+ "<|28.38|>": 51783,
1072
+ "<|28.40|>": 51784,
1073
+ "<|28.42|>": 51785,
1074
+ "<|28.44|>": 51786,
1075
+ "<|28.46|>": 51787,
1076
+ "<|28.48|>": 51788,
1077
+ "<|28.50|>": 51789,
1078
+ "<|28.52|>": 51790,
1079
+ "<|28.54|>": 51791,
1080
+ "<|28.56|>": 51792,
1081
+ "<|28.58|>": 51793,
1082
+ "<|28.60|>": 51794,
1083
+ "<|28.62|>": 51795,
1084
+ "<|28.64|>": 51796,
1085
+ "<|28.66|>": 51797,
1086
+ "<|28.68|>": 51798,
1087
+ "<|28.70|>": 51799,
1088
+ "<|28.72|>": 51800,
1089
+ "<|28.74|>": 51801,
1090
+ "<|28.76|>": 51802,
1091
+ "<|28.78|>": 51803,
1092
+ "<|28.80|>": 51804,
1093
+ "<|28.82|>": 51805,
1094
+ "<|28.84|>": 51806,
1095
+ "<|28.86|>": 51807,
1096
+ "<|28.88|>": 51808,
1097
+ "<|28.90|>": 51809,
1098
+ "<|28.92|>": 51810,
1099
+ "<|28.94|>": 51811,
1100
+ "<|28.96|>": 51812,
1101
+ "<|28.98|>": 51813,
1102
+ "<|29.00|>": 51814,
1103
+ "<|29.02|>": 51815,
1104
+ "<|29.04|>": 51816,
1105
+ "<|29.06|>": 51817,
1106
+ "<|29.08|>": 51818,
1107
+ "<|29.10|>": 51819,
1108
+ "<|29.12|>": 51820,
1109
+ "<|29.14|>": 51821,
1110
+ "<|29.16|>": 51822,
1111
+ "<|29.18|>": 51823,
1112
+ "<|29.20|>": 51824,
1113
+ "<|29.22|>": 51825,
1114
+ "<|29.24|>": 51826,
1115
+ "<|29.26|>": 51827,
1116
+ "<|29.28|>": 51828,
1117
+ "<|29.30|>": 51829,
1118
+ "<|29.32|>": 51830,
1119
+ "<|29.34|>": 51831,
1120
+ "<|29.36|>": 51832,
1121
+ "<|29.38|>": 51833,
1122
+ "<|29.40|>": 51834,
1123
+ "<|29.42|>": 51835,
1124
+ "<|29.44|>": 51836,
1125
+ "<|29.46|>": 51837,
1126
+ "<|29.48|>": 51838,
1127
+ "<|29.50|>": 51839,
1128
+ "<|29.52|>": 51840,
1129
+ "<|29.54|>": 51841,
1130
+ "<|29.56|>": 51842,
1131
+ "<|29.58|>": 51843,
1132
+ "<|29.60|>": 51844,
1133
+ "<|29.62|>": 51845,
1134
+ "<|29.64|>": 51846,
1135
+ "<|29.66|>": 51847,
1136
+ "<|29.68|>": 51848,
1137
+ "<|29.70|>": 51849,
1138
+ "<|29.72|>": 51850,
1139
+ "<|29.74|>": 51851,
1140
+ "<|29.76|>": 51852,
1141
+ "<|29.78|>": 51853,
1142
+ "<|29.80|>": 51854,
1143
+ "<|29.82|>": 51855,
1144
+ "<|29.84|>": 51856,
1145
+ "<|29.86|>": 51857,
1146
+ "<|29.88|>": 51858,
1147
+ "<|29.90|>": 51859,
1148
+ "<|29.92|>": 51860,
1149
+ "<|29.94|>": 51861,
1150
+ "<|29.96|>": 51862,
1151
+ "<|29.98|>": 51863,
1152
+ "<|3.00|>": 50514,
1153
+ "<|3.02|>": 50515,
1154
+ "<|3.04|>": 50516,
1155
+ "<|3.06|>": 50517,
1156
+ "<|3.08|>": 50518,
1157
+ "<|3.10|>": 50519,
1158
+ "<|3.12|>": 50520,
1159
+ "<|3.14|>": 50521,
1160
+ "<|3.16|>": 50522,
1161
+ "<|3.18|>": 50523,
1162
+ "<|3.20|>": 50524,
1163
+ "<|3.22|>": 50525,
1164
+ "<|3.24|>": 50526,
1165
+ "<|3.26|>": 50527,
1166
+ "<|3.28|>": 50528,
1167
+ "<|3.30|>": 50529,
1168
+ "<|3.32|>": 50530,
1169
+ "<|3.34|>": 50531,
1170
+ "<|3.36|>": 50532,
1171
+ "<|3.38|>": 50533,
1172
+ "<|3.40|>": 50534,
1173
+ "<|3.42|>": 50535,
1174
+ "<|3.44|>": 50536,
1175
+ "<|3.46|>": 50537,
1176
+ "<|3.48|>": 50538,
1177
+ "<|3.50|>": 50539,
1178
+ "<|3.52|>": 50540,
1179
+ "<|3.54|>": 50541,
1180
+ "<|3.56|>": 50542,
1181
+ "<|3.58|>": 50543,
1182
+ "<|3.60|>": 50544,
1183
+ "<|3.62|>": 50545,
1184
+ "<|3.64|>": 50546,
1185
+ "<|3.66|>": 50547,
1186
+ "<|3.68|>": 50548,
1187
+ "<|3.70|>": 50549,
1188
+ "<|3.72|>": 50550,
1189
+ "<|3.74|>": 50551,
1190
+ "<|3.76|>": 50552,
1191
+ "<|3.78|>": 50553,
1192
+ "<|3.80|>": 50554,
1193
+ "<|3.82|>": 50555,
1194
+ "<|3.84|>": 50556,
1195
+ "<|3.86|>": 50557,
1196
+ "<|3.88|>": 50558,
1197
+ "<|3.90|>": 50559,
1198
+ "<|3.92|>": 50560,
1199
+ "<|3.94|>": 50561,
1200
+ "<|3.96|>": 50562,
1201
+ "<|3.98|>": 50563,
1202
+ "<|30.00|>": 51864,
1203
+ "<|4.00|>": 50564,
1204
+ "<|4.02|>": 50565,
1205
+ "<|4.04|>": 50566,
1206
+ "<|4.06|>": 50567,
1207
+ "<|4.08|>": 50568,
1208
+ "<|4.10|>": 50569,
1209
+ "<|4.12|>": 50570,
1210
+ "<|4.14|>": 50571,
1211
+ "<|4.16|>": 50572,
1212
+ "<|4.18|>": 50573,
1213
+ "<|4.20|>": 50574,
1214
+ "<|4.22|>": 50575,
1215
+ "<|4.24|>": 50576,
1216
+ "<|4.26|>": 50577,
1217
+ "<|4.28|>": 50578,
1218
+ "<|4.30|>": 50579,
1219
+ "<|4.32|>": 50580,
1220
+ "<|4.34|>": 50581,
1221
+ "<|4.36|>": 50582,
1222
+ "<|4.38|>": 50583,
1223
+ "<|4.40|>": 50584,
1224
+ "<|4.42|>": 50585,
1225
+ "<|4.44|>": 50586,
1226
+ "<|4.46|>": 50587,
1227
+ "<|4.48|>": 50588,
1228
+ "<|4.50|>": 50589,
1229
+ "<|4.52|>": 50590,
1230
+ "<|4.54|>": 50591,
1231
+ "<|4.56|>": 50592,
1232
+ "<|4.58|>": 50593,
1233
+ "<|4.60|>": 50594,
1234
+ "<|4.62|>": 50595,
1235
+ "<|4.64|>": 50596,
1236
+ "<|4.66|>": 50597,
1237
+ "<|4.68|>": 50598,
1238
+ "<|4.70|>": 50599,
1239
+ "<|4.72|>": 50600,
1240
+ "<|4.74|>": 50601,
1241
+ "<|4.76|>": 50602,
1242
+ "<|4.78|>": 50603,
1243
+ "<|4.80|>": 50604,
1244
+ "<|4.82|>": 50605,
1245
+ "<|4.84|>": 50606,
1246
+ "<|4.86|>": 50607,
1247
+ "<|4.88|>": 50608,
1248
+ "<|4.90|>": 50609,
1249
+ "<|4.92|>": 50610,
1250
+ "<|4.94|>": 50611,
1251
+ "<|4.96|>": 50612,
1252
+ "<|4.98|>": 50613,
1253
+ "<|5.00|>": 50614,
1254
+ "<|5.02|>": 50615,
1255
+ "<|5.04|>": 50616,
1256
+ "<|5.06|>": 50617,
1257
+ "<|5.08|>": 50618,
1258
+ "<|5.10|>": 50619,
1259
+ "<|5.12|>": 50620,
1260
+ "<|5.14|>": 50621,
1261
+ "<|5.16|>": 50622,
1262
+ "<|5.18|>": 50623,
1263
+ "<|5.20|>": 50624,
1264
+ "<|5.22|>": 50625,
1265
+ "<|5.24|>": 50626,
1266
+ "<|5.26|>": 50627,
1267
+ "<|5.28|>": 50628,
1268
+ "<|5.30|>": 50629,
1269
+ "<|5.32|>": 50630,
1270
+ "<|5.34|>": 50631,
1271
+ "<|5.36|>": 50632,
1272
+ "<|5.38|>": 50633,
1273
+ "<|5.40|>": 50634,
1274
+ "<|5.42|>": 50635,
1275
+ "<|5.44|>": 50636,
1276
+ "<|5.46|>": 50637,
1277
+ "<|5.48|>": 50638,
1278
+ "<|5.50|>": 50639,
1279
+ "<|5.52|>": 50640,
1280
+ "<|5.54|>": 50641,
1281
+ "<|5.56|>": 50642,
1282
+ "<|5.58|>": 50643,
1283
+ "<|5.60|>": 50644,
1284
+ "<|5.62|>": 50645,
1285
+ "<|5.64|>": 50646,
1286
+ "<|5.66|>": 50647,
1287
+ "<|5.68|>": 50648,
1288
+ "<|5.70|>": 50649,
1289
+ "<|5.72|>": 50650,
1290
+ "<|5.74|>": 50651,
1291
+ "<|5.76|>": 50652,
1292
+ "<|5.78|>": 50653,
1293
+ "<|5.80|>": 50654,
1294
+ "<|5.82|>": 50655,
1295
+ "<|5.84|>": 50656,
1296
+ "<|5.86|>": 50657,
1297
+ "<|5.88|>": 50658,
1298
+ "<|5.90|>": 50659,
1299
+ "<|5.92|>": 50660,
1300
+ "<|5.94|>": 50661,
1301
+ "<|5.96|>": 50662,
1302
+ "<|5.98|>": 50663,
1303
+ "<|6.00|>": 50664,
1304
+ "<|6.02|>": 50665,
1305
+ "<|6.04|>": 50666,
1306
+ "<|6.06|>": 50667,
1307
+ "<|6.08|>": 50668,
1308
+ "<|6.10|>": 50669,
1309
+ "<|6.12|>": 50670,
1310
+ "<|6.14|>": 50671,
1311
+ "<|6.16|>": 50672,
1312
+ "<|6.18|>": 50673,
1313
+ "<|6.20|>": 50674,
1314
+ "<|6.22|>": 50675,
1315
+ "<|6.24|>": 50676,
1316
+ "<|6.26|>": 50677,
1317
+ "<|6.28|>": 50678,
1318
+ "<|6.30|>": 50679,
1319
+ "<|6.32|>": 50680,
1320
+ "<|6.34|>": 50681,
1321
+ "<|6.36|>": 50682,
1322
+ "<|6.38|>": 50683,
1323
+ "<|6.40|>": 50684,
1324
+ "<|6.42|>": 50685,
1325
+ "<|6.44|>": 50686,
1326
+ "<|6.46|>": 50687,
1327
+ "<|6.48|>": 50688,
1328
+ "<|6.50|>": 50689,
1329
+ "<|6.52|>": 50690,
1330
+ "<|6.54|>": 50691,
1331
+ "<|6.56|>": 50692,
1332
+ "<|6.58|>": 50693,
1333
+ "<|6.60|>": 50694,
1334
+ "<|6.62|>": 50695,
1335
+ "<|6.64|>": 50696,
1336
+ "<|6.66|>": 50697,
1337
+ "<|6.68|>": 50698,
1338
+ "<|6.70|>": 50699,
1339
+ "<|6.72|>": 50700,
1340
+ "<|6.74|>": 50701,
1341
+ "<|6.76|>": 50702,
1342
+ "<|6.78|>": 50703,
1343
+ "<|6.80|>": 50704,
1344
+ "<|6.82|>": 50705,
1345
+ "<|6.84|>": 50706,
1346
+ "<|6.86|>": 50707,
1347
+ "<|6.88|>": 50708,
1348
+ "<|6.90|>": 50709,
1349
+ "<|6.92|>": 50710,
1350
+ "<|6.94|>": 50711,
1351
+ "<|6.96|>": 50712,
1352
+ "<|6.98|>": 50713,
1353
+ "<|7.00|>": 50714,
1354
+ "<|7.02|>": 50715,
1355
+ "<|7.04|>": 50716,
1356
+ "<|7.06|>": 50717,
1357
+ "<|7.08|>": 50718,
1358
+ "<|7.10|>": 50719,
1359
+ "<|7.12|>": 50720,
1360
+ "<|7.14|>": 50721,
1361
+ "<|7.16|>": 50722,
1362
+ "<|7.18|>": 50723,
1363
+ "<|7.20|>": 50724,
1364
+ "<|7.22|>": 50725,
1365
+ "<|7.24|>": 50726,
1366
+ "<|7.26|>": 50727,
1367
+ "<|7.28|>": 50728,
1368
+ "<|7.30|>": 50729,
1369
+ "<|7.32|>": 50730,
1370
+ "<|7.34|>": 50731,
1371
+ "<|7.36|>": 50732,
1372
+ "<|7.38|>": 50733,
1373
+ "<|7.40|>": 50734,
1374
+ "<|7.42|>": 50735,
1375
+ "<|7.44|>": 50736,
1376
+ "<|7.46|>": 50737,
1377
+ "<|7.48|>": 50738,
1378
+ "<|7.50|>": 50739,
1379
+ "<|7.52|>": 50740,
1380
+ "<|7.54|>": 50741,
1381
+ "<|7.56|>": 50742,
1382
+ "<|7.58|>": 50743,
1383
+ "<|7.60|>": 50744,
1384
+ "<|7.62|>": 50745,
1385
+ "<|7.64|>": 50746,
1386
+ "<|7.66|>": 50747,
1387
+ "<|7.68|>": 50748,
1388
+ "<|7.70|>": 50749,
1389
+ "<|7.72|>": 50750,
1390
+ "<|7.74|>": 50751,
1391
+ "<|7.76|>": 50752,
1392
+ "<|7.78|>": 50753,
1393
+ "<|7.80|>": 50754,
1394
+ "<|7.82|>": 50755,
1395
+ "<|7.84|>": 50756,
1396
+ "<|7.86|>": 50757,
1397
+ "<|7.88|>": 50758,
1398
+ "<|7.90|>": 50759,
1399
+ "<|7.92|>": 50760,
1400
+ "<|7.94|>": 50761,
1401
+ "<|7.96|>": 50762,
1402
+ "<|7.98|>": 50763,
1403
+ "<|8.00|>": 50764,
1404
+ "<|8.02|>": 50765,
1405
+ "<|8.04|>": 50766,
1406
+ "<|8.06|>": 50767,
1407
+ "<|8.08|>": 50768,
1408
+ "<|8.10|>": 50769,
1409
+ "<|8.12|>": 50770,
1410
+ "<|8.14|>": 50771,
1411
+ "<|8.16|>": 50772,
1412
+ "<|8.18|>": 50773,
1413
+ "<|8.20|>": 50774,
1414
+ "<|8.22|>": 50775,
1415
+ "<|8.24|>": 50776,
1416
+ "<|8.26|>": 50777,
1417
+ "<|8.28|>": 50778,
1418
+ "<|8.30|>": 50779,
1419
+ "<|8.32|>": 50780,
1420
+ "<|8.34|>": 50781,
1421
+ "<|8.36|>": 50782,
1422
+ "<|8.38|>": 50783,
1423
+ "<|8.40|>": 50784,
1424
+ "<|8.42|>": 50785,
1425
+ "<|8.44|>": 50786,
1426
+ "<|8.46|>": 50787,
1427
+ "<|8.48|>": 50788,
1428
+ "<|8.50|>": 50789,
1429
+ "<|8.52|>": 50790,
1430
+ "<|8.54|>": 50791,
1431
+ "<|8.56|>": 50792,
1432
+ "<|8.58|>": 50793,
1433
+ "<|8.60|>": 50794,
1434
+ "<|8.62|>": 50795,
1435
+ "<|8.64|>": 50796,
1436
+ "<|8.66|>": 50797,
1437
+ "<|8.68|>": 50798,
1438
+ "<|8.70|>": 50799,
1439
+ "<|8.72|>": 50800,
1440
+ "<|8.74|>": 50801,
1441
+ "<|8.76|>": 50802,
1442
+ "<|8.78|>": 50803,
1443
+ "<|8.80|>": 50804,
1444
+ "<|8.82|>": 50805,
1445
+ "<|8.84|>": 50806,
1446
+ "<|8.86|>": 50807,
1447
+ "<|8.88|>": 50808,
1448
+ "<|8.90|>": 50809,
1449
+ "<|8.92|>": 50810,
1450
+ "<|8.94|>": 50811,
1451
+ "<|8.96|>": 50812,
1452
+ "<|8.98|>": 50813,
1453
+ "<|9.00|>": 50814,
1454
+ "<|9.02|>": 50815,
1455
+ "<|9.04|>": 50816,
1456
+ "<|9.06|>": 50817,
1457
+ "<|9.08|>": 50818,
1458
+ "<|9.10|>": 50819,
1459
+ "<|9.12|>": 50820,
1460
+ "<|9.14|>": 50821,
1461
+ "<|9.16|>": 50822,
1462
+ "<|9.18|>": 50823,
1463
+ "<|9.20|>": 50824,
1464
+ "<|9.22|>": 50825,
1465
+ "<|9.24|>": 50826,
1466
+ "<|9.26|>": 50827,
1467
+ "<|9.28|>": 50828,
1468
+ "<|9.30|>": 50829,
1469
+ "<|9.32|>": 50830,
1470
+ "<|9.34|>": 50831,
1471
+ "<|9.36|>": 50832,
1472
+ "<|9.38|>": 50833,
1473
+ "<|9.40|>": 50834,
1474
+ "<|9.42|>": 50835,
1475
+ "<|9.44|>": 50836,
1476
+ "<|9.46|>": 50837,
1477
+ "<|9.48|>": 50838,
1478
+ "<|9.50|>": 50839,
1479
+ "<|9.52|>": 50840,
1480
+ "<|9.54|>": 50841,
1481
+ "<|9.56|>": 50842,
1482
+ "<|9.58|>": 50843,
1483
+ "<|9.60|>": 50844,
1484
+ "<|9.62|>": 50845,
1485
+ "<|9.64|>": 50846,
1486
+ "<|9.66|>": 50847,
1487
+ "<|9.68|>": 50848,
1488
+ "<|9.70|>": 50849,
1489
+ "<|9.72|>": 50850,
1490
+ "<|9.74|>": 50851,
1491
+ "<|9.76|>": 50852,
1492
+ "<|9.78|>": 50853,
1493
+ "<|9.80|>": 50854,
1494
+ "<|9.82|>": 50855,
1495
+ "<|9.84|>": 50856,
1496
+ "<|9.86|>": 50857,
1497
+ "<|9.88|>": 50858,
1498
+ "<|9.90|>": 50859,
1499
+ "<|9.92|>": 50860,
1500
+ "<|9.94|>": 50861,
1501
+ "<|9.96|>": 50862,
1502
+ "<|9.98|>": 50863,
1503
+ "<|af|>": 50327,
1504
+ "<|am|>": 50334,
1505
+ "<|ar|>": 50272,
1506
+ "<|as|>": 50350,
1507
+ "<|az|>": 50304,
1508
+ "<|ba|>": 50355,
1509
+ "<|be|>": 50330,
1510
+ "<|bg|>": 50292,
1511
+ "<|bn|>": 50302,
1512
+ "<|bo|>": 50347,
1513
+ "<|br|>": 50309,
1514
+ "<|bs|>": 50315,
1515
+ "<|ca|>": 50270,
1516
+ "<|cs|>": 50283,
1517
+ "<|cy|>": 50297,
1518
+ "<|da|>": 50285,
1519
+ "<|de|>": 50261,
1520
+ "<|el|>": 50281,
1521
+ "<|en|>": 50259,
1522
+ "<|es|>": 50262,
1523
+ "<|et|>": 50307,
1524
+ "<|eu|>": 50310,
1525
+ "<|fa|>": 50300,
1526
+ "<|fi|>": 50277,
1527
+ "<|fo|>": 50338,
1528
+ "<|fr|>": 50265,
1529
+ "<|gl|>": 50319,
1530
+ "<|gu|>": 50333,
1531
+ "<|haw|>": 50352,
1532
+ "<|ha|>": 50354,
1533
+ "<|he|>": 50279,
1534
+ "<|hi|>": 50276,
1535
+ "<|hr|>": 50291,
1536
+ "<|ht|>": 50339,
1537
+ "<|hu|>": 50286,
1538
+ "<|hy|>": 50312,
1539
+ "<|id|>": 50275,
1540
+ "<|is|>": 50311,
1541
+ "<|it|>": 50274,
1542
+ "<|ja|>": 50266,
1543
+ "<|jw|>": 50356,
1544
+ "<|ka|>": 50329,
1545
+ "<|kk|>": 50316,
1546
+ "<|km|>": 50323,
1547
+ "<|kn|>": 50306,
1548
+ "<|ko|>": 50264,
1549
+ "<|la|>": 50294,
1550
+ "<|lb|>": 50345,
1551
+ "<|ln|>": 50353,
1552
+ "<|lo|>": 50336,
1553
+ "<|lt|>": 50293,
1554
+ "<|lv|>": 50301,
1555
+ "<|mg|>": 50349,
1556
+ "<|mi|>": 50295,
1557
+ "<|mk|>": 50308,
1558
+ "<|ml|>": 50296,
1559
+ "<|mn|>": 50314,
1560
+ "<|mr|>": 50320,
1561
+ "<|ms|>": 50282,
1562
+ "<|mt|>": 50343,
1563
+ "<|my|>": 50346,
1564
+ "<|ne|>": 50313,
1565
+ "<|nl|>": 50271,
1566
+ "<|nn|>": 50342,
1567
+ "<|nocaptions|>": 50362,
1568
+ "<|notimestamps|>": 50363,
1569
+ "<|no|>": 50288,
1570
+ "<|oc|>": 50328,
1571
+ "<|pa|>": 50321,
1572
+ "<|pl|>": 50269,
1573
+ "<|ps|>": 50340,
1574
+ "<|pt|>": 50267,
1575
+ "<|ro|>": 50284,
1576
+ "<|ru|>": 50263,
1577
+ "<|sa|>": 50344,
1578
+ "<|sd|>": 50332,
1579
+ "<|si|>": 50322,
1580
+ "<|sk|>": 50298,
1581
+ "<|sl|>": 50305,
1582
+ "<|sn|>": 50324,
1583
+ "<|so|>": 50326,
1584
+ "<|sq|>": 50317,
1585
+ "<|sr|>": 50303,
1586
+ "<|startoflm|>": 50360,
1587
+ "<|startofprev|>": 50361,
1588
+ "<|startoftranscript|>": 50258,
1589
+ "<|su|>": 50357,
1590
+ "<|sv|>": 50273,
1591
+ "<|sw|>": 50318,
1592
+ "<|ta|>": 50287,
1593
+ "<|te|>": 50299,
1594
+ "<|tg|>": 50331,
1595
+ "<|th|>": 50289,
1596
+ "<|tk|>": 50341,
1597
+ "<|tl|>": 50348,
1598
+ "<|transcribe|>": 50359,
1599
+ "<|translate|>": 50358,
1600
+ "<|tr|>": 50268,
1601
+ "<|tt|>": 50351,
1602
+ "<|uk|>": 50280,
1603
+ "<|ur|>": 50290,
1604
+ "<|uz|>": 50337,
1605
+ "<|vi|>": 50278,
1606
+ "<|yi|>": 50335,
1607
+ "<|yo|>": 50325,
1608
+ "<|zh|>": 50260
1609
+ }
checkpoint-10000/config.json ADDED
@@ -0,0 +1,144 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "openai/whisper-medium",
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "gelu",
5
+ "architectures": [
6
+ "WhisperForConditionalGeneration"
7
+ ],
8
+ "attention_dropout": 0.0,
9
+ "begin_suppress_tokens": [
10
+ 220,
11
+ 50257
12
+ ],
13
+ "bos_token_id": 50257,
14
+ "d_model": 1024,
15
+ "decoder_attention_heads": 16,
16
+ "decoder_ffn_dim": 4096,
17
+ "decoder_layerdrop": 0.0,
18
+ "decoder_layers": 24,
19
+ "decoder_start_token_id": 50258,
20
+ "dropout": 0.0,
21
+ "encoder_attention_heads": 16,
22
+ "encoder_ffn_dim": 4096,
23
+ "encoder_layerdrop": 0.0,
24
+ "encoder_layers": 24,
25
+ "eos_token_id": 50257,
26
+ "forced_decoder_ids": [
27
+ [
28
+ 1,
29
+ 50259
30
+ ],
31
+ [
32
+ 2,
33
+ 50359
34
+ ],
35
+ [
36
+ 3,
37
+ 50363
38
+ ]
39
+ ],
40
+ "init_std": 0.02,
41
+ "is_encoder_decoder": true,
42
+ "max_length": 448,
43
+ "max_source_positions": 1500,
44
+ "max_target_positions": 448,
45
+ "model_type": "whisper",
46
+ "num_hidden_layers": 24,
47
+ "num_mel_bins": 80,
48
+ "pad_token_id": 50257,
49
+ "scale_embedding": false,
50
+ "suppress_tokens": [
51
+ 1,
52
+ 2,
53
+ 7,
54
+ 8,
55
+ 9,
56
+ 10,
57
+ 14,
58
+ 25,
59
+ 26,
60
+ 27,
61
+ 28,
62
+ 29,
63
+ 31,
64
+ 58,
65
+ 59,
66
+ 60,
67
+ 61,
68
+ 62,
69
+ 63,
70
+ 90,
71
+ 91,
72
+ 92,
73
+ 93,
74
+ 359,
75
+ 503,
76
+ 522,
77
+ 542,
78
+ 873,
79
+ 893,
80
+ 902,
81
+ 918,
82
+ 922,
83
+ 931,
84
+ 1350,
85
+ 1853,
86
+ 1982,
87
+ 2460,
88
+ 2627,
89
+ 3246,
90
+ 3253,
91
+ 3268,
92
+ 3536,
93
+ 3846,
94
+ 3961,
95
+ 4183,
96
+ 4667,
97
+ 6585,
98
+ 6647,
99
+ 7273,
100
+ 9061,
101
+ 9383,
102
+ 10428,
103
+ 10929,
104
+ 11938,
105
+ 12033,
106
+ 12331,
107
+ 12562,
108
+ 13793,
109
+ 14157,
110
+ 14635,
111
+ 15265,
112
+ 15618,
113
+ 16553,
114
+ 16604,
115
+ 18362,
116
+ 18956,
117
+ 20075,
118
+ 21675,
119
+ 22520,
120
+ 26130,
121
+ 26161,
122
+ 26435,
123
+ 28279,
124
+ 29464,
125
+ 31650,
126
+ 32302,
127
+ 32470,
128
+ 36865,
129
+ 42863,
130
+ 47425,
131
+ 49870,
132
+ 50254,
133
+ 50258,
134
+ 50358,
135
+ 50359,
136
+ 50360,
137
+ 50361,
138
+ 50362
139
+ ],
140
+ "torch_dtype": "float32",
141
+ "transformers_version": "4.26.0",
142
+ "use_cache": true,
143
+ "vocab_size": 51865
144
+ }
checkpoint-10000/generation_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "begin_suppress_tokens": [
3
+ 220,
4
+ 50257
5
+ ],
6
+ "bos_token_id": 50257,
7
+ "decoder_start_token_id": 50258,
8
+ "eos_token_id": 50257,
9
+ "language": "spanish",
10
+ "max_length": 448,
11
+ "pad_token_id": 50257,
12
+ "suppress_tokens": [
13
+ 1,
14
+ 2,
15
+ 7,
16
+ 8,
17
+ 9,
18
+ 10,
19
+ 14,
20
+ 25,
21
+ 26,
22
+ 27,
23
+ 28,
24
+ 29,
25
+ 31,
26
+ 58,
27
+ 59,
28
+ 60,
29
+ 61,
30
+ 62,
31
+ 63,
32
+ 90,
33
+ 91,
34
+ 92,
35
+ 93,
36
+ 359,
37
+ 503,
38
+ 522,
39
+ 542,
40
+ 873,
41
+ 893,
42
+ 902,
43
+ 918,
44
+ 922,
45
+ 931,
46
+ 1350,
47
+ 1853,
48
+ 1982,
49
+ 2460,
50
+ 2627,
51
+ 3246,
52
+ 3253,
53
+ 3268,
54
+ 3536,
55
+ 3846,
56
+ 3961,
57
+ 4183,
58
+ 4667,
59
+ 6585,
60
+ 6647,
61
+ 7273,
62
+ 9061,
63
+ 9383,
64
+ 10428,
65
+ 10929,
66
+ 11938,
67
+ 12033,
68
+ 12331,
69
+ 12562,
70
+ 13793,
71
+ 14157,
72
+ 14635,
73
+ 15265,
74
+ 15618,
75
+ 16553,
76
+ 16604,
77
+ 18362,
78
+ 18956,
79
+ 20075,
80
+ 21675,
81
+ 22520,
82
+ 26130,
83
+ 26161,
84
+ 26435,
85
+ 28279,
86
+ 29464,
87
+ 31650,
88
+ 32302,
89
+ 32470,
90
+ 36865,
91
+ 42863,
92
+ 47425,
93
+ 49870,
94
+ 50254,
95
+ 50258,
96
+ 50358,
97
+ 50359,
98
+ 50360,
99
+ 50361,
100
+ 50362
101
+ ],
102
+ "task": "transcribe",
103
+ "transformers_version": "4.26.0"
104
+ }
checkpoint-10000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d102766e9f60f1cdcaad004c6a6d63c9151f16fe590af601c63f2ee50587bf4b
3
+ size 6111429144
checkpoint-10000/preprocessor_config.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-10000/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59d3e042fbb7ce3f511b23f3c8be25cb40e21af62393b2dd1c659d1685076f24
3
+ size 3055755286
checkpoint-10000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93ac3ee80ac18a266708e8ffc87605d3d369b472d57644e11b7afb17624838f3
3
+ size 14244
checkpoint-10000/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad331e3f2a5f55a3f57ef86339a7ddfcd22bbf4b538f2edade9ef66a9405a5c6
3
+ size 988
checkpoint-10000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ca0253c432c1c934329a180a1fa886f435c1635d2c93793ea5c7b353241646f
3
+ size 1064
checkpoint-10000/trainer_state.json ADDED
@@ -0,0 +1,2641 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 93.07875894988067,
3
+ "best_model_checkpoint": "ASFEMA_SEV/checkpoint-6000",
4
+ "epoch": 129.86173633440515,
5
+ "global_step": 10000,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.32,
12
+ "learning_rate": 4.6000000000000004e-07,
13
+ "loss": 4.0894,
14
+ "step": 25
15
+ },
16
+ {
17
+ "epoch": 0.64,
18
+ "learning_rate": 9.600000000000001e-07,
19
+ "loss": 3.7,
20
+ "step": 50
21
+ },
22
+ {
23
+ "epoch": 0.96,
24
+ "learning_rate": 1.46e-06,
25
+ "loss": 3.0759,
26
+ "step": 75
27
+ },
28
+ {
29
+ "epoch": 1.3,
30
+ "learning_rate": 1.9600000000000003e-06,
31
+ "loss": 2.5729,
32
+ "step": 100
33
+ },
34
+ {
35
+ "epoch": 1.62,
36
+ "learning_rate": 2.46e-06,
37
+ "loss": 2.1885,
38
+ "step": 125
39
+ },
40
+ {
41
+ "epoch": 1.94,
42
+ "learning_rate": 2.96e-06,
43
+ "loss": 1.9431,
44
+ "step": 150
45
+ },
46
+ {
47
+ "epoch": 2.27,
48
+ "learning_rate": 3.46e-06,
49
+ "loss": 1.5753,
50
+ "step": 175
51
+ },
52
+ {
53
+ "epoch": 2.59,
54
+ "learning_rate": 3.96e-06,
55
+ "loss": 1.3319,
56
+ "step": 200
57
+ },
58
+ {
59
+ "epoch": 2.91,
60
+ "learning_rate": 4.4600000000000005e-06,
61
+ "loss": 1.2076,
62
+ "step": 225
63
+ },
64
+ {
65
+ "epoch": 3.24,
66
+ "learning_rate": 4.960000000000001e-06,
67
+ "loss": 0.9211,
68
+ "step": 250
69
+ },
70
+ {
71
+ "epoch": 3.57,
72
+ "learning_rate": 5.460000000000001e-06,
73
+ "loss": 0.7332,
74
+ "step": 275
75
+ },
76
+ {
77
+ "epoch": 3.89,
78
+ "learning_rate": 5.9600000000000005e-06,
79
+ "loss": 0.6601,
80
+ "step": 300
81
+ },
82
+ {
83
+ "epoch": 4.22,
84
+ "learning_rate": 6.460000000000001e-06,
85
+ "loss": 0.4824,
86
+ "step": 325
87
+ },
88
+ {
89
+ "epoch": 4.54,
90
+ "learning_rate": 6.96e-06,
91
+ "loss": 0.3455,
92
+ "step": 350
93
+ },
94
+ {
95
+ "epoch": 4.86,
96
+ "learning_rate": 7.4600000000000006e-06,
97
+ "loss": 0.3741,
98
+ "step": 375
99
+ },
100
+ {
101
+ "epoch": 5.19,
102
+ "learning_rate": 7.960000000000002e-06,
103
+ "loss": 0.2448,
104
+ "step": 400
105
+ },
106
+ {
107
+ "epoch": 5.19,
108
+ "eval_loss": 0.9975764751434326,
109
+ "eval_runtime": 24.147,
110
+ "eval_samples_per_second": 5.756,
111
+ "eval_steps_per_second": 0.745,
112
+ "eval_wer": 132.57756563245823,
113
+ "step": 400
114
+ },
115
+ {
116
+ "epoch": 5.51,
117
+ "learning_rate": 8.46e-06,
118
+ "loss": 0.1713,
119
+ "step": 425
120
+ },
121
+ {
122
+ "epoch": 5.84,
123
+ "learning_rate": 8.96e-06,
124
+ "loss": 0.1906,
125
+ "step": 450
126
+ },
127
+ {
128
+ "epoch": 6.17,
129
+ "learning_rate": 9.460000000000001e-06,
130
+ "loss": 0.1479,
131
+ "step": 475
132
+ },
133
+ {
134
+ "epoch": 6.49,
135
+ "learning_rate": 9.960000000000001e-06,
136
+ "loss": 0.103,
137
+ "step": 500
138
+ },
139
+ {
140
+ "epoch": 6.81,
141
+ "learning_rate": 9.975789473684211e-06,
142
+ "loss": 0.1271,
143
+ "step": 525
144
+ },
145
+ {
146
+ "epoch": 7.14,
147
+ "learning_rate": 9.949473684210526e-06,
148
+ "loss": 0.1033,
149
+ "step": 550
150
+ },
151
+ {
152
+ "epoch": 7.46,
153
+ "learning_rate": 9.923157894736844e-06,
154
+ "loss": 0.0799,
155
+ "step": 575
156
+ },
157
+ {
158
+ "epoch": 7.78,
159
+ "learning_rate": 9.89684210526316e-06,
160
+ "loss": 0.0814,
161
+ "step": 600
162
+ },
163
+ {
164
+ "epoch": 8.12,
165
+ "learning_rate": 9.870526315789474e-06,
166
+ "loss": 0.069,
167
+ "step": 625
168
+ },
169
+ {
170
+ "epoch": 8.44,
171
+ "learning_rate": 9.84421052631579e-06,
172
+ "loss": 0.0546,
173
+ "step": 650
174
+ },
175
+ {
176
+ "epoch": 8.76,
177
+ "learning_rate": 9.817894736842106e-06,
178
+ "loss": 0.0627,
179
+ "step": 675
180
+ },
181
+ {
182
+ "epoch": 9.09,
183
+ "learning_rate": 9.791578947368422e-06,
184
+ "loss": 0.0459,
185
+ "step": 700
186
+ },
187
+ {
188
+ "epoch": 9.41,
189
+ "learning_rate": 9.765263157894737e-06,
190
+ "loss": 0.034,
191
+ "step": 725
192
+ },
193
+ {
194
+ "epoch": 9.73,
195
+ "learning_rate": 9.738947368421054e-06,
196
+ "loss": 0.036,
197
+ "step": 750
198
+ },
199
+ {
200
+ "epoch": 10.06,
201
+ "learning_rate": 9.712631578947369e-06,
202
+ "loss": 0.0367,
203
+ "step": 775
204
+ },
205
+ {
206
+ "epoch": 10.39,
207
+ "learning_rate": 9.686315789473684e-06,
208
+ "loss": 0.0248,
209
+ "step": 800
210
+ },
211
+ {
212
+ "epoch": 10.39,
213
+ "eval_loss": 0.8046127557754517,
214
+ "eval_runtime": 19.9036,
215
+ "eval_samples_per_second": 6.984,
216
+ "eval_steps_per_second": 0.904,
217
+ "eval_wer": 108.59188544152745,
218
+ "step": 800
219
+ },
220
+ {
221
+ "epoch": 10.71,
222
+ "learning_rate": 9.66e-06,
223
+ "loss": 0.0324,
224
+ "step": 825
225
+ },
226
+ {
227
+ "epoch": 11.04,
228
+ "learning_rate": 9.633684210526316e-06,
229
+ "loss": 0.0374,
230
+ "step": 850
231
+ },
232
+ {
233
+ "epoch": 11.36,
234
+ "learning_rate": 9.607368421052632e-06,
235
+ "loss": 0.0253,
236
+ "step": 875
237
+ },
238
+ {
239
+ "epoch": 11.68,
240
+ "learning_rate": 9.581052631578947e-06,
241
+ "loss": 0.0262,
242
+ "step": 900
243
+ },
244
+ {
245
+ "epoch": 12.01,
246
+ "learning_rate": 9.554736842105264e-06,
247
+ "loss": 0.0374,
248
+ "step": 925
249
+ },
250
+ {
251
+ "epoch": 12.33,
252
+ "learning_rate": 9.52842105263158e-06,
253
+ "loss": 0.0179,
254
+ "step": 950
255
+ },
256
+ {
257
+ "epoch": 12.66,
258
+ "learning_rate": 9.502105263157896e-06,
259
+ "loss": 0.0236,
260
+ "step": 975
261
+ },
262
+ {
263
+ "epoch": 12.98,
264
+ "learning_rate": 9.475789473684212e-06,
265
+ "loss": 0.0187,
266
+ "step": 1000
267
+ },
268
+ {
269
+ "epoch": 13.31,
270
+ "learning_rate": 9.449473684210527e-06,
271
+ "loss": 0.0186,
272
+ "step": 1025
273
+ },
274
+ {
275
+ "epoch": 13.63,
276
+ "learning_rate": 9.423157894736842e-06,
277
+ "loss": 0.0217,
278
+ "step": 1050
279
+ },
280
+ {
281
+ "epoch": 13.95,
282
+ "learning_rate": 9.396842105263159e-06,
283
+ "loss": 0.0206,
284
+ "step": 1075
285
+ },
286
+ {
287
+ "epoch": 14.28,
288
+ "learning_rate": 9.370526315789474e-06,
289
+ "loss": 0.0103,
290
+ "step": 1100
291
+ },
292
+ {
293
+ "epoch": 14.6,
294
+ "learning_rate": 9.34421052631579e-06,
295
+ "loss": 0.0168,
296
+ "step": 1125
297
+ },
298
+ {
299
+ "epoch": 14.93,
300
+ "learning_rate": 9.317894736842105e-06,
301
+ "loss": 0.0214,
302
+ "step": 1150
303
+ },
304
+ {
305
+ "epoch": 15.26,
306
+ "learning_rate": 9.291578947368422e-06,
307
+ "loss": 0.0115,
308
+ "step": 1175
309
+ },
310
+ {
311
+ "epoch": 15.58,
312
+ "learning_rate": 9.265263157894737e-06,
313
+ "loss": 0.0138,
314
+ "step": 1200
315
+ },
316
+ {
317
+ "epoch": 15.58,
318
+ "eval_loss": 0.8237143158912659,
319
+ "eval_runtime": 22.295,
320
+ "eval_samples_per_second": 6.235,
321
+ "eval_steps_per_second": 0.807,
322
+ "eval_wer": 117.89976133651552,
323
+ "step": 1200
324
+ },
325
+ {
326
+ "epoch": 15.9,
327
+ "learning_rate": 9.238947368421052e-06,
328
+ "loss": 0.0216,
329
+ "step": 1225
330
+ },
331
+ {
332
+ "epoch": 16.23,
333
+ "learning_rate": 9.21263157894737e-06,
334
+ "loss": 0.014,
335
+ "step": 1250
336
+ },
337
+ {
338
+ "epoch": 16.55,
339
+ "learning_rate": 9.186315789473685e-06,
340
+ "loss": 0.0117,
341
+ "step": 1275
342
+ },
343
+ {
344
+ "epoch": 16.87,
345
+ "learning_rate": 9.16e-06,
346
+ "loss": 0.0104,
347
+ "step": 1300
348
+ },
349
+ {
350
+ "epoch": 17.21,
351
+ "learning_rate": 9.133684210526317e-06,
352
+ "loss": 0.0086,
353
+ "step": 1325
354
+ },
355
+ {
356
+ "epoch": 17.53,
357
+ "learning_rate": 9.107368421052632e-06,
358
+ "loss": 0.0092,
359
+ "step": 1350
360
+ },
361
+ {
362
+ "epoch": 17.85,
363
+ "learning_rate": 9.081052631578949e-06,
364
+ "loss": 0.0098,
365
+ "step": 1375
366
+ },
367
+ {
368
+ "epoch": 18.18,
369
+ "learning_rate": 9.054736842105264e-06,
370
+ "loss": 0.0125,
371
+ "step": 1400
372
+ },
373
+ {
374
+ "epoch": 18.5,
375
+ "learning_rate": 9.02842105263158e-06,
376
+ "loss": 0.0055,
377
+ "step": 1425
378
+ },
379
+ {
380
+ "epoch": 18.82,
381
+ "learning_rate": 9.002105263157895e-06,
382
+ "loss": 0.011,
383
+ "step": 1450
384
+ },
385
+ {
386
+ "epoch": 19.15,
387
+ "learning_rate": 8.97578947368421e-06,
388
+ "loss": 0.0088,
389
+ "step": 1475
390
+ },
391
+ {
392
+ "epoch": 19.48,
393
+ "learning_rate": 8.949473684210527e-06,
394
+ "loss": 0.0065,
395
+ "step": 1500
396
+ },
397
+ {
398
+ "epoch": 19.8,
399
+ "learning_rate": 8.923157894736842e-06,
400
+ "loss": 0.0098,
401
+ "step": 1525
402
+ },
403
+ {
404
+ "epoch": 20.13,
405
+ "learning_rate": 8.896842105263159e-06,
406
+ "loss": 0.0038,
407
+ "step": 1550
408
+ },
409
+ {
410
+ "epoch": 20.45,
411
+ "learning_rate": 8.870526315789474e-06,
412
+ "loss": 0.0053,
413
+ "step": 1575
414
+ },
415
+ {
416
+ "epoch": 20.77,
417
+ "learning_rate": 8.84421052631579e-06,
418
+ "loss": 0.0075,
419
+ "step": 1600
420
+ },
421
+ {
422
+ "epoch": 20.77,
423
+ "eval_loss": 0.8213781118392944,
424
+ "eval_runtime": 18.9322,
425
+ "eval_samples_per_second": 7.342,
426
+ "eval_steps_per_second": 0.951,
427
+ "eval_wer": 103.5799522673031,
428
+ "step": 1600
429
+ },
430
+ {
431
+ "epoch": 21.1,
432
+ "learning_rate": 8.817894736842107e-06,
433
+ "loss": 0.0111,
434
+ "step": 1625
435
+ },
436
+ {
437
+ "epoch": 21.42,
438
+ "learning_rate": 8.791578947368422e-06,
439
+ "loss": 0.0083,
440
+ "step": 1650
441
+ },
442
+ {
443
+ "epoch": 21.75,
444
+ "learning_rate": 8.765263157894739e-06,
445
+ "loss": 0.0056,
446
+ "step": 1675
447
+ },
448
+ {
449
+ "epoch": 22.08,
450
+ "learning_rate": 8.738947368421053e-06,
451
+ "loss": 0.0055,
452
+ "step": 1700
453
+ },
454
+ {
455
+ "epoch": 22.4,
456
+ "learning_rate": 8.712631578947368e-06,
457
+ "loss": 0.0031,
458
+ "step": 1725
459
+ },
460
+ {
461
+ "epoch": 22.72,
462
+ "learning_rate": 8.686315789473685e-06,
463
+ "loss": 0.0075,
464
+ "step": 1750
465
+ },
466
+ {
467
+ "epoch": 23.05,
468
+ "learning_rate": 8.66e-06,
469
+ "loss": 0.0078,
470
+ "step": 1775
471
+ },
472
+ {
473
+ "epoch": 23.37,
474
+ "learning_rate": 8.633684210526317e-06,
475
+ "loss": 0.0063,
476
+ "step": 1800
477
+ },
478
+ {
479
+ "epoch": 23.69,
480
+ "learning_rate": 8.607368421052632e-06,
481
+ "loss": 0.0012,
482
+ "step": 1825
483
+ },
484
+ {
485
+ "epoch": 24.03,
486
+ "learning_rate": 8.581052631578948e-06,
487
+ "loss": 0.0045,
488
+ "step": 1850
489
+ },
490
+ {
491
+ "epoch": 24.35,
492
+ "learning_rate": 8.554736842105263e-06,
493
+ "loss": 0.004,
494
+ "step": 1875
495
+ },
496
+ {
497
+ "epoch": 24.67,
498
+ "learning_rate": 8.528421052631578e-06,
499
+ "loss": 0.006,
500
+ "step": 1900
501
+ },
502
+ {
503
+ "epoch": 24.99,
504
+ "learning_rate": 8.502105263157897e-06,
505
+ "loss": 0.0034,
506
+ "step": 1925
507
+ },
508
+ {
509
+ "epoch": 25.32,
510
+ "learning_rate": 8.475789473684212e-06,
511
+ "loss": 0.0024,
512
+ "step": 1950
513
+ },
514
+ {
515
+ "epoch": 25.64,
516
+ "learning_rate": 8.449473684210527e-06,
517
+ "loss": 0.0032,
518
+ "step": 1975
519
+ },
520
+ {
521
+ "epoch": 25.96,
522
+ "learning_rate": 8.423157894736843e-06,
523
+ "loss": 0.0021,
524
+ "step": 2000
525
+ },
526
+ {
527
+ "epoch": 25.96,
528
+ "eval_loss": 0.7943580746650696,
529
+ "eval_runtime": 20.6652,
530
+ "eval_samples_per_second": 6.726,
531
+ "eval_steps_per_second": 0.871,
532
+ "eval_wer": 106.92124105011933,
533
+ "step": 2000
534
+ },
535
+ {
536
+ "epoch": 26.3,
537
+ "learning_rate": 8.396842105263158e-06,
538
+ "loss": 0.0006,
539
+ "step": 2025
540
+ },
541
+ {
542
+ "epoch": 26.62,
543
+ "learning_rate": 8.370526315789475e-06,
544
+ "loss": 0.0018,
545
+ "step": 2050
546
+ },
547
+ {
548
+ "epoch": 26.94,
549
+ "learning_rate": 8.34421052631579e-06,
550
+ "loss": 0.0032,
551
+ "step": 2075
552
+ },
553
+ {
554
+ "epoch": 27.27,
555
+ "learning_rate": 8.317894736842107e-06,
556
+ "loss": 0.0034,
557
+ "step": 2100
558
+ },
559
+ {
560
+ "epoch": 27.59,
561
+ "learning_rate": 8.291578947368422e-06,
562
+ "loss": 0.0021,
563
+ "step": 2125
564
+ },
565
+ {
566
+ "epoch": 27.91,
567
+ "learning_rate": 8.265263157894737e-06,
568
+ "loss": 0.005,
569
+ "step": 2150
570
+ },
571
+ {
572
+ "epoch": 28.24,
573
+ "learning_rate": 8.238947368421053e-06,
574
+ "loss": 0.002,
575
+ "step": 2175
576
+ },
577
+ {
578
+ "epoch": 28.57,
579
+ "learning_rate": 8.212631578947368e-06,
580
+ "loss": 0.0023,
581
+ "step": 2200
582
+ },
583
+ {
584
+ "epoch": 28.89,
585
+ "learning_rate": 8.186315789473685e-06,
586
+ "loss": 0.0126,
587
+ "step": 2225
588
+ },
589
+ {
590
+ "epoch": 29.22,
591
+ "learning_rate": 8.16e-06,
592
+ "loss": 0.0064,
593
+ "step": 2250
594
+ },
595
+ {
596
+ "epoch": 29.54,
597
+ "learning_rate": 8.133684210526316e-06,
598
+ "loss": 0.004,
599
+ "step": 2275
600
+ },
601
+ {
602
+ "epoch": 29.86,
603
+ "learning_rate": 8.107368421052633e-06,
604
+ "loss": 0.003,
605
+ "step": 2300
606
+ },
607
+ {
608
+ "epoch": 30.19,
609
+ "learning_rate": 8.081052631578948e-06,
610
+ "loss": 0.0064,
611
+ "step": 2325
612
+ },
613
+ {
614
+ "epoch": 30.51,
615
+ "learning_rate": 8.054736842105265e-06,
616
+ "loss": 0.0038,
617
+ "step": 2350
618
+ },
619
+ {
620
+ "epoch": 30.84,
621
+ "learning_rate": 8.02842105263158e-06,
622
+ "loss": 0.006,
623
+ "step": 2375
624
+ },
625
+ {
626
+ "epoch": 31.17,
627
+ "learning_rate": 8.002105263157895e-06,
628
+ "loss": 0.003,
629
+ "step": 2400
630
+ },
631
+ {
632
+ "epoch": 31.17,
633
+ "eval_loss": 0.811363160610199,
634
+ "eval_runtime": 17.6674,
635
+ "eval_samples_per_second": 7.868,
636
+ "eval_steps_per_second": 1.019,
637
+ "eval_wer": 98.68735083532219,
638
+ "step": 2400
639
+ },
640
+ {
641
+ "epoch": 31.49,
642
+ "learning_rate": 7.975789473684211e-06,
643
+ "loss": 0.0021,
644
+ "step": 2425
645
+ },
646
+ {
647
+ "epoch": 31.81,
648
+ "learning_rate": 7.949473684210526e-06,
649
+ "loss": 0.0045,
650
+ "step": 2450
651
+ },
652
+ {
653
+ "epoch": 32.14,
654
+ "learning_rate": 7.923157894736843e-06,
655
+ "loss": 0.0028,
656
+ "step": 2475
657
+ },
658
+ {
659
+ "epoch": 32.46,
660
+ "learning_rate": 7.896842105263158e-06,
661
+ "loss": 0.0008,
662
+ "step": 2500
663
+ },
664
+ {
665
+ "epoch": 32.78,
666
+ "learning_rate": 7.870526315789475e-06,
667
+ "loss": 0.0006,
668
+ "step": 2525
669
+ },
670
+ {
671
+ "epoch": 33.12,
672
+ "learning_rate": 7.84421052631579e-06,
673
+ "loss": 0.0023,
674
+ "step": 2550
675
+ },
676
+ {
677
+ "epoch": 33.44,
678
+ "learning_rate": 7.817894736842105e-06,
679
+ "loss": 0.0023,
680
+ "step": 2575
681
+ },
682
+ {
683
+ "epoch": 33.76,
684
+ "learning_rate": 7.791578947368423e-06,
685
+ "loss": 0.0068,
686
+ "step": 2600
687
+ },
688
+ {
689
+ "epoch": 34.09,
690
+ "learning_rate": 7.765263157894738e-06,
691
+ "loss": 0.0009,
692
+ "step": 2625
693
+ },
694
+ {
695
+ "epoch": 34.41,
696
+ "learning_rate": 7.738947368421053e-06,
697
+ "loss": 0.0011,
698
+ "step": 2650
699
+ },
700
+ {
701
+ "epoch": 34.73,
702
+ "learning_rate": 7.71263157894737e-06,
703
+ "loss": 0.0027,
704
+ "step": 2675
705
+ },
706
+ {
707
+ "epoch": 35.06,
708
+ "learning_rate": 7.686315789473685e-06,
709
+ "loss": 0.0008,
710
+ "step": 2700
711
+ },
712
+ {
713
+ "epoch": 35.39,
714
+ "learning_rate": 7.660000000000001e-06,
715
+ "loss": 0.0021,
716
+ "step": 2725
717
+ },
718
+ {
719
+ "epoch": 35.71,
720
+ "learning_rate": 7.633684210526316e-06,
721
+ "loss": 0.0028,
722
+ "step": 2750
723
+ },
724
+ {
725
+ "epoch": 36.04,
726
+ "learning_rate": 7.607368421052632e-06,
727
+ "loss": 0.0018,
728
+ "step": 2775
729
+ },
730
+ {
731
+ "epoch": 36.36,
732
+ "learning_rate": 7.581052631578948e-06,
733
+ "loss": 0.0032,
734
+ "step": 2800
735
+ },
736
+ {
737
+ "epoch": 36.36,
738
+ "eval_loss": 0.7548059821128845,
739
+ "eval_runtime": 19.1115,
740
+ "eval_samples_per_second": 7.273,
741
+ "eval_steps_per_second": 0.942,
742
+ "eval_wer": 105.48926014319808,
743
+ "step": 2800
744
+ },
745
+ {
746
+ "epoch": 36.68,
747
+ "learning_rate": 7.554736842105264e-06,
748
+ "loss": 0.003,
749
+ "step": 2825
750
+ },
751
+ {
752
+ "epoch": 37.01,
753
+ "learning_rate": 7.5284210526315794e-06,
754
+ "loss": 0.0005,
755
+ "step": 2850
756
+ },
757
+ {
758
+ "epoch": 37.33,
759
+ "learning_rate": 7.502105263157895e-06,
760
+ "loss": 0.0038,
761
+ "step": 2875
762
+ },
763
+ {
764
+ "epoch": 37.66,
765
+ "learning_rate": 7.475789473684211e-06,
766
+ "loss": 0.0021,
767
+ "step": 2900
768
+ },
769
+ {
770
+ "epoch": 37.98,
771
+ "learning_rate": 7.449473684210526e-06,
772
+ "loss": 0.0011,
773
+ "step": 2925
774
+ },
775
+ {
776
+ "epoch": 38.31,
777
+ "learning_rate": 7.4231578947368436e-06,
778
+ "loss": 0.0023,
779
+ "step": 2950
780
+ },
781
+ {
782
+ "epoch": 38.63,
783
+ "learning_rate": 7.3968421052631585e-06,
784
+ "loss": 0.003,
785
+ "step": 2975
786
+ },
787
+ {
788
+ "epoch": 38.95,
789
+ "learning_rate": 7.370526315789474e-06,
790
+ "loss": 0.0046,
791
+ "step": 3000
792
+ },
793
+ {
794
+ "epoch": 39.28,
795
+ "learning_rate": 7.34421052631579e-06,
796
+ "loss": 0.001,
797
+ "step": 3025
798
+ },
799
+ {
800
+ "epoch": 39.6,
801
+ "learning_rate": 7.317894736842106e-06,
802
+ "loss": 0.0003,
803
+ "step": 3050
804
+ },
805
+ {
806
+ "epoch": 39.93,
807
+ "learning_rate": 7.291578947368422e-06,
808
+ "loss": 0.003,
809
+ "step": 3075
810
+ },
811
+ {
812
+ "epoch": 40.26,
813
+ "learning_rate": 7.265263157894738e-06,
814
+ "loss": 0.0013,
815
+ "step": 3100
816
+ },
817
+ {
818
+ "epoch": 40.58,
819
+ "learning_rate": 7.2389473684210534e-06,
820
+ "loss": 0.0016,
821
+ "step": 3125
822
+ },
823
+ {
824
+ "epoch": 40.9,
825
+ "learning_rate": 7.212631578947369e-06,
826
+ "loss": 0.0009,
827
+ "step": 3150
828
+ },
829
+ {
830
+ "epoch": 41.23,
831
+ "learning_rate": 7.186315789473684e-06,
832
+ "loss": 0.0004,
833
+ "step": 3175
834
+ },
835
+ {
836
+ "epoch": 41.55,
837
+ "learning_rate": 7.16e-06,
838
+ "loss": 0.0002,
839
+ "step": 3200
840
+ },
841
+ {
842
+ "epoch": 41.55,
843
+ "eval_loss": 0.7998279929161072,
844
+ "eval_runtime": 17.346,
845
+ "eval_samples_per_second": 8.013,
846
+ "eval_steps_per_second": 1.038,
847
+ "eval_wer": 99.88066825775657,
848
+ "step": 3200
849
+ },
850
+ {
851
+ "epoch": 41.87,
852
+ "learning_rate": 7.133684210526316e-06,
853
+ "loss": 0.0008,
854
+ "step": 3225
855
+ },
856
+ {
857
+ "epoch": 42.21,
858
+ "learning_rate": 7.107368421052632e-06,
859
+ "loss": 0.0014,
860
+ "step": 3250
861
+ },
862
+ {
863
+ "epoch": 42.53,
864
+ "learning_rate": 7.0810526315789475e-06,
865
+ "loss": 0.0016,
866
+ "step": 3275
867
+ },
868
+ {
869
+ "epoch": 42.85,
870
+ "learning_rate": 7.054736842105264e-06,
871
+ "loss": 0.0015,
872
+ "step": 3300
873
+ },
874
+ {
875
+ "epoch": 43.18,
876
+ "learning_rate": 7.02842105263158e-06,
877
+ "loss": 0.0032,
878
+ "step": 3325
879
+ },
880
+ {
881
+ "epoch": 43.5,
882
+ "learning_rate": 7.002105263157896e-06,
883
+ "loss": 0.0027,
884
+ "step": 3350
885
+ },
886
+ {
887
+ "epoch": 43.82,
888
+ "learning_rate": 6.975789473684212e-06,
889
+ "loss": 0.0027,
890
+ "step": 3375
891
+ },
892
+ {
893
+ "epoch": 44.15,
894
+ "learning_rate": 6.9494736842105275e-06,
895
+ "loss": 0.0052,
896
+ "step": 3400
897
+ },
898
+ {
899
+ "epoch": 44.48,
900
+ "learning_rate": 6.9231578947368424e-06,
901
+ "loss": 0.0081,
902
+ "step": 3425
903
+ },
904
+ {
905
+ "epoch": 44.8,
906
+ "learning_rate": 6.896842105263158e-06,
907
+ "loss": 0.0076,
908
+ "step": 3450
909
+ },
910
+ {
911
+ "epoch": 45.13,
912
+ "learning_rate": 6.870526315789474e-06,
913
+ "loss": 0.0074,
914
+ "step": 3475
915
+ },
916
+ {
917
+ "epoch": 45.45,
918
+ "learning_rate": 6.84421052631579e-06,
919
+ "loss": 0.003,
920
+ "step": 3500
921
+ },
922
+ {
923
+ "epoch": 45.77,
924
+ "learning_rate": 6.817894736842106e-06,
925
+ "loss": 0.0058,
926
+ "step": 3525
927
+ },
928
+ {
929
+ "epoch": 46.1,
930
+ "learning_rate": 6.7915789473684215e-06,
931
+ "loss": 0.007,
932
+ "step": 3550
933
+ },
934
+ {
935
+ "epoch": 46.42,
936
+ "learning_rate": 6.765263157894737e-06,
937
+ "loss": 0.0098,
938
+ "step": 3575
939
+ },
940
+ {
941
+ "epoch": 46.75,
942
+ "learning_rate": 6.738947368421052e-06,
943
+ "loss": 0.0077,
944
+ "step": 3600
945
+ },
946
+ {
947
+ "epoch": 46.75,
948
+ "eval_loss": 0.8303264379501343,
949
+ "eval_runtime": 22.148,
950
+ "eval_samples_per_second": 6.276,
951
+ "eval_steps_per_second": 0.813,
952
+ "eval_wer": 123.74701670644393,
953
+ "step": 3600
954
+ },
955
+ {
956
+ "epoch": 47.08,
957
+ "learning_rate": 6.71263157894737e-06,
958
+ "loss": 0.0072,
959
+ "step": 3625
960
+ },
961
+ {
962
+ "epoch": 47.4,
963
+ "learning_rate": 6.686315789473685e-06,
964
+ "loss": 0.0067,
965
+ "step": 3650
966
+ },
967
+ {
968
+ "epoch": 47.72,
969
+ "learning_rate": 6.660000000000001e-06,
970
+ "loss": 0.0031,
971
+ "step": 3675
972
+ },
973
+ {
974
+ "epoch": 48.05,
975
+ "learning_rate": 6.6336842105263164e-06,
976
+ "loss": 0.0041,
977
+ "step": 3700
978
+ },
979
+ {
980
+ "epoch": 48.37,
981
+ "learning_rate": 6.607368421052632e-06,
982
+ "loss": 0.0018,
983
+ "step": 3725
984
+ },
985
+ {
986
+ "epoch": 48.69,
987
+ "learning_rate": 6.581052631578948e-06,
988
+ "loss": 0.0032,
989
+ "step": 3750
990
+ },
991
+ {
992
+ "epoch": 49.03,
993
+ "learning_rate": 6.554736842105264e-06,
994
+ "loss": 0.0046,
995
+ "step": 3775
996
+ },
997
+ {
998
+ "epoch": 49.35,
999
+ "learning_rate": 6.52842105263158e-06,
1000
+ "loss": 0.0025,
1001
+ "step": 3800
1002
+ },
1003
+ {
1004
+ "epoch": 49.67,
1005
+ "learning_rate": 6.5021052631578955e-06,
1006
+ "loss": 0.0042,
1007
+ "step": 3825
1008
+ },
1009
+ {
1010
+ "epoch": 49.99,
1011
+ "learning_rate": 6.4757894736842105e-06,
1012
+ "loss": 0.0044,
1013
+ "step": 3850
1014
+ },
1015
+ {
1016
+ "epoch": 50.32,
1017
+ "learning_rate": 6.449473684210526e-06,
1018
+ "loss": 0.003,
1019
+ "step": 3875
1020
+ },
1021
+ {
1022
+ "epoch": 50.64,
1023
+ "learning_rate": 6.423157894736842e-06,
1024
+ "loss": 0.0023,
1025
+ "step": 3900
1026
+ },
1027
+ {
1028
+ "epoch": 50.96,
1029
+ "learning_rate": 6.396842105263158e-06,
1030
+ "loss": 0.0019,
1031
+ "step": 3925
1032
+ },
1033
+ {
1034
+ "epoch": 51.3,
1035
+ "learning_rate": 6.370526315789474e-06,
1036
+ "loss": 0.0037,
1037
+ "step": 3950
1038
+ },
1039
+ {
1040
+ "epoch": 51.62,
1041
+ "learning_rate": 6.3442105263157904e-06,
1042
+ "loss": 0.0027,
1043
+ "step": 3975
1044
+ },
1045
+ {
1046
+ "epoch": 51.94,
1047
+ "learning_rate": 6.317894736842106e-06,
1048
+ "loss": 0.0013,
1049
+ "step": 4000
1050
+ },
1051
+ {
1052
+ "epoch": 51.94,
1053
+ "eval_loss": 0.7767909169197083,
1054
+ "eval_runtime": 21.9995,
1055
+ "eval_samples_per_second": 6.318,
1056
+ "eval_steps_per_second": 0.818,
1057
+ "eval_wer": 118.49642004773268,
1058
+ "step": 4000
1059
+ },
1060
+ {
1061
+ "epoch": 52.27,
1062
+ "learning_rate": 6.291578947368422e-06,
1063
+ "loss": 0.0022,
1064
+ "step": 4025
1065
+ },
1066
+ {
1067
+ "epoch": 52.59,
1068
+ "learning_rate": 6.265263157894738e-06,
1069
+ "loss": 0.0019,
1070
+ "step": 4050
1071
+ },
1072
+ {
1073
+ "epoch": 52.91,
1074
+ "learning_rate": 6.238947368421054e-06,
1075
+ "loss": 0.0006,
1076
+ "step": 4075
1077
+ },
1078
+ {
1079
+ "epoch": 53.24,
1080
+ "learning_rate": 6.212631578947369e-06,
1081
+ "loss": 0.0018,
1082
+ "step": 4100
1083
+ },
1084
+ {
1085
+ "epoch": 53.57,
1086
+ "learning_rate": 6.1863157894736845e-06,
1087
+ "loss": 0.0027,
1088
+ "step": 4125
1089
+ },
1090
+ {
1091
+ "epoch": 53.89,
1092
+ "learning_rate": 6.16e-06,
1093
+ "loss": 0.0007,
1094
+ "step": 4150
1095
+ },
1096
+ {
1097
+ "epoch": 54.22,
1098
+ "learning_rate": 6.133684210526316e-06,
1099
+ "loss": 0.0005,
1100
+ "step": 4175
1101
+ },
1102
+ {
1103
+ "epoch": 54.54,
1104
+ "learning_rate": 6.107368421052632e-06,
1105
+ "loss": 0.0005,
1106
+ "step": 4200
1107
+ },
1108
+ {
1109
+ "epoch": 54.86,
1110
+ "learning_rate": 6.081052631578948e-06,
1111
+ "loss": 0.0002,
1112
+ "step": 4225
1113
+ },
1114
+ {
1115
+ "epoch": 55.19,
1116
+ "learning_rate": 6.054736842105264e-06,
1117
+ "loss": 0.0004,
1118
+ "step": 4250
1119
+ },
1120
+ {
1121
+ "epoch": 55.51,
1122
+ "learning_rate": 6.0284210526315786e-06,
1123
+ "loss": 0.0001,
1124
+ "step": 4275
1125
+ },
1126
+ {
1127
+ "epoch": 55.84,
1128
+ "learning_rate": 6.002105263157896e-06,
1129
+ "loss": 0.0001,
1130
+ "step": 4300
1131
+ },
1132
+ {
1133
+ "epoch": 56.17,
1134
+ "learning_rate": 5.975789473684212e-06,
1135
+ "loss": 0.0001,
1136
+ "step": 4325
1137
+ },
1138
+ {
1139
+ "epoch": 56.49,
1140
+ "learning_rate": 5.949473684210527e-06,
1141
+ "loss": 0.0001,
1142
+ "step": 4350
1143
+ },
1144
+ {
1145
+ "epoch": 56.81,
1146
+ "learning_rate": 5.923157894736843e-06,
1147
+ "loss": 0.0001,
1148
+ "step": 4375
1149
+ },
1150
+ {
1151
+ "epoch": 57.14,
1152
+ "learning_rate": 5.8968421052631585e-06,
1153
+ "loss": 0.0001,
1154
+ "step": 4400
1155
+ },
1156
+ {
1157
+ "epoch": 57.14,
1158
+ "eval_loss": 0.772087812423706,
1159
+ "eval_runtime": 20.4241,
1160
+ "eval_samples_per_second": 6.806,
1161
+ "eval_steps_per_second": 0.881,
1162
+ "eval_wer": 97.01670644391407,
1163
+ "step": 4400
1164
+ },
1165
+ {
1166
+ "epoch": 57.46,
1167
+ "learning_rate": 5.870526315789474e-06,
1168
+ "loss": 0.0001,
1169
+ "step": 4425
1170
+ },
1171
+ {
1172
+ "epoch": 57.78,
1173
+ "learning_rate": 5.84421052631579e-06,
1174
+ "loss": 0.0001,
1175
+ "step": 4450
1176
+ },
1177
+ {
1178
+ "epoch": 58.12,
1179
+ "learning_rate": 5.817894736842106e-06,
1180
+ "loss": 0.0001,
1181
+ "step": 4475
1182
+ },
1183
+ {
1184
+ "epoch": 58.44,
1185
+ "learning_rate": 5.791578947368422e-06,
1186
+ "loss": 0.0001,
1187
+ "step": 4500
1188
+ },
1189
+ {
1190
+ "epoch": 58.76,
1191
+ "learning_rate": 5.765263157894737e-06,
1192
+ "loss": 0.0001,
1193
+ "step": 4525
1194
+ },
1195
+ {
1196
+ "epoch": 59.09,
1197
+ "learning_rate": 5.7389473684210526e-06,
1198
+ "loss": 0.0001,
1199
+ "step": 4550
1200
+ },
1201
+ {
1202
+ "epoch": 59.41,
1203
+ "learning_rate": 5.712631578947368e-06,
1204
+ "loss": 0.0,
1205
+ "step": 4575
1206
+ },
1207
+ {
1208
+ "epoch": 59.73,
1209
+ "learning_rate": 5.686315789473684e-06,
1210
+ "loss": 0.0001,
1211
+ "step": 4600
1212
+ },
1213
+ {
1214
+ "epoch": 60.06,
1215
+ "learning_rate": 5.66e-06,
1216
+ "loss": 0.0001,
1217
+ "step": 4625
1218
+ },
1219
+ {
1220
+ "epoch": 60.39,
1221
+ "learning_rate": 5.633684210526317e-06,
1222
+ "loss": 0.0,
1223
+ "step": 4650
1224
+ },
1225
+ {
1226
+ "epoch": 60.71,
1227
+ "learning_rate": 5.6073684210526325e-06,
1228
+ "loss": 0.0,
1229
+ "step": 4675
1230
+ },
1231
+ {
1232
+ "epoch": 61.04,
1233
+ "learning_rate": 5.581052631578948e-06,
1234
+ "loss": 0.0,
1235
+ "step": 4700
1236
+ },
1237
+ {
1238
+ "epoch": 61.36,
1239
+ "learning_rate": 5.554736842105264e-06,
1240
+ "loss": 0.0,
1241
+ "step": 4725
1242
+ },
1243
+ {
1244
+ "epoch": 61.68,
1245
+ "learning_rate": 5.52842105263158e-06,
1246
+ "loss": 0.0,
1247
+ "step": 4750
1248
+ },
1249
+ {
1250
+ "epoch": 62.01,
1251
+ "learning_rate": 5.502105263157895e-06,
1252
+ "loss": 0.0,
1253
+ "step": 4775
1254
+ },
1255
+ {
1256
+ "epoch": 62.33,
1257
+ "learning_rate": 5.475789473684211e-06,
1258
+ "loss": 0.0,
1259
+ "step": 4800
1260
+ },
1261
+ {
1262
+ "epoch": 62.33,
1263
+ "eval_loss": 0.7726908922195435,
1264
+ "eval_runtime": 20.4469,
1265
+ "eval_samples_per_second": 6.798,
1266
+ "eval_steps_per_second": 0.88,
1267
+ "eval_wer": 97.49403341288783,
1268
+ "step": 4800
1269
+ },
1270
+ {
1271
+ "epoch": 62.66,
1272
+ "learning_rate": 5.4494736842105266e-06,
1273
+ "loss": 0.0,
1274
+ "step": 4825
1275
+ },
1276
+ {
1277
+ "epoch": 62.98,
1278
+ "learning_rate": 5.423157894736842e-06,
1279
+ "loss": 0.0,
1280
+ "step": 4850
1281
+ },
1282
+ {
1283
+ "epoch": 63.31,
1284
+ "learning_rate": 5.396842105263158e-06,
1285
+ "loss": 0.0,
1286
+ "step": 4875
1287
+ },
1288
+ {
1289
+ "epoch": 63.63,
1290
+ "learning_rate": 5.370526315789474e-06,
1291
+ "loss": 0.0,
1292
+ "step": 4900
1293
+ },
1294
+ {
1295
+ "epoch": 63.95,
1296
+ "learning_rate": 5.34421052631579e-06,
1297
+ "loss": 0.0,
1298
+ "step": 4925
1299
+ },
1300
+ {
1301
+ "epoch": 64.28,
1302
+ "learning_rate": 5.317894736842105e-06,
1303
+ "loss": 0.0,
1304
+ "step": 4950
1305
+ },
1306
+ {
1307
+ "epoch": 64.6,
1308
+ "learning_rate": 5.291578947368422e-06,
1309
+ "loss": 0.0,
1310
+ "step": 4975
1311
+ },
1312
+ {
1313
+ "epoch": 64.93,
1314
+ "learning_rate": 5.265263157894738e-06,
1315
+ "loss": 0.0,
1316
+ "step": 5000
1317
+ },
1318
+ {
1319
+ "epoch": 65.26,
1320
+ "learning_rate": 5.238947368421053e-06,
1321
+ "loss": 0.0,
1322
+ "step": 5025
1323
+ },
1324
+ {
1325
+ "epoch": 65.58,
1326
+ "learning_rate": 5.212631578947369e-06,
1327
+ "loss": 0.0,
1328
+ "step": 5050
1329
+ },
1330
+ {
1331
+ "epoch": 65.9,
1332
+ "learning_rate": 5.186315789473685e-06,
1333
+ "loss": 0.0,
1334
+ "step": 5075
1335
+ },
1336
+ {
1337
+ "epoch": 66.23,
1338
+ "learning_rate": 5.1600000000000006e-06,
1339
+ "loss": 0.0,
1340
+ "step": 5100
1341
+ },
1342
+ {
1343
+ "epoch": 66.55,
1344
+ "learning_rate": 5.133684210526316e-06,
1345
+ "loss": 0.0,
1346
+ "step": 5125
1347
+ },
1348
+ {
1349
+ "epoch": 66.87,
1350
+ "learning_rate": 5.107368421052632e-06,
1351
+ "loss": 0.0,
1352
+ "step": 5150
1353
+ },
1354
+ {
1355
+ "epoch": 67.21,
1356
+ "learning_rate": 5.081052631578948e-06,
1357
+ "loss": 0.0,
1358
+ "step": 5175
1359
+ },
1360
+ {
1361
+ "epoch": 67.53,
1362
+ "learning_rate": 5.054736842105263e-06,
1363
+ "loss": 0.0,
1364
+ "step": 5200
1365
+ },
1366
+ {
1367
+ "epoch": 67.53,
1368
+ "eval_loss": 0.7741473317146301,
1369
+ "eval_runtime": 20.4587,
1370
+ "eval_samples_per_second": 6.794,
1371
+ "eval_steps_per_second": 0.88,
1372
+ "eval_wer": 98.21002386634845,
1373
+ "step": 5200
1374
+ },
1375
+ {
1376
+ "epoch": 67.85,
1377
+ "learning_rate": 5.028421052631579e-06,
1378
+ "loss": 0.0,
1379
+ "step": 5225
1380
+ },
1381
+ {
1382
+ "epoch": 68.18,
1383
+ "learning_rate": 5.002105263157895e-06,
1384
+ "loss": 0.0,
1385
+ "step": 5250
1386
+ },
1387
+ {
1388
+ "epoch": 68.5,
1389
+ "learning_rate": 4.975789473684211e-06,
1390
+ "loss": 0.0,
1391
+ "step": 5275
1392
+ },
1393
+ {
1394
+ "epoch": 68.82,
1395
+ "learning_rate": 4.949473684210527e-06,
1396
+ "loss": 0.0,
1397
+ "step": 5300
1398
+ },
1399
+ {
1400
+ "epoch": 69.15,
1401
+ "learning_rate": 4.923157894736842e-06,
1402
+ "loss": 0.0,
1403
+ "step": 5325
1404
+ },
1405
+ {
1406
+ "epoch": 69.48,
1407
+ "learning_rate": 4.896842105263158e-06,
1408
+ "loss": 0.0,
1409
+ "step": 5350
1410
+ },
1411
+ {
1412
+ "epoch": 69.8,
1413
+ "learning_rate": 4.870526315789474e-06,
1414
+ "loss": 0.0,
1415
+ "step": 5375
1416
+ },
1417
+ {
1418
+ "epoch": 70.13,
1419
+ "learning_rate": 4.84421052631579e-06,
1420
+ "loss": 0.0,
1421
+ "step": 5400
1422
+ },
1423
+ {
1424
+ "epoch": 70.45,
1425
+ "learning_rate": 4.817894736842106e-06,
1426
+ "loss": 0.0,
1427
+ "step": 5425
1428
+ },
1429
+ {
1430
+ "epoch": 70.77,
1431
+ "learning_rate": 4.791578947368421e-06,
1432
+ "loss": 0.0,
1433
+ "step": 5450
1434
+ },
1435
+ {
1436
+ "epoch": 71.1,
1437
+ "learning_rate": 4.765263157894737e-06,
1438
+ "loss": 0.0,
1439
+ "step": 5475
1440
+ },
1441
+ {
1442
+ "epoch": 71.42,
1443
+ "learning_rate": 4.738947368421053e-06,
1444
+ "loss": 0.0,
1445
+ "step": 5500
1446
+ },
1447
+ {
1448
+ "epoch": 71.75,
1449
+ "learning_rate": 4.712631578947369e-06,
1450
+ "loss": 0.0,
1451
+ "step": 5525
1452
+ },
1453
+ {
1454
+ "epoch": 72.08,
1455
+ "learning_rate": 4.6863157894736845e-06,
1456
+ "loss": 0.0,
1457
+ "step": 5550
1458
+ },
1459
+ {
1460
+ "epoch": 72.4,
1461
+ "learning_rate": 4.66e-06,
1462
+ "loss": 0.0,
1463
+ "step": 5575
1464
+ },
1465
+ {
1466
+ "epoch": 72.72,
1467
+ "learning_rate": 4.633684210526316e-06,
1468
+ "loss": 0.0,
1469
+ "step": 5600
1470
+ },
1471
+ {
1472
+ "epoch": 72.72,
1473
+ "eval_loss": 0.7758468389511108,
1474
+ "eval_runtime": 19.2002,
1475
+ "eval_samples_per_second": 7.239,
1476
+ "eval_steps_per_second": 0.937,
1477
+ "eval_wer": 93.67541766109785,
1478
+ "step": 5600
1479
+ },
1480
+ {
1481
+ "epoch": 73.05,
1482
+ "learning_rate": 4.607368421052632e-06,
1483
+ "loss": 0.0,
1484
+ "step": 5625
1485
+ },
1486
+ {
1487
+ "epoch": 73.37,
1488
+ "learning_rate": 4.581052631578948e-06,
1489
+ "loss": 0.0,
1490
+ "step": 5650
1491
+ },
1492
+ {
1493
+ "epoch": 73.69,
1494
+ "learning_rate": 4.5547368421052636e-06,
1495
+ "loss": 0.0,
1496
+ "step": 5675
1497
+ },
1498
+ {
1499
+ "epoch": 74.03,
1500
+ "learning_rate": 4.528421052631579e-06,
1501
+ "loss": 0.0,
1502
+ "step": 5700
1503
+ },
1504
+ {
1505
+ "epoch": 74.35,
1506
+ "learning_rate": 4.502105263157895e-06,
1507
+ "loss": 0.0,
1508
+ "step": 5725
1509
+ },
1510
+ {
1511
+ "epoch": 74.67,
1512
+ "learning_rate": 4.475789473684211e-06,
1513
+ "loss": 0.0,
1514
+ "step": 5750
1515
+ },
1516
+ {
1517
+ "epoch": 74.99,
1518
+ "learning_rate": 4.449473684210527e-06,
1519
+ "loss": 0.0,
1520
+ "step": 5775
1521
+ },
1522
+ {
1523
+ "epoch": 75.32,
1524
+ "learning_rate": 4.423157894736843e-06,
1525
+ "loss": 0.0,
1526
+ "step": 5800
1527
+ },
1528
+ {
1529
+ "epoch": 75.64,
1530
+ "learning_rate": 4.3968421052631585e-06,
1531
+ "loss": 0.0,
1532
+ "step": 5825
1533
+ },
1534
+ {
1535
+ "epoch": 75.96,
1536
+ "learning_rate": 4.370526315789474e-06,
1537
+ "loss": 0.0,
1538
+ "step": 5850
1539
+ },
1540
+ {
1541
+ "epoch": 76.3,
1542
+ "learning_rate": 4.344210526315789e-06,
1543
+ "loss": 0.0,
1544
+ "step": 5875
1545
+ },
1546
+ {
1547
+ "epoch": 76.62,
1548
+ "learning_rate": 4.317894736842105e-06,
1549
+ "loss": 0.0,
1550
+ "step": 5900
1551
+ },
1552
+ {
1553
+ "epoch": 76.94,
1554
+ "learning_rate": 4.291578947368422e-06,
1555
+ "loss": 0.0,
1556
+ "step": 5925
1557
+ },
1558
+ {
1559
+ "epoch": 77.27,
1560
+ "learning_rate": 4.2652631578947376e-06,
1561
+ "loss": 0.0,
1562
+ "step": 5950
1563
+ },
1564
+ {
1565
+ "epoch": 77.59,
1566
+ "learning_rate": 4.238947368421053e-06,
1567
+ "loss": 0.0,
1568
+ "step": 5975
1569
+ },
1570
+ {
1571
+ "epoch": 77.91,
1572
+ "learning_rate": 4.212631578947368e-06,
1573
+ "loss": 0.0,
1574
+ "step": 6000
1575
+ },
1576
+ {
1577
+ "epoch": 77.91,
1578
+ "eval_loss": 0.7771462202072144,
1579
+ "eval_runtime": 19.1772,
1580
+ "eval_samples_per_second": 7.248,
1581
+ "eval_steps_per_second": 0.939,
1582
+ "eval_wer": 93.07875894988067,
1583
+ "step": 6000
1584
+ },
1585
+ {
1586
+ "epoch": 78.24,
1587
+ "learning_rate": 4.186315789473684e-06,
1588
+ "loss": 0.0,
1589
+ "step": 6025
1590
+ },
1591
+ {
1592
+ "epoch": 78.57,
1593
+ "learning_rate": 4.16e-06,
1594
+ "loss": 0.0,
1595
+ "step": 6050
1596
+ },
1597
+ {
1598
+ "epoch": 78.89,
1599
+ "learning_rate": 4.133684210526316e-06,
1600
+ "loss": 0.0,
1601
+ "step": 6075
1602
+ },
1603
+ {
1604
+ "epoch": 79.22,
1605
+ "learning_rate": 4.1073684210526325e-06,
1606
+ "loss": 0.0,
1607
+ "step": 6100
1608
+ },
1609
+ {
1610
+ "epoch": 79.54,
1611
+ "learning_rate": 4.0810526315789474e-06,
1612
+ "loss": 0.0,
1613
+ "step": 6125
1614
+ },
1615
+ {
1616
+ "epoch": 79.86,
1617
+ "learning_rate": 4.054736842105263e-06,
1618
+ "loss": 0.0,
1619
+ "step": 6150
1620
+ },
1621
+ {
1622
+ "epoch": 80.19,
1623
+ "learning_rate": 4.028421052631579e-06,
1624
+ "loss": 0.0,
1625
+ "step": 6175
1626
+ },
1627
+ {
1628
+ "epoch": 80.51,
1629
+ "learning_rate": 4.002105263157895e-06,
1630
+ "loss": 0.0,
1631
+ "step": 6200
1632
+ },
1633
+ {
1634
+ "epoch": 80.84,
1635
+ "learning_rate": 3.975789473684211e-06,
1636
+ "loss": 0.0,
1637
+ "step": 6225
1638
+ },
1639
+ {
1640
+ "epoch": 81.17,
1641
+ "learning_rate": 3.9494736842105265e-06,
1642
+ "loss": 0.0,
1643
+ "step": 6250
1644
+ },
1645
+ {
1646
+ "epoch": 81.49,
1647
+ "learning_rate": 3.923157894736842e-06,
1648
+ "loss": 0.0,
1649
+ "step": 6275
1650
+ },
1651
+ {
1652
+ "epoch": 81.81,
1653
+ "learning_rate": 3.896842105263158e-06,
1654
+ "loss": 0.0,
1655
+ "step": 6300
1656
+ },
1657
+ {
1658
+ "epoch": 82.14,
1659
+ "learning_rate": 3.870526315789474e-06,
1660
+ "loss": 0.0,
1661
+ "step": 6325
1662
+ },
1663
+ {
1664
+ "epoch": 82.46,
1665
+ "learning_rate": 3.84421052631579e-06,
1666
+ "loss": 0.0,
1667
+ "step": 6350
1668
+ },
1669
+ {
1670
+ "epoch": 82.78,
1671
+ "learning_rate": 3.817894736842106e-06,
1672
+ "loss": 0.0,
1673
+ "step": 6375
1674
+ },
1675
+ {
1676
+ "epoch": 83.12,
1677
+ "learning_rate": 3.791578947368421e-06,
1678
+ "loss": 0.0,
1679
+ "step": 6400
1680
+ },
1681
+ {
1682
+ "epoch": 83.12,
1683
+ "eval_loss": 0.7782990336418152,
1684
+ "eval_runtime": 20.1685,
1685
+ "eval_samples_per_second": 6.892,
1686
+ "eval_steps_per_second": 0.892,
1687
+ "eval_wer": 98.21002386634845,
1688
+ "step": 6400
1689
+ },
1690
+ {
1691
+ "epoch": 83.44,
1692
+ "learning_rate": 3.7652631578947373e-06,
1693
+ "loss": 0.0,
1694
+ "step": 6425
1695
+ },
1696
+ {
1697
+ "epoch": 83.76,
1698
+ "learning_rate": 3.738947368421053e-06,
1699
+ "loss": 0.0,
1700
+ "step": 6450
1701
+ },
1702
+ {
1703
+ "epoch": 84.09,
1704
+ "learning_rate": 3.712631578947369e-06,
1705
+ "loss": 0.0,
1706
+ "step": 6475
1707
+ },
1708
+ {
1709
+ "epoch": 84.41,
1710
+ "learning_rate": 3.6863157894736847e-06,
1711
+ "loss": 0.0,
1712
+ "step": 6500
1713
+ },
1714
+ {
1715
+ "epoch": 84.73,
1716
+ "learning_rate": 3.66e-06,
1717
+ "loss": 0.0,
1718
+ "step": 6525
1719
+ },
1720
+ {
1721
+ "epoch": 85.06,
1722
+ "learning_rate": 3.633684210526316e-06,
1723
+ "loss": 0.0,
1724
+ "step": 6550
1725
+ },
1726
+ {
1727
+ "epoch": 85.39,
1728
+ "learning_rate": 3.6073684210526317e-06,
1729
+ "loss": 0.0,
1730
+ "step": 6575
1731
+ },
1732
+ {
1733
+ "epoch": 85.71,
1734
+ "learning_rate": 3.581052631578948e-06,
1735
+ "loss": 0.0,
1736
+ "step": 6600
1737
+ },
1738
+ {
1739
+ "epoch": 86.04,
1740
+ "learning_rate": 3.554736842105264e-06,
1741
+ "loss": 0.0,
1742
+ "step": 6625
1743
+ },
1744
+ {
1745
+ "epoch": 86.36,
1746
+ "learning_rate": 3.528421052631579e-06,
1747
+ "loss": 0.0,
1748
+ "step": 6650
1749
+ },
1750
+ {
1751
+ "epoch": 86.68,
1752
+ "learning_rate": 3.502105263157895e-06,
1753
+ "loss": 0.0,
1754
+ "step": 6675
1755
+ },
1756
+ {
1757
+ "epoch": 87.01,
1758
+ "learning_rate": 3.475789473684211e-06,
1759
+ "loss": 0.0,
1760
+ "step": 6700
1761
+ },
1762
+ {
1763
+ "epoch": 87.33,
1764
+ "learning_rate": 3.4494736842105262e-06,
1765
+ "loss": 0.0,
1766
+ "step": 6725
1767
+ },
1768
+ {
1769
+ "epoch": 87.66,
1770
+ "learning_rate": 3.423157894736842e-06,
1771
+ "loss": 0.0,
1772
+ "step": 6750
1773
+ },
1774
+ {
1775
+ "epoch": 87.98,
1776
+ "learning_rate": 3.3968421052631583e-06,
1777
+ "loss": 0.0,
1778
+ "step": 6775
1779
+ },
1780
+ {
1781
+ "epoch": 88.31,
1782
+ "learning_rate": 3.370526315789474e-06,
1783
+ "loss": 0.0,
1784
+ "step": 6800
1785
+ },
1786
+ {
1787
+ "epoch": 88.31,
1788
+ "eval_loss": 0.7796774506568909,
1789
+ "eval_runtime": 20.1982,
1790
+ "eval_samples_per_second": 6.882,
1791
+ "eval_steps_per_second": 0.891,
1792
+ "eval_wer": 98.80668257756562,
1793
+ "step": 6800
1794
+ },
1795
+ {
1796
+ "epoch": 88.63,
1797
+ "learning_rate": 3.34421052631579e-06,
1798
+ "loss": 0.0,
1799
+ "step": 6825
1800
+ },
1801
+ {
1802
+ "epoch": 88.95,
1803
+ "learning_rate": 3.3178947368421053e-06,
1804
+ "loss": 0.0,
1805
+ "step": 6850
1806
+ },
1807
+ {
1808
+ "epoch": 89.28,
1809
+ "learning_rate": 3.291578947368421e-06,
1810
+ "loss": 0.0,
1811
+ "step": 6875
1812
+ },
1813
+ {
1814
+ "epoch": 89.6,
1815
+ "learning_rate": 3.265263157894737e-06,
1816
+ "loss": 0.0,
1817
+ "step": 6900
1818
+ },
1819
+ {
1820
+ "epoch": 89.93,
1821
+ "learning_rate": 3.238947368421053e-06,
1822
+ "loss": 0.0,
1823
+ "step": 6925
1824
+ },
1825
+ {
1826
+ "epoch": 90.26,
1827
+ "learning_rate": 3.212631578947369e-06,
1828
+ "loss": 0.0,
1829
+ "step": 6950
1830
+ },
1831
+ {
1832
+ "epoch": 90.58,
1833
+ "learning_rate": 3.1863157894736844e-06,
1834
+ "loss": 0.0,
1835
+ "step": 6975
1836
+ },
1837
+ {
1838
+ "epoch": 90.9,
1839
+ "learning_rate": 3.1600000000000002e-06,
1840
+ "loss": 0.0,
1841
+ "step": 7000
1842
+ },
1843
+ {
1844
+ "epoch": 91.23,
1845
+ "learning_rate": 3.133684210526316e-06,
1846
+ "loss": 0.0,
1847
+ "step": 7025
1848
+ },
1849
+ {
1850
+ "epoch": 91.55,
1851
+ "learning_rate": 3.107368421052632e-06,
1852
+ "loss": 0.0,
1853
+ "step": 7050
1854
+ },
1855
+ {
1856
+ "epoch": 91.87,
1857
+ "learning_rate": 3.0810526315789473e-06,
1858
+ "loss": 0.0,
1859
+ "step": 7075
1860
+ },
1861
+ {
1862
+ "epoch": 92.21,
1863
+ "learning_rate": 3.0547368421052635e-06,
1864
+ "loss": 0.0,
1865
+ "step": 7100
1866
+ },
1867
+ {
1868
+ "epoch": 92.53,
1869
+ "learning_rate": 3.0284210526315793e-06,
1870
+ "loss": 0.0,
1871
+ "step": 7125
1872
+ },
1873
+ {
1874
+ "epoch": 92.85,
1875
+ "learning_rate": 3.002105263157895e-06,
1876
+ "loss": 0.0,
1877
+ "step": 7150
1878
+ },
1879
+ {
1880
+ "epoch": 93.18,
1881
+ "learning_rate": 2.975789473684211e-06,
1882
+ "loss": 0.0,
1883
+ "step": 7175
1884
+ },
1885
+ {
1886
+ "epoch": 93.5,
1887
+ "learning_rate": 2.9494736842105264e-06,
1888
+ "loss": 0.0,
1889
+ "step": 7200
1890
+ },
1891
+ {
1892
+ "epoch": 93.5,
1893
+ "eval_loss": 0.781008243560791,
1894
+ "eval_runtime": 20.1464,
1895
+ "eval_samples_per_second": 6.899,
1896
+ "eval_steps_per_second": 0.893,
1897
+ "eval_wer": 98.80668257756562,
1898
+ "step": 7200
1899
+ },
1900
+ {
1901
+ "epoch": 93.82,
1902
+ "learning_rate": 2.923157894736842e-06,
1903
+ "loss": 0.0,
1904
+ "step": 7225
1905
+ },
1906
+ {
1907
+ "epoch": 94.15,
1908
+ "learning_rate": 2.896842105263158e-06,
1909
+ "loss": 0.0,
1910
+ "step": 7250
1911
+ },
1912
+ {
1913
+ "epoch": 94.48,
1914
+ "learning_rate": 2.8705263157894742e-06,
1915
+ "loss": 0.0,
1916
+ "step": 7275
1917
+ },
1918
+ {
1919
+ "epoch": 94.8,
1920
+ "learning_rate": 2.84421052631579e-06,
1921
+ "loss": 0.0,
1922
+ "step": 7300
1923
+ },
1924
+ {
1925
+ "epoch": 95.13,
1926
+ "learning_rate": 2.8178947368421055e-06,
1927
+ "loss": 0.0,
1928
+ "step": 7325
1929
+ },
1930
+ {
1931
+ "epoch": 95.45,
1932
+ "learning_rate": 2.7915789473684213e-06,
1933
+ "loss": 0.0,
1934
+ "step": 7350
1935
+ },
1936
+ {
1937
+ "epoch": 95.77,
1938
+ "learning_rate": 2.765263157894737e-06,
1939
+ "loss": 0.0,
1940
+ "step": 7375
1941
+ },
1942
+ {
1943
+ "epoch": 96.1,
1944
+ "learning_rate": 2.7389473684210525e-06,
1945
+ "loss": 0.0,
1946
+ "step": 7400
1947
+ },
1948
+ {
1949
+ "epoch": 96.42,
1950
+ "learning_rate": 2.7126315789473683e-06,
1951
+ "loss": 0.0,
1952
+ "step": 7425
1953
+ },
1954
+ {
1955
+ "epoch": 96.75,
1956
+ "learning_rate": 2.6863157894736845e-06,
1957
+ "loss": 0.0,
1958
+ "step": 7450
1959
+ },
1960
+ {
1961
+ "epoch": 97.08,
1962
+ "learning_rate": 2.6600000000000004e-06,
1963
+ "loss": 0.0,
1964
+ "step": 7475
1965
+ },
1966
+ {
1967
+ "epoch": 97.4,
1968
+ "learning_rate": 2.633684210526316e-06,
1969
+ "loss": 0.0,
1970
+ "step": 7500
1971
+ },
1972
+ {
1973
+ "epoch": 97.72,
1974
+ "learning_rate": 2.6073684210526316e-06,
1975
+ "loss": 0.0,
1976
+ "step": 7525
1977
+ },
1978
+ {
1979
+ "epoch": 98.05,
1980
+ "learning_rate": 2.5810526315789474e-06,
1981
+ "loss": 0.0,
1982
+ "step": 7550
1983
+ },
1984
+ {
1985
+ "epoch": 98.37,
1986
+ "learning_rate": 2.5547368421052632e-06,
1987
+ "loss": 0.0,
1988
+ "step": 7575
1989
+ },
1990
+ {
1991
+ "epoch": 98.69,
1992
+ "learning_rate": 2.528421052631579e-06,
1993
+ "loss": 0.0,
1994
+ "step": 7600
1995
+ },
1996
+ {
1997
+ "epoch": 98.69,
1998
+ "eval_loss": 0.7822578549385071,
1999
+ "eval_runtime": 20.1798,
2000
+ "eval_samples_per_second": 6.888,
2001
+ "eval_steps_per_second": 0.892,
2002
+ "eval_wer": 98.80668257756562,
2003
+ "step": 7600
2004
+ },
2005
+ {
2006
+ "epoch": 99.03,
2007
+ "learning_rate": 2.5021052631578953e-06,
2008
+ "loss": 0.0,
2009
+ "step": 7625
2010
+ },
2011
+ {
2012
+ "epoch": 99.35,
2013
+ "learning_rate": 2.4757894736842107e-06,
2014
+ "loss": 0.0,
2015
+ "step": 7650
2016
+ },
2017
+ {
2018
+ "epoch": 99.67,
2019
+ "learning_rate": 2.4494736842105265e-06,
2020
+ "loss": 0.0,
2021
+ "step": 7675
2022
+ },
2023
+ {
2024
+ "epoch": 99.99,
2025
+ "learning_rate": 2.4231578947368423e-06,
2026
+ "loss": 0.0,
2027
+ "step": 7700
2028
+ },
2029
+ {
2030
+ "epoch": 100.32,
2031
+ "learning_rate": 2.396842105263158e-06,
2032
+ "loss": 0.0,
2033
+ "step": 7725
2034
+ },
2035
+ {
2036
+ "epoch": 100.64,
2037
+ "learning_rate": 2.370526315789474e-06,
2038
+ "loss": 0.0,
2039
+ "step": 7750
2040
+ },
2041
+ {
2042
+ "epoch": 100.96,
2043
+ "learning_rate": 2.3442105263157898e-06,
2044
+ "loss": 0.0,
2045
+ "step": 7775
2046
+ },
2047
+ {
2048
+ "epoch": 101.3,
2049
+ "learning_rate": 2.317894736842105e-06,
2050
+ "loss": 0.0,
2051
+ "step": 7800
2052
+ },
2053
+ {
2054
+ "epoch": 101.62,
2055
+ "learning_rate": 2.2915789473684214e-06,
2056
+ "loss": 0.0,
2057
+ "step": 7825
2058
+ },
2059
+ {
2060
+ "epoch": 101.94,
2061
+ "learning_rate": 2.2652631578947372e-06,
2062
+ "loss": 0.0,
2063
+ "step": 7850
2064
+ },
2065
+ {
2066
+ "epoch": 102.27,
2067
+ "learning_rate": 2.2389473684210526e-06,
2068
+ "loss": 0.0,
2069
+ "step": 7875
2070
+ },
2071
+ {
2072
+ "epoch": 102.59,
2073
+ "learning_rate": 2.2126315789473684e-06,
2074
+ "loss": 0.0,
2075
+ "step": 7900
2076
+ },
2077
+ {
2078
+ "epoch": 102.91,
2079
+ "learning_rate": 2.1863157894736843e-06,
2080
+ "loss": 0.0,
2081
+ "step": 7925
2082
+ },
2083
+ {
2084
+ "epoch": 103.24,
2085
+ "learning_rate": 2.16e-06,
2086
+ "loss": 0.0,
2087
+ "step": 7950
2088
+ },
2089
+ {
2090
+ "epoch": 103.57,
2091
+ "learning_rate": 2.133684210526316e-06,
2092
+ "loss": 0.0,
2093
+ "step": 7975
2094
+ },
2095
+ {
2096
+ "epoch": 103.89,
2097
+ "learning_rate": 2.1073684210526317e-06,
2098
+ "loss": 0.0,
2099
+ "step": 8000
2100
+ },
2101
+ {
2102
+ "epoch": 103.89,
2103
+ "eval_loss": 0.7835149168968201,
2104
+ "eval_runtime": 20.2001,
2105
+ "eval_samples_per_second": 6.881,
2106
+ "eval_steps_per_second": 0.891,
2107
+ "eval_wer": 98.68735083532219,
2108
+ "step": 8000
2109
+ },
2110
+ {
2111
+ "epoch": 104.22,
2112
+ "learning_rate": 2.0810526315789475e-06,
2113
+ "loss": 0.0,
2114
+ "step": 8025
2115
+ },
2116
+ {
2117
+ "epoch": 104.54,
2118
+ "learning_rate": 2.0547368421052633e-06,
2119
+ "loss": 0.0,
2120
+ "step": 8050
2121
+ },
2122
+ {
2123
+ "epoch": 104.86,
2124
+ "learning_rate": 2.028421052631579e-06,
2125
+ "loss": 0.0,
2126
+ "step": 8075
2127
+ },
2128
+ {
2129
+ "epoch": 105.19,
2130
+ "learning_rate": 2.002105263157895e-06,
2131
+ "loss": 0.0,
2132
+ "step": 8100
2133
+ },
2134
+ {
2135
+ "epoch": 105.51,
2136
+ "learning_rate": 1.975789473684211e-06,
2137
+ "loss": 0.0,
2138
+ "step": 8125
2139
+ },
2140
+ {
2141
+ "epoch": 105.84,
2142
+ "learning_rate": 1.949473684210526e-06,
2143
+ "loss": 0.0,
2144
+ "step": 8150
2145
+ },
2146
+ {
2147
+ "epoch": 106.17,
2148
+ "learning_rate": 1.9231578947368424e-06,
2149
+ "loss": 0.0,
2150
+ "step": 8175
2151
+ },
2152
+ {
2153
+ "epoch": 106.49,
2154
+ "learning_rate": 1.896842105263158e-06,
2155
+ "loss": 0.0,
2156
+ "step": 8200
2157
+ },
2158
+ {
2159
+ "epoch": 106.81,
2160
+ "learning_rate": 1.8705263157894737e-06,
2161
+ "loss": 0.0,
2162
+ "step": 8225
2163
+ },
2164
+ {
2165
+ "epoch": 107.14,
2166
+ "learning_rate": 1.8442105263157897e-06,
2167
+ "loss": 0.0,
2168
+ "step": 8250
2169
+ },
2170
+ {
2171
+ "epoch": 107.46,
2172
+ "learning_rate": 1.8178947368421055e-06,
2173
+ "loss": 0.0,
2174
+ "step": 8275
2175
+ },
2176
+ {
2177
+ "epoch": 107.78,
2178
+ "learning_rate": 1.7915789473684211e-06,
2179
+ "loss": 0.0,
2180
+ "step": 8300
2181
+ },
2182
+ {
2183
+ "epoch": 108.12,
2184
+ "learning_rate": 1.7652631578947371e-06,
2185
+ "loss": 0.0,
2186
+ "step": 8325
2187
+ },
2188
+ {
2189
+ "epoch": 108.44,
2190
+ "learning_rate": 1.7389473684210527e-06,
2191
+ "loss": 0.0,
2192
+ "step": 8350
2193
+ },
2194
+ {
2195
+ "epoch": 108.76,
2196
+ "learning_rate": 1.7126315789473686e-06,
2197
+ "loss": 0.0,
2198
+ "step": 8375
2199
+ },
2200
+ {
2201
+ "epoch": 109.09,
2202
+ "learning_rate": 1.6863157894736842e-06,
2203
+ "loss": 0.0,
2204
+ "step": 8400
2205
+ },
2206
+ {
2207
+ "epoch": 109.09,
2208
+ "eval_loss": 0.7845548987388611,
2209
+ "eval_runtime": 20.053,
2210
+ "eval_samples_per_second": 6.932,
2211
+ "eval_steps_per_second": 0.898,
2212
+ "eval_wer": 99.16467780429595,
2213
+ "step": 8400
2214
+ },
2215
+ {
2216
+ "epoch": 109.41,
2217
+ "learning_rate": 1.6600000000000002e-06,
2218
+ "loss": 0.0,
2219
+ "step": 8425
2220
+ },
2221
+ {
2222
+ "epoch": 109.73,
2223
+ "learning_rate": 1.633684210526316e-06,
2224
+ "loss": 0.0,
2225
+ "step": 8450
2226
+ },
2227
+ {
2228
+ "epoch": 110.06,
2229
+ "learning_rate": 1.6073684210526316e-06,
2230
+ "loss": 0.0,
2231
+ "step": 8475
2232
+ },
2233
+ {
2234
+ "epoch": 110.39,
2235
+ "learning_rate": 1.5810526315789477e-06,
2236
+ "loss": 0.0,
2237
+ "step": 8500
2238
+ },
2239
+ {
2240
+ "epoch": 110.71,
2241
+ "learning_rate": 1.5547368421052633e-06,
2242
+ "loss": 0.0,
2243
+ "step": 8525
2244
+ },
2245
+ {
2246
+ "epoch": 111.04,
2247
+ "learning_rate": 1.528421052631579e-06,
2248
+ "loss": 0.0,
2249
+ "step": 8550
2250
+ },
2251
+ {
2252
+ "epoch": 111.36,
2253
+ "learning_rate": 1.5021052631578947e-06,
2254
+ "loss": 0.0,
2255
+ "step": 8575
2256
+ },
2257
+ {
2258
+ "epoch": 111.68,
2259
+ "learning_rate": 1.4757894736842107e-06,
2260
+ "loss": 0.0,
2261
+ "step": 8600
2262
+ },
2263
+ {
2264
+ "epoch": 112.01,
2265
+ "learning_rate": 1.4494736842105263e-06,
2266
+ "loss": 0.0,
2267
+ "step": 8625
2268
+ },
2269
+ {
2270
+ "epoch": 112.33,
2271
+ "learning_rate": 1.4231578947368421e-06,
2272
+ "loss": 0.0,
2273
+ "step": 8650
2274
+ },
2275
+ {
2276
+ "epoch": 112.66,
2277
+ "learning_rate": 1.3968421052631582e-06,
2278
+ "loss": 0.0,
2279
+ "step": 8675
2280
+ },
2281
+ {
2282
+ "epoch": 112.98,
2283
+ "learning_rate": 1.3705263157894738e-06,
2284
+ "loss": 0.0,
2285
+ "step": 8700
2286
+ },
2287
+ {
2288
+ "epoch": 113.31,
2289
+ "learning_rate": 1.3442105263157896e-06,
2290
+ "loss": 0.0,
2291
+ "step": 8725
2292
+ },
2293
+ {
2294
+ "epoch": 113.63,
2295
+ "learning_rate": 1.3178947368421054e-06,
2296
+ "loss": 0.0,
2297
+ "step": 8750
2298
+ },
2299
+ {
2300
+ "epoch": 113.95,
2301
+ "learning_rate": 1.2915789473684212e-06,
2302
+ "loss": 0.0,
2303
+ "step": 8775
2304
+ },
2305
+ {
2306
+ "epoch": 114.28,
2307
+ "learning_rate": 1.2652631578947368e-06,
2308
+ "loss": 0.0,
2309
+ "step": 8800
2310
+ },
2311
+ {
2312
+ "epoch": 114.28,
2313
+ "eval_loss": 0.7855232357978821,
2314
+ "eval_runtime": 20.0942,
2315
+ "eval_samples_per_second": 6.917,
2316
+ "eval_steps_per_second": 0.896,
2317
+ "eval_wer": 99.40334128878283,
2318
+ "step": 8800
2319
+ },
2320
+ {
2321
+ "epoch": 114.6,
2322
+ "learning_rate": 1.2389473684210527e-06,
2323
+ "loss": 0.0,
2324
+ "step": 8825
2325
+ },
2326
+ {
2327
+ "epoch": 114.93,
2328
+ "learning_rate": 1.2126315789473685e-06,
2329
+ "loss": 0.0,
2330
+ "step": 8850
2331
+ },
2332
+ {
2333
+ "epoch": 115.26,
2334
+ "learning_rate": 1.1863157894736843e-06,
2335
+ "loss": 0.0,
2336
+ "step": 8875
2337
+ },
2338
+ {
2339
+ "epoch": 115.58,
2340
+ "learning_rate": 1.1600000000000001e-06,
2341
+ "loss": 0.0,
2342
+ "step": 8900
2343
+ },
2344
+ {
2345
+ "epoch": 115.9,
2346
+ "learning_rate": 1.133684210526316e-06,
2347
+ "loss": 0.0,
2348
+ "step": 8925
2349
+ },
2350
+ {
2351
+ "epoch": 116.23,
2352
+ "learning_rate": 1.1073684210526318e-06,
2353
+ "loss": 0.0,
2354
+ "step": 8950
2355
+ },
2356
+ {
2357
+ "epoch": 116.55,
2358
+ "learning_rate": 1.0810526315789474e-06,
2359
+ "loss": 0.0,
2360
+ "step": 8975
2361
+ },
2362
+ {
2363
+ "epoch": 116.87,
2364
+ "learning_rate": 1.0547368421052632e-06,
2365
+ "loss": 0.0,
2366
+ "step": 9000
2367
+ },
2368
+ {
2369
+ "epoch": 117.21,
2370
+ "learning_rate": 1.028421052631579e-06,
2371
+ "loss": 0.0,
2372
+ "step": 9025
2373
+ },
2374
+ {
2375
+ "epoch": 117.53,
2376
+ "learning_rate": 1.0021052631578948e-06,
2377
+ "loss": 0.0,
2378
+ "step": 9050
2379
+ },
2380
+ {
2381
+ "epoch": 117.85,
2382
+ "learning_rate": 9.757894736842106e-07,
2383
+ "loss": 0.0,
2384
+ "step": 9075
2385
+ },
2386
+ {
2387
+ "epoch": 118.18,
2388
+ "learning_rate": 9.494736842105263e-07,
2389
+ "loss": 0.0,
2390
+ "step": 9100
2391
+ },
2392
+ {
2393
+ "epoch": 118.5,
2394
+ "learning_rate": 9.231578947368422e-07,
2395
+ "loss": 0.0,
2396
+ "step": 9125
2397
+ },
2398
+ {
2399
+ "epoch": 118.82,
2400
+ "learning_rate": 8.968421052631579e-07,
2401
+ "loss": 0.0,
2402
+ "step": 9150
2403
+ },
2404
+ {
2405
+ "epoch": 119.15,
2406
+ "learning_rate": 8.705263157894737e-07,
2407
+ "loss": 0.0,
2408
+ "step": 9175
2409
+ },
2410
+ {
2411
+ "epoch": 119.48,
2412
+ "learning_rate": 8.442105263157896e-07,
2413
+ "loss": 0.0,
2414
+ "step": 9200
2415
+ },
2416
+ {
2417
+ "epoch": 119.48,
2418
+ "eval_loss": 0.786337673664093,
2419
+ "eval_runtime": 19.9619,
2420
+ "eval_samples_per_second": 6.963,
2421
+ "eval_steps_per_second": 0.902,
2422
+ "eval_wer": 99.40334128878283,
2423
+ "step": 9200
2424
+ },
2425
+ {
2426
+ "epoch": 119.8,
2427
+ "learning_rate": 8.178947368421053e-07,
2428
+ "loss": 0.0,
2429
+ "step": 9225
2430
+ },
2431
+ {
2432
+ "epoch": 120.13,
2433
+ "learning_rate": 7.915789473684212e-07,
2434
+ "loss": 0.0,
2435
+ "step": 9250
2436
+ },
2437
+ {
2438
+ "epoch": 120.45,
2439
+ "learning_rate": 7.652631578947369e-07,
2440
+ "loss": 0.0,
2441
+ "step": 9275
2442
+ },
2443
+ {
2444
+ "epoch": 120.77,
2445
+ "learning_rate": 7.389473684210527e-07,
2446
+ "loss": 0.0,
2447
+ "step": 9300
2448
+ },
2449
+ {
2450
+ "epoch": 121.1,
2451
+ "learning_rate": 7.126315789473685e-07,
2452
+ "loss": 0.0,
2453
+ "step": 9325
2454
+ },
2455
+ {
2456
+ "epoch": 121.42,
2457
+ "learning_rate": 6.863157894736842e-07,
2458
+ "loss": 0.0,
2459
+ "step": 9350
2460
+ },
2461
+ {
2462
+ "epoch": 121.75,
2463
+ "learning_rate": 6.6e-07,
2464
+ "loss": 0.0,
2465
+ "step": 9375
2466
+ },
2467
+ {
2468
+ "epoch": 122.08,
2469
+ "learning_rate": 6.336842105263157e-07,
2470
+ "loss": 0.0,
2471
+ "step": 9400
2472
+ },
2473
+ {
2474
+ "epoch": 122.4,
2475
+ "learning_rate": 6.073684210526317e-07,
2476
+ "loss": 0.0,
2477
+ "step": 9425
2478
+ },
2479
+ {
2480
+ "epoch": 122.72,
2481
+ "learning_rate": 5.810526315789474e-07,
2482
+ "loss": 0.0,
2483
+ "step": 9450
2484
+ },
2485
+ {
2486
+ "epoch": 123.05,
2487
+ "learning_rate": 5.547368421052632e-07,
2488
+ "loss": 0.0,
2489
+ "step": 9475
2490
+ },
2491
+ {
2492
+ "epoch": 123.37,
2493
+ "learning_rate": 5.284210526315789e-07,
2494
+ "loss": 0.0,
2495
+ "step": 9500
2496
+ },
2497
+ {
2498
+ "epoch": 123.69,
2499
+ "learning_rate": 5.021052631578948e-07,
2500
+ "loss": 0.0,
2501
+ "step": 9525
2502
+ },
2503
+ {
2504
+ "epoch": 124.03,
2505
+ "learning_rate": 4.757894736842106e-07,
2506
+ "loss": 0.0,
2507
+ "step": 9550
2508
+ },
2509
+ {
2510
+ "epoch": 124.35,
2511
+ "learning_rate": 4.4947368421052637e-07,
2512
+ "loss": 0.0,
2513
+ "step": 9575
2514
+ },
2515
+ {
2516
+ "epoch": 124.67,
2517
+ "learning_rate": 4.2315789473684214e-07,
2518
+ "loss": 0.0,
2519
+ "step": 9600
2520
+ },
2521
+ {
2522
+ "epoch": 124.67,
2523
+ "eval_loss": 0.786903977394104,
2524
+ "eval_runtime": 19.962,
2525
+ "eval_samples_per_second": 6.963,
2526
+ "eval_steps_per_second": 0.902,
2527
+ "eval_wer": 99.40334128878283,
2528
+ "step": 9600
2529
+ },
2530
+ {
2531
+ "epoch": 124.99,
2532
+ "learning_rate": 3.968421052631579e-07,
2533
+ "loss": 0.0,
2534
+ "step": 9625
2535
+ },
2536
+ {
2537
+ "epoch": 125.32,
2538
+ "learning_rate": 3.7052631578947377e-07,
2539
+ "loss": 0.0,
2540
+ "step": 9650
2541
+ },
2542
+ {
2543
+ "epoch": 125.64,
2544
+ "learning_rate": 3.4421052631578954e-07,
2545
+ "loss": 0.0,
2546
+ "step": 9675
2547
+ },
2548
+ {
2549
+ "epoch": 125.96,
2550
+ "learning_rate": 3.178947368421053e-07,
2551
+ "loss": 0.0,
2552
+ "step": 9700
2553
+ },
2554
+ {
2555
+ "epoch": 126.3,
2556
+ "learning_rate": 2.9157894736842107e-07,
2557
+ "loss": 0.0,
2558
+ "step": 9725
2559
+ },
2560
+ {
2561
+ "epoch": 126.62,
2562
+ "learning_rate": 2.6526315789473684e-07,
2563
+ "loss": 0.0,
2564
+ "step": 9750
2565
+ },
2566
+ {
2567
+ "epoch": 126.94,
2568
+ "learning_rate": 2.3894736842105265e-07,
2569
+ "loss": 0.0,
2570
+ "step": 9775
2571
+ },
2572
+ {
2573
+ "epoch": 127.27,
2574
+ "learning_rate": 2.1263157894736842e-07,
2575
+ "loss": 0.0,
2576
+ "step": 9800
2577
+ },
2578
+ {
2579
+ "epoch": 127.59,
2580
+ "learning_rate": 1.8631578947368424e-07,
2581
+ "loss": 0.0,
2582
+ "step": 9825
2583
+ },
2584
+ {
2585
+ "epoch": 127.91,
2586
+ "learning_rate": 1.6e-07,
2587
+ "loss": 0.0,
2588
+ "step": 9850
2589
+ },
2590
+ {
2591
+ "epoch": 128.24,
2592
+ "learning_rate": 1.3368421052631582e-07,
2593
+ "loss": 0.0,
2594
+ "step": 9875
2595
+ },
2596
+ {
2597
+ "epoch": 128.57,
2598
+ "learning_rate": 1.0736842105263159e-07,
2599
+ "loss": 0.0,
2600
+ "step": 9900
2601
+ },
2602
+ {
2603
+ "epoch": 128.89,
2604
+ "learning_rate": 8.105263157894738e-08,
2605
+ "loss": 0.0,
2606
+ "step": 9925
2607
+ },
2608
+ {
2609
+ "epoch": 129.22,
2610
+ "learning_rate": 5.473684210526316e-08,
2611
+ "loss": 0.0,
2612
+ "step": 9950
2613
+ },
2614
+ {
2615
+ "epoch": 129.54,
2616
+ "learning_rate": 2.842105263157895e-08,
2617
+ "loss": 0.0,
2618
+ "step": 9975
2619
+ },
2620
+ {
2621
+ "epoch": 129.86,
2622
+ "learning_rate": 2.105263157894737e-09,
2623
+ "loss": 0.0,
2624
+ "step": 10000
2625
+ },
2626
+ {
2627
+ "epoch": 129.86,
2628
+ "eval_loss": 0.7871208786964417,
2629
+ "eval_runtime": 19.9495,
2630
+ "eval_samples_per_second": 6.968,
2631
+ "eval_steps_per_second": 0.902,
2632
+ "eval_wer": 99.40334128878283,
2633
+ "step": 10000
2634
+ }
2635
+ ],
2636
+ "max_steps": 10000,
2637
+ "num_train_epochs": 130,
2638
+ "total_flos": 1.6487668160004096e+20,
2639
+ "trial_name": null,
2640
+ "trial_params": null
2641
+ }
checkpoint-10000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4ad917c70902fd83abc5a9152e70fd07531fce940ccd66f6af7f3bc05e82880
3
+ size 4088
checkpoint-1200/config.json ADDED
@@ -0,0 +1,144 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "openai/whisper-medium",
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "gelu",
5
+ "architectures": [
6
+ "WhisperForConditionalGeneration"
7
+ ],
8
+ "attention_dropout": 0.0,
9
+ "begin_suppress_tokens": [
10
+ 220,
11
+ 50257
12
+ ],
13
+ "bos_token_id": 50257,
14
+ "d_model": 1024,
15
+ "decoder_attention_heads": 16,
16
+ "decoder_ffn_dim": 4096,
17
+ "decoder_layerdrop": 0.0,
18
+ "decoder_layers": 24,
19
+ "decoder_start_token_id": 50258,
20
+ "dropout": 0.0,
21
+ "encoder_attention_heads": 16,
22
+ "encoder_ffn_dim": 4096,
23
+ "encoder_layerdrop": 0.0,
24
+ "encoder_layers": 24,
25
+ "eos_token_id": 50257,
26
+ "forced_decoder_ids": [
27
+ [
28
+ 1,
29
+ 50259
30
+ ],
31
+ [
32
+ 2,
33
+ 50359
34
+ ],
35
+ [
36
+ 3,
37
+ 50363
38
+ ]
39
+ ],
40
+ "init_std": 0.02,
41
+ "is_encoder_decoder": true,
42
+ "max_length": 448,
43
+ "max_source_positions": 1500,
44
+ "max_target_positions": 448,
45
+ "model_type": "whisper",
46
+ "num_hidden_layers": 24,
47
+ "num_mel_bins": 80,
48
+ "pad_token_id": 50257,
49
+ "scale_embedding": false,
50
+ "suppress_tokens": [
51
+ 1,
52
+ 2,
53
+ 7,
54
+ 8,
55
+ 9,
56
+ 10,
57
+ 14,
58
+ 25,
59
+ 26,
60
+ 27,
61
+ 28,
62
+ 29,
63
+ 31,
64
+ 58,
65
+ 59,
66
+ 60,
67
+ 61,
68
+ 62,
69
+ 63,
70
+ 90,
71
+ 91,
72
+ 92,
73
+ 93,
74
+ 359,
75
+ 503,
76
+ 522,
77
+ 542,
78
+ 873,
79
+ 893,
80
+ 902,
81
+ 918,
82
+ 922,
83
+ 931,
84
+ 1350,
85
+ 1853,
86
+ 1982,
87
+ 2460,
88
+ 2627,
89
+ 3246,
90
+ 3253,
91
+ 3268,
92
+ 3536,
93
+ 3846,
94
+ 3961,
95
+ 4183,
96
+ 4667,
97
+ 6585,
98
+ 6647,
99
+ 7273,
100
+ 9061,
101
+ 9383,
102
+ 10428,
103
+ 10929,
104
+ 11938,
105
+ 12033,
106
+ 12331,
107
+ 12562,
108
+ 13793,
109
+ 14157,
110
+ 14635,
111
+ 15265,
112
+ 15618,
113
+ 16553,
114
+ 16604,
115
+ 18362,
116
+ 18956,
117
+ 20075,
118
+ 21675,
119
+ 22520,
120
+ 26130,
121
+ 26161,
122
+ 26435,
123
+ 28279,
124
+ 29464,
125
+ 31650,
126
+ 32302,
127
+ 32470,
128
+ 36865,
129
+ 42863,
130
+ 47425,
131
+ 49870,
132
+ 50254,
133
+ 50258,
134
+ 50358,
135
+ 50359,
136
+ 50360,
137
+ 50361,
138
+ 50362
139
+ ],
140
+ "torch_dtype": "float32",
141
+ "transformers_version": "4.26.0",
142
+ "use_cache": true,
143
+ "vocab_size": 51865
144
+ }
checkpoint-1200/generation_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "begin_suppress_tokens": [
3
+ 220,
4
+ 50257
5
+ ],
6
+ "bos_token_id": 50257,
7
+ "decoder_start_token_id": 50258,
8
+ "eos_token_id": 50257,
9
+ "language": "spanish",
10
+ "max_length": 448,
11
+ "pad_token_id": 50257,
12
+ "suppress_tokens": [
13
+ 1,
14
+ 2,
15
+ 7,
16
+ 8,
17
+ 9,
18
+ 10,
19
+ 14,
20
+ 25,
21
+ 26,
22
+ 27,
23
+ 28,
24
+ 29,
25
+ 31,
26
+ 58,
27
+ 59,
28
+ 60,
29
+ 61,
30
+ 62,
31
+ 63,
32
+ 90,
33
+ 91,
34
+ 92,
35
+ 93,
36
+ 359,
37
+ 503,
38
+ 522,
39
+ 542,
40
+ 873,
41
+ 893,
42
+ 902,
43
+ 918,
44
+ 922,
45
+ 931,
46
+ 1350,
47
+ 1853,
48
+ 1982,
49
+ 2460,
50
+ 2627,
51
+ 3246,
52
+ 3253,
53
+ 3268,
54
+ 3536,
55
+ 3846,
56
+ 3961,
57
+ 4183,
58
+ 4667,
59
+ 6585,
60
+ 6647,
61
+ 7273,
62
+ 9061,
63
+ 9383,
64
+ 10428,
65
+ 10929,
66
+ 11938,
67
+ 12033,
68
+ 12331,
69
+ 12562,
70
+ 13793,
71
+ 14157,
72
+ 14635,
73
+ 15265,
74
+ 15618,
75
+ 16553,
76
+ 16604,
77
+ 18362,
78
+ 18956,
79
+ 20075,
80
+ 21675,
81
+ 22520,
82
+ 26130,
83
+ 26161,
84
+ 26435,
85
+ 28279,
86
+ 29464,
87
+ 31650,
88
+ 32302,
89
+ 32470,
90
+ 36865,
91
+ 42863,
92
+ 47425,
93
+ 49870,
94
+ 50254,
95
+ 50258,
96
+ 50358,
97
+ 50359,
98
+ 50360,
99
+ 50361,
100
+ 50362
101
+ ],
102
+ "task": "transcribe",
103
+ "transformers_version": "4.26.0"
104
+ }
checkpoint-1200/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9d1fde05b92cdae77ddc567038ee842773a5ea4acadd293a6d6ac3990a3142c
3
+ size 6111429144
checkpoint-1200/preprocessor_config.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1200/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eda6b9c17ebef41bc3084919085e99d8bade84d9b2e1939dbd52a80f5a4aca71
3
+ size 3055755286
checkpoint-1200/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41408a0f99fc0892502deba5659743b776ae2cceccfe18859fa915f755070273
3
+ size 14244
checkpoint-1200/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6bfd54458377de7624240c383f6acb6d4099f6b47bc18401dc0c43a600f6f461
3
+ size 988
checkpoint-1200/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afca3f59d8acabf6b604171ddfcb7208597645926602a097b732af910c1344a0
3
+ size 1064
checkpoint-1200/trainer_state.json ADDED
@@ -0,0 +1,331 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 108.59188544152745,
3
+ "best_model_checkpoint": "ASFEMA_SEV/checkpoint-800",
4
+ "epoch": 15.578778135048232,
5
+ "global_step": 1200,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.32,
12
+ "learning_rate": 4.6000000000000004e-07,
13
+ "loss": 4.0894,
14
+ "step": 25
15
+ },
16
+ {
17
+ "epoch": 0.64,
18
+ "learning_rate": 9.600000000000001e-07,
19
+ "loss": 3.7,
20
+ "step": 50
21
+ },
22
+ {
23
+ "epoch": 0.96,
24
+ "learning_rate": 1.46e-06,
25
+ "loss": 3.0759,
26
+ "step": 75
27
+ },
28
+ {
29
+ "epoch": 1.3,
30
+ "learning_rate": 1.9600000000000003e-06,
31
+ "loss": 2.5729,
32
+ "step": 100
33
+ },
34
+ {
35
+ "epoch": 1.62,
36
+ "learning_rate": 2.46e-06,
37
+ "loss": 2.1885,
38
+ "step": 125
39
+ },
40
+ {
41
+ "epoch": 1.94,
42
+ "learning_rate": 2.96e-06,
43
+ "loss": 1.9431,
44
+ "step": 150
45
+ },
46
+ {
47
+ "epoch": 2.27,
48
+ "learning_rate": 3.46e-06,
49
+ "loss": 1.5753,
50
+ "step": 175
51
+ },
52
+ {
53
+ "epoch": 2.59,
54
+ "learning_rate": 3.96e-06,
55
+ "loss": 1.3319,
56
+ "step": 200
57
+ },
58
+ {
59
+ "epoch": 2.91,
60
+ "learning_rate": 4.4600000000000005e-06,
61
+ "loss": 1.2076,
62
+ "step": 225
63
+ },
64
+ {
65
+ "epoch": 3.24,
66
+ "learning_rate": 4.960000000000001e-06,
67
+ "loss": 0.9211,
68
+ "step": 250
69
+ },
70
+ {
71
+ "epoch": 3.57,
72
+ "learning_rate": 5.460000000000001e-06,
73
+ "loss": 0.7332,
74
+ "step": 275
75
+ },
76
+ {
77
+ "epoch": 3.89,
78
+ "learning_rate": 5.9600000000000005e-06,
79
+ "loss": 0.6601,
80
+ "step": 300
81
+ },
82
+ {
83
+ "epoch": 4.22,
84
+ "learning_rate": 6.460000000000001e-06,
85
+ "loss": 0.4824,
86
+ "step": 325
87
+ },
88
+ {
89
+ "epoch": 4.54,
90
+ "learning_rate": 6.96e-06,
91
+ "loss": 0.3455,
92
+ "step": 350
93
+ },
94
+ {
95
+ "epoch": 4.86,
96
+ "learning_rate": 7.4600000000000006e-06,
97
+ "loss": 0.3741,
98
+ "step": 375
99
+ },
100
+ {
101
+ "epoch": 5.19,
102
+ "learning_rate": 7.960000000000002e-06,
103
+ "loss": 0.2448,
104
+ "step": 400
105
+ },
106
+ {
107
+ "epoch": 5.19,
108
+ "eval_loss": 0.9975764751434326,
109
+ "eval_runtime": 24.147,
110
+ "eval_samples_per_second": 5.756,
111
+ "eval_steps_per_second": 0.745,
112
+ "eval_wer": 132.57756563245823,
113
+ "step": 400
114
+ },
115
+ {
116
+ "epoch": 5.51,
117
+ "learning_rate": 8.46e-06,
118
+ "loss": 0.1713,
119
+ "step": 425
120
+ },
121
+ {
122
+ "epoch": 5.84,
123
+ "learning_rate": 8.96e-06,
124
+ "loss": 0.1906,
125
+ "step": 450
126
+ },
127
+ {
128
+ "epoch": 6.17,
129
+ "learning_rate": 9.460000000000001e-06,
130
+ "loss": 0.1479,
131
+ "step": 475
132
+ },
133
+ {
134
+ "epoch": 6.49,
135
+ "learning_rate": 9.960000000000001e-06,
136
+ "loss": 0.103,
137
+ "step": 500
138
+ },
139
+ {
140
+ "epoch": 6.81,
141
+ "learning_rate": 9.975789473684211e-06,
142
+ "loss": 0.1271,
143
+ "step": 525
144
+ },
145
+ {
146
+ "epoch": 7.14,
147
+ "learning_rate": 9.949473684210526e-06,
148
+ "loss": 0.1033,
149
+ "step": 550
150
+ },
151
+ {
152
+ "epoch": 7.46,
153
+ "learning_rate": 9.923157894736844e-06,
154
+ "loss": 0.0799,
155
+ "step": 575
156
+ },
157
+ {
158
+ "epoch": 7.78,
159
+ "learning_rate": 9.89684210526316e-06,
160
+ "loss": 0.0814,
161
+ "step": 600
162
+ },
163
+ {
164
+ "epoch": 8.12,
165
+ "learning_rate": 9.870526315789474e-06,
166
+ "loss": 0.069,
167
+ "step": 625
168
+ },
169
+ {
170
+ "epoch": 8.44,
171
+ "learning_rate": 9.84421052631579e-06,
172
+ "loss": 0.0546,
173
+ "step": 650
174
+ },
175
+ {
176
+ "epoch": 8.76,
177
+ "learning_rate": 9.817894736842106e-06,
178
+ "loss": 0.0627,
179
+ "step": 675
180
+ },
181
+ {
182
+ "epoch": 9.09,
183
+ "learning_rate": 9.791578947368422e-06,
184
+ "loss": 0.0459,
185
+ "step": 700
186
+ },
187
+ {
188
+ "epoch": 9.41,
189
+ "learning_rate": 9.765263157894737e-06,
190
+ "loss": 0.034,
191
+ "step": 725
192
+ },
193
+ {
194
+ "epoch": 9.73,
195
+ "learning_rate": 9.738947368421054e-06,
196
+ "loss": 0.036,
197
+ "step": 750
198
+ },
199
+ {
200
+ "epoch": 10.06,
201
+ "learning_rate": 9.712631578947369e-06,
202
+ "loss": 0.0367,
203
+ "step": 775
204
+ },
205
+ {
206
+ "epoch": 10.39,
207
+ "learning_rate": 9.686315789473684e-06,
208
+ "loss": 0.0248,
209
+ "step": 800
210
+ },
211
+ {
212
+ "epoch": 10.39,
213
+ "eval_loss": 0.8046127557754517,
214
+ "eval_runtime": 19.9036,
215
+ "eval_samples_per_second": 6.984,
216
+ "eval_steps_per_second": 0.904,
217
+ "eval_wer": 108.59188544152745,
218
+ "step": 800
219
+ },
220
+ {
221
+ "epoch": 10.71,
222
+ "learning_rate": 9.66e-06,
223
+ "loss": 0.0324,
224
+ "step": 825
225
+ },
226
+ {
227
+ "epoch": 11.04,
228
+ "learning_rate": 9.633684210526316e-06,
229
+ "loss": 0.0374,
230
+ "step": 850
231
+ },
232
+ {
233
+ "epoch": 11.36,
234
+ "learning_rate": 9.607368421052632e-06,
235
+ "loss": 0.0253,
236
+ "step": 875
237
+ },
238
+ {
239
+ "epoch": 11.68,
240
+ "learning_rate": 9.581052631578947e-06,
241
+ "loss": 0.0262,
242
+ "step": 900
243
+ },
244
+ {
245
+ "epoch": 12.01,
246
+ "learning_rate": 9.554736842105264e-06,
247
+ "loss": 0.0374,
248
+ "step": 925
249
+ },
250
+ {
251
+ "epoch": 12.33,
252
+ "learning_rate": 9.52842105263158e-06,
253
+ "loss": 0.0179,
254
+ "step": 950
255
+ },
256
+ {
257
+ "epoch": 12.66,
258
+ "learning_rate": 9.502105263157896e-06,
259
+ "loss": 0.0236,
260
+ "step": 975
261
+ },
262
+ {
263
+ "epoch": 12.98,
264
+ "learning_rate": 9.475789473684212e-06,
265
+ "loss": 0.0187,
266
+ "step": 1000
267
+ },
268
+ {
269
+ "epoch": 13.31,
270
+ "learning_rate": 9.449473684210527e-06,
271
+ "loss": 0.0186,
272
+ "step": 1025
273
+ },
274
+ {
275
+ "epoch": 13.63,
276
+ "learning_rate": 9.423157894736842e-06,
277
+ "loss": 0.0217,
278
+ "step": 1050
279
+ },
280
+ {
281
+ "epoch": 13.95,
282
+ "learning_rate": 9.396842105263159e-06,
283
+ "loss": 0.0206,
284
+ "step": 1075
285
+ },
286
+ {
287
+ "epoch": 14.28,
288
+ "learning_rate": 9.370526315789474e-06,
289
+ "loss": 0.0103,
290
+ "step": 1100
291
+ },
292
+ {
293
+ "epoch": 14.6,
294
+ "learning_rate": 9.34421052631579e-06,
295
+ "loss": 0.0168,
296
+ "step": 1125
297
+ },
298
+ {
299
+ "epoch": 14.93,
300
+ "learning_rate": 9.317894736842105e-06,
301
+ "loss": 0.0214,
302
+ "step": 1150
303
+ },
304
+ {
305
+ "epoch": 15.26,
306
+ "learning_rate": 9.291578947368422e-06,
307
+ "loss": 0.0115,
308
+ "step": 1175
309
+ },
310
+ {
311
+ "epoch": 15.58,
312
+ "learning_rate": 9.265263157894737e-06,
313
+ "loss": 0.0138,
314
+ "step": 1200
315
+ },
316
+ {
317
+ "epoch": 15.58,
318
+ "eval_loss": 0.8237143158912659,
319
+ "eval_runtime": 22.295,
320
+ "eval_samples_per_second": 6.235,
321
+ "eval_steps_per_second": 0.807,
322
+ "eval_wer": 117.89976133651552,
323
+ "step": 1200
324
+ }
325
+ ],
326
+ "max_steps": 10000,
327
+ "num_train_epochs": 130,
328
+ "total_flos": 1.97793231077376e+19,
329
+ "trial_name": null,
330
+ "trial_params": null
331
+ }
checkpoint-1200/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4ad917c70902fd83abc5a9152e70fd07531fce940ccd66f6af7f3bc05e82880
3
+ size 4088
checkpoint-1600/config.json ADDED
@@ -0,0 +1,144 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "openai/whisper-medium",
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "gelu",
5
+ "architectures": [
6
+ "WhisperForConditionalGeneration"
7
+ ],
8
+ "attention_dropout": 0.0,
9
+ "begin_suppress_tokens": [
10
+ 220,
11
+ 50257
12
+ ],
13
+ "bos_token_id": 50257,
14
+ "d_model": 1024,
15
+ "decoder_attention_heads": 16,
16
+ "decoder_ffn_dim": 4096,
17
+ "decoder_layerdrop": 0.0,
18
+ "decoder_layers": 24,
19
+ "decoder_start_token_id": 50258,
20
+ "dropout": 0.0,
21
+ "encoder_attention_heads": 16,
22
+ "encoder_ffn_dim": 4096,
23
+ "encoder_layerdrop": 0.0,
24
+ "encoder_layers": 24,
25
+ "eos_token_id": 50257,
26
+ "forced_decoder_ids": [
27
+ [
28
+ 1,
29
+ 50259
30
+ ],
31
+ [
32
+ 2,
33
+ 50359
34
+ ],
35
+ [
36
+ 3,
37
+ 50363
38
+ ]
39
+ ],
40
+ "init_std": 0.02,
41
+ "is_encoder_decoder": true,
42
+ "max_length": 448,
43
+ "max_source_positions": 1500,
44
+ "max_target_positions": 448,
45
+ "model_type": "whisper",
46
+ "num_hidden_layers": 24,
47
+ "num_mel_bins": 80,
48
+ "pad_token_id": 50257,
49
+ "scale_embedding": false,
50
+ "suppress_tokens": [
51
+ 1,
52
+ 2,
53
+ 7,
54
+ 8,
55
+ 9,
56
+ 10,
57
+ 14,
58
+ 25,
59
+ 26,
60
+ 27,
61
+ 28,
62
+ 29,
63
+ 31,
64
+ 58,
65
+ 59,
66
+ 60,
67
+ 61,
68
+ 62,
69
+ 63,
70
+ 90,
71
+ 91,
72
+ 92,
73
+ 93,
74
+ 359,
75
+ 503,
76
+ 522,
77
+ 542,
78
+ 873,
79
+ 893,
80
+ 902,
81
+ 918,
82
+ 922,
83
+ 931,
84
+ 1350,
85
+ 1853,
86
+ 1982,
87
+ 2460,
88
+ 2627,
89
+ 3246,
90
+ 3253,
91
+ 3268,
92
+ 3536,
93
+ 3846,
94
+ 3961,
95
+ 4183,
96
+ 4667,
97
+ 6585,
98
+ 6647,
99
+ 7273,
100
+ 9061,
101
+ 9383,
102
+ 10428,
103
+ 10929,
104
+ 11938,
105
+ 12033,
106
+ 12331,
107
+ 12562,
108
+ 13793,
109
+ 14157,
110
+ 14635,
111
+ 15265,
112
+ 15618,
113
+ 16553,
114
+ 16604,
115
+ 18362,
116
+ 18956,
117
+ 20075,
118
+ 21675,
119
+ 22520,
120
+ 26130,
121
+ 26161,
122
+ 26435,
123
+ 28279,
124
+ 29464,
125
+ 31650,
126
+ 32302,
127
+ 32470,
128
+ 36865,
129
+ 42863,
130
+ 47425,
131
+ 49870,
132
+ 50254,
133
+ 50258,
134
+ 50358,
135
+ 50359,
136
+ 50360,
137
+ 50361,
138
+ 50362
139
+ ],
140
+ "torch_dtype": "float32",
141
+ "transformers_version": "4.26.0",
142
+ "use_cache": true,
143
+ "vocab_size": 51865
144
+ }
checkpoint-1600/generation_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "begin_suppress_tokens": [
3
+ 220,
4
+ 50257
5
+ ],
6
+ "bos_token_id": 50257,
7
+ "decoder_start_token_id": 50258,
8
+ "eos_token_id": 50257,
9
+ "language": "spanish",
10
+ "max_length": 448,
11
+ "pad_token_id": 50257,
12
+ "suppress_tokens": [
13
+ 1,
14
+ 2,
15
+ 7,
16
+ 8,
17
+ 9,
18
+ 10,
19
+ 14,
20
+ 25,
21
+ 26,
22
+ 27,
23
+ 28,
24
+ 29,
25
+ 31,
26
+ 58,
27
+ 59,
28
+ 60,
29
+ 61,
30
+ 62,
31
+ 63,
32
+ 90,
33
+ 91,
34
+ 92,
35
+ 93,
36
+ 359,
37
+ 503,
38
+ 522,
39
+ 542,
40
+ 873,
41
+ 893,
42
+ 902,
43
+ 918,
44
+ 922,
45
+ 931,
46
+ 1350,
47
+ 1853,
48
+ 1982,
49
+ 2460,
50
+ 2627,
51
+ 3246,
52
+ 3253,
53
+ 3268,
54
+ 3536,
55
+ 3846,
56
+ 3961,
57
+ 4183,
58
+ 4667,
59
+ 6585,
60
+ 6647,
61
+ 7273,
62
+ 9061,
63
+ 9383,
64
+ 10428,
65
+ 10929,
66
+ 11938,
67
+ 12033,
68
+ 12331,
69
+ 12562,
70
+ 13793,
71
+ 14157,
72
+ 14635,
73
+ 15265,
74
+ 15618,
75
+ 16553,
76
+ 16604,
77
+ 18362,
78
+ 18956,
79
+ 20075,
80
+ 21675,
81
+ 22520,
82
+ 26130,
83
+ 26161,
84
+ 26435,
85
+ 28279,
86
+ 29464,
87
+ 31650,
88
+ 32302,
89
+ 32470,
90
+ 36865,
91
+ 42863,
92
+ 47425,
93
+ 49870,
94
+ 50254,
95
+ 50258,
96
+ 50358,
97
+ 50359,
98
+ 50360,
99
+ 50361,
100
+ 50362
101
+ ],
102
+ "task": "transcribe",
103
+ "transformers_version": "4.26.0"
104
+ }
checkpoint-1600/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:571d7eb393dedf0f58f1b65b0be5e62d7ad4cfd766d2c157bc23285da99d3b94
3
+ size 6111429144
checkpoint-1600/preprocessor_config.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1600/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d1528b41108357c8ea8c11f403caa1fa688198fc2e96107e7edad8955b8e12d
3
+ size 3055755286
checkpoint-1600/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3440fc5c490287f3b62da52da3c577b2efac510091ede21871b6d7db5a7b71dd
3
+ size 14244
checkpoint-1600/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f0409957cdf0f0e718e42ebeafcb7f9ea32e89669cdff41ad41dfb868f925fd
3
+ size 988
checkpoint-1600/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fea7ea3f403fddb6bba4db183021b972ab6191ab93d0c7f5b85291e7f045c772
3
+ size 1064
checkpoint-1600/trainer_state.json ADDED
@@ -0,0 +1,436 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 103.5799522673031,
3
+ "best_model_checkpoint": "ASFEMA_SEV/checkpoint-1600",
4
+ "epoch": 20.771704180064308,
5
+ "global_step": 1600,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.32,
12
+ "learning_rate": 4.6000000000000004e-07,
13
+ "loss": 4.0894,
14
+ "step": 25
15
+ },
16
+ {
17
+ "epoch": 0.64,
18
+ "learning_rate": 9.600000000000001e-07,
19
+ "loss": 3.7,
20
+ "step": 50
21
+ },
22
+ {
23
+ "epoch": 0.96,
24
+ "learning_rate": 1.46e-06,
25
+ "loss": 3.0759,
26
+ "step": 75
27
+ },
28
+ {
29
+ "epoch": 1.3,
30
+ "learning_rate": 1.9600000000000003e-06,
31
+ "loss": 2.5729,
32
+ "step": 100
33
+ },
34
+ {
35
+ "epoch": 1.62,
36
+ "learning_rate": 2.46e-06,
37
+ "loss": 2.1885,
38
+ "step": 125
39
+ },
40
+ {
41
+ "epoch": 1.94,
42
+ "learning_rate": 2.96e-06,
43
+ "loss": 1.9431,
44
+ "step": 150
45
+ },
46
+ {
47
+ "epoch": 2.27,
48
+ "learning_rate": 3.46e-06,
49
+ "loss": 1.5753,
50
+ "step": 175
51
+ },
52
+ {
53
+ "epoch": 2.59,
54
+ "learning_rate": 3.96e-06,
55
+ "loss": 1.3319,
56
+ "step": 200
57
+ },
58
+ {
59
+ "epoch": 2.91,
60
+ "learning_rate": 4.4600000000000005e-06,
61
+ "loss": 1.2076,
62
+ "step": 225
63
+ },
64
+ {
65
+ "epoch": 3.24,
66
+ "learning_rate": 4.960000000000001e-06,
67
+ "loss": 0.9211,
68
+ "step": 250
69
+ },
70
+ {
71
+ "epoch": 3.57,
72
+ "learning_rate": 5.460000000000001e-06,
73
+ "loss": 0.7332,
74
+ "step": 275
75
+ },
76
+ {
77
+ "epoch": 3.89,
78
+ "learning_rate": 5.9600000000000005e-06,
79
+ "loss": 0.6601,
80
+ "step": 300
81
+ },
82
+ {
83
+ "epoch": 4.22,
84
+ "learning_rate": 6.460000000000001e-06,
85
+ "loss": 0.4824,
86
+ "step": 325
87
+ },
88
+ {
89
+ "epoch": 4.54,
90
+ "learning_rate": 6.96e-06,
91
+ "loss": 0.3455,
92
+ "step": 350
93
+ },
94
+ {
95
+ "epoch": 4.86,
96
+ "learning_rate": 7.4600000000000006e-06,
97
+ "loss": 0.3741,
98
+ "step": 375
99
+ },
100
+ {
101
+ "epoch": 5.19,
102
+ "learning_rate": 7.960000000000002e-06,
103
+ "loss": 0.2448,
104
+ "step": 400
105
+ },
106
+ {
107
+ "epoch": 5.19,
108
+ "eval_loss": 0.9975764751434326,
109
+ "eval_runtime": 24.147,
110
+ "eval_samples_per_second": 5.756,
111
+ "eval_steps_per_second": 0.745,
112
+ "eval_wer": 132.57756563245823,
113
+ "step": 400
114
+ },
115
+ {
116
+ "epoch": 5.51,
117
+ "learning_rate": 8.46e-06,
118
+ "loss": 0.1713,
119
+ "step": 425
120
+ },
121
+ {
122
+ "epoch": 5.84,
123
+ "learning_rate": 8.96e-06,
124
+ "loss": 0.1906,
125
+ "step": 450
126
+ },
127
+ {
128
+ "epoch": 6.17,
129
+ "learning_rate": 9.460000000000001e-06,
130
+ "loss": 0.1479,
131
+ "step": 475
132
+ },
133
+ {
134
+ "epoch": 6.49,
135
+ "learning_rate": 9.960000000000001e-06,
136
+ "loss": 0.103,
137
+ "step": 500
138
+ },
139
+ {
140
+ "epoch": 6.81,
141
+ "learning_rate": 9.975789473684211e-06,
142
+ "loss": 0.1271,
143
+ "step": 525
144
+ },
145
+ {
146
+ "epoch": 7.14,
147
+ "learning_rate": 9.949473684210526e-06,
148
+ "loss": 0.1033,
149
+ "step": 550
150
+ },
151
+ {
152
+ "epoch": 7.46,
153
+ "learning_rate": 9.923157894736844e-06,
154
+ "loss": 0.0799,
155
+ "step": 575
156
+ },
157
+ {
158
+ "epoch": 7.78,
159
+ "learning_rate": 9.89684210526316e-06,
160
+ "loss": 0.0814,
161
+ "step": 600
162
+ },
163
+ {
164
+ "epoch": 8.12,
165
+ "learning_rate": 9.870526315789474e-06,
166
+ "loss": 0.069,
167
+ "step": 625
168
+ },
169
+ {
170
+ "epoch": 8.44,
171
+ "learning_rate": 9.84421052631579e-06,
172
+ "loss": 0.0546,
173
+ "step": 650
174
+ },
175
+ {
176
+ "epoch": 8.76,
177
+ "learning_rate": 9.817894736842106e-06,
178
+ "loss": 0.0627,
179
+ "step": 675
180
+ },
181
+ {
182
+ "epoch": 9.09,
183
+ "learning_rate": 9.791578947368422e-06,
184
+ "loss": 0.0459,
185
+ "step": 700
186
+ },
187
+ {
188
+ "epoch": 9.41,
189
+ "learning_rate": 9.765263157894737e-06,
190
+ "loss": 0.034,
191
+ "step": 725
192
+ },
193
+ {
194
+ "epoch": 9.73,
195
+ "learning_rate": 9.738947368421054e-06,
196
+ "loss": 0.036,
197
+ "step": 750
198
+ },
199
+ {
200
+ "epoch": 10.06,
201
+ "learning_rate": 9.712631578947369e-06,
202
+ "loss": 0.0367,
203
+ "step": 775
204
+ },
205
+ {
206
+ "epoch": 10.39,
207
+ "learning_rate": 9.686315789473684e-06,
208
+ "loss": 0.0248,
209
+ "step": 800
210
+ },
211
+ {
212
+ "epoch": 10.39,
213
+ "eval_loss": 0.8046127557754517,
214
+ "eval_runtime": 19.9036,
215
+ "eval_samples_per_second": 6.984,
216
+ "eval_steps_per_second": 0.904,
217
+ "eval_wer": 108.59188544152745,
218
+ "step": 800
219
+ },
220
+ {
221
+ "epoch": 10.71,
222
+ "learning_rate": 9.66e-06,
223
+ "loss": 0.0324,
224
+ "step": 825
225
+ },
226
+ {
227
+ "epoch": 11.04,
228
+ "learning_rate": 9.633684210526316e-06,
229
+ "loss": 0.0374,
230
+ "step": 850
231
+ },
232
+ {
233
+ "epoch": 11.36,
234
+ "learning_rate": 9.607368421052632e-06,
235
+ "loss": 0.0253,
236
+ "step": 875
237
+ },
238
+ {
239
+ "epoch": 11.68,
240
+ "learning_rate": 9.581052631578947e-06,
241
+ "loss": 0.0262,
242
+ "step": 900
243
+ },
244
+ {
245
+ "epoch": 12.01,
246
+ "learning_rate": 9.554736842105264e-06,
247
+ "loss": 0.0374,
248
+ "step": 925
249
+ },
250
+ {
251
+ "epoch": 12.33,
252
+ "learning_rate": 9.52842105263158e-06,
253
+ "loss": 0.0179,
254
+ "step": 950
255
+ },
256
+ {
257
+ "epoch": 12.66,
258
+ "learning_rate": 9.502105263157896e-06,
259
+ "loss": 0.0236,
260
+ "step": 975
261
+ },
262
+ {
263
+ "epoch": 12.98,
264
+ "learning_rate": 9.475789473684212e-06,
265
+ "loss": 0.0187,
266
+ "step": 1000
267
+ },
268
+ {
269
+ "epoch": 13.31,
270
+ "learning_rate": 9.449473684210527e-06,
271
+ "loss": 0.0186,
272
+ "step": 1025
273
+ },
274
+ {
275
+ "epoch": 13.63,
276
+ "learning_rate": 9.423157894736842e-06,
277
+ "loss": 0.0217,
278
+ "step": 1050
279
+ },
280
+ {
281
+ "epoch": 13.95,
282
+ "learning_rate": 9.396842105263159e-06,
283
+ "loss": 0.0206,
284
+ "step": 1075
285
+ },
286
+ {
287
+ "epoch": 14.28,
288
+ "learning_rate": 9.370526315789474e-06,
289
+ "loss": 0.0103,
290
+ "step": 1100
291
+ },
292
+ {
293
+ "epoch": 14.6,
294
+ "learning_rate": 9.34421052631579e-06,
295
+ "loss": 0.0168,
296
+ "step": 1125
297
+ },
298
+ {
299
+ "epoch": 14.93,
300
+ "learning_rate": 9.317894736842105e-06,
301
+ "loss": 0.0214,
302
+ "step": 1150
303
+ },
304
+ {
305
+ "epoch": 15.26,
306
+ "learning_rate": 9.291578947368422e-06,
307
+ "loss": 0.0115,
308
+ "step": 1175
309
+ },
310
+ {
311
+ "epoch": 15.58,
312
+ "learning_rate": 9.265263157894737e-06,
313
+ "loss": 0.0138,
314
+ "step": 1200
315
+ },
316
+ {
317
+ "epoch": 15.58,
318
+ "eval_loss": 0.8237143158912659,
319
+ "eval_runtime": 22.295,
320
+ "eval_samples_per_second": 6.235,
321
+ "eval_steps_per_second": 0.807,
322
+ "eval_wer": 117.89976133651552,
323
+ "step": 1200
324
+ },
325
+ {
326
+ "epoch": 15.9,
327
+ "learning_rate": 9.238947368421052e-06,
328
+ "loss": 0.0216,
329
+ "step": 1225
330
+ },
331
+ {
332
+ "epoch": 16.23,
333
+ "learning_rate": 9.21263157894737e-06,
334
+ "loss": 0.014,
335
+ "step": 1250
336
+ },
337
+ {
338
+ "epoch": 16.55,
339
+ "learning_rate": 9.186315789473685e-06,
340
+ "loss": 0.0117,
341
+ "step": 1275
342
+ },
343
+ {
344
+ "epoch": 16.87,
345
+ "learning_rate": 9.16e-06,
346
+ "loss": 0.0104,
347
+ "step": 1300
348
+ },
349
+ {
350
+ "epoch": 17.21,
351
+ "learning_rate": 9.133684210526317e-06,
352
+ "loss": 0.0086,
353
+ "step": 1325
354
+ },
355
+ {
356
+ "epoch": 17.53,
357
+ "learning_rate": 9.107368421052632e-06,
358
+ "loss": 0.0092,
359
+ "step": 1350
360
+ },
361
+ {
362
+ "epoch": 17.85,
363
+ "learning_rate": 9.081052631578949e-06,
364
+ "loss": 0.0098,
365
+ "step": 1375
366
+ },
367
+ {
368
+ "epoch": 18.18,
369
+ "learning_rate": 9.054736842105264e-06,
370
+ "loss": 0.0125,
371
+ "step": 1400
372
+ },
373
+ {
374
+ "epoch": 18.5,
375
+ "learning_rate": 9.02842105263158e-06,
376
+ "loss": 0.0055,
377
+ "step": 1425
378
+ },
379
+ {
380
+ "epoch": 18.82,
381
+ "learning_rate": 9.002105263157895e-06,
382
+ "loss": 0.011,
383
+ "step": 1450
384
+ },
385
+ {
386
+ "epoch": 19.15,
387
+ "learning_rate": 8.97578947368421e-06,
388
+ "loss": 0.0088,
389
+ "step": 1475
390
+ },
391
+ {
392
+ "epoch": 19.48,
393
+ "learning_rate": 8.949473684210527e-06,
394
+ "loss": 0.0065,
395
+ "step": 1500
396
+ },
397
+ {
398
+ "epoch": 19.8,
399
+ "learning_rate": 8.923157894736842e-06,
400
+ "loss": 0.0098,
401
+ "step": 1525
402
+ },
403
+ {
404
+ "epoch": 20.13,
405
+ "learning_rate": 8.896842105263159e-06,
406
+ "loss": 0.0038,
407
+ "step": 1550
408
+ },
409
+ {
410
+ "epoch": 20.45,
411
+ "learning_rate": 8.870526315789474e-06,
412
+ "loss": 0.0053,
413
+ "step": 1575
414
+ },
415
+ {
416
+ "epoch": 20.77,
417
+ "learning_rate": 8.84421052631579e-06,
418
+ "loss": 0.0075,
419
+ "step": 1600
420
+ },
421
+ {
422
+ "epoch": 20.77,
423
+ "eval_loss": 0.8213781118392944,
424
+ "eval_runtime": 18.9322,
425
+ "eval_samples_per_second": 7.342,
426
+ "eval_steps_per_second": 0.951,
427
+ "eval_wer": 103.5799522673031,
428
+ "step": 1600
429
+ }
430
+ ],
431
+ "max_steps": 10000,
432
+ "num_train_epochs": 130,
433
+ "total_flos": 2.63724308103168e+19,
434
+ "trial_name": null,
435
+ "trial_params": null
436
+ }
checkpoint-1600/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4ad917c70902fd83abc5a9152e70fd07531fce940ccd66f6af7f3bc05e82880
3
+ size 4088
checkpoint-2000/config.json ADDED
@@ -0,0 +1,144 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "openai/whisper-medium",
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "gelu",
5
+ "architectures": [
6
+ "WhisperForConditionalGeneration"
7
+ ],
8
+ "attention_dropout": 0.0,
9
+ "begin_suppress_tokens": [
10
+ 220,
11
+ 50257
12
+ ],
13
+ "bos_token_id": 50257,
14
+ "d_model": 1024,
15
+ "decoder_attention_heads": 16,
16
+ "decoder_ffn_dim": 4096,
17
+ "decoder_layerdrop": 0.0,
18
+ "decoder_layers": 24,
19
+ "decoder_start_token_id": 50258,
20
+ "dropout": 0.0,
21
+ "encoder_attention_heads": 16,
22
+ "encoder_ffn_dim": 4096,
23
+ "encoder_layerdrop": 0.0,
24
+ "encoder_layers": 24,
25
+ "eos_token_id": 50257,
26
+ "forced_decoder_ids": [
27
+ [
28
+ 1,
29
+ 50259
30
+ ],
31
+ [
32
+ 2,
33
+ 50359
34
+ ],
35
+ [
36
+ 3,
37
+ 50363
38
+ ]
39
+ ],
40
+ "init_std": 0.02,
41
+ "is_encoder_decoder": true,
42
+ "max_length": 448,
43
+ "max_source_positions": 1500,
44
+ "max_target_positions": 448,
45
+ "model_type": "whisper",
46
+ "num_hidden_layers": 24,
47
+ "num_mel_bins": 80,
48
+ "pad_token_id": 50257,
49
+ "scale_embedding": false,
50
+ "suppress_tokens": [
51
+ 1,
52
+ 2,
53
+ 7,
54
+ 8,
55
+ 9,
56
+ 10,
57
+ 14,
58
+ 25,
59
+ 26,
60
+ 27,
61
+ 28,
62
+ 29,
63
+ 31,
64
+ 58,
65
+ 59,
66
+ 60,
67
+ 61,
68
+ 62,
69
+ 63,
70
+ 90,
71
+ 91,
72
+ 92,
73
+ 93,
74
+ 359,
75
+ 503,
76
+ 522,
77
+ 542,
78
+ 873,
79
+ 893,
80
+ 902,
81
+ 918,
82
+ 922,
83
+ 931,
84
+ 1350,
85
+ 1853,
86
+ 1982,
87
+ 2460,
88
+ 2627,
89
+ 3246,
90
+ 3253,
91
+ 3268,
92
+ 3536,
93
+ 3846,
94
+ 3961,
95
+ 4183,
96
+ 4667,
97
+ 6585,
98
+ 6647,
99
+ 7273,
100
+ 9061,
101
+ 9383,
102
+ 10428,
103
+ 10929,
104
+ 11938,
105
+ 12033,
106
+ 12331,
107
+ 12562,
108
+ 13793,
109
+ 14157,
110
+ 14635,
111
+ 15265,
112
+ 15618,
113
+ 16553,
114
+ 16604,
115
+ 18362,
116
+ 18956,
117
+ 20075,
118
+ 21675,
119
+ 22520,
120
+ 26130,
121
+ 26161,
122
+ 26435,
123
+ 28279,
124
+ 29464,
125
+ 31650,
126
+ 32302,
127
+ 32470,
128
+ 36865,
129
+ 42863,
130
+ 47425,
131
+ 49870,
132
+ 50254,
133
+ 50258,
134
+ 50358,
135
+ 50359,
136
+ 50360,
137
+ 50361,
138
+ 50362
139
+ ],
140
+ "torch_dtype": "float32",
141
+ "transformers_version": "4.26.0",
142
+ "use_cache": true,
143
+ "vocab_size": 51865
144
+ }
checkpoint-2000/generation_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "begin_suppress_tokens": [
3
+ 220,
4
+ 50257
5
+ ],
6
+ "bos_token_id": 50257,
7
+ "decoder_start_token_id": 50258,
8
+ "eos_token_id": 50257,
9
+ "language": "spanish",
10
+ "max_length": 448,
11
+ "pad_token_id": 50257,
12
+ "suppress_tokens": [
13
+ 1,
14
+ 2,
15
+ 7,
16
+ 8,
17
+ 9,
18
+ 10,
19
+ 14,
20
+ 25,
21
+ 26,
22
+ 27,
23
+ 28,
24
+ 29,
25
+ 31,
26
+ 58,
27
+ 59,
28
+ 60,
29
+ 61,
30
+ 62,
31
+ 63,
32
+ 90,
33
+ 91,
34
+ 92,
35
+ 93,
36
+ 359,
37
+ 503,
38
+ 522,
39
+ 542,
40
+ 873,
41
+ 893,
42
+ 902,
43
+ 918,
44
+ 922,
45
+ 931,
46
+ 1350,
47
+ 1853,
48
+ 1982,
49
+ 2460,
50
+ 2627,
51
+ 3246,
52
+ 3253,
53
+ 3268,
54
+ 3536,
55
+ 3846,
56
+ 3961,
57
+ 4183,
58
+ 4667,
59
+ 6585,
60
+ 6647,
61
+ 7273,
62
+ 9061,
63
+ 9383,
64
+ 10428,
65
+ 10929,
66
+ 11938,
67
+ 12033,
68
+ 12331,
69
+ 12562,
70
+ 13793,
71
+ 14157,
72
+ 14635,
73
+ 15265,
74
+ 15618,
75
+ 16553,
76
+ 16604,
77
+ 18362,
78
+ 18956,
79
+ 20075,
80
+ 21675,
81
+ 22520,
82
+ 26130,
83
+ 26161,
84
+ 26435,
85
+ 28279,
86
+ 29464,
87
+ 31650,
88
+ 32302,
89
+ 32470,
90
+ 36865,
91
+ 42863,
92
+ 47425,
93
+ 49870,
94
+ 50254,
95
+ 50258,
96
+ 50358,
97
+ 50359,
98
+ 50360,
99
+ 50361,
100
+ 50362
101
+ ],
102
+ "task": "transcribe",
103
+ "transformers_version": "4.26.0"
104
+ }
checkpoint-2000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b1de0a7da644f0d348a87cd4b7960d0c4de64a21c6b7e20a591d4f3ae963dc8
3
+ size 6111429144
checkpoint-2000/preprocessor_config.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-2000/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d61a2e241b8bfc33af219a514a1727fe940632e5e12c9547ae4b2b72aa1b69b
3
+ size 3055755286
checkpoint-2000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49475abf43029973b16239d40b9c9cdeb29859430ca8c204587c6b93589cd16b
3
+ size 14180
checkpoint-2000/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67673cade32d1cab7895a361505491a7cdd197dd49d222dc1e622c04c3b208e1
3
+ size 988
checkpoint-2000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ca61b8d50e041a9057c1f9af2cf7571ed562744478450c34ac922198ce5fd39
3
+ size 1064
checkpoint-2000/trainer_state.json ADDED
@@ -0,0 +1,541 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 103.5799522673031,
3
+ "best_model_checkpoint": "ASFEMA_SEV/checkpoint-1600",
4
+ "epoch": 25.964630225080384,
5
+ "global_step": 2000,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.32,
12
+ "learning_rate": 4.6000000000000004e-07,
13
+ "loss": 4.0894,
14
+ "step": 25
15
+ },
16
+ {
17
+ "epoch": 0.64,
18
+ "learning_rate": 9.600000000000001e-07,
19
+ "loss": 3.7,
20
+ "step": 50
21
+ },
22
+ {
23
+ "epoch": 0.96,
24
+ "learning_rate": 1.46e-06,
25
+ "loss": 3.0759,
26
+ "step": 75
27
+ },
28
+ {
29
+ "epoch": 1.3,
30
+ "learning_rate": 1.9600000000000003e-06,
31
+ "loss": 2.5729,
32
+ "step": 100
33
+ },
34
+ {
35
+ "epoch": 1.62,
36
+ "learning_rate": 2.46e-06,
37
+ "loss": 2.1885,
38
+ "step": 125
39
+ },
40
+ {
41
+ "epoch": 1.94,
42
+ "learning_rate": 2.96e-06,
43
+ "loss": 1.9431,
44
+ "step": 150
45
+ },
46
+ {
47
+ "epoch": 2.27,
48
+ "learning_rate": 3.46e-06,
49
+ "loss": 1.5753,
50
+ "step": 175
51
+ },
52
+ {
53
+ "epoch": 2.59,
54
+ "learning_rate": 3.96e-06,
55
+ "loss": 1.3319,
56
+ "step": 200
57
+ },
58
+ {
59
+ "epoch": 2.91,
60
+ "learning_rate": 4.4600000000000005e-06,
61
+ "loss": 1.2076,
62
+ "step": 225
63
+ },
64
+ {
65
+ "epoch": 3.24,
66
+ "learning_rate": 4.960000000000001e-06,
67
+ "loss": 0.9211,
68
+ "step": 250
69
+ },
70
+ {
71
+ "epoch": 3.57,
72
+ "learning_rate": 5.460000000000001e-06,
73
+ "loss": 0.7332,
74
+ "step": 275
75
+ },
76
+ {
77
+ "epoch": 3.89,
78
+ "learning_rate": 5.9600000000000005e-06,
79
+ "loss": 0.6601,
80
+ "step": 300
81
+ },
82
+ {
83
+ "epoch": 4.22,
84
+ "learning_rate": 6.460000000000001e-06,
85
+ "loss": 0.4824,
86
+ "step": 325
87
+ },
88
+ {
89
+ "epoch": 4.54,
90
+ "learning_rate": 6.96e-06,
91
+ "loss": 0.3455,
92
+ "step": 350
93
+ },
94
+ {
95
+ "epoch": 4.86,
96
+ "learning_rate": 7.4600000000000006e-06,
97
+ "loss": 0.3741,
98
+ "step": 375
99
+ },
100
+ {
101
+ "epoch": 5.19,
102
+ "learning_rate": 7.960000000000002e-06,
103
+ "loss": 0.2448,
104
+ "step": 400
105
+ },
106
+ {
107
+ "epoch": 5.19,
108
+ "eval_loss": 0.9975764751434326,
109
+ "eval_runtime": 24.147,
110
+ "eval_samples_per_second": 5.756,
111
+ "eval_steps_per_second": 0.745,
112
+ "eval_wer": 132.57756563245823,
113
+ "step": 400
114
+ },
115
+ {
116
+ "epoch": 5.51,
117
+ "learning_rate": 8.46e-06,
118
+ "loss": 0.1713,
119
+ "step": 425
120
+ },
121
+ {
122
+ "epoch": 5.84,
123
+ "learning_rate": 8.96e-06,
124
+ "loss": 0.1906,
125
+ "step": 450
126
+ },
127
+ {
128
+ "epoch": 6.17,
129
+ "learning_rate": 9.460000000000001e-06,
130
+ "loss": 0.1479,
131
+ "step": 475
132
+ },
133
+ {
134
+ "epoch": 6.49,
135
+ "learning_rate": 9.960000000000001e-06,
136
+ "loss": 0.103,
137
+ "step": 500
138
+ },
139
+ {
140
+ "epoch": 6.81,
141
+ "learning_rate": 9.975789473684211e-06,
142
+ "loss": 0.1271,
143
+ "step": 525
144
+ },
145
+ {
146
+ "epoch": 7.14,
147
+ "learning_rate": 9.949473684210526e-06,
148
+ "loss": 0.1033,
149
+ "step": 550
150
+ },
151
+ {
152
+ "epoch": 7.46,
153
+ "learning_rate": 9.923157894736844e-06,
154
+ "loss": 0.0799,
155
+ "step": 575
156
+ },
157
+ {
158
+ "epoch": 7.78,
159
+ "learning_rate": 9.89684210526316e-06,
160
+ "loss": 0.0814,
161
+ "step": 600
162
+ },
163
+ {
164
+ "epoch": 8.12,
165
+ "learning_rate": 9.870526315789474e-06,
166
+ "loss": 0.069,
167
+ "step": 625
168
+ },
169
+ {
170
+ "epoch": 8.44,
171
+ "learning_rate": 9.84421052631579e-06,
172
+ "loss": 0.0546,
173
+ "step": 650
174
+ },
175
+ {
176
+ "epoch": 8.76,
177
+ "learning_rate": 9.817894736842106e-06,
178
+ "loss": 0.0627,
179
+ "step": 675
180
+ },
181
+ {
182
+ "epoch": 9.09,
183
+ "learning_rate": 9.791578947368422e-06,
184
+ "loss": 0.0459,
185
+ "step": 700
186
+ },
187
+ {
188
+ "epoch": 9.41,
189
+ "learning_rate": 9.765263157894737e-06,
190
+ "loss": 0.034,
191
+ "step": 725
192
+ },
193
+ {
194
+ "epoch": 9.73,
195
+ "learning_rate": 9.738947368421054e-06,
196
+ "loss": 0.036,
197
+ "step": 750
198
+ },
199
+ {
200
+ "epoch": 10.06,
201
+ "learning_rate": 9.712631578947369e-06,
202
+ "loss": 0.0367,
203
+ "step": 775
204
+ },
205
+ {
206
+ "epoch": 10.39,
207
+ "learning_rate": 9.686315789473684e-06,
208
+ "loss": 0.0248,
209
+ "step": 800
210
+ },
211
+ {
212
+ "epoch": 10.39,
213
+ "eval_loss": 0.8046127557754517,
214
+ "eval_runtime": 19.9036,
215
+ "eval_samples_per_second": 6.984,
216
+ "eval_steps_per_second": 0.904,
217
+ "eval_wer": 108.59188544152745,
218
+ "step": 800
219
+ },
220
+ {
221
+ "epoch": 10.71,
222
+ "learning_rate": 9.66e-06,
223
+ "loss": 0.0324,
224
+ "step": 825
225
+ },
226
+ {
227
+ "epoch": 11.04,
228
+ "learning_rate": 9.633684210526316e-06,
229
+ "loss": 0.0374,
230
+ "step": 850
231
+ },
232
+ {
233
+ "epoch": 11.36,
234
+ "learning_rate": 9.607368421052632e-06,
235
+ "loss": 0.0253,
236
+ "step": 875
237
+ },
238
+ {
239
+ "epoch": 11.68,
240
+ "learning_rate": 9.581052631578947e-06,
241
+ "loss": 0.0262,
242
+ "step": 900
243
+ },
244
+ {
245
+ "epoch": 12.01,
246
+ "learning_rate": 9.554736842105264e-06,
247
+ "loss": 0.0374,
248
+ "step": 925
249
+ },
250
+ {
251
+ "epoch": 12.33,
252
+ "learning_rate": 9.52842105263158e-06,
253
+ "loss": 0.0179,
254
+ "step": 950
255
+ },
256
+ {
257
+ "epoch": 12.66,
258
+ "learning_rate": 9.502105263157896e-06,
259
+ "loss": 0.0236,
260
+ "step": 975
261
+ },
262
+ {
263
+ "epoch": 12.98,
264
+ "learning_rate": 9.475789473684212e-06,
265
+ "loss": 0.0187,
266
+ "step": 1000
267
+ },
268
+ {
269
+ "epoch": 13.31,
270
+ "learning_rate": 9.449473684210527e-06,
271
+ "loss": 0.0186,
272
+ "step": 1025
273
+ },
274
+ {
275
+ "epoch": 13.63,
276
+ "learning_rate": 9.423157894736842e-06,
277
+ "loss": 0.0217,
278
+ "step": 1050
279
+ },
280
+ {
281
+ "epoch": 13.95,
282
+ "learning_rate": 9.396842105263159e-06,
283
+ "loss": 0.0206,
284
+ "step": 1075
285
+ },
286
+ {
287
+ "epoch": 14.28,
288
+ "learning_rate": 9.370526315789474e-06,
289
+ "loss": 0.0103,
290
+ "step": 1100
291
+ },
292
+ {
293
+ "epoch": 14.6,
294
+ "learning_rate": 9.34421052631579e-06,
295
+ "loss": 0.0168,
296
+ "step": 1125
297
+ },
298
+ {
299
+ "epoch": 14.93,
300
+ "learning_rate": 9.317894736842105e-06,
301
+ "loss": 0.0214,
302
+ "step": 1150
303
+ },
304
+ {
305
+ "epoch": 15.26,
306
+ "learning_rate": 9.291578947368422e-06,
307
+ "loss": 0.0115,
308
+ "step": 1175
309
+ },
310
+ {
311
+ "epoch": 15.58,
312
+ "learning_rate": 9.265263157894737e-06,
313
+ "loss": 0.0138,
314
+ "step": 1200
315
+ },
316
+ {
317
+ "epoch": 15.58,
318
+ "eval_loss": 0.8237143158912659,
319
+ "eval_runtime": 22.295,
320
+ "eval_samples_per_second": 6.235,
321
+ "eval_steps_per_second": 0.807,
322
+ "eval_wer": 117.89976133651552,
323
+ "step": 1200
324
+ },
325
+ {
326
+ "epoch": 15.9,
327
+ "learning_rate": 9.238947368421052e-06,
328
+ "loss": 0.0216,
329
+ "step": 1225
330
+ },
331
+ {
332
+ "epoch": 16.23,
333
+ "learning_rate": 9.21263157894737e-06,
334
+ "loss": 0.014,
335
+ "step": 1250
336
+ },
337
+ {
338
+ "epoch": 16.55,
339
+ "learning_rate": 9.186315789473685e-06,
340
+ "loss": 0.0117,
341
+ "step": 1275
342
+ },
343
+ {
344
+ "epoch": 16.87,
345
+ "learning_rate": 9.16e-06,
346
+ "loss": 0.0104,
347
+ "step": 1300
348
+ },
349
+ {
350
+ "epoch": 17.21,
351
+ "learning_rate": 9.133684210526317e-06,
352
+ "loss": 0.0086,
353
+ "step": 1325
354
+ },
355
+ {
356
+ "epoch": 17.53,
357
+ "learning_rate": 9.107368421052632e-06,
358
+ "loss": 0.0092,
359
+ "step": 1350
360
+ },
361
+ {
362
+ "epoch": 17.85,
363
+ "learning_rate": 9.081052631578949e-06,
364
+ "loss": 0.0098,
365
+ "step": 1375
366
+ },
367
+ {
368
+ "epoch": 18.18,
369
+ "learning_rate": 9.054736842105264e-06,
370
+ "loss": 0.0125,
371
+ "step": 1400
372
+ },
373
+ {
374
+ "epoch": 18.5,
375
+ "learning_rate": 9.02842105263158e-06,
376
+ "loss": 0.0055,
377
+ "step": 1425
378
+ },
379
+ {
380
+ "epoch": 18.82,
381
+ "learning_rate": 9.002105263157895e-06,
382
+ "loss": 0.011,
383
+ "step": 1450
384
+ },
385
+ {
386
+ "epoch": 19.15,
387
+ "learning_rate": 8.97578947368421e-06,
388
+ "loss": 0.0088,
389
+ "step": 1475
390
+ },
391
+ {
392
+ "epoch": 19.48,
393
+ "learning_rate": 8.949473684210527e-06,
394
+ "loss": 0.0065,
395
+ "step": 1500
396
+ },
397
+ {
398
+ "epoch": 19.8,
399
+ "learning_rate": 8.923157894736842e-06,
400
+ "loss": 0.0098,
401
+ "step": 1525
402
+ },
403
+ {
404
+ "epoch": 20.13,
405
+ "learning_rate": 8.896842105263159e-06,
406
+ "loss": 0.0038,
407
+ "step": 1550
408
+ },
409
+ {
410
+ "epoch": 20.45,
411
+ "learning_rate": 8.870526315789474e-06,
412
+ "loss": 0.0053,
413
+ "step": 1575
414
+ },
415
+ {
416
+ "epoch": 20.77,
417
+ "learning_rate": 8.84421052631579e-06,
418
+ "loss": 0.0075,
419
+ "step": 1600
420
+ },
421
+ {
422
+ "epoch": 20.77,
423
+ "eval_loss": 0.8213781118392944,
424
+ "eval_runtime": 18.9322,
425
+ "eval_samples_per_second": 7.342,
426
+ "eval_steps_per_second": 0.951,
427
+ "eval_wer": 103.5799522673031,
428
+ "step": 1600
429
+ },
430
+ {
431
+ "epoch": 21.1,
432
+ "learning_rate": 8.817894736842107e-06,
433
+ "loss": 0.0111,
434
+ "step": 1625
435
+ },
436
+ {
437
+ "epoch": 21.42,
438
+ "learning_rate": 8.791578947368422e-06,
439
+ "loss": 0.0083,
440
+ "step": 1650
441
+ },
442
+ {
443
+ "epoch": 21.75,
444
+ "learning_rate": 8.765263157894739e-06,
445
+ "loss": 0.0056,
446
+ "step": 1675
447
+ },
448
+ {
449
+ "epoch": 22.08,
450
+ "learning_rate": 8.738947368421053e-06,
451
+ "loss": 0.0055,
452
+ "step": 1700
453
+ },
454
+ {
455
+ "epoch": 22.4,
456
+ "learning_rate": 8.712631578947368e-06,
457
+ "loss": 0.0031,
458
+ "step": 1725
459
+ },
460
+ {
461
+ "epoch": 22.72,
462
+ "learning_rate": 8.686315789473685e-06,
463
+ "loss": 0.0075,
464
+ "step": 1750
465
+ },
466
+ {
467
+ "epoch": 23.05,
468
+ "learning_rate": 8.66e-06,
469
+ "loss": 0.0078,
470
+ "step": 1775
471
+ },
472
+ {
473
+ "epoch": 23.37,
474
+ "learning_rate": 8.633684210526317e-06,
475
+ "loss": 0.0063,
476
+ "step": 1800
477
+ },
478
+ {
479
+ "epoch": 23.69,
480
+ "learning_rate": 8.607368421052632e-06,
481
+ "loss": 0.0012,
482
+ "step": 1825
483
+ },
484
+ {
485
+ "epoch": 24.03,
486
+ "learning_rate": 8.581052631578948e-06,
487
+ "loss": 0.0045,
488
+ "step": 1850
489
+ },
490
+ {
491
+ "epoch": 24.35,
492
+ "learning_rate": 8.554736842105263e-06,
493
+ "loss": 0.004,
494
+ "step": 1875
495
+ },
496
+ {
497
+ "epoch": 24.67,
498
+ "learning_rate": 8.528421052631578e-06,
499
+ "loss": 0.006,
500
+ "step": 1900
501
+ },
502
+ {
503
+ "epoch": 24.99,
504
+ "learning_rate": 8.502105263157897e-06,
505
+ "loss": 0.0034,
506
+ "step": 1925
507
+ },
508
+ {
509
+ "epoch": 25.32,
510
+ "learning_rate": 8.475789473684212e-06,
511
+ "loss": 0.0024,
512
+ "step": 1950
513
+ },
514
+ {
515
+ "epoch": 25.64,
516
+ "learning_rate": 8.449473684210527e-06,
517
+ "loss": 0.0032,
518
+ "step": 1975
519
+ },
520
+ {
521
+ "epoch": 25.96,
522
+ "learning_rate": 8.423157894736843e-06,
523
+ "loss": 0.0021,
524
+ "step": 2000
525
+ },
526
+ {
527
+ "epoch": 25.96,
528
+ "eval_loss": 0.7943580746650696,
529
+ "eval_runtime": 20.6652,
530
+ "eval_samples_per_second": 6.726,
531
+ "eval_steps_per_second": 0.871,
532
+ "eval_wer": 106.92124105011933,
533
+ "step": 2000
534
+ }
535
+ ],
536
+ "max_steps": 10000,
537
+ "num_train_epochs": 130,
538
+ "total_flos": 3.2965538512896e+19,
539
+ "trial_name": null,
540
+ "trial_params": null
541
+ }
checkpoint-2000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4ad917c70902fd83abc5a9152e70fd07531fce940ccd66f6af7f3bc05e82880
3
+ size 4088
checkpoint-2400/config.json ADDED
@@ -0,0 +1,144 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "openai/whisper-medium",
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "gelu",
5
+ "architectures": [
6
+ "WhisperForConditionalGeneration"
7
+ ],
8
+ "attention_dropout": 0.0,
9
+ "begin_suppress_tokens": [
10
+ 220,
11
+ 50257
12
+ ],
13
+ "bos_token_id": 50257,
14
+ "d_model": 1024,
15
+ "decoder_attention_heads": 16,
16
+ "decoder_ffn_dim": 4096,
17
+ "decoder_layerdrop": 0.0,
18
+ "decoder_layers": 24,
19
+ "decoder_start_token_id": 50258,
20
+ "dropout": 0.0,
21
+ "encoder_attention_heads": 16,
22
+ "encoder_ffn_dim": 4096,
23
+ "encoder_layerdrop": 0.0,
24
+ "encoder_layers": 24,
25
+ "eos_token_id": 50257,
26
+ "forced_decoder_ids": [
27
+ [
28
+ 1,
29
+ 50259
30
+ ],
31
+ [
32
+ 2,
33
+ 50359
34
+ ],
35
+ [
36
+ 3,
37
+ 50363
38
+ ]
39
+ ],
40
+ "init_std": 0.02,
41
+ "is_encoder_decoder": true,
42
+ "max_length": 448,
43
+ "max_source_positions": 1500,
44
+ "max_target_positions": 448,
45
+ "model_type": "whisper",
46
+ "num_hidden_layers": 24,
47
+ "num_mel_bins": 80,
48
+ "pad_token_id": 50257,
49
+ "scale_embedding": false,
50
+ "suppress_tokens": [
51
+ 1,
52
+ 2,
53
+ 7,
54
+ 8,
55
+ 9,
56
+ 10,
57
+ 14,
58
+ 25,
59
+ 26,
60
+ 27,
61
+ 28,
62
+ 29,
63
+ 31,
64
+ 58,
65
+ 59,
66
+ 60,
67
+ 61,
68
+ 62,
69
+ 63,
70
+ 90,
71
+ 91,
72
+ 92,
73
+ 93,
74
+ 359,
75
+ 503,
76
+ 522,
77
+ 542,
78
+ 873,
79
+ 893,
80
+ 902,
81
+ 918,
82
+ 922,
83
+ 931,
84
+ 1350,
85
+ 1853,
86
+ 1982,
87
+ 2460,
88
+ 2627,
89
+ 3246,
90
+ 3253,
91
+ 3268,
92
+ 3536,
93
+ 3846,
94
+ 3961,
95
+ 4183,
96
+ 4667,
97
+ 6585,
98
+ 6647,
99
+ 7273,
100
+ 9061,
101
+ 9383,
102
+ 10428,
103
+ 10929,
104
+ 11938,
105
+ 12033,
106
+ 12331,
107
+ 12562,
108
+ 13793,
109
+ 14157,
110
+ 14635,
111
+ 15265,
112
+ 15618,
113
+ 16553,
114
+ 16604,
115
+ 18362,
116
+ 18956,
117
+ 20075,
118
+ 21675,
119
+ 22520,
120
+ 26130,
121
+ 26161,
122
+ 26435,
123
+ 28279,
124
+ 29464,
125
+ 31650,
126
+ 32302,
127
+ 32470,
128
+ 36865,
129
+ 42863,
130
+ 47425,
131
+ 49870,
132
+ 50254,
133
+ 50258,
134
+ 50358,
135
+ 50359,
136
+ 50360,
137
+ 50361,
138
+ 50362
139
+ ],
140
+ "torch_dtype": "float32",
141
+ "transformers_version": "4.26.0",
142
+ "use_cache": true,
143
+ "vocab_size": 51865
144
+ }
checkpoint-2400/generation_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "begin_suppress_tokens": [
3
+ 220,
4
+ 50257
5
+ ],
6
+ "bos_token_id": 50257,
7
+ "decoder_start_token_id": 50258,
8
+ "eos_token_id": 50257,
9
+ "language": "spanish",
10
+ "max_length": 448,
11
+ "pad_token_id": 50257,
12
+ "suppress_tokens": [
13
+ 1,
14
+ 2,
15
+ 7,
16
+ 8,
17
+ 9,
18
+ 10,
19
+ 14,
20
+ 25,
21
+ 26,
22
+ 27,
23
+ 28,
24
+ 29,
25
+ 31,
26
+ 58,
27
+ 59,
28
+ 60,
29
+ 61,
30
+ 62,
31
+ 63,
32
+ 90,
33
+ 91,
34
+ 92,
35
+ 93,
36
+ 359,
37
+ 503,
38
+ 522,
39
+ 542,
40
+ 873,
41
+ 893,
42
+ 902,
43
+ 918,
44
+ 922,
45
+ 931,
46
+ 1350,
47
+ 1853,
48
+ 1982,
49
+ 2460,
50
+ 2627,
51
+ 3246,
52
+ 3253,
53
+ 3268,
54
+ 3536,
55
+ 3846,
56
+ 3961,
57
+ 4183,
58
+ 4667,
59
+ 6585,
60
+ 6647,
61
+ 7273,
62
+ 9061,
63
+ 9383,
64
+ 10428,
65
+ 10929,
66
+ 11938,
67
+ 12033,
68
+ 12331,
69
+ 12562,
70
+ 13793,
71
+ 14157,
72
+ 14635,
73
+ 15265,
74
+ 15618,
75
+ 16553,
76
+ 16604,
77
+ 18362,
78
+ 18956,
79
+ 20075,
80
+ 21675,
81
+ 22520,
82
+ 26130,
83
+ 26161,
84
+ 26435,
85
+ 28279,
86
+ 29464,
87
+ 31650,
88
+ 32302,
89
+ 32470,
90
+ 36865,
91
+ 42863,
92
+ 47425,
93
+ 49870,
94
+ 50254,
95
+ 50258,
96
+ 50358,
97
+ 50359,
98
+ 50360,
99
+ 50361,
100
+ 50362
101
+ ],
102
+ "task": "transcribe",
103
+ "transformers_version": "4.26.0"
104
+ }
checkpoint-2400/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d7e36c158e9c5793209c5894c7df152c0be0996bc5f391ad77c6ba754ff8df4
3
+ size 6111429144
checkpoint-2400/preprocessor_config.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-2400/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:702595ea2357de9688d5501134b5e81ba5698f204e41459032692c1bf5564556
3
+ size 3055755286
checkpoint-2400/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb6587ca15147eccce7fd6e8a67ccd1ad442a2bf0c6a52209504b7e5ce096cbb
3
+ size 14244
checkpoint-2400/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:addfb63cbfb1de3600775f954351a3baba62a2c733f97a5fd3a38405b2fb718f
3
+ size 988
checkpoint-2400/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f561268a5d80fb39bff65130dc6ed4c1fbcf80673f818359c343b2d496191792
3
+ size 1064
checkpoint-2400/trainer_state.json ADDED
@@ -0,0 +1,646 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 98.68735083532219,
3
+ "best_model_checkpoint": "ASFEMA_SEV/checkpoint-2400",
4
+ "epoch": 31.167202572347268,
5
+ "global_step": 2400,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.32,
12
+ "learning_rate": 4.6000000000000004e-07,
13
+ "loss": 4.0894,
14
+ "step": 25
15
+ },
16
+ {
17
+ "epoch": 0.64,
18
+ "learning_rate": 9.600000000000001e-07,
19
+ "loss": 3.7,
20
+ "step": 50
21
+ },
22
+ {
23
+ "epoch": 0.96,
24
+ "learning_rate": 1.46e-06,
25
+ "loss": 3.0759,
26
+ "step": 75
27
+ },
28
+ {
29
+ "epoch": 1.3,
30
+ "learning_rate": 1.9600000000000003e-06,
31
+ "loss": 2.5729,
32
+ "step": 100
33
+ },
34
+ {
35
+ "epoch": 1.62,
36
+ "learning_rate": 2.46e-06,
37
+ "loss": 2.1885,
38
+ "step": 125
39
+ },
40
+ {
41
+ "epoch": 1.94,
42
+ "learning_rate": 2.96e-06,
43
+ "loss": 1.9431,
44
+ "step": 150
45
+ },
46
+ {
47
+ "epoch": 2.27,
48
+ "learning_rate": 3.46e-06,
49
+ "loss": 1.5753,
50
+ "step": 175
51
+ },
52
+ {
53
+ "epoch": 2.59,
54
+ "learning_rate": 3.96e-06,
55
+ "loss": 1.3319,
56
+ "step": 200
57
+ },
58
+ {
59
+ "epoch": 2.91,
60
+ "learning_rate": 4.4600000000000005e-06,
61
+ "loss": 1.2076,
62
+ "step": 225
63
+ },
64
+ {
65
+ "epoch": 3.24,
66
+ "learning_rate": 4.960000000000001e-06,
67
+ "loss": 0.9211,
68
+ "step": 250
69
+ },
70
+ {
71
+ "epoch": 3.57,
72
+ "learning_rate": 5.460000000000001e-06,
73
+ "loss": 0.7332,
74
+ "step": 275
75
+ },
76
+ {
77
+ "epoch": 3.89,
78
+ "learning_rate": 5.9600000000000005e-06,
79
+ "loss": 0.6601,
80
+ "step": 300
81
+ },
82
+ {
83
+ "epoch": 4.22,
84
+ "learning_rate": 6.460000000000001e-06,
85
+ "loss": 0.4824,
86
+ "step": 325
87
+ },
88
+ {
89
+ "epoch": 4.54,
90
+ "learning_rate": 6.96e-06,
91
+ "loss": 0.3455,
92
+ "step": 350
93
+ },
94
+ {
95
+ "epoch": 4.86,
96
+ "learning_rate": 7.4600000000000006e-06,
97
+ "loss": 0.3741,
98
+ "step": 375
99
+ },
100
+ {
101
+ "epoch": 5.19,
102
+ "learning_rate": 7.960000000000002e-06,
103
+ "loss": 0.2448,
104
+ "step": 400
105
+ },
106
+ {
107
+ "epoch": 5.19,
108
+ "eval_loss": 0.9975764751434326,
109
+ "eval_runtime": 24.147,
110
+ "eval_samples_per_second": 5.756,
111
+ "eval_steps_per_second": 0.745,
112
+ "eval_wer": 132.57756563245823,
113
+ "step": 400
114
+ },
115
+ {
116
+ "epoch": 5.51,
117
+ "learning_rate": 8.46e-06,
118
+ "loss": 0.1713,
119
+ "step": 425
120
+ },
121
+ {
122
+ "epoch": 5.84,
123
+ "learning_rate": 8.96e-06,
124
+ "loss": 0.1906,
125
+ "step": 450
126
+ },
127
+ {
128
+ "epoch": 6.17,
129
+ "learning_rate": 9.460000000000001e-06,
130
+ "loss": 0.1479,
131
+ "step": 475
132
+ },
133
+ {
134
+ "epoch": 6.49,
135
+ "learning_rate": 9.960000000000001e-06,
136
+ "loss": 0.103,
137
+ "step": 500
138
+ },
139
+ {
140
+ "epoch": 6.81,
141
+ "learning_rate": 9.975789473684211e-06,
142
+ "loss": 0.1271,
143
+ "step": 525
144
+ },
145
+ {
146
+ "epoch": 7.14,
147
+ "learning_rate": 9.949473684210526e-06,
148
+ "loss": 0.1033,
149
+ "step": 550
150
+ },
151
+ {
152
+ "epoch": 7.46,
153
+ "learning_rate": 9.923157894736844e-06,
154
+ "loss": 0.0799,
155
+ "step": 575
156
+ },
157
+ {
158
+ "epoch": 7.78,
159
+ "learning_rate": 9.89684210526316e-06,
160
+ "loss": 0.0814,
161
+ "step": 600
162
+ },
163
+ {
164
+ "epoch": 8.12,
165
+ "learning_rate": 9.870526315789474e-06,
166
+ "loss": 0.069,
167
+ "step": 625
168
+ },
169
+ {
170
+ "epoch": 8.44,
171
+ "learning_rate": 9.84421052631579e-06,
172
+ "loss": 0.0546,
173
+ "step": 650
174
+ },
175
+ {
176
+ "epoch": 8.76,
177
+ "learning_rate": 9.817894736842106e-06,
178
+ "loss": 0.0627,
179
+ "step": 675
180
+ },
181
+ {
182
+ "epoch": 9.09,
183
+ "learning_rate": 9.791578947368422e-06,
184
+ "loss": 0.0459,
185
+ "step": 700
186
+ },
187
+ {
188
+ "epoch": 9.41,
189
+ "learning_rate": 9.765263157894737e-06,
190
+ "loss": 0.034,
191
+ "step": 725
192
+ },
193
+ {
194
+ "epoch": 9.73,
195
+ "learning_rate": 9.738947368421054e-06,
196
+ "loss": 0.036,
197
+ "step": 750
198
+ },
199
+ {
200
+ "epoch": 10.06,
201
+ "learning_rate": 9.712631578947369e-06,
202
+ "loss": 0.0367,
203
+ "step": 775
204
+ },
205
+ {
206
+ "epoch": 10.39,
207
+ "learning_rate": 9.686315789473684e-06,
208
+ "loss": 0.0248,
209
+ "step": 800
210
+ },
211
+ {
212
+ "epoch": 10.39,
213
+ "eval_loss": 0.8046127557754517,
214
+ "eval_runtime": 19.9036,
215
+ "eval_samples_per_second": 6.984,
216
+ "eval_steps_per_second": 0.904,
217
+ "eval_wer": 108.59188544152745,
218
+ "step": 800
219
+ },
220
+ {
221
+ "epoch": 10.71,
222
+ "learning_rate": 9.66e-06,
223
+ "loss": 0.0324,
224
+ "step": 825
225
+ },
226
+ {
227
+ "epoch": 11.04,
228
+ "learning_rate": 9.633684210526316e-06,
229
+ "loss": 0.0374,
230
+ "step": 850
231
+ },
232
+ {
233
+ "epoch": 11.36,
234
+ "learning_rate": 9.607368421052632e-06,
235
+ "loss": 0.0253,
236
+ "step": 875
237
+ },
238
+ {
239
+ "epoch": 11.68,
240
+ "learning_rate": 9.581052631578947e-06,
241
+ "loss": 0.0262,
242
+ "step": 900
243
+ },
244
+ {
245
+ "epoch": 12.01,
246
+ "learning_rate": 9.554736842105264e-06,
247
+ "loss": 0.0374,
248
+ "step": 925
249
+ },
250
+ {
251
+ "epoch": 12.33,
252
+ "learning_rate": 9.52842105263158e-06,
253
+ "loss": 0.0179,
254
+ "step": 950
255
+ },
256
+ {
257
+ "epoch": 12.66,
258
+ "learning_rate": 9.502105263157896e-06,
259
+ "loss": 0.0236,
260
+ "step": 975
261
+ },
262
+ {
263
+ "epoch": 12.98,
264
+ "learning_rate": 9.475789473684212e-06,
265
+ "loss": 0.0187,
266
+ "step": 1000
267
+ },
268
+ {
269
+ "epoch": 13.31,
270
+ "learning_rate": 9.449473684210527e-06,
271
+ "loss": 0.0186,
272
+ "step": 1025
273
+ },
274
+ {
275
+ "epoch": 13.63,
276
+ "learning_rate": 9.423157894736842e-06,
277
+ "loss": 0.0217,
278
+ "step": 1050
279
+ },
280
+ {
281
+ "epoch": 13.95,
282
+ "learning_rate": 9.396842105263159e-06,
283
+ "loss": 0.0206,
284
+ "step": 1075
285
+ },
286
+ {
287
+ "epoch": 14.28,
288
+ "learning_rate": 9.370526315789474e-06,
289
+ "loss": 0.0103,
290
+ "step": 1100
291
+ },
292
+ {
293
+ "epoch": 14.6,
294
+ "learning_rate": 9.34421052631579e-06,
295
+ "loss": 0.0168,
296
+ "step": 1125
297
+ },
298
+ {
299
+ "epoch": 14.93,
300
+ "learning_rate": 9.317894736842105e-06,
301
+ "loss": 0.0214,
302
+ "step": 1150
303
+ },
304
+ {
305
+ "epoch": 15.26,
306
+ "learning_rate": 9.291578947368422e-06,
307
+ "loss": 0.0115,
308
+ "step": 1175
309
+ },
310
+ {
311
+ "epoch": 15.58,
312
+ "learning_rate": 9.265263157894737e-06,
313
+ "loss": 0.0138,
314
+ "step": 1200
315
+ },
316
+ {
317
+ "epoch": 15.58,
318
+ "eval_loss": 0.8237143158912659,
319
+ "eval_runtime": 22.295,
320
+ "eval_samples_per_second": 6.235,
321
+ "eval_steps_per_second": 0.807,
322
+ "eval_wer": 117.89976133651552,
323
+ "step": 1200
324
+ },
325
+ {
326
+ "epoch": 15.9,
327
+ "learning_rate": 9.238947368421052e-06,
328
+ "loss": 0.0216,
329
+ "step": 1225
330
+ },
331
+ {
332
+ "epoch": 16.23,
333
+ "learning_rate": 9.21263157894737e-06,
334
+ "loss": 0.014,
335
+ "step": 1250
336
+ },
337
+ {
338
+ "epoch": 16.55,
339
+ "learning_rate": 9.186315789473685e-06,
340
+ "loss": 0.0117,
341
+ "step": 1275
342
+ },
343
+ {
344
+ "epoch": 16.87,
345
+ "learning_rate": 9.16e-06,
346
+ "loss": 0.0104,
347
+ "step": 1300
348
+ },
349
+ {
350
+ "epoch": 17.21,
351
+ "learning_rate": 9.133684210526317e-06,
352
+ "loss": 0.0086,
353
+ "step": 1325
354
+ },
355
+ {
356
+ "epoch": 17.53,
357
+ "learning_rate": 9.107368421052632e-06,
358
+ "loss": 0.0092,
359
+ "step": 1350
360
+ },
361
+ {
362
+ "epoch": 17.85,
363
+ "learning_rate": 9.081052631578949e-06,
364
+ "loss": 0.0098,
365
+ "step": 1375
366
+ },
367
+ {
368
+ "epoch": 18.18,
369
+ "learning_rate": 9.054736842105264e-06,
370
+ "loss": 0.0125,
371
+ "step": 1400
372
+ },
373
+ {
374
+ "epoch": 18.5,
375
+ "learning_rate": 9.02842105263158e-06,
376
+ "loss": 0.0055,
377
+ "step": 1425
378
+ },
379
+ {
380
+ "epoch": 18.82,
381
+ "learning_rate": 9.002105263157895e-06,
382
+ "loss": 0.011,
383
+ "step": 1450
384
+ },
385
+ {
386
+ "epoch": 19.15,
387
+ "learning_rate": 8.97578947368421e-06,
388
+ "loss": 0.0088,
389
+ "step": 1475
390
+ },
391
+ {
392
+ "epoch": 19.48,
393
+ "learning_rate": 8.949473684210527e-06,
394
+ "loss": 0.0065,
395
+ "step": 1500
396
+ },
397
+ {
398
+ "epoch": 19.8,
399
+ "learning_rate": 8.923157894736842e-06,
400
+ "loss": 0.0098,
401
+ "step": 1525
402
+ },
403
+ {
404
+ "epoch": 20.13,
405
+ "learning_rate": 8.896842105263159e-06,
406
+ "loss": 0.0038,
407
+ "step": 1550
408
+ },
409
+ {
410
+ "epoch": 20.45,
411
+ "learning_rate": 8.870526315789474e-06,
412
+ "loss": 0.0053,
413
+ "step": 1575
414
+ },
415
+ {
416
+ "epoch": 20.77,
417
+ "learning_rate": 8.84421052631579e-06,
418
+ "loss": 0.0075,
419
+ "step": 1600
420
+ },
421
+ {
422
+ "epoch": 20.77,
423
+ "eval_loss": 0.8213781118392944,
424
+ "eval_runtime": 18.9322,
425
+ "eval_samples_per_second": 7.342,
426
+ "eval_steps_per_second": 0.951,
427
+ "eval_wer": 103.5799522673031,
428
+ "step": 1600
429
+ },
430
+ {
431
+ "epoch": 21.1,
432
+ "learning_rate": 8.817894736842107e-06,
433
+ "loss": 0.0111,
434
+ "step": 1625
435
+ },
436
+ {
437
+ "epoch": 21.42,
438
+ "learning_rate": 8.791578947368422e-06,
439
+ "loss": 0.0083,
440
+ "step": 1650
441
+ },
442
+ {
443
+ "epoch": 21.75,
444
+ "learning_rate": 8.765263157894739e-06,
445
+ "loss": 0.0056,
446
+ "step": 1675
447
+ },
448
+ {
449
+ "epoch": 22.08,
450
+ "learning_rate": 8.738947368421053e-06,
451
+ "loss": 0.0055,
452
+ "step": 1700
453
+ },
454
+ {
455
+ "epoch": 22.4,
456
+ "learning_rate": 8.712631578947368e-06,
457
+ "loss": 0.0031,
458
+ "step": 1725
459
+ },
460
+ {
461
+ "epoch": 22.72,
462
+ "learning_rate": 8.686315789473685e-06,
463
+ "loss": 0.0075,
464
+ "step": 1750
465
+ },
466
+ {
467
+ "epoch": 23.05,
468
+ "learning_rate": 8.66e-06,
469
+ "loss": 0.0078,
470
+ "step": 1775
471
+ },
472
+ {
473
+ "epoch": 23.37,
474
+ "learning_rate": 8.633684210526317e-06,
475
+ "loss": 0.0063,
476
+ "step": 1800
477
+ },
478
+ {
479
+ "epoch": 23.69,
480
+ "learning_rate": 8.607368421052632e-06,
481
+ "loss": 0.0012,
482
+ "step": 1825
483
+ },
484
+ {
485
+ "epoch": 24.03,
486
+ "learning_rate": 8.581052631578948e-06,
487
+ "loss": 0.0045,
488
+ "step": 1850
489
+ },
490
+ {
491
+ "epoch": 24.35,
492
+ "learning_rate": 8.554736842105263e-06,
493
+ "loss": 0.004,
494
+ "step": 1875
495
+ },
496
+ {
497
+ "epoch": 24.67,
498
+ "learning_rate": 8.528421052631578e-06,
499
+ "loss": 0.006,
500
+ "step": 1900
501
+ },
502
+ {
503
+ "epoch": 24.99,
504
+ "learning_rate": 8.502105263157897e-06,
505
+ "loss": 0.0034,
506
+ "step": 1925
507
+ },
508
+ {
509
+ "epoch": 25.32,
510
+ "learning_rate": 8.475789473684212e-06,
511
+ "loss": 0.0024,
512
+ "step": 1950
513
+ },
514
+ {
515
+ "epoch": 25.64,
516
+ "learning_rate": 8.449473684210527e-06,
517
+ "loss": 0.0032,
518
+ "step": 1975
519
+ },
520
+ {
521
+ "epoch": 25.96,
522
+ "learning_rate": 8.423157894736843e-06,
523
+ "loss": 0.0021,
524
+ "step": 2000
525
+ },
526
+ {
527
+ "epoch": 25.96,
528
+ "eval_loss": 0.7943580746650696,
529
+ "eval_runtime": 20.6652,
530
+ "eval_samples_per_second": 6.726,
531
+ "eval_steps_per_second": 0.871,
532
+ "eval_wer": 106.92124105011933,
533
+ "step": 2000
534
+ },
535
+ {
536
+ "epoch": 26.3,
537
+ "learning_rate": 8.396842105263158e-06,
538
+ "loss": 0.0006,
539
+ "step": 2025
540
+ },
541
+ {
542
+ "epoch": 26.62,
543
+ "learning_rate": 8.370526315789475e-06,
544
+ "loss": 0.0018,
545
+ "step": 2050
546
+ },
547
+ {
548
+ "epoch": 26.94,
549
+ "learning_rate": 8.34421052631579e-06,
550
+ "loss": 0.0032,
551
+ "step": 2075
552
+ },
553
+ {
554
+ "epoch": 27.27,
555
+ "learning_rate": 8.317894736842107e-06,
556
+ "loss": 0.0034,
557
+ "step": 2100
558
+ },
559
+ {
560
+ "epoch": 27.59,
561
+ "learning_rate": 8.291578947368422e-06,
562
+ "loss": 0.0021,
563
+ "step": 2125
564
+ },
565
+ {
566
+ "epoch": 27.91,
567
+ "learning_rate": 8.265263157894737e-06,
568
+ "loss": 0.005,
569
+ "step": 2150
570
+ },
571
+ {
572
+ "epoch": 28.24,
573
+ "learning_rate": 8.238947368421053e-06,
574
+ "loss": 0.002,
575
+ "step": 2175
576
+ },
577
+ {
578
+ "epoch": 28.57,
579
+ "learning_rate": 8.212631578947368e-06,
580
+ "loss": 0.0023,
581
+ "step": 2200
582
+ },
583
+ {
584
+ "epoch": 28.89,
585
+ "learning_rate": 8.186315789473685e-06,
586
+ "loss": 0.0126,
587
+ "step": 2225
588
+ },
589
+ {
590
+ "epoch": 29.22,
591
+ "learning_rate": 8.16e-06,
592
+ "loss": 0.0064,
593
+ "step": 2250
594
+ },
595
+ {
596
+ "epoch": 29.54,
597
+ "learning_rate": 8.133684210526316e-06,
598
+ "loss": 0.004,
599
+ "step": 2275
600
+ },
601
+ {
602
+ "epoch": 29.86,
603
+ "learning_rate": 8.107368421052633e-06,
604
+ "loss": 0.003,
605
+ "step": 2300
606
+ },
607
+ {
608
+ "epoch": 30.19,
609
+ "learning_rate": 8.081052631578948e-06,
610
+ "loss": 0.0064,
611
+ "step": 2325
612
+ },
613
+ {
614
+ "epoch": 30.51,
615
+ "learning_rate": 8.054736842105265e-06,
616
+ "loss": 0.0038,
617
+ "step": 2350
618
+ },
619
+ {
620
+ "epoch": 30.84,
621
+ "learning_rate": 8.02842105263158e-06,
622
+ "loss": 0.006,
623
+ "step": 2375
624
+ },
625
+ {
626
+ "epoch": 31.17,
627
+ "learning_rate": 8.002105263157895e-06,
628
+ "loss": 0.003,
629
+ "step": 2400
630
+ },
631
+ {
632
+ "epoch": 31.17,
633
+ "eval_loss": 0.811363160610199,
634
+ "eval_runtime": 17.6674,
635
+ "eval_samples_per_second": 7.868,
636
+ "eval_steps_per_second": 1.019,
637
+ "eval_wer": 98.68735083532219,
638
+ "step": 2400
639
+ }
640
+ ],
641
+ "max_steps": 10000,
642
+ "num_train_epochs": 130,
643
+ "total_flos": 3.957089347436544e+19,
644
+ "trial_name": null,
645
+ "trial_params": null
646
+ }