dongim04 commited on
Commit
51a9fa3
·
verified ·
1 Parent(s): 78088b8

End of training

Browse files
README.md CHANGED
@@ -4,8 +4,6 @@ license: apache-2.0
4
  base_model: openai/whisper-base
5
  tags:
6
  - generated_from_trainer
7
- metrics:
8
- - wer
9
  model-index:
10
  - name: whisper-base-zh
11
  results: []
@@ -18,8 +16,13 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [openai/whisper-base](https://huggingface.co/openai/whisper-base) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.4316
22
- - Wer: 85.1032
 
 
 
 
 
23
 
24
  ## Model description
25
 
@@ -38,64 +41,19 @@ More information needed
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
41
- - learning_rate: 1e-05
42
  - train_batch_size: 16
43
  - eval_batch_size: 8
44
  - seed: 42
45
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
 
47
  - training_steps: 4000
48
  - mixed_precision_training: Native AMP
49
 
50
- ### Training results
51
-
52
- | Training Loss | Epoch | Step | Validation Loss | Wer |
53
- |:-------------:|:------:|:----:|:---------------:|:-------:|
54
- | 0.5085 | 0.2070 | 100 | 0.4641 | 89.6326 |
55
- | 0.4047 | 0.4141 | 200 | 0.4315 | 88.2235 |
56
- | 0.4156 | 0.6211 | 300 | 0.4055 | 86.8646 |
57
- | 0.3826 | 0.8282 | 400 | 0.3907 | 87.2169 |
58
- | 0.3087 | 1.0352 | 500 | 0.3806 | 86.9149 |
59
- | 0.2709 | 1.2422 | 600 | 0.3775 | 87.0659 |
60
- | 0.2707 | 1.4493 | 700 | 0.3681 | 85.7071 |
61
- | 0.273 | 1.6563 | 800 | 0.3633 | 85.5561 |
62
- | 0.2638 | 1.8634 | 900 | 0.3592 | 85.4051 |
63
- | 0.1698 | 2.0704 | 1000 | 0.3613 | 85.6568 |
64
- | 0.1849 | 2.2774 | 1100 | 0.3628 | 85.7574 |
65
- | 0.1819 | 2.4845 | 1200 | 0.3645 | 85.4555 |
66
- | 0.1712 | 2.6915 | 1300 | 0.3611 | 84.8515 |
67
- | 0.1768 | 2.8986 | 1400 | 0.3573 | 84.3986 |
68
- | 0.1265 | 3.1056 | 1500 | 0.3640 | 85.2038 |
69
- | 0.1275 | 3.3126 | 1600 | 0.3661 | 85.4051 |
70
- | 0.1277 | 3.5197 | 1700 | 0.3698 | 84.3986 |
71
- | 0.1195 | 3.7267 | 1800 | 0.3684 | 85.3045 |
72
- | 0.1176 | 3.9337 | 1900 | 0.3668 | 84.9522 |
73
- | 0.0843 | 4.1408 | 2000 | 0.3764 | 84.9522 |
74
- | 0.077 | 4.3478 | 2100 | 0.3801 | 84.8012 |
75
- | 0.0853 | 4.5549 | 2200 | 0.3804 | 85.4051 |
76
- | 0.0826 | 4.7619 | 2300 | 0.3827 | 84.8515 |
77
- | 0.0759 | 4.9689 | 2400 | 0.3803 | 84.9522 |
78
- | 0.0519 | 5.1760 | 2500 | 0.3959 | 84.9522 |
79
- | 0.0578 | 5.3830 | 2600 | 0.3946 | 85.0528 |
80
- | 0.0594 | 5.5901 | 2700 | 0.3956 | 85.0528 |
81
- | 0.0598 | 5.7971 | 2800 | 0.3971 | 85.1535 |
82
- | 0.0519 | 6.0041 | 2900 | 0.3980 | 84.7509 |
83
- | 0.0421 | 6.2112 | 3000 | 0.4103 | 85.5058 |
84
- | 0.0379 | 6.4182 | 3100 | 0.4124 | 84.9019 |
85
- | 0.0363 | 6.6253 | 3200 | 0.4131 | 84.9019 |
86
- | 0.0377 | 6.8323 | 3300 | 0.4149 | 85.2038 |
87
- | 0.0286 | 7.0393 | 3400 | 0.4205 | 84.8515 |
88
- | 0.0263 | 7.2464 | 3500 | 0.4257 | 84.6502 |
89
- | 0.0309 | 7.4534 | 3600 | 0.4276 | 85.3548 |
90
- | 0.0295 | 7.6605 | 3700 | 0.4282 | 85.3045 |
91
- | 0.0266 | 7.8675 | 3800 | 0.4291 | 85.2038 |
92
- | 0.0241 | 8.0745 | 3900 | 0.4308 | 85.1535 |
93
- | 0.025 | 8.2816 | 4000 | 0.4316 | 85.1032 |
94
-
95
-
96
  ### Framework versions
97
 
98
- - Transformers 4.44.2
99
- - Pytorch 2.5.0+cu121
100
  - Datasets 3.1.0
101
- - Tokenizers 0.19.1
 
4
  base_model: openai/whisper-base
5
  tags:
6
  - generated_from_trainer
 
 
7
  model-index:
8
  - name: whisper-base-zh
9
  results: []
 
16
 
17
  This model is a fine-tuned version of [openai/whisper-base](https://huggingface.co/openai/whisper-base) on the None dataset.
18
  It achieves the following results on the evaluation set:
19
+ - eval_loss: 0.3523
20
+ - eval_wer: 77.0525
21
+ - eval_runtime: 533.856
22
+ - eval_samples_per_second: 3.606
23
+ - eval_steps_per_second: 0.451
24
+ - epoch: 4.7718
25
+ - step: 2300
26
 
27
  ## Model description
28
 
 
41
  ### Training hyperparameters
42
 
43
  The following hyperparameters were used during training:
44
+ - learning_rate: 5e-06
45
  - train_batch_size: 16
46
  - eval_batch_size: 8
47
  - seed: 42
48
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
49
  - lr_scheduler_type: linear
50
+ - lr_scheduler_warmup_steps: 100
51
  - training_steps: 4000
52
  - mixed_precision_training: Native AMP
53
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
54
  ### Framework versions
55
 
56
+ - Transformers 4.46.2
57
+ - Pytorch 2.5.1+cu121
58
  - Datasets 3.1.0
59
+ - Tokenizers 0.20.3
generation_config.json CHANGED
@@ -158,99 +158,10 @@
158
  "pad_token_id": 50257,
159
  "prev_sot_token_id": 50361,
160
  "return_timestamps": false,
161
- "suppress_tokens": [
162
- 1,
163
- 2,
164
- 7,
165
- 8,
166
- 9,
167
- 10,
168
- 14,
169
- 25,
170
- 26,
171
- 27,
172
- 28,
173
- 29,
174
- 31,
175
- 58,
176
- 59,
177
- 60,
178
- 61,
179
- 62,
180
- 63,
181
- 90,
182
- 91,
183
- 92,
184
- 93,
185
- 359,
186
- 503,
187
- 522,
188
- 542,
189
- 873,
190
- 893,
191
- 902,
192
- 918,
193
- 922,
194
- 931,
195
- 1350,
196
- 1853,
197
- 1982,
198
- 2460,
199
- 2627,
200
- 3246,
201
- 3253,
202
- 3268,
203
- 3536,
204
- 3846,
205
- 3961,
206
- 4183,
207
- 4667,
208
- 6585,
209
- 6647,
210
- 7273,
211
- 9061,
212
- 9383,
213
- 10428,
214
- 10929,
215
- 11938,
216
- 12033,
217
- 12331,
218
- 12562,
219
- 13793,
220
- 14157,
221
- 14635,
222
- 15265,
223
- 15618,
224
- 16553,
225
- 16604,
226
- 18362,
227
- 18956,
228
- 20075,
229
- 21675,
230
- 22520,
231
- 26130,
232
- 26161,
233
- 26435,
234
- 28279,
235
- 29464,
236
- 31650,
237
- 32302,
238
- 32470,
239
- 36865,
240
- 42863,
241
- 47425,
242
- 49870,
243
- 50254,
244
- 50258,
245
- 50358,
246
- 50359,
247
- 50360,
248
- 50361,
249
- 50362
250
- ],
251
  "task_to_id": {
252
  "transcribe": 50359,
253
  "translate": 50358
254
  },
255
- "transformers_version": "4.44.2"
256
  }
 
158
  "pad_token_id": 50257,
159
  "prev_sot_token_id": 50361,
160
  "return_timestamps": false,
161
+ "suppress_tokens": [],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
162
  "task_to_id": {
163
  "transcribe": 50359,
164
  "translate": 50358
165
  },
166
+ "transformers_version": "4.46.2"
167
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e8a239321bd82d63104ac851bd9c451c442dec668a3179e9656f6da43b9e7e0
3
  size 290403936
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1580bed24f3d2b9ce788254e3db0f83017356ca8188b018e5f4b3699436816dd
3
  size 290403936
runs/Nov16_20-57-52_1e51c95c58f0/events.out.tfevents.1731790675.1e51c95c58f0.2060.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e4e0dc118e91b9bf8ef454fd299938172e1593500f84c97887aca2fc5d7ffdd7
3
- size 29116
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa6a0d0f4587fd7db38e6b127310086832f6d4eaf62e9d3df988be257279651e
3
+ size 32602