File size: 12,275 Bytes
40eb5f7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2421746
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40eb5f7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a28d0de
40eb5f7
 
 
 
 
 
 
f784d9e
 
 
 
 
 
 
 
 
 
 
 
 
9a1f2f2
a28d0de
768de51
b965f4b
ef6b677
2313ee4
2eca096
9d7b8eb
174aaf5
2421746
 
 
40eb5f7
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
---
language:
- 'no'
license: apache-2.0
tags:
- audio
- asr
- automatic-speech-recognition
- hf-asr-leaderboard
model-index:
- name: scream_duodevicesimus_working_noaudiobooks_7e5_v2
  results: []
---

<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->

# scream_duodevicesimus_working_noaudiobooks_7e5_v2

This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the NbAiLab/ncc_speech dataset.
It achieves the following results on the evaluation set:
- step: 19999
- validation_fleurs_loss: 0.3089
- train_loss: 0.7173
- validation_fleurs_wer: 8.2391
- validation_fleurs_cer: 3.7039
- validation_fleurs_exact_wer: 12.6643
- validation_fleurs_exact_cer: 4.6485
- validation_stortinget_loss: 0.2845
- validation_stortinget_wer: 13.9625
- validation_stortinget_cer: 10.0306
- validation_stortinget_exact_wer: 17.2389
- validation_stortinget_exact_cer: 10.5844
- validation_nrk_tv_loss: 0.7447
- validation_nrk_tv_wer: 40.1880
- validation_nrk_tv_cer: 31.3161
- validation_nrk_tv_exact_wer: 47.6494
- validation_nrk_tv_exact_cer: 32.4497

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- learning_rate: 7e-05
- lr_scheduler_type: linear
- per_device_train_batch_size: 32
- total_train_batch_size_per_node: 128
- total_train_batch_size: 1024
- total_optimization_steps: 20,000
- starting_optimization_step: None
- finishing_optimization_step: 20,000
- num_train_dataset_workers: 32
- num_hosts: 8
- total_num_training_examples: 20,480,000
- steps_per_epoch: 11275
- num_beams: 5
- dropout: True
- bpe_dropout_probability: 0.1
- activation_dropout_probability: 0.1

### Training results

| step  | validation_fleurs_loss | train_loss | validation_fleurs_wer | validation_fleurs_cer | validation_fleurs_exact_wer | validation_fleurs_exact_cer | validation_stortinget_loss | validation_stortinget_wer | validation_stortinget_cer | validation_stortinget_exact_wer | validation_stortinget_exact_cer | validation_nrk_tv_loss | validation_nrk_tv_wer | validation_nrk_tv_cer | validation_nrk_tv_exact_wer | validation_nrk_tv_exact_cer |
|:-----:|:----------------------:|:----------:|:---------------------:|:---------------------:|:---------------------------:|:---------------------------:|:--------------------------:|:-------------------------:|:-------------------------:|:-------------------------------:|:-------------------------------:|:----------------------:|:---------------------:|:---------------------:|:---------------------------:|:---------------------------:|
| 0     | 1.3211                 | 3.0189     | 110.1725              | 80.3659               | 196.8041                    | 131.4230                    | 1.5012                     | 76.6096                   | 51.2561                   | 82.1890                         | 54.4126                         | 1.8187                 | 259.8656              | 217.2117              | 269.5665                    | 222.7746                    |
| 1000  | 0.6977                 | 1.1353     | 13.4444               | 4.3105                | 17.5926                     | 5.2863                      | 0.4717                     | 21.7105                   | 13.9604                   | 25.3783                         | 14.6687                         | 0.9934                 | 86.4845               | 70.4142               | 93.7677                     | 73.6462                     |
| 2000  | 0.3926                 | 0.8912     | 10.5889               | 3.7088                | 14.7849                     | 4.6968                      | 0.3930                     | 18.7212                   | 12.5960                   | 22.2213                         | 13.2354                         | 0.8926                 | 49.9691               | 39.8385               | 57.6635                     | 41.2514                     |
| 3000  | 0.3620                 | 0.8106     | 10.7674               | 4.3007                | 15.0836                     | 5.2573                      | 0.3632                     | 17.5019                   | 11.9674                   | 21.0430                         | 12.5977                         | 0.8606                 | 44.9157               | 34.5510               | 52.6419                     | 35.8510                     |
| 4000  | 0.3363                 | 0.8043     | 10.3807               | 3.8518                | 14.0980                     | 4.7886                      | 0.3443                     | 16.1694                   | 11.2786                   | 19.6917                         | 11.8983                         | 0.8431                 | 44.9487               | 34.0425               | 52.5379                     | 35.4061                     |
| 5000  | 0.3060                 | 0.7682     | 9.6074                | 3.6694                | 13.8590                     | 4.5808                      | 0.3329                     | 16.0903                   | 11.1667                   | 19.5724                         | 11.7732                         | 0.8154                 | 45.4598               | 35.0224               | 52.7292                     | 36.3997                     |
| 6000  | 0.3477                 | 0.7510     | 9.2207                | 3.5510                | 13.3214                     | 4.5083                      | 0.3246                     | 15.9711                   | 11.2829                   | 19.4232                         | 11.8775                         | 0.8097                 | 43.0897               | 33.1321               | 50.5325                     | 34.3331                     |
| 7000  | 0.3152                 | 0.7608     | 9.6074                | 4.1034                | 13.7395                     | 5.0834                      | 0.3217                     | 15.1188                   | 10.6651                   | 18.5510                         | 11.2540                         | 0.7959                 | 42.0139               | 32.2852               | 49.4716                     | 33.4915                     |
| 8000  | 0.3232                 | 0.7680     | 9.8453                | 3.9258                | 13.7993                     | 4.8128                      | 0.3161                     | 15.1877                   | 10.7202                   | 18.5356                         | 11.2959                         | 0.7938                 | 42.1829               | 32.6832               | 49.6256                     | 34.2256                     |
| 9000  | 0.3376                 | 0.7386     | 8.9827                | 3.4967                | 12.8734                     | 4.4262                      | 0.3082                     | 14.8650                   | 10.5644                   | 18.2372                         | 11.1377                         | 0.7793                 | 41.7501               | 32.6252               | 49.4924                     | 33.8756                     |
| 10000 | 0.3152                 | 0.7408     | 9.0422                | 3.7335                | 13.5006                     | 4.6678                      | 0.3068                     | 14.7458                   | 10.4670                   | 18.1324                         | 11.0457                         | 0.7773                 | 41.3750               | 31.9683               | 49.1887                     | 33.1957                     |
| 11000 | 0.3167                 | 0.7246     | 9.2802                | 3.7976                | 13.2318                     | 4.7354                      | 0.3010                     | 14.4811                   | 10.3391                   | 17.8357                         | 10.9036                         | 0.7745                 | 41.2926               | 31.8771               | 48.8018                     | 33.0807                     |
| 12000 | 0.3087                 | 0.7240     | 8.7745                | 3.6447                | 12.7838                     | 4.5712                      | 0.2996                     | 14.2912                   | 10.2057                   | 17.6353                         | 10.7722                         | 0.7683                 | 41.1525               | 32.0549               | 48.8018                     | 33.2402                     |
| 13000 | 0.3330                 | 0.7061     | 8.3284                | 3.5362                | 12.4851                     | 4.4552                      | 0.2981                     | 14.3397                   | 10.2971                   | 17.7257                         | 10.8680                         | 0.7642                 | 41.1401               | 32.0379               | 48.7685                     | 33.1949                     |
| 14000 | 0.3176                 | 0.6887     | 8.8935                | 3.7680                | 12.8734                     | 4.6726                      | 0.2940                     | 14.1728                   | 10.2323                   | 17.5270                         | 10.8029                         | 0.7618                 | 39.9942               | 30.7597               | 47.3873                     | 31.9470                     |
| 15000 | 0.3113                 | 0.7296     | 8.7151                | 3.8617                | 13.0526                     | 4.8224                      | 0.2924                     | 14.0962                   | 10.1638                   | 17.4495                         | 10.7350                         | 0.7595                 | 40.2951               | 31.0757               | 47.7201                     | 32.2502                     |
| 16000 | 0.3058                 | 0.6820     | 9.2802                | 4.0688                | 13.2019                     | 4.9481                      | 0.2922                     | 14.0766                   | 10.1057                   | 17.4222                         | 10.6784                         | 0.7544                 | 40.4641               | 31.5116               | 47.9947                     | 32.7092                     |
| 17000 | 0.3030                 | 0.7120     | 8.9233                | 3.8913                | 13.0824                     | 4.8321                      | 0.2878                     | 14.1413                   | 10.1954                   | 17.4785                         | 10.7684                         | 0.7559                 | 40.3487               | 31.4791               | 48.0113                     | 32.6455                     |
| 18000 | 0.3080                 | 0.6951     | 8.5961                | 3.7138                | 12.5747                     | 4.6533                      | 0.2863                     | 13.8595                   | 9.9432                    | 17.1562                         | 10.5007                         | 0.7467                 | 40.0437               | 31.2512               | 47.5745                     | 32.4163                     |
| 19000 | 0.3104                 | 0.6771     | 8.5961                | 3.6743                | 12.7838                     | 4.6050                      | 0.2854                     | 13.9702                   | 10.0538                   | 17.2858                         | 10.6153                         | 0.7477                 | 40.2003               | 31.3663               | 47.6743                     | 32.5098                     |
| 19999 | 0.3089                 | 0.7173     | 8.2391                | 3.7039                | 12.6643                     | 4.6485                      |
| 19999 | 0.2845                 | 0.7173     | 13.9625               | 10.0306               | 17.2389                     | 10.5844                     |
| 19999 | 0.7447                 | 0.7173     | 40.1880               | 31.3161               | 47.6494                     | 32.4497                     |


### Framework versions

- Transformers 4.31.0.dev0
- Datasets 2.13.0
- Tokenizers 0.13.3