octava commited on
Commit
269f2ac
·
verified ·
1 Parent(s): ec9857c

Training in progress, step 1000

Browse files
README.md CHANGED
@@ -8,22 +8,9 @@ tags:
8
  - generated_from_trainer
9
  datasets:
10
  - octava/indonesian-voice-transcription-1.4.9a-cv-fl-slrjv-md
11
- metrics:
12
- - wer
13
  model-index:
14
  - name: Optimized Whisper Small Id for Inspirasi
15
- results:
16
- - task:
17
- name: Automatic Speech Recognition
18
- type: automatic-speech-recognition
19
- dataset:
20
- name: Extracted Youtube with CommonVoice11, Fleurs, OpenSLR, and MagicData
21
- type: octava/indonesian-voice-transcription-1.4.9a-cv-fl-slrjv-md
22
- args: 'config: id, split: train'
23
- metrics:
24
- - name: Wer
25
- type: wer
26
- value: 19.96201329534663
27
  ---
28
 
29
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -33,8 +20,13 @@ should probably proofread and complete it, then remove this comment. -->
33
 
34
  This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the Extracted Youtube with CommonVoice11, Fleurs, OpenSLR, and MagicData dataset.
35
  It achieves the following results on the evaluation set:
36
- - Loss: 0.3376
37
- - Wer: 19.9620
 
 
 
 
 
38
 
39
  ## Model description
40
 
@@ -60,25 +52,9 @@ The following hyperparameters were used during training:
60
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
61
  - lr_scheduler_type: linear
62
  - lr_scheduler_warmup_steps: 500
63
- - training_steps: 5000
64
  - mixed_precision_training: Native AMP
65
 
66
- ### Training results
67
-
68
- | Training Loss | Epoch | Step | Validation Loss | Wer |
69
- |:-------------:|:------:|:----:|:---------------:|:-------:|
70
- | 0.4122 | 0.1686 | 500 | 0.3999 | 24.8908 |
71
- | 0.2737 | 0.3373 | 1000 | 0.3655 | 22.4691 |
72
- | 0.2311 | 0.5059 | 1500 | 0.3491 | 21.5195 |
73
- | 0.1947 | 0.6745 | 2000 | 0.3339 | 21.5100 |
74
- | 0.169 | 0.8432 | 2500 | 0.3408 | 20.6363 |
75
- | 0.0875 | 1.0118 | 3000 | 0.3429 | 21.2726 |
76
- | 0.0877 | 1.1804 | 3500 | 0.3430 | 20.4748 |
77
- | 0.0726 | 1.3491 | 4000 | 0.3396 | 20.2469 |
78
- | 0.0741 | 1.5177 | 4500 | 0.3378 | 20.2754 |
79
- | 0.0675 | 1.6863 | 5000 | 0.3376 | 19.9620 |
80
-
81
-
82
  ### Framework versions
83
 
84
  - Transformers 4.49.0
 
8
  - generated_from_trainer
9
  datasets:
10
  - octava/indonesian-voice-transcription-1.4.9a-cv-fl-slrjv-md
 
 
11
  model-index:
12
  - name: Optimized Whisper Small Id for Inspirasi
13
+ results: []
 
 
 
 
 
 
 
 
 
 
 
14
  ---
15
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
20
 
21
  This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the Extracted Youtube with CommonVoice11, Fleurs, OpenSLR, and MagicData dataset.
22
  It achieves the following results on the evaluation set:
23
+ - eval_loss: 0.3722
24
+ - eval_wer: 20.3324
25
+ - eval_runtime: 495.0068
26
+ - eval_samples_per_second: 3.527
27
+ - eval_steps_per_second: 0.222
28
+ - epoch: 3.0354
29
+ - step: 9000
30
 
31
  ## Model description
32
 
 
52
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
53
  - lr_scheduler_type: linear
54
  - lr_scheduler_warmup_steps: 500
55
+ - training_steps: 10000
56
  - mixed_precision_training: Native AMP
57
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
58
  ### Framework versions
59
 
60
  - Transformers 4.49.0
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bd3795ab541d4aea8220fe95dd5fdbb682a76136800fb4d1d76ee6f83c62a598
3
  size 966995080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:971a159cfc8a3081e4307ba7c8bd4e9aff289b0459e8cd56a72a6b607f63aa99
3
  size 966995080
runs/Dec06_07-03-01_060d7252a7cc/events.out.tfevents.1733468582.060d7252a7cc.146.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1f4ac057c818ea4be17ff8d2642489ad9ac5d35fe61ce75930e63e43af4d2c7
3
+ size 15532
runs/Nov29_07-23-40_21eadb60b9df/events.out.tfevents.1732865022.21eadb60b9df.2526.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1302f82b838cd2f1577c379d219d318a655e53f8be2420f27e4029b357055116
3
- size 76838
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62c38870e8fc95aca3d0dc6cb308df8de61927328ebbe69758e94f3c2ca64244
3
+ size 91293
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a6400f370e9e5da0218d021c4e96f42ab67889b6bdfdfc2526b3f73656f1febc
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b746e91e564ecec27b8a6425b7314deb7e82f3b3171c86b778141455a4233453
3
  size 5496