arun100 commited on
Commit
ad39415
1 Parent(s): bdb69a1

End of training

Browse files
README.md CHANGED
@@ -1,21 +1,24 @@
1
  ---
 
 
2
  license: mit
3
  base_model: bangla-speech-processing/BanglaASR
4
  tags:
 
5
  - generated_from_trainer
6
  datasets:
7
- - common_voice_16_0
8
  metrics:
9
  - wer
10
  model-index:
11
- - name: arun100/whisper-small-derived-bn-1
12
  results:
13
  - task:
14
  name: Automatic Speech Recognition
15
  type: automatic-speech-recognition
16
  dataset:
17
- name: common_voice_16_0
18
- type: common_voice_16_0
19
  config: bn
20
  split: test
21
  args: bn
@@ -28,9 +31,9 @@ model-index:
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
29
  should probably proofread and complete it, then remove this comment. -->
30
 
31
- # bangla-speech-processing/BanglaASR
32
 
33
- This model is a fine-tuned version of [bangla-speech-processing/BanglaASR](https://huggingface.co/bangla-speech-processing/BanglaASR) on the common_voice_16_0 dataset.
34
  It achieves the following results on the evaluation set:
35
  - Loss: 0.1281
36
  - Wer: 3.7265
 
1
  ---
2
+ language:
3
+ - bn
4
  license: mit
5
  base_model: bangla-speech-processing/BanglaASR
6
  tags:
7
+ - whisper-event
8
  - generated_from_trainer
9
  datasets:
10
+ - mozilla-foundation/common_voice_16_0
11
  metrics:
12
  - wer
13
  model-index:
14
+ - name: Whisper Base Hindi
15
  results:
16
  - task:
17
  name: Automatic Speech Recognition
18
  type: automatic-speech-recognition
19
  dataset:
20
+ name: mozilla-foundation/common_voice_16_0 bn
21
+ type: mozilla-foundation/common_voice_16_0
22
  config: bn
23
  split: test
24
  args: bn
 
31
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
32
  should probably proofread and complete it, then remove this comment. -->
33
 
34
+ # Whisper Base Hindi
35
 
36
+ This model is a fine-tuned version of [bangla-speech-processing/BanglaASR](https://huggingface.co/bangla-speech-processing/BanglaASR) on the mozilla-foundation/common_voice_16_0 bn dataset.
37
  It achieves the following results on the evaluation set:
38
  - Loss: 0.1281
39
  - Wer: 3.7265
all_results.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "eval_loss": 0.12812651693820953,
4
+ "eval_runtime": 5982.3035,
5
+ "eval_samples_per_second": 1.557,
6
+ "eval_steps_per_second": 0.097,
7
+ "eval_wer": 3.7265218830814386,
8
+ "train_loss": 0.144483003616333,
9
+ "train_runtime": 13002.902,
10
+ "train_samples_per_second": 0.492,
11
+ "train_steps_per_second": 0.008
12
+ }
eval_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "eval_loss": 0.12812651693820953,
4
+ "eval_runtime": 5982.3035,
5
+ "eval_samples_per_second": 1.557,
6
+ "eval_steps_per_second": 0.097,
7
+ "eval_wer": 3.7265218830814386
8
+ }
runs/Jan21_22-04-28_ip-172-31-23-125/events.out.tfevents.1705893691.ip-172-31-23-125.29447.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4070c463c0dfc4c8706c616deeac87bd7526a09cda9cc074f8e56917cd78d173
3
+ size 400
train_results.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "train_loss": 0.144483003616333,
4
+ "train_runtime": 13002.902,
5
+ "train_samples_per_second": 0.492,
6
+ "train_steps_per_second": 0.008
7
+ }
trainer_state.json ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 3.7265218830814386,
3
+ "best_model_checkpoint": "./checkpoint-100",
4
+ "epoch": 1.0,
5
+ "eval_steps": 50,
6
+ "global_step": 100,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.25,
13
+ "learning_rate": 5e-08,
14
+ "loss": 0.1602,
15
+ "step": 25
16
+ },
17
+ {
18
+ "epoch": 0.5,
19
+ "learning_rate": 1e-07,
20
+ "loss": 0.1588,
21
+ "step": 50
22
+ },
23
+ {
24
+ "epoch": 0.5,
25
+ "eval_loss": 0.14945247769355774,
26
+ "eval_runtime": 6007.4304,
27
+ "eval_samples_per_second": 1.55,
28
+ "eval_steps_per_second": 0.097,
29
+ "eval_wer": 3.7649995973262462,
30
+ "step": 50
31
+ },
32
+ {
33
+ "epoch": 0.75,
34
+ "learning_rate": 5e-08,
35
+ "loss": 0.1321,
36
+ "step": 75
37
+ },
38
+ {
39
+ "epoch": 1.0,
40
+ "learning_rate": 0.0,
41
+ "loss": 0.1267,
42
+ "step": 100
43
+ },
44
+ {
45
+ "epoch": 1.0,
46
+ "eval_loss": 0.12812651693820953,
47
+ "eval_runtime": 5985.3223,
48
+ "eval_samples_per_second": 1.556,
49
+ "eval_steps_per_second": 0.097,
50
+ "eval_wer": 3.7265218830814386,
51
+ "step": 100
52
+ },
53
+ {
54
+ "epoch": 1.0,
55
+ "step": 100,
56
+ "total_flos": 1.846946562048e+18,
57
+ "train_loss": 0.144483003616333,
58
+ "train_runtime": 13002.902,
59
+ "train_samples_per_second": 0.492,
60
+ "train_steps_per_second": 0.008
61
+ }
62
+ ],
63
+ "logging_steps": 25,
64
+ "max_steps": 100,
65
+ "num_input_tokens_seen": 0,
66
+ "num_train_epochs": 9223372036854775807,
67
+ "save_steps": 50,
68
+ "total_flos": 1.846946562048e+18,
69
+ "train_batch_size": 16,
70
+ "trial_name": null,
71
+ "trial_params": null
72
+ }