ganga4364 commited on
Commit
c57b21d
·
verified ·
1 Parent(s): c4a31db

Model save

Browse files
Files changed (5) hide show
  1. README.md +31 -18
  2. config.json +1 -1
  3. generation_config.json +1 -1
  4. model.safetensors +1 -1
  5. training_args.bin +1 -1
README.md CHANGED
@@ -18,8 +18,8 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 2.3193
22
- - Wer: 100
23
 
24
  ## Model description
25
 
@@ -39,32 +39,45 @@ More information needed
39
 
40
  The following hyperparameters were used during training:
41
  - learning_rate: 5e-05
42
- - train_batch_size: 32
43
- - eval_batch_size: 32
44
  - seed: 42
45
  - gradient_accumulation_steps: 2
46
- - total_train_batch_size: 64
47
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: linear
49
  - lr_scheduler_warmup_steps: 5
50
- - num_epochs: 2
51
  - mixed_precision_training: Native AMP
52
 
53
  ### Training results
54
 
55
- | Training Loss | Epoch | Step | Validation Loss | Wer |
56
- |:-------------:|:------:|:----:|:---------------:|:---:|
57
- | 6.0361 | 0.3125 | 5 | 5.3765 | 100 |
58
- | 4.4117 | 0.625 | 10 | 3.4696 | 100 |
59
- | 3.0052 | 0.9375 | 15 | 2.8653 | 100 |
60
- | 2.4717 | 1.25 | 20 | 2.5730 | 100 |
61
- | 2.2188 | 1.5625 | 25 | 2.3918 | 100 |
62
- | 2.085 | 1.875 | 30 | 2.3193 | 100 |
 
 
 
 
 
 
 
 
 
 
 
 
 
63
 
64
 
65
  ### Framework versions
66
 
67
- - Transformers 4.47.0.dev0
68
- - Pytorch 2.5.0+cu124
69
- - Datasets 3.0.1
70
- - Tokenizers 0.20.1
 
18
 
19
  This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.2998
22
+ - Wer: 37.0027
23
 
24
  ## Model description
25
 
 
39
 
40
  The following hyperparameters were used during training:
41
  - learning_rate: 5e-05
42
+ - train_batch_size: 8
43
+ - eval_batch_size: 8
44
  - seed: 42
45
  - gradient_accumulation_steps: 2
46
+ - total_train_batch_size: 16
47
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: linear
49
  - lr_scheduler_warmup_steps: 5
50
+ - num_epochs: 10
51
  - mixed_precision_training: Native AMP
52
 
53
  ### Training results
54
 
55
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
56
+ |:-------------:|:------:|:-----:|:---------------:|:-------:|
57
+ | 0.3985 | 0.2458 | 5000 | 0.4006 | 100 |
58
+ | 0.3535 | 0.4917 | 10000 | 0.3515 | 41.9359 |
59
+ | 0.3329 | 0.7375 | 15000 | 0.3267 | 41.2601 |
60
+ | 0.3088 | 0.9834 | 20000 | 0.3122 | 40.8518 |
61
+ | 0.2702 | 1.2292 | 25000 | 0.3052 | 39.1540 |
62
+ | 0.2682 | 1.4750 | 30000 | 0.2963 | 39.5239 |
63
+ | 0.2606 | 1.7208 | 35000 | 0.2891 | 38.2537 |
64
+ | 0.2622 | 1.9667 | 40000 | 0.2837 | 38.0593 |
65
+ | 0.2164 | 2.2125 | 45000 | 0.2854 | 37.4512 |
66
+ | 0.222 | 2.4583 | 50000 | 0.2817 | 37.9314 |
67
+ | 0.2233 | 2.7042 | 55000 | 0.2788 | 37.6204 |
68
+ | 0.2218 | 2.9500 | 60000 | 0.2755 | 36.9669 |
69
+ | 0.1752 | 3.1958 | 65000 | 0.2871 | 36.6628 |
70
+ | 0.1802 | 3.4417 | 70000 | 0.2860 | 36.9139 |
71
+ | 0.1864 | 3.6875 | 75000 | 0.2818 | 37.8713 |
72
+ | 0.1832 | 3.9334 | 80000 | 0.2804 | 38.1352 |
73
+ | 0.1396 | 4.1792 | 85000 | 0.3042 | 37.6950 |
74
+ | 0.1398 | 4.4250 | 90000 | 0.3042 | 37.7528 |
75
+ | 0.1482 | 4.6708 | 95000 | 0.2998 | 37.0027 |
76
 
77
 
78
  ### Framework versions
79
 
80
+ - Transformers 4.48.0.dev0
81
+ - Pytorch 2.5.1+cu124
82
+ - Datasets 3.1.0
83
+ - Tokenizers 0.21.0
config.json CHANGED
@@ -41,7 +41,7 @@
41
  "pad_token_id": 50257,
42
  "scale_embedding": false,
43
  "torch_dtype": "float32",
44
- "transformers_version": "4.47.0.dev0",
45
  "use_cache": true,
46
  "use_weighted_layer_sum": false,
47
  "vocab_size": 51865
 
41
  "pad_token_id": 50257,
42
  "scale_embedding": false,
43
  "torch_dtype": "float32",
44
+ "transformers_version": "4.48.0.dev0",
45
  "use_cache": true,
46
  "use_weighted_layer_sum": false,
47
  "vocab_size": 51865
generation_config.json CHANGED
@@ -172,5 +172,5 @@
172
  "transcribe": 50359,
173
  "translate": 50358
174
  },
175
- "transformers_version": "4.47.0.dev0"
176
  }
 
172
  "transcribe": 50359,
173
  "translate": 50358
174
  },
175
+ "transformers_version": "4.48.0.dev0"
176
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6fe1d6005746b9cdf30c1267d8d7dcd69c8973b5895d0f8af2ad74bc70825604
3
  size 966995080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd55b173c1cacf6b49c7bc229649a37007cd2ddcfbf5e871dc97d3f1e1d8887b
3
  size 966995080
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e1c6b0c1bb6ed1b21ce222912e7884d208760c979dc4e9cd6516dd38bb9b8922
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4774929ca81e78f05e3e89058efd8115f295b3804fa56b4a48688fc3459f6c3
3
  size 5496