Model save
Browse files- README.md +31 -18
- config.json +1 -1
- generation_config.json +1 -1
- model.safetensors +1 -1
- training_args.bin +1 -1
README.md
CHANGED
@@ -18,8 +18,8 @@ should probably proofread and complete it, then remove this comment. -->
|
|
18 |
|
19 |
This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the None dataset.
|
20 |
It achieves the following results on the evaluation set:
|
21 |
-
- Loss:
|
22 |
-
- Wer:
|
23 |
|
24 |
## Model description
|
25 |
|
@@ -39,32 +39,45 @@ More information needed
|
|
39 |
|
40 |
The following hyperparameters were used during training:
|
41 |
- learning_rate: 5e-05
|
42 |
-
- train_batch_size:
|
43 |
-
- eval_batch_size:
|
44 |
- seed: 42
|
45 |
- gradient_accumulation_steps: 2
|
46 |
-
- total_train_batch_size:
|
47 |
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
48 |
- lr_scheduler_type: linear
|
49 |
- lr_scheduler_warmup_steps: 5
|
50 |
-
- num_epochs:
|
51 |
- mixed_precision_training: Native AMP
|
52 |
|
53 |
### Training results
|
54 |
|
55 |
-
| Training Loss | Epoch | Step
|
56 |
-
|
57 |
-
|
|
58 |
-
|
|
59 |
-
|
|
60 |
-
|
|
61 |
-
|
|
62 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
63 |
|
64 |
|
65 |
### Framework versions
|
66 |
|
67 |
-
- Transformers 4.
|
68 |
-
- Pytorch 2.5.
|
69 |
-
- Datasets 3.0
|
70 |
-
- Tokenizers 0.
|
|
|
18 |
|
19 |
This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the None dataset.
|
20 |
It achieves the following results on the evaluation set:
|
21 |
+
- Loss: 0.2998
|
22 |
+
- Wer: 37.0027
|
23 |
|
24 |
## Model description
|
25 |
|
|
|
39 |
|
40 |
The following hyperparameters were used during training:
|
41 |
- learning_rate: 5e-05
|
42 |
+
- train_batch_size: 8
|
43 |
+
- eval_batch_size: 8
|
44 |
- seed: 42
|
45 |
- gradient_accumulation_steps: 2
|
46 |
+
- total_train_batch_size: 16
|
47 |
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
48 |
- lr_scheduler_type: linear
|
49 |
- lr_scheduler_warmup_steps: 5
|
50 |
+
- num_epochs: 10
|
51 |
- mixed_precision_training: Native AMP
|
52 |
|
53 |
### Training results
|
54 |
|
55 |
+
| Training Loss | Epoch | Step | Validation Loss | Wer |
|
56 |
+
|:-------------:|:------:|:-----:|:---------------:|:-------:|
|
57 |
+
| 0.3985 | 0.2458 | 5000 | 0.4006 | 100 |
|
58 |
+
| 0.3535 | 0.4917 | 10000 | 0.3515 | 41.9359 |
|
59 |
+
| 0.3329 | 0.7375 | 15000 | 0.3267 | 41.2601 |
|
60 |
+
| 0.3088 | 0.9834 | 20000 | 0.3122 | 40.8518 |
|
61 |
+
| 0.2702 | 1.2292 | 25000 | 0.3052 | 39.1540 |
|
62 |
+
| 0.2682 | 1.4750 | 30000 | 0.2963 | 39.5239 |
|
63 |
+
| 0.2606 | 1.7208 | 35000 | 0.2891 | 38.2537 |
|
64 |
+
| 0.2622 | 1.9667 | 40000 | 0.2837 | 38.0593 |
|
65 |
+
| 0.2164 | 2.2125 | 45000 | 0.2854 | 37.4512 |
|
66 |
+
| 0.222 | 2.4583 | 50000 | 0.2817 | 37.9314 |
|
67 |
+
| 0.2233 | 2.7042 | 55000 | 0.2788 | 37.6204 |
|
68 |
+
| 0.2218 | 2.9500 | 60000 | 0.2755 | 36.9669 |
|
69 |
+
| 0.1752 | 3.1958 | 65000 | 0.2871 | 36.6628 |
|
70 |
+
| 0.1802 | 3.4417 | 70000 | 0.2860 | 36.9139 |
|
71 |
+
| 0.1864 | 3.6875 | 75000 | 0.2818 | 37.8713 |
|
72 |
+
| 0.1832 | 3.9334 | 80000 | 0.2804 | 38.1352 |
|
73 |
+
| 0.1396 | 4.1792 | 85000 | 0.3042 | 37.6950 |
|
74 |
+
| 0.1398 | 4.4250 | 90000 | 0.3042 | 37.7528 |
|
75 |
+
| 0.1482 | 4.6708 | 95000 | 0.2998 | 37.0027 |
|
76 |
|
77 |
|
78 |
### Framework versions
|
79 |
|
80 |
+
- Transformers 4.48.0.dev0
|
81 |
+
- Pytorch 2.5.1+cu124
|
82 |
+
- Datasets 3.1.0
|
83 |
+
- Tokenizers 0.21.0
|
config.json
CHANGED
@@ -41,7 +41,7 @@
|
|
41 |
"pad_token_id": 50257,
|
42 |
"scale_embedding": false,
|
43 |
"torch_dtype": "float32",
|
44 |
-
"transformers_version": "4.
|
45 |
"use_cache": true,
|
46 |
"use_weighted_layer_sum": false,
|
47 |
"vocab_size": 51865
|
|
|
41 |
"pad_token_id": 50257,
|
42 |
"scale_embedding": false,
|
43 |
"torch_dtype": "float32",
|
44 |
+
"transformers_version": "4.48.0.dev0",
|
45 |
"use_cache": true,
|
46 |
"use_weighted_layer_sum": false,
|
47 |
"vocab_size": 51865
|
generation_config.json
CHANGED
@@ -172,5 +172,5 @@
|
|
172 |
"transcribe": 50359,
|
173 |
"translate": 50358
|
174 |
},
|
175 |
-
"transformers_version": "4.
|
176 |
}
|
|
|
172 |
"transcribe": 50359,
|
173 |
"translate": 50358
|
174 |
},
|
175 |
+
"transformers_version": "4.48.0.dev0"
|
176 |
}
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 966995080
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cd55b173c1cacf6b49c7bc229649a37007cd2ddcfbf5e871dc97d3f1e1d8887b
|
3 |
size 966995080
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5496
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f4774929ca81e78f05e3e89058efd8115f295b3804fa56b4a48688fc3459f6c3
|
3 |
size 5496
|