metadata
library_name: transformers
language:
- sn
license: cc-by-nc-4.0
base_model: facebook/mms-300m
tags:
- generated_from_trainer
datasets:
- DigitalUmuganda_Afrivoice/Shona
metrics:
- wer
model-index:
- name: facebook/mms-300m
results:
- task:
name: Automatic Speech Recognition
type: automatic-speech-recognition
dataset:
name: DigitalUmuganda
type: DigitalUmuganda_Afrivoice/Shona
metrics:
- name: Wer
type: wer
value: 0.3497254482725108
facebook/mms-300m
This model is a fine-tuned version of facebook/mms-300m on the DigitalUmuganda dataset. It achieves the following results on the evaluation set:
- Loss: 0.6190
- Wer: 0.3497
- Cer: 0.0791
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0003
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 32
- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.03
- num_epochs: 150
- mixed_precision_training: Native AMP
Training results
Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
---|---|---|---|---|---|
22.4497 | 1.0 | 273 | 2.9247 | 1.0 | 1.0 |
8.283 | 2.0 | 546 | 0.5185 | 0.5978 | 0.1175 |
1.5316 | 3.0 | 819 | 0.2757 | 0.3547 | 0.0632 |
1.0241 | 4.0 | 1092 | 0.2358 | 0.3222 | 0.0570 |
0.9072 | 5.0 | 1365 | 0.2283 | 0.3081 | 0.0560 |
0.8161 | 6.0 | 1638 | 0.2201 | 0.2945 | 0.0527 |
0.7469 | 7.0 | 1911 | 0.2122 | 0.2818 | 0.0487 |
0.6981 | 8.0 | 2184 | 0.2191 | 0.2946 | 0.0519 |
0.6454 | 9.0 | 2457 | 0.2122 | 0.2892 | 0.0504 |
0.6094 | 10.0 | 2730 | 0.2206 | 0.2952 | 0.0498 |
0.5649 | 11.0 | 3003 | 0.2164 | 0.2732 | 0.0482 |
0.5197 | 12.0 | 3276 | 0.2143 | 0.2728 | 0.0479 |
0.4793 | 13.0 | 3549 | 0.2344 | 0.2766 | 0.0485 |
0.4368 | 14.0 | 3822 | 0.2266 | 0.2751 | 0.0486 |
0.3977 | 15.0 | 4095 | 0.2358 | 0.2669 | 0.0458 |
0.3667 | 16.0 | 4368 | 0.2447 | 0.2793 | 0.0474 |
0.3335 | 17.0 | 4641 | 0.2547 | 0.2692 | 0.0461 |
0.3044 | 18.0 | 4914 | 0.2566 | 0.2684 | 0.0463 |
0.2801 | 19.0 | 5187 | 0.2698 | 0.2737 | 0.0473 |
0.2514 | 20.0 | 5460 | 0.2967 | 0.2790 | 0.0480 |
0.2314 | 21.0 | 5733 | 0.2855 | 0.2708 | 0.0462 |
0.2092 | 22.0 | 6006 | 0.3074 | 0.2785 | 0.0476 |
0.1991 | 23.0 | 6279 | 0.2953 | 0.2681 | 0.0449 |
0.183 | 24.0 | 6552 | 0.3118 | 0.2712 | 0.0457 |
0.1706 | 25.0 | 6825 | 0.3108 | 0.2654 | 0.0454 |
0.1573 | 26.0 | 7098 | 0.3310 | 0.2730 | 0.0463 |
0.1538 | 27.0 | 7371 | 0.3142 | 0.2767 | 0.0463 |
0.1489 | 28.0 | 7644 | 0.3337 | 0.2666 | 0.0458 |
0.1404 | 29.0 | 7917 | 0.3176 | 0.2728 | 0.0462 |
0.1309 | 30.0 | 8190 | 0.3431 | 0.2707 | 0.0461 |
0.1246 | 31.0 | 8463 | 0.3368 | 0.2725 | 0.0458 |
0.1216 | 32.0 | 8736 | 0.3490 | 0.2648 | 0.0448 |
0.1205 | 33.0 | 9009 | 0.3436 | 0.2627 | 0.0447 |
0.1145 | 34.0 | 9282 | 0.3384 | 0.2693 | 0.0454 |
0.1117 | 35.0 | 9555 | 0.3500 | 0.2610 | 0.0443 |
0.1058 | 36.0 | 9828 | 0.3478 | 0.2608 | 0.0441 |
0.1039 | 37.0 | 10101 | 0.3434 | 0.2590 | 0.0438 |
0.1 | 38.0 | 10374 | 0.3507 | 0.2609 | 0.0438 |
0.0987 | 39.0 | 10647 | 0.3553 | 0.2610 | 0.0443 |
0.0977 | 40.0 | 10920 | 0.3691 | 0.2683 | 0.0451 |
0.0972 | 41.0 | 11193 | 0.3508 | 0.2657 | 0.0450 |
0.0946 | 42.0 | 11466 | 0.3445 | 0.2605 | 0.0445 |
0.0934 | 43.0 | 11739 | 0.3541 | 0.2628 | 0.0445 |
0.0896 | 44.0 | 12012 | 0.3424 | 0.2600 | 0.0437 |
0.0844 | 45.0 | 12285 | 0.3504 | 0.2599 | 0.0439 |
0.0847 | 46.0 | 12558 | 0.3630 | 0.2588 | 0.0433 |
0.0809 | 47.0 | 12831 | 0.3650 | 0.2565 | 0.0435 |
0.0812 | 48.0 | 13104 | 0.3632 | 0.2665 | 0.0446 |
0.0783 | 49.0 | 13377 | 0.3585 | 0.2629 | 0.0435 |
0.0739 | 50.0 | 13650 | 0.3722 | 0.2650 | 0.0435 |
0.0772 | 51.0 | 13923 | 0.3627 | 0.2607 | 0.0433 |
0.0746 | 52.0 | 14196 | 0.3712 | 0.2600 | 0.0439 |
0.0713 | 53.0 | 14469 | 0.3540 | 0.2664 | 0.0445 |
0.073 | 54.0 | 14742 | 0.3764 | 0.2571 | 0.0433 |
0.0712 | 55.0 | 15015 | 0.3625 | 0.2533 | 0.0428 |
0.0679 | 56.0 | 15288 | 0.3818 | 0.2603 | 0.0435 |
0.0701 | 57.0 | 15561 | 0.3921 | 0.2565 | 0.0434 |
0.0672 | 58.0 | 15834 | 0.3807 | 0.2584 | 0.0434 |
0.0681 | 59.0 | 16107 | 0.3787 | 0.2575 | 0.0438 |
0.063 | 60.0 | 16380 | 0.3818 | 0.2561 | 0.0428 |
0.062 | 61.0 | 16653 | 0.3919 | 0.2567 | 0.0429 |
0.0613 | 62.0 | 16926 | 0.4038 | 0.2598 | 0.0431 |
0.0617 | 63.0 | 17199 | 0.3885 | 0.2566 | 0.0431 |
0.0618 | 64.0 | 17472 | 0.3949 | 0.2557 | 0.0430 |
0.0572 | 65.0 | 17745 | 0.3839 | 0.2529 | 0.0418 |
Framework versions
- Transformers 4.47.0.dev0
- Pytorch 2.1.0+cu118
- Datasets 3.0.2
- Tokenizers 0.20.1