t5-small-finetuned-hausa-to-english
This model is a fine-tuned version of google-t5/t5-small on the None dataset. It achieves the following results on the evaluation set:
- Loss: 1.5734
- Bleu: 71.4633
- Gen Len: 6.984
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0008
- train_batch_size: 32
- eval_batch_size: 32
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_steps: 3000
- num_epochs: 30
- mixed_precision_training: Native AMP
Training results
Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
---|---|---|---|---|---|
3.4817 | 1.0 | 749 | 2.2949 | 34.5266 | 7.5496 |
2.0957 | 2.0 | 1498 | 1.6167 | 53.0797 | 6.8785 |
1.562 | 3.0 | 2247 | 1.3650 | 60.3083 | 7.0095 |
1.2877 | 4.0 | 2996 | 1.2149 | 63.8437 | 6.9444 |
1.1026 | 5.0 | 3745 | 1.1452 | 66.7883 | 7.13 |
0.9531 | 6.0 | 4494 | 1.1028 | 67.3774 | 6.911 |
0.8402 | 7.0 | 5243 | 1.0995 | 68.0354 | 6.8114 |
0.7513 | 8.0 | 5992 | 1.0878 | 69.4876 | 7.0216 |
0.6746 | 9.0 | 6741 | 1.1109 | 69.7327 | 7.1134 |
0.6073 | 10.0 | 7490 | 1.1167 | 70.1607 | 7.0526 |
0.5531 | 11.0 | 8239 | 1.1468 | 69.8006 | 6.8101 |
0.4981 | 12.0 | 8988 | 1.1856 | 70.5423 | 6.8789 |
0.4544 | 13.0 | 9737 | 1.2019 | 70.5876 | 6.9313 |
0.4095 | 14.0 | 10486 | 1.2347 | 70.7996 | 6.8371 |
0.373 | 15.0 | 11235 | 1.2734 | 71.0903 | 7.0274 |
0.3408 | 16.0 | 11984 | 1.2974 | 71.104 | 7.0025 |
0.3096 | 17.0 | 12733 | 1.3313 | 70.7308 | 6.925 |
0.2856 | 18.0 | 13482 | 1.3820 | 70.9862 | 6.9656 |
0.2601 | 19.0 | 14231 | 1.4016 | 71.1836 | 7.0082 |
0.2404 | 20.0 | 14980 | 1.4483 | 71.0219 | 6.9268 |
0.2241 | 21.0 | 15729 | 1.4714 | 71.2721 | 6.9552 |
0.2065 | 22.0 | 16478 | 1.4814 | 71.3874 | 6.9968 |
0.1942 | 23.0 | 17227 | 1.5090 | 71.4722 | 6.9404 |
0.1831 | 24.0 | 17976 | 1.5265 | 71.4556 | 6.9771 |
0.173 | 25.0 | 18725 | 1.5379 | 71.4026 | 6.9998 |
0.1662 | 26.0 | 19474 | 1.5530 | 71.4843 | 6.9932 |
0.159 | 27.0 | 20223 | 1.5668 | 71.3663 | 6.9784 |
0.1568 | 28.0 | 20972 | 1.5742 | 71.3261 | 6.9734 |
0.1566 | 29.0 | 21721 | 1.5739 | 71.4435 | 6.9843 |
0.156 | 30.0 | 22470 | 1.5734 | 71.4633 | 6.984 |
Framework versions
- Transformers 4.44.2
- Pytorch 2.4.0+cu121
- Datasets 2.21.0
- Tokenizers 0.19.1
- Downloads last month
- 40
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for Kumshe/t5-small-finetuned-hausa-to-english
Base model
google-t5/t5-small