metadata
language: ms
t5-base-bahasa-cased
Pretrained T5 base language model for Malay.
Pretraining Corpus
t5-base-bahasa-cased
model was pretrained on multiple tasks. Below is list of tasks we trained on,
- Language masking task on bahasa news, bahasa Wikipedia, bahasa Academia.edu, bahasa parliament and translated The Pile.
- News title prediction on bahasa news.
- Next sentence prediction on bahasa news, bahasa Wikipedia, bahasa Academia.edu, bahasa parliament and translated The Pile.
- Translated QA Natural.
- Text Similarity task on translated SNLI and translated MNLI.
- EN-MS translation.
- MS-EN translation.
- Abstractive Summarization.
- Knowledge Graph triples generation.
- Paraphrase.
Preparing steps can reproduce at https://github.com/huseinzol05/malaya/tree/master/pretrained-model/t5/prepare
Pretraining details
- This model was trained using Google T5 repository https://github.com/google-research/text-to-text-transfer-transformer, on v3-8 TPU.
- All steps can reproduce from here, https://github.com/huseinzol05/Malaya/tree/master/pretrained-model/t5