Edit model card

indosum-pt-pl50-0

This model is a fine-tuned version of LazarusNLP/IndoNanoT5-base on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 1.1937
  • Rouge1: 67.2533
  • Rouge2: 57.3905
  • Rougel: 64.0732
  • Rougelsum: 66.2476
  • Gen Len: 97.596

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.001
  • train_batch_size: 16
  • eval_batch_size: 32
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 5.0

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
2.9265 1.0 892 1.8737 55.0839 40.3522 51.1983 53.7369 83.836
2.2409 2.0 1784 1.5733 61.4245 48.8132 57.8573 60.2997 97.0253
1.9661 3.0 2676 1.3583 63.476 51.6887 59.9726 62.3509 98.7573
1.7713 4.0 3568 1.2569 65.7891 54.9944 62.429 64.7377 98.7987
1.6308 5.0 4460 1.1937 66.4804 56.0803 63.1939 65.4418 100.6973

Framework versions

  • Transformers 4.40.2
  • Pytorch 2.3.1+cu121
  • Datasets 2.20.0
  • Tokenizers 0.19.1
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference API
Unable to determine this model's library. Check the docs .

Model tree for apwic/indosum-pt-pl50-0

Finetuned
(53)
this model