PhongLT's picture
End of training
b173df0
|
raw
history blame
2.54 kB
metadata
base_model: vinai/bartpho-syllable-base
tags:
  - generated_from_trainer
metrics:
  - rouge
model-index:
  - name: bartpho-syllable-base-ceg-pair-lexnorm-wiki20k
    results: []

bartpho-syllable-base-ceg-pair-lexnorm-wiki20k

This model is a fine-tuned version of vinai/bartpho-syllable-base on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 0.0441
  • Rouge1: 83.8906
  • Rouge2: 76.8215
  • Rougel: 83.4733
  • Rougelsum: 83.4783
  • Gen Len: 14.9751

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 32
  • eval_batch_size: 32
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
0.1329 1.0 625 0.0572 80.5125 71.817 80.0125 79.9934 14.9828
0.0424 2.0 1250 0.0447 81.8041 73.7589 81.3035 81.2949 14.8986
0.0253 3.0 1875 0.0449 82.4467 74.7259 81.9169 81.906 15.0067
0.0167 4.0 2500 0.0427 83.3339 76.0786 82.8455 82.8229 15.0010
0.0116 5.0 3125 0.0444 83.4771 76.234 83.0177 83.008 15.0144
0.0082 6.0 3750 0.0438 83.6459 76.467 83.2461 83.2372 14.9933
0.0061 7.0 4375 0.0440 83.699 76.5119 83.3283 83.3175 14.9684
0.0047 8.0 5000 0.0447 83.7618 76.6774 83.346 83.3564 14.9349
0.0038 9.0 5625 0.0436 83.9239 76.845 83.4885 83.5062 14.9732
0.0032 10.0 6250 0.0441 83.8906 76.8215 83.4733 83.4783 14.9751

Framework versions

  • Transformers 4.36.2
  • Pytorch 2.1.2
  • Datasets 2.15.0
  • Tokenizers 0.15.0