Edit model card

t5-small-indosum-3

This model is a fine-tuned version of t5-small on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4340
  • Rouge1: 15.1875
  • Rouge2: 11.795
  • Rougel: 14.9384
  • Rougelsum: 15.0579
  • Gen Len: 19.0

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 40
  • eval_batch_size: 40
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 3
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
0.5356 1.0 1784 0.4647 15.1653 11.7743 14.9193 15.0383 19.0
0.4791 2.0 3568 0.4401 15.175 11.789 14.9281 15.0459 19.0
0.4698 3.0 5352 0.4340 15.1875 11.795 14.9384 15.0579 19.0

Framework versions

  • Transformers 4.28.0
  • Pytorch 2.0.1+cu118
  • Datasets 2.14.3
  • Tokenizers 0.13.3
Downloads last month
3