Edit model card

test

This model is a fine-tuned version of google/flan-t5-base on the samsum dataset. It achieves the following results on the evaluation set:

  • Loss: 1.4073
  • Rouge1: 47.697
  • Rouge2: 22.4202
  • Rougel: 39.7706
  • Rougelsum: 43.9975
  • Gen Len: 17.4

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • gradient_accumulation_steps: 8
  • total_train_batch_size: 64
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 5

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
No log 0.96 15 1.4247 47.3573 22.6495 39.4919 43.7222 17.29
No log 1.98 31 1.4103 47.6181 22.9152 39.9759 43.9923 17.52
No log 2.94 46 1.4079 46.9127 22.0069 39.1927 43.296 17.44
No log 3.97 62 1.4074 47.7178 22.552 39.8294 44.0775 17.42
No log 4.8 75 1.4073 47.697 22.4202 39.7706 43.9975 17.4

Framework versions

  • Transformers 4.28.0
  • Pytorch 2.0.1+cu118
  • Datasets 2.12.0
  • Tokenizers 0.13.3
Downloads last month
1

Dataset used to train srikant-personal/test

Evaluation results