Edit model card

flan-t5-large-finetuned-samsum-2

This model is a fine-tuned version of google/flan-t5-large on the samsum dataset. It achieves the following results on the evaluation set:

  • Loss: 1.2143
  • Rouge1: 49.3339
  • Rouge2: 25.6711
  • Rougel: 41.9672
  • Rougelsum: 45.5636
  • Gen Len: 17.3504

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-06
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 5

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
1.1906 1.0 1842 1.2211 49.2035 25.36 41.7117 45.3746 17.3761
1.1441 2.0 3684 1.2168 49.3485 25.488 41.8996 45.4621 17.2882
1.1327 3.0 5526 1.2150 49.6279 25.8807 42.1862 45.7906 17.3858
1.1181 4.0 7368 1.2143 49.3339 25.6711 41.9672 45.5636 17.3504
1.0973 5.0 9210 1.2149 49.3509 25.7362 41.9252 45.5451 17.3480

Framework versions

  • Transformers 4.25.1
  • Pytorch 1.13.0+cu116
  • Datasets 2.8.0
  • Tokenizers 0.13.2
Downloads last month
4
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train mrm8488/flan-t5-large-finetuned-samsum-2

Evaluation results