Edit model card

flan-t5-base-samsum

This model is a fine-tuned version of google/flan-t5-base on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 1.3733
  • Rouge1: 47.0915
  • Rouge2: 23.4118
  • Rougel: 39.8368
  • Rougelsum: 43.3312
  • Gen Len: 17.2308

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 32
  • eval_batch_size: 32
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 5

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
No log 1.0 461 1.3902 46.725 23.0515 39.1832 42.944 17.2601
1.4553 2.0 922 1.3787 47.2726 23.4894 39.5801 43.4158 17.1355
1.3763 3.0 1383 1.3737 47.3267 23.6022 39.7992 43.4532 17.1966
1.3412 4.0 1844 1.3750 47.2869 23.7053 39.9717 43.5825 17.3431
1.3045 5.0 2305 1.3733 47.0915 23.4118 39.8368 43.3312 17.2308

Framework versions

  • Transformers 4.39.1
  • Pytorch 2.0.1+cu118
  • Datasets 2.18.0
  • Tokenizers 0.15.2
Downloads last month
1
Safetensors
Model size
248M params
Tensor type
F32
·

Finetuned from