Edit model card

flan-t5-base-updated_data

This model is a fine-tuned version of google/flan-t5-base on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 2.1855
  • Rouge1: 30.1504
  • Rouge2: 15.0601
  • Rougel: 25.9757
  • Rougelsum: 26.4304
  • Gen Len: 18.18

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 5

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
No log 1.0 112 2.4245 30.8943 14.0099 26.1186 26.6014 18.0467
No log 2.0 224 2.3033 31.2618 14.9632 26.6116 27.0315 18.0933
No log 3.0 336 2.2369 29.9047 14.7151 25.5824 26.0168 17.9667
No log 4.0 448 2.1954 30.2469 15.1358 26.1566 26.5722 18.18
2.45 5.0 560 2.1855 30.1504 15.0601 25.9757 26.4304 18.18

Framework versions

  • Transformers 4.35.2
  • Pytorch 2.1.0+cu118
  • Datasets 2.15.0
  • Tokenizers 0.15.0
Downloads last month
8
Safetensors
Model size
248M params
Tensor type
F32
·

Finetuned from