JimmyPin's picture
Training complete
430a455 verified
|
raw
history blame
3.48 kB
metadata
library_name: transformers
license: apache-2.0
base_model: google/mt5-base
tags:
  - summarization
  - generated_from_trainer
metrics:
  - rouge
model-index:
  - name: mt5-finetuned-summarize
    results: []

mt5-finetuned-summarize

This model is a fine-tuned version of google/mt5-base on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: nan
  • Rouge1: 0.0
  • Rouge2: 0.0
  • Rougel: 0.0
  • Rougelsum: 0.0

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 4
  • eval_batch_size: 4
  • seed: 42
  • optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • num_epochs: 20
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum
0.0 0.9969 317 nan 0.0 0.0 0.0 0.0
0.0 1.9937 634 nan 0.0 0.0 0.0 0.0
0.0 2.9906 951 nan 0.0 0.0 0.0 0.0
0.0 3.9874 1268 nan 0.0 0.0 0.0 0.0
0.0 4.9843 1585 nan 0.0 0.0 0.0 0.0
0.0 5.9811 1902 nan 0.0 0.0 0.0 0.0
0.0 6.9780 2219 nan 0.0 0.0 0.0 0.0
0.0 7.9748 2536 nan 0.0 0.0 0.0 0.0
0.0 8.9717 2853 nan 0.0 0.0 0.0 0.0
0.0 9.9686 3170 nan 0.0 0.0 0.0 0.0
0.0 10.9654 3487 nan 0.0 0.0 0.0 0.0
0.0 11.9623 3804 nan 0.0 0.0 0.0 0.0
0.0 12.9591 4121 nan 0.0 0.0 0.0 0.0
0.0 13.9560 4438 nan 0.0 0.0 0.0 0.0
0.0 14.9528 4755 nan 0.0 0.0 0.0 0.0
0.0 15.9497 5072 nan 0.0 0.0 0.0 0.0
0.0 16.9465 5389 nan 0.0 0.0 0.0 0.0
0.0 17.9434 5706 nan 0.0 0.0 0.0 0.0
0.0 18.9403 6023 nan 0.0 0.0 0.0 0.0
0.0 19.9371 6340 nan 0.0 0.0 0.0 0.0

Framework versions

  • Transformers 4.47.1
  • Pytorch 2.5.1+cu121
  • Datasets 3.2.0
  • Tokenizers 0.21.0