3_loa / README.md
eschorn's picture
update model card README.md
3c9c8c4
|
raw
history blame
No virus
3.45 kB
metadata
license: apache-2.0
base_model: google/flan-t5-small
tags:
  - generated_from_trainer
datasets:
  - billsum
metrics:
  - rouge
model-index:
  - name: 3_loa
    results: []

3_loa

This model is a fine-tuned version of google/flan-t5-small on the billsum dataset. It achieves the following results on the evaluation set:

  • Loss: 2.1919
  • Rouge1: 0.1973
  • Rouge2: 0.1007
  • Rougel: 0.1708
  • Rougelsum: 0.1711
  • Gen Len: 19.0

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 20

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
No log 1.0 124 2.2622 0.1909 0.0921 0.1656 0.1659 19.0
No log 2.0 248 2.2534 0.1931 0.0956 0.1679 0.1681 19.0
No log 3.0 372 2.2433 0.1952 0.0967 0.1697 0.1699 19.0
No log 4.0 496 2.2358 0.1953 0.0978 0.1701 0.1702 19.0
2.4755 5.0 620 2.2323 0.1951 0.0981 0.1705 0.1706 19.0
2.4755 6.0 744 2.2253 0.1962 0.0996 0.1712 0.1714 19.0
2.4755 7.0 868 2.2199 0.1968 0.1003 0.1719 0.172 19.0
2.4755 8.0 992 2.2170 0.1963 0.0999 0.1717 0.1717 19.0
2.4416 9.0 1116 2.2134 0.1971 0.1002 0.1723 0.1724 19.0
2.4416 10.0 1240 2.2069 0.1967 0.0995 0.1715 0.1716 19.0
2.4416 11.0 1364 2.2053 0.1983 0.102 0.1729 0.1732 19.0
2.4416 12.0 1488 2.2034 0.1976 0.1018 0.1722 0.1725 19.0
2.4153 13.0 1612 2.1995 0.1985 0.1019 0.1725 0.1727 19.0
2.4153 14.0 1736 2.1980 0.198 0.1016 0.1721 0.1722 19.0
2.4153 15.0 1860 2.1961 0.1983 0.1017 0.172 0.1721 19.0
2.4153 16.0 1984 2.1947 0.1977 0.1013 0.1715 0.1717 19.0
2.4069 17.0 2108 2.1936 0.1976 0.101 0.1714 0.1716 19.0
2.4069 18.0 2232 2.1925 0.1977 0.1013 0.1713 0.1715 19.0
2.4069 19.0 2356 2.1918 0.1973 0.1007 0.1709 0.1711 19.0
2.4069 20.0 2480 2.1919 0.1973 0.1007 0.1708 0.1711 19.0

Framework versions

  • Transformers 4.31.0
  • Pytorch 1.13.1.post200
  • Datasets 2.10.0
  • Tokenizers 0.13.2