3_loa / README.md
eschorn's picture
update model card README.md
1abff9b
---
license: apache-2.0
base_model: google/flan-t5-large
tags:
- generated_from_trainer
datasets:
- billsum
metrics:
- rouge
model-index:
- name: 3_loa
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# 3_loa
This model is a fine-tuned version of [google/flan-t5-large](https://huggingface.co/google/flan-t5-large) on the billsum dataset.
It achieves the following results on the evaluation set:
- Loss: 1.4825
- Rouge1: 0.201
- Rouge2: 0.1132
- Rougel: 0.1753
- Rougelsum: 0.1755
- Gen Len: 19.0
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 1
- eval_batch_size: 1
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 20
### Training results
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
|:-------------:|:-----:|:-----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|
| 2.1079 | 1.0 | 989 | 1.6673 | 0.2028 | 0.1092 | 0.1748 | 0.1751 | 19.0 |
| 1.8481 | 2.0 | 1978 | 1.6150 | 0.1979 | 0.1061 | 0.1715 | 0.1717 | 19.0 |
| 1.7889 | 3.0 | 2967 | 1.5833 | 0.1994 | 0.11 | 0.1727 | 0.1727 | 19.0 |
| 1.7319 | 4.0 | 3956 | 1.5584 | 0.1978 | 0.1084 | 0.1718 | 0.1718 | 19.0 |
| 1.7279 | 5.0 | 4945 | 1.5440 | 0.2016 | 0.1106 | 0.1755 | 0.1756 | 19.0 |
| 1.7386 | 6.0 | 5934 | 1.5326 | 0.1991 | 0.1086 | 0.1734 | 0.1736 | 19.0 |
| 1.6972 | 7.0 | 6923 | 1.5251 | 0.2013 | 0.1122 | 0.1759 | 0.176 | 19.0 |
| 1.6732 | 8.0 | 7912 | 1.5145 | 0.2024 | 0.1123 | 0.1766 | 0.1766 | 19.0 |
| 1.6597 | 9.0 | 8901 | 1.5079 | 0.2019 | 0.1125 | 0.1751 | 0.1753 | 19.0 |
| 1.6151 | 10.0 | 9890 | 1.5045 | 0.201 | 0.1123 | 0.1758 | 0.1761 | 19.0 |
| 1.6381 | 11.0 | 10879 | 1.4997 | 0.2009 | 0.1116 | 0.1755 | 0.1756 | 19.0 |
| 1.6148 | 12.0 | 11868 | 1.4974 | 0.2018 | 0.1133 | 0.1763 | 0.1765 | 19.0 |
| 1.6196 | 13.0 | 12857 | 1.4940 | 0.2014 | 0.1129 | 0.1756 | 0.1756 | 19.0 |
| 1.6137 | 14.0 | 13846 | 1.4914 | 0.2025 | 0.1136 | 0.1766 | 0.1768 | 19.0 |
| 1.6313 | 15.0 | 14835 | 1.4873 | 0.2032 | 0.114 | 0.1769 | 0.1771 | 19.0 |
| 1.6098 | 16.0 | 15824 | 1.4847 | 0.2012 | 0.1133 | 0.175 | 0.1754 | 19.0 |
| 1.6061 | 17.0 | 16813 | 1.4845 | 0.2019 | 0.1138 | 0.1752 | 0.1755 | 19.0 |
| 1.5918 | 18.0 | 17802 | 1.4833 | 0.2011 | 0.1129 | 0.1747 | 0.175 | 19.0 |
| 1.5842 | 19.0 | 18791 | 1.4824 | 0.2013 | 0.1133 | 0.1753 | 0.1755 | 19.0 |
| 1.5964 | 20.0 | 19780 | 1.4825 | 0.201 | 0.1132 | 0.1753 | 0.1755 | 19.0 |
### Framework versions
- Transformers 4.31.0
- Pytorch 1.13.1.post200
- Datasets 2.10.0
- Tokenizers 0.13.2