flan-t5-base-extraction-cnndm_4000-all-hint_precision-ep50
This model is a fine-tuned version of google/flan-t5-base on the None dataset. It achieves the following results on the evaluation set:
- Loss: 1.8125
- Hint Hit Num: 2.3766
- Hint Precision: 0.4326
- Num: 5.5019
- Gen Len: 18.9989
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 32
- eval_batch_size: 100
- seed: 1799
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 50
Training results
Training Loss | Epoch | Step | Validation Loss | Hint Hit Num | Hint Precision | Num | Gen Len |
---|---|---|---|---|---|---|---|
2.321 | 1.6 | 200 | 1.8768 | 2.2035 | 0.4107 | 5.3583 | 18.9861 |
2.0736 | 3.2 | 400 | 1.8125 | 2.3766 | 0.4326 | 5.5019 | 18.9989 |
1.9909 | 4.8 | 600 | 1.7994 | 2.2962 | 0.4233 | 5.4165 | 18.9978 |
1.9262 | 6.4 | 800 | 1.7771 | 2.3816 | 0.4323 | 5.5021 | 18.9999 |
1.8874 | 8.0 | 1000 | 1.7705 | 2.3657 | 0.4295 | 5.493 | 18.9997 |
1.8487 | 9.6 | 1200 | 1.7739 | 2.3792 | 0.4299 | 5.5187 | 18.9994 |
1.8073 | 11.2 | 1400 | 1.7743 | 2.3692 | 0.4302 | 5.4875 | 18.9993 |
Framework versions
- Transformers 4.18.0
- Pytorch 1.10.0+cu111
- Datasets 2.5.1
- Tokenizers 0.12.1
- Downloads last month
- 1