--- base_model: google/flan-t5-large license: apache-2.0 metrics: - rouge tags: - generated_from_trainer model-index: - name: flan-t5-large-invoice results: [] --- # flan-t5-large-invoice This model is a fine-tuned version of [google/flan-t5-large](https://huggingface.co/google/flan-t5-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.7345 - Rouge1: 20.6872 - Rouge2: 8.404 - Rougel: 16.4285 - Rougelsum: 18.8189 - Gen Len: 19.0 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 1 - eval_batch_size: 1 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 20 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:-----:|:---------------:|:-------:|:------:|:-------:|:---------:|:-------:| | 1.779 | 1.0 | 933 | 1.3862 | 19.5863 | 7.7096 | 15.7946 | 18.3055 | 19.0 | | 1.3799 | 2.0 | 1866 | 1.3274 | 19.6876 | 6.7441 | 15.804 | 17.8499 | 19.0 | | 1.2505 | 3.0 | 2799 | 1.3099 | 19.8912 | 7.0276 | 16.3826 | 17.8505 | 19.0 | | 1.1726 | 4.0 | 3732 | 1.3136 | 20.9896 | 8.5214 | 16.7944 | 19.2926 | 19.0 | | 1.0985 | 5.0 | 4665 | 1.3205 | 20.5499 | 7.9224 | 16.3301 | 18.519 | 19.0 | | 1.0439 | 6.0 | 5598 | 1.3297 | 20.8897 | 7.6685 | 16.6002 | 19.1149 | 19.0 | | 0.9912 | 7.0 | 6531 | 1.3515 | 21.0205 | 7.9648 | 16.6663 | 19.1507 | 19.0 | | 0.922 | 8.0 | 7464 | 1.3819 | 20.0508 | 7.0989 | 16.2521 | 18.2977 | 19.0 | | 0.8729 | 9.0 | 8397 | 1.4222 | 20.3315 | 8.1219 | 16.5236 | 18.8624 | 19.0 | | 0.8353 | 10.0 | 9330 | 1.4372 | 19.3277 | 7.4633 | 15.5985 | 17.6354 | 19.0 | | 0.7928 | 11.0 | 10263 | 1.4851 | 20.1063 | 7.9988 | 16.0191 | 18.3571 | 19.0 | | 0.7561 | 12.0 | 11196 | 1.5084 | 20.6322 | 7.2781 | 16.3066 | 18.8342 | 19.0 | | 0.7251 | 13.0 | 12129 | 1.5437 | 21.4814 | 8.1047 | 17.0497 | 19.8034 | 19.0 | | 0.692 | 14.0 | 13062 | 1.5802 | 21.1577 | 8.0934 | 16.738 | 19.2418 | 19.0 | | 0.668 | 15.0 | 13995 | 1.6349 | 19.8489 | 7.7261 | 15.8122 | 18.0971 | 19.0 | | 0.6237 | 16.0 | 14928 | 1.6246 | 20.8915 | 8.2484 | 16.8627 | 19.0533 | 19.0 | | 0.6166 | 17.0 | 15861 | 1.6662 | 20.3532 | 8.2927 | 16.2535 | 18.5981 | 19.0 | | 0.5973 | 18.0 | 16794 | 1.6963 | 20.4374 | 7.9296 | 16.2445 | 18.4927 | 19.0 | | 0.581 | 19.0 | 17727 | 1.7168 | 20.6517 | 8.3175 | 16.4803 | 18.8024 | 19.0 | | 0.5704 | 20.0 | 18660 | 1.7345 | 20.6872 | 8.404 | 16.4285 | 18.8189 | 19.0 | ### Framework versions - Transformers 4.43.2 - Pytorch 2.2.0a0+81ea7a4 - Datasets 2.20.0 - Tokenizers 0.19.1