### google/flan-t5-base finetuned on xsum dataset | |
#### train args<br> | |
max_input_length: 512<br> | |
max_tgt_length: 128<br> | |
epoch: 3<br> | |
optimizer: AdamW<br> | |
lr: 2e-5<br> | |
weight_decay: 1e-3<br> | |
fp16: False<br> | |
prefix: "summarize: "<br> | |
#### performance<br> | |
eval_rouge1: 38.6648<br> | |
eval_rouge2: 15.5661<br> | |
eval_rougeL: 30.6158<br> | |
#### usage<br> | |
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM<br> | |