gsarti commited on
Commit
3e463ac
1 Parent(s): c32c4dc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -5
README.md CHANGED
@@ -47,7 +47,7 @@ thumbnail: https://gsarti.com/publication/it5/featured.png
47
  ---
48
  # IT5 Base for News Summarization ✂️🗞️ 🇮🇹
49
 
50
- This repository contains the checkpoint for the [IT5 Base](https://huggingface.co/gsarti/it5-base) model fine-tuned on news summarization on the [Fanpage](https://huggingface.co/datasets/ARTeLab/fanpage) and [Il Post](https://huggingface.co/datasets/ARTeLab/ilpost) corpora as part of the experiments of the paper [IT5: Large-scale Text-to-text Pretraining for Italian Language Understanding and Generation](https://arxiv.org) by [Gabriele Sarti](https://gsarti.com) and [Malvina Nissim](https://malvinanissim.github.io).
51
 
52
  A comprehensive overview of other released materials is provided in the [gsarti/it5](https://github.com/gsarti/it5) repository. Refer to the paper for additional details concerning the reported scores and the evaluation approach.
53
 
@@ -76,10 +76,11 @@ If you use this model in your research, please cite our work as:
76
 
77
  ```bibtex
78
  @article{sarti-nissim-2022-it5,
79
- title={IT5: Large-scale Text-to-text Pretraining for Italian Language Understanding and Generation},
80
  author={Sarti, Gabriele and Nissim, Malvina},
81
- journal={ArXiv preprint TBD},
82
- url={TBD},
83
- year={2022}
 
84
  }
85
  ```
 
47
  ---
48
  # IT5 Base for News Summarization ✂️🗞️ 🇮🇹
49
 
50
+ This repository contains the checkpoint for the [IT5 Base](https://huggingface.co/gsarti/it5-base) model fine-tuned on news summarization on the [Fanpage](https://huggingface.co/datasets/ARTeLab/fanpage) and [Il Post](https://huggingface.co/datasets/ARTeLab/ilpost) corpora as part of the experiments of the paper [IT5: Large-scale Text-to-text Pretraining for Italian Language Understanding and Generation](https://arxiv.org/abs/2203.03759) by [Gabriele Sarti](https://gsarti.com) and [Malvina Nissim](https://malvinanissim.github.io).
51
 
52
  A comprehensive overview of other released materials is provided in the [gsarti/it5](https://github.com/gsarti/it5) repository. Refer to the paper for additional details concerning the reported scores and the evaluation approach.
53
 
 
76
 
77
  ```bibtex
78
  @article{sarti-nissim-2022-it5,
79
+ title={{IT5}: Large-scale Text-to-text Pretraining for Italian Language Understanding and Generation},
80
  author={Sarti, Gabriele and Nissim, Malvina},
81
+ journal={ArXiv preprint 2203.03759},
82
+ url={https://arxiv.org/abs/2203.03759},
83
+ year={2022},
84
+ month={mar}
85
  }
86
  ```