Update README.md
Browse files
README.md
CHANGED
@@ -55,7 +55,7 @@ co2_eq_emissions:
|
|
55 |
|
56 |
# mT5 Base for Informal-to-formal Style Transfer 🧐
|
57 |
|
58 |
-
This repository contains the checkpoint for the [mT5 Base](https://huggingface.co/google/mt5-base) model fine-tuned on Informal-to-formal style transfer on the Italian subset of the XFORMAL dataset as part of the experiments of the paper [IT5: Large-scale Text-to-text Pretraining for Italian Language Understanding and Generation](https://arxiv.org) by [Gabriele Sarti](https://gsarti.com) and [Malvina Nissim](https://malvinanissim.github.io).
|
59 |
|
60 |
A comprehensive overview of other released materials is provided in the [gsarti/it5](https://github.com/gsarti/it5) repository. Refer to the paper for additional details concerning the reported scores and the evaluation approach.
|
61 |
|
@@ -84,10 +84,11 @@ If you use this model in your research, please cite our work as:
|
|
84 |
|
85 |
```bibtex
|
86 |
@article{sarti-nissim-2022-it5,
|
87 |
-
title={IT5: Large-scale Text-to-text Pretraining for Italian Language Understanding and Generation},
|
88 |
author={Sarti, Gabriele and Nissim, Malvina},
|
89 |
-
journal={ArXiv preprint
|
90 |
-
url={
|
91 |
-
year={2022}
|
|
|
92 |
}
|
93 |
```
|
|
|
55 |
|
56 |
# mT5 Base for Informal-to-formal Style Transfer 🧐
|
57 |
|
58 |
+
This repository contains the checkpoint for the [mT5 Base](https://huggingface.co/google/mt5-base) model fine-tuned on Informal-to-formal style transfer on the Italian subset of the XFORMAL dataset as part of the experiments of the paper [IT5: Large-scale Text-to-text Pretraining for Italian Language Understanding and Generation](https://arxiv.org/abs/2203.03759) by [Gabriele Sarti](https://gsarti.com) and [Malvina Nissim](https://malvinanissim.github.io).
|
59 |
|
60 |
A comprehensive overview of other released materials is provided in the [gsarti/it5](https://github.com/gsarti/it5) repository. Refer to the paper for additional details concerning the reported scores and the evaluation approach.
|
61 |
|
|
|
84 |
|
85 |
```bibtex
|
86 |
@article{sarti-nissim-2022-it5,
|
87 |
+
title={{IT5}: Large-scale Text-to-text Pretraining for Italian Language Understanding and Generation},
|
88 |
author={Sarti, Gabriele and Nissim, Malvina},
|
89 |
+
journal={ArXiv preprint 2203.03759},
|
90 |
+
url={https://arxiv.org/abs/2203.03759},
|
91 |
+
year={2022},
|
92 |
+
month={mar}
|
93 |
}
|
94 |
```
|