Spanish T5 (small) trained on large_spanish_corpus.

This is a Spanish T5 (small arch) trained from scratch on the large_spanish_corpus aka BETO's corpus with Flax

This is part of the Flax/Jax Community Week, organised by HuggingFace and TPU usage sponsored by Google.


The dataset is about 20 GB. 95% of the data was used for training and the rest 5% for validation.

Metrics (on evaluation dataset)

  • Accuracy: 0.675

Team members


Select AutoNLP in the “Train” menu to fine-tune this model automatically.

Downloads last month
Hosted inference API
Text2Text Generation
This model can be loaded on the Inference API on-demand.