RobeCzech model

RobeCzech is a monolingual RoBERTa language representation model trained on Czech data.

RobeCzech model is released publicly at LINDAT and Hugging Face.

Please cite the corresponding publication:

  • Milan Straka, Jakub Náplava, Jana Straková and David Samuel: Czech RoBERTa, a monolingual contextualized language representation model. Accepted to TSD 2021.

Preprint of the paper is available at


Select AutoNLP in the “Train” menu to fine-tune this model automatically.

Downloads last month
Hosted inference API
Mask token: <mask>
This model can be loaded on the Inference API on-demand.