RobeCzech model

RobeCzech is a monolingual RoBERTa language representation model trained on Czech data.

RobeCzech model is released publicly at LINDAT and Hugging Face.

Please cite the corresponding publication:

  • Milan Straka, Jakub Náplava, Jana Straková and David Samuel: Czech RoBERTa, a monolingual contextualized language representation model. Accepted to TSD 2021.

Preprint of the paper is available at

New: fine-tune this model in a few clicks by selecting AutoNLP in the "Train" menu!
Downloads last month
Hosted inference API
Mask token: <mask>
This model can be loaded on the Inference API on-demand.