# Work in Progress Polish RoBERTa | |
The model has been trained for about 5% time of the target. We will publish new increments as they will be trained. | |
The model pre-trained on KGR10 corpora. | |
More about model at [CLARIN-dspace](https://huggingface.co/clarin/roberta-polish-v1) | |
## Usage | |
## Huggingface model hub | |
## Acknowledgments | |
[CLARIN-PL and CLARIN-BIZ project](https://clarin-pl.eu/) |