File size: 749 Bytes
d22bd1f 7de52db e7d29dc 1f83bbd e7d29dc |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 |
---
license: apache-2.0
datasets:
- tay-yozhik/SyntheticTexts
language:
- ru
---
# SyntheticRoBERTa
This is a pre-trained model of type [RoBERTa](https://arxiv.org/abs/1907.11692).
SyntheticRoBERTa is built on a dataset of texts in Russian, which were generated according to [the rules written in context-free grammar](https://github.com/GorbachevaTaisia/JSGF_generative_grammar).
# Evaluation
This model was evaluated on [RussianSuperGLUE tests](https://russiansuperglue.com/):
| Task | Result | Metrics |
|-------|----------|---------|
| LiDiRus | 0,0 | Matthews Correlation Coefficient |
| RCB | 0,091 / 0,158 | F1 / Accuracy |
| PARus | 0,502 | Accuracy |
| TERRa | 0,487 | Accuracy |
| RUSSE | 0,587 | Accuracy |
| RWSD | 0,331 | Accuracy | |