File size: 1,170 Bytes
680d85d
 
 
 
 
 
d594fe9
 
680d85d
d594fe9
680d85d
 
efdc76b
 
 
680d85d
 
efdc76b
680d85d
 
419a854
 
 
 
efdc76b
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
---
language:
- ru
tags:
- PyTorch
- Transformers
- bert
- exbert
thumbnail: "https://github.com/sberbank-ai/model-zoo"
pipeline_tag: fill-mask
---
# ruBert-large
The model architecture design, pretraining, and evaluation are documented in our preprint: [**A Family of Pretrained Transformer Language Models for Russian**](https://arxiv.org/abs/2309.10931).

The model is pretrained by the [SberDevices](https://sberdevices.ru/) team.  
* Task: `mask filling`
* Type: `encoder`
* Tokenizer: `BPE`
* Dict size: `120 138`
* Num Parameters: `427 M`	
* Training Data Volume `30 GB`

# Authors
+ NLP core team RnD [Telegram channel](https://t.me/nlpcoreteam):
  + Dmitry Zmitrovich
 
# Cite us
```
@misc{zmitrovich2023family,
      title={A Family of Pretrained Transformer Language Models for Russian}, 
      author={Dmitry Zmitrovich and Alexander Abramov and Andrey Kalmykov and Maria Tikhonova and Ekaterina Taktasheva and Danil Astafurov and Mark Baushenko and Artem Snegirev and Tatiana Shavrina and Sergey Markov and Vladislav Mikhailov and Alena Fenogenova},
      year={2023},
      eprint={2309.10931},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}
```