ai-forever vmkhlv commited on
Commit
efdc76b
1 Parent(s): 419a854

Update README.md (#3)

Browse files

- Update README.md (6045ef6d58217161fbbc5fcbe00478b2fa9b891e)


Co-authored-by: Vladislav Mikhailov <vmkhlv@users.noreply.huggingface.co>

Files changed (1) hide show
  1. README.md +17 -3
README.md CHANGED
@@ -10,14 +10,28 @@ thumbnail: "https://github.com/sberbank-ai/model-zoo"
10
  pipeline_tag: fill-mask
11
  ---
12
  # ruBert-large
13
- Model was trained by [SberDevices](https://sberdevices.ru/) team.
 
 
14
  * Task: `mask filling`
15
  * Type: `encoder`
16
- * Tokenizer: `bpe`
17
  * Dict size: `120 138`
18
  * Num Parameters: `427 M`
19
  * Training Data Volume `30 GB`
20
 
21
  # Authors
22
  + NLP core team RnD [Telegram channel](https://t.me/nlpcoreteam):
23
- + Dmitry Zmitrovich
 
 
 
 
 
 
 
 
 
 
 
 
10
  pipeline_tag: fill-mask
11
  ---
12
  # ruBert-large
13
+ The model architecture design, pretraining, and evaluation are documented in our preprint: [**A Family of Pretrained Transformer Language Models for Russian**](https://arxiv.org/abs/2309.10931).
14
+
15
+ The model is pretrained by the [SberDevices](https://sberdevices.ru/) team.
16
  * Task: `mask filling`
17
  * Type: `encoder`
18
+ * Tokenizer: `BPE`
19
  * Dict size: `120 138`
20
  * Num Parameters: `427 M`
21
  * Training Data Volume `30 GB`
22
 
23
  # Authors
24
  + NLP core team RnD [Telegram channel](https://t.me/nlpcoreteam):
25
+ + Dmitry Zmitrovich
26
+
27
+ # Cite us
28
+ ```
29
+ @misc{zmitrovich2023family,
30
+ title={A Family of Pretrained Transformer Language Models for Russian},
31
+ author={Dmitry Zmitrovich and Alexander Abramov and Andrey Kalmykov and Maria Tikhonova and Ekaterina Taktasheva and Danil Astafurov and Mark Baushenko and Artem Snegirev and Tatiana Shavrina and Sergey Markov and Vladislav Mikhailov and Alena Fenogenova},
32
+ year={2023},
33
+ eprint={2309.10931},
34
+ archivePrefix={arXiv},
35
+ primaryClass={cs.CL}
36
+ }
37
+ ```