Update README.md
Browse files
README.md
CHANGED
@@ -3,8 +3,7 @@ language:
|
|
3 |
- ru
|
4 |
---
|
5 |
# distilrubert-tiny-cased-conversational
|
6 |
-
Conversational DistilRuBERT-tiny \(Russian, cased, 3‑layers, 264‑hidden, 12‑heads,
|
7 |
-
|
8 |
|
9 |
Our DistilRuBERT-tiny is highly inspired by \[3\], \[4\] and architecture is very close to \[5\]. Namely, we use
|
10 |
* MLM loss (between token labels and student output distribution)
|
@@ -47,6 +46,28 @@ We used `PyTorchBenchmark` from `transformers` to evaluate model's performance a
|
|
47 |
|
48 |
To evaluate model quality, we fine-tuned DistilRuBERT-tiny on classification (RuSentiment, ParaPhraser), NER and question answering data sets for Russian and obtained scores very similar to the [Conversational DistilRuBERT-small](https://huggingface.co/DeepPavlov/distilrubert-tiny-cased-conversational).
|
49 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
50 |
\[1\]: P. Lison and J. Tiedemann, 2016, OpenSubtitles2016: Extracting Large Parallel Corpora from Movie and TV Subtitles. In Proceedings of the 10th International Conference on Language Resources and Evaluation \(LREC 2016\)
|
51 |
|
52 |
\[2\]: Shavrina T., Shapovalova O. \(2017\) TO THE METHODOLOGY OF CORPUS CONSTRUCTION FOR MACHINE LEARNING: «TAIGA» SYNTAX TREE CORPUS AND PARSER. in proc. of “CORPORA2017”, international conference , Saint-Petersbourg, 2017.
|
|
|
3 |
- ru
|
4 |
---
|
5 |
# distilrubert-tiny-cased-conversational
|
6 |
+
Conversational DistilRuBERT-tiny \(Russian, cased, 3‑layers, 264‑hidden, 12‑heads, 10.4M parameters\) was trained on OpenSubtitles\[1\], [Dirty](https://d3.ru/), [Pikabu](https://pikabu.ru/), and a Social Media segment of Taiga corpus\[2\] (as [Conversational RuBERT](https://huggingface.co/DeepPavlov/rubert-base-cased-conversational)). It can be considered as tiny copy of [Conversational DistilRuBERT-small](https://huggingface.co/DeepPavlov/distilrubert-tiny-cased-conversational).
|
|
|
7 |
|
8 |
Our DistilRuBERT-tiny is highly inspired by \[3\], \[4\] and architecture is very close to \[5\]. Namely, we use
|
9 |
* MLM loss (between token labels and student output distribution)
|
|
|
46 |
|
47 |
To evaluate model quality, we fine-tuned DistilRuBERT-tiny on classification (RuSentiment, ParaPhraser), NER and question answering data sets for Russian and obtained scores very similar to the [Conversational DistilRuBERT-small](https://huggingface.co/DeepPavlov/distilrubert-tiny-cased-conversational).
|
48 |
|
49 |
+
# Citation
|
50 |
+
If you found the model useful for your research, we are kindly ask to cite [this](https://arxiv.org/abs/2205.02340) paper:
|
51 |
+
```
|
52 |
+
@misc{https://doi.org/10.48550/arxiv.2205.02340,
|
53 |
+
doi = {10.48550/ARXIV.2205.02340},
|
54 |
+
|
55 |
+
url = {https://arxiv.org/abs/2205.02340},
|
56 |
+
|
57 |
+
author = {Kolesnikova, Alina and Kuratov, Yuri and Konovalov, Vasily and Burtsev, Mikhail},
|
58 |
+
|
59 |
+
keywords = {Computation and Language (cs.CL), Machine Learning (cs.LG), FOS: Computer and information sciences, FOS: Computer and information sciences},
|
60 |
+
|
61 |
+
title = {Knowledge Distillation of Russian Language Models with Reduction of Vocabulary},
|
62 |
+
|
63 |
+
publisher = {arXiv},
|
64 |
+
|
65 |
+
year = {2022},
|
66 |
+
|
67 |
+
copyright = {arXiv.org perpetual, non-exclusive license}
|
68 |
+
}
|
69 |
+
```
|
70 |
+
|
71 |
\[1\]: P. Lison and J. Tiedemann, 2016, OpenSubtitles2016: Extracting Large Parallel Corpora from Movie and TV Subtitles. In Proceedings of the 10th International Conference on Language Resources and Evaluation \(LREC 2016\)
|
72 |
|
73 |
\[2\]: Shavrina T., Shapovalova O. \(2017\) TO THE METHODOLOGY OF CORPUS CONSTRUCTION FOR MACHINE LEARNING: «TAIGA» SYNTAX TREE CORPUS AND PARSER. in proc. of “CORPORA2017”, international conference , Saint-Petersbourg, 2017.
|