Fill-Mask
Transformers
PyTorch
Bulgarian
bert
torch
rmihaylov commited on
Commit
985c9aa
1 Parent(s): 5d8623a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -1
README.md CHANGED
@@ -16,7 +16,13 @@ tags:
16
  Pretrained model on Bulgarian language using a masked language modeling (MLM) objective. It was introduced in
17
  [this paper](https://arxiv.org/abs/1810.04805) and first released in
18
  [this repository](https://github.com/google-research/bert). This model is cased: it does make a difference
19
- between bulgarian and Bulgarian. The training data is Bulgarian text from [OSCAR](https://oscar-corpus.com/post/oscar-2019/), [Chitanka](https://chitanka.info/) and [Wikipedia](https://bg.wikipedia.org/).
 
 
 
 
 
 
20
 
21
  ### How to use
22
 
 
16
  Pretrained model on Bulgarian language using a masked language modeling (MLM) objective. It was introduced in
17
  [this paper](https://arxiv.org/abs/1810.04805) and first released in
18
  [this repository](https://github.com/google-research/bert). This model is cased: it does make a difference
19
+ between bulgarian and Bulgarian.
20
+
21
+ ## Model description
22
+
23
+ The model was trained similarly to [RuBert](https://arxiv.org/pdf/1905.07213.pdf) wherein the Multilingual Bert was adapted for the Russian language.
24
+
25
+ The training data was Bulgarian text from [OSCAR](https://oscar-corpus.com/post/oscar-2019/), [Chitanka](https://chitanka.info/) and [Wikipedia](https://bg.wikipedia.org/).
26
 
27
  ### How to use
28