File size: 868 Bytes
b6fb017
 
 
 
 
 
 
 
 
 
 
 
02f63ad
b6fb017
2b024b0
b6fb017
 
 
 
 
170d422
b6fb017
 
 
 
 
d6c8b30
7fe5722
d6c8b30
b6fb017
 
0e2c7ea
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
---
license: apache-2.0

widget:
- text: 横浜国立大学は日本の[MASK]奈川県にある。
---


This is RoBERTa model pretrained on texts in the Japanese language.

3.45GB wikipedia text

 trained 125M step
 
  use the BERT BPE tokenizer.
  
  If you want to fine-tune model. Please use 
```python


from transformers import BertTokenizer, RoFormerModel
BertTokenizer.from_pretrained('Roformer-base-japanese')
RoFormerModel.from_pretrained('Roformer-base-japanese')
```
 
  
  The accuracy in JGLUE-marc-ja-v1.0 binary sentiment classification 95.12%

  
  
  
 Contribute by Yokohama Nationaly University Mori Lab
 
 @article{su2021roformer,
  title={Roformer: Enhanced transformer with rotary position embedding},
  author={Su, Jianlin and Lu, Yu and Pan, Shengfeng and Wen, Bo and Liu, Yunfeng},
  journal={arXiv preprint arXiv:2104.09864},
  year={2021}
}