File size: 868 Bytes
b6fb017 02f63ad b6fb017 2b024b0 b6fb017 170d422 b6fb017 d6c8b30 7fe5722 d6c8b30 b6fb017 0e2c7ea |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 |
---
license: apache-2.0
widget:
- text: 横浜国立大学は日本の[MASK]奈川県にある。
---
This is RoBERTa model pretrained on texts in the Japanese language.
3.45GB wikipedia text
trained 125M step
use the BERT BPE tokenizer.
If you want to fine-tune model. Please use
```python
from transformers import BertTokenizer, RoFormerModel
BertTokenizer.from_pretrained('Roformer-base-japanese')
RoFormerModel.from_pretrained('Roformer-base-japanese')
```
The accuracy in JGLUE-marc-ja-v1.0 binary sentiment classification 95.12%
Contribute by Yokohama Nationaly University Mori Lab
@article{su2021roformer,
title={Roformer: Enhanced transformer with rotary position embedding},
author={Su, Jianlin and Lu, Yu and Pan, Shengfeng and Wen, Bo and Liu, Yunfeng},
journal={arXiv preprint arXiv:2104.09864},
year={2021}
} |