|
--- |
|
license: apache-2.0 |
|
|
|
widget: |
|
- text: 横浜国立大学は日本の[MASK]奈川県にある。 |
|
--- |
|
|
|
|
|
This is RoBERTa model pretrained on texts in the Japanese language. |
|
|
|
3.45GB wikipedia text |
|
|
|
trained 125M step |
|
|
|
use the BERT BPE tokenizer. |
|
|
|
If you want to fine-tune model. Please use |
|
```python |
|
|
|
|
|
from transformers import BertTokenizer, RoFormerModel |
|
BertTokenizer.from_pretrained('Roformer-base-japanese') |
|
RoFormerModel.from_pretrained('Roformer-base-japanese') |
|
``` |
|
|
|
|
|
The accuracy in JGLUE-marc-ja-v1.0 binary sentiment classification 95.12% |
|
|
|
|
|
|
|
|
|
Contribute by Yokohama Nationaly University Mori Lab |
|
|
|
@article{su2021roformer, |
|
title={Roformer: Enhanced transformer with rotary position embedding}, |
|
author={Su, Jianlin and Lu, Yu and Pan, Shengfeng and Wen, Bo and Liu, Yunfeng}, |
|
journal={arXiv preprint arXiv:2104.09864}, |
|
year={2021} |
|
} |