Update README.md
Browse files
README.md
CHANGED
@@ -14,7 +14,7 @@ widget:
|
|
14 |
|
15 |
## Model description
|
16 |
|
17 |
-
This is a Japanese BigBird base model pretrained on Japanese Wikipedia, the Japanese portion of CC-100, and the Japanese portion of
|
18 |
|
19 |
## How to use
|
20 |
|
@@ -41,7 +41,7 @@ The vocabulary consists of 32000 tokens including words ([JumanDIC](https://gith
|
|
41 |
|
42 |
## Training procedure
|
43 |
|
44 |
-
This model was trained on Japanese Wikipedia (as of 20221101), the Japanese portion of CC-100, and the and the Japanese portion of
|
45 |
|
46 |
The following hyperparameters were used during pretraining:
|
47 |
- learning_rate: 1e-4
|
@@ -59,7 +59,7 @@ The following hyperparameters were used during pretraining:
|
|
59 |
We fine-tuned the following models and evaluated them on the dev set of JGLUE.
|
60 |
We tuned learning rate and training epochs for each model and task following [the JGLUE paper](https://www.jstage.jst.go.jp/article/jnlp/30/1/30_63/_pdf/-char/ja).
|
61 |
|
62 |
-
For tasks other than MARC-ja, the maximum length is short, so the attention_type was set to "original_full" and fine-tuning was performed
|
63 |
|
64 |
| Model | MARC-ja/acc | JSTS/pearson | JSTS/spearman | JNLI/acc | JSQuAD/EM | JSQuAD/F1 | JComQA/acc |
|
65 |
|-------------------------------|--------------|---------------|----------|-----------|-----------|------------|------------|
|
|
|
14 |
|
15 |
## Model description
|
16 |
|
17 |
+
This is a Japanese BigBird base model pretrained on Japanese Wikipedia, the Japanese portion of CC-100, and the Japanese portion of OSCAR.
|
18 |
|
19 |
## How to use
|
20 |
|
|
|
41 |
|
42 |
## Training procedure
|
43 |
|
44 |
+
This model was trained on Japanese Wikipedia (as of 20221101), the Japanese portion of CC-100, and the and the Japanese portion of OSCAR. It took two weeks using 16 NVIDIA A100 GPUs using [transformers](https://github.com/huggingface/transformers) and [DeepSpeed](https://github.com/microsoft/DeepSpeed).
|
45 |
|
46 |
The following hyperparameters were used during pretraining:
|
47 |
- learning_rate: 1e-4
|
|
|
59 |
We fine-tuned the following models and evaluated them on the dev set of JGLUE.
|
60 |
We tuned learning rate and training epochs for each model and task following [the JGLUE paper](https://www.jstage.jst.go.jp/article/jnlp/30/1/30_63/_pdf/-char/ja).
|
61 |
|
62 |
+
For the tasks other than MARC-ja, the maximum length is short, so the attention_type was set to "original_full", and fine-tuning was performed. For MARC-ja, both "block_sparse" and "original_full" were used.
|
63 |
|
64 |
| Model | MARC-ja/acc | JSTS/pearson | JSTS/spearman | JNLI/acc | JSQuAD/EM | JSQuAD/F1 | JComQA/acc |
|
65 |
|-------------------------------|--------------|---------------|----------|-----------|-----------|------------|------------|
|