mkmkmkmk commited on
Commit
9dcb5bf
1 Parent(s): 83d6d92

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -14,7 +14,7 @@ widget:
14
 
15
  ## Model description
16
 
17
- This is a Japanese BigBird base model pretrained on Japanese Wikipedia, the Japanese portion of CC-100, and the Japanese portion of oscar.
18
 
19
  ## How to use
20
 
@@ -41,7 +41,7 @@ The vocabulary consists of 32000 tokens including words ([JumanDIC](https://gith
41
 
42
  ## Training procedure
43
 
44
- This model was trained on Japanese Wikipedia (as of 20221101), the Japanese portion of CC-100, and the and the Japanese portion of oscar. It took two weeks using 16 NVIDIA A100 GPUs using [transformers](https://github.com/huggingface/transformers) and [DeepSpeed](https://github.com/microsoft/DeepSpeed).
45
 
46
  The following hyperparameters were used during pretraining:
47
  - learning_rate: 1e-4
@@ -59,7 +59,7 @@ The following hyperparameters were used during pretraining:
59
  We fine-tuned the following models and evaluated them on the dev set of JGLUE.
60
  We tuned learning rate and training epochs for each model and task following [the JGLUE paper](https://www.jstage.jst.go.jp/article/jnlp/30/1/30_63/_pdf/-char/ja).
61
 
62
- For tasks other than MARC-ja, the maximum length is short, so the attention_type was set to "original_full" and fine-tuning was performed for tasks other than MARC-ja. For MARC-ja, both "block_sparse" and "original_full" were used.
63
 
64
  | Model | MARC-ja/acc | JSTS/pearson | JSTS/spearman | JNLI/acc | JSQuAD/EM | JSQuAD/F1 | JComQA/acc |
65
  |-------------------------------|--------------|---------------|----------|-----------|-----------|------------|------------|
 
14
 
15
  ## Model description
16
 
17
+ This is a Japanese BigBird base model pretrained on Japanese Wikipedia, the Japanese portion of CC-100, and the Japanese portion of OSCAR.
18
 
19
  ## How to use
20
 
 
41
 
42
  ## Training procedure
43
 
44
+ This model was trained on Japanese Wikipedia (as of 20221101), the Japanese portion of CC-100, and the and the Japanese portion of OSCAR. It took two weeks using 16 NVIDIA A100 GPUs using [transformers](https://github.com/huggingface/transformers) and [DeepSpeed](https://github.com/microsoft/DeepSpeed).
45
 
46
  The following hyperparameters were used during pretraining:
47
  - learning_rate: 1e-4
 
59
  We fine-tuned the following models and evaluated them on the dev set of JGLUE.
60
  We tuned learning rate and training epochs for each model and task following [the JGLUE paper](https://www.jstage.jst.go.jp/article/jnlp/30/1/30_63/_pdf/-char/ja).
61
 
62
+ For the tasks other than MARC-ja, the maximum length is short, so the attention_type was set to "original_full", and fine-tuning was performed. For MARC-ja, both "block_sparse" and "original_full" were used.
63
 
64
  | Model | MARC-ja/acc | JSTS/pearson | JSTS/spearman | JNLI/acc | JSQuAD/EM | JSQuAD/F1 | JComQA/acc |
65
  |-------------------------------|--------------|---------------|----------|-----------|-----------|------------|------------|