Update README.md
Browse files
README.md
CHANGED
@@ -60,7 +60,7 @@ We first segmented texts in the corpora into words using [Juman++](https://githu
|
|
60 |
Then, we built a sentencepiece model with 32000 tokens including words ([JumanDIC](https://github.com/ku-nlp/JumanDIC)) and subwords induced by the unigram language model of [sentencepiece](https://github.com/google/sentencepiece).
|
61 |
|
62 |
We tokenized the segmented corpora into subwords using the sentencepiece model and trained the Japanese DeBERTa model using [transformers](https://github.com/huggingface/transformers) library.
|
63 |
-
The training took
|
64 |
|
65 |
The following hyperparameters were used during pre-training:
|
66 |
|
|
|
60 |
Then, we built a sentencepiece model with 32000 tokens including words ([JumanDIC](https://github.com/ku-nlp/JumanDIC)) and subwords induced by the unigram language model of [sentencepiece](https://github.com/google/sentencepiece).
|
61 |
|
62 |
We tokenized the segmented corpora into subwords using the sentencepiece model and trained the Japanese DeBERTa model using [transformers](https://github.com/huggingface/transformers) library.
|
63 |
+
The training took 33 hours using 8 NVIDIA A100-SXM4-40GB GPUs.
|
64 |
|
65 |
The following hyperparameters were used during pre-training:
|
66 |
|