Update README.md
Browse files
README.md
CHANGED
@@ -112,7 +112,7 @@ print(f'*cosine_score:{cosine_scores[0]}')
|
|
112 |
|
113 |
**MLM(Masked Langeuage Model) 훈련**
|
114 |
- 입력 모델 : bert-base-multilingual-cased(vocab(119,548개))
|
115 |
-
- 말뭉치 : 훈련 : bongsoo/moco-corpus-kowiki2022(7.6M) , 평가: bongsoo/bongevalsmall
|
116 |
- HyperParameter : LearningRate : 5e-5, epochs: 8, batchsize: 32, max_token_len : 128
|
117 |
- vocab : 152,537개 (기존 119,548 에 32,989 신규 vocab 추가)
|
118 |
- 출력 모델 : mbertV2.0 (size: 813MB)
|
|
|
112 |
|
113 |
**MLM(Masked Langeuage Model) 훈련**
|
114 |
- 입력 모델 : bert-base-multilingual-cased(vocab(119,548개))
|
115 |
+
- 말뭉치 : 훈련 : bongsoo/moco-corpus-kowiki2022(7.6M) , 평가: bongsoo/bongevalsmall(200)
|
116 |
- HyperParameter : LearningRate : 5e-5, epochs: 8, batchsize: 32, max_token_len : 128
|
117 |
- vocab : 152,537개 (기존 119,548 에 32,989 신규 vocab 추가)
|
118 |
- 출력 모델 : mbertV2.0 (size: 813MB)
|