elmadany commited on
Commit
0641745
1 Parent(s): 72c8287

commit from elmadany

Browse files
Files changed (1) hide show
  1. README.md +1 -0
README.md CHANGED
@@ -7,6 +7,7 @@ We find that results with ARBERT and MARBERT on QA are not competitive, a clear
7
  To rectify this, we further pre-train the stronger model, MARBERT, on the same MSA data as ARBERT in addition to AraNews dataset but with a bigger sequence length of 512 tokens for 40 epochs. We call this
8
  further pre-trained model **MARBERTv2**, noting it has **29B tokens**. MARBERTv2 acquires best performance on all but one test set, where XLM-RLarge marginally outperforms us (only in F1).
9
 
 
10
 
11
 
12
 
 
7
  To rectify this, we further pre-train the stronger model, MARBERT, on the same MSA data as ARBERT in addition to AraNews dataset but with a bigger sequence length of 512 tokens for 40 epochs. We call this
8
  further pre-trained model **MARBERTv2**, noting it has **29B tokens**. MARBERTv2 acquires best performance on all but one test set, where XLM-RLarge marginally outperforms us (only in F1).
9
 
10
+ For more information, please visit our own GitHub [repo](https://github.com/UBC-NLP/marbert).
11
 
12
 
13