ctoraman commited on
Commit
8ff6a58
1 Parent(s): 4cb9519

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -15,7 +15,7 @@ The pretrained corpus is OSCAR's Turkish split, but it is further filtered and c
15
 
16
  Model architecture is similar to bert-medium (8 layers, 8 heads, and 512 hidden size). Tokenization algorithm is Word-level, which means text is split by white space. Vocabulary size is 66.7k.
17
 
18
- The details can be found at this paper:
19
  https://arxiv.org/abs/2204.08832
20
 
21
  The following code can be used for model loading and tokenization, example max length (514) can be changed:
 
15
 
16
  Model architecture is similar to bert-medium (8 layers, 8 heads, and 512 hidden size). Tokenization algorithm is Word-level, which means text is split by white space. Vocabulary size is 66.7k.
17
 
18
+ The details and performance comparisons can be found at this paper:
19
  https://arxiv.org/abs/2204.08832
20
 
21
  The following code can be used for model loading and tokenization, example max length (514) can be changed: