Ebtihal commited on
Commit
fd6a60a
1 Parent(s): 19e898c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +36 -12
README.md CHANGED
@@ -1,15 +1,39 @@
1
- Arabic Model AraBertMo_base_V3
2
  ---
3
- language:
4
- - ar
5
- tags:
6
- - Fill-Mask
7
- license: apache-2.0
8
- datasets:
9
- - OSCAR
10
  widget:
11
  - text: " السلام عليكم ورحمة[MASK] وبركاتة"
12
- example_title: "Example 1"
13
- - text: "مرحبا بك عزيزي الزائر [MASK] موقعنا"
14
- example_title: "Example 2"
15
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language: ar
3
+ tags: Fill-Mask
4
+ datasets: OSCAR
 
 
 
 
5
  widget:
6
  - text: " السلام عليكم ورحمة[MASK] وبركاتة"
7
+ - text: " اهلا وسهلا بكم في [MASK] من سيربح المليون "
8
+ ---
9
+ # Arabic BERT Model
10
+ **AraBERTMo** is an Arabic pre-trained language model based on [Google's BERT architechture](https://github.com/google-research/bert).
11
+ AraBERTMo_base uses the same BERT-Base config.
12
+ AraBERTMo_base now comes in 10 new variants
13
+ All models are available on the `HuggingFace` model page under the [Ebtihal](https://huggingface.co/Ebtihal/) name.
14
+ Checkpoints are available in PyTorch formats.
15
+
16
+ ## Pretraining Corpus
17
+ `AraBertMo_base_V3' model was pre-trained on ~3 million words:
18
+ - [OSCAR](https://traces1.inria.fr/oscar/) - Arabic version "unshuffled_deduplicated_ar".
19
+
20
+ ## Training results
21
+ this model achieves the following results:
22
+
23
+ | Task | Num examples | Num Epochs | Batch Size | steps | Wall time | training loss|
24
+ |:----:|:----:|:----:|:----:|:-----:|:----:|:-----:|
25
+ | Fill-Mask| 30024| 3 | 64 | 1410 | 3h 10m 31s | 8.0201 |
26
+
27
+ ## Load Pretrained Model
28
+ You can use this model by installing `torch` or `tensorflow` and Huggingface library `transformers`. And you can use it directly by initializing it like this:
29
+ ```python
30
+ from transformers import AutoTokenizer, AutoModel
31
+ tokenizer = AutoTokenizer.from_pretrained("Ebtihal/AraBertMo_base_V3")
32
+ model = AutoModelForMaskedLM.from_pretrained("Ebtihal/AraBertMo_base_V3")
33
+ ```
34
+
35
+ ## This model was built for master's degree research in an organization:
36
+ - [University of kufa](https://uokufa.edu.iq/).
37
+ - [Faculty of Computer Science and Mathematics](https://mathcomp.uokufa.edu.iq/).
38
+ - **Department of Computer Science**
39
+