amine commited on
Commit
dbce128
1 Parent(s): f48769d

docs(readme): update email

Browse files
Files changed (1) hide show
  1. README.md +60 -60
README.md CHANGED
@@ -1,60 +1,60 @@
1
- ---
2
- language:
3
- - ar
4
- - dz
5
-
6
- tags:
7
- - pytorch
8
- - bert
9
- - multilingual
10
- - ar
11
- - dz
12
-
13
- license: apache-2.0
14
-
15
- widget:
16
- - text: " أنا من الجزائر من ولاية [MASK] "
17
- - text: "rabi [MASK] khouya sami"
18
- - text: " ربي [MASK] خويا لعزيز"
19
- - text: "tahya el [MASK]."
20
- - text: "rouhi ya dzayer [MASK]"
21
-
22
- inference: true
23
- ---
24
-
25
- <img src="https://raw.githubusercontent.com/alger-ia/dziribert/main/dziribert_drawing.png" alt="drawing" width="25%" height="25%" align="right"/>
26
-
27
-
28
- # DziriBERT
29
-
30
-
31
- DziriBERT is the first Transformer-based Language Model that has been pre-trained specifically for the Algerian Dialect. It handles Algerian text contents written using both Arabic and Latin characters. It sets new state of the art results on Algerian text classification datasets, even if it has been pre-trained on much less data (~1 million tweets).
32
-
33
- For more information, please visit our paper: https://arxiv.org/pdf/2109.12346.pdf.
34
-
35
- ## How to use
36
-
37
- ```python
38
- from transformers import BertTokenizer, BertForMaskedLM
39
-
40
- tokenizer = BertTokenizer.from_pretrained("alger-ia/dziribert")
41
- model = BertForMaskedLM.from_pretrained("alger-ia/dziribert")
42
-
43
- ```
44
-
45
- You can find a fine-tuning script in our Github repo: https://github.com/alger-ia/dziribert
46
-
47
- ### How to cite
48
-
49
- ```bibtex
50
- @article{dziribert,
51
- title={DziriBERT: a Pre-trained Language Model for the Algerian Dialect},
52
- author={Abdaoui, Amine and Berrimi, Mohamed and Oussalah, Mourad and Moussaoui, Abdelouahab},
53
- journal={arXiv preprint arXiv:2109.12346},
54
- year={2021}
55
- }
56
- ```
57
-
58
- ## Contact
59
-
60
- Please contact amine.abdaoui@huawei.com for any question, feedback or request.
 
1
+ ---
2
+ language:
3
+ - ar
4
+ - dz
5
+
6
+ tags:
7
+ - pytorch
8
+ - bert
9
+ - multilingual
10
+ - ar
11
+ - dz
12
+
13
+ license: apache-2.0
14
+
15
+ widget:
16
+ - text: " أنا من الجزائر من ولاية [MASK] "
17
+ - text: "rabi [MASK] khouya sami"
18
+ - text: " ربي [MASK] خويا لعزيز"
19
+ - text: "tahya el [MASK]."
20
+ - text: "rouhi ya dzayer [MASK]"
21
+
22
+ inference: true
23
+ ---
24
+
25
+ <img src="https://raw.githubusercontent.com/alger-ia/dziribert/main/dziribert_drawing.png" alt="drawing" width="25%" height="25%" align="right"/>
26
+
27
+
28
+ # DziriBERT
29
+
30
+
31
+ DziriBERT is the first Transformer-based Language Model that has been pre-trained specifically for the Algerian Dialect. It handles Algerian text contents written using both Arabic and Latin characters. It sets new state of the art results on Algerian text classification datasets, even if it has been pre-trained on much less data (~1 million tweets).
32
+
33
+ For more information, please visit our paper: https://arxiv.org/pdf/2109.12346.pdf.
34
+
35
+ ## How to use
36
+
37
+ ```python
38
+ from transformers import BertTokenizer, BertForMaskedLM
39
+
40
+ tokenizer = BertTokenizer.from_pretrained("alger-ia/dziribert")
41
+ model = BertForMaskedLM.from_pretrained("alger-ia/dziribert")
42
+
43
+ ```
44
+
45
+ You can find a fine-tuning script in our Github repo: https://github.com/alger-ia/dziribert
46
+
47
+ ### How to cite
48
+
49
+ ```bibtex
50
+ @article{dziribert,
51
+ title={DziriBERT: a Pre-trained Language Model for the Algerian Dialect},
52
+ author={Abdaoui, Amine and Berrimi, Mohamed and Oussalah, Mourad and Moussaoui, Abdelouahab},
53
+ journal={arXiv preprint arXiv:2109.12346},
54
+ year={2021}
55
+ }
56
+ ```
57
+
58
+ ## Contact
59
+
60
+ Please contact amine.abdaoui.nlp@gmail.com for any question, feedback or request.