amine-private commited on
Commit
fda800c
1 Parent(s): 3ed253a

First upload from amine

Browse files
Files changed (5) hide show
  1. README.md +54 -0
  2. config.json +24 -0
  3. pytorch_model.bin +3 -0
  4. tokenizer_config.json +1 -0
  5. vocab.txt +0 -0
README.md ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: multilingual
3
+
4
+ datasets: wikipedia
5
+
6
+ license: apache-2.0
7
+
8
+ widget:
9
+ - text: "Google generated 46 billion [MASK] in revenue."
10
+ - text: "Paris is the capital of [MASK]."
11
+ - text: "Algiers is the largest city in [MASK]."
12
+ - text: "Paris est la [MASK] de la France."
13
+ - text: "Paris est la capitale de la [MASK]."
14
+ - text: "L'élection américaine a eu [MASK] en novembre 2020."
15
+ - text: "تقع سويسرا في [MASK] أوروبا"
16
+ - text: "إسمي محمد وأسكن في [MASK]."
17
+ ---
18
+
19
+ # distilbert-base-25lang-cased
20
+
21
+ We are sharing smaller versions of [distilbert-base-multilingual-cased](https://huggingface.co/distilbert-base-multilingual-cased) that handle a custom number of languages.
22
+
23
+ Our versions give exactly the same representations produced by the original model which preserves the original accuracy.
24
+
25
+ Handled languages: en, fr, es, de, zh, ar, ru, vi, el, bg, th, tr, hi, ur, sw, nl, uk, ro, pt, it, lt, no, pl, da and ja.
26
+
27
+ For more information please visit our paper: [Load What You Need: Smaller Versions of Multilingual BERT](https://www.aclweb.org/anthology/2020.sustainlp-1.16.pdf).
28
+
29
+ ## How to use
30
+
31
+ ```python
32
+ from transformers import AutoTokenizer, AutoModel
33
+
34
+ tokenizer = AutoTokenizer.from_pretrained("Geotrend/distilbert-base-25lang-cased")
35
+ model = AutoModel.from_pretrained("Geotrend/distilbert-base-25lang-cased")
36
+
37
+ ```
38
+
39
+ To generate other smaller versions of multilingual transformers please visit [our Github repo](https://github.com/Geotrend-research/smaller-transformers).
40
+
41
+ ### How to cite
42
+
43
+ ```bibtex
44
+ @inproceedings{smallermbert,
45
+ title={Load What You Need: Smaller Versions of Multilingual BERT},
46
+ author={Abdaoui, Amine and Pradel, Camille and Sigel, Grégoire},
47
+ booktitle={SustaiNLP / EMNLP},
48
+ year={2020}
49
+ }
50
+ ```
51
+
52
+ ## Contact
53
+
54
+ Please contact amine@geotrend.fr for any question, feedback or request.
config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-multilingual-cased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForMaskedLM"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "initializer_range": 0.02,
12
+ "max_position_embeddings": 512,
13
+ "model_type": "distilbert",
14
+ "n_heads": 12,
15
+ "n_layers": 6,
16
+ "output_past": true,
17
+ "pad_token_id": 0,
18
+ "qa_dropout": 0.1,
19
+ "seq_classif_dropout": 0.2,
20
+ "sinusoidal_pos_embds": false,
21
+ "tie_weights_": true,
22
+ "transformers_version": "4.4.1",
23
+ "vocab_size": 84985
24
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23959580a1928f0d2c1d16cebd5430476679dc05079d05a665aa909e82f92516
3
+ size 435511219
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
1
+ {"do_lower_case": false, "model_max_length": 512}
vocab.txt ADDED
The diff for this file is too large to render. See raw diff