ryo0634 commited on
Commit
f1b6f9f
1 Parent(s): 3aa774a

initial commit

Browse files
.gitattributes CHANGED
@@ -25,3 +25,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
  *.zstandard filter=lfs diff=lfs merge=lfs -text
27
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
  *.zstandard filter=lfs diff=lfs merge=lfs -text
27
  *tfevents* filter=lfs diff=lfs merge=lfs -text
28
+ pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
29
+ entity_vocab.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: multilingual
3
+ thumbnail: https://github.com/studio-ousia/luke/raw/master/resources/luke_logo.png
4
+ tags:
5
+ - luke
6
+ - named entity recognition
7
+ - relation classification
8
+ - question answering
9
+ license: apache-2.0
10
+ ---
11
+
12
+ ## mLUKE
13
+
14
+ **mLUKE** (multilingual LUKE) is a multilingual extension of LUKE.
15
+
16
+ Please check the [official repository](https://github.com/studio-ousia/luke) for
17
+ more details and updates.
18
+
19
+ This is the mLUKE base model with 12 hidden layers, 768 hidden size. The total number
20
+ of parameters in this model is 585M (278M for the word embeddings and encoder, 307M for the entity embeddings).
21
+ The model was initialized with the weights of XLM-RoBERTa(base) and trained using December 2020 version of Wikipedia in 24 languages.
22
+
23
+ ### Citation
24
+
25
+ If you find mLUKE useful for your work, please cite the following paper:
26
+
27
+ ```latex
28
+ @inproceedings{ri2021mluke,
29
+ title={mLUKE: The Power of Entity Representations in Multilingual Pretrained Language Models},
30
+ author={Ryokan Ri, Ikuya Yamada, Yoshimasa Tsuruoka},
31
+ booktitle={arXiv},
32
+ year={2021}
33
+ }
34
+ ```
added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"<ent>": 250002, "<ent2>": 250003}
config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LukeForMaskedLM"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bert_model_name": "xlm-roberta-base",
7
+ "bos_token_id": 0,
8
+ "cls_entity_prediction": false,
9
+ "deepspeed_transformer_layer_args": {},
10
+ "entity_emb_size": 256,
11
+ "entity_vocab_size": 1200001,
12
+ "eos_token_id": 2,
13
+ "gradient_checkpointing": false,
14
+ "hidden_act": "gelu",
15
+ "hidden_dropout_prob": 0.1,
16
+ "hidden_size": 768,
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 3072,
19
+ "layer_norm_eps": 1e-05,
20
+ "max_position_embeddings": 514,
21
+ "model_type": "luke",
22
+ "num_attention_heads": 12,
23
+ "num_hidden_layers": 12,
24
+ "output_past": true,
25
+ "pad_token_id": 1,
26
+ "position_embedding_type": "absolute",
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.13.0.dev0",
29
+ "type_vocab_size": 1,
30
+ "use_cache": true,
31
+ "use_deepspeed_transformer_layer": false,
32
+ "use_entity_aware_attention": true,
33
+ "vocab_size": 250004
34
+ }
entity_vocab.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc661ccd50c8856fec1c1937275801ac148dadea1835bd15ba37037853b73cc5
3
+ size 438520875
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56588c40f41cff82e5174f1cd790bf59151f82c434407fd103d6ede9addeee5a
3
+ size 2437467971
sentencepiece.bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfc8146abe2a0488e9e2a0c56de7952f7c11ab059eca145a0a727afce0db2865
3
+ size 5069051
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true}, "additional_special_tokens": [{"content": "<ent>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<ent2>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "cls_token": "<s>", "pad_token": "<pad>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "sp_model_kwargs": {}, "model_max_length": 512, "special_tokens_map_file": null, "tokenizer_file": "/Users/linghan/.cache/huggingface/transformers/daeda8d936162ca65fe6dd158ecce1d8cb56c17d89b78ab86be1558eaef1d76a.a984cf52fc87644bd4a2165f1e07e0ac880272c1e82d648b4674907056912bd7", "name_or_path": "xlm-roberta-base", "tokenizer_class": "MLukeTokenizer"}