asahi417 commited on
Commit
a8e263d
1 Parent(s): 807f0c7
config.json ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "xlm-roberta-large",
3
+ "architectures": [
4
+ "XLMRobertaForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "eos_token_id": 2,
9
+ "gradient_checkpointing": false,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 1024,
13
+ "id2label": {
14
+ "0": "O",
15
+ "1": "B-cardinal number",
16
+ "2": "B-date",
17
+ "3": "I-date",
18
+ "4": "B-person",
19
+ "5": "I-person",
20
+ "6": "B-group",
21
+ "7": "B-geopolitical area",
22
+ "8": "I-geopolitical area",
23
+ "9": "B-law",
24
+ "10": "I-law",
25
+ "11": "B-organization",
26
+ "12": "I-organization",
27
+ "13": "B-percent",
28
+ "14": "I-percent",
29
+ "15": "B-ordinal number",
30
+ "16": "B-money",
31
+ "17": "I-money",
32
+ "18": "B-work of art",
33
+ "19": "I-work of art",
34
+ "20": "B-facility",
35
+ "21": "B-time",
36
+ "22": "I-cardinal number",
37
+ "23": "B-location",
38
+ "24": "B-quantity",
39
+ "25": "I-quantity",
40
+ "26": "I-group",
41
+ "27": "I-location",
42
+ "28": "B-product",
43
+ "29": "I-time",
44
+ "30": "B-event",
45
+ "31": "I-event",
46
+ "32": "I-facility",
47
+ "33": "B-language",
48
+ "34": "I-product",
49
+ "35": "I-ordinal number",
50
+ "36": "I-language"
51
+ },
52
+ "initializer_range": 0.02,
53
+ "intermediate_size": 4096,
54
+ "label2id": {
55
+ "B-cardinal number": 1,
56
+ "B-date": 2,
57
+ "B-event": 30,
58
+ "B-facility": 20,
59
+ "B-geopolitical area": 7,
60
+ "B-group": 6,
61
+ "B-language": 33,
62
+ "B-law": 9,
63
+ "B-location": 23,
64
+ "B-money": 16,
65
+ "B-ordinal number": 15,
66
+ "B-organization": 11,
67
+ "B-percent": 13,
68
+ "B-person": 4,
69
+ "B-product": 28,
70
+ "B-quantity": 24,
71
+ "B-time": 21,
72
+ "B-work of art": 18,
73
+ "I-cardinal number": 22,
74
+ "I-date": 3,
75
+ "I-event": 31,
76
+ "I-facility": 32,
77
+ "I-geopolitical area": 8,
78
+ "I-group": 26,
79
+ "I-language": 36,
80
+ "I-law": 10,
81
+ "I-location": 27,
82
+ "I-money": 17,
83
+ "I-ordinal number": 35,
84
+ "I-organization": 12,
85
+ "I-percent": 14,
86
+ "I-person": 5,
87
+ "I-product": 34,
88
+ "I-quantity": 25,
89
+ "I-time": 29,
90
+ "I-work of art": 19,
91
+ "O": 0
92
+ },
93
+ "layer_norm_eps": 1e-05,
94
+ "max_position_embeddings": 514,
95
+ "model_type": "xlm-roberta",
96
+ "num_attention_heads": 16,
97
+ "num_hidden_layers": 24,
98
+ "output_past": true,
99
+ "pad_token_id": 1,
100
+ "type_vocab_size": 1,
101
+ "vocab_size": 250002
102
+ }
parameter.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dataset": ["ontonotes5"], "transformers_model": "xlm-roberta-large", "random_seed": 1234, "lr": 1e-05, "total_step": 5000, "warmup_step": 700, "weight_decay": 1e-07, "batch_size": 32, "max_seq_length": 128, "fp16": false, "max_grad_norm": 1.0, "lower_case": true}
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d05059e0deacdda29518a446f3bbad28b5dd6cba754ee3c63525759be28abe01
3
+ size 2235678522
sentencepiece.bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfc8146abe2a0488e9e2a0c56de7952f7c11ab059eca145a0a727afce0db2865
3
+ size 5069051
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": "<mask>"}
test_bc5cdr_span_lower.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"valid": {"f1": 0.0, "recall": 0.0, "precision": 0.0, "summary": ""}, "test": {"f1": 0.0, "recall": 0.0, "precision": 0.0, "summary": ""}}
test_bionlp2004_span_lower.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"valid": {"f1": 0.0, "recall": 0.0, "precision": 0.0, "summary": ""}}
test_conll2003_span_lower.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"valid": {"f1": 62.173140390962175, "recall": 49.00940564338603, "precision": 85.00520652551198, "summary": " precision recall f1-score support\n\n entity 0.85 0.49 0.62 4997\n\n micro avg 0.85 0.49 0.62 4997\n macro avg 0.85 0.49 0.62 4997\nweighted avg 0.85 0.49 0.62 4997\n"}, "test": {"f1": 59.989819292440814, "recall": 47.848152659358504, "precision": 80.38881309686221, "summary": " precision recall f1-score support\n\n entity 0.80 0.48 0.60 4926\n\n micro avg 0.80 0.48 0.60 4926\n macro avg 0.80 0.48 0.60 4926\nweighted avg 0.80 0.48 0.60 4926\n"}}
test_fin_span_lower.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"valid": {"f1": 15.151515151515154, "recall": 9.84251968503937, "precision": 32.89473684210527, "summary": " precision recall f1-score support\n\n entity 0.33 0.10 0.15 254\n\n micro avg 0.33 0.10 0.15 254\n macro avg 0.33 0.10 0.15 254\nweighted avg 0.33 0.10 0.15 254\n"}}
test_mit_movie_trivia_span_lower.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"valid": {"f1": 88.60759493670885, "recall": 95.31013615733737, "precision": 82.78580814717476, "summary": " precision recall f1-score support\n\n entity 0.83 0.95 0.89 661\n\n micro avg 0.83 0.95 0.89 661\n macro avg 0.83 0.95 0.89 661\nweighted avg 0.83 0.95 0.89 661\n"}}
test_mit_restaurant_span_lower.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"valid": {"f1": 4.585798816568047, "recall": 2.594142259414226, "precision": 19.745222929936308, "summary": " precision recall f1-score support\n\n entity 0.20 0.03 0.05 1195\n\n micro avg 0.20 0.03 0.05 1195\n macro avg 0.20 0.03 0.05 1195\nweighted avg 0.20 0.03 0.05 1195\n"}}
test_ontonotes5_lower.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"valid": {"f1": 85.86261980830672, "recall": 87.65174850516398, "precision": 84.14506870760133, "summary": " precision recall f1-score support\n\n cardinal number 0.83 0.88 0.85 937\n date 0.81 0.87 0.84 1507\n event 0.65 0.52 0.58 143\n facility 0.42 0.53 0.47 115\ngeopolitical area 0.92 0.94 0.93 2261\n group 0.88 0.92 0.90 847\n language 0.85 0.88 0.87 33\n law 0.47 0.70 0.56 40\n location 0.61 0.67 0.64 203\n money 0.86 0.91 0.88 274\n ordinal number 0.84 0.87 0.85 232\n organization 0.82 0.82 0.82 1728\n percent 0.91 0.90 0.91 177\n person 0.93 0.95 0.94 2014\n product 0.52 0.61 0.56 72\n quantity 0.77 0.80 0.79 99\n time 0.68 0.77 0.72 214\n work of art 0.40 0.50 0.45 142\n\n micro avg 0.84 0.88 0.86 11038\n macro avg 0.73 0.78 0.75 11038\n weighted avg 0.84 0.88 0.86 11038\n"}, "test": {"f1": 87.02697967764541, "recall": 88.30326193227269, "precision": 85.78706502029185, "summary": " precision recall f1-score support\n\n cardinal number 0.84 0.86 0.85 934\n date 0.83 0.89 0.86 1601\n event 0.58 0.60 0.59 63\n facility 0.59 0.61 0.60 135\ngeopolitical area 0.95 0.95 0.95 2240\n group 0.84 0.90 0.87 841\n language 0.75 0.68 0.71 22\n law 0.50 0.62 0.56 40\n location 0.60 0.63 0.61 179\n money 0.87 0.91 0.89 314\n ordinal number 0.80 0.92 0.86 195\n organization 0.85 0.86 0.85 1792\n percent 0.87 0.89 0.88 348\n person 0.94 0.94 0.94 1988\n product 0.71 0.71 0.71 76\n quantity 0.72 0.79 0.75 105\n time 0.61 0.66 0.63 212\n work of art 0.52 0.54 0.53 166\n\n micro avg 0.86 0.88 0.87 11251\n macro avg 0.74 0.78 0.76 11251\n weighted avg 0.86 0.88 0.87 11251\n"}}
test_ontonotes5_span_lower.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"valid": {"f1": 89.1553376420771, "recall": 90.6051820982062, "precision": 87.7511625866456, "summary": " precision recall f1-score support\n\n entity 0.88 0.91 0.89 11038\n\n micro avg 0.88 0.91 0.89 11038\n macro avg 0.88 0.91 0.89 11038\nweighted avg 0.88 0.91 0.89 11038\n"}, "test": {"f1": 89.36469553748077, "recall": 90.32974846680295, "precision": 88.42004524099531, "summary": " precision recall f1-score support\n\n entity 0.88 0.90 0.89 11251\n\n micro avg 0.88 0.90 0.89 11251\n macro avg 0.88 0.90 0.89 11251\nweighted avg 0.88 0.90 0.89 11251\n"}}
test_panx_dataset-en_span_lower.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"valid": {"f1": 45.25706010137582, "recall": 35.44088460447973, "precision": 62.59389083625438, "summary": " precision recall f1-score support\n\n entity 0.63 0.35 0.45 14108\n\n micro avg 0.63 0.35 0.45 14108\n macro avg 0.63 0.35 0.45 14108\nweighted avg 0.63 0.35 0.45 14108\n"}, "test": {"f1": 44.75627092218096, "recall": 35.15090398328892, "precision": 61.58505805148915, "summary": " precision recall f1-score support\n\n entity 0.62 0.35 0.45 13883\n\n micro avg 0.62 0.35 0.45 13883\n macro avg 0.62 0.35 0.45 13883\nweighted avg 0.62 0.35 0.45 13883\n"}}
test_wnut2017_span_lower.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"valid": {"f1": 66.30944407233757, "recall": 61.720698254364095, "precision": 71.63531114327061, "summary": " precision recall f1-score support\n\n entity 0.72 0.62 0.66 802\n\n micro avg 0.72 0.62 0.66 802\n macro avg 0.72 0.62 0.66 802\nweighted avg 0.72 0.62 0.66 802\n"}, "test": {"f1": 50.140845070422536, "recall": 44.01582591493571, "precision": 58.24607329842932, "summary": " precision recall f1-score support\n\n entity 0.58 0.44 0.50 1011\n\n micro avg 0.58 0.44 0.50 1011\n macro avg 0.58 0.44 0.50 1011\nweighted avg 0.58 0.44 0.50 1011\n"}}
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "sep_token": "</s>", "cls_token": "<s>", "unk_token": "<unk>", "pad_token": "<pad>", "mask_token": "<mask>", "model_max_length": 512, "name_or_path": "xlm-roberta-large"}