asahi417 commited on
Commit
73756a0
1 Parent(s): f57e1b5
Files changed (2) hide show
  1. config.json +159 -0
  2. pytorch_model.bin +3 -0
config.json ADDED
@@ -0,0 +1,159 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "tner_ckpt/bc5cdr_roberta_large/best_model",
3
+ "architectures": [
4
+ "RobertaForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "crf_state_dict": {
10
+ "_constraint_mask": [
11
+ [
12
+ 1.0,
13
+ 1.0,
14
+ 1.0,
15
+ 0.0,
16
+ 0.0,
17
+ 0.0,
18
+ 1.0
19
+ ],
20
+ [
21
+ 1.0,
22
+ 1.0,
23
+ 1.0,
24
+ 0.0,
25
+ 1.0,
26
+ 0.0,
27
+ 1.0
28
+ ],
29
+ [
30
+ 1.0,
31
+ 1.0,
32
+ 1.0,
33
+ 1.0,
34
+ 0.0,
35
+ 0.0,
36
+ 1.0
37
+ ],
38
+ [
39
+ 1.0,
40
+ 1.0,
41
+ 1.0,
42
+ 1.0,
43
+ 0.0,
44
+ 0.0,
45
+ 1.0
46
+ ],
47
+ [
48
+ 1.0,
49
+ 1.0,
50
+ 1.0,
51
+ 0.0,
52
+ 1.0,
53
+ 0.0,
54
+ 1.0
55
+ ],
56
+ [
57
+ 1.0,
58
+ 1.0,
59
+ 1.0,
60
+ 0.0,
61
+ 0.0,
62
+ 0.0,
63
+ 0.0
64
+ ],
65
+ [
66
+ 0.0,
67
+ 0.0,
68
+ 0.0,
69
+ 0.0,
70
+ 0.0,
71
+ 0.0,
72
+ 0.0
73
+ ]
74
+ ],
75
+ "end_transitions": [
76
+ 1.1829675436019897,
77
+ 0.338004469871521,
78
+ 1.610476016998291,
79
+ -2.1025612354278564,
80
+ 0.026044141501188278
81
+ ],
82
+ "start_transitions": [
83
+ 0.23535622656345367,
84
+ -0.5871618390083313,
85
+ -0.908195972442627,
86
+ 1.289412021636963,
87
+ 0.7491621971130371
88
+ ],
89
+ "transitions": [
90
+ [
91
+ 0.08883310854434967,
92
+ -0.34687548875808716,
93
+ -0.5557966232299805,
94
+ -0.6514758467674255,
95
+ -0.32734575867652893
96
+ ],
97
+ [
98
+ -0.020529085770249367,
99
+ -1.1056996583938599,
100
+ -0.6824190020561218,
101
+ 0.38456156849861145,
102
+ -0.19291424751281738
103
+ ],
104
+ [
105
+ -0.6466730237007141,
106
+ -0.2793288826942444,
107
+ -0.8555130958557129,
108
+ 0.10881402343511581,
109
+ 0.8075306415557861
110
+ ],
111
+ [
112
+ -0.09906835854053497,
113
+ -0.14599257707595825,
114
+ 0.10467313230037689,
115
+ -0.26882439851760864,
116
+ -0.12745390832424164
117
+ ],
118
+ [
119
+ 0.11021356284618378,
120
+ -0.1971757709980011,
121
+ 0.632551372051239,
122
+ 0.9238606095314026,
123
+ -0.20630259811878204
124
+ ]
125
+ ]
126
+ },
127
+ "eos_token_id": 2,
128
+ "hidden_act": "gelu",
129
+ "hidden_dropout_prob": 0.1,
130
+ "hidden_size": 1024,
131
+ "id2label": {
132
+ "0": "O",
133
+ "1": "B-Chemical",
134
+ "2": "B-Disease",
135
+ "3": "I-Disease",
136
+ "4": "I-Chemical"
137
+ },
138
+ "initializer_range": 0.02,
139
+ "intermediate_size": 4096,
140
+ "label2id": {
141
+ "B-Chemical": 1,
142
+ "B-Disease": 2,
143
+ "I-Chemical": 4,
144
+ "I-Disease": 3,
145
+ "O": 0
146
+ },
147
+ "layer_norm_eps": 1e-05,
148
+ "max_position_embeddings": 514,
149
+ "model_type": "roberta",
150
+ "num_attention_heads": 16,
151
+ "num_hidden_layers": 24,
152
+ "pad_token_id": 1,
153
+ "position_embedding_type": "absolute",
154
+ "torch_dtype": "float32",
155
+ "transformers_version": "4.20.1",
156
+ "type_vocab_size": 1,
157
+ "use_cache": true,
158
+ "vocab_size": 50265
159
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99a9fd565e76d1775117d515580168d79cac62b61740feb840c1211233a02b33
3
+ size 1417392177