1 {
2 "architectures": [
3 "BertForTokenClassification"
4 ],
5 "attention_probs_dropout_prob": 0.1,
6 "classifier_dropout": null,
7 "directionality": "bidi",
8 "finetuning_task": "ner",
9 "gradient_checkpointing": false,
10 "hidden_act": "gelu",
11 "hidden_dropout_prob": 0.1,
12 "hidden_size": 768,
13 "id2label": {
14 "0": "ADJ",
15 "1": "ADP",
16 "2": "ADV",
17 "3": "AUX",
18 "4": "B-ADJ",
19 "5": "B-ADP",
20 "6": "B-ADV",
21 "7": "B-AUX",
22 "8": "B-CCONJ",
23 "9": "B-DET",
24 "10": "B-NOUN",
25 "11": "B-NUM",
26 "12": "B-PART",
27 "13": "B-PRON",
28 "14": "B-PROPN",
29 "15": "B-PUNCT",
30 "16": "B-SYM",
31 "17": "B-VERB",
32 "18": "B-X",
33 "19": "CCONJ",
34 "20": "DET",
35 "21": "I-ADJ",
36 "22": "I-ADP",
37 "23": "I-ADV",
38 "24": "I-AUX",
39 "25": "I-CCONJ",
40 "26": "I-DET",
41 "27": "I-NOUN",
42 "28": "I-NUM",
43 "29": "I-PART",
44 "30": "I-PRON",
45 "31": "I-PROPN",
46 "32": "I-PUNCT",
47 "33": "I-SYM",
48 "34": "I-VERB",
49 "35": "I-X",
50 "36": "NOUN",
51 "37": "NUM",
52 "38": "PART",
53 "39": "PRON",
54 "40": "PROPN",
55 "41": "PUNCT",
56 "42": "SYM",
57 "43": "VERB",
58 "44": "X"
59 },
60 "initializer_range": 0.02,
61 "intermediate_size": 3072,
62 "label2id": {
63 "ADJ": 0,
64 "ADP": 1,
65 "ADV": 2,
66 "AUX": 3,
67 "B-ADJ": 4,
68 "B-ADP": 5,
69 "B-ADV": 6,
70 "B-AUX": 7,
71 "B-CCONJ": 8,
72 "B-DET": 9,
73 "B-NOUN": 10,
74 "B-NUM": 11,
75 "B-PART": 12,
76 "B-PRON": 13,
77 "B-PROPN": 14,
78 "B-PUNCT": 15,
79 "B-SYM": 16,
80 "B-VERB": 17,
81 "B-X": 18,
82 "CCONJ": 19,
83 "DET": 20,
84 "I-ADJ": 21,
85 "I-ADP": 22,
86 "I-ADV": 23,
87 "I-AUX": 24,
88 "I-CCONJ": 25,
89 "I-DET": 26,
90 "I-NOUN": 27,
91 "I-NUM": 28,
92 "I-PART": 29,
93 "I-PRON": 30,
94 "I-PROPN": 31,
95 "I-PUNCT": 32,
96 "I-SYM": 33,
97 "I-VERB": 34,
98 "I-X": 35,
99 "NOUN": 36,
100 "NUM": 37,
101 "PART": 38,
102 "PRON": 39,
103 "PROPN": 40,
104 "PUNCT": 41,
105 "SYM": 42,
106 "VERB": 43,
107 "X": 44
108 },
109 "layer_norm_eps": 1e-12,
110 "max_position_embeddings": 512,
111 "model_type": "bert",
112 "num_attention_heads": 12,
113 "num_hidden_layers": 12,
114 "output_past": true,
115 "pad_token_id": 0,
116 "pooler_fc_size": 768,
117 "pooler_num_attention_heads": 12,
118 "pooler_num_fc_layers": 3,
119 "pooler_size_per_head": 128,
120 "pooler_type": "first_token_transform",
121 "position_embedding_type": "absolute",
122 "torch_dtype": "float32",
123 "transformers_version": "4.10.0",
124 "type_vocab_size": 2,
125 "use_cache": true,
126 "vocab_size": 21128
127 }
128