{ "_num_labels": 61, "architectures": [ "BertForTokenClassification" ], "attention_probs_dropout_prob": 0.1, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "O", "1": "AO", "2": "AQ", "3": "CC", "4": "CS", "5": "DA", "6": "DD", "7": "DE", "8": "DI", "9": "DN", "10": "DP", "11": "DT", "12": "Faa", "13": "Fat", "14": "Fc", "15": "Fd", "16": "Fe", "17": "Fg", "18": "Fh", "19": "Fia", "20": "Fit", "21": "Fp", "22": "Fpa", "23": "Fpt", "24": "Fs", "25": "Ft", "26": "Fx", "27": "Fz", "28": "I", "29": "NC", "30": "NP", "31": "P0", "32": "PD", "33": "PI", "34": "PN", "35": "PP", "36": "PR", "37": "PT", "38": "PX", "39": "RG", "40": "RN", "41": "SP", "42": "VAI", "43": "VAM", "44": "VAN", "45": "VAP", "46": "VAS", "47": "VMG", "48": "VMI", "49": "VMM", "50": "VMN", "51": "VMP", "52": "VMS", "53": "VSG", "54": "VSI", "55": "VSM", "56": "VSN", "57": "VSP", "58": "VSS", "59": "Y", "60": "Z" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "AO": 1, "AQ": 2, "CC": 3, "CS": 4, "DA": 5, "DD": 6, "DE": 7, "DI": 8, "DN": 9, "DP": 10, "DT": 11, "Faa": 12, "Fat": 13, "Fc": 14, "Fd": 15, "Fe": 16, "Fg": 17, "Fh": 18, "Fia": 19, "Fit": 20, "Fp": 21, "Fpa": 22, "Fpt": 23, "Fs": 24, "Ft": 25, "Fx": 26, "Fz": 27, "I": 28, "NC": 29, "NP": 30, "O": 0, "P0": 31, "PD": 32, "PI": 33, "PN": 34, "PP": 35, "PR": 36, "PT": 37, "PX": 38, "RG": 39, "RN": 40, "SP": 41, "VAI": 42, "VAM": 43, "VAN": 44, "VAP": 45, "VAS": 46, "VMG": 47, "VMI": 48, "VMM": 49, "VMN": 50, "VMP": 51, "VMS": 52, "VSG": 53, "VSI": 54, "VSM": 55, "VSN": 56, "VSP": 57, "VSS": 58, "Y": 59, "Z": 60 }, "layer_norm_eps": 1e-12, "max_position_embeddings": 512, "model_type": "bert", "num_attention_heads": 12, "num_hidden_layers": 12, "output_past": true, "pad_token_id": 1, "type_vocab_size": 2, "vocab_size": 31002 }