{ "architectures": [ "BertForTokenClassification" ], "attention_probs_dropout_prob": 0.1, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "_", "1": "2", "10": "Lni", "11": "N", "12": "p", "13": "P", "14": "S", "15": "T", "16": "V", "17": "X", "18": "Y", "19": "Z", "2": "A", "3": "D", "4": "G", "5": "I", "6": "J", "7": "K", "8": "ks", "9": "L0" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "2": 1, "A": 2, "D": 3, "G": 4, "I": 5, "J": 6, "K": 7, "L0": 9, "Lni": 10, "N": 11, "P": 13, "S": 14, "T": 15, "V": 16, "X": 17, "Y": 18, "Z": 19, "_": 0, "ks": 8, "p": 12 }, "layer_norm_eps": 1e-12, "max_position_embeddings": 512, "model_type": "bert", "num_attention_heads": 12, "num_hidden_layers": 12, "pad_token_id": 0, "type_vocab_size": 2, "vocab_size": 50000 }