{ "attention_probs_dropout_prob": 0.1, "finetuning_task": "ner", "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "EVN", "1": "LOC", "2": "LOC/LOC", "3": "LOC/ORG", "4": "LOC/PRS", "5": "MSR", "6": "O", "7": "OBJ", "8": "OBJ/ORG", "9": "ORG", "10": "ORG/PRS", "11": "PRS", "12": "PRS/WRK", "13": "TME", "14": "WRK" }, "initializer_range": 0.02, "intermediate_size": 3072, "is_decoder": false, "label2id": { "EVN": 0, "LOC": 1, "LOC/LOC": 2, "LOC/ORG": 3, "LOC/PRS": 4, "MSR": 5, "O": 6, "OBJ": 7, "OBJ/ORG": 8, "ORG": 9, "ORG/PRS": 10, "PRS": 11, "PRS/WRK": 12, "TME": 13, "WRK": 14 }, "layer_norm_eps": 1e-12, "max_position_embeddings": 512, "num_attention_heads": 12, "num_hidden_layers": 12, "num_labels": 15, "output_attentions": false, "output_hidden_states": false, "output_past": true, "pruned_heads": {}, "torchscript": false, "type_vocab_size": 2, "use_bfloat16": false, "vocab_size": 50002 }