{ "_name_or_path": "models/CNEC1_1_nested_xlm-roberta-large", "architectures": [ "XLMRobertaForTokenClassification" ], "attention_probs_dropout_prob": 0.15, "bos_token_id": 0, "classifier_dropout": null, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.15, "hidden_size": 1024, "id2label": { "0": "O", "1": "B-P", "2": "I-P", "3": "B-T", "4": "I-T", "5": "B-A", "6": "I-A", "7": "B-C", "8": "I-C", "9": "B-ah", "10": "I-ah", "11": "B-at", "12": "I-at", "13": "B-az", "14": "I-az", "15": "B-g_", "16": "I-g_", "17": "B-gc", "18": "I-gc", "19": "B-gh", "20": "I-gh", "21": "B-gl", "22": "I-gl", "23": "B-gp", "24": "I-gp", "25": "B-gq", "26": "I-gq", "27": "B-gr", "28": "I-gr", "29": "B-gs", "30": "I-gs", "31": "B-gt", "32": "I-gt", "33": "B-gu", "34": "I-gu", "35": "B-i_", "36": "I-i_", "37": "B-ia", "38": "I-ia", "39": "B-ic", "40": "I-ic", "41": "B-if", "42": "I-if", "43": "B-io", "44": "I-io", "45": "B-mn", "46": "I-mn", "47": "B-mt", "48": "I-mt", "49": "B-mr", "50": "I-mr", "51": "B-o_", "52": "I-o_", "53": "B-oa", "54": "I-oa", "55": "B-oc", "56": "I-oc", "57": "B-oe", "58": "I-oe", "59": "B-om", "60": "I-om", "61": "B-op", "62": "I-op", "63": "B-or", "64": "I-or", "65": "B-p_", "66": "I-p_", "67": "B-pb", "68": "I-pb", "69": "B-pc", "70": "I-pc", "71": "B-pd", "72": "I-pd", "73": "B-pf", "74": "I-pf", "75": "B-pm", "76": "I-pm", "77": "B-pp", "78": "I-pp", "79": "B-ps", "80": "I-ps", "81": "B-td", "82": "I-td", "83": "B-tf", "84": "I-tf", "85": "B-th", "86": "I-th", "87": "B-ti", "88": "I-ti", "89": "B-tm", "90": "I-tm", "91": "B-ty", "92": "I-ty" }, "initializer_range": 0.02, "intermediate_size": 4096, "label2id": { "B-A": 5, "B-C": 7, "B-P": 1, "B-T": 3, "B-ah": 9, "B-at": 11, "B-az": 13, "B-g_": 15, "B-gc": 17, "B-gh": 19, "B-gl": 21, "B-gp": 23, "B-gq": 25, "B-gr": 27, "B-gs": 29, "B-gt": 31, "B-gu": 33, "B-i_": 35, "B-ia": 37, "B-ic": 39, "B-if": 41, "B-io": 43, "B-mn": 45, "B-mr": 49, "B-mt": 47, "B-o_": 51, "B-oa": 53, "B-oc": 55, "B-oe": 57, "B-om": 59, "B-op": 61, "B-or": 63, "B-p_": 65, "B-pb": 67, "B-pc": 69, "B-pd": 71, "B-pf": 73, "B-pm": 75, "B-pp": 77, "B-ps": 79, "B-td": 81, "B-tf": 83, "B-th": 85, "B-ti": 87, "B-tm": 89, "B-ty": 91, "I-A": 6, "I-C": 8, "I-P": 2, "I-T": 4, "I-ah": 10, "I-at": 12, "I-az": 14, "I-g_": 16, "I-gc": 18, "I-gh": 20, "I-gl": 22, "I-gp": 24, "I-gq": 26, "I-gr": 28, "I-gs": 30, "I-gt": 32, "I-gu": 34, "I-i_": 36, "I-ia": 38, "I-ic": 40, "I-if": 42, "I-io": 44, "I-mn": 46, "I-mr": 50, "I-mt": 48, "I-o_": 52, "I-oa": 54, "I-oc": 56, "I-oe": 58, "I-om": 60, "I-op": 62, "I-or": 64, "I-p_": 66, "I-pb": 68, "I-pc": 70, "I-pd": 72, "I-pf": 74, "I-pm": 76, "I-pp": 78, "I-ps": 80, "I-td": 82, "I-tf": 84, "I-th": 86, "I-ti": 88, "I-tm": 90, "I-ty": 92, "O": 0 }, "layer_norm_eps": 1e-05, "max_position_embeddings": 514, "model_type": "xlm-roberta", "num_attention_heads": 16, "num_hidden_layers": 24, "output_past": true, "pad_token_id": 1, "position_embedding_type": "absolute", "torch_dtype": "float32", "transformers_version": "4.36.2", "type_vocab_size": 1, "use_cache": true, "vocab_size": 250002 }