{ "_name_or_path": "PlanTL-GOB-ES/roberta-large-bne", "architectures": [ "RobertaForTokenClassification" ], "attention_probs_dropout_prob": 0.0, "bos_token_id": 0, "classifier_dropout": null, "eos_token_id": 2, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.0, "hidden_size": 1024, "id2label": { "0": "Z", "1": "vas", "2": "pe", "3": "Faa", "4": "pd", "5": "vsn", "6": "Fp", "7": "Fit", "8": "vsm", "9": "dp", "10": "Fd", "11": "I", "12": "de", "13": "vam", "14": "vmm", "15": "vsg", "16": "dt", "17": "Fat", "18": "Fh", "19": "W", "20": "vmi", "21": "Y", "22": "dn", "23": "da", "24": "Zm", "25": "X", "26": "vmn", "27": "dd", "28": "vag", "29": "Fx", "30": "rn", "31": "pr", "32": "ao", "33": "Fz", "34": "Fs", "35": "di", "36": "pn", "37": "Fe", "38": "vmp", "39": "sp", "40": "aq", "41": "pt", "42": "pi", "43": "vms", "44": "Fg", "45": "Fpt", "46": "px", "47": "Zp", "48": "np", "49": "van", "50": "rg", "51": "Fia", "52": "p0", "53": "Fc", "54": "Fpa", "55": "vsp", "56": "vap", "57": "cc", "58": "cs", "59": "vss", "60": "vmg", "61": "vai", "62": "nc", "63": "pp", "64": "i", "65": "vsi" }, "initializer_range": 0.02, "intermediate_size": 4096, "label2id": { "Faa": 3, "Fat": 17, "Fc": 53, "Fd": 10, "Fe": 37, "Fg": 44, "Fh": 18, "Fia": 51, "Fit": 7, "Fp": 6, "Fpa": 54, "Fpt": 45, "Fs": 34, "Fx": 29, "Fz": 33, "I": 11, "W": 19, "X": 25, "Y": 21, "Z": 0, "Zm": 24, "Zp": 47, "ao": 32, "aq": 40, "cc": 57, "cs": 58, "da": 23, "dd": 27, "de": 12, "di": 35, "dn": 22, "dp": 9, "dt": 16, "i": 64, "nc": 62, "np": 48, "p0": 52, "pd": 4, "pe": 2, "pi": 42, "pn": 36, "pp": 63, "pr": 31, "pt": 41, "px": 46, "rg": 50, "rn": 30, "sp": 39, "vag": 28, "vai": 61, "vam": 13, "van": 49, "vap": 56, "vas": 1, "vmg": 60, "vmi": 20, "vmm": 14, "vmn": 26, "vmp": 38, "vms": 43, "vsg": 15, "vsi": 65, "vsm": 8, "vsn": 5, "vsp": 55, "vss": 59 }, "layer_norm_eps": 1e-05, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 16, "num_hidden_layers": 24, "pad_token_id": 1, "position_embedding_type": "absolute", "torch_dtype": "float32", "transformers_version": "4.44.2", "type_vocab_size": 1, "use_cache": true, "vocab_size": 50262 }