{ "_name_or_path": "PlanTL-GOB-ES/roberta-large-bne", "architectures": [ "RobertaForTokenClassification" ], "attention_probs_dropout_prob": 0.0, "bos_token_id": 0, "classifier_dropout": null, "eos_token_id": 2, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.0, "hidden_size": 1024, "id2label": { "0": "i", "1": "Zp", "2": "rg", "3": "vmm", "4": "Fs", "5": "vsi", "6": "dd", "7": "vms", "8": "Fh", "9": "Fe", "10": "rn", "11": "aq", "12": "vag", "13": "Z", "14": "W", "15": "p0", "16": "cc", "17": "da", "18": "di", "19": "X", "20": "vsn", "21": "np", "22": "dn", "23": "Fg", "24": "Fat", "25": "Fx", "26": "I", "27": "vap", "28": "Fd", "29": "van", "30": "vmg", "31": "vsm", "32": "Fc", "33": "sp", "34": "Fit", "35": "pd", "36": "px", "37": "nc", "38": "vai", "39": "de", "40": "Fia", "41": "pe", "42": "dp", "43": "vsp", "44": "pp", "45": "dt", "46": "ao", "47": "Fp", "48": "Fpt", "49": "Faa", "50": "pn", "51": "pr", "52": "vss", "53": "vam", "54": "vmn", "55": "Zm", "56": "vmp", "57": "Y", "58": "vsg", "59": "vas", "60": "pi", "61": "vmi", "62": "pt", "63": "Fz", "64": "Fpa", "65": "cs" }, "initializer_range": 0.02, "intermediate_size": 4096, "label2id": { "Faa": 49, "Fat": 24, "Fc": 32, "Fd": 28, "Fe": 9, "Fg": 23, "Fh": 8, "Fia": 40, "Fit": 34, "Fp": 47, "Fpa": 64, "Fpt": 48, "Fs": 4, "Fx": 25, "Fz": 63, "I": 26, "W": 14, "X": 19, "Y": 57, "Z": 13, "Zm": 55, "Zp": 1, "ao": 46, "aq": 11, "cc": 16, "cs": 65, "da": 17, "dd": 6, "de": 39, "di": 18, "dn": 22, "dp": 42, "dt": 45, "i": 0, "nc": 37, "np": 21, "p0": 15, "pd": 35, "pe": 41, "pi": 60, "pn": 50, "pp": 44, "pr": 51, "pt": 62, "px": 36, "rg": 2, "rn": 10, "sp": 33, "vag": 12, "vai": 38, "vam": 53, "van": 29, "vap": 27, "vas": 59, "vmg": 30, "vmi": 61, "vmm": 3, "vmn": 54, "vmp": 56, "vms": 7, "vsg": 58, "vsi": 5, "vsm": 31, "vsn": 20, "vsp": 43, "vss": 52 }, "layer_norm_eps": 1e-05, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 16, "num_hidden_layers": 24, "pad_token_id": 1, "position_embedding_type": "absolute", "torch_dtype": "float32", "transformers_version": "4.44.2", "type_vocab_size": 1, "use_cache": true, "vocab_size": 50262 }