{ "_name_or_path": "dccuchile/bert-base-spanish-wwm-cased", "architectures": [ "BertForTokenClassification" ], "attention_probs_dropout_prob": 0.1, "classifier_dropout": null, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "Fe", "1": "vmg", "2": "de", "3": "W", "4": "np", "5": "vai", "6": "Fh", "7": "vsi", "8": "vsg", "9": "Fat", "10": "Fs", "11": "dt", "12": "sp", "13": "Fc", "14": "pt", "15": "pi", "16": "rn", "17": "vas", "18": "Zm", "19": "X", "20": "vms", "21": "rg", "22": "vsn", "23": "da", "24": "vsm", "25": "nc", "26": "vss", "27": "pe", "28": "Fg", "29": "Fx", "30": "vmp", "31": "px", "32": "aq", "33": "pn", "34": "dn", "35": "Fd", "36": "ao", "37": "Fp", "38": "Zp", "39": "vap", "40": "Y", "41": "I", "42": "cs", "43": "pr", "44": "Z", "45": "vmm", "46": "vmi", "47": "Fpt", "48": "Fit", "49": "van", "50": "vag", "51": "vmn", "52": "p0", "53": "Fia", "54": "i", "55": "Faa", "56": "vam", "57": "Fpa", "58": "pp", "59": "cc", "60": "pd", "61": "vsp", "62": "dp", "63": "Fz", "64": "dd", "65": "di" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "Faa": 55, "Fat": 9, "Fc": 13, "Fd": 35, "Fe": 0, "Fg": 28, "Fh": 6, "Fia": 53, "Fit": 48, "Fp": 37, "Fpa": 57, "Fpt": 47, "Fs": 10, "Fx": 29, "Fz": 63, "I": 41, "W": 3, "X": 19, "Y": 40, "Z": 44, "Zm": 18, "Zp": 38, "ao": 36, "aq": 32, "cc": 59, "cs": 42, "da": 23, "dd": 64, "de": 2, "di": 65, "dn": 34, "dp": 62, "dt": 11, "i": 54, "nc": 25, "np": 4, "p0": 52, "pd": 60, "pe": 27, "pi": 15, "pn": 33, "pp": 58, "pr": 43, "pt": 14, "px": 31, "rg": 21, "rn": 16, "sp": 12, "vag": 50, "vai": 5, "vam": 56, "van": 49, "vap": 39, "vas": 17, "vmg": 1, "vmi": 46, "vmm": 45, "vmn": 51, "vmp": 30, "vms": 20, "vsg": 8, "vsi": 7, "vsm": 24, "vsn": 22, "vsp": 61, "vss": 26 }, "layer_norm_eps": 1e-12, "max_position_embeddings": 512, "model_type": "bert", "num_attention_heads": 12, "num_hidden_layers": 12, "output_past": true, "pad_token_id": 1, "position_embedding_type": "absolute", "torch_dtype": "float32", "transformers_version": "4.44.2", "type_vocab_size": 2, "use_cache": true, "vocab_size": 31002 }