{ "_name_or_path": "microsoft/deberta-v3-base", "architectures": [ "DebertaV2ForTokenClassification" ], "attention_probs_dropout_prob": 0.1, "finetuning_task": "ner", "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "B-Company", "1": "B-Date", "2": "B-Degree", "3": "B-Education", "4": "B-Email", "5": "B-Experiments", "6": "B-Gender", "7": "B-Institution", "8": "B-Job", "9": "B-Language", "10": "B-Location", "11": "B-Name", "12": "B-Overall", "13": "B-Phone", "14": "B-Position", "15": "B-Program", "16": "B-Project", "17": "B-Skill", "18": "B-Training", "19": "I-Company", "20": "I-Date", "21": "I-Degree", "22": "I-Education", "23": "I-Email", "24": "I-Experiments", "25": "I-Gender", "26": "I-Institution", "27": "I-Job", "28": "I-Language", "29": "I-Location", "30": "I-Name", "31": "I-Overall", "32": "I-Phone", "33": "I-Position", "34": "I-Program", "35": "I-Project", "36": "I-Skill", "37": "I-Training", "38": "O" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "B-Company": 0, "B-Date": 1, "B-Degree": 2, "B-Education": 3, "B-Email": 4, "B-Experiments": 5, "B-Gender": 6, "B-Institution": 7, "B-Job": 8, "B-Language": 9, "B-Location": 10, "B-Name": 11, "B-Overall": 12, "B-Phone": 13, "B-Position": 14, "B-Program": 15, "B-Project": 16, "B-Skill": 17, "B-Training": 18, "I-Company": 19, "I-Date": 20, "I-Degree": 21, "I-Education": 22, "I-Email": 23, "I-Experiments": 24, "I-Gender": 25, "I-Institution": 26, "I-Job": 27, "I-Language": 28, "I-Location": 29, "I-Name": 30, "I-Overall": 31, "I-Phone": 32, "I-Position": 33, "I-Program": 34, "I-Project": 35, "I-Skill": 36, "I-Training": 37, "O": 38 }, "layer_norm_eps": 1e-07, "max_position_embeddings": 512, "max_relative_positions": -1, "model_type": "deberta-v2", "norm_rel_ebd": "layer_norm", "num_attention_heads": 12, "num_hidden_layers": 12, "pad_token_id": 0, "pooler_dropout": 0, "pooler_hidden_act": "gelu", "pooler_hidden_size": 768, "pos_att_type": [ "p2c", "c2p" ], "position_biased_input": false, "position_buckets": 256, "relative_attention": true, "share_att_key": true, "torch_dtype": "float32", "transformers_version": "4.22.2", "type_vocab_size": 0, "vocab_size": 128100 }