{ "architectures": [ "DebertaV2ForTokenClassification" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "B-ADJ|_", "1": "B-ADP|_", "2": "B-ADV|_", "3": "B-AUX|Polarity=Neg", "4": "B-AUX|_", "5": "B-CCONJ|_", "6": "B-DET|_", "7": "B-INTJ|_", "8": "B-NOUN|Polarity=Neg", "9": "B-NOUN|_", "10": "B-NUM|_", "11": "B-PART|_", "12": "B-PRON|_", "13": "B-PROPN|_", "14": "B-PUNCT|_", "15": "B-SCONJ|_", "16": "B-SYM|_", "17": "B-VERB|_", "18": "B-X|_", "19": "I-ADJ|_", "20": "I-ADP|_", "21": "I-ADV|_", "22": "I-AUX|Polarity=Neg", "23": "I-AUX|_", "24": "I-CCONJ|_", "25": "I-INTJ|_", "26": "I-NOUN|_", "27": "I-NUM|_", "28": "I-PART|_", "29": "I-PRON|_", "30": "I-PROPN|_", "31": "I-PUNCT|_", "32": "I-SCONJ|_", "33": "I-SYM|_", "34": "I-VERB|_", "35": "I-X|_" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "B-ADJ|_": 0, "B-ADP|_": 1, "B-ADV|_": 2, "B-AUX|Polarity=Neg": 3, "B-AUX|_": 4, "B-CCONJ|_": 5, "B-DET|_": 6, "B-INTJ|_": 7, "B-NOUN|Polarity=Neg": 8, "B-NOUN|_": 9, "B-NUM|_": 10, "B-PART|_": 11, "B-PRON|_": 12, "B-PROPN|_": 13, "B-PUNCT|_": 14, "B-SCONJ|_": 15, "B-SYM|_": 16, "B-VERB|_": 17, "B-X|_": 18, "I-ADJ|_": 19, "I-ADP|_": 20, "I-ADV|_": 21, "I-AUX|Polarity=Neg": 22, "I-AUX|_": 23, "I-CCONJ|_": 24, "I-INTJ|_": 25, "I-NOUN|_": 26, "I-NUM|_": 27, "I-PART|_": 28, "I-PRON|_": 29, "I-PROPN|_": 30, "I-PUNCT|_": 31, "I-SCONJ|_": 32, "I-SYM|_": 33, "I-VERB|_": 34, "I-X|_": 35 }, "layer_norm_eps": 1e-07, "max_position_embeddings": 512, "max_relative_positions": -1, "model_type": "deberta-v2", "num_attention_heads": 12, "num_hidden_layers": 12, "pad_token_id": 1, "pooler_dropout": 0, "pooler_hidden_act": "gelu", "pooler_hidden_size": 768, "pos_att_type": null, "position_biased_input": true, "relative_attention": false, "tokenizer_class": "BertJapaneseTokenizer", "torch_dtype": "float32", "transformers_version": "4.19.4", "type_vocab_size": 0, "vocab_size": 32000 }