{ "_name_or_path": "microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract-fulltext", "architectures": [ "BertForTokenClassification" ], "attention_probs_dropout_prob": 0.1, "classifier_dropout": null, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "O", "1": "B-chemical", "2": "B-species", "3": "B-gene", "4": "I-gene", "5": "B-taxonomy_domain", "6": "I-chemical", "7": "B-protein_type", "8": "B-experimental_method", "9": "I-experimental_method", "10": "I-protein_type", "11": "I-species", "12": "B-evidence", "13": "I-evidence", "14": "B-protein", "15": "I-protein", "16": "B-protein_state", "17": "B-site", "18": "I-site", "19": "B-structure_element", "20": "I-structure_element", "21": "B-complex_assembly", "22": "I-protein_state", "23": "B-mutant", "24": "I-mutant", "25": "B-oligomeric_state", "26": "I-complex_assembly", "27": "B-residue_range", "28": "I-residue_range", "29": "B-residue_name_number", "30": "B-residue_name", "31": "I-taxonomy_domain", "32": "I-oligomeric_state", "33": "B-ptm", "34": "I-ptm", "35": "B-residue_number", "36": "I-residue_name_number", "37": "I-residue_number" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "B-chemical": 1, "B-complex_assembly": 21, "B-evidence": 12, "B-experimental_method": 8, "B-gene": 3, "B-mutant": 23, "B-oligomeric_state": 25, "B-protein": 14, "B-protein_state": 16, "B-protein_type": 7, "B-ptm": 33, "B-residue_name": 30, "B-residue_name_number": 29, "B-residue_number": 35, "B-residue_range": 27, "B-site": 17, "B-species": 2, "B-structure_element": 19, "B-taxonomy_domain": 5, "I-chemical": 6, "I-complex_assembly": 26, "I-evidence": 13, "I-experimental_method": 9, "I-gene": 4, "I-mutant": 24, "I-oligomeric_state": 32, "I-protein": 15, "I-protein_state": 22, "I-protein_type": 10, "I-ptm": 34, "I-residue_name_number": 36, "I-residue_number": 37, "I-residue_range": 28, "I-site": 18, "I-species": 11, "I-structure_element": 20, "I-taxonomy_domain": 31, "O": 0 }, "layer_norm_eps": 1e-12, "max_position_embeddings": 512, "model_type": "bert", "num_attention_heads": 12, "num_hidden_layers": 12, "pad_token_id": 0, "position_embedding_type": "absolute", "torch_dtype": "float32", "transformers_version": "4.28.0", "type_vocab_size": 2, "use_cache": true, "vocab_size": 30522 }