{ "_name_or_path": "jam15/bert-finetuned-combineP5", "architectures": [ "BertForSequenceClassification" ], "attention_probs_dropout_prob": 0.1, "classifier_dropout": null, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "loc1lay1", "1": "loc1lay2", "10": "loc3lay3", "11": "loc3lay4", "12": "loc4lay1", "13": "loc4lay2", "14": "loc4lay3", "15": "loc4lay4", "16": "loc5+", "17": "loc1", "18": "loc2", "19": "loc3", "2": "loc1lay3", "20": "loc4", "21": "nfw", "22": "tfw", "23": "tfwc", "24": "tfwp", "25": "fwb", "26": "lay1", "27": "lay2", "28": "lay3", "29": "lay4", "3": "loc1lay4", "4": "loc2lay1", "5": "loc2lay2", "6": "loc2lay3", "7": "loc2lay4", "8": "loc3lay1", "9": "loc3lay2" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "fwb": 25, "lay1": 26, "lay2": 27, "lay3": 28, "lay4": 29, "loc1": 17, "loc1lay1": 0, "loc1lay2": 1, "loc1lay3": 2, "loc1lay4": 3, "loc2": 18, "loc2lay1": 4, "loc2lay2": 5, "loc2lay3": 6, "loc2lay4": 7, "loc3": 19, "loc3lay1": 8, "loc3lay2": 9, "loc3lay3": 10, "loc3lay4": 11, "loc4": 20, "loc4lay1": 12, "loc4lay2": 13, "loc4lay3": 14, "loc4lay4": 15, "loc5+": 16, "nfw": 21, "tfw": 22, "tfwc": 23, "tfwp": 24 }, "layer_norm_eps": 1e-12, "max_position_embeddings": 512, "model_type": "bert", "num_attention_heads": 12, "num_hidden_layers": 12, "pad_token_id": 0, "position_embedding_type": "absolute", "problem_type": "single_label_classification", "torch_dtype": "float32", "transformers_version": "4.41.2", "type_vocab_size": 2, "use_cache": true, "vocab_size": 30522 }