{ "_name_or_path": "microsoft/deberta-v3-small", "architectures": [ "DebertaV2ForSequenceClassification" ], "attention_probs_dropout_prob": 0.1, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "C&B - Payroll", "1": "Talent as a Service", "2": "HRDT - HRBP Requests", "3": "HRDT - HRIS", "4": "HRDT - Data Updating", "5": "HRDT - Requested Report", "6": "HRDT - Others", "7": "P&C Allowance", "8": "P&C Employee Movement/EAA Request", "9": "P&C HR Documents", "10": "P&C HRIS", "11": "P&C Incident Report/Employee Discipline", "12": "P&C Inquiries", "13": "P&C Internship", "14": "P&C Leave Benefits", "15": "P&C Mental Health / MindNation", "16": "P&C Payroll", "17": "P&C Performance/Coaching", "18": "P&C Reports", "19": "P&C Resignation/Offboarding", "20": "P&C SSS | Pag-Ibig | Philhealth | BIR", "21": "P&C Verification", "22": "P&C Others", "23": "People Experience" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "C&B - Payroll": 0, "HRDT - Data Updating": 4, "HRDT - HRBP Requests": 2, "HRDT - HRIS": 3, "HRDT - Others": 6, "HRDT - Requested Report": 5, "P&C Allowance": 7, "P&C Employee Movement/EAA Request": 8, "P&C HR Documents": 9, "P&C HRIS": 10, "P&C Incident Report/Employee Discipline": 11, "P&C Inquiries": 12, "P&C Internship": 13, "P&C Leave Benefits": 14, "P&C Mental Health / MindNation": 15, "P&C Others": 22, "P&C Payroll": 16, "P&C Performance/Coaching": 17, "P&C Reports": 18, "P&C Resignation/Offboarding": 19, "P&C SSS | Pag-Ibig | Philhealth | BIR": 20, "P&C Verification": 21, "People Experience": 23, "Talent as a Service": 1 }, "layer_norm_eps": 1e-07, "max_position_embeddings": 512, "max_relative_positions": -1, "model_type": "deberta-v2", "norm_rel_ebd": "layer_norm", "num_attention_heads": 12, "num_hidden_layers": 6, "pad_token_id": 0, "pooler_dropout": 0, "pooler_hidden_act": "gelu", "pooler_hidden_size": 768, "pos_att_type": [ "p2c", "c2p" ], "position_biased_input": false, "position_buckets": 256, "problem_type": "multi_label_classification", "relative_attention": true, "share_att_key": true, "torch_dtype": "float32", "transformers_version": "4.42.4", "type_vocab_size": 0, "vocab_size": 128100 }