{ "_name_or_path": "AutoTrain", "_num_labels": 23, "architectures": [ "DebertaV2ForSequenceClassification" ], "attention_probs_dropout_prob": 0.1, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "Access to Care", "1": "Access to Care Contradiction", "2": "Depression", "3": "Depression Contradiction", "4": "Economic Instability", "5": "Economic Instability Contradiction", "6": "Employment Stress", "7": "Employment Stress Contradiction", "8": "Exposure to Violence", "9": "Exposure to Violence Contradiction", "10": "Food Insecurity ", "11": "Food Insecurity Contradiction", "12": "Housing Instability", "13": "Housing Instability Contradiction", "14": "Limited Language (English) Proficiency", "15": "Limited Language Proficiency Contradiction", "16": "Neutral", "17": "Social Context", "18": "Social Context Contradiction", "19": "Substance Abuse", "20": "Substance Abuse Contradiction", "21": "Transportation", "22": "Transportation Contradiction" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "Access to Care": 0, "Access to Care Contradiction": 1, "Depression": 2, "Depression Contradiction": 3, "Economic Instability": 4, "Economic Instability Contradiction": 5, "Employment Stress": 6, "Employment Stress Contradiction": 7, "Exposure to Violence": 8, "Exposure to Violence Contradiction": 9, "Food Insecurity ": 10, "Food Insecurity Contradiction": 11, "Housing Instability": 12, "Housing Instability Contradiction": 13, "Limited Language (English) Proficiency": 14, "Limited Language Proficiency Contradiction": 15, "Neutral": 16, "Social Context": 17, "Social Context Contradiction": 18, "Substance Abuse": 19, "Substance Abuse Contradiction": 20, "Transportation": 21, "Transportation Contradiction": 22 }, "layer_norm_eps": 1e-07, "max_length": 64, "max_position_embeddings": 512, "max_relative_positions": -1, "model_type": "deberta-v2", "norm_rel_ebd": "layer_norm", "num_attention_heads": 12, "num_hidden_layers": 12, "pad_token_id": 0, "padding": "max_length", "pooler_dropout": 0, "pooler_hidden_act": "gelu", "pooler_hidden_size": 768, "pos_att_type": [ "p2c", "c2p" ], "position_biased_input": false, "position_buckets": 256, "relative_attention": true, "share_att_key": true, "torch_dtype": "float32", "transformers_version": "4.25.1", "type_vocab_size": 0, "vocab_size": 128100 }