{ "_name_or_path": "AutoNLP", "_num_labels": 77, "architectures": [ "RobertaForSequenceClassification" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "eos_token_id": 2, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "0", "1": "1", "2": "10", "3": "11", "4": "12", "5": "13", "6": "14", "7": "15", "8": "16", "9": "17", "10": "18", "11": "19", "12": "2", "13": "20", "14": "21", "15": "22", "16": "23", "17": "24", "18": "25", "19": "26", "20": "27", "21": "28", "22": "29", "23": "3", "24": "30", "25": "31", "26": "32", "27": "33", "28": "34", "29": "35", "30": "36", "31": "37", "32": "38", "33": "39", "34": "4", "35": "40", "36": "41", "37": "42", "38": "43", "39": "44", "40": "45", "41": "46", "42": "47", "43": "48", "44": "49", "45": "5", "46": "50", "47": "51", "48": "52", "49": "53", "50": "54", "51": "55", "52": "56", "53": "57", "54": "58", "55": "59", "56": "6", "57": "60", "58": "61", "59": "62", "60": "63", "61": "64", "62": "65", "63": "66", "64": "67", "65": "68", "66": "69", "67": "7", "68": "70", "69": "71", "70": "72", "71": "73", "72": "74", "73": "75", "74": "76", "75": "8", "76": "9" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "0": 0, "1": 1, "10": 2, "11": 3, "12": 4, "13": 5, "14": 6, "15": 7, "16": 8, "17": 9, "18": 10, "19": 11, "2": 12, "20": 13, "21": 14, "22": 15, "23": 16, "24": 17, "25": 18, "26": 19, "27": 20, "28": 21, "29": 22, "3": 23, "30": 24, "31": 25, "32": 26, "33": 27, "34": 28, "35": 29, "36": 30, "37": 31, "38": 32, "39": 33, "4": 34, "40": 35, "41": 36, "42": 37, "43": 38, "44": 39, "45": 40, "46": 41, "47": 42, "48": 43, "49": 44, "5": 45, "50": 46, "51": 47, "52": 48, "53": 49, "54": 50, "55": 51, "56": 52, "57": 53, "58": 54, "59": 55, "6": 56, "60": 57, "61": 58, "62": 59, "63": 60, "64": 61, "65": 62, "66": 63, "67": 64, "68": 65, "69": 66, "7": 67, "70": 68, "71": 69, "72": 70, "73": 71, "74": 72, "75": 73, "76": 74, "8": 75, "9": 76 }, "layer_norm_eps": 1e-05, "max_length": 96, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 12, "num_hidden_layers": 6, "pad_token_id": 1, "padding": "max_length", "position_embedding_type": "absolute", "problem_type": "single_label_classification", "transformers_version": "4.8.0", "type_vocab_size": 1, "use_cache": true, "vocab_size": 50265 }