{ "_name_or_path": "xlm-roberta-base", "architectures": [ "XLMRobertaForSequenceClassification" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "classifier_dropout": null, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "af", "1": "am", "2": "ar", "3": "as", "4": "az", "5": "bg", "6": "bn", "7": "bs", "8": "ca", "9": "cs", "10": "de", "11": "el", "12": "en", "13": "es", "14": "et", "15": "fi", "16": "fr", "17": "gl", "18": "gu", "19": "hi", "20": "hr", "21": "ht", "22": "hu", "23": "hy", "24": "id", "25": "is", "26": "it", "27": "iw", "28": "ja", "29": "ka", "30": "kk", "31": "km", "32": "kn", "33": "ko", "34": "ky", "35": "lt", "36": "lv", "37": "ml", "38": "mn", "39": "mr", "40": "ms", "41": "my", "42": "ne", "43": "nl", "44": "no", "45": "or", "46": "other", "47": "pa", "48": "pl", "49": "ps", "50": "pt", "51": "ro", "52": "ru", "53": "rw", "54": "si", "55": "sk", "56": "sl", "57": "so", "58": "sq", "59": "sv", "60": "sw", "61": "ta", "62": "te", "63": "th", "64": "tl", "65": "tr", "66": "ug", "67": "uk", "68": "ur", "69": "uz", "70": "vi", "71": "zh", "72": "zh-Hant", "73": "zu" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "af": 0, "am": 1, "ar": 2, "as": 3, "az": 4, "bg": 5, "bn": 6, "bs": 7, "ca": 8, "cs": 9, "de": 10, "el": 11, "en": 12, "es": 13, "et": 14, "fi": 15, "fr": 16, "gl": 17, "gu": 18, "hi": 19, "hr": 20, "ht": 21, "hu": 22, "hy": 23, "id": 24, "is": 25, "it": 26, "iw": 27, "ja": 28, "ka": 29, "kk": 30, "km": 31, "kn": 32, "ko": 33, "ky": 34, "lt": 35, "lv": 36, "ml": 37, "mn": 38, "mr": 39, "ms": 40, "my": 41, "ne": 42, "nl": 43, "no": 44, "or": 45, "other": 46, "pa": 47, "pl": 48, "ps": 49, "pt": 50, "ro": 51, "ru": 52, "rw": 53, "si": 54, "sk": 55, "sl": 56, "so": 57, "sq": 58, "sv": 59, "sw": 60, "ta": 61, "te": 62, "th": 63, "tl": 64, "tr": 65, "ug": 66, "uk": 67, "ur": 68, "uz": 69, "vi": 70, "zh": 71, "zh-Hant": 72, "zu": 73 }, "layer_norm_eps": 1e-05, "max_position_embeddings": 514, "model_type": "xlm-roberta", "num_attention_heads": 12, "num_hidden_layers": 12, "output_past": true, "pad_token_id": 1, "position_embedding_type": "absolute", "problem_type": "single_label_classification", "torch_dtype": "float32", "transformers_version": "4.27.3", "type_vocab_size": 1, "use_cache": true, "vocab_size": 250002 }