{ "_name_or_path": "xlm-roberta-base", "architectures": [ "XLMRobertaForSequenceClassification" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "classifier_dropout": null, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "bg", "1": "hr", "2": "cs", "3": "da", "4": "nl", "5": "en", "6": "et", "7": "fi", "8": "fr", "9": "de", "10": "el", "11": "hu", "12": "it", "13": "lv", "14": "lt", "15": "mt", "16": "no", "17": "pl", "18": "pt", "19": "ro", "20": "sk", "21": "sl", "22": "es", "23": "sv", "24": "ru", "25": "tr", "26": "eu", "27": "ca", "28": "sq", "29": "se", "30": "uk", "31": "ar", "32": "zh", "33": "he", "34": "mhd", "35": "la", "36": "gml", "37": "fro", "38": "chu", "39": "fnhd", "40": "grc" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "ar": 31, "bg": 0, "ca": 27, "chu": 38, "cs": 2, "da": 3, "de": 9, "el": 10, "en": 5, "es": 22, "et": 6, "eu": 26, "fi": 7, "fnhd": 39, "fr": 8, "fro": 37, "gml": 36, "grc": 40, "he": 33, "hr": 1, "hu": 11, "it": 12, "la": 35, "lt": 14, "lv": 13, "mhd": 34, "mt": 15, "nl": 4, "no": 16, "pl": 17, "pt": 18, "ro": 19, "ru": 24, "se": 29, "sk": 20, "sl": 21, "sq": 28, "sv": 23, "tr": 25, "uk": 30, "zh": 32 }, "layer_norm_eps": 1e-05, "max_position_embeddings": 514, "model_type": "xlm-roberta", "num_attention_heads": 12, "num_hidden_layers": 12, "output_past": true, "pad_token_id": 1, "position_embedding_type": "absolute", "problem_type": "single_label_classification", "torch_dtype": "float32", "transformers_version": "4.24.0", "type_vocab_size": 1, "use_cache": true, "vocab_size": 250002 }