{ | |
"_name_or_path": "xlm-roberta-base", | |
"architectures": [ | |
"XLMRobertaForSequenceClassification" | |
], | |
"attention_probs_dropout_prob": 0.1, | |
"bos_token_id": 0, | |
"classifier_dropout": null, | |
"eos_token_id": 2, | |
"hidden_act": "gelu", | |
"hidden_dropout_prob": 0.1, | |
"hidden_size": 768, | |
"id2label": { | |
"0": "af", | |
"1": "am", | |
"2": "ar", | |
"3": "as", | |
"4": "az", | |
"5": "bg", | |
"6": "bn", | |
"7": "bs", | |
"8": "ca", | |
"9": "cs", | |
"10": "de", | |
"11": "el", | |
"12": "en", | |
"13": "es", | |
"14": "et", | |
"15": "fi", | |
"16": "fr", | |
"17": "gl", | |
"18": "gu", | |
"19": "hi", | |
"20": "hr", | |
"21": "ht", | |
"22": "hu", | |
"23": "hy", | |
"24": "id", | |
"25": "is", | |
"26": "it", | |
"27": "iw", | |
"28": "ja", | |
"29": "ka", | |
"30": "kk", | |
"31": "km", | |
"32": "kn", | |
"33": "ko", | |
"34": "ky", | |
"35": "lt", | |
"36": "lv", | |
"37": "ml", | |
"38": "mn", | |
"39": "mr", | |
"40": "ms", | |
"41": "my", | |
"42": "ne", | |
"43": "nl", | |
"44": "no", | |
"45": "or", | |
"46": "other", | |
"47": "pa", | |
"48": "pl", | |
"49": "ps", | |
"50": "pt", | |
"51": "ro", | |
"52": "ru", | |
"53": "rw", | |
"54": "si", | |
"55": "sk", | |
"56": "sl", | |
"57": "so", | |
"58": "sq", | |
"59": "sv", | |
"60": "sw", | |
"61": "ta", | |
"62": "te", | |
"63": "th", | |
"64": "tl", | |
"65": "tr", | |
"66": "ug", | |
"67": "uk", | |
"68": "ur", | |
"69": "uz", | |
"70": "vi", | |
"71": "zh", | |
"72": "zh-Hant", | |
"73": "zu" | |
}, | |
"initializer_range": 0.02, | |
"intermediate_size": 3072, | |
"label2id": { | |
"af": 0, | |
"am": 1, | |
"ar": 2, | |
"as": 3, | |
"az": 4, | |
"bg": 5, | |
"bn": 6, | |
"bs": 7, | |
"ca": 8, | |
"cs": 9, | |
"de": 10, | |
"el": 11, | |
"en": 12, | |
"es": 13, | |
"et": 14, | |
"fi": 15, | |
"fr": 16, | |
"gl": 17, | |
"gu": 18, | |
"hi": 19, | |
"hr": 20, | |
"ht": 21, | |
"hu": 22, | |
"hy": 23, | |
"id": 24, | |
"is": 25, | |
"it": 26, | |
"iw": 27, | |
"ja": 28, | |
"ka": 29, | |
"kk": 30, | |
"km": 31, | |
"kn": 32, | |
"ko": 33, | |
"ky": 34, | |
"lt": 35, | |
"lv": 36, | |
"ml": 37, | |
"mn": 38, | |
"mr": 39, | |
"ms": 40, | |
"my": 41, | |
"ne": 42, | |
"nl": 43, | |
"no": 44, | |
"or": 45, | |
"other": 46, | |
"pa": 47, | |
"pl": 48, | |
"ps": 49, | |
"pt": 50, | |
"ro": 51, | |
"ru": 52, | |
"rw": 53, | |
"si": 54, | |
"sk": 55, | |
"sl": 56, | |
"so": 57, | |
"sq": 58, | |
"sv": 59, | |
"sw": 60, | |
"ta": 61, | |
"te": 62, | |
"th": 63, | |
"tl": 64, | |
"tr": 65, | |
"ug": 66, | |
"uk": 67, | |
"ur": 68, | |
"uz": 69, | |
"vi": 70, | |
"zh": 71, | |
"zh-Hant": 72, | |
"zu": 73 | |
}, | |
"layer_norm_eps": 1e-05, | |
"max_position_embeddings": 514, | |
"model_type": "xlm-roberta", | |
"num_attention_heads": 12, | |
"num_hidden_layers": 12, | |
"output_past": true, | |
"pad_token_id": 1, | |
"position_embedding_type": "absolute", | |
"problem_type": "single_label_classification", | |
"torch_dtype": "float32", | |
"transformers_version": "4.27.3", | |
"type_vocab_size": 1, | |
"use_cache": true, | |
"vocab_size": 250002 | |
} | |