xlmr-tatoeba / config.json
bigpang's picture
Upload XLMRobertaForSequenceClassification
e8132ad
{
"_name_or_path": "xlm-roberta-base",
"architectures": [
"XLMRobertaForSequenceClassification"
],
"attention_probs_dropout_prob": 0.1,
"bos_token_id": 0,
"classifier_dropout": null,
"eos_token_id": 2,
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 768,
"id2label": {
"0": "af",
"1": "am",
"2": "ar",
"3": "as",
"4": "az",
"5": "bg",
"6": "bn",
"7": "bs",
"8": "ca",
"9": "cs",
"10": "de",
"11": "el",
"12": "en",
"13": "es",
"14": "et",
"15": "fi",
"16": "fr",
"17": "gl",
"18": "gu",
"19": "hi",
"20": "hr",
"21": "ht",
"22": "hu",
"23": "hy",
"24": "id",
"25": "is",
"26": "it",
"27": "iw",
"28": "ja",
"29": "ka",
"30": "kk",
"31": "km",
"32": "kn",
"33": "ko",
"34": "ky",
"35": "lt",
"36": "lv",
"37": "ml",
"38": "mn",
"39": "mr",
"40": "ms",
"41": "my",
"42": "ne",
"43": "nl",
"44": "no",
"45": "or",
"46": "other",
"47": "pa",
"48": "pl",
"49": "ps",
"50": "pt",
"51": "ro",
"52": "ru",
"53": "rw",
"54": "si",
"55": "sk",
"56": "sl",
"57": "so",
"58": "sq",
"59": "sv",
"60": "sw",
"61": "ta",
"62": "te",
"63": "th",
"64": "tl",
"65": "tr",
"66": "ug",
"67": "uk",
"68": "ur",
"69": "uz",
"70": "vi",
"71": "zh",
"72": "zh-Hant",
"73": "zu"
},
"initializer_range": 0.02,
"intermediate_size": 3072,
"label2id": {
"af": 0,
"am": 1,
"ar": 2,
"as": 3,
"az": 4,
"bg": 5,
"bn": 6,
"bs": 7,
"ca": 8,
"cs": 9,
"de": 10,
"el": 11,
"en": 12,
"es": 13,
"et": 14,
"fi": 15,
"fr": 16,
"gl": 17,
"gu": 18,
"hi": 19,
"hr": 20,
"ht": 21,
"hu": 22,
"hy": 23,
"id": 24,
"is": 25,
"it": 26,
"iw": 27,
"ja": 28,
"ka": 29,
"kk": 30,
"km": 31,
"kn": 32,
"ko": 33,
"ky": 34,
"lt": 35,
"lv": 36,
"ml": 37,
"mn": 38,
"mr": 39,
"ms": 40,
"my": 41,
"ne": 42,
"nl": 43,
"no": 44,
"or": 45,
"other": 46,
"pa": 47,
"pl": 48,
"ps": 49,
"pt": 50,
"ro": 51,
"ru": 52,
"rw": 53,
"si": 54,
"sk": 55,
"sl": 56,
"so": 57,
"sq": 58,
"sv": 59,
"sw": 60,
"ta": 61,
"te": 62,
"th": 63,
"tl": 64,
"tr": 65,
"ug": 66,
"uk": 67,
"ur": 68,
"uz": 69,
"vi": 70,
"zh": 71,
"zh-Hant": 72,
"zu": 73
},
"layer_norm_eps": 1e-05,
"max_position_embeddings": 514,
"model_type": "xlm-roberta",
"num_attention_heads": 12,
"num_hidden_layers": 12,
"output_past": true,
"pad_token_id": 1,
"position_embedding_type": "absolute",
"problem_type": "single_label_classification",
"torch_dtype": "float32",
"transformers_version": "4.27.3",
"type_vocab_size": 1,
"use_cache": true,
"vocab_size": 250002
}