{ "_name_or_path": "microsoft/Multilingual-MiniLM-L12-H384", "architectures": [ "BertForSequenceClassification" ], "attention_probs_dropout_prob": 0.1, "classifier_dropout": null, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 384, "id2label": { "0": "❀", "1": "😍", "2": "πŸ˜‚", "3": "πŸ’•", "4": "πŸ”₯", "5": "😊", "6": "😎", "7": "✨", "8": "πŸ’™", "9": "😘", "10": "πŸ“·", "11": "πŸ‡ΊπŸ‡Έ", "12": "β˜€", "13": "πŸ’œ", "14": "πŸ˜‰", "15": "πŸ’―", "16": "😁", "17": "πŸŽ„", "18": "πŸ“Έ", "19": "😜" }, "initializer_range": 0.02, "intermediate_size": 1536, "label2id": { "❀": 0, "😍": 1, "πŸ˜‚": 2, "πŸ’•": 3, "πŸ”₯": 4, "😊": 5, "😎": 6, "✨": 7, "πŸ’™": 8, "😘": 9, "πŸ“·": 10, "πŸ‡ΊπŸ‡Έ": 11, "β˜€": 12, "πŸ’œ": 13, "πŸ˜‰": 14, "πŸ’―": 15, "😁": 16, "πŸŽ„": 17, "πŸ“Έ": 18, "😜": 19 }, "layer_norm_eps": 1e-12, "max_position_embeddings": 512, "model_type": "bert", "num_attention_heads": 12, "num_hidden_layers": 12, "pad_token_id": 0, "position_embedding_type": "absolute", "problem_type": "single_label_classification", "tokenizer_class": "XLMRobertaTokenizer", "torch_dtype": "float32", "transformers_version": "4.25.1", "type_vocab_size": 2, "use_cache": true, "vocab_size": 250037 }