{ "_name_or_path": "pjox/dalembert", "architectures": [ "RobertaForSequenceClassification" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "classifier_dropout": null, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "leroy_cg", "1": "bordeu", "2": "liebault", "3": "paillasson", "4": "lavirotte", "5": "meyzieu", "6": "lachapelle", "7": "watelet", "8": "morellet", "9": "lacondamine", "10": "lefevre", "11": "faiguet", "12": "rousseau", "13": "abbes", "14": "holbach", "15": "cahusac", "16": "pezay", "17": "sauvages", "18": "dumarsais", "19": "louis", "20": "perronet", "21": "romilly_j", "22": "tarin", "23": "deleyre", "24": "genson", "25": "romilly_je", "26": "formey", "27": "venel", "28": "blondel", "29": "boucher", "30": "duclos", "31": "berthoud", "32": "naigeon", "33": "vandenesse", "34": "monnoye", "35": "desmahis", "36": "damilaville", "37": "brulle", "38": "kurdwanowski", "39": "bouchaud", "40": "pesselier", "41": "petit", "42": "allut", "43": "fenouillot", "44": "toussaint", "45": "leblond", "46": "quesnay", "47": "montet", "48": "jaucourt", "49": "bertrand", "50": "leromain", "51": "grimm", "52": "willermoz", "53": "yvon", "54": "desmarest", "55": "pestre", "56": "eidous", "57": "forbonnais", "58": "landois", "59": "tressan", "60": "lenglet", "61": "morand", "62": "robert", "63": "daubenton_ljm", "64": "millot", "65": "collot", "66": "lucotte_jr", "67": "bouillet_j", "68": "boullanger", "69": "dufour", "70": "montdorge", "71": "soubeyran", "72": "mallet", "73": "barthez_de_marmorieres", "74": "gueneau", "75": "bellin", "76": "durival_j", "77": "authville", "78": "penchenier", "79": "david", "80": "falconet", "81": "lebreton", "82": "tronchin", "83": "margency", "84": "bourgelat", "85": "argenville", "86": "papillon", "87": "lesage", "88": "lamotte", "89": "diderot", "90": "goussier", "91": "villiers", "92": "rallier", "93": "barthez", "94": "ratte", "95": "leroy_jb", "96": "menuret", "97": "beauzee", "98": "aumont", "99": "grosley", "100": "roux", "101": "dalembert", "102": "voltaire", "103": "malouin", "104": "brisson", "105": "marmontel", "106": "necker", "107": "voglie", "108": "anville", "109": "fouquet", "110": "daubenton_p", "111": "lemonnier", "112": "bouchu" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "abbes": 13, "allut": 42, "anville": 108, "argenville": 85, "aumont": 98, "authville": 77, "barthez": 93, "barthez_de_marmorieres": 73, "beauzee": 97, "bellin": 75, "berthoud": 31, "bertrand": 49, "blondel": 28, "bordeu": 1, "bouchaud": 39, "boucher": 29, "bouchu": 112, "bouillet_j": 67, "boullanger": 68, "bourgelat": 84, "brisson": 104, "brulle": 37, "cahusac": 15, "collot": 65, "dalembert": 101, "damilaville": 36, "daubenton_ljm": 63, "daubenton_p": 110, "david": 79, "deleyre": 23, "desmahis": 35, "desmarest": 54, "diderot": 89, "duclos": 30, "dufour": 69, "dumarsais": 18, "durival_j": 76, "eidous": 56, "faiguet": 11, "falconet": 80, "fenouillot": 43, "forbonnais": 57, "formey": 26, "fouquet": 109, "genson": 24, "goussier": 90, "grimm": 51, "grosley": 99, "gueneau": 74, "holbach": 14, "jaucourt": 48, "kurdwanowski": 38, "lachapelle": 6, "lacondamine": 9, "lamotte": 88, "landois": 58, "lavirotte": 4, "leblond": 45, "lebreton": 81, "lefevre": 10, "lemonnier": 111, "lenglet": 60, "leromain": 50, "leroy_cg": 0, "leroy_jb": 95, "lesage": 87, "liebault": 2, "louis": 19, "lucotte_jr": 66, "mallet": 72, "malouin": 103, "margency": 83, "marmontel": 105, "menuret": 96, "meyzieu": 5, "millot": 64, "monnoye": 34, "montdorge": 70, "montet": 47, "morand": 61, "morellet": 8, "naigeon": 32, "necker": 106, "paillasson": 3, "papillon": 86, "penchenier": 78, "perronet": 20, "pesselier": 40, "pestre": 55, "petit": 41, "pezay": 16, "quesnay": 46, "rallier": 92, "ratte": 94, "robert": 62, "romilly_j": 21, "romilly_je": 25, "rousseau": 12, "roux": 100, "sauvages": 17, "soubeyran": 71, "tarin": 22, "toussaint": 44, "tressan": 59, "tronchin": 82, "vandenesse": 33, "venel": 27, "villiers": 91, "voglie": 107, "voltaire": 102, "watelet": 7, "willermoz": 52, "yvon": 53 }, "layer_norm_eps": 1e-05, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 12, "num_hidden_layers": 12, "pad_token_id": 1, "position_embedding_type": "absolute", "problem_type": "single_label_classification", "torch_dtype": "float32", "transformers_version": "4.42.4", "type_vocab_size": 1, "use_cache": true, "vocab_size": 32768 }