{ "_name_or_path": "pjox/dalembert", "architectures": [ "RobertaForSequenceClassification" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "classifier_dropout": null, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "deleyre", "1": "leroy_jb", "2": "bouchaud", "3": "watelet", "4": "landois", "5": "villiers", "6": "liebault", "7": "barthez", "8": "ratte", "9": "grosley", "10": "eidous", "11": "sauvages", "12": "argenville", "13": "lachapelle", "14": "bouchu", "15": "lacondamine", "16": "pestre", "17": "lebreton", "18": "romilly_j", "19": "lucotte_jr", "20": "dalembert", "21": "brisson", "22": "montdorge", "23": "david", "24": "dumarsais", "25": "montet", "26": "jaucourt", "27": "fenouillot", "28": "pezay", "29": "necker", "30": "lesage", "31": "tarin", "32": "lavirotte", "33": "morellet", "34": "menuret", "35": "millot", "36": "paillasson", "37": "monnoye", "38": "leromain", "39": "boucher", "40": "bordeu", "41": "robert", "42": "toussaint", "43": "lenglet", "44": "bellin", "45": "willermoz", "46": "boullanger", "47": "forbonnais", "48": "mallet", "49": "malouin", "50": "blondel", "51": "damilaville", "52": "margency", "53": "barthez_de_marmorieres", "54": "tressan", "55": "tronchin", "56": "fouquet", "57": "holbach", "58": "perronet", "59": "leblond", "60": "petit", "61": "goussier", "62": "romilly_je", "63": "bourgelat", "64": "aumont", "65": "louis", "66": "formey", "67": "papillon", "68": "daubenton_p", "69": "collot", "70": "allut", "71": "vandenesse", "72": "faiguet", "73": "meyzieu", "74": "naigeon", "75": "pesselier", "76": "lemonnier", "77": "durival_j", "78": "rousseau", "79": "leroy_cg", "80": "abbes", "81": "soubeyran", "82": "dufour", "83": "brulle", "84": "beauzee", "85": "cahusac", "86": "bertrand", "87": "yvon", "88": "lefevre", "89": "rallier", "90": "grimm", "91": "diderot", "92": "lamotte", "93": "berthoud", "94": "desmarest", "95": "bouillet_j", "96": "quesnay", "97": "kurdwanowski", "98": "voglie", "99": "morand", "100": "venel", "101": "genson", "102": "authville", "103": "anville", "104": "voltaire", "105": "marmontel", "106": "duclos", "107": "roux", "108": "desmahis", "109": "falconet", "110": "gueneau", "111": "daubenton_ljm", "112": "penchenier" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "abbes": 80, "allut": 70, "anville": 103, "argenville": 12, "aumont": 64, "authville": 102, "barthez": 7, "barthez_de_marmorieres": 53, "beauzee": 84, "bellin": 44, "berthoud": 93, "bertrand": 86, "blondel": 50, "bordeu": 40, "bouchaud": 2, "boucher": 39, "bouchu": 14, "bouillet_j": 95, "boullanger": 46, "bourgelat": 63, "brisson": 21, "brulle": 83, "cahusac": 85, "collot": 69, "dalembert": 20, "damilaville": 51, "daubenton_ljm": 111, "daubenton_p": 68, "david": 23, "deleyre": 0, "desmahis": 108, "desmarest": 94, "diderot": 91, "duclos": 106, "dufour": 82, "dumarsais": 24, "durival_j": 77, "eidous": 10, "faiguet": 72, "falconet": 109, "fenouillot": 27, "forbonnais": 47, "formey": 66, "fouquet": 56, "genson": 101, "goussier": 61, "grimm": 90, "grosley": 9, "gueneau": 110, "holbach": 57, "jaucourt": 26, "kurdwanowski": 97, "lachapelle": 13, "lacondamine": 15, "lamotte": 92, "landois": 4, "lavirotte": 32, "leblond": 59, "lebreton": 17, "lefevre": 88, "lemonnier": 76, "lenglet": 43, "leromain": 38, "leroy_cg": 79, "leroy_jb": 1, "lesage": 30, "liebault": 6, "louis": 65, "lucotte_jr": 19, "mallet": 48, "malouin": 49, "margency": 52, "marmontel": 105, "menuret": 34, "meyzieu": 73, "millot": 35, "monnoye": 37, "montdorge": 22, "montet": 25, "morand": 99, "morellet": 33, "naigeon": 74, "necker": 29, "paillasson": 36, "papillon": 67, "penchenier": 112, "perronet": 58, "pesselier": 75, "pestre": 16, "petit": 60, "pezay": 28, "quesnay": 96, "rallier": 89, "ratte": 8, "robert": 41, "romilly_j": 18, "romilly_je": 62, "rousseau": 78, "roux": 107, "sauvages": 11, "soubeyran": 81, "tarin": 31, "toussaint": 42, "tressan": 54, "tronchin": 55, "vandenesse": 71, "venel": 100, "villiers": 5, "voglie": 98, "voltaire": 104, "watelet": 3, "willermoz": 45, "yvon": 87 }, "layer_norm_eps": 1e-05, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 12, "num_hidden_layers": 12, "pad_token_id": 1, "position_embedding_type": "absolute", "problem_type": "single_label_classification", "torch_dtype": "float32", "transformers_version": "4.47.1", "type_vocab_size": 1, "use_cache": true, "vocab_size": 32768 }