systemslibrarian commited on
Commit
f61584d
·
verified ·
1 Parent(s): 8196217

Upload folder using huggingface_hub

Browse files
checkpoint-1251/config.json ADDED
@@ -0,0 +1,194 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation": "gelu",
3
+ "architectures": [
4
+ "DistilBertForSequenceClassification"
5
+ ],
6
+ "attention_dropout": 0.1,
7
+ "bos_token_id": null,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "dtype": "float32",
11
+ "eos_token_id": null,
12
+ "hidden_dim": 3072,
13
+ "id2label": {
14
+ "0": "adfgvx",
15
+ "1": "adfgx",
16
+ "2": "aeneas_tacticus",
17
+ "3": "affine",
18
+ "4": "alberti_disk",
19
+ "5": "argenti",
20
+ "6": "arnold_andre",
21
+ "7": "atbash",
22
+ "8": "autokey",
23
+ "9": "babington",
24
+ "10": "bacon_cipher",
25
+ "11": "bazeries",
26
+ "12": "beaufort",
27
+ "13": "bifid",
28
+ "14": "book_cipher",
29
+ "15": "caesar",
30
+ "16": "cardano_autokey",
31
+ "17": "chaocipher",
32
+ "18": "chinese_telegraph",
33
+ "19": "columnar_transposition",
34
+ "20": "commercial_code",
35
+ "21": "confederate_vigenere",
36
+ "22": "copiale",
37
+ "23": "culper_ring",
38
+ "24": "diana",
39
+ "25": "double_transposition",
40
+ "26": "enigma",
41
+ "27": "fialka",
42
+ "28": "four_square",
43
+ "29": "fractionated_morse",
44
+ "30": "geez_monastic",
45
+ "31": "geheimschreiber",
46
+ "32": "great_cipher",
47
+ "33": "gronsfeld",
48
+ "34": "hill",
49
+ "35": "homophonic",
50
+ "36": "jefferson_disk",
51
+ "37": "jn25",
52
+ "38": "joseon_yeokhak",
53
+ "39": "kama_sutra",
54
+ "40": "kl7",
55
+ "41": "kryha",
56
+ "42": "kryptos",
57
+ "43": "lorenz",
58
+ "44": "m209",
59
+ "45": "m94",
60
+ "46": "monoalphabetic",
61
+ "47": "morse_code",
62
+ "48": "navajo_code",
63
+ "49": "nihilist",
64
+ "50": "nomenclator",
65
+ "51": "null_cipher",
66
+ "52": "one_time_pad",
67
+ "53": "pigpen",
68
+ "54": "playfair",
69
+ "55": "polybius",
70
+ "56": "porta",
71
+ "57": "purple",
72
+ "58": "rail_fence",
73
+ "59": "red_type_a",
74
+ "60": "rot13",
75
+ "61": "running_key",
76
+ "62": "scytale",
77
+ "63": "sigaba",
78
+ "64": "slidex",
79
+ "65": "solitaire",
80
+ "66": "stager_route",
81
+ "67": "straddling_checkerboard",
82
+ "68": "tap_code",
83
+ "69": "trifid",
84
+ "70": "trithemius",
85
+ "71": "two_square",
86
+ "72": "typex",
87
+ "73": "venona_pad_reuse",
88
+ "74": "vernam",
89
+ "75": "vic",
90
+ "76": "vigenere",
91
+ "77": "voynich_render",
92
+ "78": "wallis_cipher",
93
+ "79": "wheatstone",
94
+ "80": "zimmermann"
95
+ },
96
+ "initializer_range": 0.02,
97
+ "label2id": {
98
+ "adfgvx": 0,
99
+ "adfgx": 1,
100
+ "aeneas_tacticus": 2,
101
+ "affine": 3,
102
+ "alberti_disk": 4,
103
+ "argenti": 5,
104
+ "arnold_andre": 6,
105
+ "atbash": 7,
106
+ "autokey": 8,
107
+ "babington": 9,
108
+ "bacon_cipher": 10,
109
+ "bazeries": 11,
110
+ "beaufort": 12,
111
+ "bifid": 13,
112
+ "book_cipher": 14,
113
+ "caesar": 15,
114
+ "cardano_autokey": 16,
115
+ "chaocipher": 17,
116
+ "chinese_telegraph": 18,
117
+ "columnar_transposition": 19,
118
+ "commercial_code": 20,
119
+ "confederate_vigenere": 21,
120
+ "copiale": 22,
121
+ "culper_ring": 23,
122
+ "diana": 24,
123
+ "double_transposition": 25,
124
+ "enigma": 26,
125
+ "fialka": 27,
126
+ "four_square": 28,
127
+ "fractionated_morse": 29,
128
+ "geez_monastic": 30,
129
+ "geheimschreiber": 31,
130
+ "great_cipher": 32,
131
+ "gronsfeld": 33,
132
+ "hill": 34,
133
+ "homophonic": 35,
134
+ "jefferson_disk": 36,
135
+ "jn25": 37,
136
+ "joseon_yeokhak": 38,
137
+ "kama_sutra": 39,
138
+ "kl7": 40,
139
+ "kryha": 41,
140
+ "kryptos": 42,
141
+ "lorenz": 43,
142
+ "m209": 44,
143
+ "m94": 45,
144
+ "monoalphabetic": 46,
145
+ "morse_code": 47,
146
+ "navajo_code": 48,
147
+ "nihilist": 49,
148
+ "nomenclator": 50,
149
+ "null_cipher": 51,
150
+ "one_time_pad": 52,
151
+ "pigpen": 53,
152
+ "playfair": 54,
153
+ "polybius": 55,
154
+ "porta": 56,
155
+ "purple": 57,
156
+ "rail_fence": 58,
157
+ "red_type_a": 59,
158
+ "rot13": 60,
159
+ "running_key": 61,
160
+ "scytale": 62,
161
+ "sigaba": 63,
162
+ "slidex": 64,
163
+ "solitaire": 65,
164
+ "stager_route": 66,
165
+ "straddling_checkerboard": 67,
166
+ "tap_code": 68,
167
+ "trifid": 69,
168
+ "trithemius": 70,
169
+ "two_square": 71,
170
+ "typex": 72,
171
+ "venona_pad_reuse": 73,
172
+ "vernam": 74,
173
+ "vic": 75,
174
+ "vigenere": 76,
175
+ "voynich_render": 77,
176
+ "wallis_cipher": 78,
177
+ "wheatstone": 79,
178
+ "zimmermann": 80
179
+ },
180
+ "max_position_embeddings": 512,
181
+ "model_type": "distilbert",
182
+ "n_heads": 12,
183
+ "n_layers": 6,
184
+ "pad_token_id": 0,
185
+ "problem_type": "single_label_classification",
186
+ "qa_dropout": 0.1,
187
+ "seq_classif_dropout": 0.2,
188
+ "sinusoidal_pos_embds": false,
189
+ "tie_weights_": true,
190
+ "tie_word_embeddings": true,
191
+ "transformers_version": "5.8.0",
192
+ "use_cache": false,
193
+ "vocab_size": 30522
194
+ }
checkpoint-1251/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b5188ce6cd1a7ab0beea37097723e66223ee3458744b052d229122df3aff39d
3
+ size 268075580
checkpoint-1251/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18c994d470b86116b4929bea77ab319b2994b6d3cad2735997b97b946bfbcd58
3
+ size 536215243
checkpoint-1251/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8129a65f2943c3f1125b5547fbdd4bc15ec5ab5372edd6800166fdd54d8c70eb
3
+ size 14645
checkpoint-1251/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b162ccbdfb641b869a0738d8b8861b094215a8daccd86f21565528b0050f7ec
3
+ size 1465
checkpoint-1251/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1251/tokenizer_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "cls_token": "[CLS]",
4
+ "do_lower_case": true,
5
+ "is_local": false,
6
+ "local_files_only": false,
7
+ "mask_token": "[MASK]",
8
+ "model_max_length": 512,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "strip_accents": null,
12
+ "tokenize_chinese_chars": true,
13
+ "tokenizer_class": "BertTokenizer",
14
+ "unk_token": "[UNK]"
15
+ }
checkpoint-1251/trainer_state.json ADDED
@@ -0,0 +1,221 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 1251,
3
+ "best_metric": 0.42778234953459543,
4
+ "best_model_checkpoint": "/tmp/cipher_model/checkpoint-1251",
5
+ "epoch": 1.0,
6
+ "eval_steps": 500,
7
+ "global_step": 1251,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.03996802557953637,
14
+ "grad_norm": 2.1291019916534424,
15
+ "learning_rate": 1.97388755662137e-05,
16
+ "loss": 4.206104736328125,
17
+ "step": 50
18
+ },
19
+ {
20
+ "epoch": 0.07993605115907274,
21
+ "grad_norm": 2.337212085723877,
22
+ "learning_rate": 1.9472422062350123e-05,
23
+ "loss": 3.785872802734375,
24
+ "step": 100
25
+ },
26
+ {
27
+ "epoch": 0.11990407673860912,
28
+ "grad_norm": 2.5010275840759277,
29
+ "learning_rate": 1.9205968558486547e-05,
30
+ "loss": 3.458525390625,
31
+ "step": 150
32
+ },
33
+ {
34
+ "epoch": 0.15987210231814547,
35
+ "grad_norm": 3.1468915939331055,
36
+ "learning_rate": 1.893951505462297e-05,
37
+ "loss": 3.239898681640625,
38
+ "step": 200
39
+ },
40
+ {
41
+ "epoch": 0.19984012789768185,
42
+ "grad_norm": 2.6385934352874756,
43
+ "learning_rate": 1.8673061550759395e-05,
44
+ "loss": 3.1392282104492186,
45
+ "step": 250
46
+ },
47
+ {
48
+ "epoch": 0.23980815347721823,
49
+ "grad_norm": 3.423367738723755,
50
+ "learning_rate": 1.8406608046895816e-05,
51
+ "loss": 2.9791189575195314,
52
+ "step": 300
53
+ },
54
+ {
55
+ "epoch": 0.2797761790567546,
56
+ "grad_norm": 4.398775100708008,
57
+ "learning_rate": 1.8140154543032244e-05,
58
+ "loss": 2.8875311279296874,
59
+ "step": 350
60
+ },
61
+ {
62
+ "epoch": 0.31974420463629094,
63
+ "grad_norm": 3.074233055114746,
64
+ "learning_rate": 1.7873701039168668e-05,
65
+ "loss": 2.837044982910156,
66
+ "step": 400
67
+ },
68
+ {
69
+ "epoch": 0.3597122302158273,
70
+ "grad_norm": 3.2560489177703857,
71
+ "learning_rate": 1.760724753530509e-05,
72
+ "loss": 2.7165292358398436,
73
+ "step": 450
74
+ },
75
+ {
76
+ "epoch": 0.3996802557953637,
77
+ "grad_norm": 3.481449842453003,
78
+ "learning_rate": 1.7340794031441516e-05,
79
+ "loss": 2.6779739379882814,
80
+ "step": 500
81
+ },
82
+ {
83
+ "epoch": 0.4396482813749001,
84
+ "grad_norm": 3.3513357639312744,
85
+ "learning_rate": 1.707434052757794e-05,
86
+ "loss": 2.617684326171875,
87
+ "step": 550
88
+ },
89
+ {
90
+ "epoch": 0.47961630695443647,
91
+ "grad_norm": 3.7497808933258057,
92
+ "learning_rate": 1.6807887023714364e-05,
93
+ "loss": 2.546351013183594,
94
+ "step": 600
95
+ },
96
+ {
97
+ "epoch": 0.5195843325339728,
98
+ "grad_norm": 4.766043663024902,
99
+ "learning_rate": 1.654143351985079e-05,
100
+ "loss": 2.5230949401855467,
101
+ "step": 650
102
+ },
103
+ {
104
+ "epoch": 0.5595523581135092,
105
+ "grad_norm": 4.27366304397583,
106
+ "learning_rate": 1.6274980015987213e-05,
107
+ "loss": 2.4813865661621093,
108
+ "step": 700
109
+ },
110
+ {
111
+ "epoch": 0.5995203836930456,
112
+ "grad_norm": 4.00847864151001,
113
+ "learning_rate": 1.6008526512123637e-05,
114
+ "loss": 2.4810533142089843,
115
+ "step": 750
116
+ },
117
+ {
118
+ "epoch": 0.6394884092725819,
119
+ "grad_norm": 4.365473747253418,
120
+ "learning_rate": 1.5742073008260058e-05,
121
+ "loss": 2.392391357421875,
122
+ "step": 800
123
+ },
124
+ {
125
+ "epoch": 0.6794564348521183,
126
+ "grad_norm": 3.571737289428711,
127
+ "learning_rate": 1.5475619504396485e-05,
128
+ "loss": 2.4272395324707032,
129
+ "step": 850
130
+ },
131
+ {
132
+ "epoch": 0.7194244604316546,
133
+ "grad_norm": 4.288852691650391,
134
+ "learning_rate": 1.520916600053291e-05,
135
+ "loss": 2.320939636230469,
136
+ "step": 900
137
+ },
138
+ {
139
+ "epoch": 0.759392486011191,
140
+ "grad_norm": 3.860504627227783,
141
+ "learning_rate": 1.4942712496669332e-05,
142
+ "loss": 2.423423156738281,
143
+ "step": 950
144
+ },
145
+ {
146
+ "epoch": 0.7993605115907274,
147
+ "grad_norm": 5.249766826629639,
148
+ "learning_rate": 1.4676258992805756e-05,
149
+ "loss": 2.35907470703125,
150
+ "step": 1000
151
+ },
152
+ {
153
+ "epoch": 0.8393285371702638,
154
+ "grad_norm": 5.544083118438721,
155
+ "learning_rate": 1.4409805488942182e-05,
156
+ "loss": 2.3059541320800783,
157
+ "step": 1050
158
+ },
159
+ {
160
+ "epoch": 0.8792965627498002,
161
+ "grad_norm": 3.9598212242126465,
162
+ "learning_rate": 1.4143351985078604e-05,
163
+ "loss": 2.298104248046875,
164
+ "step": 1100
165
+ },
166
+ {
167
+ "epoch": 0.9192645883293366,
168
+ "grad_norm": 5.253625392913818,
169
+ "learning_rate": 1.3876898481215029e-05,
170
+ "loss": 2.3257647705078126,
171
+ "step": 1150
172
+ },
173
+ {
174
+ "epoch": 0.9592326139088729,
175
+ "grad_norm": 6.435204029083252,
176
+ "learning_rate": 1.3610444977351453e-05,
177
+ "loss": 2.2559954833984377,
178
+ "step": 1200
179
+ },
180
+ {
181
+ "epoch": 0.9992006394884093,
182
+ "grad_norm": 4.584034442901611,
183
+ "learning_rate": 1.3343991473487879e-05,
184
+ "loss": 2.224993133544922,
185
+ "step": 1250
186
+ },
187
+ {
188
+ "epoch": 1.0,
189
+ "eval_accuracy": 0.4229731080675797,
190
+ "eval_loss": 2.1663317680358887,
191
+ "eval_macro_f1": 0.42778234953459543,
192
+ "eval_macro_precision": 0.4744399485043334,
193
+ "eval_macro_recall": 0.43904318815193216,
194
+ "eval_runtime": 98.525,
195
+ "eval_samples_per_second": 203.055,
196
+ "eval_steps_per_second": 3.177,
197
+ "step": 1251
198
+ }
199
+ ],
200
+ "logging_steps": 50,
201
+ "max_steps": 3753,
202
+ "num_input_tokens_seen": 0,
203
+ "num_train_epochs": 3,
204
+ "save_steps": 500,
205
+ "stateful_callbacks": {
206
+ "TrainerControl": {
207
+ "args": {
208
+ "should_epoch_stop": false,
209
+ "should_evaluate": false,
210
+ "should_log": false,
211
+ "should_save": true,
212
+ "should_training_stop": false
213
+ },
214
+ "attributes": {}
215
+ }
216
+ },
217
+ "total_flos": 4103579740725360.0,
218
+ "train_batch_size": 64,
219
+ "trial_name": null,
220
+ "trial_params": null
221
+ }
checkpoint-1251/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f18c5dfb7dc4eb65becd7f75437e97d7a12dc629e02e3a7f3952ff22db75c9d
3
+ size 5265
checkpoint-2502/config.json ADDED
@@ -0,0 +1,194 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation": "gelu",
3
+ "architectures": [
4
+ "DistilBertForSequenceClassification"
5
+ ],
6
+ "attention_dropout": 0.1,
7
+ "bos_token_id": null,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "dtype": "float32",
11
+ "eos_token_id": null,
12
+ "hidden_dim": 3072,
13
+ "id2label": {
14
+ "0": "adfgvx",
15
+ "1": "adfgx",
16
+ "2": "aeneas_tacticus",
17
+ "3": "affine",
18
+ "4": "alberti_disk",
19
+ "5": "argenti",
20
+ "6": "arnold_andre",
21
+ "7": "atbash",
22
+ "8": "autokey",
23
+ "9": "babington",
24
+ "10": "bacon_cipher",
25
+ "11": "bazeries",
26
+ "12": "beaufort",
27
+ "13": "bifid",
28
+ "14": "book_cipher",
29
+ "15": "caesar",
30
+ "16": "cardano_autokey",
31
+ "17": "chaocipher",
32
+ "18": "chinese_telegraph",
33
+ "19": "columnar_transposition",
34
+ "20": "commercial_code",
35
+ "21": "confederate_vigenere",
36
+ "22": "copiale",
37
+ "23": "culper_ring",
38
+ "24": "diana",
39
+ "25": "double_transposition",
40
+ "26": "enigma",
41
+ "27": "fialka",
42
+ "28": "four_square",
43
+ "29": "fractionated_morse",
44
+ "30": "geez_monastic",
45
+ "31": "geheimschreiber",
46
+ "32": "great_cipher",
47
+ "33": "gronsfeld",
48
+ "34": "hill",
49
+ "35": "homophonic",
50
+ "36": "jefferson_disk",
51
+ "37": "jn25",
52
+ "38": "joseon_yeokhak",
53
+ "39": "kama_sutra",
54
+ "40": "kl7",
55
+ "41": "kryha",
56
+ "42": "kryptos",
57
+ "43": "lorenz",
58
+ "44": "m209",
59
+ "45": "m94",
60
+ "46": "monoalphabetic",
61
+ "47": "morse_code",
62
+ "48": "navajo_code",
63
+ "49": "nihilist",
64
+ "50": "nomenclator",
65
+ "51": "null_cipher",
66
+ "52": "one_time_pad",
67
+ "53": "pigpen",
68
+ "54": "playfair",
69
+ "55": "polybius",
70
+ "56": "porta",
71
+ "57": "purple",
72
+ "58": "rail_fence",
73
+ "59": "red_type_a",
74
+ "60": "rot13",
75
+ "61": "running_key",
76
+ "62": "scytale",
77
+ "63": "sigaba",
78
+ "64": "slidex",
79
+ "65": "solitaire",
80
+ "66": "stager_route",
81
+ "67": "straddling_checkerboard",
82
+ "68": "tap_code",
83
+ "69": "trifid",
84
+ "70": "trithemius",
85
+ "71": "two_square",
86
+ "72": "typex",
87
+ "73": "venona_pad_reuse",
88
+ "74": "vernam",
89
+ "75": "vic",
90
+ "76": "vigenere",
91
+ "77": "voynich_render",
92
+ "78": "wallis_cipher",
93
+ "79": "wheatstone",
94
+ "80": "zimmermann"
95
+ },
96
+ "initializer_range": 0.02,
97
+ "label2id": {
98
+ "adfgvx": 0,
99
+ "adfgx": 1,
100
+ "aeneas_tacticus": 2,
101
+ "affine": 3,
102
+ "alberti_disk": 4,
103
+ "argenti": 5,
104
+ "arnold_andre": 6,
105
+ "atbash": 7,
106
+ "autokey": 8,
107
+ "babington": 9,
108
+ "bacon_cipher": 10,
109
+ "bazeries": 11,
110
+ "beaufort": 12,
111
+ "bifid": 13,
112
+ "book_cipher": 14,
113
+ "caesar": 15,
114
+ "cardano_autokey": 16,
115
+ "chaocipher": 17,
116
+ "chinese_telegraph": 18,
117
+ "columnar_transposition": 19,
118
+ "commercial_code": 20,
119
+ "confederate_vigenere": 21,
120
+ "copiale": 22,
121
+ "culper_ring": 23,
122
+ "diana": 24,
123
+ "double_transposition": 25,
124
+ "enigma": 26,
125
+ "fialka": 27,
126
+ "four_square": 28,
127
+ "fractionated_morse": 29,
128
+ "geez_monastic": 30,
129
+ "geheimschreiber": 31,
130
+ "great_cipher": 32,
131
+ "gronsfeld": 33,
132
+ "hill": 34,
133
+ "homophonic": 35,
134
+ "jefferson_disk": 36,
135
+ "jn25": 37,
136
+ "joseon_yeokhak": 38,
137
+ "kama_sutra": 39,
138
+ "kl7": 40,
139
+ "kryha": 41,
140
+ "kryptos": 42,
141
+ "lorenz": 43,
142
+ "m209": 44,
143
+ "m94": 45,
144
+ "monoalphabetic": 46,
145
+ "morse_code": 47,
146
+ "navajo_code": 48,
147
+ "nihilist": 49,
148
+ "nomenclator": 50,
149
+ "null_cipher": 51,
150
+ "one_time_pad": 52,
151
+ "pigpen": 53,
152
+ "playfair": 54,
153
+ "polybius": 55,
154
+ "porta": 56,
155
+ "purple": 57,
156
+ "rail_fence": 58,
157
+ "red_type_a": 59,
158
+ "rot13": 60,
159
+ "running_key": 61,
160
+ "scytale": 62,
161
+ "sigaba": 63,
162
+ "slidex": 64,
163
+ "solitaire": 65,
164
+ "stager_route": 66,
165
+ "straddling_checkerboard": 67,
166
+ "tap_code": 68,
167
+ "trifid": 69,
168
+ "trithemius": 70,
169
+ "two_square": 71,
170
+ "typex": 72,
171
+ "venona_pad_reuse": 73,
172
+ "vernam": 74,
173
+ "vic": 75,
174
+ "vigenere": 76,
175
+ "voynich_render": 77,
176
+ "wallis_cipher": 78,
177
+ "wheatstone": 79,
178
+ "zimmermann": 80
179
+ },
180
+ "max_position_embeddings": 512,
181
+ "model_type": "distilbert",
182
+ "n_heads": 12,
183
+ "n_layers": 6,
184
+ "pad_token_id": 0,
185
+ "problem_type": "single_label_classification",
186
+ "qa_dropout": 0.1,
187
+ "seq_classif_dropout": 0.2,
188
+ "sinusoidal_pos_embds": false,
189
+ "tie_weights_": true,
190
+ "tie_word_embeddings": true,
191
+ "transformers_version": "5.8.0",
192
+ "use_cache": false,
193
+ "vocab_size": 30522
194
+ }
checkpoint-2502/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7cbe8dcfb3092f726f4bdd2722d285fde06b135d5e08ca9930983072949cc56
3
+ size 268075580
checkpoint-2502/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8fbf0c94ab4a504744982fba8192ccaafd523af51bb715cf508c4762a2e7d458
3
+ size 536215243
checkpoint-2502/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ff0c275148b013a74e746c28106321b7c590e47ffb8aa1e2594086a28dcb21f
3
+ size 14645
checkpoint-2502/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a22124a458473173e737464cfba92b278fdba3f6440e4296352f5edec9e5dc5
3
+ size 1465
checkpoint-2502/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-2502/tokenizer_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "cls_token": "[CLS]",
4
+ "do_lower_case": true,
5
+ "is_local": false,
6
+ "local_files_only": false,
7
+ "mask_token": "[MASK]",
8
+ "model_max_length": 512,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "strip_accents": null,
12
+ "tokenize_chinese_chars": true,
13
+ "tokenizer_class": "BertTokenizer",
14
+ "unk_token": "[UNK]"
15
+ }
checkpoint-2502/trainer_state.json ADDED
@@ -0,0 +1,408 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 2502,
3
+ "best_metric": 0.5002139201674197,
4
+ "best_model_checkpoint": "/tmp/cipher_model/checkpoint-2502",
5
+ "epoch": 2.0,
6
+ "eval_steps": 500,
7
+ "global_step": 2502,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.03996802557953637,
14
+ "grad_norm": 2.1291019916534424,
15
+ "learning_rate": 1.97388755662137e-05,
16
+ "loss": 4.206104736328125,
17
+ "step": 50
18
+ },
19
+ {
20
+ "epoch": 0.07993605115907274,
21
+ "grad_norm": 2.337212085723877,
22
+ "learning_rate": 1.9472422062350123e-05,
23
+ "loss": 3.785872802734375,
24
+ "step": 100
25
+ },
26
+ {
27
+ "epoch": 0.11990407673860912,
28
+ "grad_norm": 2.5010275840759277,
29
+ "learning_rate": 1.9205968558486547e-05,
30
+ "loss": 3.458525390625,
31
+ "step": 150
32
+ },
33
+ {
34
+ "epoch": 0.15987210231814547,
35
+ "grad_norm": 3.1468915939331055,
36
+ "learning_rate": 1.893951505462297e-05,
37
+ "loss": 3.239898681640625,
38
+ "step": 200
39
+ },
40
+ {
41
+ "epoch": 0.19984012789768185,
42
+ "grad_norm": 2.6385934352874756,
43
+ "learning_rate": 1.8673061550759395e-05,
44
+ "loss": 3.1392282104492186,
45
+ "step": 250
46
+ },
47
+ {
48
+ "epoch": 0.23980815347721823,
49
+ "grad_norm": 3.423367738723755,
50
+ "learning_rate": 1.8406608046895816e-05,
51
+ "loss": 2.9791189575195314,
52
+ "step": 300
53
+ },
54
+ {
55
+ "epoch": 0.2797761790567546,
56
+ "grad_norm": 4.398775100708008,
57
+ "learning_rate": 1.8140154543032244e-05,
58
+ "loss": 2.8875311279296874,
59
+ "step": 350
60
+ },
61
+ {
62
+ "epoch": 0.31974420463629094,
63
+ "grad_norm": 3.074233055114746,
64
+ "learning_rate": 1.7873701039168668e-05,
65
+ "loss": 2.837044982910156,
66
+ "step": 400
67
+ },
68
+ {
69
+ "epoch": 0.3597122302158273,
70
+ "grad_norm": 3.2560489177703857,
71
+ "learning_rate": 1.760724753530509e-05,
72
+ "loss": 2.7165292358398436,
73
+ "step": 450
74
+ },
75
+ {
76
+ "epoch": 0.3996802557953637,
77
+ "grad_norm": 3.481449842453003,
78
+ "learning_rate": 1.7340794031441516e-05,
79
+ "loss": 2.6779739379882814,
80
+ "step": 500
81
+ },
82
+ {
83
+ "epoch": 0.4396482813749001,
84
+ "grad_norm": 3.3513357639312744,
85
+ "learning_rate": 1.707434052757794e-05,
86
+ "loss": 2.617684326171875,
87
+ "step": 550
88
+ },
89
+ {
90
+ "epoch": 0.47961630695443647,
91
+ "grad_norm": 3.7497808933258057,
92
+ "learning_rate": 1.6807887023714364e-05,
93
+ "loss": 2.546351013183594,
94
+ "step": 600
95
+ },
96
+ {
97
+ "epoch": 0.5195843325339728,
98
+ "grad_norm": 4.766043663024902,
99
+ "learning_rate": 1.654143351985079e-05,
100
+ "loss": 2.5230949401855467,
101
+ "step": 650
102
+ },
103
+ {
104
+ "epoch": 0.5595523581135092,
105
+ "grad_norm": 4.27366304397583,
106
+ "learning_rate": 1.6274980015987213e-05,
107
+ "loss": 2.4813865661621093,
108
+ "step": 700
109
+ },
110
+ {
111
+ "epoch": 0.5995203836930456,
112
+ "grad_norm": 4.00847864151001,
113
+ "learning_rate": 1.6008526512123637e-05,
114
+ "loss": 2.4810533142089843,
115
+ "step": 750
116
+ },
117
+ {
118
+ "epoch": 0.6394884092725819,
119
+ "grad_norm": 4.365473747253418,
120
+ "learning_rate": 1.5742073008260058e-05,
121
+ "loss": 2.392391357421875,
122
+ "step": 800
123
+ },
124
+ {
125
+ "epoch": 0.6794564348521183,
126
+ "grad_norm": 3.571737289428711,
127
+ "learning_rate": 1.5475619504396485e-05,
128
+ "loss": 2.4272395324707032,
129
+ "step": 850
130
+ },
131
+ {
132
+ "epoch": 0.7194244604316546,
133
+ "grad_norm": 4.288852691650391,
134
+ "learning_rate": 1.520916600053291e-05,
135
+ "loss": 2.320939636230469,
136
+ "step": 900
137
+ },
138
+ {
139
+ "epoch": 0.759392486011191,
140
+ "grad_norm": 3.860504627227783,
141
+ "learning_rate": 1.4942712496669332e-05,
142
+ "loss": 2.423423156738281,
143
+ "step": 950
144
+ },
145
+ {
146
+ "epoch": 0.7993605115907274,
147
+ "grad_norm": 5.249766826629639,
148
+ "learning_rate": 1.4676258992805756e-05,
149
+ "loss": 2.35907470703125,
150
+ "step": 1000
151
+ },
152
+ {
153
+ "epoch": 0.8393285371702638,
154
+ "grad_norm": 5.544083118438721,
155
+ "learning_rate": 1.4409805488942182e-05,
156
+ "loss": 2.3059541320800783,
157
+ "step": 1050
158
+ },
159
+ {
160
+ "epoch": 0.8792965627498002,
161
+ "grad_norm": 3.9598212242126465,
162
+ "learning_rate": 1.4143351985078604e-05,
163
+ "loss": 2.298104248046875,
164
+ "step": 1100
165
+ },
166
+ {
167
+ "epoch": 0.9192645883293366,
168
+ "grad_norm": 5.253625392913818,
169
+ "learning_rate": 1.3876898481215029e-05,
170
+ "loss": 2.3257647705078126,
171
+ "step": 1150
172
+ },
173
+ {
174
+ "epoch": 0.9592326139088729,
175
+ "grad_norm": 6.435204029083252,
176
+ "learning_rate": 1.3610444977351453e-05,
177
+ "loss": 2.2559954833984377,
178
+ "step": 1200
179
+ },
180
+ {
181
+ "epoch": 0.9992006394884093,
182
+ "grad_norm": 4.584034442901611,
183
+ "learning_rate": 1.3343991473487879e-05,
184
+ "loss": 2.224993133544922,
185
+ "step": 1250
186
+ },
187
+ {
188
+ "epoch": 1.0,
189
+ "eval_accuracy": 0.4229731080675797,
190
+ "eval_loss": 2.1663317680358887,
191
+ "eval_macro_f1": 0.42778234953459543,
192
+ "eval_macro_precision": 0.4744399485043334,
193
+ "eval_macro_recall": 0.43904318815193216,
194
+ "eval_runtime": 98.525,
195
+ "eval_samples_per_second": 203.055,
196
+ "eval_steps_per_second": 3.177,
197
+ "step": 1251
198
+ },
199
+ {
200
+ "epoch": 1.0391686650679457,
201
+ "grad_norm": 4.706632614135742,
202
+ "learning_rate": 1.3077537969624301e-05,
203
+ "loss": 2.1520008850097656,
204
+ "step": 1300
205
+ },
206
+ {
207
+ "epoch": 1.079136690647482,
208
+ "grad_norm": 4.258380889892578,
209
+ "learning_rate": 1.2811084465760725e-05,
210
+ "loss": 2.1391026306152345,
211
+ "step": 1350
212
+ },
213
+ {
214
+ "epoch": 1.1191047162270185,
215
+ "grad_norm": 5.545387268066406,
216
+ "learning_rate": 1.2544630961897151e-05,
217
+ "loss": 2.1541981506347656,
218
+ "step": 1400
219
+ },
220
+ {
221
+ "epoch": 1.1590727418065547,
222
+ "grad_norm": 4.906599998474121,
223
+ "learning_rate": 1.2278177458033574e-05,
224
+ "loss": 2.1915689086914063,
225
+ "step": 1450
226
+ },
227
+ {
228
+ "epoch": 1.1990407673860912,
229
+ "grad_norm": 4.406611442565918,
230
+ "learning_rate": 1.2011723954169998e-05,
231
+ "loss": 2.1513375854492187,
232
+ "step": 1500
233
+ },
234
+ {
235
+ "epoch": 1.2390087929656275,
236
+ "grad_norm": 4.207244873046875,
237
+ "learning_rate": 1.1745270450306424e-05,
238
+ "loss": 2.146766815185547,
239
+ "step": 1550
240
+ },
241
+ {
242
+ "epoch": 1.2789768185451638,
243
+ "grad_norm": 5.66338586807251,
244
+ "learning_rate": 1.1478816946442846e-05,
245
+ "loss": 2.1281501770019533,
246
+ "step": 1600
247
+ },
248
+ {
249
+ "epoch": 1.3189448441247003,
250
+ "grad_norm": 5.967006206512451,
251
+ "learning_rate": 1.121236344257927e-05,
252
+ "loss": 2.115397644042969,
253
+ "step": 1650
254
+ },
255
+ {
256
+ "epoch": 1.3589128697042367,
257
+ "grad_norm": 4.559199333190918,
258
+ "learning_rate": 1.0945909938715694e-05,
259
+ "loss": 2.1187815856933594,
260
+ "step": 1700
261
+ },
262
+ {
263
+ "epoch": 1.398880895283773,
264
+ "grad_norm": 6.57089900970459,
265
+ "learning_rate": 1.067945643485212e-05,
266
+ "loss": 2.127105407714844,
267
+ "step": 1750
268
+ },
269
+ {
270
+ "epoch": 1.4388489208633093,
271
+ "grad_norm": 5.105190277099609,
272
+ "learning_rate": 1.0413002930988543e-05,
273
+ "loss": 2.080175323486328,
274
+ "step": 1800
275
+ },
276
+ {
277
+ "epoch": 1.4788169464428458,
278
+ "grad_norm": 6.654059886932373,
279
+ "learning_rate": 1.0146549427124967e-05,
280
+ "loss": 2.0765325927734377,
281
+ "step": 1850
282
+ },
283
+ {
284
+ "epoch": 1.518784972022382,
285
+ "grad_norm": 4.843920707702637,
286
+ "learning_rate": 9.880095923261391e-06,
287
+ "loss": 2.078269195556641,
288
+ "step": 1900
289
+ },
290
+ {
291
+ "epoch": 1.5587529976019185,
292
+ "grad_norm": 5.037885665893555,
293
+ "learning_rate": 9.613642419397815e-06,
294
+ "loss": 2.0772938537597656,
295
+ "step": 1950
296
+ },
297
+ {
298
+ "epoch": 1.5987210231814548,
299
+ "grad_norm": 6.010105133056641,
300
+ "learning_rate": 9.34718891553424e-06,
301
+ "loss": 2.07054931640625,
302
+ "step": 2000
303
+ },
304
+ {
305
+ "epoch": 1.638689048760991,
306
+ "grad_norm": 4.658566474914551,
307
+ "learning_rate": 9.080735411670665e-06,
308
+ "loss": 2.0459413146972656,
309
+ "step": 2050
310
+ },
311
+ {
312
+ "epoch": 1.6786570743405276,
313
+ "grad_norm": 5.411528587341309,
314
+ "learning_rate": 8.814281907807088e-06,
315
+ "loss": 2.0509548950195313,
316
+ "step": 2100
317
+ },
318
+ {
319
+ "epoch": 1.718625099920064,
320
+ "grad_norm": 6.495298385620117,
321
+ "learning_rate": 8.547828403943512e-06,
322
+ "loss": 2.0080926513671873,
323
+ "step": 2150
324
+ },
325
+ {
326
+ "epoch": 1.7585931254996003,
327
+ "grad_norm": 7.016653537750244,
328
+ "learning_rate": 8.281374900079936e-06,
329
+ "loss": 2.0153048706054686,
330
+ "step": 2200
331
+ },
332
+ {
333
+ "epoch": 1.7985611510791366,
334
+ "grad_norm": 5.021507263183594,
335
+ "learning_rate": 8.01492139621636e-06,
336
+ "loss": 2.055178680419922,
337
+ "step": 2250
338
+ },
339
+ {
340
+ "epoch": 1.838529176658673,
341
+ "grad_norm": 5.355196475982666,
342
+ "learning_rate": 7.748467892352786e-06,
343
+ "loss": 2.0198463439941405,
344
+ "step": 2300
345
+ },
346
+ {
347
+ "epoch": 1.8784972022382096,
348
+ "grad_norm": 6.557318687438965,
349
+ "learning_rate": 7.48201438848921e-06,
350
+ "loss": 2.0120050048828126,
351
+ "step": 2350
352
+ },
353
+ {
354
+ "epoch": 1.9184652278177459,
355
+ "grad_norm": 6.802421569824219,
356
+ "learning_rate": 7.215560884625633e-06,
357
+ "loss": 2.03742431640625,
358
+ "step": 2400
359
+ },
360
+ {
361
+ "epoch": 1.9584332533972821,
362
+ "grad_norm": 6.564506530761719,
363
+ "learning_rate": 6.949107380762058e-06,
364
+ "loss": 1.9714970397949219,
365
+ "step": 2450
366
+ },
367
+ {
368
+ "epoch": 1.9984012789768184,
369
+ "grad_norm": 6.481470108032227,
370
+ "learning_rate": 6.682653876898481e-06,
371
+ "loss": 2.002969512939453,
372
+ "step": 2500
373
+ },
374
+ {
375
+ "epoch": 2.0,
376
+ "eval_accuracy": 0.4633110066979906,
377
+ "eval_loss": 1.9806830883026123,
378
+ "eval_macro_f1": 0.5002139201674197,
379
+ "eval_macro_precision": 0.5580379243961209,
380
+ "eval_macro_recall": 0.49887820516729536,
381
+ "eval_runtime": 98.8336,
382
+ "eval_samples_per_second": 202.421,
383
+ "eval_steps_per_second": 3.167,
384
+ "step": 2502
385
+ }
386
+ ],
387
+ "logging_steps": 50,
388
+ "max_steps": 3753,
389
+ "num_input_tokens_seen": 0,
390
+ "num_train_epochs": 3,
391
+ "save_steps": 500,
392
+ "stateful_callbacks": {
393
+ "TrainerControl": {
394
+ "args": {
395
+ "should_epoch_stop": false,
396
+ "should_evaluate": false,
397
+ "should_log": false,
398
+ "should_save": true,
399
+ "should_training_stop": false
400
+ },
401
+ "attributes": {}
402
+ }
403
+ },
404
+ "total_flos": 8192407939245936.0,
405
+ "train_batch_size": 64,
406
+ "trial_name": null,
407
+ "trial_params": null
408
+ }
checkpoint-2502/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f18c5dfb7dc4eb65becd7f75437e97d7a12dc629e02e3a7f3952ff22db75c9d
3
+ size 5265
checkpoint-3753/config.json ADDED
@@ -0,0 +1,194 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation": "gelu",
3
+ "architectures": [
4
+ "DistilBertForSequenceClassification"
5
+ ],
6
+ "attention_dropout": 0.1,
7
+ "bos_token_id": null,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "dtype": "float32",
11
+ "eos_token_id": null,
12
+ "hidden_dim": 3072,
13
+ "id2label": {
14
+ "0": "adfgvx",
15
+ "1": "adfgx",
16
+ "2": "aeneas_tacticus",
17
+ "3": "affine",
18
+ "4": "alberti_disk",
19
+ "5": "argenti",
20
+ "6": "arnold_andre",
21
+ "7": "atbash",
22
+ "8": "autokey",
23
+ "9": "babington",
24
+ "10": "bacon_cipher",
25
+ "11": "bazeries",
26
+ "12": "beaufort",
27
+ "13": "bifid",
28
+ "14": "book_cipher",
29
+ "15": "caesar",
30
+ "16": "cardano_autokey",
31
+ "17": "chaocipher",
32
+ "18": "chinese_telegraph",
33
+ "19": "columnar_transposition",
34
+ "20": "commercial_code",
35
+ "21": "confederate_vigenere",
36
+ "22": "copiale",
37
+ "23": "culper_ring",
38
+ "24": "diana",
39
+ "25": "double_transposition",
40
+ "26": "enigma",
41
+ "27": "fialka",
42
+ "28": "four_square",
43
+ "29": "fractionated_morse",
44
+ "30": "geez_monastic",
45
+ "31": "geheimschreiber",
46
+ "32": "great_cipher",
47
+ "33": "gronsfeld",
48
+ "34": "hill",
49
+ "35": "homophonic",
50
+ "36": "jefferson_disk",
51
+ "37": "jn25",
52
+ "38": "joseon_yeokhak",
53
+ "39": "kama_sutra",
54
+ "40": "kl7",
55
+ "41": "kryha",
56
+ "42": "kryptos",
57
+ "43": "lorenz",
58
+ "44": "m209",
59
+ "45": "m94",
60
+ "46": "monoalphabetic",
61
+ "47": "morse_code",
62
+ "48": "navajo_code",
63
+ "49": "nihilist",
64
+ "50": "nomenclator",
65
+ "51": "null_cipher",
66
+ "52": "one_time_pad",
67
+ "53": "pigpen",
68
+ "54": "playfair",
69
+ "55": "polybius",
70
+ "56": "porta",
71
+ "57": "purple",
72
+ "58": "rail_fence",
73
+ "59": "red_type_a",
74
+ "60": "rot13",
75
+ "61": "running_key",
76
+ "62": "scytale",
77
+ "63": "sigaba",
78
+ "64": "slidex",
79
+ "65": "solitaire",
80
+ "66": "stager_route",
81
+ "67": "straddling_checkerboard",
82
+ "68": "tap_code",
83
+ "69": "trifid",
84
+ "70": "trithemius",
85
+ "71": "two_square",
86
+ "72": "typex",
87
+ "73": "venona_pad_reuse",
88
+ "74": "vernam",
89
+ "75": "vic",
90
+ "76": "vigenere",
91
+ "77": "voynich_render",
92
+ "78": "wallis_cipher",
93
+ "79": "wheatstone",
94
+ "80": "zimmermann"
95
+ },
96
+ "initializer_range": 0.02,
97
+ "label2id": {
98
+ "adfgvx": 0,
99
+ "adfgx": 1,
100
+ "aeneas_tacticus": 2,
101
+ "affine": 3,
102
+ "alberti_disk": 4,
103
+ "argenti": 5,
104
+ "arnold_andre": 6,
105
+ "atbash": 7,
106
+ "autokey": 8,
107
+ "babington": 9,
108
+ "bacon_cipher": 10,
109
+ "bazeries": 11,
110
+ "beaufort": 12,
111
+ "bifid": 13,
112
+ "book_cipher": 14,
113
+ "caesar": 15,
114
+ "cardano_autokey": 16,
115
+ "chaocipher": 17,
116
+ "chinese_telegraph": 18,
117
+ "columnar_transposition": 19,
118
+ "commercial_code": 20,
119
+ "confederate_vigenere": 21,
120
+ "copiale": 22,
121
+ "culper_ring": 23,
122
+ "diana": 24,
123
+ "double_transposition": 25,
124
+ "enigma": 26,
125
+ "fialka": 27,
126
+ "four_square": 28,
127
+ "fractionated_morse": 29,
128
+ "geez_monastic": 30,
129
+ "geheimschreiber": 31,
130
+ "great_cipher": 32,
131
+ "gronsfeld": 33,
132
+ "hill": 34,
133
+ "homophonic": 35,
134
+ "jefferson_disk": 36,
135
+ "jn25": 37,
136
+ "joseon_yeokhak": 38,
137
+ "kama_sutra": 39,
138
+ "kl7": 40,
139
+ "kryha": 41,
140
+ "kryptos": 42,
141
+ "lorenz": 43,
142
+ "m209": 44,
143
+ "m94": 45,
144
+ "monoalphabetic": 46,
145
+ "morse_code": 47,
146
+ "navajo_code": 48,
147
+ "nihilist": 49,
148
+ "nomenclator": 50,
149
+ "null_cipher": 51,
150
+ "one_time_pad": 52,
151
+ "pigpen": 53,
152
+ "playfair": 54,
153
+ "polybius": 55,
154
+ "porta": 56,
155
+ "purple": 57,
156
+ "rail_fence": 58,
157
+ "red_type_a": 59,
158
+ "rot13": 60,
159
+ "running_key": 61,
160
+ "scytale": 62,
161
+ "sigaba": 63,
162
+ "slidex": 64,
163
+ "solitaire": 65,
164
+ "stager_route": 66,
165
+ "straddling_checkerboard": 67,
166
+ "tap_code": 68,
167
+ "trifid": 69,
168
+ "trithemius": 70,
169
+ "two_square": 71,
170
+ "typex": 72,
171
+ "venona_pad_reuse": 73,
172
+ "vernam": 74,
173
+ "vic": 75,
174
+ "vigenere": 76,
175
+ "voynich_render": 77,
176
+ "wallis_cipher": 78,
177
+ "wheatstone": 79,
178
+ "zimmermann": 80
179
+ },
180
+ "max_position_embeddings": 512,
181
+ "model_type": "distilbert",
182
+ "n_heads": 12,
183
+ "n_layers": 6,
184
+ "pad_token_id": 0,
185
+ "problem_type": "single_label_classification",
186
+ "qa_dropout": 0.1,
187
+ "seq_classif_dropout": 0.2,
188
+ "sinusoidal_pos_embds": false,
189
+ "tie_weights_": true,
190
+ "tie_word_embeddings": true,
191
+ "transformers_version": "5.8.0",
192
+ "use_cache": false,
193
+ "vocab_size": 30522
194
+ }
checkpoint-3753/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2943abbfbe864d54f7dd6febfe1e9f33c4aa5e49ac2c2e840204148469537444
3
+ size 268075580
checkpoint-3753/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c58174614b8ea9771bed55b45a451c18b2c80e511e64c5c56089986420df937e
3
+ size 536215243
checkpoint-3753/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1bf1afdfee0d3afdff955f0267833c42ba5ca9dd882983adfb7a1a6086fe59a5
3
+ size 14645
checkpoint-3753/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11fc4f7050e8d3a59f201411117d5c1f5563dada2852c11087f446f80525fde3
3
+ size 1465
checkpoint-3753/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-3753/tokenizer_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "cls_token": "[CLS]",
4
+ "do_lower_case": true,
5
+ "is_local": false,
6
+ "local_files_only": false,
7
+ "mask_token": "[MASK]",
8
+ "model_max_length": 512,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "strip_accents": null,
12
+ "tokenize_chinese_chars": true,
13
+ "tokenizer_class": "BertTokenizer",
14
+ "unk_token": "[UNK]"
15
+ }
checkpoint-3753/trainer_state.json ADDED
@@ -0,0 +1,595 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 3753,
3
+ "best_metric": 0.5181542733976486,
4
+ "best_model_checkpoint": "/tmp/cipher_model/checkpoint-3753",
5
+ "epoch": 3.0,
6
+ "eval_steps": 500,
7
+ "global_step": 3753,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.03996802557953637,
14
+ "grad_norm": 2.1291019916534424,
15
+ "learning_rate": 1.97388755662137e-05,
16
+ "loss": 4.206104736328125,
17
+ "step": 50
18
+ },
19
+ {
20
+ "epoch": 0.07993605115907274,
21
+ "grad_norm": 2.337212085723877,
22
+ "learning_rate": 1.9472422062350123e-05,
23
+ "loss": 3.785872802734375,
24
+ "step": 100
25
+ },
26
+ {
27
+ "epoch": 0.11990407673860912,
28
+ "grad_norm": 2.5010275840759277,
29
+ "learning_rate": 1.9205968558486547e-05,
30
+ "loss": 3.458525390625,
31
+ "step": 150
32
+ },
33
+ {
34
+ "epoch": 0.15987210231814547,
35
+ "grad_norm": 3.1468915939331055,
36
+ "learning_rate": 1.893951505462297e-05,
37
+ "loss": 3.239898681640625,
38
+ "step": 200
39
+ },
40
+ {
41
+ "epoch": 0.19984012789768185,
42
+ "grad_norm": 2.6385934352874756,
43
+ "learning_rate": 1.8673061550759395e-05,
44
+ "loss": 3.1392282104492186,
45
+ "step": 250
46
+ },
47
+ {
48
+ "epoch": 0.23980815347721823,
49
+ "grad_norm": 3.423367738723755,
50
+ "learning_rate": 1.8406608046895816e-05,
51
+ "loss": 2.9791189575195314,
52
+ "step": 300
53
+ },
54
+ {
55
+ "epoch": 0.2797761790567546,
56
+ "grad_norm": 4.398775100708008,
57
+ "learning_rate": 1.8140154543032244e-05,
58
+ "loss": 2.8875311279296874,
59
+ "step": 350
60
+ },
61
+ {
62
+ "epoch": 0.31974420463629094,
63
+ "grad_norm": 3.074233055114746,
64
+ "learning_rate": 1.7873701039168668e-05,
65
+ "loss": 2.837044982910156,
66
+ "step": 400
67
+ },
68
+ {
69
+ "epoch": 0.3597122302158273,
70
+ "grad_norm": 3.2560489177703857,
71
+ "learning_rate": 1.760724753530509e-05,
72
+ "loss": 2.7165292358398436,
73
+ "step": 450
74
+ },
75
+ {
76
+ "epoch": 0.3996802557953637,
77
+ "grad_norm": 3.481449842453003,
78
+ "learning_rate": 1.7340794031441516e-05,
79
+ "loss": 2.6779739379882814,
80
+ "step": 500
81
+ },
82
+ {
83
+ "epoch": 0.4396482813749001,
84
+ "grad_norm": 3.3513357639312744,
85
+ "learning_rate": 1.707434052757794e-05,
86
+ "loss": 2.617684326171875,
87
+ "step": 550
88
+ },
89
+ {
90
+ "epoch": 0.47961630695443647,
91
+ "grad_norm": 3.7497808933258057,
92
+ "learning_rate": 1.6807887023714364e-05,
93
+ "loss": 2.546351013183594,
94
+ "step": 600
95
+ },
96
+ {
97
+ "epoch": 0.5195843325339728,
98
+ "grad_norm": 4.766043663024902,
99
+ "learning_rate": 1.654143351985079e-05,
100
+ "loss": 2.5230949401855467,
101
+ "step": 650
102
+ },
103
+ {
104
+ "epoch": 0.5595523581135092,
105
+ "grad_norm": 4.27366304397583,
106
+ "learning_rate": 1.6274980015987213e-05,
107
+ "loss": 2.4813865661621093,
108
+ "step": 700
109
+ },
110
+ {
111
+ "epoch": 0.5995203836930456,
112
+ "grad_norm": 4.00847864151001,
113
+ "learning_rate": 1.6008526512123637e-05,
114
+ "loss": 2.4810533142089843,
115
+ "step": 750
116
+ },
117
+ {
118
+ "epoch": 0.6394884092725819,
119
+ "grad_norm": 4.365473747253418,
120
+ "learning_rate": 1.5742073008260058e-05,
121
+ "loss": 2.392391357421875,
122
+ "step": 800
123
+ },
124
+ {
125
+ "epoch": 0.6794564348521183,
126
+ "grad_norm": 3.571737289428711,
127
+ "learning_rate": 1.5475619504396485e-05,
128
+ "loss": 2.4272395324707032,
129
+ "step": 850
130
+ },
131
+ {
132
+ "epoch": 0.7194244604316546,
133
+ "grad_norm": 4.288852691650391,
134
+ "learning_rate": 1.520916600053291e-05,
135
+ "loss": 2.320939636230469,
136
+ "step": 900
137
+ },
138
+ {
139
+ "epoch": 0.759392486011191,
140
+ "grad_norm": 3.860504627227783,
141
+ "learning_rate": 1.4942712496669332e-05,
142
+ "loss": 2.423423156738281,
143
+ "step": 950
144
+ },
145
+ {
146
+ "epoch": 0.7993605115907274,
147
+ "grad_norm": 5.249766826629639,
148
+ "learning_rate": 1.4676258992805756e-05,
149
+ "loss": 2.35907470703125,
150
+ "step": 1000
151
+ },
152
+ {
153
+ "epoch": 0.8393285371702638,
154
+ "grad_norm": 5.544083118438721,
155
+ "learning_rate": 1.4409805488942182e-05,
156
+ "loss": 2.3059541320800783,
157
+ "step": 1050
158
+ },
159
+ {
160
+ "epoch": 0.8792965627498002,
161
+ "grad_norm": 3.9598212242126465,
162
+ "learning_rate": 1.4143351985078604e-05,
163
+ "loss": 2.298104248046875,
164
+ "step": 1100
165
+ },
166
+ {
167
+ "epoch": 0.9192645883293366,
168
+ "grad_norm": 5.253625392913818,
169
+ "learning_rate": 1.3876898481215029e-05,
170
+ "loss": 2.3257647705078126,
171
+ "step": 1150
172
+ },
173
+ {
174
+ "epoch": 0.9592326139088729,
175
+ "grad_norm": 6.435204029083252,
176
+ "learning_rate": 1.3610444977351453e-05,
177
+ "loss": 2.2559954833984377,
178
+ "step": 1200
179
+ },
180
+ {
181
+ "epoch": 0.9992006394884093,
182
+ "grad_norm": 4.584034442901611,
183
+ "learning_rate": 1.3343991473487879e-05,
184
+ "loss": 2.224993133544922,
185
+ "step": 1250
186
+ },
187
+ {
188
+ "epoch": 1.0,
189
+ "eval_accuracy": 0.4229731080675797,
190
+ "eval_loss": 2.1663317680358887,
191
+ "eval_macro_f1": 0.42778234953459543,
192
+ "eval_macro_precision": 0.4744399485043334,
193
+ "eval_macro_recall": 0.43904318815193216,
194
+ "eval_runtime": 98.525,
195
+ "eval_samples_per_second": 203.055,
196
+ "eval_steps_per_second": 3.177,
197
+ "step": 1251
198
+ },
199
+ {
200
+ "epoch": 1.0391686650679457,
201
+ "grad_norm": 4.706632614135742,
202
+ "learning_rate": 1.3077537969624301e-05,
203
+ "loss": 2.1520008850097656,
204
+ "step": 1300
205
+ },
206
+ {
207
+ "epoch": 1.079136690647482,
208
+ "grad_norm": 4.258380889892578,
209
+ "learning_rate": 1.2811084465760725e-05,
210
+ "loss": 2.1391026306152345,
211
+ "step": 1350
212
+ },
213
+ {
214
+ "epoch": 1.1191047162270185,
215
+ "grad_norm": 5.545387268066406,
216
+ "learning_rate": 1.2544630961897151e-05,
217
+ "loss": 2.1541981506347656,
218
+ "step": 1400
219
+ },
220
+ {
221
+ "epoch": 1.1590727418065547,
222
+ "grad_norm": 4.906599998474121,
223
+ "learning_rate": 1.2278177458033574e-05,
224
+ "loss": 2.1915689086914063,
225
+ "step": 1450
226
+ },
227
+ {
228
+ "epoch": 1.1990407673860912,
229
+ "grad_norm": 4.406611442565918,
230
+ "learning_rate": 1.2011723954169998e-05,
231
+ "loss": 2.1513375854492187,
232
+ "step": 1500
233
+ },
234
+ {
235
+ "epoch": 1.2390087929656275,
236
+ "grad_norm": 4.207244873046875,
237
+ "learning_rate": 1.1745270450306424e-05,
238
+ "loss": 2.146766815185547,
239
+ "step": 1550
240
+ },
241
+ {
242
+ "epoch": 1.2789768185451638,
243
+ "grad_norm": 5.66338586807251,
244
+ "learning_rate": 1.1478816946442846e-05,
245
+ "loss": 2.1281501770019533,
246
+ "step": 1600
247
+ },
248
+ {
249
+ "epoch": 1.3189448441247003,
250
+ "grad_norm": 5.967006206512451,
251
+ "learning_rate": 1.121236344257927e-05,
252
+ "loss": 2.115397644042969,
253
+ "step": 1650
254
+ },
255
+ {
256
+ "epoch": 1.3589128697042367,
257
+ "grad_norm": 4.559199333190918,
258
+ "learning_rate": 1.0945909938715694e-05,
259
+ "loss": 2.1187815856933594,
260
+ "step": 1700
261
+ },
262
+ {
263
+ "epoch": 1.398880895283773,
264
+ "grad_norm": 6.57089900970459,
265
+ "learning_rate": 1.067945643485212e-05,
266
+ "loss": 2.127105407714844,
267
+ "step": 1750
268
+ },
269
+ {
270
+ "epoch": 1.4388489208633093,
271
+ "grad_norm": 5.105190277099609,
272
+ "learning_rate": 1.0413002930988543e-05,
273
+ "loss": 2.080175323486328,
274
+ "step": 1800
275
+ },
276
+ {
277
+ "epoch": 1.4788169464428458,
278
+ "grad_norm": 6.654059886932373,
279
+ "learning_rate": 1.0146549427124967e-05,
280
+ "loss": 2.0765325927734377,
281
+ "step": 1850
282
+ },
283
+ {
284
+ "epoch": 1.518784972022382,
285
+ "grad_norm": 4.843920707702637,
286
+ "learning_rate": 9.880095923261391e-06,
287
+ "loss": 2.078269195556641,
288
+ "step": 1900
289
+ },
290
+ {
291
+ "epoch": 1.5587529976019185,
292
+ "grad_norm": 5.037885665893555,
293
+ "learning_rate": 9.613642419397815e-06,
294
+ "loss": 2.0772938537597656,
295
+ "step": 1950
296
+ },
297
+ {
298
+ "epoch": 1.5987210231814548,
299
+ "grad_norm": 6.010105133056641,
300
+ "learning_rate": 9.34718891553424e-06,
301
+ "loss": 2.07054931640625,
302
+ "step": 2000
303
+ },
304
+ {
305
+ "epoch": 1.638689048760991,
306
+ "grad_norm": 4.658566474914551,
307
+ "learning_rate": 9.080735411670665e-06,
308
+ "loss": 2.0459413146972656,
309
+ "step": 2050
310
+ },
311
+ {
312
+ "epoch": 1.6786570743405276,
313
+ "grad_norm": 5.411528587341309,
314
+ "learning_rate": 8.814281907807088e-06,
315
+ "loss": 2.0509548950195313,
316
+ "step": 2100
317
+ },
318
+ {
319
+ "epoch": 1.718625099920064,
320
+ "grad_norm": 6.495298385620117,
321
+ "learning_rate": 8.547828403943512e-06,
322
+ "loss": 2.0080926513671873,
323
+ "step": 2150
324
+ },
325
+ {
326
+ "epoch": 1.7585931254996003,
327
+ "grad_norm": 7.016653537750244,
328
+ "learning_rate": 8.281374900079936e-06,
329
+ "loss": 2.0153048706054686,
330
+ "step": 2200
331
+ },
332
+ {
333
+ "epoch": 1.7985611510791366,
334
+ "grad_norm": 5.021507263183594,
335
+ "learning_rate": 8.01492139621636e-06,
336
+ "loss": 2.055178680419922,
337
+ "step": 2250
338
+ },
339
+ {
340
+ "epoch": 1.838529176658673,
341
+ "grad_norm": 5.355196475982666,
342
+ "learning_rate": 7.748467892352786e-06,
343
+ "loss": 2.0198463439941405,
344
+ "step": 2300
345
+ },
346
+ {
347
+ "epoch": 1.8784972022382096,
348
+ "grad_norm": 6.557318687438965,
349
+ "learning_rate": 7.48201438848921e-06,
350
+ "loss": 2.0120050048828126,
351
+ "step": 2350
352
+ },
353
+ {
354
+ "epoch": 1.9184652278177459,
355
+ "grad_norm": 6.802421569824219,
356
+ "learning_rate": 7.215560884625633e-06,
357
+ "loss": 2.03742431640625,
358
+ "step": 2400
359
+ },
360
+ {
361
+ "epoch": 1.9584332533972821,
362
+ "grad_norm": 6.564506530761719,
363
+ "learning_rate": 6.949107380762058e-06,
364
+ "loss": 1.9714970397949219,
365
+ "step": 2450
366
+ },
367
+ {
368
+ "epoch": 1.9984012789768184,
369
+ "grad_norm": 6.481470108032227,
370
+ "learning_rate": 6.682653876898481e-06,
371
+ "loss": 2.002969512939453,
372
+ "step": 2500
373
+ },
374
+ {
375
+ "epoch": 2.0,
376
+ "eval_accuracy": 0.4633110066979906,
377
+ "eval_loss": 1.9806830883026123,
378
+ "eval_macro_f1": 0.5002139201674197,
379
+ "eval_macro_precision": 0.5580379243961209,
380
+ "eval_macro_recall": 0.49887820516729536,
381
+ "eval_runtime": 98.8336,
382
+ "eval_samples_per_second": 202.421,
383
+ "eval_steps_per_second": 3.167,
384
+ "step": 2502
385
+ },
386
+ {
387
+ "epoch": 2.038369304556355,
388
+ "grad_norm": 5.825824737548828,
389
+ "learning_rate": 6.4162003730349054e-06,
390
+ "loss": 1.996568145751953,
391
+ "step": 2550
392
+ },
393
+ {
394
+ "epoch": 2.0783373301358914,
395
+ "grad_norm": 5.599540710449219,
396
+ "learning_rate": 6.1497468691713305e-06,
397
+ "loss": 1.9537928771972657,
398
+ "step": 2600
399
+ },
400
+ {
401
+ "epoch": 2.1183053557154277,
402
+ "grad_norm": 7.352904796600342,
403
+ "learning_rate": 5.883293365307754e-06,
404
+ "loss": 1.93813232421875,
405
+ "step": 2650
406
+ },
407
+ {
408
+ "epoch": 2.158273381294964,
409
+ "grad_norm": 6.629299640655518,
410
+ "learning_rate": 5.616839861444179e-06,
411
+ "loss": 1.9688923645019532,
412
+ "step": 2700
413
+ },
414
+ {
415
+ "epoch": 2.1982414068745,
416
+ "grad_norm": 6.491769790649414,
417
+ "learning_rate": 5.350386357580602e-06,
418
+ "loss": 1.9486038208007812,
419
+ "step": 2750
420
+ },
421
+ {
422
+ "epoch": 2.238209432454037,
423
+ "grad_norm": 6.566427230834961,
424
+ "learning_rate": 5.083932853717026e-06,
425
+ "loss": 2.001356506347656,
426
+ "step": 2800
427
+ },
428
+ {
429
+ "epoch": 2.278177458033573,
430
+ "grad_norm": 5.103731155395508,
431
+ "learning_rate": 4.817479349853451e-06,
432
+ "loss": 1.9097093200683595,
433
+ "step": 2850
434
+ },
435
+ {
436
+ "epoch": 2.3181454836131095,
437
+ "grad_norm": 8.290349960327148,
438
+ "learning_rate": 4.5510258459898755e-06,
439
+ "loss": 1.9358775329589843,
440
+ "step": 2900
441
+ },
442
+ {
443
+ "epoch": 2.3581135091926457,
444
+ "grad_norm": 7.596254348754883,
445
+ "learning_rate": 4.284572342126299e-06,
446
+ "loss": 1.8779676818847657,
447
+ "step": 2950
448
+ },
449
+ {
450
+ "epoch": 2.3980815347721824,
451
+ "grad_norm": 6.9873151779174805,
452
+ "learning_rate": 4.018118838262724e-06,
453
+ "loss": 1.9217027282714845,
454
+ "step": 3000
455
+ },
456
+ {
457
+ "epoch": 2.4380495603517187,
458
+ "grad_norm": 6.62452507019043,
459
+ "learning_rate": 3.7516653343991476e-06,
460
+ "loss": 1.941250762939453,
461
+ "step": 3050
462
+ },
463
+ {
464
+ "epoch": 2.478017585931255,
465
+ "grad_norm": 4.5438313484191895,
466
+ "learning_rate": 3.485211830535572e-06,
467
+ "loss": 1.9345573425292968,
468
+ "step": 3100
469
+ },
470
+ {
471
+ "epoch": 2.5179856115107913,
472
+ "grad_norm": 5.7735595703125,
473
+ "learning_rate": 3.2187583266719955e-06,
474
+ "loss": 1.918298797607422,
475
+ "step": 3150
476
+ },
477
+ {
478
+ "epoch": 2.5579536370903275,
479
+ "grad_norm": 8.303494453430176,
480
+ "learning_rate": 2.95230482280842e-06,
481
+ "loss": 1.8983050537109376,
482
+ "step": 3200
483
+ },
484
+ {
485
+ "epoch": 2.5979216626698642,
486
+ "grad_norm": 7.226261615753174,
487
+ "learning_rate": 2.6858513189448443e-06,
488
+ "loss": 1.9115919494628906,
489
+ "step": 3250
490
+ },
491
+ {
492
+ "epoch": 2.6378896882494005,
493
+ "grad_norm": 7.166338920593262,
494
+ "learning_rate": 2.4193978150812685e-06,
495
+ "loss": 1.9131546020507812,
496
+ "step": 3300
497
+ },
498
+ {
499
+ "epoch": 2.677857713828937,
500
+ "grad_norm": 6.6787261962890625,
501
+ "learning_rate": 2.1529443112176926e-06,
502
+ "loss": 1.9326390075683593,
503
+ "step": 3350
504
+ },
505
+ {
506
+ "epoch": 2.7178257394084735,
507
+ "grad_norm": 7.1048264503479,
508
+ "learning_rate": 1.8864908073541168e-06,
509
+ "loss": 1.8645829772949218,
510
+ "step": 3400
511
+ },
512
+ {
513
+ "epoch": 2.7577937649880093,
514
+ "grad_norm": 5.909543037414551,
515
+ "learning_rate": 1.6200373034905412e-06,
516
+ "loss": 1.9545182800292968,
517
+ "step": 3450
518
+ },
519
+ {
520
+ "epoch": 2.797761790567546,
521
+ "grad_norm": 4.829916477203369,
522
+ "learning_rate": 1.3535837996269651e-06,
523
+ "loss": 1.9301649475097655,
524
+ "step": 3500
525
+ },
526
+ {
527
+ "epoch": 2.8377298161470823,
528
+ "grad_norm": 7.5515289306640625,
529
+ "learning_rate": 1.0871302957633893e-06,
530
+ "loss": 1.9069998168945312,
531
+ "step": 3550
532
+ },
533
+ {
534
+ "epoch": 2.8776978417266186,
535
+ "grad_norm": 7.264887809753418,
536
+ "learning_rate": 8.206767918998136e-07,
537
+ "loss": 1.8916896057128907,
538
+ "step": 3600
539
+ },
540
+ {
541
+ "epoch": 2.9176658673061553,
542
+ "grad_norm": 5.612414836883545,
543
+ "learning_rate": 5.542232880362377e-07,
544
+ "loss": 1.8344429016113282,
545
+ "step": 3650
546
+ },
547
+ {
548
+ "epoch": 2.9576338928856916,
549
+ "grad_norm": 5.269375324249268,
550
+ "learning_rate": 2.877697841726619e-07,
551
+ "loss": 1.9053506469726562,
552
+ "step": 3700
553
+ },
554
+ {
555
+ "epoch": 2.997601918465228,
556
+ "grad_norm": 5.398017883300781,
557
+ "learning_rate": 2.1316280309086065e-08,
558
+ "loss": 1.878918914794922,
559
+ "step": 3750
560
+ },
561
+ {
562
+ "epoch": 3.0,
563
+ "eval_accuracy": 0.4828551434569629,
564
+ "eval_loss": 1.9081437587738037,
565
+ "eval_macro_f1": 0.5181542733976486,
566
+ "eval_macro_precision": 0.5688194962198772,
567
+ "eval_macro_recall": 0.5197711618330794,
568
+ "eval_runtime": 98.5634,
569
+ "eval_samples_per_second": 202.976,
570
+ "eval_steps_per_second": 3.176,
571
+ "step": 3753
572
+ }
573
+ ],
574
+ "logging_steps": 50,
575
+ "max_steps": 3753,
576
+ "num_input_tokens_seen": 0,
577
+ "num_train_epochs": 3,
578
+ "save_steps": 500,
579
+ "stateful_callbacks": {
580
+ "TrainerControl": {
581
+ "args": {
582
+ "should_epoch_stop": false,
583
+ "should_evaluate": false,
584
+ "should_log": false,
585
+ "should_save": true,
586
+ "should_training_stop": true
587
+ },
588
+ "attributes": {}
589
+ }
590
+ },
591
+ "total_flos": 1.2246069348274104e+16,
592
+ "train_batch_size": 64,
593
+ "trial_name": null,
594
+ "trial_params": null
595
+ }
checkpoint-3753/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f18c5dfb7dc4eb65becd7f75437e97d7a12dc629e02e3a7f3952ff22db75c9d
3
+ size 5265
config.json ADDED
@@ -0,0 +1,194 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation": "gelu",
3
+ "architectures": [
4
+ "DistilBertForSequenceClassification"
5
+ ],
6
+ "attention_dropout": 0.1,
7
+ "bos_token_id": null,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "dtype": "float32",
11
+ "eos_token_id": null,
12
+ "hidden_dim": 3072,
13
+ "id2label": {
14
+ "0": "adfgvx",
15
+ "1": "adfgx",
16
+ "2": "aeneas_tacticus",
17
+ "3": "affine",
18
+ "4": "alberti_disk",
19
+ "5": "argenti",
20
+ "6": "arnold_andre",
21
+ "7": "atbash",
22
+ "8": "autokey",
23
+ "9": "babington",
24
+ "10": "bacon_cipher",
25
+ "11": "bazeries",
26
+ "12": "beaufort",
27
+ "13": "bifid",
28
+ "14": "book_cipher",
29
+ "15": "caesar",
30
+ "16": "cardano_autokey",
31
+ "17": "chaocipher",
32
+ "18": "chinese_telegraph",
33
+ "19": "columnar_transposition",
34
+ "20": "commercial_code",
35
+ "21": "confederate_vigenere",
36
+ "22": "copiale",
37
+ "23": "culper_ring",
38
+ "24": "diana",
39
+ "25": "double_transposition",
40
+ "26": "enigma",
41
+ "27": "fialka",
42
+ "28": "four_square",
43
+ "29": "fractionated_morse",
44
+ "30": "geez_monastic",
45
+ "31": "geheimschreiber",
46
+ "32": "great_cipher",
47
+ "33": "gronsfeld",
48
+ "34": "hill",
49
+ "35": "homophonic",
50
+ "36": "jefferson_disk",
51
+ "37": "jn25",
52
+ "38": "joseon_yeokhak",
53
+ "39": "kama_sutra",
54
+ "40": "kl7",
55
+ "41": "kryha",
56
+ "42": "kryptos",
57
+ "43": "lorenz",
58
+ "44": "m209",
59
+ "45": "m94",
60
+ "46": "monoalphabetic",
61
+ "47": "morse_code",
62
+ "48": "navajo_code",
63
+ "49": "nihilist",
64
+ "50": "nomenclator",
65
+ "51": "null_cipher",
66
+ "52": "one_time_pad",
67
+ "53": "pigpen",
68
+ "54": "playfair",
69
+ "55": "polybius",
70
+ "56": "porta",
71
+ "57": "purple",
72
+ "58": "rail_fence",
73
+ "59": "red_type_a",
74
+ "60": "rot13",
75
+ "61": "running_key",
76
+ "62": "scytale",
77
+ "63": "sigaba",
78
+ "64": "slidex",
79
+ "65": "solitaire",
80
+ "66": "stager_route",
81
+ "67": "straddling_checkerboard",
82
+ "68": "tap_code",
83
+ "69": "trifid",
84
+ "70": "trithemius",
85
+ "71": "two_square",
86
+ "72": "typex",
87
+ "73": "venona_pad_reuse",
88
+ "74": "vernam",
89
+ "75": "vic",
90
+ "76": "vigenere",
91
+ "77": "voynich_render",
92
+ "78": "wallis_cipher",
93
+ "79": "wheatstone",
94
+ "80": "zimmermann"
95
+ },
96
+ "initializer_range": 0.02,
97
+ "label2id": {
98
+ "adfgvx": 0,
99
+ "adfgx": 1,
100
+ "aeneas_tacticus": 2,
101
+ "affine": 3,
102
+ "alberti_disk": 4,
103
+ "argenti": 5,
104
+ "arnold_andre": 6,
105
+ "atbash": 7,
106
+ "autokey": 8,
107
+ "babington": 9,
108
+ "bacon_cipher": 10,
109
+ "bazeries": 11,
110
+ "beaufort": 12,
111
+ "bifid": 13,
112
+ "book_cipher": 14,
113
+ "caesar": 15,
114
+ "cardano_autokey": 16,
115
+ "chaocipher": 17,
116
+ "chinese_telegraph": 18,
117
+ "columnar_transposition": 19,
118
+ "commercial_code": 20,
119
+ "confederate_vigenere": 21,
120
+ "copiale": 22,
121
+ "culper_ring": 23,
122
+ "diana": 24,
123
+ "double_transposition": 25,
124
+ "enigma": 26,
125
+ "fialka": 27,
126
+ "four_square": 28,
127
+ "fractionated_morse": 29,
128
+ "geez_monastic": 30,
129
+ "geheimschreiber": 31,
130
+ "great_cipher": 32,
131
+ "gronsfeld": 33,
132
+ "hill": 34,
133
+ "homophonic": 35,
134
+ "jefferson_disk": 36,
135
+ "jn25": 37,
136
+ "joseon_yeokhak": 38,
137
+ "kama_sutra": 39,
138
+ "kl7": 40,
139
+ "kryha": 41,
140
+ "kryptos": 42,
141
+ "lorenz": 43,
142
+ "m209": 44,
143
+ "m94": 45,
144
+ "monoalphabetic": 46,
145
+ "morse_code": 47,
146
+ "navajo_code": 48,
147
+ "nihilist": 49,
148
+ "nomenclator": 50,
149
+ "null_cipher": 51,
150
+ "one_time_pad": 52,
151
+ "pigpen": 53,
152
+ "playfair": 54,
153
+ "polybius": 55,
154
+ "porta": 56,
155
+ "purple": 57,
156
+ "rail_fence": 58,
157
+ "red_type_a": 59,
158
+ "rot13": 60,
159
+ "running_key": 61,
160
+ "scytale": 62,
161
+ "sigaba": 63,
162
+ "slidex": 64,
163
+ "solitaire": 65,
164
+ "stager_route": 66,
165
+ "straddling_checkerboard": 67,
166
+ "tap_code": 68,
167
+ "trifid": 69,
168
+ "trithemius": 70,
169
+ "two_square": 71,
170
+ "typex": 72,
171
+ "venona_pad_reuse": 73,
172
+ "vernam": 74,
173
+ "vic": 75,
174
+ "vigenere": 76,
175
+ "voynich_render": 77,
176
+ "wallis_cipher": 78,
177
+ "wheatstone": 79,
178
+ "zimmermann": 80
179
+ },
180
+ "max_position_embeddings": 512,
181
+ "model_type": "distilbert",
182
+ "n_heads": 12,
183
+ "n_layers": 6,
184
+ "pad_token_id": 0,
185
+ "problem_type": "single_label_classification",
186
+ "qa_dropout": 0.1,
187
+ "seq_classif_dropout": 0.2,
188
+ "sinusoidal_pos_embds": false,
189
+ "tie_weights_": true,
190
+ "tie_word_embeddings": true,
191
+ "transformers_version": "5.8.0",
192
+ "use_cache": false,
193
+ "vocab_size": 30522
194
+ }
label_mapping.json ADDED
@@ -0,0 +1,168 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "label2id": {
3
+ "adfgvx": 0,
4
+ "adfgx": 1,
5
+ "aeneas_tacticus": 2,
6
+ "affine": 3,
7
+ "alberti_disk": 4,
8
+ "argenti": 5,
9
+ "arnold_andre": 6,
10
+ "atbash": 7,
11
+ "autokey": 8,
12
+ "babington": 9,
13
+ "bacon_cipher": 10,
14
+ "bazeries": 11,
15
+ "beaufort": 12,
16
+ "bifid": 13,
17
+ "book_cipher": 14,
18
+ "caesar": 15,
19
+ "cardano_autokey": 16,
20
+ "chaocipher": 17,
21
+ "chinese_telegraph": 18,
22
+ "columnar_transposition": 19,
23
+ "commercial_code": 20,
24
+ "confederate_vigenere": 21,
25
+ "copiale": 22,
26
+ "culper_ring": 23,
27
+ "diana": 24,
28
+ "double_transposition": 25,
29
+ "enigma": 26,
30
+ "fialka": 27,
31
+ "four_square": 28,
32
+ "fractionated_morse": 29,
33
+ "geez_monastic": 30,
34
+ "geheimschreiber": 31,
35
+ "great_cipher": 32,
36
+ "gronsfeld": 33,
37
+ "hill": 34,
38
+ "homophonic": 35,
39
+ "jefferson_disk": 36,
40
+ "jn25": 37,
41
+ "joseon_yeokhak": 38,
42
+ "kama_sutra": 39,
43
+ "kl7": 40,
44
+ "kryha": 41,
45
+ "kryptos": 42,
46
+ "lorenz": 43,
47
+ "m209": 44,
48
+ "m94": 45,
49
+ "monoalphabetic": 46,
50
+ "morse_code": 47,
51
+ "navajo_code": 48,
52
+ "nihilist": 49,
53
+ "nomenclator": 50,
54
+ "null_cipher": 51,
55
+ "one_time_pad": 52,
56
+ "pigpen": 53,
57
+ "playfair": 54,
58
+ "polybius": 55,
59
+ "porta": 56,
60
+ "purple": 57,
61
+ "rail_fence": 58,
62
+ "red_type_a": 59,
63
+ "rot13": 60,
64
+ "running_key": 61,
65
+ "scytale": 62,
66
+ "sigaba": 63,
67
+ "slidex": 64,
68
+ "solitaire": 65,
69
+ "stager_route": 66,
70
+ "straddling_checkerboard": 67,
71
+ "tap_code": 68,
72
+ "trifid": 69,
73
+ "trithemius": 70,
74
+ "two_square": 71,
75
+ "typex": 72,
76
+ "venona_pad_reuse": 73,
77
+ "vernam": 74,
78
+ "vic": 75,
79
+ "vigenere": 76,
80
+ "voynich_render": 77,
81
+ "wallis_cipher": 78,
82
+ "wheatstone": 79,
83
+ "zimmermann": 80
84
+ },
85
+ "id2label": {
86
+ "0": "adfgvx",
87
+ "1": "adfgx",
88
+ "2": "aeneas_tacticus",
89
+ "3": "affine",
90
+ "4": "alberti_disk",
91
+ "5": "argenti",
92
+ "6": "arnold_andre",
93
+ "7": "atbash",
94
+ "8": "autokey",
95
+ "9": "babington",
96
+ "10": "bacon_cipher",
97
+ "11": "bazeries",
98
+ "12": "beaufort",
99
+ "13": "bifid",
100
+ "14": "book_cipher",
101
+ "15": "caesar",
102
+ "16": "cardano_autokey",
103
+ "17": "chaocipher",
104
+ "18": "chinese_telegraph",
105
+ "19": "columnar_transposition",
106
+ "20": "commercial_code",
107
+ "21": "confederate_vigenere",
108
+ "22": "copiale",
109
+ "23": "culper_ring",
110
+ "24": "diana",
111
+ "25": "double_transposition",
112
+ "26": "enigma",
113
+ "27": "fialka",
114
+ "28": "four_square",
115
+ "29": "fractionated_morse",
116
+ "30": "geez_monastic",
117
+ "31": "geheimschreiber",
118
+ "32": "great_cipher",
119
+ "33": "gronsfeld",
120
+ "34": "hill",
121
+ "35": "homophonic",
122
+ "36": "jefferson_disk",
123
+ "37": "jn25",
124
+ "38": "joseon_yeokhak",
125
+ "39": "kama_sutra",
126
+ "40": "kl7",
127
+ "41": "kryha",
128
+ "42": "kryptos",
129
+ "43": "lorenz",
130
+ "44": "m209",
131
+ "45": "m94",
132
+ "46": "monoalphabetic",
133
+ "47": "morse_code",
134
+ "48": "navajo_code",
135
+ "49": "nihilist",
136
+ "50": "nomenclator",
137
+ "51": "null_cipher",
138
+ "52": "one_time_pad",
139
+ "53": "pigpen",
140
+ "54": "playfair",
141
+ "55": "polybius",
142
+ "56": "porta",
143
+ "57": "purple",
144
+ "58": "rail_fence",
145
+ "59": "red_type_a",
146
+ "60": "rot13",
147
+ "61": "running_key",
148
+ "62": "scytale",
149
+ "63": "sigaba",
150
+ "64": "slidex",
151
+ "65": "solitaire",
152
+ "66": "stager_route",
153
+ "67": "straddling_checkerboard",
154
+ "68": "tap_code",
155
+ "69": "trifid",
156
+ "70": "trithemius",
157
+ "71": "two_square",
158
+ "72": "typex",
159
+ "73": "venona_pad_reuse",
160
+ "74": "vernam",
161
+ "75": "vic",
162
+ "76": "vigenere",
163
+ "77": "voynich_render",
164
+ "78": "wallis_cipher",
165
+ "79": "wheatstone",
166
+ "80": "zimmermann"
167
+ }
168
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2943abbfbe864d54f7dd6febfe1e9f33c4aa5e49ac2c2e840204148469537444
3
+ size 268075580
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "cls_token": "[CLS]",
4
+ "do_lower_case": true,
5
+ "is_local": false,
6
+ "local_files_only": false,
7
+ "mask_token": "[MASK]",
8
+ "model_max_length": 512,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "strip_accents": null,
12
+ "tokenize_chinese_chars": true,
13
+ "tokenizer_class": "BertTokenizer",
14
+ "unk_token": "[UNK]"
15
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f18c5dfb7dc4eb65becd7f75437e97d7a12dc629e02e3a7f3952ff22db75c9d
3
+ size 5265
training_metrics.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_loss": 1.9081437587738037,
3
+ "eval_accuracy": 0.4828551434569629,
4
+ "eval_macro_precision": 0.5688194962198772,
5
+ "eval_macro_recall": 0.5197711618330794,
6
+ "eval_macro_f1": 0.5181542733976486,
7
+ "eval_runtime": 98.5423,
8
+ "eval_samples_per_second": 203.019,
9
+ "eval_steps_per_second": 3.176,
10
+ "epoch": 3.0
11
+ }