garrettmeek123 commited on
Commit
71bae0b
·
verified ·
1 Parent(s): 3d1b617

Upload folder using huggingface_hub

Browse files
config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "RobertaForSequenceClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": 0,
7
+ "classifier_dropout": null,
8
+ "dtype": "float32",
9
+ "eos_token_id": 2,
10
+ "gradient_checkpointing": false,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 3072,
16
+ "layer_norm_eps": 1e-05,
17
+ "max_position_embeddings": 514,
18
+ "model_type": "roberta",
19
+ "num_attention_heads": 12,
20
+ "num_hidden_layers": 12,
21
+ "output_past": true,
22
+ "pad_token_id": 1,
23
+ "position_embedding_type": "absolute",
24
+ "problem_type": "single_label_classification",
25
+ "transformers_version": "4.57.1",
26
+ "type_vocab_size": 1,
27
+ "use_cache": true,
28
+ "vocab_size": 50265
29
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69db38b9800cfeb7bd3cbbde613cdd4243be06ab69d00422037974e40fead155
3
+ size 498612824
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25f7bddd970da66471c2d9252fd65a76769409bcbf914f2d99dc42728ce2598a
3
+ size 997348747
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbbd94508f58871476757381a84e67b19b54c13ad44f83fcdbaf28ef1d1c3c6b
3
+ size 14709
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6577149420304ede87aa59bf995d3982b1e1dce8c799a97d829347f2ac691371
3
+ size 1465
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": true,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": true,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": true,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "<unk>",
46
+ "lstrip": false,
47
+ "normalized": true,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<pad>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "50264": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
+ "normalized": true,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ }
44
+ },
45
+ "bos_token": "<s>",
46
+ "clean_up_tokenization_spaces": false,
47
+ "cls_token": "<s>",
48
+ "eos_token": "</s>",
49
+ "errors": "replace",
50
+ "extra_special_tokens": {},
51
+ "mask_token": "<mask>",
52
+ "model_max_length": 512,
53
+ "pad_token": "<pad>",
54
+ "sep_token": "</s>",
55
+ "tokenizer_class": "RobertaTokenizer",
56
+ "trim_offsets": true,
57
+ "unk_token": "<unk>"
58
+ }
trainer_state.json ADDED
@@ -0,0 +1,555 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 2049,
3
+ "best_metric": 0.6610541727672035,
4
+ "best_model_checkpoint": "./saved_models/graphcodebert/checkpoint-2049",
5
+ "epoch": 5.0,
6
+ "eval_steps": 500,
7
+ "global_step": 3415,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.07320644216691069,
14
+ "grad_norm": 1.5853087902069092,
15
+ "learning_rate": 1.971303074670571e-05,
16
+ "loss": 0.6901,
17
+ "step": 50
18
+ },
19
+ {
20
+ "epoch": 0.14641288433382138,
21
+ "grad_norm": 2.419438123703003,
22
+ "learning_rate": 1.942020497803807e-05,
23
+ "loss": 0.6902,
24
+ "step": 100
25
+ },
26
+ {
27
+ "epoch": 0.21961932650073207,
28
+ "grad_norm": 1.6440658569335938,
29
+ "learning_rate": 1.9127379209370426e-05,
30
+ "loss": 0.685,
31
+ "step": 150
32
+ },
33
+ {
34
+ "epoch": 0.29282576866764276,
35
+ "grad_norm": 1.883244514465332,
36
+ "learning_rate": 1.8834553440702785e-05,
37
+ "loss": 0.6857,
38
+ "step": 200
39
+ },
40
+ {
41
+ "epoch": 0.36603221083455345,
42
+ "grad_norm": 3.6438748836517334,
43
+ "learning_rate": 1.854172767203514e-05,
44
+ "loss": 0.682,
45
+ "step": 250
46
+ },
47
+ {
48
+ "epoch": 0.43923865300146414,
49
+ "grad_norm": 3.641892194747925,
50
+ "learning_rate": 1.8248901903367496e-05,
51
+ "loss": 0.6656,
52
+ "step": 300
53
+ },
54
+ {
55
+ "epoch": 0.5124450951683748,
56
+ "grad_norm": 2.7403852939605713,
57
+ "learning_rate": 1.7956076134699855e-05,
58
+ "loss": 0.6656,
59
+ "step": 350
60
+ },
61
+ {
62
+ "epoch": 0.5856515373352855,
63
+ "grad_norm": 1.8151828050613403,
64
+ "learning_rate": 1.766325036603221e-05,
65
+ "loss": 0.6762,
66
+ "step": 400
67
+ },
68
+ {
69
+ "epoch": 0.6588579795021962,
70
+ "grad_norm": 2.336029291152954,
71
+ "learning_rate": 1.737042459736457e-05,
72
+ "loss": 0.6701,
73
+ "step": 450
74
+ },
75
+ {
76
+ "epoch": 0.7320644216691069,
77
+ "grad_norm": 2.9843106269836426,
78
+ "learning_rate": 1.7077598828696925e-05,
79
+ "loss": 0.6661,
80
+ "step": 500
81
+ },
82
+ {
83
+ "epoch": 0.8052708638360175,
84
+ "grad_norm": 2.4253265857696533,
85
+ "learning_rate": 1.6784773060029284e-05,
86
+ "loss": 0.6776,
87
+ "step": 550
88
+ },
89
+ {
90
+ "epoch": 0.8784773060029283,
91
+ "grad_norm": 3.049626350402832,
92
+ "learning_rate": 1.649194729136164e-05,
93
+ "loss": 0.6626,
94
+ "step": 600
95
+ },
96
+ {
97
+ "epoch": 0.9516837481698389,
98
+ "grad_norm": 2.690331220626831,
99
+ "learning_rate": 1.6199121522694e-05,
100
+ "loss": 0.658,
101
+ "step": 650
102
+ },
103
+ {
104
+ "epoch": 1.0,
105
+ "eval_accuracy": 0.6332357247437774,
106
+ "eval_loss": 0.6321774125099182,
107
+ "eval_runtime": 11.6506,
108
+ "eval_samples_per_second": 234.494,
109
+ "eval_steps_per_second": 3.691,
110
+ "step": 683
111
+ },
112
+ {
113
+ "epoch": 1.0248901903367496,
114
+ "grad_norm": 3.389944553375244,
115
+ "learning_rate": 1.5906295754026355e-05,
116
+ "loss": 0.645,
117
+ "step": 700
118
+ },
119
+ {
120
+ "epoch": 1.0980966325036603,
121
+ "grad_norm": 2.976720094680786,
122
+ "learning_rate": 1.5613469985358714e-05,
123
+ "loss": 0.633,
124
+ "step": 750
125
+ },
126
+ {
127
+ "epoch": 1.171303074670571,
128
+ "grad_norm": 3.046492576599121,
129
+ "learning_rate": 1.532064421669107e-05,
130
+ "loss": 0.63,
131
+ "step": 800
132
+ },
133
+ {
134
+ "epoch": 1.2445095168374818,
135
+ "grad_norm": 2.867762804031372,
136
+ "learning_rate": 1.5027818448023428e-05,
137
+ "loss": 0.6293,
138
+ "step": 850
139
+ },
140
+ {
141
+ "epoch": 1.3177159590043923,
142
+ "grad_norm": 3.334984064102173,
143
+ "learning_rate": 1.4734992679355784e-05,
144
+ "loss": 0.6251,
145
+ "step": 900
146
+ },
147
+ {
148
+ "epoch": 1.390922401171303,
149
+ "grad_norm": 4.373927116394043,
150
+ "learning_rate": 1.4442166910688143e-05,
151
+ "loss": 0.6224,
152
+ "step": 950
153
+ },
154
+ {
155
+ "epoch": 1.4641288433382138,
156
+ "grad_norm": 3.5130934715270996,
157
+ "learning_rate": 1.4149341142020499e-05,
158
+ "loss": 0.6182,
159
+ "step": 1000
160
+ },
161
+ {
162
+ "epoch": 1.5373352855051245,
163
+ "grad_norm": 3.466965675354004,
164
+ "learning_rate": 1.3856515373352856e-05,
165
+ "loss": 0.6139,
166
+ "step": 1050
167
+ },
168
+ {
169
+ "epoch": 1.610541727672035,
170
+ "grad_norm": 4.486567974090576,
171
+ "learning_rate": 1.3563689604685213e-05,
172
+ "loss": 0.6221,
173
+ "step": 1100
174
+ },
175
+ {
176
+ "epoch": 1.6837481698389458,
177
+ "grad_norm": 3.9357879161834717,
178
+ "learning_rate": 1.327086383601757e-05,
179
+ "loss": 0.6185,
180
+ "step": 1150
181
+ },
182
+ {
183
+ "epoch": 1.7569546120058566,
184
+ "grad_norm": 3.294029474258423,
185
+ "learning_rate": 1.2978038067349928e-05,
186
+ "loss": 0.6048,
187
+ "step": 1200
188
+ },
189
+ {
190
+ "epoch": 1.830161054172767,
191
+ "grad_norm": 4.328296661376953,
192
+ "learning_rate": 1.2685212298682286e-05,
193
+ "loss": 0.6191,
194
+ "step": 1250
195
+ },
196
+ {
197
+ "epoch": 1.903367496339678,
198
+ "grad_norm": 3.7641713619232178,
199
+ "learning_rate": 1.2392386530014641e-05,
200
+ "loss": 0.6028,
201
+ "step": 1300
202
+ },
203
+ {
204
+ "epoch": 1.9765739385065886,
205
+ "grad_norm": 4.414918422698975,
206
+ "learning_rate": 1.2099560761347e-05,
207
+ "loss": 0.5892,
208
+ "step": 1350
209
+ },
210
+ {
211
+ "epoch": 2.0,
212
+ "eval_accuracy": 0.636896046852123,
213
+ "eval_loss": 0.6083204746246338,
214
+ "eval_runtime": 11.6592,
215
+ "eval_samples_per_second": 234.32,
216
+ "eval_steps_per_second": 3.688,
217
+ "step": 1366
218
+ },
219
+ {
220
+ "epoch": 2.049780380673499,
221
+ "grad_norm": 5.577649116516113,
222
+ "learning_rate": 1.1806734992679356e-05,
223
+ "loss": 0.5804,
224
+ "step": 1400
225
+ },
226
+ {
227
+ "epoch": 2.12298682284041,
228
+ "grad_norm": 5.713629722595215,
229
+ "learning_rate": 1.1513909224011715e-05,
230
+ "loss": 0.5494,
231
+ "step": 1450
232
+ },
233
+ {
234
+ "epoch": 2.1961932650073206,
235
+ "grad_norm": 14.552216529846191,
236
+ "learning_rate": 1.122108345534407e-05,
237
+ "loss": 0.5611,
238
+ "step": 1500
239
+ },
240
+ {
241
+ "epoch": 2.269399707174231,
242
+ "grad_norm": 4.35590124130249,
243
+ "learning_rate": 1.092825768667643e-05,
244
+ "loss": 0.5653,
245
+ "step": 1550
246
+ },
247
+ {
248
+ "epoch": 2.342606149341142,
249
+ "grad_norm": 4.740884304046631,
250
+ "learning_rate": 1.0635431918008785e-05,
251
+ "loss": 0.5658,
252
+ "step": 1600
253
+ },
254
+ {
255
+ "epoch": 2.4158125915080526,
256
+ "grad_norm": 6.103859901428223,
257
+ "learning_rate": 1.0342606149341143e-05,
258
+ "loss": 0.5605,
259
+ "step": 1650
260
+ },
261
+ {
262
+ "epoch": 2.4890190336749636,
263
+ "grad_norm": 5.405401229858398,
264
+ "learning_rate": 1.00497803806735e-05,
265
+ "loss": 0.5657,
266
+ "step": 1700
267
+ },
268
+ {
269
+ "epoch": 2.562225475841874,
270
+ "grad_norm": 5.416690349578857,
271
+ "learning_rate": 9.756954612005857e-06,
272
+ "loss": 0.545,
273
+ "step": 1750
274
+ },
275
+ {
276
+ "epoch": 2.6354319180087846,
277
+ "grad_norm": 3.8133316040039062,
278
+ "learning_rate": 9.464128843338215e-06,
279
+ "loss": 0.5518,
280
+ "step": 1800
281
+ },
282
+ {
283
+ "epoch": 2.7086383601756956,
284
+ "grad_norm": 11.942543029785156,
285
+ "learning_rate": 9.171303074670572e-06,
286
+ "loss": 0.5516,
287
+ "step": 1850
288
+ },
289
+ {
290
+ "epoch": 2.781844802342606,
291
+ "grad_norm": 7.067141532897949,
292
+ "learning_rate": 8.87847730600293e-06,
293
+ "loss": 0.5597,
294
+ "step": 1900
295
+ },
296
+ {
297
+ "epoch": 2.855051244509517,
298
+ "grad_norm": 5.501221656799316,
299
+ "learning_rate": 8.585651537335287e-06,
300
+ "loss": 0.536,
301
+ "step": 1950
302
+ },
303
+ {
304
+ "epoch": 2.9282576866764276,
305
+ "grad_norm": 9.329773902893066,
306
+ "learning_rate": 8.292825768667644e-06,
307
+ "loss": 0.5298,
308
+ "step": 2000
309
+ },
310
+ {
311
+ "epoch": 3.0,
312
+ "eval_accuracy": 0.6610541727672035,
313
+ "eval_loss": 0.6209500432014465,
314
+ "eval_runtime": 11.6196,
315
+ "eval_samples_per_second": 235.121,
316
+ "eval_steps_per_second": 3.701,
317
+ "step": 2049
318
+ },
319
+ {
320
+ "epoch": 3.001464128843338,
321
+ "grad_norm": 4.468164443969727,
322
+ "learning_rate": 8.000000000000001e-06,
323
+ "loss": 0.5594,
324
+ "step": 2050
325
+ },
326
+ {
327
+ "epoch": 3.074670571010249,
328
+ "grad_norm": 6.163565158843994,
329
+ "learning_rate": 7.707174231332359e-06,
330
+ "loss": 0.488,
331
+ "step": 2100
332
+ },
333
+ {
334
+ "epoch": 3.1478770131771596,
335
+ "grad_norm": 6.804717063903809,
336
+ "learning_rate": 7.414348462664715e-06,
337
+ "loss": 0.491,
338
+ "step": 2150
339
+ },
340
+ {
341
+ "epoch": 3.22108345534407,
342
+ "grad_norm": 7.049874782562256,
343
+ "learning_rate": 7.1215226939970725e-06,
344
+ "loss": 0.4866,
345
+ "step": 2200
346
+ },
347
+ {
348
+ "epoch": 3.294289897510981,
349
+ "grad_norm": 10.319636344909668,
350
+ "learning_rate": 6.82869692532943e-06,
351
+ "loss": 0.4837,
352
+ "step": 2250
353
+ },
354
+ {
355
+ "epoch": 3.3674963396778916,
356
+ "grad_norm": 7.2100677490234375,
357
+ "learning_rate": 6.535871156661787e-06,
358
+ "loss": 0.4773,
359
+ "step": 2300
360
+ },
361
+ {
362
+ "epoch": 3.440702781844802,
363
+ "grad_norm": 10.271724700927734,
364
+ "learning_rate": 6.2430453879941446e-06,
365
+ "loss": 0.5051,
366
+ "step": 2350
367
+ },
368
+ {
369
+ "epoch": 3.513909224011713,
370
+ "grad_norm": 6.702558517456055,
371
+ "learning_rate": 5.950219619326502e-06,
372
+ "loss": 0.4979,
373
+ "step": 2400
374
+ },
375
+ {
376
+ "epoch": 3.5871156661786237,
377
+ "grad_norm": 7.939746856689453,
378
+ "learning_rate": 5.657393850658858e-06,
379
+ "loss": 0.5013,
380
+ "step": 2450
381
+ },
382
+ {
383
+ "epoch": 3.660322108345534,
384
+ "grad_norm": 7.762535095214844,
385
+ "learning_rate": 5.364568081991216e-06,
386
+ "loss": 0.4999,
387
+ "step": 2500
388
+ },
389
+ {
390
+ "epoch": 3.733528550512445,
391
+ "grad_norm": 7.71613073348999,
392
+ "learning_rate": 5.071742313323573e-06,
393
+ "loss": 0.4935,
394
+ "step": 2550
395
+ },
396
+ {
397
+ "epoch": 3.8067349926793557,
398
+ "grad_norm": 6.267000675201416,
399
+ "learning_rate": 4.77891654465593e-06,
400
+ "loss": 0.4855,
401
+ "step": 2600
402
+ },
403
+ {
404
+ "epoch": 3.8799414348462666,
405
+ "grad_norm": 10.031179428100586,
406
+ "learning_rate": 4.486090775988287e-06,
407
+ "loss": 0.4932,
408
+ "step": 2650
409
+ },
410
+ {
411
+ "epoch": 3.953147877013177,
412
+ "grad_norm": 5.606414794921875,
413
+ "learning_rate": 4.193265007320644e-06,
414
+ "loss": 0.4849,
415
+ "step": 2700
416
+ },
417
+ {
418
+ "epoch": 4.0,
419
+ "eval_accuracy": 0.6442166910688141,
420
+ "eval_loss": 0.6546101570129395,
421
+ "eval_runtime": 11.6102,
422
+ "eval_samples_per_second": 235.31,
423
+ "eval_steps_per_second": 3.704,
424
+ "step": 2732
425
+ },
426
+ {
427
+ "epoch": 4.026354319180088,
428
+ "grad_norm": 7.047442436218262,
429
+ "learning_rate": 3.900439238653002e-06,
430
+ "loss": 0.4756,
431
+ "step": 2750
432
+ },
433
+ {
434
+ "epoch": 4.099560761346998,
435
+ "grad_norm": 8.444013595581055,
436
+ "learning_rate": 3.607613469985359e-06,
437
+ "loss": 0.4465,
438
+ "step": 2800
439
+ },
440
+ {
441
+ "epoch": 4.172767203513909,
442
+ "grad_norm": 8.391129493713379,
443
+ "learning_rate": 3.314787701317716e-06,
444
+ "loss": 0.4532,
445
+ "step": 2850
446
+ },
447
+ {
448
+ "epoch": 4.24597364568082,
449
+ "grad_norm": 8.165177345275879,
450
+ "learning_rate": 3.0219619326500732e-06,
451
+ "loss": 0.4591,
452
+ "step": 2900
453
+ },
454
+ {
455
+ "epoch": 4.31918008784773,
456
+ "grad_norm": 10.9652099609375,
457
+ "learning_rate": 2.7291361639824306e-06,
458
+ "loss": 0.4126,
459
+ "step": 2950
460
+ },
461
+ {
462
+ "epoch": 4.392386530014641,
463
+ "grad_norm": 7.355696201324463,
464
+ "learning_rate": 2.436310395314788e-06,
465
+ "loss": 0.4195,
466
+ "step": 3000
467
+ },
468
+ {
469
+ "epoch": 4.465592972181552,
470
+ "grad_norm": 9.898347854614258,
471
+ "learning_rate": 2.1434846266471453e-06,
472
+ "loss": 0.4476,
473
+ "step": 3050
474
+ },
475
+ {
476
+ "epoch": 4.538799414348462,
477
+ "grad_norm": 12.907100677490234,
478
+ "learning_rate": 1.8506588579795024e-06,
479
+ "loss": 0.441,
480
+ "step": 3100
481
+ },
482
+ {
483
+ "epoch": 4.612005856515373,
484
+ "grad_norm": 6.766877174377441,
485
+ "learning_rate": 1.5578330893118595e-06,
486
+ "loss": 0.4316,
487
+ "step": 3150
488
+ },
489
+ {
490
+ "epoch": 4.685212298682284,
491
+ "grad_norm": 6.144525527954102,
492
+ "learning_rate": 1.2650073206442169e-06,
493
+ "loss": 0.4415,
494
+ "step": 3200
495
+ },
496
+ {
497
+ "epoch": 4.758418740849194,
498
+ "grad_norm": 12.745665550231934,
499
+ "learning_rate": 9.72181551976574e-07,
500
+ "loss": 0.4525,
501
+ "step": 3250
502
+ },
503
+ {
504
+ "epoch": 4.831625183016105,
505
+ "grad_norm": 8.909607887268066,
506
+ "learning_rate": 6.793557833089313e-07,
507
+ "loss": 0.4426,
508
+ "step": 3300
509
+ },
510
+ {
511
+ "epoch": 4.904831625183016,
512
+ "grad_norm": 8.444921493530273,
513
+ "learning_rate": 3.865300146412885e-07,
514
+ "loss": 0.4379,
515
+ "step": 3350
516
+ },
517
+ {
518
+ "epoch": 4.978038067349927,
519
+ "grad_norm": 9.61002254486084,
520
+ "learning_rate": 9.370424597364569e-08,
521
+ "loss": 0.4603,
522
+ "step": 3400
523
+ },
524
+ {
525
+ "epoch": 5.0,
526
+ "eval_accuracy": 0.6548316251830161,
527
+ "eval_loss": 0.687939465045929,
528
+ "eval_runtime": 11.6277,
529
+ "eval_samples_per_second": 234.957,
530
+ "eval_steps_per_second": 3.698,
531
+ "step": 3415
532
+ }
533
+ ],
534
+ "logging_steps": 50,
535
+ "max_steps": 3415,
536
+ "num_input_tokens_seen": 0,
537
+ "num_train_epochs": 5,
538
+ "save_steps": 500,
539
+ "stateful_callbacks": {
540
+ "TrainerControl": {
541
+ "args": {
542
+ "should_epoch_stop": false,
543
+ "should_evaluate": false,
544
+ "should_log": false,
545
+ "should_save": true,
546
+ "should_training_stop": true
547
+ },
548
+ "attributes": {}
549
+ }
550
+ },
551
+ "total_flos": 1.43750725095936e+16,
552
+ "train_batch_size": 32,
553
+ "trial_name": null,
554
+ "trial_params": null
555
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3789039dd21b3973d94df9e7f2fa70d455e3183669f066a49936555978f42d7f
3
+ size 5777
vocab.json ADDED
The diff for this file is too large to render. See raw diff