ChiefTheLord commited on
Commit
f143704
·
verified ·
1 Parent(s): cc46ea0

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -76,3 +76,4 @@ checkpoints-v4.7/checkpoint-21504/eval_state.json filter=lfs diff=lfs merge=lfs
76
  checkpoints-d1.0/checkpoint-10240/eval_state.json filter=lfs diff=lfs merge=lfs -text
77
  checkpoints-d1.1/checkpoint-11264/eval_state.json filter=lfs diff=lfs merge=lfs -text
78
  checkpoints-d1.2/checkpoint-20480/eval_state.json filter=lfs diff=lfs merge=lfs -text
 
 
76
  checkpoints-d1.0/checkpoint-10240/eval_state.json filter=lfs diff=lfs merge=lfs -text
77
  checkpoints-d1.1/checkpoint-11264/eval_state.json filter=lfs diff=lfs merge=lfs -text
78
  checkpoints-d1.2/checkpoint-20480/eval_state.json filter=lfs diff=lfs merge=lfs -text
79
+ checkpoints-v1.1-d/checkpoint-21504/eval_state.json filter=lfs diff=lfs merge=lfs -text
checkpoints-v1.1-d/checkpoint-21504/eval_state.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6c474f7741c2445bf97dd93ee767318799592b991e90015c084cdee8a6c5632
3
+ size 54234900
checkpoints-v1.1-d/checkpoint-21504/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b745834763c9fdb686337985ad6605f8834611b1af73235af9f14be626a5868f
3
+ size 32318104
checkpoints-v1.1-d/checkpoint-21504/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12e1a1de04448acc8768ded2e3c981c82a5a0986e2ca1631b98efa665413712b
3
+ size 10010635
checkpoints-v1.1-d/checkpoint-21504/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a273e21470eca5ea2d8a06e9d04419be82b7b4a59ce87e44386508d82e9eeef
3
+ size 14645
checkpoints-v1.1-d/checkpoint-21504/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78fad30fa53fb3fbf65481787a67ed34358edcc175c7aca6268825c960127df7
3
+ size 1383
checkpoints-v1.1-d/checkpoint-21504/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8d0f20f63f45b1b1775e9f97d9c6a1a2904fa04b7d5a41f30f11b9aef0479eb
3
+ size 1465
checkpoints-v1.1-d/checkpoint-21504/trainer_state.json ADDED
@@ -0,0 +1,979 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 0.9932104752667313,
6
+ "eval_steps": 1024,
7
+ "global_step": 21504,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.011823934229365849,
14
+ "grad_norm": 0.5449081659317017,
15
+ "learning_rate": 0.000498046875,
16
+ "loss": 9.256452560424805,
17
+ "step": 256
18
+ },
19
+ {
20
+ "epoch": 0.023647868458731697,
21
+ "grad_norm": 0.5607944130897522,
22
+ "learning_rate": 0.000998046875,
23
+ "loss": 5.257084369659424,
24
+ "step": 512
25
+ },
26
+ {
27
+ "epoch": 0.03547180268809755,
28
+ "grad_norm": 0.15684783458709717,
29
+ "learning_rate": 0.000999640996023194,
30
+ "loss": 1.0911091566085815,
31
+ "step": 768
32
+ },
33
+ {
34
+ "epoch": 0.047295736917463395,
35
+ "grad_norm": 0.0985269844532013,
36
+ "learning_rate": 0.0009985588674043958,
37
+ "loss": 0.41666868329048157,
38
+ "step": 1024
39
+ },
40
+ {
41
+ "epoch": 0.047295736917463395,
42
+ "eval_bleu": 0.27265830639135136,
43
+ "eval_ce_loss": 0.3059718095480579,
44
+ "eval_loss": 0.3059718095480579,
45
+ "step": 1024
46
+ },
47
+ {
48
+ "epoch": 0.047295736917463395,
49
+ "eval_bleu": 0.27265830639135136,
50
+ "eval_ce_loss": 0.3059718095480579,
51
+ "eval_loss": 0.3059718095480579,
52
+ "eval_runtime": 113.3557,
53
+ "eval_samples_per_second": 246.948,
54
+ "eval_steps_per_second": 3.864,
55
+ "step": 1024
56
+ },
57
+ {
58
+ "epoch": 0.05911967114682925,
59
+ "grad_norm": 0.0791890025138855,
60
+ "learning_rate": 0.0009967551747861387,
61
+ "loss": 0.24503706395626068,
62
+ "step": 1280
63
+ },
64
+ {
65
+ "epoch": 0.0709436053761951,
66
+ "grad_norm": 0.07241348922252655,
67
+ "learning_rate": 0.000994232528651847,
68
+ "loss": 0.16681283712387085,
69
+ "step": 1536
70
+ },
71
+ {
72
+ "epoch": 0.08276753960556095,
73
+ "grad_norm": 0.06362218409776688,
74
+ "learning_rate": 0.0009909945800260092,
75
+ "loss": 0.12249665707349777,
76
+ "step": 1792
77
+ },
78
+ {
79
+ "epoch": 0.09459147383492679,
80
+ "grad_norm": 0.06142408400774002,
81
+ "learning_rate": 0.0009870460151900522,
82
+ "loss": 0.09425020217895508,
83
+ "step": 2048
84
+ },
85
+ {
86
+ "epoch": 0.09459147383492679,
87
+ "eval_bleu": 0.2645654621946291,
88
+ "eval_ce_loss": 0.0849783036211446,
89
+ "eval_loss": 0.0849783036211446,
90
+ "step": 2048
91
+ },
92
+ {
93
+ "epoch": 0.09459147383492679,
94
+ "eval_bleu": 0.2645654621946291,
95
+ "eval_ce_loss": 0.0849783036211446,
96
+ "eval_loss": 0.0849783036211446,
97
+ "eval_runtime": 107.8993,
98
+ "eval_samples_per_second": 259.436,
99
+ "eval_steps_per_second": 4.059,
100
+ "step": 2048
101
+ },
102
+ {
103
+ "epoch": 0.10641540806429264,
104
+ "grad_norm": 0.058095596730709076,
105
+ "learning_rate": 0.0009823925488998885,
106
+ "loss": 0.07650011032819748,
107
+ "step": 2304
108
+ },
109
+ {
110
+ "epoch": 0.1182393422936585,
111
+ "grad_norm": 0.05368533730506897,
112
+ "learning_rate": 0.0009770409161149525,
113
+ "loss": 0.06166598200798035,
114
+ "step": 2560
115
+ },
116
+ {
117
+ "epoch": 0.13006327652302435,
118
+ "grad_norm": 0.04492728039622307,
119
+ "learning_rate": 0.0009709988622506973,
120
+ "loss": 0.051953624933958054,
121
+ "step": 2816
122
+ },
123
+ {
124
+ "epoch": 0.1418872107523902,
125
+ "grad_norm": 0.038033194839954376,
126
+ "learning_rate": 0.000964275131968659,
127
+ "loss": 0.0445532500743866,
128
+ "step": 3072
129
+ },
130
+ {
131
+ "epoch": 0.1418872107523902,
132
+ "eval_bleu": 0.26919279555175596,
133
+ "eval_ce_loss": 0.04093987658858027,
134
+ "eval_loss": 0.04093987658858027,
135
+ "step": 3072
136
+ },
137
+ {
138
+ "epoch": 0.1418872107523902,
139
+ "eval_bleu": 0.26919279555175596,
140
+ "eval_ce_loss": 0.04093987658858027,
141
+ "eval_loss": 0.04093987658858027,
142
+ "eval_runtime": 110.2876,
143
+ "eval_samples_per_second": 253.818,
144
+ "eval_steps_per_second": 3.971,
145
+ "step": 3072
146
+ },
147
+ {
148
+ "epoch": 0.15371114498175603,
149
+ "grad_norm": 0.036269549280405045,
150
+ "learning_rate": 0.0009568794565203123,
151
+ "loss": 0.03803830221295357,
152
+ "step": 3328
153
+ },
154
+ {
155
+ "epoch": 0.1655350792111219,
156
+ "grad_norm": 0.03661005198955536,
157
+ "learning_rate": 0.0009488225396630347,
158
+ "loss": 0.03269872069358826,
159
+ "step": 3584
160
+ },
161
+ {
162
+ "epoch": 0.17735901344048774,
163
+ "grad_norm": 0.03384782001376152,
164
+ "learning_rate": 0.0009401160421685646,
165
+ "loss": 0.02887391857802868,
166
+ "step": 3840
167
+ },
168
+ {
169
+ "epoch": 0.18918294766985358,
170
+ "grad_norm": 0.03782425448298454,
171
+ "learning_rate": 0.0009307725649463714,
172
+ "loss": 0.025456363335251808,
173
+ "step": 4096
174
+ },
175
+ {
176
+ "epoch": 0.18918294766985358,
177
+ "eval_bleu": 0.26767754742363126,
178
+ "eval_ce_loss": 0.023863823824991647,
179
+ "eval_loss": 0.023863823824991647,
180
+ "step": 4096
181
+ },
182
+ {
183
+ "epoch": 0.18918294766985358,
184
+ "eval_bleu": 0.26767754742363126,
185
+ "eval_ce_loss": 0.023863823824991647,
186
+ "eval_loss": 0.023863823824991647,
187
+ "eval_runtime": 110.9469,
188
+ "eval_samples_per_second": 252.31,
189
+ "eval_steps_per_second": 3.948,
190
+ "step": 4096
191
+ },
192
+ {
193
+ "epoch": 0.20100688189921945,
194
+ "grad_norm": 0.043788664042949677,
195
+ "learning_rate": 0.0009208056308063659,
196
+ "loss": 0.022065965458750725,
197
+ "step": 4352
198
+ },
199
+ {
200
+ "epoch": 0.2128308161285853,
201
+ "grad_norm": 0.04368242621421814,
202
+ "learning_rate": 0.0009102296648873445,
203
+ "loss": 0.020814381539821625,
204
+ "step": 4608
205
+ },
206
+ {
207
+ "epoch": 0.22465475035795113,
208
+ "grad_norm": 0.024551063776016235,
209
+ "learning_rate": 0.0008990599737794927,
210
+ "loss": 0.01824565976858139,
211
+ "step": 4864
212
+ },
213
+ {
214
+ "epoch": 0.236478684587317,
215
+ "grad_norm": 0.02257447876036167,
216
+ "learning_rate": 0.0008873127233711644,
217
+ "loss": 0.016040779650211334,
218
+ "step": 5120
219
+ },
220
+ {
221
+ "epoch": 0.236478684587317,
222
+ "eval_bleu": 0.27054794345363903,
223
+ "eval_ce_loss": 0.01571538393092278,
224
+ "eval_loss": 0.01571538393092278,
225
+ "step": 5120
226
+ },
227
+ {
228
+ "epoch": 0.236478684587317,
229
+ "eval_bleu": 0.27054794345363903,
230
+ "eval_ce_loss": 0.01571538393092278,
231
+ "eval_loss": 0.01571538393092278,
232
+ "eval_runtime": 111.5952,
233
+ "eval_samples_per_second": 250.844,
234
+ "eval_steps_per_second": 3.925,
235
+ "step": 5120
236
+ },
237
+ {
238
+ "epoch": 0.24830261881668284,
239
+ "grad_norm": 0.021084846928715706,
240
+ "learning_rate": 0.0008750049154520011,
241
+ "loss": 0.014501616358757019,
242
+ "step": 5376
243
+ },
244
+ {
245
+ "epoch": 0.2601265530460487,
246
+ "grad_norm": 0.025962376967072487,
247
+ "learning_rate": 0.0008621543631062487,
248
+ "loss": 0.013666299171745777,
249
+ "step": 5632
250
+ },
251
+ {
252
+ "epoch": 0.27195048727541454,
253
+ "grad_norm": 0.058145921677351,
254
+ "learning_rate": 0.0008487796649318904,
255
+ "loss": 0.012593724764883518,
256
+ "step": 5888
257
+ },
258
+ {
259
+ "epoch": 0.2837744215047804,
260
+ "grad_norm": 0.02390468120574951,
261
+ "learning_rate": 0.0008349001781229053,
262
+ "loss": 0.011387365870177746,
263
+ "step": 6144
264
+ },
265
+ {
266
+ "epoch": 0.2837744215047804,
267
+ "eval_bleu": 0.2698466629682675,
268
+ "eval_ce_loss": 0.010894727121324164,
269
+ "eval_loss": 0.010894727121324164,
270
+ "step": 6144
271
+ },
272
+ {
273
+ "epoch": 0.2837744215047804,
274
+ "eval_bleu": 0.2698466629682675,
275
+ "eval_ce_loss": 0.010894727121324164,
276
+ "eval_loss": 0.010894727121324164,
277
+ "eval_runtime": 111.9404,
278
+ "eval_samples_per_second": 250.071,
279
+ "eval_steps_per_second": 3.913,
280
+ "step": 6144
281
+ },
282
+ {
283
+ "epoch": 0.2955983557341462,
284
+ "grad_norm": 0.01789081282913685,
285
+ "learning_rate": 0.0008205359904536107,
286
+ "loss": 0.010051091201603413,
287
+ "step": 6400
288
+ },
289
+ {
290
+ "epoch": 0.30742228996351206,
291
+ "grad_norm": 0.024055052548646927,
292
+ "learning_rate": 0.0008057078912056363,
293
+ "loss": 0.009441766887903214,
294
+ "step": 6656
295
+ },
296
+ {
297
+ "epoch": 0.3192462241928779,
298
+ "grad_norm": 0.025855517014861107,
299
+ "learning_rate": 0.0007904373410796086,
300
+ "loss": 0.009011849761009216,
301
+ "step": 6912
302
+ },
303
+ {
304
+ "epoch": 0.3310701584222438,
305
+ "grad_norm": 0.02206815779209137,
306
+ "learning_rate": 0.0007747464411350876,
307
+ "loss": 0.007886786945164204,
308
+ "step": 7168
309
+ },
310
+ {
311
+ "epoch": 0.3310701584222438,
312
+ "eval_bleu": 0.26735583207895347,
313
+ "eval_ce_loss": 0.007830630816449092,
314
+ "eval_loss": 0.007830630816449092,
315
+ "step": 7168
316
+ },
317
+ {
318
+ "epoch": 0.3310701584222438,
319
+ "eval_bleu": 0.26735583207895347,
320
+ "eval_ce_loss": 0.007830630816449092,
321
+ "eval_loss": 0.007830630816449092,
322
+ "eval_runtime": 115.3929,
323
+ "eval_samples_per_second": 242.589,
324
+ "eval_steps_per_second": 3.796,
325
+ "step": 7168
326
+ },
327
+ {
328
+ "epoch": 0.34289409265160964,
329
+ "grad_norm": 0.020352263003587723,
330
+ "learning_rate": 0.000758657900803716,
331
+ "loss": 0.007837384939193726,
332
+ "step": 7424
333
+ },
334
+ {
335
+ "epoch": 0.3547180268809755,
336
+ "grad_norm": 0.028763771057128906,
337
+ "learning_rate": 0.000742195005021869,
338
+ "loss": 0.007105502299964428,
339
+ "step": 7680
340
+ },
341
+ {
342
+ "epoch": 0.3665419611103413,
343
+ "grad_norm": 0.011777768842875957,
344
+ "learning_rate": 0.0007253815805303786,
345
+ "loss": 0.00655593303963542,
346
+ "step": 7936
347
+ },
348
+ {
349
+ "epoch": 0.37836589533970716,
350
+ "grad_norm": 0.014801163226366043,
351
+ "learning_rate": 0.0007082419613901028,
352
+ "loss": 0.006127453874796629,
353
+ "step": 8192
354
+ },
355
+ {
356
+ "epoch": 0.37836589533970716,
357
+ "eval_bleu": 0.26779506983265766,
358
+ "eval_ce_loss": 0.006108720087580229,
359
+ "eval_loss": 0.006108720087580229,
360
+ "step": 8192
361
+ },
362
+ {
363
+ "epoch": 0.37836589533970716,
364
+ "eval_bleu": 0.26779506983265766,
365
+ "eval_ce_loss": 0.006108720087580229,
366
+ "eval_loss": 0.006108720087580229,
367
+ "eval_runtime": 111.3262,
368
+ "eval_samples_per_second": 251.45,
369
+ "eval_steps_per_second": 3.934,
370
+ "step": 8192
371
+ },
372
+ {
373
+ "epoch": 0.390189829569073,
374
+ "grad_norm": 0.012766832485795021,
375
+ "learning_rate": 0.0006908009537632514,
376
+ "loss": 0.005777300801128149,
377
+ "step": 8448
378
+ },
379
+ {
380
+ "epoch": 0.4020137637984389,
381
+ "grad_norm": 0.010622252710163593,
382
+ "learning_rate": 0.0006730838000114403,
383
+ "loss": 0.005370937753468752,
384
+ "step": 8704
385
+ },
386
+ {
387
+ "epoch": 0.41383769802780473,
388
+ "grad_norm": 0.018417516723275185,
389
+ "learning_rate": 0.0006551161421624341,
390
+ "loss": 0.004708444699645042,
391
+ "step": 8960
392
+ },
393
+ {
394
+ "epoch": 0.4256616322571706,
395
+ "grad_norm": 0.01107096392661333,
396
+ "learning_rate": 0.0006369239847984517,
397
+ "loss": 0.004844233393669128,
398
+ "step": 9216
399
+ },
400
+ {
401
+ "epoch": 0.4256616322571706,
402
+ "eval_bleu": 0.27117611081612686,
403
+ "eval_ce_loss": 0.004634970387913814,
404
+ "eval_loss": 0.004634970387913814,
405
+ "step": 9216
406
+ },
407
+ {
408
+ "epoch": 0.4256616322571706,
409
+ "eval_bleu": 0.27117611081612686,
410
+ "eval_ce_loss": 0.004634970387913814,
411
+ "eval_loss": 0.004634970387913814,
412
+ "eval_runtime": 110.365,
413
+ "eval_samples_per_second": 253.64,
414
+ "eval_steps_per_second": 3.969,
415
+ "step": 9216
416
+ },
417
+ {
418
+ "epoch": 0.4374855664865364,
419
+ "grad_norm": 0.016654323786497116,
420
+ "learning_rate": 0.0006185336574197479,
421
+ "loss": 0.004370348993688822,
422
+ "step": 9472
423
+ },
424
+ {
425
+ "epoch": 0.44930950071590225,
426
+ "grad_norm": 0.009538416750729084,
427
+ "learning_rate": 0.0005999717763379407,
428
+ "loss": 0.004241208545863628,
429
+ "step": 9728
430
+ },
431
+ {
432
+ "epoch": 0.4611334349452681,
433
+ "grad_norm": 0.019346073269844055,
434
+ "learning_rate": 0.0005812652061542363,
435
+ "loss": 0.004045420326292515,
436
+ "step": 9984
437
+ },
438
+ {
439
+ "epoch": 0.472957369174634,
440
+ "grad_norm": 0.009237069636583328,
441
+ "learning_rate": 0.0005624410208783071,
442
+ "loss": 0.0038289830554276705,
443
+ "step": 10240
444
+ },
445
+ {
446
+ "epoch": 0.472957369174634,
447
+ "eval_bleu": 0.2709327535478657,
448
+ "eval_ce_loss": 0.0036890122933966028,
449
+ "eval_loss": 0.0036890122933966028,
450
+ "step": 10240
451
+ },
452
+ {
453
+ "epoch": 0.472957369174634,
454
+ "eval_bleu": 0.2709327535478657,
455
+ "eval_ce_loss": 0.0036890122933966028,
456
+ "eval_loss": 0.0036890122933966028,
457
+ "eval_runtime": 109.8588,
458
+ "eval_samples_per_second": 254.809,
459
+ "eval_steps_per_second": 3.987,
460
+ "step": 10240
461
+ },
462
+ {
463
+ "epoch": 0.48478130340399983,
464
+ "grad_norm": 0.018487900495529175,
465
+ "learning_rate": 0.0005435264647440881,
466
+ "loss": 0.0035509562585502863,
467
+ "step": 10496
468
+ },
469
+ {
470
+ "epoch": 0.49660523763336567,
471
+ "grad_norm": 0.007689731661230326,
472
+ "learning_rate": 0.000524548912779213,
473
+ "loss": 0.003250380977988243,
474
+ "step": 10752
475
+ },
476
+ {
477
+ "epoch": 0.5084291718627315,
478
+ "grad_norm": 0.013474254868924618,
479
+ "learning_rate": 0.0005055358311851499,
480
+ "loss": 0.003267573891207576,
481
+ "step": 11008
482
+ },
483
+ {
484
+ "epoch": 0.5202531060920974,
485
+ "grad_norm": 0.0173040684312582,
486
+ "learning_rate": 0.0004865147375853812,
487
+ "loss": 0.0030927686020731926,
488
+ "step": 11264
489
+ },
490
+ {
491
+ "epoch": 0.5202531060920974,
492
+ "eval_bleu": 0.27271556678043923,
493
+ "eval_ce_loss": 0.0031219599181127023,
494
+ "eval_loss": 0.0031219599181127023,
495
+ "step": 11264
496
+ },
497
+ {
498
+ "epoch": 0.5202531060920974,
499
+ "eval_bleu": 0.27271556678043923,
500
+ "eval_ce_loss": 0.0031219599181127023,
501
+ "eval_loss": 0.0031219599181127023,
502
+ "eval_runtime": 109.1112,
503
+ "eval_samples_per_second": 256.555,
504
+ "eval_steps_per_second": 4.014,
505
+ "step": 11264
506
+ },
507
+ {
508
+ "epoch": 0.5320770403214632,
509
+ "grad_norm": 0.01329875085502863,
510
+ "learning_rate": 0.0004675131611991607,
511
+ "loss": 0.002897108905017376,
512
+ "step": 11520
513
+ },
514
+ {
515
+ "epoch": 0.5439009745508291,
516
+ "grad_norm": 0.01354902796447277,
517
+ "learning_rate": 0.0004485586029984899,
518
+ "loss": 0.003055332228541374,
519
+ "step": 11776
520
+ },
521
+ {
522
+ "epoch": 0.5557249087801949,
523
+ "grad_norm": 0.008980591781437397,
524
+ "learning_rate": 0.00042967849590597266,
525
+ "loss": 0.0028114793822169304,
526
+ "step": 12032
527
+ },
528
+ {
529
+ "epoch": 0.5675488430095608,
530
+ "grad_norm": 0.006398347206413746,
531
+ "learning_rate": 0.0004109001650911621,
532
+ "loss": 0.002583935856819153,
533
+ "step": 12288
534
+ },
535
+ {
536
+ "epoch": 0.5675488430095608,
537
+ "eval_bleu": 0.2711715005619301,
538
+ "eval_ce_loss": 0.002555297637415718,
539
+ "eval_loss": 0.002555297637415718,
540
+ "step": 12288
541
+ },
542
+ {
543
+ "epoch": 0.5675488430095608,
544
+ "eval_bleu": 0.2711715005619301,
545
+ "eval_ce_loss": 0.002555297637415718,
546
+ "eval_loss": 0.002555297637415718,
547
+ "eval_runtime": 110.1449,
548
+ "eval_samples_per_second": 254.147,
549
+ "eval_steps_per_second": 3.977,
550
+ "step": 12288
551
+ },
552
+ {
553
+ "epoch": 0.5793727772389267,
554
+ "grad_norm": 0.006243300624191761,
555
+ "learning_rate": 0.0003922507884228551,
556
+ "loss": 0.002527546603232622,
557
+ "step": 12544
558
+ },
559
+ {
560
+ "epoch": 0.5911967114682924,
561
+ "grad_norm": 0.013985877856612206,
562
+ "learning_rate": 0.00037375735713457723,
563
+ "loss": 0.002268948359414935,
564
+ "step": 12800
565
+ },
566
+ {
567
+ "epoch": 0.6030206456976583,
568
+ "grad_norm": 0.007294178940355778,
569
+ "learning_rate": 0.00035544663676018276,
570
+ "loss": 0.0022150948643684387,
571
+ "step": 13056
572
+ },
573
+ {
574
+ "epoch": 0.6148445799270241,
575
+ "grad_norm": 0.00554023077711463,
576
+ "learning_rate": 0.00033734512839611255,
577
+ "loss": 0.0021498501300811768,
578
+ "step": 13312
579
+ },
580
+ {
581
+ "epoch": 0.6148445799270241,
582
+ "eval_bleu": 0.2683290253865085,
583
+ "eval_ce_loss": 0.0022093608622775693,
584
+ "eval_loss": 0.0022093608622775693,
585
+ "step": 13312
586
+ },
587
+ {
588
+ "epoch": 0.6148445799270241,
589
+ "eval_bleu": 0.2683290253865085,
590
+ "eval_ce_loss": 0.0022093608622775693,
591
+ "eval_loss": 0.0022093608622775693,
592
+ "eval_runtime": 110.7199,
593
+ "eval_samples_per_second": 252.827,
594
+ "eval_steps_per_second": 3.956,
595
+ "step": 13312
596
+ },
597
+ {
598
+ "epoch": 0.62666851415639,
599
+ "grad_norm": 0.013757260516285896,
600
+ "learning_rate": 0.0003194790303463687,
601
+ "loss": 0.0020817620679736137,
602
+ "step": 13568
603
+ },
604
+ {
605
+ "epoch": 0.6384924483857558,
606
+ "grad_norm": 0.019224034622311592,
607
+ "learning_rate": 0.00030187420020572406,
608
+ "loss": 0.002039685845375061,
609
+ "step": 13824
610
+ },
611
+ {
612
+ "epoch": 0.6503163826151217,
613
+ "grad_norm": 0.01222301833331585,
614
+ "learning_rate": 0.00028455611743603626,
615
+ "loss": 0.0020018373616039753,
616
+ "step": 14080
617
+ },
618
+ {
619
+ "epoch": 0.6621403168444876,
620
+ "grad_norm": 0.021646995097398758,
621
+ "learning_rate": 0.0002675498464898373,
622
+ "loss": 0.00218460732139647,
623
+ "step": 14336
624
+ },
625
+ {
626
+ "epoch": 0.6621403168444876,
627
+ "eval_bleu": 0.26793733719293295,
628
+ "eval_ce_loss": 0.0019528514221664114,
629
+ "eval_loss": 0.0019528514221664114,
630
+ "step": 14336
631
+ },
632
+ {
633
+ "epoch": 0.6621403168444876,
634
+ "eval_bleu": 0.26793733719293295,
635
+ "eval_ce_loss": 0.0019528514221664114,
636
+ "eval_loss": 0.0019528514221664114,
637
+ "eval_runtime": 110.9086,
638
+ "eval_samples_per_second": 252.397,
639
+ "eval_steps_per_second": 3.949,
640
+ "step": 14336
641
+ },
642
+ {
643
+ "epoch": 0.6739642510738534,
644
+ "grad_norm": 0.006785638630390167,
645
+ "learning_rate": 0.0002508800005345623,
646
+ "loss": 0.001926972414366901,
647
+ "step": 14592
648
+ },
649
+ {
650
+ "epoch": 0.6857881853032193,
651
+ "grad_norm": 0.008890391327440739,
652
+ "learning_rate": 0.00023457070582992562,
653
+ "loss": 0.001943480921909213,
654
+ "step": 14848
655
+ },
656
+ {
657
+ "epoch": 0.6976121195325851,
658
+ "grad_norm": 0.006759077310562134,
659
+ "learning_rate": 0.00021864556680999692,
660
+ "loss": 0.00178119249176234,
661
+ "step": 15104
662
+ },
663
+ {
664
+ "epoch": 0.709436053761951,
665
+ "grad_norm": 0.006032236386090517,
666
+ "learning_rate": 0.0002031276319205152,
667
+ "loss": 0.0017275057034566998,
668
+ "step": 15360
669
+ },
670
+ {
671
+ "epoch": 0.709436053761951,
672
+ "eval_bleu": 0.27057655967695254,
673
+ "eval_ce_loss": 0.0017077088214219951,
674
+ "eval_loss": 0.0017077088214219951,
675
+ "step": 15360
676
+ },
677
+ {
678
+ "epoch": 0.709436053761951,
679
+ "eval_bleu": 0.27057655967695254,
680
+ "eval_ce_loss": 0.0017077088214219951,
681
+ "eval_loss": 0.0017077088214219951,
682
+ "eval_runtime": 111.5469,
683
+ "eval_samples_per_second": 250.953,
684
+ "eval_steps_per_second": 3.927,
685
+ "step": 15360
686
+ },
687
+ {
688
+ "epoch": 0.7212599879913169,
689
+ "grad_norm": 0.008526836521923542,
690
+ "learning_rate": 0.00018803936026088542,
691
+ "loss": 0.001562677789479494,
692
+ "step": 15616
693
+ },
694
+ {
695
+ "epoch": 0.7330839222206826,
696
+ "grad_norm": 0.007385567296296358,
697
+ "learning_rate": 0.00017340258907913464,
698
+ "loss": 0.0016144757391884923,
699
+ "step": 15872
700
+ },
701
+ {
702
+ "epoch": 0.7449078564500485,
703
+ "grad_norm": 0.0050244200974702835,
704
+ "learning_rate": 0.0001592385021668743,
705
+ "loss": 0.0016120458021759987,
706
+ "step": 16128
707
+ },
708
+ {
709
+ "epoch": 0.7567317906794143,
710
+ "grad_norm": 0.005648311693221331,
711
+ "learning_rate": 0.0001455675992000087,
712
+ "loss": 0.001692429999820888,
713
+ "step": 16384
714
+ },
715
+ {
716
+ "epoch": 0.7567317906794143,
717
+ "eval_bleu": 0.2685717071524248,
718
+ "eval_ce_loss": 0.0015959215974420593,
719
+ "eval_loss": 0.0015959215974420593,
720
+ "step": 16384
721
+ },
722
+ {
723
+ "epoch": 0.7567317906794143,
724
+ "eval_bleu": 0.2685717071524248,
725
+ "eval_ce_loss": 0.0015959215974420593,
726
+ "eval_loss": 0.0015959215974420593,
727
+ "eval_runtime": 109.7001,
728
+ "eval_samples_per_second": 255.177,
729
+ "eval_steps_per_second": 3.993,
730
+ "step": 16384
731
+ },
732
+ {
733
+ "epoch": 0.7685557249087802,
734
+ "grad_norm": 0.00418821582570672,
735
+ "learning_rate": 0.000132409666069565,
736
+ "loss": 0.001514198025688529,
737
+ "step": 16640
738
+ },
739
+ {
740
+ "epoch": 0.780379659138146,
741
+ "grad_norm": 0.008380233310163021,
742
+ "learning_rate": 0.0001197837462455823,
743
+ "loss": 0.0014031081227585673,
744
+ "step": 16896
745
+ },
746
+ {
747
+ "epoch": 0.7922035933675119,
748
+ "grad_norm": 0.008456946350634098,
749
+ "learning_rate": 0.00010770811321550749,
750
+ "loss": 0.0014507079031318426,
751
+ "step": 17152
752
+ },
753
+ {
754
+ "epoch": 0.8040275275968778,
755
+ "grad_norm": 0.01046363078057766,
756
+ "learning_rate": 9.620024403698591e-05,
757
+ "loss": 0.0016408010851591825,
758
+ "step": 17408
759
+ },
760
+ {
761
+ "epoch": 0.8040275275968778,
762
+ "eval_bleu": 0.2652117482429029,
763
+ "eval_ce_loss": 0.0014940057051086057,
764
+ "eval_loss": 0.0014940057051086057,
765
+ "step": 17408
766
+ },
767
+ {
768
+ "epoch": 0.8040275275968778,
769
+ "eval_bleu": 0.2652117482429029,
770
+ "eval_ce_loss": 0.0014940057051086057,
771
+ "eval_loss": 0.0014940057051086057,
772
+ "eval_runtime": 109.1029,
773
+ "eval_samples_per_second": 256.574,
774
+ "eval_steps_per_second": 4.015,
775
+ "step": 17408
776
+ },
777
+ {
778
+ "epoch": 0.8158514618262436,
779
+ "grad_norm": 0.003936219960451126,
780
+ "learning_rate": 8.527679404332429e-05,
781
+ "loss": 0.001487646484747529,
782
+ "step": 17664
783
+ },
784
+ {
785
+ "epoch": 0.8276753960556095,
786
+ "grad_norm": 0.030263634398579597,
787
+ "learning_rate": 7.495357273823544e-05,
788
+ "loss": 0.0014763937797397375,
789
+ "step": 17920
790
+ },
791
+ {
792
+ "epoch": 0.8394993302849753,
793
+ "grad_norm": 0.021031692624092102,
794
+ "learning_rate": 6.524552091475183e-05,
795
+ "loss": 0.001435705809853971,
796
+ "step": 18176
797
+ },
798
+ {
799
+ "epoch": 0.8513232645143411,
800
+ "grad_norm": 0.010445632040500641,
801
+ "learning_rate": 5.6166689031422024e-05,
802
+ "loss": 0.001405209768563509,
803
+ "step": 18432
804
+ },
805
+ {
806
+ "epoch": 0.8513232645143411,
807
+ "eval_bleu": 0.2666050549594915,
808
+ "eval_ce_loss": 0.0014094488985837038,
809
+ "eval_loss": 0.0014094488985837038,
810
+ "step": 18432
811
+ },
812
+ {
813
+ "epoch": 0.8513232645143411,
814
+ "eval_bleu": 0.2666050549594915,
815
+ "eval_ce_loss": 0.0014094488985837038,
816
+ "eval_loss": 0.0014094488985837038,
817
+ "eval_runtime": 110.9263,
818
+ "eval_samples_per_second": 252.357,
819
+ "eval_steps_per_second": 3.949,
820
+ "step": 18432
821
+ },
822
+ {
823
+ "epoch": 0.8631471987437069,
824
+ "grad_norm": 0.0071349553763866425,
825
+ "learning_rate": 4.773021687709067e-05,
826
+ "loss": 0.0014939571265131235,
827
+ "step": 18688
828
+ },
829
+ {
830
+ "epoch": 0.8749711329730728,
831
+ "grad_norm": 0.0075807152315974236,
832
+ "learning_rate": 3.994831455368719e-05,
833
+ "loss": 0.0016243808204308152,
834
+ "step": 18944
835
+ },
836
+ {
837
+ "epoch": 0.8867950672024387,
838
+ "grad_norm": 0.003486819099634886,
839
+ "learning_rate": 3.283224480455282e-05,
840
+ "loss": 0.0014689115341752768,
841
+ "step": 19200
842
+ },
843
+ {
844
+ "epoch": 0.8986190014318045,
845
+ "grad_norm": 0.004220427479594946,
846
+ "learning_rate": 2.639230671387627e-05,
847
+ "loss": 0.0012729011941701174,
848
+ "step": 19456
849
+ },
850
+ {
851
+ "epoch": 0.8986190014318045,
852
+ "eval_bleu": 0.2695832110812109,
853
+ "eval_ce_loss": 0.0013992262525565297,
854
+ "eval_loss": 0.0013992262525565297,
855
+ "step": 19456
856
+ },
857
+ {
858
+ "epoch": 0.8986190014318045,
859
+ "eval_bleu": 0.2695832110812109,
860
+ "eval_ce_loss": 0.0013992262525565297,
861
+ "eval_loss": 0.0013992262525565297,
862
+ "eval_runtime": 110.239,
863
+ "eval_samples_per_second": 253.93,
864
+ "eval_steps_per_second": 3.973,
865
+ "step": 19456
866
+ },
867
+ {
868
+ "epoch": 0.9104429356611704,
869
+ "grad_norm": 0.013082730583846569,
870
+ "learning_rate": 2.063782080083576e-05,
871
+ "loss": 0.0013599519152194262,
872
+ "step": 19712
873
+ },
874
+ {
875
+ "epoch": 0.9222668698905362,
876
+ "grad_norm": 0.00748586468398571,
877
+ "learning_rate": 1.557711553001523e-05,
878
+ "loss": 0.0015039942227303982,
879
+ "step": 19968
880
+ },
881
+ {
882
+ "epoch": 0.9340908041199021,
883
+ "grad_norm": 0.003521893871948123,
884
+ "learning_rate": 1.1217515257622269e-05,
885
+ "loss": 0.0014425483532249928,
886
+ "step": 20224
887
+ },
888
+ {
889
+ "epoch": 0.945914738349268,
890
+ "grad_norm": 0.008369974792003632,
891
+ "learning_rate": 7.565329630950746e-06,
892
+ "loss": 0.0014663139590993524,
893
+ "step": 20480
894
+ },
895
+ {
896
+ "epoch": 0.945914738349268,
897
+ "eval_bleu": 0.2669610603654314,
898
+ "eval_ce_loss": 0.0013518765415306859,
899
+ "eval_loss": 0.0013518765415306859,
900
+ "step": 20480
901
+ },
902
+ {
903
+ "epoch": 0.945914738349268,
904
+ "eval_bleu": 0.2669610603654314,
905
+ "eval_ce_loss": 0.0013518765415306859,
906
+ "eval_loss": 0.0013518765415306859,
907
+ "eval_runtime": 112.2477,
908
+ "eval_samples_per_second": 249.386,
909
+ "eval_steps_per_second": 3.902,
910
+ "step": 20480
911
+ },
912
+ {
913
+ "epoch": 0.9577386725786338,
914
+ "grad_norm": 0.008028030395507812,
915
+ "learning_rate": 4.62584445643166e-06,
916
+ "loss": 0.0015790105098858476,
917
+ "step": 20736
918
+ },
919
+ {
920
+ "epoch": 0.9695626068079997,
921
+ "grad_norm": 0.005876564886420965,
922
+ "learning_rate": 2.40331404948807e-06,
923
+ "loss": 0.001455229939892888,
924
+ "step": 20992
925
+ },
926
+ {
927
+ "epoch": 0.9813865410373654,
928
+ "grad_norm": 0.008718357421457767,
929
+ "learning_rate": 9.009550772663965e-07,
930
+ "loss": 0.0013052865397185087,
931
+ "step": 21248
932
+ },
933
+ {
934
+ "epoch": 0.9932104752667313,
935
+ "grad_norm": 0.0048894439823925495,
936
+ "learning_rate": 1.2094190315575791e-07,
937
+ "loss": 0.0014210316585376859,
938
+ "step": 21504
939
+ },
940
+ {
941
+ "epoch": 0.9932104752667313,
942
+ "eval_bleu": 0.2713391630531368,
943
+ "eval_ce_loss": 0.001370158953104858,
944
+ "eval_loss": 0.001370158953104858,
945
+ "step": 21504
946
+ },
947
+ {
948
+ "epoch": 0.9932104752667313,
949
+ "eval_bleu": 0.2713391630531368,
950
+ "eval_ce_loss": 0.001370158953104858,
951
+ "eval_loss": 0.001370158953104858,
952
+ "eval_runtime": 108.048,
953
+ "eval_samples_per_second": 259.079,
954
+ "eval_steps_per_second": 4.054,
955
+ "step": 21504
956
+ }
957
+ ],
958
+ "logging_steps": 256,
959
+ "max_steps": 21651,
960
+ "num_input_tokens_seen": 0,
961
+ "num_train_epochs": 1,
962
+ "save_steps": 1024,
963
+ "stateful_callbacks": {
964
+ "TrainerControl": {
965
+ "args": {
966
+ "should_epoch_stop": false,
967
+ "should_evaluate": false,
968
+ "should_log": false,
969
+ "should_save": true,
970
+ "should_training_stop": false
971
+ },
972
+ "attributes": {}
973
+ }
974
+ },
975
+ "total_flos": 0.0,
976
+ "train_batch_size": 64,
977
+ "trial_name": null,
978
+ "trial_params": null
979
+ }
checkpoints-v1.1-d/checkpoint-21504/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61a16bb839f687414b8e48611327c4b9cfddeefe38c031ca70808f9a97c476b7
3
+ size 5137