SummerSigh commited on
Commit
83a9cd4
1 Parent(s): daa6298

Upload 8 files

Browse files
Files changed (7) hide show
  1. config.json +1 -0
  2. model.safetensors +1 -1
  3. optimizer.pt +1 -1
  4. rng_state.pth +1 -1
  5. scheduler.pt +1 -1
  6. trainer_state.json +1149 -405
  7. training_args.bin +1 -1
config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "architectures": [
3
  "LlamaForCausalLM"
4
  ],
 
1
  {
2
+ "_name_or_path": "SummerSigh/SimpleLlamaSentences",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d3c2e20e14ad149e73575ba64b05ad353625b797ea427a99990d68008d3d3cb
3
  size 18494040
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63c707b72c3779662a4c5641912d3bd484e848cda99f1f5fa45be2dfd8a0fef8
3
  size 18494040
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9bda737f67c64070b80a0458ecda2fcea611bd2d8344185d36fb69035c97d221
3
  size 37035002
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadf442c9909d4e1c2b9e49f1d628a9826771fbfb353eea8aaebd90c70f08957
3
  size 37035002
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1ff264f99d31b522cc7e2a4eac9d38606d0c58a34c0adc74d71e0ca8b371dc36
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:386fcc8cc1089aade9450d86fb239ea3483f455fd2d78d8378645feecfec9d69
3
  size 14244
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1af92ed44d240f5ab8a8e4f68304482a678ec63e16e1ccdf7c4a711760c469a9
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7341b3083234687d0b9b4d7741daf02731dfc2e5ecfee8b661951c827d79431
3
  size 1064
trainer_state.json CHANGED
@@ -1,657 +1,1401 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.5045606996207255,
5
  "eval_steps": 500,
6
- "global_step": 12000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.018807008745259066,
13
- "grad_norm": 0.8045865297317505,
14
- "learning_rate": 0.00015902345219681452,
15
- "loss": 9.3828,
16
- "num_input_tokens_seen": 19660800,
17
  "step": 150
18
  },
19
  {
20
- "epoch": 0.03761401749051813,
21
- "grad_norm": 0.8349803686141968,
22
- "learning_rate": 0.0001580201496592952,
23
- "loss": 8.0829,
24
- "num_input_tokens_seen": 39321600,
25
  "step": 300
26
  },
27
  {
28
- "epoch": 0.0564210262357772,
29
- "grad_norm": 0.6253471374511719,
30
- "learning_rate": 0.00015701684712177586,
31
- "loss": 7.4091,
32
- "num_input_tokens_seen": 58982400,
33
  "step": 450
34
  },
35
  {
36
- "epoch": 0.07522803498103627,
37
- "grad_norm": 0.5699061751365662,
38
- "learning_rate": 0.00015601354458425653,
39
- "loss": 7.0279,
40
- "num_input_tokens_seen": 78643200,
41
  "step": 600
42
  },
43
  {
44
- "epoch": 0.09403504372629533,
45
- "grad_norm": 0.6017518043518066,
46
- "learning_rate": 0.0001550102420467372,
47
- "loss": 6.732,
48
- "num_input_tokens_seen": 98304000,
49
  "step": 750
50
  },
51
  {
52
- "epoch": 0.1128420524715544,
53
- "grad_norm": 0.6045653223991394,
54
- "learning_rate": 0.00015400693950921786,
55
- "loss": 6.4612,
56
- "num_input_tokens_seen": 117964800,
57
  "step": 900
58
  },
59
  {
60
- "epoch": 0.13164906121681347,
61
- "grad_norm": 0.8820130825042725,
62
- "learning_rate": 0.00015300363697169853,
63
- "loss": 6.2193,
64
- "num_input_tokens_seen": 137625600,
65
  "step": 1050
66
  },
67
  {
68
- "epoch": 0.15045606996207253,
69
- "grad_norm": 0.8934310674667358,
70
- "learning_rate": 0.00015200033443417917,
71
- "loss": 6.0184,
72
- "num_input_tokens_seen": 157286400,
73
  "step": 1200
74
  },
75
  {
76
- "epoch": 0.1692630787073316,
77
- "grad_norm": 1.082834243774414,
78
- "learning_rate": 0.00015099703189665986,
79
- "loss": 5.8664,
80
- "num_input_tokens_seen": 176947200,
81
  "step": 1350
82
  },
83
  {
84
- "epoch": 0.18807008745259066,
85
- "grad_norm": 1.1755659580230713,
86
- "learning_rate": 0.00014999372935914053,
87
- "loss": 5.7258,
88
- "num_input_tokens_seen": 196608000,
89
  "step": 1500
90
  },
91
  {
92
- "epoch": 0.20687709619784972,
93
- "grad_norm": 1.1867622137069702,
94
- "learning_rate": 0.00014899042682162117,
95
- "loss": 5.6018,
96
- "num_input_tokens_seen": 216268800,
97
  "step": 1650
98
  },
99
  {
100
- "epoch": 0.2256841049431088,
101
- "grad_norm": 1.2599449157714844,
102
- "learning_rate": 0.00014798712428410184,
103
- "loss": 5.5016,
104
- "num_input_tokens_seen": 235929600,
105
  "step": 1800
106
  },
107
  {
108
- "epoch": 0.24449111368836787,
109
- "grad_norm": 1.292974591255188,
110
- "learning_rate": 0.0001469838217465825,
111
- "loss": 5.413,
112
- "num_input_tokens_seen": 255590400,
113
  "step": 1950
114
  },
115
  {
116
- "epoch": 0.26329812243362694,
117
- "grad_norm": 1.325179100036621,
118
- "learning_rate": 0.00014598051920906318,
119
- "loss": 5.3436,
120
- "num_input_tokens_seen": 275251200,
121
  "step": 2100
122
  },
123
  {
124
- "epoch": 0.28210513117888597,
125
- "grad_norm": 1.3993710279464722,
126
- "learning_rate": 0.00014497721667154384,
127
- "loss": 5.2768,
128
- "num_input_tokens_seen": 294912000,
129
  "step": 2250
130
  },
131
  {
132
- "epoch": 0.30091213992414506,
133
- "grad_norm": 1.5274803638458252,
134
- "learning_rate": 0.0001439739141340245,
135
- "loss": 5.2122,
136
- "num_input_tokens_seen": 314572800,
137
  "step": 2400
138
  },
139
  {
140
- "epoch": 0.31971914866940415,
141
- "grad_norm": 1.475855827331543,
142
- "learning_rate": 0.00014297061159650518,
143
- "loss": 5.1714,
144
- "num_input_tokens_seen": 334233600,
145
  "step": 2550
146
  },
147
  {
148
- "epoch": 0.3385261574146632,
149
- "grad_norm": 1.5435775518417358,
150
- "learning_rate": 0.00014196730905898585,
151
- "loss": 5.1111,
152
- "num_input_tokens_seen": 353894400,
153
  "step": 2700
154
  },
155
  {
156
- "epoch": 0.3573331661599223,
157
- "grad_norm": 1.6586663722991943,
158
- "learning_rate": 0.0001409640065214665,
159
- "loss": 5.0548,
160
- "num_input_tokens_seen": 373555200,
161
  "step": 2850
162
  },
163
  {
164
- "epoch": 0.3761401749051813,
165
- "grad_norm": 1.5785890817642212,
166
- "learning_rate": 0.00013996070398394718,
167
- "loss": 5.0055,
168
- "num_input_tokens_seen": 393216000,
169
  "step": 3000
170
  },
171
  {
172
- "epoch": 0.3949471836504404,
173
- "grad_norm": 1.6534926891326904,
174
- "learning_rate": 0.00013895740144642785,
175
- "loss": 4.973,
176
- "num_input_tokens_seen": 412876800,
177
  "step": 3150
178
  },
179
  {
180
- "epoch": 0.41375419239569944,
181
- "grad_norm": 1.82899010181427,
182
- "learning_rate": 0.0001379540989089085,
183
- "loss": 4.9336,
184
- "num_input_tokens_seen": 432537600,
185
  "step": 3300
186
  },
187
  {
188
- "epoch": 0.43256120114095853,
189
- "grad_norm": 1.7137963771820068,
190
- "learning_rate": 0.00013695079637138916,
191
- "loss": 4.9158,
192
- "num_input_tokens_seen": 452198400,
193
  "step": 3450
194
  },
195
  {
196
- "epoch": 0.4513682098862176,
197
- "grad_norm": 2.001296043395996,
198
- "learning_rate": 0.00013594749383386985,
199
- "loss": 4.8633,
200
- "num_input_tokens_seen": 471859200,
201
  "step": 3600
202
  },
203
  {
204
- "epoch": 0.47017521863147665,
205
- "grad_norm": 1.7802987098693848,
206
- "learning_rate": 0.0001349441912963505,
207
- "loss": 4.8431,
208
- "num_input_tokens_seen": 491520000,
209
  "step": 3750
210
  },
211
  {
212
- "epoch": 0.48898222737673575,
213
- "grad_norm": 1.8677113056182861,
214
- "learning_rate": 0.00013394088875883116,
215
- "loss": 4.8243,
216
- "num_input_tokens_seen": 511180800,
217
  "step": 3900
218
  },
219
  {
220
- "epoch": 0.5077892361219948,
221
- "grad_norm": 1.9465229511260986,
222
- "learning_rate": 0.00013293758622131183,
223
- "loss": 4.7799,
224
- "num_input_tokens_seen": 530841600,
225
  "step": 4050
226
  },
227
  {
228
- "epoch": 0.5265962448672539,
229
- "grad_norm": 1.9226826429367065,
230
- "learning_rate": 0.0001319342836837925,
231
- "loss": 4.7513,
232
- "num_input_tokens_seen": 550502400,
233
  "step": 4200
234
  },
235
  {
236
- "epoch": 0.545403253612513,
237
- "grad_norm": 2.018714666366577,
238
- "learning_rate": 0.00013093098114627316,
239
- "loss": 4.7226,
240
- "num_input_tokens_seen": 570163200,
241
  "step": 4350
242
  },
243
  {
244
- "epoch": 0.5642102623577719,
245
- "grad_norm": 1.975203275680542,
246
- "learning_rate": 0.00012992767860875383,
247
- "loss": 4.7069,
248
- "num_input_tokens_seen": 589824000,
249
  "step": 4500
250
  },
251
  {
252
- "epoch": 0.583017271103031,
253
- "grad_norm": 1.9473543167114258,
254
- "learning_rate": 0.0001289243760712345,
255
- "loss": 4.6781,
256
- "num_input_tokens_seen": 609484800,
257
  "step": 4650
258
  },
259
  {
260
- "epoch": 0.6018242798482901,
261
- "grad_norm": 1.9783662557601929,
262
- "learning_rate": 0.00012792107353371517,
263
- "loss": 4.653,
264
- "num_input_tokens_seen": 629145600,
265
  "step": 4800
266
  },
267
  {
268
- "epoch": 0.6206312885935492,
269
- "grad_norm": 2.184589147567749,
270
- "learning_rate": 0.00012691777099619583,
271
- "loss": 4.6403,
272
- "num_input_tokens_seen": 648806400,
273
  "step": 4950
274
  },
275
  {
276
- "epoch": 0.6394382973388083,
277
- "grad_norm": 2.109097480773926,
278
- "learning_rate": 0.00012591446845867647,
279
- "loss": 4.613,
280
- "num_input_tokens_seen": 668467200,
281
  "step": 5100
282
  },
283
  {
284
- "epoch": 0.6582453060840673,
285
- "grad_norm": 2.1296463012695312,
286
- "learning_rate": 0.00012491116592115717,
287
- "loss": 4.5984,
288
- "num_input_tokens_seen": 688128000,
289
  "step": 5250
290
  },
291
  {
292
- "epoch": 0.6770523148293264,
293
- "grad_norm": 2.2577426433563232,
294
- "learning_rate": 0.0001239078633836378,
295
- "loss": 4.5816,
296
- "num_input_tokens_seen": 707788800,
297
  "step": 5400
298
  },
299
  {
300
- "epoch": 0.6958593235745855,
301
- "grad_norm": 2.186983585357666,
302
- "learning_rate": 0.00012290456084611848,
303
- "loss": 4.5545,
304
- "num_input_tokens_seen": 727449600,
305
  "step": 5550
306
  },
307
  {
308
- "epoch": 0.7146663323198446,
309
- "grad_norm": 2.1596477031707764,
310
- "learning_rate": 0.00012190125830859916,
311
- "loss": 4.5328,
312
- "num_input_tokens_seen": 747110400,
313
  "step": 5700
314
  },
315
  {
316
- "epoch": 0.7334733410651036,
317
- "grad_norm": 2.112995147705078,
318
- "learning_rate": 0.00012089795577107981,
319
- "loss": 4.5176,
320
- "num_input_tokens_seen": 766771200,
321
  "step": 5850
322
  },
323
  {
324
- "epoch": 0.7522803498103626,
325
- "grad_norm": 2.322986125946045,
326
- "learning_rate": 0.00011989465323356048,
327
- "loss": 4.5145,
328
- "num_input_tokens_seen": 786432000,
329
  "step": 6000
330
  },
331
  {
332
- "epoch": 0.7710873585556217,
333
- "grad_norm": 2.2006442546844482,
334
- "learning_rate": 0.00011889135069604115,
335
- "loss": 4.4968,
336
- "num_input_tokens_seen": 806092800,
337
  "step": 6150
338
  },
339
  {
340
- "epoch": 0.7898943673008808,
341
- "grad_norm": 2.222529411315918,
342
- "learning_rate": 0.0001178880481585218,
343
- "loss": 4.4817,
344
- "num_input_tokens_seen": 825753600,
345
  "step": 6300
346
  },
347
  {
348
- "epoch": 0.8087013760461399,
349
- "grad_norm": 2.286618947982788,
350
- "learning_rate": 0.00011688474562100248,
351
- "loss": 4.4596,
352
- "num_input_tokens_seen": 845414400,
353
  "step": 6450
354
  },
355
  {
356
- "epoch": 0.8275083847913989,
357
- "grad_norm": 2.36089825630188,
358
- "learning_rate": 0.0001158881317670666,
359
- "loss": 4.4455,
360
- "num_input_tokens_seen": 865075200,
361
  "step": 6600
362
  },
363
  {
364
- "epoch": 0.846315393536658,
365
- "grad_norm": 2.4029579162597656,
366
- "learning_rate": 0.00011488482922954726,
367
- "loss": 4.4359,
368
- "num_input_tokens_seen": 884736000,
369
  "step": 6750
370
  },
371
  {
372
- "epoch": 0.8651224022819171,
373
- "grad_norm": 2.3895576000213623,
374
- "learning_rate": 0.00011388152669202794,
375
- "loss": 4.4152,
376
- "num_input_tokens_seen": 904396800,
377
  "step": 6900
378
  },
379
  {
380
- "epoch": 0.8839294110271761,
381
- "grad_norm": 2.2175564765930176,
382
- "learning_rate": 0.0001128782241545086,
383
- "loss": 4.4096,
384
- "num_input_tokens_seen": 924057600,
385
  "step": 7050
386
  },
387
  {
388
- "epoch": 0.9027364197724352,
389
- "grad_norm": 2.4382975101470947,
390
- "learning_rate": 0.00011187492161698926,
391
- "loss": 4.3976,
392
- "num_input_tokens_seen": 943718400,
393
  "step": 7200
394
  },
395
  {
396
- "epoch": 0.9215434285176942,
397
- "grad_norm": 2.533416986465454,
398
- "learning_rate": 0.00011087161907946993,
399
- "loss": 4.384,
400
- "num_input_tokens_seen": 963379200,
401
  "step": 7350
402
  },
403
  {
404
- "epoch": 0.9403504372629533,
405
- "grad_norm": 2.362147808074951,
406
- "learning_rate": 0.00010986831654195061,
407
- "loss": 4.3608,
408
- "num_input_tokens_seen": 983040000,
409
  "step": 7500
410
  },
411
  {
412
- "epoch": 0.9591574460082124,
413
- "grad_norm": 2.4153311252593994,
414
- "learning_rate": 0.00010886501400443126,
415
- "loss": 4.3604,
416
- "num_input_tokens_seen": 1002700800,
417
  "step": 7650
418
  },
419
  {
420
- "epoch": 0.9779644547534715,
421
- "grad_norm": 2.541653871536255,
422
- "learning_rate": 0.00010786171146691193,
423
- "loss": 4.3495,
424
- "num_input_tokens_seen": 1022361600,
425
  "step": 7800
426
  },
427
  {
428
- "epoch": 0.9967714634987306,
429
- "grad_norm": 2.507594108581543,
430
- "learning_rate": 0.0001068584089293926,
431
- "loss": 4.3303,
432
- "num_input_tokens_seen": 1042022400,
433
  "step": 7950
434
  },
435
  {
436
- "epoch": 1.0155784722439896,
437
- "grad_norm": 2.4842772483825684,
438
- "learning_rate": 0.00010585510639187325,
439
- "loss": 4.3045,
440
- "num_input_tokens_seen": 1061654528,
441
  "step": 8100
442
  },
443
  {
444
- "epoch": 1.0343854809892488,
445
- "grad_norm": 2.3852617740631104,
446
- "learning_rate": 0.00010485180385435392,
447
- "loss": 4.3033,
448
- "num_input_tokens_seen": 1081315328,
449
  "step": 8250
450
  },
451
  {
452
- "epoch": 1.0531924897345077,
453
- "grad_norm": 2.6390843391418457,
454
- "learning_rate": 0.00010384850131683457,
455
- "loss": 4.2792,
456
- "num_input_tokens_seen": 1100976128,
457
  "step": 8400
458
  },
459
  {
460
- "epoch": 1.0719994984797667,
461
- "grad_norm": 2.5094408988952637,
462
- "learning_rate": 0.00010284519877931526,
463
- "loss": 4.2826,
464
- "num_input_tokens_seen": 1120636928,
465
  "step": 8550
466
  },
467
  {
468
- "epoch": 1.090806507225026,
469
- "grad_norm": 2.564629554748535,
470
- "learning_rate": 0.00010184858492537939,
471
- "loss": 4.2866,
472
- "num_input_tokens_seen": 1140297728,
473
  "step": 8700
474
  },
475
  {
476
- "epoch": 1.109613515970285,
477
- "grad_norm": 2.6148977279663086,
478
- "learning_rate": 0.00010084528238786004,
479
- "loss": 4.2614,
480
- "num_input_tokens_seen": 1159958528,
481
  "step": 8850
482
  },
483
  {
484
- "epoch": 1.128420524715544,
485
- "grad_norm": 2.5449106693267822,
486
- "learning_rate": 9.984197985034071e-05,
487
- "loss": 4.2622,
488
- "num_input_tokens_seen": 1179619328,
489
  "step": 9000
490
  },
491
  {
492
- "epoch": 1.147227533460803,
493
- "grad_norm": 2.903010129928589,
494
- "learning_rate": 9.883867731282138e-05,
495
- "loss": 4.2427,
496
- "num_input_tokens_seen": 1199280128,
497
  "step": 9150
498
  },
499
  {
500
- "epoch": 1.166034542206062,
501
- "grad_norm": 2.6503586769104004,
502
- "learning_rate": 9.783537477530203e-05,
503
- "loss": 4.2433,
504
- "num_input_tokens_seen": 1218940928,
505
  "step": 9300
506
  },
507
  {
508
- "epoch": 1.1848415509513213,
509
- "grad_norm": 2.6648051738739014,
510
- "learning_rate": 9.68320722377827e-05,
511
- "loss": 4.223,
512
- "num_input_tokens_seen": 1238601728,
513
  "step": 9450
514
  },
515
  {
516
- "epoch": 1.2036485596965802,
517
- "grad_norm": 2.688499689102173,
518
- "learning_rate": 9.582876970026338e-05,
519
- "loss": 4.2232,
520
- "num_input_tokens_seen": 1258262528,
521
  "step": 9600
522
  },
523
  {
524
- "epoch": 1.2224555684418394,
525
- "grad_norm": 2.5007975101470947,
526
- "learning_rate": 9.482546716274404e-05,
527
- "loss": 4.2098,
528
- "num_input_tokens_seen": 1277923328,
529
  "step": 9750
530
  },
531
  {
532
- "epoch": 1.2412625771870984,
533
- "grad_norm": 2.682922124862671,
534
- "learning_rate": 9.38221646252247e-05,
535
- "loss": 4.2144,
536
- "num_input_tokens_seen": 1297584128,
537
  "step": 9900
538
  },
539
  {
540
- "epoch": 1.2600695859323574,
541
- "grad_norm": 2.683807611465454,
542
- "learning_rate": 9.281886208770537e-05,
543
- "loss": 4.1964,
544
- "num_input_tokens_seen": 1317244928,
545
  "step": 10050
546
  },
547
  {
548
- "epoch": 1.2788765946776166,
549
- "grad_norm": 2.7165095806121826,
550
- "learning_rate": 9.181555955018603e-05,
551
- "loss": 4.1939,
552
- "num_input_tokens_seen": 1336905728,
553
  "step": 10200
554
  },
555
  {
556
- "epoch": 1.2976836034228756,
557
- "grad_norm": 2.791292667388916,
558
- "learning_rate": 9.081225701266671e-05,
559
- "loss": 4.1796,
560
- "num_input_tokens_seen": 1356566528,
561
  "step": 10350
562
  },
563
  {
564
- "epoch": 1.3164906121681348,
565
- "grad_norm": 2.750148057937622,
566
- "learning_rate": 8.980895447514737e-05,
567
- "loss": 4.1778,
568
- "num_input_tokens_seen": 1376227328,
569
  "step": 10500
570
  },
571
  {
572
- "epoch": 1.3352976209133938,
573
- "grad_norm": 2.673332929611206,
574
- "learning_rate": 8.880565193762803e-05,
575
- "loss": 4.1669,
576
- "num_input_tokens_seen": 1395888128,
577
  "step": 10650
578
  },
579
  {
580
- "epoch": 1.3541046296586527,
581
- "grad_norm": 2.8640031814575195,
582
- "learning_rate": 8.780903808369216e-05,
583
- "loss": 4.167,
584
- "num_input_tokens_seen": 1415548928,
585
  "step": 10800
586
  },
587
  {
588
- "epoch": 1.3729116384039117,
589
- "grad_norm": 2.737783432006836,
590
- "learning_rate": 8.680573554617283e-05,
591
- "loss": 4.1697,
592
- "num_input_tokens_seen": 1435209728,
593
  "step": 10950
594
  },
595
  {
596
- "epoch": 1.391718647149171,
597
- "grad_norm": 2.7067887783050537,
598
- "learning_rate": 8.580243300865349e-05,
599
- "loss": 4.1542,
600
- "num_input_tokens_seen": 1454870528,
601
  "step": 11100
602
  },
603
  {
604
- "epoch": 1.41052565589443,
605
- "grad_norm": 2.843256711959839,
606
- "learning_rate": 8.479913047113415e-05,
607
- "loss": 4.1566,
608
- "num_input_tokens_seen": 1474531328,
609
  "step": 11250
610
  },
611
  {
612
- "epoch": 1.429332664639689,
613
- "grad_norm": 2.783827543258667,
614
- "learning_rate": 8.379582793361483e-05,
615
- "loss": 4.1527,
616
- "num_input_tokens_seen": 1494192128,
617
  "step": 11400
618
  },
619
  {
620
- "epoch": 1.448139673384948,
621
- "grad_norm": 2.912080764770508,
622
- "learning_rate": 8.279252539609549e-05,
623
- "loss": 4.1308,
624
- "num_input_tokens_seen": 1513852928,
625
  "step": 11550
626
  },
627
  {
628
- "epoch": 1.466946682130207,
629
- "grad_norm": 2.7934062480926514,
630
- "learning_rate": 8.178922285857616e-05,
631
- "loss": 4.134,
632
- "num_input_tokens_seen": 1533513728,
633
  "step": 11700
634
  },
635
  {
636
- "epoch": 1.4857536908754663,
637
- "grad_norm": 3.1042609214782715,
638
- "learning_rate": 8.078592032105681e-05,
639
- "loss": 4.1365,
640
- "num_input_tokens_seen": 1553174528,
641
  "step": 11850
642
  },
643
  {
644
- "epoch": 1.5045606996207255,
645
- "grad_norm": 3.0361480712890625,
646
- "learning_rate": 7.978261778353748e-05,
647
- "loss": 4.1343,
648
- "num_input_tokens_seen": 1572835328,
649
  "step": 12000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
650
  }
651
  ],
652
  "logging_steps": 150,
653
- "max_steps": 23925,
654
- "num_input_tokens_seen": 1572835328,
655
  "num_train_epochs": 3,
656
  "save_steps": 500,
657
  "stateful_callbacks": {
@@ -666,7 +1410,7 @@
666
  "attributes": {}
667
  }
668
  },
669
- "total_flos": 2.428558407892992e+16,
670
  "train_batch_size": 32,
671
  "trial_name": null,
672
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.28651874218272183,
5
  "eval_steps": 500,
6
+ "global_step": 26000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.0016529927433618567,
13
+ "grad_norm": 2.137389659881592,
14
+ "learning_rate": 0.0001599141895763845,
15
+ "loss": 6.3906,
16
+ "num_input_tokens_seen": 872064,
17
  "step": 150
18
  },
19
  {
20
+ "epoch": 0.0033059854867237134,
21
+ "grad_norm": 1.725138783454895,
22
+ "learning_rate": 0.000159826028182259,
23
+ "loss": 5.6466,
24
+ "num_input_tokens_seen": 1744032,
25
  "step": 300
26
  },
27
  {
28
+ "epoch": 0.00495897823008557,
29
+ "grad_norm": 1.9830477237701416,
30
+ "learning_rate": 0.00015973786678813348,
31
+ "loss": 5.4449,
32
+ "num_input_tokens_seen": 2615552,
33
  "step": 450
34
  },
35
  {
36
+ "epoch": 0.006611970973447427,
37
+ "grad_norm": 1.8168050050735474,
38
+ "learning_rate": 0.00015964970539400798,
39
+ "loss": 5.3521,
40
+ "num_input_tokens_seen": 3469792,
41
  "step": 600
42
  },
43
  {
44
+ "epoch": 0.008264963716809284,
45
+ "grad_norm": 1.9247260093688965,
46
+ "learning_rate": 0.00015956154399988245,
47
+ "loss": 5.2863,
48
+ "num_input_tokens_seen": 4342304,
49
  "step": 750
50
  },
51
  {
52
+ "epoch": 0.00991795646017114,
53
+ "grad_norm": 1.9191653728485107,
54
+ "learning_rate": 0.00015947338260575695,
55
+ "loss": 5.2405,
56
+ "num_input_tokens_seen": 5211616,
57
  "step": 900
58
  },
59
  {
60
+ "epoch": 0.011570949203532997,
61
+ "grad_norm": 1.748758316040039,
62
+ "learning_rate": 0.00015938522121163143,
63
+ "loss": 5.1998,
64
+ "num_input_tokens_seen": 6098432,
65
  "step": 1050
66
  },
67
  {
68
+ "epoch": 0.013223941946894854,
69
+ "grad_norm": 2.057206869125366,
70
+ "learning_rate": 0.00015929705981750593,
71
+ "loss": 5.1706,
72
+ "num_input_tokens_seen": 6967904,
73
  "step": 1200
74
  },
75
  {
76
+ "epoch": 0.014876934690256709,
77
+ "grad_norm": 2.0983762741088867,
78
+ "learning_rate": 0.0001592088984233804,
79
+ "loss": 5.1521,
80
+ "num_input_tokens_seen": 7851072,
81
  "step": 1350
82
  },
83
  {
84
+ "epoch": 0.016529927433618568,
85
+ "grad_norm": 1.9689010381698608,
86
+ "learning_rate": 0.0001591207370292549,
87
+ "loss": 5.1314,
88
+ "num_input_tokens_seen": 8719744,
89
  "step": 1500
90
  },
91
  {
92
+ "epoch": 0.01818292017698042,
93
+ "grad_norm": 1.7962827682495117,
94
+ "learning_rate": 0.00015903257563512938,
95
+ "loss": 5.0878,
96
+ "num_input_tokens_seen": 9566912,
97
  "step": 1650
98
  },
99
  {
100
+ "epoch": 0.01983591292034228,
101
+ "grad_norm": 1.8615788221359253,
102
+ "learning_rate": 0.00015894441424100389,
103
+ "loss": 5.0739,
104
+ "num_input_tokens_seen": 10450048,
105
  "step": 1800
106
  },
107
  {
108
+ "epoch": 0.021488905663704136,
109
+ "grad_norm": 1.8449132442474365,
110
+ "learning_rate": 0.00015885625284687836,
111
+ "loss": 5.0658,
112
+ "num_input_tokens_seen": 11348640,
113
  "step": 1950
114
  },
115
  {
116
+ "epoch": 0.023141898407065993,
117
+ "grad_norm": 1.8516219854354858,
118
+ "learning_rate": 0.00015876809145275284,
119
+ "loss": 5.0462,
120
+ "num_input_tokens_seen": 12226144,
121
  "step": 2100
122
  },
123
  {
124
+ "epoch": 0.02479489115042785,
125
+ "grad_norm": 1.9487632513046265,
126
+ "learning_rate": 0.00015867993005862734,
127
+ "loss": 5.0265,
128
+ "num_input_tokens_seen": 13119040,
129
  "step": 2250
130
  },
131
  {
132
+ "epoch": 0.026447883893789707,
133
+ "grad_norm": 1.8343034982681274,
134
+ "learning_rate": 0.0001585917686645018,
135
+ "loss": 5.0056,
136
+ "num_input_tokens_seen": 13993280,
137
  "step": 2400
138
  },
139
  {
140
+ "epoch": 0.028100876637151564,
141
+ "grad_norm": 2.002856731414795,
142
+ "learning_rate": 0.00015850360727037632,
143
+ "loss": 4.9993,
144
+ "num_input_tokens_seen": 14876768,
145
  "step": 2550
146
  },
147
  {
148
+ "epoch": 0.029753869380513418,
149
+ "grad_norm": 1.981166124343872,
150
+ "learning_rate": 0.0001584154458762508,
151
+ "loss": 4.9943,
152
+ "num_input_tokens_seen": 15749088,
153
  "step": 2700
154
  },
155
  {
156
+ "epoch": 0.03140686212387528,
157
+ "grad_norm": 1.9732400178909302,
158
+ "learning_rate": 0.0001583272844821253,
159
+ "loss": 4.9762,
160
+ "num_input_tokens_seen": 16634176,
161
  "step": 2850
162
  },
163
  {
164
+ "epoch": 0.033059854867237136,
165
+ "grad_norm": 1.8369622230529785,
166
+ "learning_rate": 0.00015823912308799977,
167
+ "loss": 4.9525,
168
+ "num_input_tokens_seen": 17509824,
169
  "step": 3000
170
  },
171
  {
172
+ "epoch": 0.034712847610598986,
173
+ "grad_norm": 1.9628883600234985,
174
+ "learning_rate": 0.00015815096169387427,
175
+ "loss": 4.9481,
176
+ "num_input_tokens_seen": 18398112,
177
  "step": 3150
178
  },
179
  {
180
+ "epoch": 0.03636584035396084,
181
+ "grad_norm": 1.8523181676864624,
182
+ "learning_rate": 0.00015806280029974874,
183
+ "loss": 4.9306,
184
+ "num_input_tokens_seen": 19272928,
185
  "step": 3300
186
  },
187
  {
188
+ "epoch": 0.0380188330973227,
189
+ "grad_norm": 1.9627933502197266,
190
+ "learning_rate": 0.00015797463890562325,
191
+ "loss": 4.9217,
192
+ "num_input_tokens_seen": 20162880,
193
  "step": 3450
194
  },
195
  {
196
+ "epoch": 0.03967182584068456,
197
+ "grad_norm": 1.8966543674468994,
198
+ "learning_rate": 0.00015788647751149772,
199
+ "loss": 4.919,
200
+ "num_input_tokens_seen": 21041888,
201
  "step": 3600
202
  },
203
  {
204
+ "epoch": 0.041324818584046415,
205
+ "grad_norm": 1.9131779670715332,
206
+ "learning_rate": 0.00015779831611737222,
207
+ "loss": 4.9118,
208
+ "num_input_tokens_seen": 21914272,
209
  "step": 3750
210
  },
211
  {
212
+ "epoch": 0.04297781132740827,
213
+ "grad_norm": 1.8262194395065308,
214
+ "learning_rate": 0.0001577101547232467,
215
+ "loss": 4.9139,
216
+ "num_input_tokens_seen": 22802432,
217
  "step": 3900
218
  },
219
  {
220
+ "epoch": 0.04463080407077013,
221
+ "grad_norm": 1.9549835920333862,
222
+ "learning_rate": 0.0001576219933291212,
223
+ "loss": 4.8919,
224
+ "num_input_tokens_seen": 23680544,
225
  "step": 4050
226
  },
227
  {
228
+ "epoch": 0.046283796814131986,
229
+ "grad_norm": 1.9537177085876465,
230
+ "learning_rate": 0.00015753383193499568,
231
+ "loss": 4.8895,
232
+ "num_input_tokens_seen": 24572928,
233
  "step": 4200
234
  },
235
  {
236
+ "epoch": 0.04793678955749384,
237
+ "grad_norm": 1.9916348457336426,
238
+ "learning_rate": 0.00015744567054087018,
239
+ "loss": 4.8798,
240
+ "num_input_tokens_seen": 25458752,
241
  "step": 4350
242
  },
243
  {
244
+ "epoch": 0.0495897823008557,
245
+ "grad_norm": 1.9964395761489868,
246
+ "learning_rate": 0.00015735750914674465,
247
+ "loss": 4.8734,
248
+ "num_input_tokens_seen": 26339424,
249
  "step": 4500
250
  },
251
  {
252
+ "epoch": 0.05124277504421756,
253
+ "grad_norm": 1.9595707654953003,
254
+ "learning_rate": 0.00015726934775261916,
255
+ "loss": 4.8654,
256
+ "num_input_tokens_seen": 27208928,
257
  "step": 4650
258
  },
259
  {
260
+ "epoch": 0.052895767787579415,
261
+ "grad_norm": 2.002746820449829,
262
+ "learning_rate": 0.00015718118635849363,
263
+ "loss": 4.8536,
264
+ "num_input_tokens_seen": 28083488,
265
  "step": 4800
266
  },
267
  {
268
+ "epoch": 0.05454876053094127,
269
+ "grad_norm": 2.014301300048828,
270
+ "learning_rate": 0.00015709302496436813,
271
+ "loss": 4.8513,
272
+ "num_input_tokens_seen": 28948832,
273
  "step": 4950
274
  },
275
  {
276
+ "epoch": 0.05620175327430313,
277
+ "grad_norm": 1.82748544216156,
278
+ "learning_rate": 0.0001570048635702426,
279
+ "loss": 4.8477,
280
+ "num_input_tokens_seen": 29830752,
281
  "step": 5100
282
  },
283
  {
284
+ "epoch": 0.05785474601766498,
285
+ "grad_norm": 1.907245397567749,
286
+ "learning_rate": 0.0001569167021761171,
287
+ "loss": 4.8445,
288
+ "num_input_tokens_seen": 30709248,
289
  "step": 5250
290
  },
291
  {
292
+ "epoch": 0.059507738761026836,
293
+ "grad_norm": 1.9649808406829834,
294
+ "learning_rate": 0.00015682854078199158,
295
+ "loss": 4.8313,
296
+ "num_input_tokens_seen": 31597856,
297
  "step": 5400
298
  },
299
  {
300
+ "epoch": 0.06116073150438869,
301
+ "grad_norm": 1.9375178813934326,
302
+ "learning_rate": 0.00015674037938786606,
303
+ "loss": 4.8206,
304
+ "num_input_tokens_seen": 32485120,
305
  "step": 5550
306
  },
307
  {
308
+ "epoch": 0.06281372424775056,
309
+ "grad_norm": 1.8886380195617676,
310
+ "learning_rate": 0.00015665221799374056,
311
+ "loss": 4.8152,
312
+ "num_input_tokens_seen": 33354688,
313
  "step": 5700
314
  },
315
  {
316
+ "epoch": 0.06446671699111241,
317
+ "grad_norm": 1.8993780612945557,
318
+ "learning_rate": 0.00015656405659961504,
319
+ "loss": 4.8155,
320
+ "num_input_tokens_seen": 34230592,
321
  "step": 5850
322
  },
323
  {
324
+ "epoch": 0.06611970973447427,
325
+ "grad_norm": 1.8930308818817139,
326
+ "learning_rate": 0.00015647589520548954,
327
+ "loss": 4.8173,
328
+ "num_input_tokens_seen": 35090336,
329
  "step": 6000
330
  },
331
  {
332
+ "epoch": 0.06777270247783612,
333
+ "grad_norm": 1.951819658279419,
334
+ "learning_rate": 0.00015638773381136401,
335
+ "loss": 4.8118,
336
+ "num_input_tokens_seen": 35973024,
337
  "step": 6150
338
  },
339
  {
340
+ "epoch": 0.06942569522119797,
341
+ "grad_norm": 1.9142402410507202,
342
+ "learning_rate": 0.00015629957241723852,
343
+ "loss": 4.8079,
344
+ "num_input_tokens_seen": 36855936,
345
  "step": 6300
346
  },
347
  {
348
+ "epoch": 0.07107868796455984,
349
+ "grad_norm": 1.9393310546875,
350
+ "learning_rate": 0.000156211411023113,
351
+ "loss": 4.7883,
352
+ "num_input_tokens_seen": 37722848,
353
  "step": 6450
354
  },
355
  {
356
+ "epoch": 0.07273168070792169,
357
+ "grad_norm": 1.8511933088302612,
358
+ "learning_rate": 0.00015612383737161498,
359
+ "loss": 4.8043,
360
+ "num_input_tokens_seen": 38597600,
361
  "step": 6600
362
  },
363
  {
364
+ "epoch": 0.07438467345128355,
365
+ "grad_norm": 1.8763892650604248,
366
+ "learning_rate": 0.00015603567597748946,
367
+ "loss": 4.7932,
368
+ "num_input_tokens_seen": 39493152,
369
  "step": 6750
370
  },
371
  {
372
+ "epoch": 0.0760376661946454,
373
+ "grad_norm": 1.9806557893753052,
374
+ "learning_rate": 0.00015594751458336396,
375
+ "loss": 4.7813,
376
+ "num_input_tokens_seen": 40352640,
377
  "step": 6900
378
  },
379
  {
380
+ "epoch": 0.07769065893800726,
381
+ "grad_norm": 2.001722574234009,
382
+ "learning_rate": 0.00015585935318923843,
383
+ "loss": 4.7941,
384
+ "num_input_tokens_seen": 41236160,
385
  "step": 7050
386
  },
387
  {
388
+ "epoch": 0.07934365168136912,
389
+ "grad_norm": 2.1065292358398438,
390
+ "learning_rate": 0.00015577119179511294,
391
+ "loss": 4.7819,
392
+ "num_input_tokens_seen": 42111296,
393
  "step": 7200
394
  },
395
  {
396
+ "epoch": 0.08099664442473098,
397
+ "grad_norm": 1.8941328525543213,
398
+ "learning_rate": 0.0001556830304009874,
399
+ "loss": 4.7737,
400
+ "num_input_tokens_seen": 42992864,
401
  "step": 7350
402
  },
403
  {
404
+ "epoch": 0.08264963716809283,
405
+ "grad_norm": 1.8765467405319214,
406
+ "learning_rate": 0.00015559486900686191,
407
+ "loss": 4.764,
408
+ "num_input_tokens_seen": 43871808,
409
  "step": 7500
410
  },
411
  {
412
+ "epoch": 0.0843026299114547,
413
+ "grad_norm": 1.9826706647872925,
414
+ "learning_rate": 0.0001555067076127364,
415
+ "loss": 4.7805,
416
+ "num_input_tokens_seen": 44742496,
417
  "step": 7650
418
  },
419
  {
420
+ "epoch": 0.08595562265481654,
421
+ "grad_norm": 1.9296499490737915,
422
+ "learning_rate": 0.0001554185462186109,
423
+ "loss": 4.7585,
424
+ "num_input_tokens_seen": 45594112,
425
  "step": 7800
426
  },
427
  {
428
+ "epoch": 0.08760861539817841,
429
+ "grad_norm": 1.9379116296768188,
430
+ "learning_rate": 0.00015533038482448537,
431
+ "loss": 4.7682,
432
+ "num_input_tokens_seen": 46465952,
433
  "step": 7950
434
  },
435
  {
436
+ "epoch": 0.08926160814154026,
437
+ "grad_norm": 1.8769218921661377,
438
+ "learning_rate": 0.00015524222343035987,
439
+ "loss": 4.7574,
440
+ "num_input_tokens_seen": 47325664,
441
  "step": 8100
442
  },
443
  {
444
+ "epoch": 0.09091460088490212,
445
+ "grad_norm": 1.8942108154296875,
446
+ "learning_rate": 0.00015515406203623434,
447
+ "loss": 4.7493,
448
+ "num_input_tokens_seen": 48202944,
449
  "step": 8250
450
  },
451
  {
452
+ "epoch": 0.09256759362826397,
453
+ "grad_norm": 1.84010648727417,
454
+ "learning_rate": 0.00015506590064210885,
455
+ "loss": 4.7515,
456
+ "num_input_tokens_seen": 49074048,
457
  "step": 8400
458
  },
459
  {
460
+ "epoch": 0.09422058637162582,
461
+ "grad_norm": 1.8978796005249023,
462
+ "learning_rate": 0.00015497773924798332,
463
+ "loss": 4.7512,
464
+ "num_input_tokens_seen": 49959008,
465
  "step": 8550
466
  },
467
  {
468
+ "epoch": 0.09587357911498769,
469
+ "grad_norm": 1.9536223411560059,
470
+ "learning_rate": 0.0001548895778538578,
471
+ "loss": 4.758,
472
+ "num_input_tokens_seen": 50863648,
473
  "step": 8700
474
  },
475
  {
476
+ "epoch": 0.09752657185834954,
477
+ "grad_norm": 2.0626060962677,
478
+ "learning_rate": 0.0001548014164597323,
479
+ "loss": 4.7434,
480
+ "num_input_tokens_seen": 51730944,
481
  "step": 8850
482
  },
483
  {
484
+ "epoch": 0.0991795646017114,
485
+ "grad_norm": 1.9423109292984009,
486
+ "learning_rate": 0.00015471325506560677,
487
+ "loss": 4.736,
488
+ "num_input_tokens_seen": 52593024,
489
  "step": 9000
490
  },
491
  {
492
+ "epoch": 0.10083255734507325,
493
+ "grad_norm": 1.9180619716644287,
494
+ "learning_rate": 0.00015462509367148127,
495
+ "loss": 4.7187,
496
+ "num_input_tokens_seen": 53470528,
497
  "step": 9150
498
  },
499
  {
500
+ "epoch": 0.10248555008843512,
501
+ "grad_norm": 1.8776642084121704,
502
+ "learning_rate": 0.00015453693227735575,
503
+ "loss": 4.7382,
504
+ "num_input_tokens_seen": 54362720,
505
  "step": 9300
506
  },
507
  {
508
+ "epoch": 0.10413854283179697,
509
+ "grad_norm": 1.9289714097976685,
510
+ "learning_rate": 0.00015444877088323025,
511
+ "loss": 4.7181,
512
+ "num_input_tokens_seen": 55238304,
513
  "step": 9450
514
  },
515
  {
516
+ "epoch": 0.10579153557515883,
517
+ "grad_norm": 1.9489550590515137,
518
+ "learning_rate": 0.00015436060948910473,
519
+ "loss": 4.7287,
520
+ "num_input_tokens_seen": 56107040,
521
  "step": 9600
522
  },
523
  {
524
+ "epoch": 0.10744452831852068,
525
+ "grad_norm": 2.01839280128479,
526
+ "learning_rate": 0.00015427244809497923,
527
+ "loss": 4.7097,
528
+ "num_input_tokens_seen": 56995456,
529
  "step": 9750
530
  },
531
  {
532
+ "epoch": 0.10909752106188254,
533
+ "grad_norm": 1.9155646562576294,
534
+ "learning_rate": 0.0001541842867008537,
535
+ "loss": 4.7153,
536
+ "num_input_tokens_seen": 57852640,
537
  "step": 9900
538
  },
539
  {
540
+ "epoch": 0.1107505138052444,
541
+ "grad_norm": 2.008150100708008,
542
+ "learning_rate": 0.0001540961253067282,
543
+ "loss": 4.7139,
544
+ "num_input_tokens_seen": 58719840,
545
  "step": 10050
546
  },
547
  {
548
+ "epoch": 0.11240350654860626,
549
+ "grad_norm": 1.9440505504608154,
550
+ "learning_rate": 0.00015400796391260268,
551
+ "loss": 4.7184,
552
+ "num_input_tokens_seen": 59594784,
553
  "step": 10200
554
  },
555
  {
556
+ "epoch": 0.11405649929196811,
557
+ "grad_norm": 1.9298348426818848,
558
+ "learning_rate": 0.00015391980251847718,
559
+ "loss": 4.708,
560
+ "num_input_tokens_seen": 60451712,
561
  "step": 10350
562
  },
563
  {
564
+ "epoch": 0.11570949203532996,
565
+ "grad_norm": 1.9444379806518555,
566
+ "learning_rate": 0.00015383164112435166,
567
+ "loss": 4.6979,
568
+ "num_input_tokens_seen": 61335008,
569
  "step": 10500
570
  },
571
  {
572
+ "epoch": 0.11736248477869182,
573
+ "grad_norm": 2.0216357707977295,
574
+ "learning_rate": 0.00015374406747285365,
575
+ "loss": 4.7055,
576
+ "num_input_tokens_seen": 62197344,
577
  "step": 10650
578
  },
579
  {
580
+ "epoch": 0.11901547752205367,
581
+ "grad_norm": 1.9788328409194946,
582
+ "learning_rate": 0.00015365590607872815,
583
+ "loss": 4.6948,
584
+ "num_input_tokens_seen": 63058464,
585
  "step": 10800
586
  },
587
  {
588
+ "epoch": 0.12066847026541554,
589
+ "grad_norm": 2.0648193359375,
590
+ "learning_rate": 0.00015356774468460263,
591
+ "loss": 4.7058,
592
+ "num_input_tokens_seen": 63942112,
593
  "step": 10950
594
  },
595
  {
596
+ "epoch": 0.12232146300877739,
597
+ "grad_norm": 1.9497121572494507,
598
+ "learning_rate": 0.00015348017103310462,
599
+ "loss": 4.6924,
600
+ "num_input_tokens_seen": 64815232,
601
  "step": 11100
602
  },
603
  {
604
+ "epoch": 0.12397445575213925,
605
+ "grad_norm": 1.9825148582458496,
606
+ "learning_rate": 0.0001533920096389791,
607
+ "loss": 4.7021,
608
+ "num_input_tokens_seen": 65689856,
609
  "step": 11250
610
  },
611
  {
612
+ "epoch": 0.12562744849550112,
613
+ "grad_norm": 1.9766299724578857,
614
+ "learning_rate": 0.0001533038482448536,
615
+ "loss": 4.6921,
616
+ "num_input_tokens_seen": 66570272,
617
  "step": 11400
618
  },
619
  {
620
+ "epoch": 0.12728044123886295,
621
+ "grad_norm": 1.9706653356552124,
622
+ "learning_rate": 0.00015321568685072807,
623
+ "loss": 4.6847,
624
+ "num_input_tokens_seen": 67436096,
625
  "step": 11550
626
  },
627
  {
628
+ "epoch": 0.12893343398222482,
629
+ "grad_norm": 1.9741766452789307,
630
+ "learning_rate": 0.00015312811319923006,
631
+ "loss": 4.6835,
632
+ "num_input_tokens_seen": 68326816,
633
  "step": 11700
634
  },
635
  {
636
+ "epoch": 0.13058642672558668,
637
+ "grad_norm": 1.850825548171997,
638
+ "learning_rate": 0.00015303995180510456,
639
+ "loss": 4.6874,
640
+ "num_input_tokens_seen": 69203328,
641
  "step": 11850
642
  },
643
  {
644
+ "epoch": 0.13223941946894854,
645
+ "grad_norm": 2.0040206909179688,
646
+ "learning_rate": 0.00015295179041097904,
647
+ "loss": 4.6805,
648
+ "num_input_tokens_seen": 70049696,
649
  "step": 12000
650
+ },
651
+ {
652
+ "epoch": 0.13389241221231038,
653
+ "grad_norm": 1.9326891899108887,
654
+ "learning_rate": 0.00015286362901685354,
655
+ "loss": 4.6781,
656
+ "num_input_tokens_seen": 70929856,
657
+ "step": 12150
658
+ },
659
+ {
660
+ "epoch": 0.13554540495567224,
661
+ "grad_norm": 1.8233270645141602,
662
+ "learning_rate": 0.00015277546762272802,
663
+ "loss": 4.6732,
664
+ "num_input_tokens_seen": 71797184,
665
+ "step": 12300
666
+ },
667
+ {
668
+ "epoch": 0.1371983976990341,
669
+ "grad_norm": 2.026263475418091,
670
+ "learning_rate": 0.00015268730622860252,
671
+ "loss": 4.672,
672
+ "num_input_tokens_seen": 72662528,
673
+ "step": 12450
674
+ },
675
+ {
676
+ "epoch": 0.13885139044239594,
677
+ "grad_norm": 1.8338570594787598,
678
+ "learning_rate": 0.000152599144834477,
679
+ "loss": 4.6718,
680
+ "num_input_tokens_seen": 73543840,
681
+ "step": 12600
682
+ },
683
+ {
684
+ "epoch": 0.1405043831857578,
685
+ "grad_norm": 1.934313416481018,
686
+ "learning_rate": 0.0001525109834403515,
687
+ "loss": 4.6595,
688
+ "num_input_tokens_seen": 74426752,
689
+ "step": 12750
690
+ },
691
+ {
692
+ "epoch": 0.14215737592911967,
693
+ "grad_norm": 1.861647367477417,
694
+ "learning_rate": 0.00015242282204622597,
695
+ "loss": 4.6692,
696
+ "num_input_tokens_seen": 75309888,
697
+ "step": 12900
698
+ },
699
+ {
700
+ "epoch": 0.14381036867248154,
701
+ "grad_norm": 1.9282541275024414,
702
+ "learning_rate": 0.00015233466065210047,
703
+ "loss": 4.6604,
704
+ "num_input_tokens_seen": 76194976,
705
+ "step": 13050
706
+ },
707
+ {
708
+ "epoch": 0.14546336141584337,
709
+ "grad_norm": 1.975542664527893,
710
+ "learning_rate": 0.00015224649925797495,
711
+ "loss": 4.662,
712
+ "num_input_tokens_seen": 77078048,
713
+ "step": 13200
714
+ },
715
+ {
716
+ "epoch": 0.14711635415920524,
717
+ "grad_norm": 1.8979029655456543,
718
+ "learning_rate": 0.00015215833786384945,
719
+ "loss": 4.6554,
720
+ "num_input_tokens_seen": 77940000,
721
+ "step": 13350
722
+ },
723
+ {
724
+ "epoch": 0.1487693469025671,
725
+ "grad_norm": 1.875108242034912,
726
+ "learning_rate": 0.00015207017646972392,
727
+ "loss": 4.6599,
728
+ "num_input_tokens_seen": 78822336,
729
+ "step": 13500
730
+ },
731
+ {
732
+ "epoch": 0.15042233964592897,
733
+ "grad_norm": 1.9476161003112793,
734
+ "learning_rate": 0.0001519820150755984,
735
+ "loss": 4.6688,
736
+ "num_input_tokens_seen": 79708256,
737
+ "step": 13650
738
+ },
739
+ {
740
+ "epoch": 0.1520753323892908,
741
+ "grad_norm": 1.9902242422103882,
742
+ "learning_rate": 0.0001518938536814729,
743
+ "loss": 4.6626,
744
+ "num_input_tokens_seen": 80595168,
745
+ "step": 13800
746
+ },
747
+ {
748
+ "epoch": 0.15372832513265267,
749
+ "grad_norm": 1.7958108186721802,
750
+ "learning_rate": 0.00015180569228734738,
751
+ "loss": 4.662,
752
+ "num_input_tokens_seen": 81484416,
753
+ "step": 13950
754
+ },
755
+ {
756
+ "epoch": 0.15538131787601453,
757
+ "grad_norm": 1.8727210760116577,
758
+ "learning_rate": 0.00015171753089322188,
759
+ "loss": 4.6579,
760
+ "num_input_tokens_seen": 82355712,
761
+ "step": 14100
762
+ },
763
+ {
764
+ "epoch": 0.1570343106193764,
765
+ "grad_norm": 2.0186071395874023,
766
+ "learning_rate": 0.00015162936949909635,
767
+ "loss": 4.6494,
768
+ "num_input_tokens_seen": 83223136,
769
+ "step": 14250
770
+ },
771
+ {
772
+ "epoch": 0.15868730336273823,
773
+ "grad_norm": 1.8837051391601562,
774
+ "learning_rate": 0.00015154120810497086,
775
+ "loss": 4.6463,
776
+ "num_input_tokens_seen": 84097152,
777
+ "step": 14400
778
+ },
779
+ {
780
+ "epoch": 0.1603402961061001,
781
+ "grad_norm": 1.8699517250061035,
782
+ "learning_rate": 0.00015145304671084533,
783
+ "loss": 4.6489,
784
+ "num_input_tokens_seen": 84976576,
785
+ "step": 14550
786
+ },
787
+ {
788
+ "epoch": 0.16199328884946196,
789
+ "grad_norm": 1.956932783126831,
790
+ "learning_rate": 0.00015136488531671983,
791
+ "loss": 4.649,
792
+ "num_input_tokens_seen": 85868224,
793
+ "step": 14700
794
+ },
795
+ {
796
+ "epoch": 0.1636462815928238,
797
+ "grad_norm": 2.020624876022339,
798
+ "learning_rate": 0.0001512767239225943,
799
+ "loss": 4.6458,
800
+ "num_input_tokens_seen": 86746944,
801
+ "step": 14850
802
+ },
803
+ {
804
+ "epoch": 0.16529927433618566,
805
+ "grad_norm": 1.9445135593414307,
806
+ "learning_rate": 0.0001511885625284688,
807
+ "loss": 4.6387,
808
+ "num_input_tokens_seen": 87626976,
809
+ "step": 15000
810
+ },
811
+ {
812
+ "epoch": 0.16695226707954752,
813
+ "grad_norm": 1.9843000173568726,
814
+ "learning_rate": 0.00015110040113434328,
815
+ "loss": 4.6481,
816
+ "num_input_tokens_seen": 88483616,
817
+ "step": 15150
818
+ },
819
+ {
820
+ "epoch": 0.1686052598229094,
821
+ "grad_norm": 2.0259897708892822,
822
+ "learning_rate": 0.00015101282748284528,
823
+ "loss": 4.6317,
824
+ "num_input_tokens_seen": 89374048,
825
+ "step": 15300
826
+ },
827
+ {
828
+ "epoch": 0.17025825256627122,
829
+ "grad_norm": 1.8472915887832642,
830
+ "learning_rate": 0.00015092466608871975,
831
+ "loss": 4.6468,
832
+ "num_input_tokens_seen": 90265376,
833
+ "step": 15450
834
+ },
835
+ {
836
+ "epoch": 0.1719112453096331,
837
+ "grad_norm": 1.9485039710998535,
838
+ "learning_rate": 0.00015083650469459425,
839
+ "loss": 4.6272,
840
+ "num_input_tokens_seen": 91161504,
841
+ "step": 15600
842
+ },
843
+ {
844
+ "epoch": 0.17356423805299495,
845
+ "grad_norm": 2.0340664386749268,
846
+ "learning_rate": 0.00015074834330046873,
847
+ "loss": 4.6252,
848
+ "num_input_tokens_seen": 92033536,
849
+ "step": 15750
850
+ },
851
+ {
852
+ "epoch": 0.17521723079635682,
853
+ "grad_norm": 1.8034217357635498,
854
+ "learning_rate": 0.00015066018190634323,
855
+ "loss": 4.6347,
856
+ "num_input_tokens_seen": 92906464,
857
+ "step": 15900
858
+ },
859
+ {
860
+ "epoch": 0.17687022353971865,
861
+ "grad_norm": 1.9323750734329224,
862
+ "learning_rate": 0.0001505720205122177,
863
+ "loss": 4.624,
864
+ "num_input_tokens_seen": 93773664,
865
+ "step": 16050
866
+ },
867
+ {
868
+ "epoch": 0.17852321628308052,
869
+ "grad_norm": 1.9791151285171509,
870
+ "learning_rate": 0.00015048385911809218,
871
+ "loss": 4.6184,
872
+ "num_input_tokens_seen": 94646528,
873
+ "step": 16200
874
+ },
875
+ {
876
+ "epoch": 0.18017620902644238,
877
+ "grad_norm": 2.0325284004211426,
878
+ "learning_rate": 0.00015039569772396668,
879
+ "loss": 4.6129,
880
+ "num_input_tokens_seen": 95522304,
881
+ "step": 16350
882
+ },
883
+ {
884
+ "epoch": 0.18182920176980424,
885
+ "grad_norm": 1.773972511291504,
886
+ "learning_rate": 0.00015030753632984116,
887
+ "loss": 4.6283,
888
+ "num_input_tokens_seen": 96397632,
889
+ "step": 16500
890
+ },
891
+ {
892
+ "epoch": 0.18348219451316608,
893
+ "grad_norm": 1.792601466178894,
894
+ "learning_rate": 0.00015021937493571566,
895
+ "loss": 4.6201,
896
+ "num_input_tokens_seen": 97283072,
897
+ "step": 16650
898
+ },
899
+ {
900
+ "epoch": 0.18513518725652794,
901
+ "grad_norm": 1.9488441944122314,
902
+ "learning_rate": 0.00015013121354159013,
903
+ "loss": 4.6174,
904
+ "num_input_tokens_seen": 98164960,
905
+ "step": 16800
906
+ },
907
+ {
908
+ "epoch": 0.1867881799998898,
909
+ "grad_norm": 1.8708151578903198,
910
+ "learning_rate": 0.00015004305214746464,
911
+ "loss": 4.6156,
912
+ "num_input_tokens_seen": 99030464,
913
+ "step": 16950
914
+ },
915
+ {
916
+ "epoch": 0.18844117274325164,
917
+ "grad_norm": 1.9848783016204834,
918
+ "learning_rate": 0.0001499548907533391,
919
+ "loss": 4.6069,
920
+ "num_input_tokens_seen": 99913184,
921
+ "step": 17100
922
+ },
923
+ {
924
+ "epoch": 0.1900941654866135,
925
+ "grad_norm": 1.9591269493103027,
926
+ "learning_rate": 0.00014986672935921361,
927
+ "loss": 4.6195,
928
+ "num_input_tokens_seen": 100797056,
929
+ "step": 17250
930
+ },
931
+ {
932
+ "epoch": 0.19174715822997537,
933
+ "grad_norm": 1.9400300979614258,
934
+ "learning_rate": 0.0001497785679650881,
935
+ "loss": 4.6167,
936
+ "num_input_tokens_seen": 101678336,
937
+ "step": 17400
938
+ },
939
+ {
940
+ "epoch": 0.19340015097333724,
941
+ "grad_norm": 1.9163286685943604,
942
+ "learning_rate": 0.0001496904065709626,
943
+ "loss": 4.6135,
944
+ "num_input_tokens_seen": 102532640,
945
+ "step": 17550
946
+ },
947
+ {
948
+ "epoch": 0.19505314371669907,
949
+ "grad_norm": 1.86648690700531,
950
+ "learning_rate": 0.00014960224517683707,
951
+ "loss": 4.6063,
952
+ "num_input_tokens_seen": 103403264,
953
+ "step": 17700
954
+ },
955
+ {
956
+ "epoch": 0.19670613646006094,
957
+ "grad_norm": 1.9310001134872437,
958
+ "learning_rate": 0.00014951408378271157,
959
+ "loss": 4.6143,
960
+ "num_input_tokens_seen": 104304224,
961
+ "step": 17850
962
+ },
963
+ {
964
+ "epoch": 0.1983591292034228,
965
+ "grad_norm": 1.9832515716552734,
966
+ "learning_rate": 0.00014942592238858604,
967
+ "loss": 4.6138,
968
+ "num_input_tokens_seen": 105184128,
969
+ "step": 18000
970
+ },
971
+ {
972
+ "epoch": 0.20001212194678467,
973
+ "grad_norm": 1.9453548192977905,
974
+ "learning_rate": 0.00014933776099446055,
975
+ "loss": 4.6143,
976
+ "num_input_tokens_seen": 106070880,
977
+ "step": 18150
978
+ },
979
+ {
980
+ "epoch": 0.2016651146901465,
981
+ "grad_norm": 1.8135297298431396,
982
+ "learning_rate": 0.00014924959960033502,
983
+ "loss": 4.6197,
984
+ "num_input_tokens_seen": 106944960,
985
+ "step": 18300
986
+ },
987
+ {
988
+ "epoch": 0.20331810743350837,
989
+ "grad_norm": 1.892717719078064,
990
+ "learning_rate": 0.00014916143820620952,
991
+ "loss": 4.6039,
992
+ "num_input_tokens_seen": 107808512,
993
+ "step": 18450
994
+ },
995
+ {
996
+ "epoch": 0.20497110017687023,
997
+ "grad_norm": 1.9304077625274658,
998
+ "learning_rate": 0.000149073276812084,
999
+ "loss": 4.6103,
1000
+ "num_input_tokens_seen": 108679584,
1001
+ "step": 18600
1002
+ },
1003
+ {
1004
+ "epoch": 0.20662409292023207,
1005
+ "grad_norm": 1.874104380607605,
1006
+ "learning_rate": 0.0001489851154179585,
1007
+ "loss": 4.6092,
1008
+ "num_input_tokens_seen": 109561440,
1009
+ "step": 18750
1010
+ },
1011
+ {
1012
+ "epoch": 0.20827708566359393,
1013
+ "grad_norm": 1.9672309160232544,
1014
+ "learning_rate": 0.00014889695402383297,
1015
+ "loss": 4.6027,
1016
+ "num_input_tokens_seen": 110445760,
1017
+ "step": 18900
1018
+ },
1019
+ {
1020
+ "epoch": 0.2099300784069558,
1021
+ "grad_norm": 1.9013960361480713,
1022
+ "learning_rate": 0.00014880879262970748,
1023
+ "loss": 4.6161,
1024
+ "num_input_tokens_seen": 111310304,
1025
+ "step": 19050
1026
+ },
1027
+ {
1028
+ "epoch": 0.21158307115031766,
1029
+ "grad_norm": 1.902948021888733,
1030
+ "learning_rate": 0.00014872063123558195,
1031
+ "loss": 4.6118,
1032
+ "num_input_tokens_seen": 112181440,
1033
+ "step": 19200
1034
+ },
1035
+ {
1036
+ "epoch": 0.2132360638936795,
1037
+ "grad_norm": 1.9160059690475464,
1038
+ "learning_rate": 0.00014863246984145645,
1039
+ "loss": 4.5929,
1040
+ "num_input_tokens_seen": 113074496,
1041
+ "step": 19350
1042
+ },
1043
+ {
1044
+ "epoch": 0.21488905663704136,
1045
+ "grad_norm": 1.843983769416809,
1046
+ "learning_rate": 0.00014854489618995845,
1047
+ "loss": 4.5982,
1048
+ "num_input_tokens_seen": 113972512,
1049
+ "step": 19500
1050
+ },
1051
+ {
1052
+ "epoch": 0.21654204938040322,
1053
+ "grad_norm": 1.83791184425354,
1054
+ "learning_rate": 0.00014845673479583292,
1055
+ "loss": 4.6046,
1056
+ "num_input_tokens_seen": 114839680,
1057
+ "step": 19650
1058
+ },
1059
+ {
1060
+ "epoch": 0.2181950421237651,
1061
+ "grad_norm": 1.8458436727523804,
1062
+ "learning_rate": 0.00014836857340170742,
1063
+ "loss": 4.5987,
1064
+ "num_input_tokens_seen": 115716000,
1065
+ "step": 19800
1066
+ },
1067
+ {
1068
+ "epoch": 0.21984803486712692,
1069
+ "grad_norm": 2.0030035972595215,
1070
+ "learning_rate": 0.0001482804120075819,
1071
+ "loss": 4.5873,
1072
+ "num_input_tokens_seen": 116581408,
1073
+ "step": 19950
1074
+ },
1075
+ {
1076
+ "epoch": 0.2215010276104888,
1077
+ "grad_norm": 1.8120313882827759,
1078
+ "learning_rate": 0.00014819225061345637,
1079
+ "loss": 4.5893,
1080
+ "num_input_tokens_seen": 117446368,
1081
+ "step": 20100
1082
+ },
1083
+ {
1084
+ "epoch": 0.22315402035385065,
1085
+ "grad_norm": 1.8799773454666138,
1086
+ "learning_rate": 0.00014810408921933087,
1087
+ "loss": 4.5746,
1088
+ "num_input_tokens_seen": 118320096,
1089
+ "step": 20250
1090
+ },
1091
+ {
1092
+ "epoch": 0.22480701309721252,
1093
+ "grad_norm": 1.9042309522628784,
1094
+ "learning_rate": 0.00014801592782520535,
1095
+ "loss": 4.5851,
1096
+ "num_input_tokens_seen": 119192128,
1097
+ "step": 20400
1098
+ },
1099
+ {
1100
+ "epoch": 0.22646000584057435,
1101
+ "grad_norm": 1.8850473165512085,
1102
+ "learning_rate": 0.00014792776643107985,
1103
+ "loss": 4.5883,
1104
+ "num_input_tokens_seen": 120065888,
1105
+ "step": 20550
1106
+ },
1107
+ {
1108
+ "epoch": 0.22811299858393622,
1109
+ "grad_norm": 1.8963854312896729,
1110
+ "learning_rate": 0.00014783960503695433,
1111
+ "loss": 4.5869,
1112
+ "num_input_tokens_seen": 120921504,
1113
+ "step": 20700
1114
+ },
1115
+ {
1116
+ "epoch": 0.22976599132729808,
1117
+ "grad_norm": 1.8145036697387695,
1118
+ "learning_rate": 0.00014775144364282883,
1119
+ "loss": 4.5857,
1120
+ "num_input_tokens_seen": 121819936,
1121
+ "step": 20850
1122
+ },
1123
+ {
1124
+ "epoch": 0.23141898407065992,
1125
+ "grad_norm": 1.8780988454818726,
1126
+ "learning_rate": 0.0001476632822487033,
1127
+ "loss": 4.5793,
1128
+ "num_input_tokens_seen": 122700576,
1129
+ "step": 21000
1130
+ },
1131
+ {
1132
+ "epoch": 0.23307197681402178,
1133
+ "grad_norm": 1.8859424591064453,
1134
+ "learning_rate": 0.00014757512085457778,
1135
+ "loss": 4.5847,
1136
+ "num_input_tokens_seen": 123578848,
1137
+ "step": 21150
1138
+ },
1139
+ {
1140
+ "epoch": 0.23472496955738364,
1141
+ "grad_norm": 1.8556190729141235,
1142
+ "learning_rate": 0.00014748695946045228,
1143
+ "loss": 4.5915,
1144
+ "num_input_tokens_seen": 124451552,
1145
+ "step": 21300
1146
+ },
1147
+ {
1148
+ "epoch": 0.2363779623007455,
1149
+ "grad_norm": 1.8445396423339844,
1150
+ "learning_rate": 0.00014739879806632676,
1151
+ "loss": 4.5857,
1152
+ "num_input_tokens_seen": 125326208,
1153
+ "step": 21450
1154
+ },
1155
+ {
1156
+ "epoch": 0.23803095504410735,
1157
+ "grad_norm": 1.903262972831726,
1158
+ "learning_rate": 0.00014731063667220126,
1159
+ "loss": 4.5811,
1160
+ "num_input_tokens_seen": 126206560,
1161
+ "step": 21600
1162
+ },
1163
+ {
1164
+ "epoch": 0.2396839477874692,
1165
+ "grad_norm": 1.7595880031585693,
1166
+ "learning_rate": 0.00014722247527807573,
1167
+ "loss": 4.567,
1168
+ "num_input_tokens_seen": 127077440,
1169
+ "step": 21750
1170
+ },
1171
+ {
1172
+ "epoch": 0.24133694053083107,
1173
+ "grad_norm": 1.8828771114349365,
1174
+ "learning_rate": 0.00014713431388395024,
1175
+ "loss": 4.5792,
1176
+ "num_input_tokens_seen": 127955584,
1177
+ "step": 21900
1178
+ },
1179
+ {
1180
+ "epoch": 0.24298993327419294,
1181
+ "grad_norm": 1.8850219249725342,
1182
+ "learning_rate": 0.0001470461524898247,
1183
+ "loss": 4.5749,
1184
+ "num_input_tokens_seen": 128807200,
1185
+ "step": 22050
1186
+ },
1187
+ {
1188
+ "epoch": 0.24464292601755477,
1189
+ "grad_norm": 1.9162580966949463,
1190
+ "learning_rate": 0.00014695799109569919,
1191
+ "loss": 4.5725,
1192
+ "num_input_tokens_seen": 129672192,
1193
+ "step": 22200
1194
+ },
1195
+ {
1196
+ "epoch": 0.24629591876091664,
1197
+ "grad_norm": 1.866351842880249,
1198
+ "learning_rate": 0.0001468698297015737,
1199
+ "loss": 4.569,
1200
+ "num_input_tokens_seen": 130541696,
1201
+ "step": 22350
1202
+ },
1203
+ {
1204
+ "epoch": 0.2479489115042785,
1205
+ "grad_norm": 1.849186658859253,
1206
+ "learning_rate": 0.00014678166830744816,
1207
+ "loss": 4.5832,
1208
+ "num_input_tokens_seen": 131420992,
1209
+ "step": 22500
1210
+ },
1211
+ {
1212
+ "epoch": 0.24960190424764037,
1213
+ "grad_norm": 1.8402087688446045,
1214
+ "learning_rate": 0.00014669350691332266,
1215
+ "loss": 4.5664,
1216
+ "num_input_tokens_seen": 132312352,
1217
+ "step": 22650
1218
+ },
1219
+ {
1220
+ "epoch": 0.25125489699100223,
1221
+ "grad_norm": 1.8887277841567993,
1222
+ "learning_rate": 0.00014660534551919714,
1223
+ "loss": 4.5771,
1224
+ "num_input_tokens_seen": 133195968,
1225
+ "step": 22800
1226
+ },
1227
+ {
1228
+ "epoch": 0.25290788973436407,
1229
+ "grad_norm": 2.029491424560547,
1230
+ "learning_rate": 0.00014651718412507164,
1231
+ "loss": 4.5718,
1232
+ "num_input_tokens_seen": 134079040,
1233
+ "step": 22950
1234
+ },
1235
+ {
1236
+ "epoch": 0.2545608824777259,
1237
+ "grad_norm": 1.9162187576293945,
1238
+ "learning_rate": 0.00014642902273094612,
1239
+ "loss": 4.5746,
1240
+ "num_input_tokens_seen": 134957440,
1241
+ "step": 23100
1242
+ },
1243
+ {
1244
+ "epoch": 0.2562138752210878,
1245
+ "grad_norm": 1.8331772089004517,
1246
+ "learning_rate": 0.00014634086133682062,
1247
+ "loss": 4.5736,
1248
+ "num_input_tokens_seen": 135835328,
1249
+ "step": 23250
1250
+ },
1251
+ {
1252
+ "epoch": 0.25786686796444963,
1253
+ "grad_norm": 1.9076178073883057,
1254
+ "learning_rate": 0.0001462526999426951,
1255
+ "loss": 4.5647,
1256
+ "num_input_tokens_seen": 136701184,
1257
+ "step": 23400
1258
+ },
1259
+ {
1260
+ "epoch": 0.25951986070781147,
1261
+ "grad_norm": 1.9407544136047363,
1262
+ "learning_rate": 0.0001461645385485696,
1263
+ "loss": 4.5607,
1264
+ "num_input_tokens_seen": 137547264,
1265
+ "step": 23550
1266
+ },
1267
+ {
1268
+ "epoch": 0.26117285345117336,
1269
+ "grad_norm": 1.9287118911743164,
1270
+ "learning_rate": 0.0001460769648970716,
1271
+ "loss": 4.5642,
1272
+ "num_input_tokens_seen": 138417888,
1273
+ "step": 23700
1274
+ },
1275
+ {
1276
+ "epoch": 0.2628258461945352,
1277
+ "grad_norm": 2.0106074810028076,
1278
+ "learning_rate": 0.00014598939124557358,
1279
+ "loss": 4.5689,
1280
+ "num_input_tokens_seen": 139297088,
1281
+ "step": 23850
1282
+ },
1283
+ {
1284
+ "epoch": 0.2644788389378971,
1285
+ "grad_norm": 1.8796470165252686,
1286
+ "learning_rate": 0.00014590122985144808,
1287
+ "loss": 4.5736,
1288
+ "num_input_tokens_seen": 140165984,
1289
+ "step": 24000
1290
+ },
1291
+ {
1292
+ "epoch": 0.2661318316812589,
1293
+ "grad_norm": 1.8495882749557495,
1294
+ "learning_rate": 0.00014581306845732256,
1295
+ "loss": 4.5613,
1296
+ "num_input_tokens_seen": 141043680,
1297
+ "step": 24150
1298
+ },
1299
+ {
1300
+ "epoch": 0.26778482442462076,
1301
+ "grad_norm": 1.8603812456130981,
1302
+ "learning_rate": 0.00014572490706319706,
1303
+ "loss": 4.5707,
1304
+ "num_input_tokens_seen": 141913088,
1305
+ "step": 24300
1306
+ },
1307
+ {
1308
+ "epoch": 0.26943781716798265,
1309
+ "grad_norm": 1.8450992107391357,
1310
+ "learning_rate": 0.00014563674566907153,
1311
+ "loss": 4.5685,
1312
+ "num_input_tokens_seen": 142780992,
1313
+ "step": 24450
1314
+ },
1315
+ {
1316
+ "epoch": 0.2710908099113445,
1317
+ "grad_norm": 1.9459301233291626,
1318
+ "learning_rate": 0.000145548584274946,
1319
+ "loss": 4.5624,
1320
+ "num_input_tokens_seen": 143655008,
1321
+ "step": 24600
1322
+ },
1323
+ {
1324
+ "epoch": 0.2727438026547063,
1325
+ "grad_norm": 1.87797212600708,
1326
+ "learning_rate": 0.0001454604228808205,
1327
+ "loss": 4.5753,
1328
+ "num_input_tokens_seen": 144533760,
1329
+ "step": 24750
1330
+ },
1331
+ {
1332
+ "epoch": 0.2743967953980682,
1333
+ "grad_norm": 1.9368420839309692,
1334
+ "learning_rate": 0.00014537226148669498,
1335
+ "loss": 4.5473,
1336
+ "num_input_tokens_seen": 145405376,
1337
+ "step": 24900
1338
+ },
1339
+ {
1340
+ "epoch": 0.27604978814143005,
1341
+ "grad_norm": 1.9425833225250244,
1342
+ "learning_rate": 0.0001452841000925695,
1343
+ "loss": 4.5621,
1344
+ "num_input_tokens_seen": 146278176,
1345
+ "step": 25050
1346
+ },
1347
+ {
1348
+ "epoch": 0.2777027808847919,
1349
+ "grad_norm": 1.8315942287445068,
1350
+ "learning_rate": 0.00014519593869844396,
1351
+ "loss": 4.5441,
1352
+ "num_input_tokens_seen": 147150880,
1353
+ "step": 25200
1354
+ },
1355
+ {
1356
+ "epoch": 0.2793557736281538,
1357
+ "grad_norm": 1.865020990371704,
1358
+ "learning_rate": 0.00014510777730431846,
1359
+ "loss": 4.5681,
1360
+ "num_input_tokens_seen": 148039488,
1361
+ "step": 25350
1362
+ },
1363
+ {
1364
+ "epoch": 0.2810087663715156,
1365
+ "grad_norm": 1.9058725833892822,
1366
+ "learning_rate": 0.00014501961591019294,
1367
+ "loss": 4.55,
1368
+ "num_input_tokens_seen": 148898720,
1369
+ "step": 25500
1370
+ },
1371
+ {
1372
+ "epoch": 0.2826617591148775,
1373
+ "grad_norm": 1.9358283281326294,
1374
+ "learning_rate": 0.00014493145451606741,
1375
+ "loss": 4.5608,
1376
+ "num_input_tokens_seen": 149776576,
1377
+ "step": 25650
1378
+ },
1379
+ {
1380
+ "epoch": 0.28431475185823935,
1381
+ "grad_norm": 1.871090292930603,
1382
+ "learning_rate": 0.00014484329312194192,
1383
+ "loss": 4.539,
1384
+ "num_input_tokens_seen": 150667296,
1385
+ "step": 25800
1386
+ },
1387
+ {
1388
+ "epoch": 0.2859677446016012,
1389
+ "grad_norm": 1.885185956954956,
1390
+ "learning_rate": 0.0001447551317278164,
1391
+ "loss": 4.5421,
1392
+ "num_input_tokens_seen": 151536512,
1393
+ "step": 25950
1394
  }
1395
  ],
1396
  "logging_steps": 150,
1397
+ "max_steps": 272232,
1398
+ "num_input_tokens_seen": 151830560,
1399
  "num_train_epochs": 3,
1400
  "save_steps": 500,
1401
  "stateful_callbacks": {
 
1410
  "attributes": {}
1411
  }
1412
  },
1413
+ "total_flos": 2344361017958400.0,
1414
  "train_batch_size": 32,
1415
  "trial_name": null,
1416
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5e1abd80795eaf5b1bdba294c57094293cff0723779ae399c08b79b9b31e5ed2
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8a97e7cc354f265709c30fdb8ff115daee4c5dd804d2b9b98d7f91932d23661
3
  size 5112