gemma2b-summarize-gpt4o-128k / trainer_state.json
chansung's picture
Model save
19f9b6c verified
raw
history blame
No virus
149 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 15.0,
"eval_steps": 500,
"global_step": 4395,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0034129692832764505,
"grad_norm": 1.8359375,
"learning_rate": 4.545454545454545e-07,
"loss": 3.0499,
"step": 1
},
{
"epoch": 0.017064846416382253,
"grad_norm": 2.234375,
"learning_rate": 2.2727272727272728e-06,
"loss": 3.0434,
"step": 5
},
{
"epoch": 0.034129692832764506,
"grad_norm": 2.078125,
"learning_rate": 4.5454545454545455e-06,
"loss": 3.0699,
"step": 10
},
{
"epoch": 0.051194539249146756,
"grad_norm": 1.8515625,
"learning_rate": 6.818181818181818e-06,
"loss": 3.0656,
"step": 15
},
{
"epoch": 0.06825938566552901,
"grad_norm": 3.0,
"learning_rate": 9.090909090909091e-06,
"loss": 3.0526,
"step": 20
},
{
"epoch": 0.08532423208191127,
"grad_norm": 2.015625,
"learning_rate": 1.1363636363636365e-05,
"loss": 3.0382,
"step": 25
},
{
"epoch": 0.10238907849829351,
"grad_norm": 17.125,
"learning_rate": 1.3636363636363637e-05,
"loss": 2.982,
"step": 30
},
{
"epoch": 0.11945392491467577,
"grad_norm": 2.03125,
"learning_rate": 1.590909090909091e-05,
"loss": 2.9332,
"step": 35
},
{
"epoch": 0.13651877133105803,
"grad_norm": 3.140625,
"learning_rate": 1.8181818181818182e-05,
"loss": 2.8934,
"step": 40
},
{
"epoch": 0.15358361774744028,
"grad_norm": 3.25,
"learning_rate": 2.0454545454545457e-05,
"loss": 2.7804,
"step": 45
},
{
"epoch": 0.17064846416382254,
"grad_norm": 1.3984375,
"learning_rate": 2.272727272727273e-05,
"loss": 2.7194,
"step": 50
},
{
"epoch": 0.18771331058020477,
"grad_norm": 1.296875,
"learning_rate": 2.5e-05,
"loss": 2.5961,
"step": 55
},
{
"epoch": 0.20477815699658702,
"grad_norm": 1.6640625,
"learning_rate": 2.7272727272727273e-05,
"loss": 2.5046,
"step": 60
},
{
"epoch": 0.22184300341296928,
"grad_norm": 1.0234375,
"learning_rate": 2.954545454545455e-05,
"loss": 2.3975,
"step": 65
},
{
"epoch": 0.23890784982935154,
"grad_norm": 1.390625,
"learning_rate": 3.181818181818182e-05,
"loss": 2.3091,
"step": 70
},
{
"epoch": 0.25597269624573377,
"grad_norm": 1.1171875,
"learning_rate": 3.409090909090909e-05,
"loss": 2.2036,
"step": 75
},
{
"epoch": 0.27303754266211605,
"grad_norm": 0.84375,
"learning_rate": 3.6363636363636364e-05,
"loss": 2.113,
"step": 80
},
{
"epoch": 0.2901023890784983,
"grad_norm": 1.7421875,
"learning_rate": 3.8636363636363636e-05,
"loss": 2.043,
"step": 85
},
{
"epoch": 0.30716723549488056,
"grad_norm": 6.375,
"learning_rate": 4.0909090909090915e-05,
"loss": 1.9568,
"step": 90
},
{
"epoch": 0.3242320819112628,
"grad_norm": 0.71484375,
"learning_rate": 4.318181818181819e-05,
"loss": 1.8927,
"step": 95
},
{
"epoch": 0.3412969283276451,
"grad_norm": 1.5078125,
"learning_rate": 4.545454545454546e-05,
"loss": 1.8394,
"step": 100
},
{
"epoch": 0.3583617747440273,
"grad_norm": 25.625,
"learning_rate": 4.772727272727273e-05,
"loss": 1.7808,
"step": 105
},
{
"epoch": 0.37542662116040953,
"grad_norm": 0.54296875,
"learning_rate": 5e-05,
"loss": 1.7467,
"step": 110
},
{
"epoch": 0.3924914675767918,
"grad_norm": 0.8671875,
"learning_rate": 5.2272727272727274e-05,
"loss": 1.6988,
"step": 115
},
{
"epoch": 0.40955631399317405,
"grad_norm": 0.609375,
"learning_rate": 5.4545454545454546e-05,
"loss": 1.6442,
"step": 120
},
{
"epoch": 0.42662116040955633,
"grad_norm": 0.5078125,
"learning_rate": 5.6818181818181825e-05,
"loss": 1.5875,
"step": 125
},
{
"epoch": 0.44368600682593856,
"grad_norm": 0.890625,
"learning_rate": 5.90909090909091e-05,
"loss": 1.5646,
"step": 130
},
{
"epoch": 0.46075085324232085,
"grad_norm": 0.53125,
"learning_rate": 6.136363636363636e-05,
"loss": 1.5244,
"step": 135
},
{
"epoch": 0.4778156996587031,
"grad_norm": 0.39453125,
"learning_rate": 6.363636363636364e-05,
"loss": 1.4945,
"step": 140
},
{
"epoch": 0.4948805460750853,
"grad_norm": 0.447265625,
"learning_rate": 6.59090909090909e-05,
"loss": 1.469,
"step": 145
},
{
"epoch": 0.5119453924914675,
"grad_norm": 0.75390625,
"learning_rate": 6.818181818181818e-05,
"loss": 1.4478,
"step": 150
},
{
"epoch": 0.5290102389078498,
"grad_norm": 0.392578125,
"learning_rate": 7.045454545454546e-05,
"loss": 1.4291,
"step": 155
},
{
"epoch": 0.5460750853242321,
"grad_norm": 0.314453125,
"learning_rate": 7.272727272727273e-05,
"loss": 1.4056,
"step": 160
},
{
"epoch": 0.5631399317406144,
"grad_norm": 0.326171875,
"learning_rate": 7.500000000000001e-05,
"loss": 1.3839,
"step": 165
},
{
"epoch": 0.5802047781569966,
"grad_norm": 0.283203125,
"learning_rate": 7.727272727272727e-05,
"loss": 1.3664,
"step": 170
},
{
"epoch": 0.5972696245733788,
"grad_norm": 0.39453125,
"learning_rate": 7.954545454545455e-05,
"loss": 1.3557,
"step": 175
},
{
"epoch": 0.6143344709897611,
"grad_norm": 0.33984375,
"learning_rate": 8.181818181818183e-05,
"loss": 1.3317,
"step": 180
},
{
"epoch": 0.6313993174061433,
"grad_norm": 0.3125,
"learning_rate": 8.40909090909091e-05,
"loss": 1.3323,
"step": 185
},
{
"epoch": 0.6484641638225256,
"grad_norm": 0.384765625,
"learning_rate": 8.636363636363637e-05,
"loss": 1.3129,
"step": 190
},
{
"epoch": 0.6655290102389079,
"grad_norm": 0.435546875,
"learning_rate": 8.863636363636364e-05,
"loss": 1.3194,
"step": 195
},
{
"epoch": 0.6825938566552902,
"grad_norm": 0.4140625,
"learning_rate": 9.090909090909092e-05,
"loss": 1.2992,
"step": 200
},
{
"epoch": 0.6996587030716723,
"grad_norm": 0.296875,
"learning_rate": 9.318181818181818e-05,
"loss": 1.2934,
"step": 205
},
{
"epoch": 0.7167235494880546,
"grad_norm": 0.515625,
"learning_rate": 9.545454545454546e-05,
"loss": 1.2759,
"step": 210
},
{
"epoch": 0.7337883959044369,
"grad_norm": 0.52734375,
"learning_rate": 9.772727272727274e-05,
"loss": 1.2775,
"step": 215
},
{
"epoch": 0.7508532423208191,
"grad_norm": 0.357421875,
"learning_rate": 0.0001,
"loss": 1.2696,
"step": 220
},
{
"epoch": 0.7679180887372014,
"grad_norm": 0.29296875,
"learning_rate": 0.00010227272727272727,
"loss": 1.2621,
"step": 225
},
{
"epoch": 0.7849829351535836,
"grad_norm": 0.3359375,
"learning_rate": 0.00010454545454545455,
"loss": 1.251,
"step": 230
},
{
"epoch": 0.8020477815699659,
"grad_norm": 0.419921875,
"learning_rate": 0.00010681818181818181,
"loss": 1.2544,
"step": 235
},
{
"epoch": 0.8191126279863481,
"grad_norm": 0.48046875,
"learning_rate": 0.00010909090909090909,
"loss": 1.2528,
"step": 240
},
{
"epoch": 0.8361774744027304,
"grad_norm": 0.5234375,
"learning_rate": 0.00011136363636363636,
"loss": 1.2459,
"step": 245
},
{
"epoch": 0.8532423208191127,
"grad_norm": 0.455078125,
"learning_rate": 0.00011363636363636365,
"loss": 1.2322,
"step": 250
},
{
"epoch": 0.8703071672354948,
"grad_norm": 0.451171875,
"learning_rate": 0.00011590909090909093,
"loss": 1.2154,
"step": 255
},
{
"epoch": 0.8873720136518771,
"grad_norm": 0.44140625,
"learning_rate": 0.0001181818181818182,
"loss": 1.2258,
"step": 260
},
{
"epoch": 0.9044368600682594,
"grad_norm": 0.56640625,
"learning_rate": 0.00012045454545454546,
"loss": 1.213,
"step": 265
},
{
"epoch": 0.9215017064846417,
"grad_norm": 0.46875,
"learning_rate": 0.00012272727272727272,
"loss": 1.224,
"step": 270
},
{
"epoch": 0.9385665529010239,
"grad_norm": 0.51171875,
"learning_rate": 0.000125,
"loss": 1.2093,
"step": 275
},
{
"epoch": 0.9556313993174061,
"grad_norm": 0.90234375,
"learning_rate": 0.00012727272727272728,
"loss": 1.2132,
"step": 280
},
{
"epoch": 0.9726962457337884,
"grad_norm": 0.63671875,
"learning_rate": 0.00012954545454545456,
"loss": 1.2083,
"step": 285
},
{
"epoch": 0.9897610921501706,
"grad_norm": 0.671875,
"learning_rate": 0.0001318181818181818,
"loss": 1.2085,
"step": 290
},
{
"epoch": 1.0,
"eval_loss": 2.486323833465576,
"eval_runtime": 0.5451,
"eval_samples_per_second": 18.345,
"eval_steps_per_second": 1.834,
"step": 293
},
{
"epoch": 1.006825938566553,
"grad_norm": 0.52734375,
"learning_rate": 0.0001340909090909091,
"loss": 1.1892,
"step": 295
},
{
"epoch": 1.023890784982935,
"grad_norm": 0.4296875,
"learning_rate": 0.00013636363636363637,
"loss": 1.191,
"step": 300
},
{
"epoch": 1.0409556313993173,
"grad_norm": 0.5390625,
"learning_rate": 0.00013863636363636365,
"loss": 1.18,
"step": 305
},
{
"epoch": 1.0580204778156996,
"grad_norm": 0.703125,
"learning_rate": 0.00014090909090909093,
"loss": 1.1964,
"step": 310
},
{
"epoch": 1.075085324232082,
"grad_norm": 0.435546875,
"learning_rate": 0.0001431818181818182,
"loss": 1.1877,
"step": 315
},
{
"epoch": 1.0921501706484642,
"grad_norm": 0.59375,
"learning_rate": 0.00014545454545454546,
"loss": 1.1846,
"step": 320
},
{
"epoch": 1.1092150170648465,
"grad_norm": 0.8125,
"learning_rate": 0.00014772727272727274,
"loss": 1.1833,
"step": 325
},
{
"epoch": 1.1262798634812285,
"grad_norm": 0.9296875,
"learning_rate": 0.00015000000000000001,
"loss": 1.1704,
"step": 330
},
{
"epoch": 1.1433447098976108,
"grad_norm": 1.5703125,
"learning_rate": 0.00015227272727272727,
"loss": 1.1886,
"step": 335
},
{
"epoch": 1.1604095563139931,
"grad_norm": 0.4609375,
"learning_rate": 0.00015454545454545454,
"loss": 1.1759,
"step": 340
},
{
"epoch": 1.1774744027303754,
"grad_norm": 0.337890625,
"learning_rate": 0.00015681818181818182,
"loss": 1.1712,
"step": 345
},
{
"epoch": 1.1945392491467577,
"grad_norm": 0.48046875,
"learning_rate": 0.0001590909090909091,
"loss": 1.1637,
"step": 350
},
{
"epoch": 1.21160409556314,
"grad_norm": 0.58203125,
"learning_rate": 0.00016136363636363635,
"loss": 1.1657,
"step": 355
},
{
"epoch": 1.2286689419795223,
"grad_norm": 0.54296875,
"learning_rate": 0.00016363636363636366,
"loss": 1.1745,
"step": 360
},
{
"epoch": 1.2457337883959045,
"grad_norm": 0.421875,
"learning_rate": 0.00016590909090909094,
"loss": 1.1496,
"step": 365
},
{
"epoch": 1.2627986348122868,
"grad_norm": 0.546875,
"learning_rate": 0.0001681818181818182,
"loss": 1.1653,
"step": 370
},
{
"epoch": 1.2798634812286689,
"grad_norm": 0.5078125,
"learning_rate": 0.00017045454545454547,
"loss": 1.1702,
"step": 375
},
{
"epoch": 1.2969283276450512,
"grad_norm": 0.53125,
"learning_rate": 0.00017272727272727275,
"loss": 1.15,
"step": 380
},
{
"epoch": 1.3139931740614335,
"grad_norm": 1.2421875,
"learning_rate": 0.000175,
"loss": 1.1615,
"step": 385
},
{
"epoch": 1.3310580204778157,
"grad_norm": 1.625,
"learning_rate": 0.00017727272727272728,
"loss": 1.1662,
"step": 390
},
{
"epoch": 1.348122866894198,
"grad_norm": 0.5,
"learning_rate": 0.00017954545454545456,
"loss": 1.1579,
"step": 395
},
{
"epoch": 1.36518771331058,
"grad_norm": 0.90234375,
"learning_rate": 0.00018181818181818183,
"loss": 1.1628,
"step": 400
},
{
"epoch": 1.3822525597269624,
"grad_norm": 0.3515625,
"learning_rate": 0.00018409090909090909,
"loss": 1.1521,
"step": 405
},
{
"epoch": 1.3993174061433447,
"grad_norm": 0.455078125,
"learning_rate": 0.00018636363636363636,
"loss": 1.1422,
"step": 410
},
{
"epoch": 1.416382252559727,
"grad_norm": 0.52734375,
"learning_rate": 0.00018863636363636364,
"loss": 1.1408,
"step": 415
},
{
"epoch": 1.4334470989761092,
"grad_norm": 0.53515625,
"learning_rate": 0.00019090909090909092,
"loss": 1.1356,
"step": 420
},
{
"epoch": 1.4505119453924915,
"grad_norm": 0.46875,
"learning_rate": 0.0001931818181818182,
"loss": 1.1497,
"step": 425
},
{
"epoch": 1.4675767918088738,
"grad_norm": 0.47265625,
"learning_rate": 0.00019545454545454548,
"loss": 1.1437,
"step": 430
},
{
"epoch": 1.484641638225256,
"grad_norm": 0.42578125,
"learning_rate": 0.00019772727272727273,
"loss": 1.1518,
"step": 435
},
{
"epoch": 1.5017064846416384,
"grad_norm": 0.38671875,
"learning_rate": 0.0002,
"loss": 1.1518,
"step": 440
},
{
"epoch": 1.5187713310580204,
"grad_norm": 0.3984375,
"learning_rate": 0.000199999211292062,
"loss": 1.1498,
"step": 445
},
{
"epoch": 1.5358361774744027,
"grad_norm": 0.388671875,
"learning_rate": 0.00019999684518068916,
"loss": 1.1378,
"step": 450
},
{
"epoch": 1.552901023890785,
"grad_norm": 0.87109375,
"learning_rate": 0.00019999290170320485,
"loss": 1.1434,
"step": 455
},
{
"epoch": 1.5699658703071673,
"grad_norm": 0.6953125,
"learning_rate": 0.00019998738092181421,
"loss": 1.1417,
"step": 460
},
{
"epoch": 1.5870307167235493,
"grad_norm": 0.79296875,
"learning_rate": 0.00019998028292360286,
"loss": 1.1329,
"step": 465
},
{
"epoch": 1.6040955631399316,
"grad_norm": 0.5625,
"learning_rate": 0.00019997160782053578,
"loss": 1.1339,
"step": 470
},
{
"epoch": 1.621160409556314,
"grad_norm": 0.384765625,
"learning_rate": 0.00019996135574945544,
"loss": 1.1273,
"step": 475
},
{
"epoch": 1.6382252559726962,
"grad_norm": 0.455078125,
"learning_rate": 0.00019994952687207954,
"loss": 1.1343,
"step": 480
},
{
"epoch": 1.6552901023890785,
"grad_norm": 0.69921875,
"learning_rate": 0.00019993612137499876,
"loss": 1.1374,
"step": 485
},
{
"epoch": 1.6723549488054608,
"grad_norm": 0.88671875,
"learning_rate": 0.00019992113946967353,
"loss": 1.1368,
"step": 490
},
{
"epoch": 1.689419795221843,
"grad_norm": 0.490234375,
"learning_rate": 0.00019990458139243077,
"loss": 1.1289,
"step": 495
},
{
"epoch": 1.7064846416382253,
"grad_norm": 0.53515625,
"learning_rate": 0.00019988644740446022,
"loss": 1.1255,
"step": 500
},
{
"epoch": 1.7235494880546076,
"grad_norm": 0.5234375,
"learning_rate": 0.00019986673779181033,
"loss": 1.1149,
"step": 505
},
{
"epoch": 1.74061433447099,
"grad_norm": 0.41015625,
"learning_rate": 0.0001998454528653836,
"loss": 1.1241,
"step": 510
},
{
"epoch": 1.757679180887372,
"grad_norm": 0.41796875,
"learning_rate": 0.0001998225929609319,
"loss": 1.1252,
"step": 515
},
{
"epoch": 1.7747440273037542,
"grad_norm": 0.458984375,
"learning_rate": 0.00019979815843905097,
"loss": 1.1292,
"step": 520
},
{
"epoch": 1.7918088737201365,
"grad_norm": 0.400390625,
"learning_rate": 0.0001997721496851748,
"loss": 1.1147,
"step": 525
},
{
"epoch": 1.8088737201365188,
"grad_norm": 0.53125,
"learning_rate": 0.00019974456710956964,
"loss": 1.1155,
"step": 530
},
{
"epoch": 1.8259385665529009,
"grad_norm": 0.546875,
"learning_rate": 0.00019971541114732741,
"loss": 1.1213,
"step": 535
},
{
"epoch": 1.8430034129692832,
"grad_norm": 0.40234375,
"learning_rate": 0.0001996846822583589,
"loss": 1.1257,
"step": 540
},
{
"epoch": 1.8600682593856654,
"grad_norm": 0.38671875,
"learning_rate": 0.00019965238092738643,
"loss": 1.1217,
"step": 545
},
{
"epoch": 1.8771331058020477,
"grad_norm": 0.5390625,
"learning_rate": 0.0001996185076639364,
"loss": 1.122,
"step": 550
},
{
"epoch": 1.89419795221843,
"grad_norm": 0.390625,
"learning_rate": 0.00019958306300233098,
"loss": 1.1236,
"step": 555
},
{
"epoch": 1.9112627986348123,
"grad_norm": 0.5390625,
"learning_rate": 0.00019954604750167993,
"loss": 1.122,
"step": 560
},
{
"epoch": 1.9283276450511946,
"grad_norm": 0.66796875,
"learning_rate": 0.00019950746174587163,
"loss": 1.1271,
"step": 565
},
{
"epoch": 1.9453924914675769,
"grad_norm": 0.47265625,
"learning_rate": 0.0001994673063435639,
"loss": 1.1064,
"step": 570
},
{
"epoch": 1.9624573378839592,
"grad_norm": 0.3359375,
"learning_rate": 0.0001994255819281744,
"loss": 1.1186,
"step": 575
},
{
"epoch": 1.9795221843003414,
"grad_norm": 0.63671875,
"learning_rate": 0.0001993822891578708,
"loss": 1.1054,
"step": 580
},
{
"epoch": 1.9965870307167235,
"grad_norm": 0.68359375,
"learning_rate": 0.00019933742871556,
"loss": 1.1135,
"step": 585
},
{
"epoch": 2.0,
"eval_loss": 2.4516425132751465,
"eval_runtime": 0.5387,
"eval_samples_per_second": 18.563,
"eval_steps_per_second": 1.856,
"step": 586
},
{
"epoch": 2.013651877133106,
"grad_norm": 0.66015625,
"learning_rate": 0.00019929100130887782,
"loss": 1.1079,
"step": 590
},
{
"epoch": 2.030716723549488,
"grad_norm": 0.94140625,
"learning_rate": 0.0001992430076701775,
"loss": 1.088,
"step": 595
},
{
"epoch": 2.04778156996587,
"grad_norm": 0.400390625,
"learning_rate": 0.00019919344855651833,
"loss": 1.0921,
"step": 600
},
{
"epoch": 2.0648464163822524,
"grad_norm": 0.59375,
"learning_rate": 0.00019914232474965365,
"loss": 1.0909,
"step": 605
},
{
"epoch": 2.0819112627986347,
"grad_norm": 0.42578125,
"learning_rate": 0.00019908963705601846,
"loss": 1.0986,
"step": 610
},
{
"epoch": 2.098976109215017,
"grad_norm": 0.435546875,
"learning_rate": 0.0001990353863067169,
"loss": 1.0925,
"step": 615
},
{
"epoch": 2.1160409556313993,
"grad_norm": 0.640625,
"learning_rate": 0.00019897957335750878,
"loss": 1.0887,
"step": 620
},
{
"epoch": 2.1331058020477816,
"grad_norm": 0.5078125,
"learning_rate": 0.00019892219908879653,
"loss": 1.0991,
"step": 625
},
{
"epoch": 2.150170648464164,
"grad_norm": 0.416015625,
"learning_rate": 0.00019886326440561093,
"loss": 1.0949,
"step": 630
},
{
"epoch": 2.167235494880546,
"grad_norm": 0.373046875,
"learning_rate": 0.00019880277023759702,
"loss": 1.0841,
"step": 635
},
{
"epoch": 2.1843003412969284,
"grad_norm": 0.78515625,
"learning_rate": 0.0001987407175389994,
"loss": 1.0947,
"step": 640
},
{
"epoch": 2.2013651877133107,
"grad_norm": 0.42578125,
"learning_rate": 0.0001986771072886472,
"loss": 1.1026,
"step": 645
},
{
"epoch": 2.218430034129693,
"grad_norm": 0.392578125,
"learning_rate": 0.00019861194048993863,
"loss": 1.0918,
"step": 650
},
{
"epoch": 2.2354948805460753,
"grad_norm": 0.41015625,
"learning_rate": 0.0001985452181708251,
"loss": 1.0903,
"step": 655
},
{
"epoch": 2.252559726962457,
"grad_norm": 0.7109375,
"learning_rate": 0.00019847694138379506,
"loss": 1.0978,
"step": 660
},
{
"epoch": 2.26962457337884,
"grad_norm": 0.4609375,
"learning_rate": 0.0001984071112058574,
"loss": 1.0864,
"step": 665
},
{
"epoch": 2.2866894197952217,
"grad_norm": 0.341796875,
"learning_rate": 0.00019833572873852444,
"loss": 1.0896,
"step": 670
},
{
"epoch": 2.303754266211604,
"grad_norm": 0.53125,
"learning_rate": 0.00019826279510779454,
"loss": 1.0962,
"step": 675
},
{
"epoch": 2.3208191126279862,
"grad_norm": 0.54296875,
"learning_rate": 0.00019818831146413434,
"loss": 1.0766,
"step": 680
},
{
"epoch": 2.3378839590443685,
"grad_norm": 0.337890625,
"learning_rate": 0.0001981122789824607,
"loss": 1.0853,
"step": 685
},
{
"epoch": 2.354948805460751,
"grad_norm": 0.69140625,
"learning_rate": 0.0001980346988621221,
"loss": 1.0788,
"step": 690
},
{
"epoch": 2.372013651877133,
"grad_norm": 0.96875,
"learning_rate": 0.00019795557232687956,
"loss": 1.0804,
"step": 695
},
{
"epoch": 2.3890784982935154,
"grad_norm": 0.470703125,
"learning_rate": 0.0001978749006248877,
"loss": 1.0674,
"step": 700
},
{
"epoch": 2.4061433447098977,
"grad_norm": 0.326171875,
"learning_rate": 0.00019779268502867473,
"loss": 1.0931,
"step": 705
},
{
"epoch": 2.42320819112628,
"grad_norm": 0.458984375,
"learning_rate": 0.0001977089268351225,
"loss": 1.0854,
"step": 710
},
{
"epoch": 2.4402730375426622,
"grad_norm": 0.43359375,
"learning_rate": 0.00019762362736544607,
"loss": 1.0858,
"step": 715
},
{
"epoch": 2.4573378839590445,
"grad_norm": 0.396484375,
"learning_rate": 0.00019753678796517282,
"loss": 1.0835,
"step": 720
},
{
"epoch": 2.474402730375427,
"grad_norm": 0.59375,
"learning_rate": 0.00019744841000412123,
"loss": 1.0881,
"step": 725
},
{
"epoch": 2.491467576791809,
"grad_norm": 0.6171875,
"learning_rate": 0.00019735849487637929,
"loss": 1.091,
"step": 730
},
{
"epoch": 2.508532423208191,
"grad_norm": 0.5625,
"learning_rate": 0.0001972670440002825,
"loss": 1.0877,
"step": 735
},
{
"epoch": 2.5255972696245736,
"grad_norm": 0.419921875,
"learning_rate": 0.00019717405881839145,
"loss": 1.0777,
"step": 740
},
{
"epoch": 2.5426621160409555,
"grad_norm": 0.380859375,
"learning_rate": 0.00019707954079746927,
"loss": 1.0934,
"step": 745
},
{
"epoch": 2.5597269624573378,
"grad_norm": 0.439453125,
"learning_rate": 0.00019698349142845814,
"loss": 1.085,
"step": 750
},
{
"epoch": 2.57679180887372,
"grad_norm": 0.38671875,
"learning_rate": 0.00019688591222645607,
"loss": 1.0744,
"step": 755
},
{
"epoch": 2.5938566552901023,
"grad_norm": 0.4375,
"learning_rate": 0.00019678680473069293,
"loss": 1.0818,
"step": 760
},
{
"epoch": 2.6109215017064846,
"grad_norm": 0.3984375,
"learning_rate": 0.00019668617050450603,
"loss": 1.0824,
"step": 765
},
{
"epoch": 2.627986348122867,
"grad_norm": 0.4921875,
"learning_rate": 0.00019658401113531565,
"loss": 1.0828,
"step": 770
},
{
"epoch": 2.645051194539249,
"grad_norm": 1.09375,
"learning_rate": 0.00019648032823459994,
"loss": 1.0884,
"step": 775
},
{
"epoch": 2.6621160409556315,
"grad_norm": 0.55859375,
"learning_rate": 0.00019637512343786937,
"loss": 1.0835,
"step": 780
},
{
"epoch": 2.6791808873720138,
"grad_norm": 0.484375,
"learning_rate": 0.00019626839840464119,
"loss": 1.0828,
"step": 785
},
{
"epoch": 2.696245733788396,
"grad_norm": 0.376953125,
"learning_rate": 0.0001961601548184129,
"loss": 1.0881,
"step": 790
},
{
"epoch": 2.7133105802047783,
"grad_norm": 0.35546875,
"learning_rate": 0.00019605039438663614,
"loss": 1.0772,
"step": 795
},
{
"epoch": 2.73037542662116,
"grad_norm": 0.349609375,
"learning_rate": 0.0001959391188406893,
"loss": 1.0677,
"step": 800
},
{
"epoch": 2.747440273037543,
"grad_norm": 0.486328125,
"learning_rate": 0.00019582632993585052,
"loss": 1.0815,
"step": 805
},
{
"epoch": 2.7645051194539247,
"grad_norm": 0.470703125,
"learning_rate": 0.00019571202945126994,
"loss": 1.0763,
"step": 810
},
{
"epoch": 2.781569965870307,
"grad_norm": 0.396484375,
"learning_rate": 0.0001955962191899415,
"loss": 1.0684,
"step": 815
},
{
"epoch": 2.7986348122866893,
"grad_norm": 0.373046875,
"learning_rate": 0.00019547890097867468,
"loss": 1.0847,
"step": 820
},
{
"epoch": 2.8156996587030716,
"grad_norm": 0.474609375,
"learning_rate": 0.00019536007666806556,
"loss": 1.071,
"step": 825
},
{
"epoch": 2.832764505119454,
"grad_norm": 0.380859375,
"learning_rate": 0.00019523974813246767,
"loss": 1.0873,
"step": 830
},
{
"epoch": 2.849829351535836,
"grad_norm": 0.40234375,
"learning_rate": 0.00019511791726996243,
"loss": 1.0676,
"step": 835
},
{
"epoch": 2.8668941979522184,
"grad_norm": 0.51953125,
"learning_rate": 0.0001949945860023292,
"loss": 1.0748,
"step": 840
},
{
"epoch": 2.8839590443686007,
"grad_norm": 0.384765625,
"learning_rate": 0.00019486975627501502,
"loss": 1.0716,
"step": 845
},
{
"epoch": 2.901023890784983,
"grad_norm": 0.38671875,
"learning_rate": 0.0001947434300571038,
"loss": 1.0777,
"step": 850
},
{
"epoch": 2.9180887372013653,
"grad_norm": 0.365234375,
"learning_rate": 0.00019461560934128533,
"loss": 1.0733,
"step": 855
},
{
"epoch": 2.9351535836177476,
"grad_norm": 0.42578125,
"learning_rate": 0.0001944862961438239,
"loss": 1.0582,
"step": 860
},
{
"epoch": 2.9522184300341294,
"grad_norm": 0.462890625,
"learning_rate": 0.00019435549250452645,
"loss": 1.0657,
"step": 865
},
{
"epoch": 2.969283276450512,
"grad_norm": 1.1171875,
"learning_rate": 0.0001942232004867103,
"loss": 1.0746,
"step": 870
},
{
"epoch": 2.986348122866894,
"grad_norm": 0.466796875,
"learning_rate": 0.0001940894221771708,
"loss": 1.0715,
"step": 875
},
{
"epoch": 3.0,
"eval_loss": 2.447284698486328,
"eval_runtime": 0.553,
"eval_samples_per_second": 18.083,
"eval_steps_per_second": 1.808,
"step": 879
},
{
"epoch": 3.0034129692832763,
"grad_norm": 0.80859375,
"learning_rate": 0.00019395415968614813,
"loss": 1.0736,
"step": 880
},
{
"epoch": 3.0204778156996586,
"grad_norm": 0.47265625,
"learning_rate": 0.00019381741514729443,
"loss": 1.0618,
"step": 885
},
{
"epoch": 3.037542662116041,
"grad_norm": 0.390625,
"learning_rate": 0.0001936791907176397,
"loss": 1.0571,
"step": 890
},
{
"epoch": 3.054607508532423,
"grad_norm": 0.46484375,
"learning_rate": 0.00019353948857755803,
"loss": 1.0626,
"step": 895
},
{
"epoch": 3.0716723549488054,
"grad_norm": 0.357421875,
"learning_rate": 0.00019339831093073318,
"loss": 1.053,
"step": 900
},
{
"epoch": 3.0887372013651877,
"grad_norm": 0.380859375,
"learning_rate": 0.00019325566000412376,
"loss": 1.06,
"step": 905
},
{
"epoch": 3.10580204778157,
"grad_norm": 0.38671875,
"learning_rate": 0.0001931115380479281,
"loss": 1.0452,
"step": 910
},
{
"epoch": 3.1228668941979523,
"grad_norm": 0.515625,
"learning_rate": 0.00019296594733554892,
"loss": 1.0642,
"step": 915
},
{
"epoch": 3.1399317406143346,
"grad_norm": 0.5,
"learning_rate": 0.0001928188901635571,
"loss": 1.0474,
"step": 920
},
{
"epoch": 3.156996587030717,
"grad_norm": 0.380859375,
"learning_rate": 0.00019267036885165588,
"loss": 1.0526,
"step": 925
},
{
"epoch": 3.174061433447099,
"grad_norm": 0.4296875,
"learning_rate": 0.00019252038574264405,
"loss": 1.061,
"step": 930
},
{
"epoch": 3.1911262798634814,
"grad_norm": 0.443359375,
"learning_rate": 0.00019236894320237894,
"loss": 1.0519,
"step": 935
},
{
"epoch": 3.2081911262798632,
"grad_norm": 0.458984375,
"learning_rate": 0.00019221604361973919,
"loss": 1.0479,
"step": 940
},
{
"epoch": 3.2252559726962455,
"grad_norm": 0.50390625,
"learning_rate": 0.00019206168940658712,
"loss": 1.049,
"step": 945
},
{
"epoch": 3.242320819112628,
"grad_norm": 0.462890625,
"learning_rate": 0.00019190588299773062,
"loss": 1.0474,
"step": 950
},
{
"epoch": 3.25938566552901,
"grad_norm": 0.462890625,
"learning_rate": 0.00019174862685088472,
"loss": 1.06,
"step": 955
},
{
"epoch": 3.2764505119453924,
"grad_norm": 0.373046875,
"learning_rate": 0.0001915899234466328,
"loss": 1.0464,
"step": 960
},
{
"epoch": 3.2935153583617747,
"grad_norm": 0.48046875,
"learning_rate": 0.00019142977528838762,
"loss": 1.0531,
"step": 965
},
{
"epoch": 3.310580204778157,
"grad_norm": 0.380859375,
"learning_rate": 0.0001912681849023516,
"loss": 1.0518,
"step": 970
},
{
"epoch": 3.3276450511945392,
"grad_norm": 0.447265625,
"learning_rate": 0.00019110515483747716,
"loss": 1.0535,
"step": 975
},
{
"epoch": 3.3447098976109215,
"grad_norm": 0.625,
"learning_rate": 0.0001909406876654264,
"loss": 1.0559,
"step": 980
},
{
"epoch": 3.361774744027304,
"grad_norm": 0.51953125,
"learning_rate": 0.00019077478598053063,
"loss": 1.0528,
"step": 985
},
{
"epoch": 3.378839590443686,
"grad_norm": 0.46875,
"learning_rate": 0.00019060745239974936,
"loss": 1.0431,
"step": 990
},
{
"epoch": 3.3959044368600684,
"grad_norm": 0.63671875,
"learning_rate": 0.0001904386895626291,
"loss": 1.0456,
"step": 995
},
{
"epoch": 3.4129692832764507,
"grad_norm": 0.48828125,
"learning_rate": 0.00019026850013126157,
"loss": 1.0579,
"step": 1000
},
{
"epoch": 3.430034129692833,
"grad_norm": 0.625,
"learning_rate": 0.0001900968867902419,
"loss": 1.0592,
"step": 1005
},
{
"epoch": 3.4470989761092152,
"grad_norm": 0.51171875,
"learning_rate": 0.00018992385224662623,
"loss": 1.0476,
"step": 1010
},
{
"epoch": 3.464163822525597,
"grad_norm": 0.470703125,
"learning_rate": 0.00018974939922988883,
"loss": 1.0517,
"step": 1015
},
{
"epoch": 3.4812286689419794,
"grad_norm": 0.423828125,
"learning_rate": 0.00018957353049187936,
"loss": 1.0607,
"step": 1020
},
{
"epoch": 3.4982935153583616,
"grad_norm": 0.4765625,
"learning_rate": 0.00018939624880677918,
"loss": 1.0502,
"step": 1025
},
{
"epoch": 3.515358361774744,
"grad_norm": 0.3671875,
"learning_rate": 0.0001892175569710577,
"loss": 1.041,
"step": 1030
},
{
"epoch": 3.532423208191126,
"grad_norm": 0.52734375,
"learning_rate": 0.00018903745780342839,
"loss": 1.0382,
"step": 1035
},
{
"epoch": 3.5494880546075085,
"grad_norm": 0.3984375,
"learning_rate": 0.00018885595414480405,
"loss": 1.0426,
"step": 1040
},
{
"epoch": 3.5665529010238908,
"grad_norm": 0.400390625,
"learning_rate": 0.0001886730488582522,
"loss": 1.0524,
"step": 1045
},
{
"epoch": 3.583617747440273,
"grad_norm": 0.58203125,
"learning_rate": 0.00018848874482894993,
"loss": 1.0371,
"step": 1050
},
{
"epoch": 3.6006825938566553,
"grad_norm": 0.412109375,
"learning_rate": 0.00018830304496413822,
"loss": 1.0571,
"step": 1055
},
{
"epoch": 3.6177474402730376,
"grad_norm": 0.33984375,
"learning_rate": 0.00018811595219307622,
"loss": 1.0458,
"step": 1060
},
{
"epoch": 3.63481228668942,
"grad_norm": 0.455078125,
"learning_rate": 0.000187927469466995,
"loss": 1.0474,
"step": 1065
},
{
"epoch": 3.651877133105802,
"grad_norm": 0.37109375,
"learning_rate": 0.00018773759975905098,
"loss": 1.0438,
"step": 1070
},
{
"epoch": 3.6689419795221845,
"grad_norm": 0.384765625,
"learning_rate": 0.00018754634606427914,
"loss": 1.0577,
"step": 1075
},
{
"epoch": 3.6860068259385663,
"grad_norm": 0.435546875,
"learning_rate": 0.00018735371139954558,
"loss": 1.0522,
"step": 1080
},
{
"epoch": 3.703071672354949,
"grad_norm": 0.55859375,
"learning_rate": 0.0001871596988035001,
"loss": 1.0622,
"step": 1085
},
{
"epoch": 3.720136518771331,
"grad_norm": 0.53125,
"learning_rate": 0.00018696431133652817,
"loss": 1.0404,
"step": 1090
},
{
"epoch": 3.737201365187713,
"grad_norm": 0.41796875,
"learning_rate": 0.00018676755208070275,
"loss": 1.0576,
"step": 1095
},
{
"epoch": 3.7542662116040955,
"grad_norm": 0.396484375,
"learning_rate": 0.00018656942413973555,
"loss": 1.0525,
"step": 1100
},
{
"epoch": 3.7713310580204777,
"grad_norm": 0.392578125,
"learning_rate": 0.0001863699306389282,
"loss": 1.047,
"step": 1105
},
{
"epoch": 3.78839590443686,
"grad_norm": 0.54296875,
"learning_rate": 0.0001861690747251228,
"loss": 1.0547,
"step": 1110
},
{
"epoch": 3.8054607508532423,
"grad_norm": 0.455078125,
"learning_rate": 0.00018596685956665245,
"loss": 1.0366,
"step": 1115
},
{
"epoch": 3.8225255972696246,
"grad_norm": 0.373046875,
"learning_rate": 0.00018576328835329117,
"loss": 1.0444,
"step": 1120
},
{
"epoch": 3.839590443686007,
"grad_norm": 0.498046875,
"learning_rate": 0.00018555836429620358,
"loss": 1.0428,
"step": 1125
},
{
"epoch": 3.856655290102389,
"grad_norm": 0.4453125,
"learning_rate": 0.00018535209062789433,
"loss": 1.0425,
"step": 1130
},
{
"epoch": 3.8737201365187715,
"grad_norm": 0.392578125,
"learning_rate": 0.00018514447060215698,
"loss": 1.0503,
"step": 1135
},
{
"epoch": 3.8907849829351537,
"grad_norm": 0.384765625,
"learning_rate": 0.00018493550749402278,
"loss": 1.0376,
"step": 1140
},
{
"epoch": 3.9078498293515356,
"grad_norm": 0.3984375,
"learning_rate": 0.00018472520459970898,
"loss": 1.054,
"step": 1145
},
{
"epoch": 3.9249146757679183,
"grad_norm": 0.44921875,
"learning_rate": 0.0001845135652365668,
"loss": 1.0491,
"step": 1150
},
{
"epoch": 3.9419795221843,
"grad_norm": 0.37890625,
"learning_rate": 0.00018430059274302917,
"loss": 1.0454,
"step": 1155
},
{
"epoch": 3.9590443686006824,
"grad_norm": 0.365234375,
"learning_rate": 0.00018408629047855804,
"loss": 1.0466,
"step": 1160
},
{
"epoch": 3.9761092150170647,
"grad_norm": 0.34765625,
"learning_rate": 0.00018387066182359133,
"loss": 1.0356,
"step": 1165
},
{
"epoch": 3.993174061433447,
"grad_norm": 0.357421875,
"learning_rate": 0.00018365371017948964,
"loss": 1.0471,
"step": 1170
},
{
"epoch": 4.0,
"eval_loss": 2.452413558959961,
"eval_runtime": 0.5427,
"eval_samples_per_second": 18.427,
"eval_steps_per_second": 1.843,
"step": 1172
},
{
"epoch": 4.010238907849829,
"grad_norm": 0.47265625,
"learning_rate": 0.00018343543896848273,
"loss": 1.0282,
"step": 1175
},
{
"epoch": 4.027303754266212,
"grad_norm": 0.41796875,
"learning_rate": 0.00018321585163361527,
"loss": 1.0262,
"step": 1180
},
{
"epoch": 4.044368600682594,
"grad_norm": 0.365234375,
"learning_rate": 0.00018299495163869275,
"loss": 1.0263,
"step": 1185
},
{
"epoch": 4.061433447098976,
"grad_norm": 0.359375,
"learning_rate": 0.0001827727424682268,
"loss": 1.0265,
"step": 1190
},
{
"epoch": 4.078498293515358,
"grad_norm": 0.375,
"learning_rate": 0.00018254922762738008,
"loss": 1.0266,
"step": 1195
},
{
"epoch": 4.09556313993174,
"grad_norm": 0.3828125,
"learning_rate": 0.00018232441064191125,
"loss": 1.0326,
"step": 1200
},
{
"epoch": 4.112627986348123,
"grad_norm": 0.3828125,
"learning_rate": 0.0001820982950581191,
"loss": 1.0278,
"step": 1205
},
{
"epoch": 4.129692832764505,
"grad_norm": 0.46484375,
"learning_rate": 0.00018187088444278674,
"loss": 1.0206,
"step": 1210
},
{
"epoch": 4.146757679180888,
"grad_norm": 0.4140625,
"learning_rate": 0.00018164218238312535,
"loss": 1.037,
"step": 1215
},
{
"epoch": 4.163822525597269,
"grad_norm": 0.3671875,
"learning_rate": 0.00018141219248671745,
"loss": 1.0229,
"step": 1220
},
{
"epoch": 4.180887372013652,
"grad_norm": 0.376953125,
"learning_rate": 0.00018118091838146029,
"loss": 1.0223,
"step": 1225
},
{
"epoch": 4.197952218430034,
"grad_norm": 0.373046875,
"learning_rate": 0.00018094836371550824,
"loss": 1.0175,
"step": 1230
},
{
"epoch": 4.215017064846417,
"grad_norm": 0.380859375,
"learning_rate": 0.00018071453215721554,
"loss": 1.0369,
"step": 1235
},
{
"epoch": 4.2320819112627985,
"grad_norm": 0.41015625,
"learning_rate": 0.00018047942739507836,
"loss": 1.0182,
"step": 1240
},
{
"epoch": 4.249146757679181,
"grad_norm": 0.421875,
"learning_rate": 0.00018024305313767646,
"loss": 1.0192,
"step": 1245
},
{
"epoch": 4.266211604095563,
"grad_norm": 0.40625,
"learning_rate": 0.000180005413113615,
"loss": 1.0427,
"step": 1250
},
{
"epoch": 4.283276450511945,
"grad_norm": 0.42578125,
"learning_rate": 0.00017976651107146533,
"loss": 1.0313,
"step": 1255
},
{
"epoch": 4.300341296928328,
"grad_norm": 0.359375,
"learning_rate": 0.0001795263507797063,
"loss": 1.0195,
"step": 1260
},
{
"epoch": 4.3174061433447095,
"grad_norm": 0.453125,
"learning_rate": 0.00017928493602666445,
"loss": 1.0222,
"step": 1265
},
{
"epoch": 4.334470989761092,
"grad_norm": 0.5546875,
"learning_rate": 0.00017904227062045437,
"loss": 1.0183,
"step": 1270
},
{
"epoch": 4.351535836177474,
"grad_norm": 0.6328125,
"learning_rate": 0.00017879835838891875,
"loss": 1.0321,
"step": 1275
},
{
"epoch": 4.368600682593857,
"grad_norm": 0.7265625,
"learning_rate": 0.00017855320317956784,
"loss": 1.0241,
"step": 1280
},
{
"epoch": 4.385665529010239,
"grad_norm": 0.380859375,
"learning_rate": 0.00017830680885951887,
"loss": 1.019,
"step": 1285
},
{
"epoch": 4.402730375426621,
"grad_norm": 0.7265625,
"learning_rate": 0.00017805917931543492,
"loss": 1.0291,
"step": 1290
},
{
"epoch": 4.419795221843003,
"grad_norm": 0.8671875,
"learning_rate": 0.00017781031845346375,
"loss": 1.0254,
"step": 1295
},
{
"epoch": 4.436860068259386,
"grad_norm": 0.38671875,
"learning_rate": 0.00017756023019917607,
"loss": 1.0232,
"step": 1300
},
{
"epoch": 4.453924914675768,
"grad_norm": 0.384765625,
"learning_rate": 0.00017730891849750377,
"loss": 1.0267,
"step": 1305
},
{
"epoch": 4.4709897610921505,
"grad_norm": 0.38671875,
"learning_rate": 0.0001770563873126775,
"loss": 1.0282,
"step": 1310
},
{
"epoch": 4.488054607508532,
"grad_norm": 0.357421875,
"learning_rate": 0.0001768026406281642,
"loss": 1.0384,
"step": 1315
},
{
"epoch": 4.505119453924914,
"grad_norm": 0.37109375,
"learning_rate": 0.00017654768244660448,
"loss": 1.0197,
"step": 1320
},
{
"epoch": 4.522184300341297,
"grad_norm": 0.458984375,
"learning_rate": 0.00017629151678974907,
"loss": 1.023,
"step": 1325
},
{
"epoch": 4.53924914675768,
"grad_norm": 0.359375,
"learning_rate": 0.00017603414769839577,
"loss": 1.0289,
"step": 1330
},
{
"epoch": 4.5563139931740615,
"grad_norm": 0.72265625,
"learning_rate": 0.00017577557923232546,
"loss": 1.0222,
"step": 1335
},
{
"epoch": 4.573378839590443,
"grad_norm": 0.5,
"learning_rate": 0.00017551581547023819,
"loss": 1.0285,
"step": 1340
},
{
"epoch": 4.590443686006826,
"grad_norm": 0.392578125,
"learning_rate": 0.00017525486050968875,
"loss": 1.0288,
"step": 1345
},
{
"epoch": 4.607508532423208,
"grad_norm": 0.37890625,
"learning_rate": 0.00017499271846702213,
"loss": 1.0302,
"step": 1350
},
{
"epoch": 4.624573378839591,
"grad_norm": 0.419921875,
"learning_rate": 0.00017472939347730856,
"loss": 1.0358,
"step": 1355
},
{
"epoch": 4.6416382252559725,
"grad_norm": 0.451171875,
"learning_rate": 0.0001744648896942782,
"loss": 1.0278,
"step": 1360
},
{
"epoch": 4.658703071672355,
"grad_norm": 0.38671875,
"learning_rate": 0.00017419921129025576,
"loss": 1.0171,
"step": 1365
},
{
"epoch": 4.675767918088737,
"grad_norm": 0.376953125,
"learning_rate": 0.0001739323624560945,
"loss": 1.0152,
"step": 1370
},
{
"epoch": 4.69283276450512,
"grad_norm": 0.384765625,
"learning_rate": 0.00017366434740111037,
"loss": 1.0247,
"step": 1375
},
{
"epoch": 4.709897610921502,
"grad_norm": 0.431640625,
"learning_rate": 0.00017339517035301532,
"loss": 1.0212,
"step": 1380
},
{
"epoch": 4.726962457337884,
"grad_norm": 0.3828125,
"learning_rate": 0.00017312483555785086,
"loss": 1.0309,
"step": 1385
},
{
"epoch": 4.744027303754266,
"grad_norm": 0.353515625,
"learning_rate": 0.000172853347279921,
"loss": 1.0298,
"step": 1390
},
{
"epoch": 4.761092150170649,
"grad_norm": 0.373046875,
"learning_rate": 0.00017258070980172494,
"loss": 1.0215,
"step": 1395
},
{
"epoch": 4.778156996587031,
"grad_norm": 0.4453125,
"learning_rate": 0.0001723069274238895,
"loss": 1.0249,
"step": 1400
},
{
"epoch": 4.795221843003413,
"grad_norm": 0.4921875,
"learning_rate": 0.0001720320044651014,
"loss": 1.0259,
"step": 1405
},
{
"epoch": 4.812286689419795,
"grad_norm": 0.380859375,
"learning_rate": 0.00017175594526203905,
"loss": 1.0215,
"step": 1410
},
{
"epoch": 4.829351535836177,
"grad_norm": 0.42578125,
"learning_rate": 0.00017147875416930416,
"loss": 1.0272,
"step": 1415
},
{
"epoch": 4.84641638225256,
"grad_norm": 0.34765625,
"learning_rate": 0.00017120043555935298,
"loss": 1.0365,
"step": 1420
},
{
"epoch": 4.863481228668942,
"grad_norm": 0.36328125,
"learning_rate": 0.00017092099382242748,
"loss": 1.02,
"step": 1425
},
{
"epoch": 4.8805460750853245,
"grad_norm": 0.455078125,
"learning_rate": 0.00017064043336648599,
"loss": 1.021,
"step": 1430
},
{
"epoch": 4.897610921501706,
"grad_norm": 0.400390625,
"learning_rate": 0.0001703587586171337,
"loss": 1.0156,
"step": 1435
},
{
"epoch": 4.914675767918089,
"grad_norm": 0.375,
"learning_rate": 0.00017007597401755276,
"loss": 1.0283,
"step": 1440
},
{
"epoch": 4.931740614334471,
"grad_norm": 0.443359375,
"learning_rate": 0.00016979208402843237,
"loss": 1.0194,
"step": 1445
},
{
"epoch": 4.948805460750854,
"grad_norm": 0.57421875,
"learning_rate": 0.00016950709312789833,
"loss": 1.0198,
"step": 1450
},
{
"epoch": 4.965870307167235,
"grad_norm": 0.37890625,
"learning_rate": 0.00016922100581144228,
"loss": 1.028,
"step": 1455
},
{
"epoch": 4.982935153583618,
"grad_norm": 0.4765625,
"learning_rate": 0.00016893382659185105,
"loss": 1.0157,
"step": 1460
},
{
"epoch": 5.0,
"grad_norm": 0.416015625,
"learning_rate": 0.00016864555999913518,
"loss": 1.0357,
"step": 1465
},
{
"epoch": 5.0,
"eval_loss": 2.468480110168457,
"eval_runtime": 0.549,
"eval_samples_per_second": 18.214,
"eval_steps_per_second": 1.821,
"step": 1465
},
{
"epoch": 5.017064846416382,
"grad_norm": 0.380859375,
"learning_rate": 0.0001683562105804577,
"loss": 1.0001,
"step": 1470
},
{
"epoch": 5.034129692832765,
"grad_norm": 0.5078125,
"learning_rate": 0.00016806578290006225,
"loss": 0.9998,
"step": 1475
},
{
"epoch": 5.051194539249146,
"grad_norm": 0.400390625,
"learning_rate": 0.0001677742815392012,
"loss": 0.9999,
"step": 1480
},
{
"epoch": 5.068259385665529,
"grad_norm": 0.43359375,
"learning_rate": 0.00016748171109606328,
"loss": 1.0085,
"step": 1485
},
{
"epoch": 5.085324232081911,
"grad_norm": 0.416015625,
"learning_rate": 0.00016718807618570106,
"loss": 1.0018,
"step": 1490
},
{
"epoch": 5.102389078498294,
"grad_norm": 0.453125,
"learning_rate": 0.00016689338143995833,
"loss": 0.9997,
"step": 1495
},
{
"epoch": 5.1194539249146755,
"grad_norm": 0.4453125,
"learning_rate": 0.00016659763150739677,
"loss": 1.009,
"step": 1500
},
{
"epoch": 5.136518771331058,
"grad_norm": 0.357421875,
"learning_rate": 0.00016630083105322266,
"loss": 1.0047,
"step": 1505
},
{
"epoch": 5.15358361774744,
"grad_norm": 0.33984375,
"learning_rate": 0.00016600298475921365,
"loss": 1.004,
"step": 1510
},
{
"epoch": 5.170648464163823,
"grad_norm": 0.400390625,
"learning_rate": 0.00016570409732364437,
"loss": 1.0022,
"step": 1515
},
{
"epoch": 5.187713310580205,
"grad_norm": 0.427734375,
"learning_rate": 0.0001654041734612127,
"loss": 1.0113,
"step": 1520
},
{
"epoch": 5.204778156996587,
"grad_norm": 0.3828125,
"learning_rate": 0.00016510321790296525,
"loss": 1.0171,
"step": 1525
},
{
"epoch": 5.221843003412969,
"grad_norm": 0.462890625,
"learning_rate": 0.00016480123539622281,
"loss": 1.0146,
"step": 1530
},
{
"epoch": 5.238907849829351,
"grad_norm": 0.38671875,
"learning_rate": 0.00016449823070450531,
"loss": 1.0005,
"step": 1535
},
{
"epoch": 5.255972696245734,
"grad_norm": 0.3515625,
"learning_rate": 0.00016419420860745699,
"loss": 1.0093,
"step": 1540
},
{
"epoch": 5.273037542662116,
"grad_norm": 0.439453125,
"learning_rate": 0.00016388917390077054,
"loss": 0.9965,
"step": 1545
},
{
"epoch": 5.290102389078498,
"grad_norm": 0.466796875,
"learning_rate": 0.00016358313139611195,
"loss": 1.0153,
"step": 1550
},
{
"epoch": 5.30716723549488,
"grad_norm": 0.376953125,
"learning_rate": 0.0001632760859210442,
"loss": 1.0094,
"step": 1555
},
{
"epoch": 5.324232081911263,
"grad_norm": 0.5234375,
"learning_rate": 0.00016296804231895142,
"loss": 0.9984,
"step": 1560
},
{
"epoch": 5.341296928327645,
"grad_norm": 0.37109375,
"learning_rate": 0.00016265900544896225,
"loss": 1.0066,
"step": 1565
},
{
"epoch": 5.3583617747440275,
"grad_norm": 0.470703125,
"learning_rate": 0.00016234898018587337,
"loss": 1.0027,
"step": 1570
},
{
"epoch": 5.375426621160409,
"grad_norm": 0.470703125,
"learning_rate": 0.0001620379714200725,
"loss": 1.014,
"step": 1575
},
{
"epoch": 5.392491467576792,
"grad_norm": 0.39453125,
"learning_rate": 0.00016172598405746124,
"loss": 1.0085,
"step": 1580
},
{
"epoch": 5.409556313993174,
"grad_norm": 0.51171875,
"learning_rate": 0.00016141302301937786,
"loss": 0.9999,
"step": 1585
},
{
"epoch": 5.426621160409557,
"grad_norm": 0.54296875,
"learning_rate": 0.0001610990932425194,
"loss": 1.0199,
"step": 1590
},
{
"epoch": 5.4436860068259385,
"grad_norm": 0.447265625,
"learning_rate": 0.00016078419967886402,
"loss": 1.0137,
"step": 1595
},
{
"epoch": 5.460750853242321,
"grad_norm": 0.408203125,
"learning_rate": 0.0001604683472955928,
"loss": 1.0057,
"step": 1600
},
{
"epoch": 5.477815699658703,
"grad_norm": 0.419921875,
"learning_rate": 0.00016015154107501133,
"loss": 1.0099,
"step": 1605
},
{
"epoch": 5.494880546075085,
"grad_norm": 0.455078125,
"learning_rate": 0.00015983378601447127,
"loss": 1.0066,
"step": 1610
},
{
"epoch": 5.511945392491468,
"grad_norm": 0.412109375,
"learning_rate": 0.0001595150871262914,
"loss": 1.0134,
"step": 1615
},
{
"epoch": 5.5290102389078495,
"grad_norm": 0.37890625,
"learning_rate": 0.00015919544943767856,
"loss": 1.0108,
"step": 1620
},
{
"epoch": 5.546075085324232,
"grad_norm": 0.40625,
"learning_rate": 0.00015887487799064838,
"loss": 1.0229,
"step": 1625
},
{
"epoch": 5.563139931740614,
"grad_norm": 0.56640625,
"learning_rate": 0.00015855337784194577,
"loss": 1.0126,
"step": 1630
},
{
"epoch": 5.580204778156997,
"grad_norm": 0.37890625,
"learning_rate": 0.00015823095406296514,
"loss": 0.9947,
"step": 1635
},
{
"epoch": 5.597269624573379,
"grad_norm": 0.373046875,
"learning_rate": 0.00015790761173967036,
"loss": 1.0063,
"step": 1640
},
{
"epoch": 5.614334470989761,
"grad_norm": 0.416015625,
"learning_rate": 0.00015758335597251458,
"loss": 1.0132,
"step": 1645
},
{
"epoch": 5.631399317406143,
"grad_norm": 0.4375,
"learning_rate": 0.00015725819187635968,
"loss": 1.0173,
"step": 1650
},
{
"epoch": 5.648464163822526,
"grad_norm": 0.388671875,
"learning_rate": 0.00015693212458039584,
"loss": 1.0115,
"step": 1655
},
{
"epoch": 5.665529010238908,
"grad_norm": 0.42578125,
"learning_rate": 0.00015660515922806027,
"loss": 0.9966,
"step": 1660
},
{
"epoch": 5.6825938566552905,
"grad_norm": 0.349609375,
"learning_rate": 0.00015627730097695638,
"loss": 1.0058,
"step": 1665
},
{
"epoch": 5.699658703071672,
"grad_norm": 0.427734375,
"learning_rate": 0.0001559485549987723,
"loss": 1.0143,
"step": 1670
},
{
"epoch": 5.716723549488055,
"grad_norm": 0.384765625,
"learning_rate": 0.0001556189264791992,
"loss": 1.0124,
"step": 1675
},
{
"epoch": 5.733788395904437,
"grad_norm": 0.40625,
"learning_rate": 0.0001552884206178498,
"loss": 1.0119,
"step": 1680
},
{
"epoch": 5.750853242320819,
"grad_norm": 0.412109375,
"learning_rate": 0.00015495704262817597,
"loss": 1.0061,
"step": 1685
},
{
"epoch": 5.7679180887372015,
"grad_norm": 0.3828125,
"learning_rate": 0.0001546247977373867,
"loss": 1.0054,
"step": 1690
},
{
"epoch": 5.784982935153583,
"grad_norm": 0.37109375,
"learning_rate": 0.00015429169118636566,
"loss": 1.0021,
"step": 1695
},
{
"epoch": 5.802047781569966,
"grad_norm": 0.392578125,
"learning_rate": 0.00015395772822958845,
"loss": 1.0037,
"step": 1700
},
{
"epoch": 5.819112627986348,
"grad_norm": 0.408203125,
"learning_rate": 0.00015362291413503984,
"loss": 1.0054,
"step": 1705
},
{
"epoch": 5.836177474402731,
"grad_norm": 0.345703125,
"learning_rate": 0.00015328725418413045,
"loss": 1.0132,
"step": 1710
},
{
"epoch": 5.853242320819112,
"grad_norm": 0.341796875,
"learning_rate": 0.00015295075367161367,
"loss": 1.0041,
"step": 1715
},
{
"epoch": 5.870307167235495,
"grad_norm": 0.34375,
"learning_rate": 0.00015261341790550196,
"loss": 1.001,
"step": 1720
},
{
"epoch": 5.887372013651877,
"grad_norm": 0.373046875,
"learning_rate": 0.0001522752522069833,
"loss": 1.0102,
"step": 1725
},
{
"epoch": 5.90443686006826,
"grad_norm": 0.38671875,
"learning_rate": 0.00015193626191033712,
"loss": 0.996,
"step": 1730
},
{
"epoch": 5.921501706484642,
"grad_norm": 0.37109375,
"learning_rate": 0.0001515964523628501,
"loss": 1.0052,
"step": 1735
},
{
"epoch": 5.938566552901024,
"grad_norm": 0.3984375,
"learning_rate": 0.00015125582892473204,
"loss": 1.0118,
"step": 1740
},
{
"epoch": 5.955631399317406,
"grad_norm": 0.392578125,
"learning_rate": 0.00015091439696903115,
"loss": 0.998,
"step": 1745
},
{
"epoch": 5.972696245733788,
"grad_norm": 0.388671875,
"learning_rate": 0.00015057216188154928,
"loss": 0.9925,
"step": 1750
},
{
"epoch": 5.989761092150171,
"grad_norm": 0.5,
"learning_rate": 0.00015022912906075702,
"loss": 0.993,
"step": 1755
},
{
"epoch": 6.0,
"eval_loss": 2.4702811241149902,
"eval_runtime": 0.5473,
"eval_samples_per_second": 18.272,
"eval_steps_per_second": 1.827,
"step": 1758
},
{
"epoch": 6.006825938566553,
"grad_norm": 0.48046875,
"learning_rate": 0.00014988530391770856,
"loss": 0.9939,
"step": 1760
},
{
"epoch": 6.023890784982935,
"grad_norm": 0.396484375,
"learning_rate": 0.00014954069187595633,
"loss": 0.9904,
"step": 1765
},
{
"epoch": 6.040955631399317,
"grad_norm": 0.396484375,
"learning_rate": 0.00014919529837146528,
"loss": 0.982,
"step": 1770
},
{
"epoch": 6.0580204778157,
"grad_norm": 0.421875,
"learning_rate": 0.0001488491288525275,
"loss": 0.9741,
"step": 1775
},
{
"epoch": 6.075085324232082,
"grad_norm": 0.439453125,
"learning_rate": 0.0001485021887796759,
"loss": 0.995,
"step": 1780
},
{
"epoch": 6.092150170648464,
"grad_norm": 0.39453125,
"learning_rate": 0.00014815448362559826,
"loss": 0.9931,
"step": 1785
},
{
"epoch": 6.109215017064846,
"grad_norm": 0.373046875,
"learning_rate": 0.00014780601887505088,
"loss": 1.0001,
"step": 1790
},
{
"epoch": 6.126279863481229,
"grad_norm": 0.384765625,
"learning_rate": 0.00014745680002477203,
"loss": 0.9913,
"step": 1795
},
{
"epoch": 6.143344709897611,
"grad_norm": 0.36328125,
"learning_rate": 0.00014710683258339536,
"loss": 0.9883,
"step": 1800
},
{
"epoch": 6.160409556313994,
"grad_norm": 0.427734375,
"learning_rate": 0.0001467561220713628,
"loss": 0.9835,
"step": 1805
},
{
"epoch": 6.177474402730375,
"grad_norm": 0.421875,
"learning_rate": 0.0001464046740208377,
"loss": 0.9894,
"step": 1810
},
{
"epoch": 6.194539249146757,
"grad_norm": 0.357421875,
"learning_rate": 0.00014605249397561736,
"loss": 0.9833,
"step": 1815
},
{
"epoch": 6.21160409556314,
"grad_norm": 0.4140625,
"learning_rate": 0.00014569958749104575,
"loss": 0.9942,
"step": 1820
},
{
"epoch": 6.228668941979522,
"grad_norm": 0.58203125,
"learning_rate": 0.00014534596013392575,
"loss": 0.9937,
"step": 1825
},
{
"epoch": 6.2457337883959045,
"grad_norm": 0.6328125,
"learning_rate": 0.00014499161748243147,
"loss": 0.9852,
"step": 1830
},
{
"epoch": 6.262798634812286,
"grad_norm": 0.6640625,
"learning_rate": 0.0001446365651260201,
"loss": 0.9886,
"step": 1835
},
{
"epoch": 6.279863481228669,
"grad_norm": 0.4375,
"learning_rate": 0.00014428080866534396,
"loss": 0.9893,
"step": 1840
},
{
"epoch": 6.296928327645051,
"grad_norm": 0.376953125,
"learning_rate": 0.00014392435371216185,
"loss": 0.9951,
"step": 1845
},
{
"epoch": 6.313993174061434,
"grad_norm": 0.34765625,
"learning_rate": 0.0001435672058892509,
"loss": 0.9877,
"step": 1850
},
{
"epoch": 6.3310580204778155,
"grad_norm": 0.390625,
"learning_rate": 0.00014320937083031748,
"loss": 0.9922,
"step": 1855
},
{
"epoch": 6.348122866894198,
"grad_norm": 0.38671875,
"learning_rate": 0.0001428508541799086,
"loss": 0.9939,
"step": 1860
},
{
"epoch": 6.36518771331058,
"grad_norm": 0.3828125,
"learning_rate": 0.0001424916615933229,
"loss": 0.994,
"step": 1865
},
{
"epoch": 6.382252559726963,
"grad_norm": 0.404296875,
"learning_rate": 0.00014213179873652127,
"loss": 0.993,
"step": 1870
},
{
"epoch": 6.399317406143345,
"grad_norm": 0.408203125,
"learning_rate": 0.00014177127128603745,
"loss": 0.9982,
"step": 1875
},
{
"epoch": 6.4163822525597265,
"grad_norm": 0.40625,
"learning_rate": 0.0001414100849288888,
"loss": 0.9926,
"step": 1880
},
{
"epoch": 6.433447098976109,
"grad_norm": 0.416015625,
"learning_rate": 0.00014104824536248614,
"loss": 0.995,
"step": 1885
},
{
"epoch": 6.450511945392491,
"grad_norm": 0.40625,
"learning_rate": 0.00014068575829454436,
"loss": 0.9894,
"step": 1890
},
{
"epoch": 6.467576791808874,
"grad_norm": 0.359375,
"learning_rate": 0.00014032262944299194,
"loss": 0.997,
"step": 1895
},
{
"epoch": 6.484641638225256,
"grad_norm": 0.392578125,
"learning_rate": 0.00013995886453588104,
"loss": 0.9861,
"step": 1900
},
{
"epoch": 6.501706484641638,
"grad_norm": 0.34765625,
"learning_rate": 0.00013959446931129704,
"loss": 0.9896,
"step": 1905
},
{
"epoch": 6.51877133105802,
"grad_norm": 0.380859375,
"learning_rate": 0.0001392294495172681,
"loss": 0.9969,
"step": 1910
},
{
"epoch": 6.535836177474403,
"grad_norm": 0.392578125,
"learning_rate": 0.0001388638109116744,
"loss": 0.9902,
"step": 1915
},
{
"epoch": 6.552901023890785,
"grad_norm": 0.375,
"learning_rate": 0.00013849755926215735,
"loss": 0.9995,
"step": 1920
},
{
"epoch": 6.5699658703071675,
"grad_norm": 0.384765625,
"learning_rate": 0.00013813070034602863,
"loss": 0.9935,
"step": 1925
},
{
"epoch": 6.587030716723549,
"grad_norm": 0.466796875,
"learning_rate": 0.00013776323995017898,
"loss": 0.9799,
"step": 1930
},
{
"epoch": 6.604095563139932,
"grad_norm": 0.345703125,
"learning_rate": 0.00013739518387098705,
"loss": 0.9959,
"step": 1935
},
{
"epoch": 6.621160409556314,
"grad_norm": 0.388671875,
"learning_rate": 0.0001370265379142279,
"loss": 0.9897,
"step": 1940
},
{
"epoch": 6.638225255972696,
"grad_norm": 0.443359375,
"learning_rate": 0.0001366573078949813,
"loss": 0.9829,
"step": 1945
},
{
"epoch": 6.6552901023890785,
"grad_norm": 0.44921875,
"learning_rate": 0.00013628749963754026,
"loss": 0.9963,
"step": 1950
},
{
"epoch": 6.672354948805461,
"grad_norm": 0.52734375,
"learning_rate": 0.0001359171189753189,
"loss": 0.999,
"step": 1955
},
{
"epoch": 6.689419795221843,
"grad_norm": 0.6484375,
"learning_rate": 0.00013554617175076062,
"loss": 0.9806,
"step": 1960
},
{
"epoch": 6.706484641638225,
"grad_norm": 0.388671875,
"learning_rate": 0.0001351746638152458,
"loss": 0.9903,
"step": 1965
},
{
"epoch": 6.723549488054608,
"grad_norm": 0.4765625,
"learning_rate": 0.00013480260102899966,
"loss": 1.0009,
"step": 1970
},
{
"epoch": 6.7406143344709895,
"grad_norm": 0.44140625,
"learning_rate": 0.0001344299892609996,
"loss": 0.9879,
"step": 1975
},
{
"epoch": 6.757679180887372,
"grad_norm": 0.392578125,
"learning_rate": 0.00013405683438888282,
"loss": 0.9966,
"step": 1980
},
{
"epoch": 6.774744027303754,
"grad_norm": 0.4140625,
"learning_rate": 0.00013368314229885347,
"loss": 0.988,
"step": 1985
},
{
"epoch": 6.791808873720137,
"grad_norm": 0.36328125,
"learning_rate": 0.00013330891888559002,
"loss": 0.9835,
"step": 1990
},
{
"epoch": 6.808873720136519,
"grad_norm": 0.421875,
"learning_rate": 0.00013293417005215188,
"loss": 0.9922,
"step": 1995
},
{
"epoch": 6.825938566552901,
"grad_norm": 0.40234375,
"learning_rate": 0.0001325589017098867,
"loss": 0.9893,
"step": 2000
},
{
"epoch": 6.843003412969283,
"grad_norm": 0.3828125,
"learning_rate": 0.00013218311977833687,
"loss": 0.9965,
"step": 2005
},
{
"epoch": 6.860068259385666,
"grad_norm": 0.365234375,
"learning_rate": 0.0001318068301851463,
"loss": 0.9843,
"step": 2010
},
{
"epoch": 6.877133105802048,
"grad_norm": 0.390625,
"learning_rate": 0.00013143003886596669,
"loss": 0.9845,
"step": 2015
},
{
"epoch": 6.8941979522184305,
"grad_norm": 0.3515625,
"learning_rate": 0.0001310527517643642,
"loss": 0.9909,
"step": 2020
},
{
"epoch": 6.911262798634812,
"grad_norm": 0.359375,
"learning_rate": 0.00013067497483172538,
"loss": 0.9885,
"step": 2025
},
{
"epoch": 6.928327645051194,
"grad_norm": 0.375,
"learning_rate": 0.00013029671402716366,
"loss": 0.9879,
"step": 2030
},
{
"epoch": 6.945392491467577,
"grad_norm": 0.380859375,
"learning_rate": 0.00012991797531742492,
"loss": 0.9891,
"step": 2035
},
{
"epoch": 6.962457337883959,
"grad_norm": 0.34375,
"learning_rate": 0.00012953876467679373,
"loss": 0.9972,
"step": 2040
},
{
"epoch": 6.979522184300341,
"grad_norm": 0.369140625,
"learning_rate": 0.00012915908808699893,
"loss": 0.9962,
"step": 2045
},
{
"epoch": 6.996587030716723,
"grad_norm": 0.44921875,
"learning_rate": 0.00012877895153711935,
"loss": 0.9941,
"step": 2050
},
{
"epoch": 7.0,
"eval_loss": 2.49063777923584,
"eval_runtime": 0.554,
"eval_samples_per_second": 18.051,
"eval_steps_per_second": 1.805,
"step": 2051
},
{
"epoch": 7.013651877133106,
"grad_norm": 0.52734375,
"learning_rate": 0.00012839836102348926,
"loss": 0.9759,
"step": 2055
},
{
"epoch": 7.030716723549488,
"grad_norm": 0.365234375,
"learning_rate": 0.00012801732254960388,
"loss": 0.9703,
"step": 2060
},
{
"epoch": 7.047781569965871,
"grad_norm": 0.375,
"learning_rate": 0.00012763584212602453,
"loss": 0.9643,
"step": 2065
},
{
"epoch": 7.064846416382252,
"grad_norm": 0.41796875,
"learning_rate": 0.00012725392577028402,
"loss": 0.9646,
"step": 2070
},
{
"epoch": 7.081911262798635,
"grad_norm": 0.400390625,
"learning_rate": 0.0001268715795067916,
"loss": 0.9732,
"step": 2075
},
{
"epoch": 7.098976109215017,
"grad_norm": 0.380859375,
"learning_rate": 0.00012648880936673787,
"loss": 0.9786,
"step": 2080
},
{
"epoch": 7.1160409556314,
"grad_norm": 0.423828125,
"learning_rate": 0.00012610562138799978,
"loss": 0.9733,
"step": 2085
},
{
"epoch": 7.1331058020477816,
"grad_norm": 0.357421875,
"learning_rate": 0.00012572202161504543,
"loss": 0.9808,
"step": 2090
},
{
"epoch": 7.150170648464163,
"grad_norm": 0.4609375,
"learning_rate": 0.00012533801609883842,
"loss": 0.9762,
"step": 2095
},
{
"epoch": 7.167235494880546,
"grad_norm": 0.38671875,
"learning_rate": 0.00012495361089674285,
"loss": 0.9809,
"step": 2100
},
{
"epoch": 7.184300341296928,
"grad_norm": 0.3984375,
"learning_rate": 0.00012456881207242732,
"loss": 0.9821,
"step": 2105
},
{
"epoch": 7.201365187713311,
"grad_norm": 0.400390625,
"learning_rate": 0.00012418362569576965,
"loss": 0.9873,
"step": 2110
},
{
"epoch": 7.2184300341296925,
"grad_norm": 0.55078125,
"learning_rate": 0.00012379805784276082,
"loss": 0.9727,
"step": 2115
},
{
"epoch": 7.235494880546075,
"grad_norm": 0.515625,
"learning_rate": 0.0001234121145954094,
"loss": 0.9827,
"step": 2120
},
{
"epoch": 7.252559726962457,
"grad_norm": 0.3828125,
"learning_rate": 0.00012302580204164541,
"loss": 0.9846,
"step": 2125
},
{
"epoch": 7.26962457337884,
"grad_norm": 0.42578125,
"learning_rate": 0.0001226391262752245,
"loss": 0.9736,
"step": 2130
},
{
"epoch": 7.286689419795222,
"grad_norm": 0.5078125,
"learning_rate": 0.00012225209339563145,
"loss": 0.9743,
"step": 2135
},
{
"epoch": 7.303754266211604,
"grad_norm": 0.419921875,
"learning_rate": 0.00012186470950798445,
"loss": 0.9787,
"step": 2140
},
{
"epoch": 7.320819112627986,
"grad_norm": 0.490234375,
"learning_rate": 0.00012147698072293842,
"loss": 0.9788,
"step": 2145
},
{
"epoch": 7.337883959044369,
"grad_norm": 0.380859375,
"learning_rate": 0.00012108891315658879,
"loss": 0.967,
"step": 2150
},
{
"epoch": 7.354948805460751,
"grad_norm": 0.396484375,
"learning_rate": 0.00012070051293037492,
"loss": 0.9792,
"step": 2155
},
{
"epoch": 7.372013651877133,
"grad_norm": 0.43359375,
"learning_rate": 0.00012031178617098371,
"loss": 0.9905,
"step": 2160
},
{
"epoch": 7.389078498293515,
"grad_norm": 0.400390625,
"learning_rate": 0.00011992273901025269,
"loss": 0.9873,
"step": 2165
},
{
"epoch": 7.406143344709897,
"grad_norm": 0.4453125,
"learning_rate": 0.0001195333775850736,
"loss": 0.9872,
"step": 2170
},
{
"epoch": 7.42320819112628,
"grad_norm": 0.44140625,
"learning_rate": 0.00011914370803729533,
"loss": 0.98,
"step": 2175
},
{
"epoch": 7.440273037542662,
"grad_norm": 0.361328125,
"learning_rate": 0.00011875373651362727,
"loss": 0.9827,
"step": 2180
},
{
"epoch": 7.4573378839590445,
"grad_norm": 0.474609375,
"learning_rate": 0.00011836346916554205,
"loss": 0.9738,
"step": 2185
},
{
"epoch": 7.474402730375426,
"grad_norm": 0.421875,
"learning_rate": 0.00011797291214917881,
"loss": 0.9762,
"step": 2190
},
{
"epoch": 7.491467576791809,
"grad_norm": 0.41796875,
"learning_rate": 0.00011758207162524598,
"loss": 0.9675,
"step": 2195
},
{
"epoch": 7.508532423208191,
"grad_norm": 0.384765625,
"learning_rate": 0.00011719095375892396,
"loss": 0.9923,
"step": 2200
},
{
"epoch": 7.525597269624574,
"grad_norm": 0.482421875,
"learning_rate": 0.00011679956471976814,
"loss": 0.9818,
"step": 2205
},
{
"epoch": 7.5426621160409555,
"grad_norm": 0.3671875,
"learning_rate": 0.0001164079106816113,
"loss": 0.9783,
"step": 2210
},
{
"epoch": 7.559726962457338,
"grad_norm": 0.376953125,
"learning_rate": 0.00011601599782246646,
"loss": 0.9735,
"step": 2215
},
{
"epoch": 7.57679180887372,
"grad_norm": 0.443359375,
"learning_rate": 0.00011562383232442926,
"loss": 0.9751,
"step": 2220
},
{
"epoch": 7.593856655290102,
"grad_norm": 0.3671875,
"learning_rate": 0.0001152314203735805,
"loss": 0.9734,
"step": 2225
},
{
"epoch": 7.610921501706485,
"grad_norm": 0.439453125,
"learning_rate": 0.00011483876815988867,
"loss": 0.9706,
"step": 2230
},
{
"epoch": 7.627986348122867,
"grad_norm": 0.44140625,
"learning_rate": 0.00011444588187711205,
"loss": 0.9727,
"step": 2235
},
{
"epoch": 7.645051194539249,
"grad_norm": 0.41796875,
"learning_rate": 0.00011405276772270126,
"loss": 0.9774,
"step": 2240
},
{
"epoch": 7.662116040955631,
"grad_norm": 0.353515625,
"learning_rate": 0.0001136594318977014,
"loss": 0.9815,
"step": 2245
},
{
"epoch": 7.679180887372014,
"grad_norm": 0.412109375,
"learning_rate": 0.0001132658806066542,
"loss": 0.9835,
"step": 2250
},
{
"epoch": 7.696245733788396,
"grad_norm": 0.384765625,
"learning_rate": 0.00011287212005750024,
"loss": 0.9773,
"step": 2255
},
{
"epoch": 7.713310580204778,
"grad_norm": 0.42578125,
"learning_rate": 0.00011247815646148087,
"loss": 0.9835,
"step": 2260
},
{
"epoch": 7.73037542662116,
"grad_norm": 0.56640625,
"learning_rate": 0.00011208399603304047,
"loss": 0.9832,
"step": 2265
},
{
"epoch": 7.747440273037543,
"grad_norm": 0.38671875,
"learning_rate": 0.00011168964498972818,
"loss": 0.9701,
"step": 2270
},
{
"epoch": 7.764505119453925,
"grad_norm": 0.3671875,
"learning_rate": 0.00011129510955209996,
"loss": 0.9832,
"step": 2275
},
{
"epoch": 7.7815699658703075,
"grad_norm": 0.546875,
"learning_rate": 0.00011090039594362045,
"loss": 0.9861,
"step": 2280
},
{
"epoch": 7.798634812286689,
"grad_norm": 0.5078125,
"learning_rate": 0.00011050551039056479,
"loss": 0.9881,
"step": 2285
},
{
"epoch": 7.815699658703072,
"grad_norm": 0.375,
"learning_rate": 0.00011011045912192035,
"loss": 0.9872,
"step": 2290
},
{
"epoch": 7.832764505119454,
"grad_norm": 0.373046875,
"learning_rate": 0.0001097152483692886,
"loss": 0.9819,
"step": 2295
},
{
"epoch": 7.849829351535837,
"grad_norm": 0.375,
"learning_rate": 0.00010931988436678666,
"loss": 0.9756,
"step": 2300
},
{
"epoch": 7.8668941979522184,
"grad_norm": 0.40234375,
"learning_rate": 0.00010892437335094912,
"loss": 0.9662,
"step": 2305
},
{
"epoch": 7.8839590443686,
"grad_norm": 0.427734375,
"learning_rate": 0.00010852872156062946,
"loss": 0.9669,
"step": 2310
},
{
"epoch": 7.901023890784983,
"grad_norm": 0.388671875,
"learning_rate": 0.00010813293523690191,
"loss": 0.9755,
"step": 2315
},
{
"epoch": 7.918088737201365,
"grad_norm": 0.423828125,
"learning_rate": 0.00010773702062296273,
"loss": 0.9916,
"step": 2320
},
{
"epoch": 7.935153583617748,
"grad_norm": 0.396484375,
"learning_rate": 0.00010734098396403192,
"loss": 0.9869,
"step": 2325
},
{
"epoch": 7.952218430034129,
"grad_norm": 0.447265625,
"learning_rate": 0.00010694483150725458,
"loss": 0.978,
"step": 2330
},
{
"epoch": 7.969283276450512,
"grad_norm": 0.451171875,
"learning_rate": 0.00010654856950160253,
"loss": 0.9711,
"step": 2335
},
{
"epoch": 7.986348122866894,
"grad_norm": 0.392578125,
"learning_rate": 0.00010615220419777548,
"loss": 0.9844,
"step": 2340
},
{
"epoch": 8.0,
"eval_loss": 2.489572525024414,
"eval_runtime": 0.5472,
"eval_samples_per_second": 18.276,
"eval_steps_per_second": 1.828,
"step": 2344
},
{
"epoch": 8.003412969283277,
"grad_norm": 0.44140625,
"learning_rate": 0.00010575574184810269,
"loss": 0.9713,
"step": 2345
},
{
"epoch": 8.020477815699659,
"grad_norm": 0.3984375,
"learning_rate": 0.0001053591887064442,
"loss": 0.9647,
"step": 2350
},
{
"epoch": 8.03754266211604,
"grad_norm": 0.453125,
"learning_rate": 0.00010496255102809223,
"loss": 0.9709,
"step": 2355
},
{
"epoch": 8.054607508532424,
"grad_norm": 0.431640625,
"learning_rate": 0.00010456583506967248,
"loss": 0.9701,
"step": 2360
},
{
"epoch": 8.071672354948806,
"grad_norm": 0.44921875,
"learning_rate": 0.00010416904708904548,
"loss": 0.9662,
"step": 2365
},
{
"epoch": 8.088737201365188,
"grad_norm": 0.46875,
"learning_rate": 0.00010377219334520783,
"loss": 0.9616,
"step": 2370
},
{
"epoch": 8.10580204778157,
"grad_norm": 0.4140625,
"learning_rate": 0.00010337528009819344,
"loss": 0.9609,
"step": 2375
},
{
"epoch": 8.122866894197951,
"grad_norm": 0.42578125,
"learning_rate": 0.00010297831360897492,
"loss": 0.9714,
"step": 2380
},
{
"epoch": 8.139931740614335,
"grad_norm": 0.40234375,
"learning_rate": 0.00010258130013936474,
"loss": 0.9718,
"step": 2385
},
{
"epoch": 8.156996587030717,
"grad_norm": 0.4296875,
"learning_rate": 0.00010218424595191631,
"loss": 0.963,
"step": 2390
},
{
"epoch": 8.174061433447099,
"grad_norm": 0.361328125,
"learning_rate": 0.00010178715730982549,
"loss": 0.9612,
"step": 2395
},
{
"epoch": 8.19112627986348,
"grad_norm": 0.451171875,
"learning_rate": 0.00010139004047683151,
"loss": 0.9757,
"step": 2400
},
{
"epoch": 8.208191126279864,
"grad_norm": 0.62890625,
"learning_rate": 0.00010099290171711841,
"loss": 0.961,
"step": 2405
},
{
"epoch": 8.225255972696246,
"grad_norm": 0.419921875,
"learning_rate": 0.00010059574729521595,
"loss": 0.962,
"step": 2410
},
{
"epoch": 8.242320819112628,
"grad_norm": 0.51171875,
"learning_rate": 0.0001001985834759011,
"loss": 0.9761,
"step": 2415
},
{
"epoch": 8.25938566552901,
"grad_norm": 0.390625,
"learning_rate": 9.980141652409895e-05,
"loss": 0.9718,
"step": 2420
},
{
"epoch": 8.276450511945393,
"grad_norm": 0.41796875,
"learning_rate": 9.940425270478407e-05,
"loss": 0.9672,
"step": 2425
},
{
"epoch": 8.293515358361775,
"grad_norm": 0.431640625,
"learning_rate": 9.900709828288164e-05,
"loss": 0.9658,
"step": 2430
},
{
"epoch": 8.310580204778157,
"grad_norm": 0.4140625,
"learning_rate": 9.860995952316851e-05,
"loss": 0.9776,
"step": 2435
},
{
"epoch": 8.327645051194539,
"grad_norm": 0.37890625,
"learning_rate": 9.821284269017455e-05,
"loss": 0.9664,
"step": 2440
},
{
"epoch": 8.344709897610922,
"grad_norm": 0.380859375,
"learning_rate": 9.781575404808371e-05,
"loss": 0.9672,
"step": 2445
},
{
"epoch": 8.361774744027304,
"grad_norm": 0.3828125,
"learning_rate": 9.741869986063526e-05,
"loss": 0.9778,
"step": 2450
},
{
"epoch": 8.378839590443686,
"grad_norm": 0.361328125,
"learning_rate": 9.702168639102509e-05,
"loss": 0.9659,
"step": 2455
},
{
"epoch": 8.395904436860068,
"grad_norm": 0.392578125,
"learning_rate": 9.662471990180657e-05,
"loss": 0.9623,
"step": 2460
},
{
"epoch": 8.41296928327645,
"grad_norm": 0.365234375,
"learning_rate": 9.622780665479222e-05,
"loss": 0.9657,
"step": 2465
},
{
"epoch": 8.430034129692833,
"grad_norm": 0.40234375,
"learning_rate": 9.583095291095453e-05,
"loss": 0.9679,
"step": 2470
},
{
"epoch": 8.447098976109215,
"grad_norm": 0.443359375,
"learning_rate": 9.543416493032757e-05,
"loss": 0.9686,
"step": 2475
},
{
"epoch": 8.464163822525597,
"grad_norm": 0.404296875,
"learning_rate": 9.503744897190778e-05,
"loss": 0.9679,
"step": 2480
},
{
"epoch": 8.481228668941979,
"grad_norm": 0.396484375,
"learning_rate": 9.464081129355586e-05,
"loss": 0.9588,
"step": 2485
},
{
"epoch": 8.498293515358363,
"grad_norm": 0.431640625,
"learning_rate": 9.424425815189733e-05,
"loss": 0.9775,
"step": 2490
},
{
"epoch": 8.515358361774744,
"grad_norm": 0.384765625,
"learning_rate": 9.384779580222453e-05,
"loss": 0.9668,
"step": 2495
},
{
"epoch": 8.532423208191126,
"grad_norm": 0.447265625,
"learning_rate": 9.345143049839749e-05,
"loss": 0.9677,
"step": 2500
},
{
"epoch": 8.549488054607508,
"grad_norm": 0.48828125,
"learning_rate": 9.305516849274541e-05,
"loss": 0.9603,
"step": 2505
},
{
"epoch": 8.56655290102389,
"grad_norm": 0.427734375,
"learning_rate": 9.265901603596811e-05,
"loss": 0.9688,
"step": 2510
},
{
"epoch": 8.583617747440274,
"grad_norm": 0.498046875,
"learning_rate": 9.226297937703728e-05,
"loss": 0.9645,
"step": 2515
},
{
"epoch": 8.600682593856655,
"grad_norm": 0.431640625,
"learning_rate": 9.186706476309812e-05,
"loss": 0.967,
"step": 2520
},
{
"epoch": 8.617747440273037,
"grad_norm": 0.423828125,
"learning_rate": 9.147127843937055e-05,
"loss": 0.9711,
"step": 2525
},
{
"epoch": 8.634812286689419,
"grad_norm": 0.455078125,
"learning_rate": 9.107562664905093e-05,
"loss": 0.971,
"step": 2530
},
{
"epoch": 8.651877133105803,
"grad_norm": 0.484375,
"learning_rate": 9.068011563321336e-05,
"loss": 0.9722,
"step": 2535
},
{
"epoch": 8.668941979522184,
"grad_norm": 0.435546875,
"learning_rate": 9.028475163071141e-05,
"loss": 0.9747,
"step": 2540
},
{
"epoch": 8.686006825938566,
"grad_norm": 0.4140625,
"learning_rate": 8.988954087807968e-05,
"loss": 0.9638,
"step": 2545
},
{
"epoch": 8.703071672354948,
"grad_norm": 0.400390625,
"learning_rate": 8.949448960943524e-05,
"loss": 0.9625,
"step": 2550
},
{
"epoch": 8.720136518771332,
"grad_norm": 0.49609375,
"learning_rate": 8.909960405637958e-05,
"loss": 0.9568,
"step": 2555
},
{
"epoch": 8.737201365187714,
"grad_norm": 0.435546875,
"learning_rate": 8.870489044790006e-05,
"loss": 0.9766,
"step": 2560
},
{
"epoch": 8.754266211604095,
"grad_norm": 0.41015625,
"learning_rate": 8.831035501027186e-05,
"loss": 0.967,
"step": 2565
},
{
"epoch": 8.771331058020477,
"grad_norm": 0.376953125,
"learning_rate": 8.791600396695954e-05,
"loss": 0.9686,
"step": 2570
},
{
"epoch": 8.788395904436861,
"grad_norm": 0.373046875,
"learning_rate": 8.752184353851916e-05,
"loss": 0.9684,
"step": 2575
},
{
"epoch": 8.805460750853243,
"grad_norm": 0.435546875,
"learning_rate": 8.712787994249979e-05,
"loss": 0.977,
"step": 2580
},
{
"epoch": 8.822525597269625,
"grad_norm": 0.419921875,
"learning_rate": 8.673411939334581e-05,
"loss": 0.9712,
"step": 2585
},
{
"epoch": 8.839590443686006,
"grad_norm": 0.478515625,
"learning_rate": 8.634056810229862e-05,
"loss": 0.9692,
"step": 2590
},
{
"epoch": 8.856655290102388,
"grad_norm": 0.404296875,
"learning_rate": 8.594723227729875e-05,
"loss": 0.9639,
"step": 2595
},
{
"epoch": 8.873720136518772,
"grad_norm": 0.447265625,
"learning_rate": 8.555411812288798e-05,
"loss": 0.974,
"step": 2600
},
{
"epoch": 8.890784982935154,
"grad_norm": 0.392578125,
"learning_rate": 8.516123184011135e-05,
"loss": 0.9589,
"step": 2605
},
{
"epoch": 8.907849829351536,
"grad_norm": 0.43359375,
"learning_rate": 8.47685796264195e-05,
"loss": 0.968,
"step": 2610
},
{
"epoch": 8.924914675767917,
"grad_norm": 0.396484375,
"learning_rate": 8.437616767557077e-05,
"loss": 0.9693,
"step": 2615
},
{
"epoch": 8.941979522184301,
"grad_norm": 0.5390625,
"learning_rate": 8.398400217753357e-05,
"loss": 0.9727,
"step": 2620
},
{
"epoch": 8.959044368600683,
"grad_norm": 0.419921875,
"learning_rate": 8.359208931838871e-05,
"loss": 0.9708,
"step": 2625
},
{
"epoch": 8.976109215017065,
"grad_norm": 0.427734375,
"learning_rate": 8.320043528023188e-05,
"loss": 0.9607,
"step": 2630
},
{
"epoch": 8.993174061433447,
"grad_norm": 0.455078125,
"learning_rate": 8.280904624107606e-05,
"loss": 0.9779,
"step": 2635
},
{
"epoch": 9.0,
"eval_loss": 2.502519130706787,
"eval_runtime": 0.5483,
"eval_samples_per_second": 18.238,
"eval_steps_per_second": 1.824,
"step": 2637
},
{
"epoch": 9.01023890784983,
"grad_norm": 0.3828125,
"learning_rate": 8.241792837475405e-05,
"loss": 0.9673,
"step": 2640
},
{
"epoch": 9.027303754266212,
"grad_norm": 0.42578125,
"learning_rate": 8.202708785082121e-05,
"loss": 0.9481,
"step": 2645
},
{
"epoch": 9.044368600682594,
"grad_norm": 0.39453125,
"learning_rate": 8.163653083445799e-05,
"loss": 0.9694,
"step": 2650
},
{
"epoch": 9.061433447098976,
"grad_norm": 0.392578125,
"learning_rate": 8.124626348637279e-05,
"loss": 0.9651,
"step": 2655
},
{
"epoch": 9.078498293515358,
"grad_norm": 0.376953125,
"learning_rate": 8.085629196270469e-05,
"loss": 0.9561,
"step": 2660
},
{
"epoch": 9.095563139931741,
"grad_norm": 0.408203125,
"learning_rate": 8.046662241492645e-05,
"loss": 0.9617,
"step": 2665
},
{
"epoch": 9.112627986348123,
"grad_norm": 0.408203125,
"learning_rate": 8.007726098974734e-05,
"loss": 0.9636,
"step": 2670
},
{
"epoch": 9.129692832764505,
"grad_norm": 0.390625,
"learning_rate": 7.96882138290163e-05,
"loss": 0.9661,
"step": 2675
},
{
"epoch": 9.146757679180887,
"grad_norm": 0.396484375,
"learning_rate": 7.929948706962508e-05,
"loss": 0.9577,
"step": 2680
},
{
"epoch": 9.16382252559727,
"grad_norm": 0.41796875,
"learning_rate": 7.891108684341121e-05,
"loss": 0.961,
"step": 2685
},
{
"epoch": 9.180887372013652,
"grad_norm": 0.37109375,
"learning_rate": 7.852301927706159e-05,
"loss": 0.9602,
"step": 2690
},
{
"epoch": 9.197952218430034,
"grad_norm": 0.396484375,
"learning_rate": 7.813529049201556e-05,
"loss": 0.9544,
"step": 2695
},
{
"epoch": 9.215017064846416,
"grad_norm": 0.470703125,
"learning_rate": 7.774790660436858e-05,
"loss": 0.9569,
"step": 2700
},
{
"epoch": 9.2320819112628,
"grad_norm": 0.375,
"learning_rate": 7.736087372477554e-05,
"loss": 0.9636,
"step": 2705
},
{
"epoch": 9.249146757679181,
"grad_norm": 0.37109375,
"learning_rate": 7.69741979583546e-05,
"loss": 0.9574,
"step": 2710
},
{
"epoch": 9.266211604095563,
"grad_norm": 0.390625,
"learning_rate": 7.658788540459062e-05,
"loss": 0.9536,
"step": 2715
},
{
"epoch": 9.283276450511945,
"grad_norm": 0.388671875,
"learning_rate": 7.620194215723919e-05,
"loss": 0.9598,
"step": 2720
},
{
"epoch": 9.300341296928327,
"grad_norm": 0.3828125,
"learning_rate": 7.581637430423037e-05,
"loss": 0.9657,
"step": 2725
},
{
"epoch": 9.31740614334471,
"grad_norm": 0.435546875,
"learning_rate": 7.543118792757266e-05,
"loss": 0.9639,
"step": 2730
},
{
"epoch": 9.334470989761092,
"grad_norm": 0.408203125,
"learning_rate": 7.504638910325717e-05,
"loss": 0.9625,
"step": 2735
},
{
"epoch": 9.351535836177474,
"grad_norm": 0.37109375,
"learning_rate": 7.466198390116158e-05,
"loss": 0.9585,
"step": 2740
},
{
"epoch": 9.368600682593856,
"grad_norm": 0.447265625,
"learning_rate": 7.427797838495463e-05,
"loss": 0.9634,
"step": 2745
},
{
"epoch": 9.38566552901024,
"grad_norm": 0.41796875,
"learning_rate": 7.389437861200024e-05,
"loss": 0.9624,
"step": 2750
},
{
"epoch": 9.402730375426621,
"grad_norm": 0.408203125,
"learning_rate": 7.35111906332622e-05,
"loss": 0.9555,
"step": 2755
},
{
"epoch": 9.419795221843003,
"grad_norm": 0.435546875,
"learning_rate": 7.312842049320844e-05,
"loss": 0.9575,
"step": 2760
},
{
"epoch": 9.436860068259385,
"grad_norm": 0.42578125,
"learning_rate": 7.2746074229716e-05,
"loss": 0.9598,
"step": 2765
},
{
"epoch": 9.453924914675769,
"grad_norm": 0.423828125,
"learning_rate": 7.236415787397548e-05,
"loss": 0.9594,
"step": 2770
},
{
"epoch": 9.47098976109215,
"grad_norm": 0.408203125,
"learning_rate": 7.198267745039612e-05,
"loss": 0.9571,
"step": 2775
},
{
"epoch": 9.488054607508532,
"grad_norm": 0.41015625,
"learning_rate": 7.160163897651075e-05,
"loss": 0.9582,
"step": 2780
},
{
"epoch": 9.505119453924914,
"grad_norm": 0.453125,
"learning_rate": 7.122104846288064e-05,
"loss": 0.9583,
"step": 2785
},
{
"epoch": 9.522184300341298,
"grad_norm": 0.474609375,
"learning_rate": 7.08409119130011e-05,
"loss": 0.9713,
"step": 2790
},
{
"epoch": 9.53924914675768,
"grad_norm": 0.388671875,
"learning_rate": 7.04612353232063e-05,
"loss": 0.9538,
"step": 2795
},
{
"epoch": 9.556313993174061,
"grad_norm": 0.41796875,
"learning_rate": 7.008202468257514e-05,
"loss": 0.9572,
"step": 2800
},
{
"epoch": 9.573378839590443,
"grad_norm": 0.41015625,
"learning_rate": 6.970328597283637e-05,
"loss": 0.9483,
"step": 2805
},
{
"epoch": 9.590443686006825,
"grad_norm": 0.40234375,
"learning_rate": 6.932502516827461e-05,
"loss": 0.9521,
"step": 2810
},
{
"epoch": 9.607508532423209,
"grad_norm": 0.38671875,
"learning_rate": 6.894724823563583e-05,
"loss": 0.9534,
"step": 2815
},
{
"epoch": 9.62457337883959,
"grad_norm": 0.41015625,
"learning_rate": 6.85699611340333e-05,
"loss": 0.9611,
"step": 2820
},
{
"epoch": 9.641638225255972,
"grad_norm": 0.369140625,
"learning_rate": 6.819316981485372e-05,
"loss": 0.9499,
"step": 2825
},
{
"epoch": 9.658703071672354,
"grad_norm": 0.361328125,
"learning_rate": 6.781688022166311e-05,
"loss": 0.9689,
"step": 2830
},
{
"epoch": 9.675767918088738,
"grad_norm": 0.40234375,
"learning_rate": 6.744109829011332e-05,
"loss": 0.9492,
"step": 2835
},
{
"epoch": 9.69283276450512,
"grad_norm": 0.384765625,
"learning_rate": 6.706582994784814e-05,
"loss": 0.9626,
"step": 2840
},
{
"epoch": 9.709897610921502,
"grad_norm": 0.408203125,
"learning_rate": 6.669108111441003e-05,
"loss": 0.9641,
"step": 2845
},
{
"epoch": 9.726962457337883,
"grad_norm": 0.40234375,
"learning_rate": 6.631685770114654e-05,
"loss": 0.9578,
"step": 2850
},
{
"epoch": 9.744027303754265,
"grad_norm": 0.37890625,
"learning_rate": 6.594316561111724e-05,
"loss": 0.9648,
"step": 2855
},
{
"epoch": 9.761092150170649,
"grad_norm": 0.390625,
"learning_rate": 6.557001073900044e-05,
"loss": 0.957,
"step": 2860
},
{
"epoch": 9.77815699658703,
"grad_norm": 0.375,
"learning_rate": 6.519739897100034e-05,
"loss": 0.9513,
"step": 2865
},
{
"epoch": 9.795221843003413,
"grad_norm": 0.453125,
"learning_rate": 6.482533618475422e-05,
"loss": 0.9591,
"step": 2870
},
{
"epoch": 9.812286689419794,
"grad_norm": 0.369140625,
"learning_rate": 6.445382824923938e-05,
"loss": 0.9625,
"step": 2875
},
{
"epoch": 9.829351535836178,
"grad_norm": 0.37109375,
"learning_rate": 6.408288102468113e-05,
"loss": 0.9606,
"step": 2880
},
{
"epoch": 9.84641638225256,
"grad_norm": 0.37890625,
"learning_rate": 6.371250036245976e-05,
"loss": 0.9662,
"step": 2885
},
{
"epoch": 9.863481228668942,
"grad_norm": 0.373046875,
"learning_rate": 6.334269210501875e-05,
"loss": 0.9635,
"step": 2890
},
{
"epoch": 9.880546075085324,
"grad_norm": 0.365234375,
"learning_rate": 6.297346208577213e-05,
"loss": 0.9649,
"step": 2895
},
{
"epoch": 9.897610921501707,
"grad_norm": 0.390625,
"learning_rate": 6.260481612901299e-05,
"loss": 0.9516,
"step": 2900
},
{
"epoch": 9.914675767918089,
"grad_norm": 0.3828125,
"learning_rate": 6.223676004982105e-05,
"loss": 0.9601,
"step": 2905
},
{
"epoch": 9.93174061433447,
"grad_norm": 0.5625,
"learning_rate": 6.18692996539714e-05,
"loss": 0.9611,
"step": 2910
},
{
"epoch": 9.948805460750853,
"grad_norm": 0.39453125,
"learning_rate": 6.150244073784266e-05,
"loss": 0.9742,
"step": 2915
},
{
"epoch": 9.965870307167236,
"grad_norm": 0.4296875,
"learning_rate": 6.113618908832561e-05,
"loss": 0.9666,
"step": 2920
},
{
"epoch": 9.982935153583618,
"grad_norm": 0.447265625,
"learning_rate": 6.0770550482731924e-05,
"loss": 0.9684,
"step": 2925
},
{
"epoch": 10.0,
"grad_norm": 0.41015625,
"learning_rate": 6.0405530688702986e-05,
"loss": 0.9639,
"step": 2930
},
{
"epoch": 10.0,
"eval_loss": 2.512617588043213,
"eval_runtime": 0.5446,
"eval_samples_per_second": 18.362,
"eval_steps_per_second": 1.836,
"step": 2930
},
{
"epoch": 10.017064846416382,
"grad_norm": 0.427734375,
"learning_rate": 6.0041135464119024e-05,
"loss": 0.9618,
"step": 2935
},
{
"epoch": 10.034129692832764,
"grad_norm": 0.384765625,
"learning_rate": 5.9677370557008104e-05,
"loss": 0.9433,
"step": 2940
},
{
"epoch": 10.051194539249147,
"grad_norm": 0.478515625,
"learning_rate": 5.9314241705455674e-05,
"loss": 0.9543,
"step": 2945
},
{
"epoch": 10.06825938566553,
"grad_norm": 0.408203125,
"learning_rate": 5.895175463751385e-05,
"loss": 0.9579,
"step": 2950
},
{
"epoch": 10.085324232081911,
"grad_norm": 0.380859375,
"learning_rate": 5.858991507111122e-05,
"loss": 0.9506,
"step": 2955
},
{
"epoch": 10.102389078498293,
"grad_norm": 0.3828125,
"learning_rate": 5.8228728713962543e-05,
"loss": 0.9582,
"step": 2960
},
{
"epoch": 10.119453924914676,
"grad_norm": 0.38671875,
"learning_rate": 5.786820126347876e-05,
"loss": 0.9576,
"step": 2965
},
{
"epoch": 10.136518771331058,
"grad_norm": 0.4140625,
"learning_rate": 5.750833840667711e-05,
"loss": 0.9506,
"step": 2970
},
{
"epoch": 10.15358361774744,
"grad_norm": 0.390625,
"learning_rate": 5.7149145820091385e-05,
"loss": 0.952,
"step": 2975
},
{
"epoch": 10.170648464163822,
"grad_norm": 0.38671875,
"learning_rate": 5.6790629169682564e-05,
"loss": 0.9532,
"step": 2980
},
{
"epoch": 10.187713310580206,
"grad_norm": 0.396484375,
"learning_rate": 5.6432794110749134e-05,
"loss": 0.9459,
"step": 2985
},
{
"epoch": 10.204778156996587,
"grad_norm": 0.490234375,
"learning_rate": 5.607564628783817e-05,
"loss": 0.9513,
"step": 2990
},
{
"epoch": 10.22184300341297,
"grad_norm": 0.41796875,
"learning_rate": 5.571919133465605e-05,
"loss": 0.9499,
"step": 2995
},
{
"epoch": 10.238907849829351,
"grad_norm": 0.392578125,
"learning_rate": 5.5363434873979903e-05,
"loss": 0.9481,
"step": 3000
},
{
"epoch": 10.255972696245733,
"grad_norm": 0.380859375,
"learning_rate": 5.500838251756857e-05,
"loss": 0.9501,
"step": 3005
},
{
"epoch": 10.273037542662117,
"grad_norm": 0.3671875,
"learning_rate": 5.465403986607426e-05,
"loss": 0.9498,
"step": 3010
},
{
"epoch": 10.290102389078498,
"grad_norm": 0.396484375,
"learning_rate": 5.430041250895428e-05,
"loss": 0.947,
"step": 3015
},
{
"epoch": 10.30716723549488,
"grad_norm": 0.42578125,
"learning_rate": 5.3947506024382665e-05,
"loss": 0.9581,
"step": 3020
},
{
"epoch": 10.324232081911262,
"grad_norm": 0.408203125,
"learning_rate": 5.359532597916233e-05,
"loss": 0.9549,
"step": 3025
},
{
"epoch": 10.341296928327646,
"grad_norm": 0.40625,
"learning_rate": 5.324387792863719e-05,
"loss": 0.968,
"step": 3030
},
{
"epoch": 10.358361774744028,
"grad_norm": 0.404296875,
"learning_rate": 5.289316741660466e-05,
"loss": 0.9499,
"step": 3035
},
{
"epoch": 10.37542662116041,
"grad_norm": 0.3828125,
"learning_rate": 5.254319997522796e-05,
"loss": 0.9639,
"step": 3040
},
{
"epoch": 10.392491467576791,
"grad_norm": 0.404296875,
"learning_rate": 5.21939811249492e-05,
"loss": 0.9555,
"step": 3045
},
{
"epoch": 10.409556313993175,
"grad_norm": 0.38671875,
"learning_rate": 5.1845516374401784e-05,
"loss": 0.9533,
"step": 3050
},
{
"epoch": 10.426621160409557,
"grad_norm": 0.421875,
"learning_rate": 5.14978112203241e-05,
"loss": 0.9632,
"step": 3055
},
{
"epoch": 10.443686006825939,
"grad_norm": 0.380859375,
"learning_rate": 5.11508711474725e-05,
"loss": 0.9596,
"step": 3060
},
{
"epoch": 10.46075085324232,
"grad_norm": 0.4140625,
"learning_rate": 5.080470162853472e-05,
"loss": 0.963,
"step": 3065
},
{
"epoch": 10.477815699658702,
"grad_norm": 0.412109375,
"learning_rate": 5.0459308124043715e-05,
"loss": 0.9602,
"step": 3070
},
{
"epoch": 10.494880546075086,
"grad_norm": 0.4375,
"learning_rate": 5.0114696082291425e-05,
"loss": 0.9429,
"step": 3075
},
{
"epoch": 10.511945392491468,
"grad_norm": 0.3828125,
"learning_rate": 4.9770870939242986e-05,
"loss": 0.9569,
"step": 3080
},
{
"epoch": 10.52901023890785,
"grad_norm": 0.396484375,
"learning_rate": 4.942783811845074e-05,
"loss": 0.945,
"step": 3085
},
{
"epoch": 10.546075085324231,
"grad_norm": 0.38671875,
"learning_rate": 4.908560303096887e-05,
"loss": 0.955,
"step": 3090
},
{
"epoch": 10.563139931740615,
"grad_norm": 0.404296875,
"learning_rate": 4.874417107526795e-05,
"loss": 0.9583,
"step": 3095
},
{
"epoch": 10.580204778156997,
"grad_norm": 0.38671875,
"learning_rate": 4.840354763714991e-05,
"loss": 0.9499,
"step": 3100
},
{
"epoch": 10.597269624573379,
"grad_norm": 0.41015625,
"learning_rate": 4.8063738089662926e-05,
"loss": 0.9528,
"step": 3105
},
{
"epoch": 10.61433447098976,
"grad_norm": 0.373046875,
"learning_rate": 4.772474779301669e-05,
"loss": 0.9581,
"step": 3110
},
{
"epoch": 10.631399317406144,
"grad_norm": 0.4140625,
"learning_rate": 4.738658209449805e-05,
"loss": 0.9456,
"step": 3115
},
{
"epoch": 10.648464163822526,
"grad_norm": 0.384765625,
"learning_rate": 4.704924632838636e-05,
"loss": 0.9507,
"step": 3120
},
{
"epoch": 10.665529010238908,
"grad_norm": 0.384765625,
"learning_rate": 4.671274581586958e-05,
"loss": 0.9586,
"step": 3125
},
{
"epoch": 10.68259385665529,
"grad_norm": 0.375,
"learning_rate": 4.637708586496018e-05,
"loss": 0.9487,
"step": 3130
},
{
"epoch": 10.699658703071673,
"grad_norm": 0.38671875,
"learning_rate": 4.604227177041156e-05,
"loss": 0.9511,
"step": 3135
},
{
"epoch": 10.716723549488055,
"grad_norm": 0.404296875,
"learning_rate": 4.570830881363439e-05,
"loss": 0.9529,
"step": 3140
},
{
"epoch": 10.733788395904437,
"grad_norm": 0.5078125,
"learning_rate": 4.537520226261333e-05,
"loss": 0.962,
"step": 3145
},
{
"epoch": 10.750853242320819,
"grad_norm": 0.396484375,
"learning_rate": 4.5042957371824057e-05,
"loss": 0.9551,
"step": 3150
},
{
"epoch": 10.7679180887372,
"grad_norm": 0.42578125,
"learning_rate": 4.471157938215017e-05,
"loss": 0.9537,
"step": 3155
},
{
"epoch": 10.784982935153584,
"grad_norm": 0.3984375,
"learning_rate": 4.438107352080076e-05,
"loss": 0.9573,
"step": 3160
},
{
"epoch": 10.802047781569966,
"grad_norm": 0.384765625,
"learning_rate": 4.405144500122772e-05,
"loss": 0.9615,
"step": 3165
},
{
"epoch": 10.819112627986348,
"grad_norm": 0.365234375,
"learning_rate": 4.372269902304363e-05,
"loss": 0.9592,
"step": 3170
},
{
"epoch": 10.83617747440273,
"grad_norm": 0.38671875,
"learning_rate": 4.339484077193974e-05,
"loss": 0.9518,
"step": 3175
},
{
"epoch": 10.853242320819113,
"grad_norm": 0.423828125,
"learning_rate": 4.3067875419604184e-05,
"loss": 0.953,
"step": 3180
},
{
"epoch": 10.870307167235495,
"grad_norm": 0.376953125,
"learning_rate": 4.2741808123640335e-05,
"loss": 0.9578,
"step": 3185
},
{
"epoch": 10.887372013651877,
"grad_norm": 0.36328125,
"learning_rate": 4.241664402748544e-05,
"loss": 0.9548,
"step": 3190
},
{
"epoch": 10.904436860068259,
"grad_norm": 0.361328125,
"learning_rate": 4.209238826032965e-05,
"loss": 0.955,
"step": 3195
},
{
"epoch": 10.921501706484642,
"grad_norm": 0.380859375,
"learning_rate": 4.1769045937034876e-05,
"loss": 0.9591,
"step": 3200
},
{
"epoch": 10.938566552901024,
"grad_norm": 0.43359375,
"learning_rate": 4.144662215805426e-05,
"loss": 0.9544,
"step": 3205
},
{
"epoch": 10.955631399317406,
"grad_norm": 0.58984375,
"learning_rate": 4.1125122009351634e-05,
"loss": 0.9539,
"step": 3210
},
{
"epoch": 10.972696245733788,
"grad_norm": 0.416015625,
"learning_rate": 4.080455056232147e-05,
"loss": 0.9497,
"step": 3215
},
{
"epoch": 10.98976109215017,
"grad_norm": 0.421875,
"learning_rate": 4.048491287370863e-05,
"loss": 0.952,
"step": 3220
},
{
"epoch": 11.0,
"eval_loss": 2.519228935241699,
"eval_runtime": 0.5351,
"eval_samples_per_second": 18.688,
"eval_steps_per_second": 1.869,
"step": 3223
},
{
"epoch": 11.006825938566553,
"grad_norm": 0.404296875,
"learning_rate": 4.016621398552877e-05,
"loss": 0.954,
"step": 3225
},
{
"epoch": 11.023890784982935,
"grad_norm": 0.390625,
"learning_rate": 3.9848458924988684e-05,
"loss": 0.9494,
"step": 3230
},
{
"epoch": 11.040955631399317,
"grad_norm": 0.404296875,
"learning_rate": 3.953165270440721e-05,
"loss": 0.9434,
"step": 3235
},
{
"epoch": 11.058020477815699,
"grad_norm": 0.38671875,
"learning_rate": 3.921580032113602e-05,
"loss": 0.9542,
"step": 3240
},
{
"epoch": 11.075085324232083,
"grad_norm": 0.388671875,
"learning_rate": 3.8900906757480614e-05,
"loss": 0.9519,
"step": 3245
},
{
"epoch": 11.092150170648464,
"grad_norm": 0.388671875,
"learning_rate": 3.858697698062217e-05,
"loss": 0.9597,
"step": 3250
},
{
"epoch": 11.109215017064846,
"grad_norm": 0.373046875,
"learning_rate": 3.8274015942538745e-05,
"loss": 0.9437,
"step": 3255
},
{
"epoch": 11.126279863481228,
"grad_norm": 0.37890625,
"learning_rate": 3.7962028579927555e-05,
"loss": 0.9545,
"step": 3260
},
{
"epoch": 11.143344709897612,
"grad_norm": 0.392578125,
"learning_rate": 3.7651019814126654e-05,
"loss": 0.9524,
"step": 3265
},
{
"epoch": 11.160409556313994,
"grad_norm": 0.37890625,
"learning_rate": 3.734099455103779e-05,
"loss": 0.9591,
"step": 3270
},
{
"epoch": 11.177474402730375,
"grad_norm": 0.38671875,
"learning_rate": 3.7031957681048604e-05,
"loss": 0.9503,
"step": 3275
},
{
"epoch": 11.194539249146757,
"grad_norm": 0.384765625,
"learning_rate": 3.6723914078955825e-05,
"loss": 0.9456,
"step": 3280
},
{
"epoch": 11.211604095563139,
"grad_norm": 0.380859375,
"learning_rate": 3.64168686038881e-05,
"loss": 0.9426,
"step": 3285
},
{
"epoch": 11.228668941979523,
"grad_norm": 0.390625,
"learning_rate": 3.6110826099229453e-05,
"loss": 0.9496,
"step": 3290
},
{
"epoch": 11.245733788395905,
"grad_norm": 0.37109375,
"learning_rate": 3.580579139254303e-05,
"loss": 0.9515,
"step": 3295
},
{
"epoch": 11.262798634812286,
"grad_norm": 0.3828125,
"learning_rate": 3.550176929549468e-05,
"loss": 0.9535,
"step": 3300
},
{
"epoch": 11.279863481228668,
"grad_norm": 0.3671875,
"learning_rate": 3.5198764603777235e-05,
"loss": 0.9575,
"step": 3305
},
{
"epoch": 11.296928327645052,
"grad_norm": 0.376953125,
"learning_rate": 3.489678209703475e-05,
"loss": 0.9468,
"step": 3310
},
{
"epoch": 11.313993174061434,
"grad_norm": 0.396484375,
"learning_rate": 3.459582653878731e-05,
"loss": 0.9536,
"step": 3315
},
{
"epoch": 11.331058020477816,
"grad_norm": 0.39453125,
"learning_rate": 3.429590267635565e-05,
"loss": 0.9575,
"step": 3320
},
{
"epoch": 11.348122866894197,
"grad_norm": 0.38671875,
"learning_rate": 3.399701524078635e-05,
"loss": 0.9533,
"step": 3325
},
{
"epoch": 11.365187713310581,
"grad_norm": 0.380859375,
"learning_rate": 3.369916894677733e-05,
"loss": 0.9414,
"step": 3330
},
{
"epoch": 11.382252559726963,
"grad_norm": 0.421875,
"learning_rate": 3.340236849260324e-05,
"loss": 0.9494,
"step": 3335
},
{
"epoch": 11.399317406143345,
"grad_norm": 0.419921875,
"learning_rate": 3.31066185600417e-05,
"loss": 0.9457,
"step": 3340
},
{
"epoch": 11.416382252559726,
"grad_norm": 0.384765625,
"learning_rate": 3.281192381429894e-05,
"loss": 0.9403,
"step": 3345
},
{
"epoch": 11.43344709897611,
"grad_norm": 0.375,
"learning_rate": 3.251828890393677e-05,
"loss": 0.9489,
"step": 3350
},
{
"epoch": 11.450511945392492,
"grad_norm": 0.412109375,
"learning_rate": 3.222571846079881e-05,
"loss": 0.9525,
"step": 3355
},
{
"epoch": 11.467576791808874,
"grad_norm": 0.37109375,
"learning_rate": 3.193421709993779e-05,
"loss": 0.9574,
"step": 3360
},
{
"epoch": 11.484641638225256,
"grad_norm": 0.390625,
"learning_rate": 3.1643789419542324e-05,
"loss": 0.9453,
"step": 3365
},
{
"epoch": 11.501706484641637,
"grad_norm": 0.3828125,
"learning_rate": 3.135444000086485e-05,
"loss": 0.9462,
"step": 3370
},
{
"epoch": 11.518771331058021,
"grad_norm": 0.384765625,
"learning_rate": 3.1066173408148955e-05,
"loss": 0.9551,
"step": 3375
},
{
"epoch": 11.535836177474403,
"grad_norm": 0.404296875,
"learning_rate": 3.077899418855772e-05,
"loss": 0.9504,
"step": 3380
},
{
"epoch": 11.552901023890785,
"grad_norm": 0.400390625,
"learning_rate": 3.04929068721017e-05,
"loss": 0.9496,
"step": 3385
},
{
"epoch": 11.569965870307167,
"grad_norm": 0.380859375,
"learning_rate": 3.0207915971567624e-05,
"loss": 0.9426,
"step": 3390
},
{
"epoch": 11.58703071672355,
"grad_norm": 0.384765625,
"learning_rate": 2.992402598244727e-05,
"loss": 0.9458,
"step": 3395
},
{
"epoch": 11.604095563139932,
"grad_norm": 0.384765625,
"learning_rate": 2.9641241382866348e-05,
"loss": 0.9525,
"step": 3400
},
{
"epoch": 11.621160409556314,
"grad_norm": 0.400390625,
"learning_rate": 2.9359566633514037e-05,
"loss": 0.9449,
"step": 3405
},
{
"epoch": 11.638225255972696,
"grad_norm": 0.380859375,
"learning_rate": 2.907900617757252e-05,
"loss": 0.9526,
"step": 3410
},
{
"epoch": 11.655290102389078,
"grad_norm": 0.373046875,
"learning_rate": 2.879956444064703e-05,
"loss": 0.9598,
"step": 3415
},
{
"epoch": 11.672354948805461,
"grad_norm": 0.388671875,
"learning_rate": 2.8521245830695864e-05,
"loss": 0.9484,
"step": 3420
},
{
"epoch": 11.689419795221843,
"grad_norm": 0.3828125,
"learning_rate": 2.8244054737960935e-05,
"loss": 0.9431,
"step": 3425
},
{
"epoch": 11.706484641638225,
"grad_norm": 0.365234375,
"learning_rate": 2.7967995534898596e-05,
"loss": 0.9554,
"step": 3430
},
{
"epoch": 11.723549488054607,
"grad_norm": 0.390625,
"learning_rate": 2.7693072576110514e-05,
"loss": 0.9519,
"step": 3435
},
{
"epoch": 11.74061433447099,
"grad_norm": 0.365234375,
"learning_rate": 2.7419290198275095e-05,
"loss": 0.9509,
"step": 3440
},
{
"epoch": 11.757679180887372,
"grad_norm": 0.40234375,
"learning_rate": 2.7146652720079003e-05,
"loss": 0.9578,
"step": 3445
},
{
"epoch": 11.774744027303754,
"grad_norm": 0.376953125,
"learning_rate": 2.6875164442149147e-05,
"loss": 0.9449,
"step": 3450
},
{
"epoch": 11.791808873720136,
"grad_norm": 0.40625,
"learning_rate": 2.6604829646984686e-05,
"loss": 0.9505,
"step": 3455
},
{
"epoch": 11.80887372013652,
"grad_norm": 0.3984375,
"learning_rate": 2.6335652598889683e-05,
"loss": 0.9433,
"step": 3460
},
{
"epoch": 11.825938566552901,
"grad_norm": 0.380859375,
"learning_rate": 2.60676375439055e-05,
"loss": 0.9464,
"step": 3465
},
{
"epoch": 11.843003412969283,
"grad_norm": 0.384765625,
"learning_rate": 2.5800788709744227e-05,
"loss": 0.955,
"step": 3470
},
{
"epoch": 11.860068259385665,
"grad_norm": 0.380859375,
"learning_rate": 2.5535110305721776e-05,
"loss": 0.9458,
"step": 3475
},
{
"epoch": 11.877133105802049,
"grad_norm": 0.3828125,
"learning_rate": 2.5270606522691443e-05,
"loss": 0.9544,
"step": 3480
},
{
"epoch": 11.89419795221843,
"grad_norm": 0.408203125,
"learning_rate": 2.500728153297788e-05,
"loss": 0.9534,
"step": 3485
},
{
"epoch": 11.911262798634812,
"grad_norm": 0.373046875,
"learning_rate": 2.4745139490311254e-05,
"loss": 0.9521,
"step": 3490
},
{
"epoch": 11.928327645051194,
"grad_norm": 0.392578125,
"learning_rate": 2.4484184529761834e-05,
"loss": 0.948,
"step": 3495
},
{
"epoch": 11.945392491467576,
"grad_norm": 0.39453125,
"learning_rate": 2.4224420767674562e-05,
"loss": 0.9543,
"step": 3500
},
{
"epoch": 11.96245733788396,
"grad_norm": 0.375,
"learning_rate": 2.3965852301604254e-05,
"loss": 0.959,
"step": 3505
},
{
"epoch": 11.979522184300341,
"grad_norm": 0.375,
"learning_rate": 2.370848321025093e-05,
"loss": 0.9599,
"step": 3510
},
{
"epoch": 11.996587030716723,
"grad_norm": 0.37109375,
"learning_rate": 2.345231755339554e-05,
"loss": 0.9505,
"step": 3515
},
{
"epoch": 12.0,
"eval_loss": 2.520477771759033,
"eval_runtime": 0.5502,
"eval_samples_per_second": 18.175,
"eval_steps_per_second": 1.818,
"step": 3516
},
{
"epoch": 12.013651877133105,
"grad_norm": 0.43359375,
"learning_rate": 2.3197359371835802e-05,
"loss": 0.9615,
"step": 3520
},
{
"epoch": 12.030716723549489,
"grad_norm": 0.376953125,
"learning_rate": 2.2943612687322525e-05,
"loss": 0.9485,
"step": 3525
},
{
"epoch": 12.04778156996587,
"grad_norm": 0.384765625,
"learning_rate": 2.2691081502496246e-05,
"loss": 0.9475,
"step": 3530
},
{
"epoch": 12.064846416382252,
"grad_norm": 0.388671875,
"learning_rate": 2.243976980082394e-05,
"loss": 0.9393,
"step": 3535
},
{
"epoch": 12.081911262798634,
"grad_norm": 0.39453125,
"learning_rate": 2.218968154653629e-05,
"loss": 0.9466,
"step": 3540
},
{
"epoch": 12.098976109215018,
"grad_norm": 0.376953125,
"learning_rate": 2.194082068456509e-05,
"loss": 0.9537,
"step": 3545
},
{
"epoch": 12.1160409556314,
"grad_norm": 0.36328125,
"learning_rate": 2.169319114048114e-05,
"loss": 0.961,
"step": 3550
},
{
"epoch": 12.133105802047782,
"grad_norm": 0.38671875,
"learning_rate": 2.1446796820432167e-05,
"loss": 0.9493,
"step": 3555
},
{
"epoch": 12.150170648464163,
"grad_norm": 0.384765625,
"learning_rate": 2.1201641611081246e-05,
"loss": 0.948,
"step": 3560
},
{
"epoch": 12.167235494880545,
"grad_norm": 0.373046875,
"learning_rate": 2.0957729379545655e-05,
"loss": 0.9584,
"step": 3565
},
{
"epoch": 12.184300341296929,
"grad_norm": 0.380859375,
"learning_rate": 2.0715063973335568e-05,
"loss": 0.9503,
"step": 3570
},
{
"epoch": 12.20136518771331,
"grad_norm": 0.388671875,
"learning_rate": 2.04736492202937e-05,
"loss": 0.9498,
"step": 3575
},
{
"epoch": 12.218430034129693,
"grad_norm": 0.392578125,
"learning_rate": 2.0233488928534673e-05,
"loss": 0.9553,
"step": 3580
},
{
"epoch": 12.235494880546074,
"grad_norm": 0.396484375,
"learning_rate": 1.9994586886385046e-05,
"loss": 0.9438,
"step": 3585
},
{
"epoch": 12.252559726962458,
"grad_norm": 0.369140625,
"learning_rate": 1.9756946862323535e-05,
"loss": 0.9489,
"step": 3590
},
{
"epoch": 12.26962457337884,
"grad_norm": 0.369140625,
"learning_rate": 1.9520572604921672e-05,
"loss": 0.9477,
"step": 3595
},
{
"epoch": 12.286689419795222,
"grad_norm": 0.375,
"learning_rate": 1.9285467842784467e-05,
"loss": 0.9457,
"step": 3600
},
{
"epoch": 12.303754266211604,
"grad_norm": 0.380859375,
"learning_rate": 1.9051636284491757e-05,
"loss": 0.9541,
"step": 3605
},
{
"epoch": 12.320819112627987,
"grad_norm": 0.365234375,
"learning_rate": 1.8819081618539723e-05,
"loss": 0.9393,
"step": 3610
},
{
"epoch": 12.337883959044369,
"grad_norm": 0.375,
"learning_rate": 1.858780751328255e-05,
"loss": 0.949,
"step": 3615
},
{
"epoch": 12.35494880546075,
"grad_norm": 0.384765625,
"learning_rate": 1.8357817616874694e-05,
"loss": 0.9537,
"step": 3620
},
{
"epoch": 12.372013651877133,
"grad_norm": 0.3671875,
"learning_rate": 1.8129115557213262e-05,
"loss": 0.9505,
"step": 3625
},
{
"epoch": 12.389078498293514,
"grad_norm": 0.3671875,
"learning_rate": 1.7901704941880914e-05,
"loss": 0.9447,
"step": 3630
},
{
"epoch": 12.406143344709898,
"grad_norm": 0.3671875,
"learning_rate": 1.7675589358088763e-05,
"loss": 0.9526,
"step": 3635
},
{
"epoch": 12.42320819112628,
"grad_norm": 0.376953125,
"learning_rate": 1.745077237261994e-05,
"loss": 0.9592,
"step": 3640
},
{
"epoch": 12.440273037542662,
"grad_norm": 0.40234375,
"learning_rate": 1.7227257531773223e-05,
"loss": 0.9515,
"step": 3645
},
{
"epoch": 12.457337883959044,
"grad_norm": 0.408203125,
"learning_rate": 1.7005048361307262e-05,
"loss": 0.9504,
"step": 3650
},
{
"epoch": 12.474402730375427,
"grad_norm": 0.388671875,
"learning_rate": 1.6784148366384754e-05,
"loss": 0.9462,
"step": 3655
},
{
"epoch": 12.491467576791809,
"grad_norm": 0.384765625,
"learning_rate": 1.656456103151728e-05,
"loss": 0.9456,
"step": 3660
},
{
"epoch": 12.508532423208191,
"grad_norm": 0.375,
"learning_rate": 1.6346289820510363e-05,
"loss": 0.9475,
"step": 3665
},
{
"epoch": 12.525597269624573,
"grad_norm": 0.384765625,
"learning_rate": 1.612933817640868e-05,
"loss": 0.9478,
"step": 3670
},
{
"epoch": 12.542662116040956,
"grad_norm": 0.3671875,
"learning_rate": 1.5913709521441988e-05,
"loss": 0.9415,
"step": 3675
},
{
"epoch": 12.559726962457338,
"grad_norm": 0.375,
"learning_rate": 1.5699407256970833e-05,
"loss": 0.9452,
"step": 3680
},
{
"epoch": 12.57679180887372,
"grad_norm": 0.375,
"learning_rate": 1.5486434763433222e-05,
"loss": 0.9479,
"step": 3685
},
{
"epoch": 12.593856655290102,
"grad_norm": 0.38671875,
"learning_rate": 1.527479540029104e-05,
"loss": 0.9495,
"step": 3690
},
{
"epoch": 12.610921501706486,
"grad_norm": 0.3828125,
"learning_rate": 1.5064492505977234e-05,
"loss": 0.936,
"step": 3695
},
{
"epoch": 12.627986348122867,
"grad_norm": 0.392578125,
"learning_rate": 1.4855529397843038e-05,
"loss": 0.9476,
"step": 3700
},
{
"epoch": 12.64505119453925,
"grad_norm": 0.380859375,
"learning_rate": 1.4647909372105672e-05,
"loss": 0.9525,
"step": 3705
},
{
"epoch": 12.662116040955631,
"grad_norm": 0.41796875,
"learning_rate": 1.4441635703796408e-05,
"loss": 0.9477,
"step": 3710
},
{
"epoch": 12.679180887372013,
"grad_norm": 0.3984375,
"learning_rate": 1.4236711646708844e-05,
"loss": 0.9505,
"step": 3715
},
{
"epoch": 12.696245733788396,
"grad_norm": 0.384765625,
"learning_rate": 1.4033140433347569e-05,
"loss": 0.9464,
"step": 3720
},
{
"epoch": 12.713310580204778,
"grad_norm": 0.384765625,
"learning_rate": 1.3830925274877216e-05,
"loss": 0.9392,
"step": 3725
},
{
"epoch": 12.73037542662116,
"grad_norm": 0.37890625,
"learning_rate": 1.363006936107183e-05,
"loss": 0.9495,
"step": 3730
},
{
"epoch": 12.747440273037542,
"grad_norm": 0.3828125,
"learning_rate": 1.343057586026446e-05,
"loss": 0.9423,
"step": 3735
},
{
"epoch": 12.764505119453926,
"grad_norm": 0.416015625,
"learning_rate": 1.3232447919297274e-05,
"loss": 0.9448,
"step": 3740
},
{
"epoch": 12.781569965870307,
"grad_norm": 0.404296875,
"learning_rate": 1.3035688663471834e-05,
"loss": 0.9544,
"step": 3745
},
{
"epoch": 12.79863481228669,
"grad_norm": 0.37109375,
"learning_rate": 1.2840301196499893e-05,
"loss": 0.9548,
"step": 3750
},
{
"epoch": 12.815699658703071,
"grad_norm": 0.376953125,
"learning_rate": 1.2646288600454448e-05,
"loss": 0.9492,
"step": 3755
},
{
"epoch": 12.832764505119453,
"grad_norm": 0.373046875,
"learning_rate": 1.2453653935720867e-05,
"loss": 0.9506,
"step": 3760
},
{
"epoch": 12.849829351535837,
"grad_norm": 0.388671875,
"learning_rate": 1.2262400240949023e-05,
"loss": 0.9543,
"step": 3765
},
{
"epoch": 12.866894197952218,
"grad_norm": 0.369140625,
"learning_rate": 1.2072530533005012e-05,
"loss": 0.9418,
"step": 3770
},
{
"epoch": 12.8839590443686,
"grad_norm": 0.369140625,
"learning_rate": 1.1884047806923815e-05,
"loss": 0.9475,
"step": 3775
},
{
"epoch": 12.901023890784982,
"grad_norm": 0.39453125,
"learning_rate": 1.169695503586179e-05,
"loss": 0.9428,
"step": 3780
},
{
"epoch": 12.918088737201366,
"grad_norm": 0.38671875,
"learning_rate": 1.1511255171050084e-05,
"loss": 0.9529,
"step": 3785
},
{
"epoch": 12.935153583617748,
"grad_norm": 0.376953125,
"learning_rate": 1.1326951141747788e-05,
"loss": 0.9455,
"step": 3790
},
{
"epoch": 12.95221843003413,
"grad_norm": 0.376953125,
"learning_rate": 1.1144045855195973e-05,
"loss": 0.9537,
"step": 3795
},
{
"epoch": 12.969283276450511,
"grad_norm": 0.396484375,
"learning_rate": 1.0962542196571634e-05,
"loss": 0.9426,
"step": 3800
},
{
"epoch": 12.986348122866895,
"grad_norm": 0.373046875,
"learning_rate": 1.078244302894229e-05,
"loss": 0.9442,
"step": 3805
},
{
"epoch": 13.0,
"eval_loss": 2.522336959838867,
"eval_runtime": 0.5484,
"eval_samples_per_second": 18.236,
"eval_steps_per_second": 1.824,
"step": 3809
},
{
"epoch": 13.003412969283277,
"grad_norm": 0.376953125,
"learning_rate": 1.0603751193220846e-05,
"loss": 0.956,
"step": 3810
},
{
"epoch": 13.020477815699659,
"grad_norm": 0.392578125,
"learning_rate": 1.0426469508120662e-05,
"loss": 0.9449,
"step": 3815
},
{
"epoch": 13.03754266211604,
"grad_norm": 0.390625,
"learning_rate": 1.0250600770111185e-05,
"loss": 0.9479,
"step": 3820
},
{
"epoch": 13.054607508532424,
"grad_norm": 0.392578125,
"learning_rate": 1.0076147753373789e-05,
"loss": 0.953,
"step": 3825
},
{
"epoch": 13.071672354948806,
"grad_norm": 0.388671875,
"learning_rate": 9.903113209758096e-06,
"loss": 0.9436,
"step": 3830
},
{
"epoch": 13.088737201365188,
"grad_norm": 0.380859375,
"learning_rate": 9.731499868738447e-06,
"loss": 0.9454,
"step": 3835
},
{
"epoch": 13.10580204778157,
"grad_norm": 0.3828125,
"learning_rate": 9.561310437370907e-06,
"loss": 0.9556,
"step": 3840
},
{
"epoch": 13.122866894197951,
"grad_norm": 0.373046875,
"learning_rate": 9.392547600250634e-06,
"loss": 0.949,
"step": 3845
},
{
"epoch": 13.139931740614335,
"grad_norm": 0.380859375,
"learning_rate": 9.225214019469385e-06,
"loss": 0.9382,
"step": 3850
},
{
"epoch": 13.156996587030717,
"grad_norm": 0.40234375,
"learning_rate": 9.059312334573633e-06,
"loss": 0.943,
"step": 3855
},
{
"epoch": 13.174061433447099,
"grad_norm": 0.3828125,
"learning_rate": 8.89484516252287e-06,
"loss": 0.9534,
"step": 3860
},
{
"epoch": 13.19112627986348,
"grad_norm": 0.369140625,
"learning_rate": 8.731815097648433e-06,
"loss": 0.9526,
"step": 3865
},
{
"epoch": 13.208191126279864,
"grad_norm": 0.392578125,
"learning_rate": 8.570224711612385e-06,
"loss": 0.9419,
"step": 3870
},
{
"epoch": 13.225255972696246,
"grad_norm": 0.373046875,
"learning_rate": 8.410076553367208e-06,
"loss": 0.9511,
"step": 3875
},
{
"epoch": 13.242320819112628,
"grad_norm": 0.380859375,
"learning_rate": 8.251373149115293e-06,
"loss": 0.9489,
"step": 3880
},
{
"epoch": 13.25938566552901,
"grad_norm": 0.36328125,
"learning_rate": 8.094117002269363e-06,
"loss": 0.9428,
"step": 3885
},
{
"epoch": 13.276450511945393,
"grad_norm": 0.443359375,
"learning_rate": 7.938310593412879e-06,
"loss": 0.9485,
"step": 3890
},
{
"epoch": 13.293515358361775,
"grad_norm": 0.3671875,
"learning_rate": 7.783956380260837e-06,
"loss": 0.955,
"step": 3895
},
{
"epoch": 13.310580204778157,
"grad_norm": 0.384765625,
"learning_rate": 7.631056797621106e-06,
"loss": 0.9566,
"step": 3900
},
{
"epoch": 13.327645051194539,
"grad_norm": 0.369140625,
"learning_rate": 7.479614257355971e-06,
"loss": 0.9495,
"step": 3905
},
{
"epoch": 13.344709897610922,
"grad_norm": 0.376953125,
"learning_rate": 7.329631148344118e-06,
"loss": 0.9535,
"step": 3910
},
{
"epoch": 13.361774744027304,
"grad_norm": 0.375,
"learning_rate": 7.181109836442912e-06,
"loss": 0.9473,
"step": 3915
},
{
"epoch": 13.378839590443686,
"grad_norm": 0.37890625,
"learning_rate": 7.034052664451118e-06,
"loss": 0.946,
"step": 3920
},
{
"epoch": 13.395904436860068,
"grad_norm": 0.380859375,
"learning_rate": 6.88846195207189e-06,
"loss": 0.9526,
"step": 3925
},
{
"epoch": 13.41296928327645,
"grad_norm": 0.365234375,
"learning_rate": 6.7443399958762584e-06,
"loss": 0.9416,
"step": 3930
},
{
"epoch": 13.430034129692833,
"grad_norm": 0.365234375,
"learning_rate": 6.6016890692668364e-06,
"loss": 0.9529,
"step": 3935
},
{
"epoch": 13.447098976109215,
"grad_norm": 0.376953125,
"learning_rate": 6.460511422441984e-06,
"loss": 0.9427,
"step": 3940
},
{
"epoch": 13.464163822525597,
"grad_norm": 0.37890625,
"learning_rate": 6.320809282360319e-06,
"loss": 0.9516,
"step": 3945
},
{
"epoch": 13.481228668941979,
"grad_norm": 0.380859375,
"learning_rate": 6.1825848527055865e-06,
"loss": 0.9448,
"step": 3950
},
{
"epoch": 13.498293515358363,
"grad_norm": 0.384765625,
"learning_rate": 6.04584031385188e-06,
"loss": 0.9542,
"step": 3955
},
{
"epoch": 13.515358361774744,
"grad_norm": 0.376953125,
"learning_rate": 5.910577822829233e-06,
"loss": 0.9525,
"step": 3960
},
{
"epoch": 13.532423208191126,
"grad_norm": 0.3671875,
"learning_rate": 5.77679951328971e-06,
"loss": 0.9502,
"step": 3965
},
{
"epoch": 13.549488054607508,
"grad_norm": 0.373046875,
"learning_rate": 5.644507495473572e-06,
"loss": 0.9464,
"step": 3970
},
{
"epoch": 13.56655290102389,
"grad_norm": 0.37890625,
"learning_rate": 5.5137038561761115e-06,
"loss": 0.9531,
"step": 3975
},
{
"epoch": 13.583617747440274,
"grad_norm": 0.375,
"learning_rate": 5.3843906587146886e-06,
"loss": 0.9498,
"step": 3980
},
{
"epoch": 13.600682593856655,
"grad_norm": 0.37890625,
"learning_rate": 5.256569942896217e-06,
"loss": 0.945,
"step": 3985
},
{
"epoch": 13.617747440273037,
"grad_norm": 0.365234375,
"learning_rate": 5.130243724984995e-06,
"loss": 0.9468,
"step": 3990
},
{
"epoch": 13.634812286689419,
"grad_norm": 0.369140625,
"learning_rate": 5.005413997670816e-06,
"loss": 0.9517,
"step": 3995
},
{
"epoch": 13.651877133105803,
"grad_norm": 0.365234375,
"learning_rate": 4.8820827300376075e-06,
"loss": 0.9502,
"step": 4000
},
{
"epoch": 13.668941979522184,
"grad_norm": 0.369140625,
"learning_rate": 4.760251867532362e-06,
"loss": 0.9462,
"step": 4005
},
{
"epoch": 13.686006825938566,
"grad_norm": 0.384765625,
"learning_rate": 4.639923331934471e-06,
"loss": 0.9476,
"step": 4010
},
{
"epoch": 13.703071672354948,
"grad_norm": 0.369140625,
"learning_rate": 4.521099021325336e-06,
"loss": 0.9556,
"step": 4015
},
{
"epoch": 13.720136518771332,
"grad_norm": 0.390625,
"learning_rate": 4.403780810058511e-06,
"loss": 0.9438,
"step": 4020
},
{
"epoch": 13.737201365187714,
"grad_norm": 0.470703125,
"learning_rate": 4.287970548730069e-06,
"loss": 0.9495,
"step": 4025
},
{
"epoch": 13.754266211604095,
"grad_norm": 0.36328125,
"learning_rate": 4.173670064149482e-06,
"loss": 0.934,
"step": 4030
},
{
"epoch": 13.771331058020477,
"grad_norm": 0.384765625,
"learning_rate": 4.060881159310725e-06,
"loss": 0.9502,
"step": 4035
},
{
"epoch": 13.788395904436861,
"grad_norm": 0.388671875,
"learning_rate": 3.949605613363882e-06,
"loss": 0.939,
"step": 4040
},
{
"epoch": 13.805460750853243,
"grad_norm": 0.37890625,
"learning_rate": 3.839845181587098e-06,
"loss": 0.9559,
"step": 4045
},
{
"epoch": 13.822525597269625,
"grad_norm": 0.376953125,
"learning_rate": 3.7316015953588467e-06,
"loss": 0.9547,
"step": 4050
},
{
"epoch": 13.839590443686006,
"grad_norm": 0.384765625,
"learning_rate": 3.6248765621306414e-06,
"loss": 0.9463,
"step": 4055
},
{
"epoch": 13.856655290102388,
"grad_norm": 0.376953125,
"learning_rate": 3.519671765400079e-06,
"loss": 0.9454,
"step": 4060
},
{
"epoch": 13.873720136518772,
"grad_norm": 0.373046875,
"learning_rate": 3.4159888646843495e-06,
"loss": 0.9485,
"step": 4065
},
{
"epoch": 13.890784982935154,
"grad_norm": 0.375,
"learning_rate": 3.313829495493992e-06,
"loss": 0.9455,
"step": 4070
},
{
"epoch": 13.907849829351536,
"grad_norm": 0.37890625,
"learning_rate": 3.2131952693070898e-06,
"loss": 0.9409,
"step": 4075
},
{
"epoch": 13.924914675767917,
"grad_norm": 0.396484375,
"learning_rate": 3.1140877735439387e-06,
"loss": 0.9468,
"step": 4080
},
{
"epoch": 13.941979522184301,
"grad_norm": 0.375,
"learning_rate": 3.0165085715418763e-06,
"loss": 0.9434,
"step": 4085
},
{
"epoch": 13.959044368600683,
"grad_norm": 0.3671875,
"learning_rate": 2.9204592025307566e-06,
"loss": 0.9455,
"step": 4090
},
{
"epoch": 13.976109215017065,
"grad_norm": 0.369140625,
"learning_rate": 2.8259411816085492e-06,
"loss": 0.9437,
"step": 4095
},
{
"epoch": 13.993174061433447,
"grad_norm": 0.478515625,
"learning_rate": 2.732955999717546e-06,
"loss": 0.9469,
"step": 4100
},
{
"epoch": 14.0,
"eval_loss": 2.5227127075195312,
"eval_runtime": 0.542,
"eval_samples_per_second": 18.45,
"eval_steps_per_second": 1.845,
"step": 4102
},
{
"epoch": 14.01023890784983,
"grad_norm": 0.376953125,
"learning_rate": 2.6415051236207355e-06,
"loss": 0.9508,
"step": 4105
},
{
"epoch": 14.027303754266212,
"grad_norm": 0.375,
"learning_rate": 2.551589995878789e-06,
"loss": 0.9459,
"step": 4110
},
{
"epoch": 14.044368600682594,
"grad_norm": 0.380859375,
"learning_rate": 2.4632120348272003e-06,
"loss": 0.9465,
"step": 4115
},
{
"epoch": 14.061433447098976,
"grad_norm": 0.37890625,
"learning_rate": 2.376372634553936e-06,
"loss": 0.9475,
"step": 4120
},
{
"epoch": 14.078498293515358,
"grad_norm": 0.376953125,
"learning_rate": 2.291073164877511e-06,
"loss": 0.9435,
"step": 4125
},
{
"epoch": 14.095563139931741,
"grad_norm": 0.37890625,
"learning_rate": 2.207314971325292e-06,
"loss": 0.9546,
"step": 4130
},
{
"epoch": 14.112627986348123,
"grad_norm": 0.400390625,
"learning_rate": 2.125099375112316e-06,
"loss": 0.9496,
"step": 4135
},
{
"epoch": 14.129692832764505,
"grad_norm": 0.3671875,
"learning_rate": 2.0444276731204415e-06,
"loss": 0.9592,
"step": 4140
},
{
"epoch": 14.146757679180887,
"grad_norm": 0.37890625,
"learning_rate": 1.9653011378779283e-06,
"loss": 0.9446,
"step": 4145
},
{
"epoch": 14.16382252559727,
"grad_norm": 0.5625,
"learning_rate": 1.88772101753929e-06,
"loss": 0.9374,
"step": 4150
},
{
"epoch": 14.180887372013652,
"grad_norm": 0.37890625,
"learning_rate": 1.8116885358656744e-06,
"loss": 0.9543,
"step": 4155
},
{
"epoch": 14.197952218430034,
"grad_norm": 0.37109375,
"learning_rate": 1.7372048922054906e-06,
"loss": 0.9488,
"step": 4160
},
{
"epoch": 14.215017064846416,
"grad_norm": 0.373046875,
"learning_rate": 1.6642712614755695e-06,
"loss": 0.9466,
"step": 4165
},
{
"epoch": 14.2320819112628,
"grad_norm": 0.396484375,
"learning_rate": 1.5928887941426107e-06,
"loss": 0.9482,
"step": 4170
},
{
"epoch": 14.249146757679181,
"grad_norm": 0.373046875,
"learning_rate": 1.523058616204942e-06,
"loss": 0.9449,
"step": 4175
},
{
"epoch": 14.266211604095563,
"grad_norm": 0.3984375,
"learning_rate": 1.4547818291749115e-06,
"loss": 0.9562,
"step": 4180
},
{
"epoch": 14.283276450511945,
"grad_norm": 0.388671875,
"learning_rate": 1.3880595100613792e-06,
"loss": 0.9445,
"step": 4185
},
{
"epoch": 14.300341296928327,
"grad_norm": 0.376953125,
"learning_rate": 1.3228927113528189e-06,
"loss": 0.9457,
"step": 4190
},
{
"epoch": 14.31740614334471,
"grad_norm": 0.388671875,
"learning_rate": 1.2592824610006215e-06,
"loss": 0.9488,
"step": 4195
},
{
"epoch": 14.334470989761092,
"grad_norm": 0.38671875,
"learning_rate": 1.1972297624030072e-06,
"loss": 0.9437,
"step": 4200
},
{
"epoch": 14.351535836177474,
"grad_norm": 0.3671875,
"learning_rate": 1.1367355943890823e-06,
"loss": 0.9459,
"step": 4205
},
{
"epoch": 14.368600682593856,
"grad_norm": 0.396484375,
"learning_rate": 1.0778009112034748e-06,
"loss": 0.9477,
"step": 4210
},
{
"epoch": 14.38566552901024,
"grad_norm": 0.375,
"learning_rate": 1.0204266424912123e-06,
"loss": 0.95,
"step": 4215
},
{
"epoch": 14.402730375426621,
"grad_norm": 0.40625,
"learning_rate": 9.64613693283123e-07,
"loss": 0.9477,
"step": 4220
},
{
"epoch": 14.419795221843003,
"grad_norm": 0.375,
"learning_rate": 9.103629439815354e-07,
"loss": 0.9461,
"step": 4225
},
{
"epoch": 14.436860068259385,
"grad_norm": 0.3828125,
"learning_rate": 8.57675250346368e-07,
"loss": 0.9585,
"step": 4230
},
{
"epoch": 14.453924914675769,
"grad_norm": 0.443359375,
"learning_rate": 8.065514434816845e-07,
"loss": 0.9434,
"step": 4235
},
{
"epoch": 14.47098976109215,
"grad_norm": 0.396484375,
"learning_rate": 7.569923298225146e-07,
"loss": 0.941,
"step": 4240
},
{
"epoch": 14.488054607508532,
"grad_norm": 0.375,
"learning_rate": 7.08998691122198e-07,
"loss": 0.9527,
"step": 4245
},
{
"epoch": 14.505119453924914,
"grad_norm": 0.380859375,
"learning_rate": 6.625712844400056e-07,
"loss": 0.9484,
"step": 4250
},
{
"epoch": 14.522184300341298,
"grad_norm": 0.390625,
"learning_rate": 6.177108421292266e-07,
"loss": 0.9453,
"step": 4255
},
{
"epoch": 14.53924914675768,
"grad_norm": 0.404296875,
"learning_rate": 5.744180718255776e-07,
"loss": 0.9464,
"step": 4260
},
{
"epoch": 14.556313993174061,
"grad_norm": 0.375,
"learning_rate": 5.326936564361118e-07,
"loss": 0.943,
"step": 4265
},
{
"epoch": 14.573378839590443,
"grad_norm": 0.369140625,
"learning_rate": 4.92538254128383e-07,
"loss": 0.9422,
"step": 4270
},
{
"epoch": 14.590443686006825,
"grad_norm": 0.390625,
"learning_rate": 4.5395249832007604e-07,
"loss": 0.9591,
"step": 4275
},
{
"epoch": 14.607508532423209,
"grad_norm": 0.396484375,
"learning_rate": 4.1693699766902626e-07,
"loss": 0.9475,
"step": 4280
},
{
"epoch": 14.62457337883959,
"grad_norm": 0.369140625,
"learning_rate": 3.814923360636158e-07,
"loss": 0.9391,
"step": 4285
},
{
"epoch": 14.641638225255972,
"grad_norm": 0.36328125,
"learning_rate": 3.4761907261356976e-07,
"loss": 0.9574,
"step": 4290
},
{
"epoch": 14.658703071672354,
"grad_norm": 0.388671875,
"learning_rate": 3.1531774164111903e-07,
"loss": 0.9495,
"step": 4295
},
{
"epoch": 14.675767918088738,
"grad_norm": 0.373046875,
"learning_rate": 2.8458885267260705e-07,
"loss": 0.9537,
"step": 4300
},
{
"epoch": 14.69283276450512,
"grad_norm": 0.38671875,
"learning_rate": 2.554328904303738e-07,
"loss": 0.9435,
"step": 4305
},
{
"epoch": 14.709897610921502,
"grad_norm": 0.39453125,
"learning_rate": 2.2785031482521758e-07,
"loss": 0.9474,
"step": 4310
},
{
"epoch": 14.726962457337883,
"grad_norm": 0.376953125,
"learning_rate": 2.0184156094905648e-07,
"loss": 0.947,
"step": 4315
},
{
"epoch": 14.744027303754265,
"grad_norm": 0.3671875,
"learning_rate": 1.7740703906810042e-07,
"loss": 0.9431,
"step": 4320
},
{
"epoch": 14.761092150170649,
"grad_norm": 0.384765625,
"learning_rate": 1.545471346164007e-07,
"loss": 0.9431,
"step": 4325
},
{
"epoch": 14.77815699658703,
"grad_norm": 0.37890625,
"learning_rate": 1.3326220818968838e-07,
"loss": 0.9455,
"step": 4330
},
{
"epoch": 14.795221843003413,
"grad_norm": 0.375,
"learning_rate": 1.1355259553978981e-07,
"loss": 0.9512,
"step": 4335
},
{
"epoch": 14.812286689419794,
"grad_norm": 0.390625,
"learning_rate": 9.541860756925314e-08,
"loss": 0.9439,
"step": 4340
},
{
"epoch": 14.829351535836178,
"grad_norm": 0.37109375,
"learning_rate": 7.886053032649665e-08,
"loss": 0.9548,
"step": 4345
},
{
"epoch": 14.84641638225256,
"grad_norm": 0.4921875,
"learning_rate": 6.387862500125685e-08,
"loss": 0.9437,
"step": 4350
},
{
"epoch": 14.863481228668942,
"grad_norm": 0.380859375,
"learning_rate": 5.047312792046954e-08,
"loss": 0.9512,
"step": 4355
},
{
"epoch": 14.880546075085324,
"grad_norm": 0.39453125,
"learning_rate": 3.8644250544594975e-08,
"loss": 0.9478,
"step": 4360
},
{
"epoch": 14.897610921501707,
"grad_norm": 0.380859375,
"learning_rate": 2.839217946422057e-08,
"loss": 0.9362,
"step": 4365
},
{
"epoch": 14.914675767918089,
"grad_norm": 0.380859375,
"learning_rate": 1.971707639712994e-08,
"loss": 0.9507,
"step": 4370
},
{
"epoch": 14.93174061433447,
"grad_norm": 0.37109375,
"learning_rate": 1.2619078185793776e-08,
"loss": 0.948,
"step": 4375
},
{
"epoch": 14.948805460750853,
"grad_norm": 0.400390625,
"learning_rate": 7.098296795138293e-09,
"loss": 0.9524,
"step": 4380
},
{
"epoch": 14.965870307167236,
"grad_norm": 0.36328125,
"learning_rate": 3.154819310868806e-09,
"loss": 0.9497,
"step": 4385
},
{
"epoch": 14.982935153583618,
"grad_norm": 0.400390625,
"learning_rate": 7.887079380153317e-10,
"loss": 0.9536,
"step": 4390
},
{
"epoch": 15.0,
"grad_norm": 0.373046875,
"learning_rate": 0.0,
"loss": 0.9444,
"step": 4395
},
{
"epoch": 15.0,
"eval_loss": 2.523277521133423,
"eval_runtime": 0.5592,
"eval_samples_per_second": 17.883,
"eval_steps_per_second": 1.788,
"step": 4395
},
{
"epoch": 15.0,
"step": 4395,
"total_flos": 2.581505823377195e+18,
"train_loss": 1.0488379673203783,
"train_runtime": 23446.7186,
"train_samples_per_second": 8.983,
"train_steps_per_second": 0.187
}
],
"logging_steps": 5,
"max_steps": 4395,
"num_input_tokens_seen": 0,
"num_train_epochs": 15,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 2.581505823377195e+18,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}