jetpack-model-pt / checkpoint-6500 /trainer_state.json
goldsounds's picture
Upload folder using huggingface_hub
7111b87 verified
raw
history blame contribute delete
No virus
221 kB
{
"best_metric": 0.7146974802017212,
"best_model_checkpoint": "saves/starcoder2-7b/lora/sft/checkpoint-6000",
"epoch": 1.037509976057462,
"eval_steps": 100,
"global_step": 6500,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0,
"grad_norm": 0.628385603427887,
"learning_rate": 4.999999126897802e-05,
"loss": 1.2582,
"step": 5
},
{
"epoch": 0.0,
"grad_norm": 1.0855119228363037,
"learning_rate": 4.999996507591817e-05,
"loss": 0.801,
"step": 10
},
{
"epoch": 0.0,
"grad_norm": 1.5689586400985718,
"learning_rate": 4.9999921420838745e-05,
"loss": 1.067,
"step": 15
},
{
"epoch": 0.0,
"grad_norm": 2.0851330757141113,
"learning_rate": 4.999986030377024e-05,
"loss": 1.2953,
"step": 20
},
{
"epoch": 0.0,
"grad_norm": 1.397479772567749,
"learning_rate": 4.999978172475535e-05,
"loss": 0.9826,
"step": 25
},
{
"epoch": 0.0,
"grad_norm": 1.344118595123291,
"learning_rate": 4.9999685683848954e-05,
"loss": 0.9485,
"step": 30
},
{
"epoch": 0.01,
"grad_norm": 1.158163070678711,
"learning_rate": 4.9999596278606616e-05,
"loss": 0.8103,
"step": 35
},
{
"epoch": 0.01,
"grad_norm": 1.602233648300171,
"learning_rate": 4.999946880647276e-05,
"loss": 0.8648,
"step": 40
},
{
"epoch": 0.01,
"grad_norm": 1.557242751121521,
"learning_rate": 4.999932387266596e-05,
"loss": 1.0198,
"step": 45
},
{
"epoch": 0.01,
"grad_norm": 1.36068856716156,
"learning_rate": 4.999916147728746e-05,
"loss": 0.9367,
"step": 50
},
{
"epoch": 0.01,
"grad_norm": 1.3263639211654663,
"learning_rate": 4.999898162045068e-05,
"loss": 0.9695,
"step": 55
},
{
"epoch": 0.01,
"grad_norm": 1.333601474761963,
"learning_rate": 4.999878430228126e-05,
"loss": 1.1509,
"step": 60
},
{
"epoch": 0.01,
"grad_norm": 1.4753800630569458,
"learning_rate": 4.999856952291702e-05,
"loss": 1.1461,
"step": 65
},
{
"epoch": 0.01,
"grad_norm": 1.5096240043640137,
"learning_rate": 4.9998337282507965e-05,
"loss": 1.1722,
"step": 70
},
{
"epoch": 0.01,
"grad_norm": 1.189892053604126,
"learning_rate": 4.999808758121633e-05,
"loss": 1.1834,
"step": 75
},
{
"epoch": 0.01,
"grad_norm": 0.9292634725570679,
"learning_rate": 4.999782041921651e-05,
"loss": 0.9498,
"step": 80
},
{
"epoch": 0.01,
"grad_norm": 2.1775777339935303,
"learning_rate": 4.9997535796695134e-05,
"loss": 0.9346,
"step": 85
},
{
"epoch": 0.01,
"grad_norm": 1.6854296922683716,
"learning_rate": 4.999723371385099e-05,
"loss": 1.119,
"step": 90
},
{
"epoch": 0.02,
"grad_norm": 1.4571490287780762,
"learning_rate": 4.999691417089507e-05,
"loss": 0.8671,
"step": 95
},
{
"epoch": 0.02,
"grad_norm": 1.277044653892517,
"learning_rate": 4.999657716805059e-05,
"loss": 1.2469,
"step": 100
},
{
"epoch": 0.02,
"eval_loss": 0.8478816747665405,
"eval_runtime": 96.2736,
"eval_samples_per_second": 7.24,
"eval_steps_per_second": 7.24,
"step": 100
},
{
"epoch": 0.02,
"grad_norm": 0.6687743067741394,
"learning_rate": 4.9996222705552933e-05,
"loss": 0.735,
"step": 105
},
{
"epoch": 0.02,
"grad_norm": 1.3488354682922363,
"learning_rate": 4.9995850783649665e-05,
"loss": 0.8344,
"step": 110
},
{
"epoch": 0.02,
"grad_norm": 1.1043323278427124,
"learning_rate": 4.9995461402600593e-05,
"loss": 0.8254,
"step": 115
},
{
"epoch": 0.02,
"grad_norm": 0.9382895827293396,
"learning_rate": 4.9995054562677684e-05,
"loss": 0.9179,
"step": 120
},
{
"epoch": 0.02,
"grad_norm": 1.2824612855911255,
"learning_rate": 4.9994630264165107e-05,
"loss": 0.8663,
"step": 125
},
{
"epoch": 0.02,
"grad_norm": 1.0491925477981567,
"learning_rate": 4.999418850735923e-05,
"loss": 0.9247,
"step": 130
},
{
"epoch": 0.02,
"grad_norm": 1.3642233610153198,
"learning_rate": 4.99937292925686e-05,
"loss": 0.8253,
"step": 135
},
{
"epoch": 0.02,
"grad_norm": 3.747757911682129,
"learning_rate": 4.9993252620113976e-05,
"loss": 1.0245,
"step": 140
},
{
"epoch": 0.02,
"grad_norm": 1.299494981765747,
"learning_rate": 4.999275849032832e-05,
"loss": 0.8723,
"step": 145
},
{
"epoch": 0.02,
"grad_norm": 1.7195830345153809,
"learning_rate": 4.999224690355675e-05,
"loss": 1.0524,
"step": 150
},
{
"epoch": 0.02,
"grad_norm": 0.9922987222671509,
"learning_rate": 4.9991717860156616e-05,
"loss": 0.9502,
"step": 155
},
{
"epoch": 0.03,
"grad_norm": 1.0577458143234253,
"learning_rate": 4.9991171360497437e-05,
"loss": 1.0115,
"step": 160
},
{
"epoch": 0.03,
"grad_norm": 1.0001195669174194,
"learning_rate": 4.999060740496093e-05,
"loss": 1.1999,
"step": 165
},
{
"epoch": 0.03,
"grad_norm": 1.2456804513931274,
"learning_rate": 4.999002599394102e-05,
"loss": 0.8882,
"step": 170
},
{
"epoch": 0.03,
"grad_norm": 1.0445325374603271,
"learning_rate": 4.9989427127843814e-05,
"loss": 1.0615,
"step": 175
},
{
"epoch": 0.03,
"grad_norm": 1.2410887479782104,
"learning_rate": 4.9988810807087584e-05,
"loss": 1.1068,
"step": 180
},
{
"epoch": 0.03,
"grad_norm": 0.8935971260070801,
"learning_rate": 4.998817703210285e-05,
"loss": 0.6683,
"step": 185
},
{
"epoch": 0.03,
"grad_norm": 1.1614488363265991,
"learning_rate": 4.9987525803332265e-05,
"loss": 0.7446,
"step": 190
},
{
"epoch": 0.03,
"grad_norm": 0.9392004013061523,
"learning_rate": 4.998685712123072e-05,
"loss": 0.7397,
"step": 195
},
{
"epoch": 0.03,
"grad_norm": 1.0314444303512573,
"learning_rate": 4.9986170986265266e-05,
"loss": 1.3584,
"step": 200
},
{
"epoch": 0.03,
"eval_loss": 0.8368077278137207,
"eval_runtime": 96.5262,
"eval_samples_per_second": 7.221,
"eval_steps_per_second": 7.221,
"step": 200
},
{
"epoch": 0.03,
"grad_norm": 0.8964811563491821,
"learning_rate": 4.998546739891516e-05,
"loss": 0.9546,
"step": 205
},
{
"epoch": 0.03,
"grad_norm": 1.0679796934127808,
"learning_rate": 4.998474635967185e-05,
"loss": 0.864,
"step": 210
},
{
"epoch": 0.03,
"grad_norm": 1.2340985536575317,
"learning_rate": 4.998400786903896e-05,
"loss": 0.885,
"step": 215
},
{
"epoch": 0.04,
"grad_norm": 1.7219617366790771,
"learning_rate": 4.9983251927532315e-05,
"loss": 1.1069,
"step": 220
},
{
"epoch": 0.04,
"grad_norm": 1.1480705738067627,
"learning_rate": 4.9982478535679924e-05,
"loss": 1.0416,
"step": 225
},
{
"epoch": 0.04,
"grad_norm": 1.515589714050293,
"learning_rate": 4.9981687694021996e-05,
"loss": 1.1844,
"step": 230
},
{
"epoch": 0.04,
"grad_norm": 1.6687963008880615,
"learning_rate": 4.998087940311091e-05,
"loss": 0.8664,
"step": 235
},
{
"epoch": 0.04,
"grad_norm": 1.9256645441055298,
"learning_rate": 4.998005366351125e-05,
"loss": 1.0125,
"step": 240
},
{
"epoch": 0.04,
"grad_norm": 1.2500052452087402,
"learning_rate": 4.997921047579978e-05,
"loss": 1.1374,
"step": 245
},
{
"epoch": 0.04,
"grad_norm": 1.0543216466903687,
"learning_rate": 4.9978349840565434e-05,
"loss": 0.8502,
"step": 250
},
{
"epoch": 0.04,
"grad_norm": 1.3009012937545776,
"learning_rate": 4.997747175840937e-05,
"loss": 1.0357,
"step": 255
},
{
"epoch": 0.04,
"grad_norm": 0.8456661105155945,
"learning_rate": 4.997657622994491e-05,
"loss": 0.6883,
"step": 260
},
{
"epoch": 0.04,
"grad_norm": 0.5856515765190125,
"learning_rate": 4.9975663255797555e-05,
"loss": 0.7656,
"step": 265
},
{
"epoch": 0.04,
"grad_norm": 0.973818302154541,
"learning_rate": 4.997473283660501e-05,
"loss": 0.823,
"step": 270
},
{
"epoch": 0.04,
"grad_norm": 0.9960187673568726,
"learning_rate": 4.997378497301715e-05,
"loss": 0.8726,
"step": 275
},
{
"epoch": 0.04,
"grad_norm": 1.2900679111480713,
"learning_rate": 4.997281966569604e-05,
"loss": 0.9781,
"step": 280
},
{
"epoch": 0.05,
"grad_norm": 1.828894853591919,
"learning_rate": 4.9971836915315926e-05,
"loss": 0.8932,
"step": 285
},
{
"epoch": 0.05,
"grad_norm": 1.239621877670288,
"learning_rate": 4.9970836722563256e-05,
"loss": 1.2022,
"step": 290
},
{
"epoch": 0.05,
"grad_norm": 1.0117149353027344,
"learning_rate": 4.996981908813664e-05,
"loss": 0.8032,
"step": 295
},
{
"epoch": 0.05,
"grad_norm": 0.8861119747161865,
"learning_rate": 4.996878401274687e-05,
"loss": 1.0651,
"step": 300
},
{
"epoch": 0.05,
"eval_loss": 0.8281473517417908,
"eval_runtime": 96.5283,
"eval_samples_per_second": 7.221,
"eval_steps_per_second": 7.221,
"step": 300
},
{
"epoch": 0.05,
"grad_norm": 0.8583046197891235,
"learning_rate": 4.996773149711693e-05,
"loss": 0.8784,
"step": 305
},
{
"epoch": 0.05,
"grad_norm": 2.5717499256134033,
"learning_rate": 4.9966661541981984e-05,
"loss": 0.8395,
"step": 310
},
{
"epoch": 0.05,
"grad_norm": 0.982342004776001,
"learning_rate": 4.9965574148089376e-05,
"loss": 0.9869,
"step": 315
},
{
"epoch": 0.05,
"grad_norm": 0.9000777006149292,
"learning_rate": 4.9964469316198633e-05,
"loss": 0.8435,
"step": 320
},
{
"epoch": 0.05,
"grad_norm": 0.8733209371566772,
"learning_rate": 4.9963347047081464e-05,
"loss": 0.7281,
"step": 325
},
{
"epoch": 0.05,
"grad_norm": 3.323739767074585,
"learning_rate": 4.9962207341521746e-05,
"loss": 1.1013,
"step": 330
},
{
"epoch": 0.05,
"grad_norm": 1.7102876901626587,
"learning_rate": 4.996105020031554e-05,
"loss": 0.8276,
"step": 335
},
{
"epoch": 0.05,
"grad_norm": 0.9196123480796814,
"learning_rate": 4.995987562427109e-05,
"loss": 0.8274,
"step": 340
},
{
"epoch": 0.06,
"grad_norm": 1.210099458694458,
"learning_rate": 4.995868361420883e-05,
"loss": 1.3257,
"step": 345
},
{
"epoch": 0.06,
"grad_norm": 0.8923581838607788,
"learning_rate": 4.9957474170961335e-05,
"loss": 0.6815,
"step": 350
},
{
"epoch": 0.06,
"grad_norm": 0.9576735496520996,
"learning_rate": 4.9956247295373396e-05,
"loss": 1.23,
"step": 355
},
{
"epoch": 0.06,
"grad_norm": 1.3774089813232422,
"learning_rate": 4.995500298830196e-05,
"loss": 1.0556,
"step": 360
},
{
"epoch": 0.06,
"grad_norm": 1.1523677110671997,
"learning_rate": 4.995374125061614e-05,
"loss": 1.1787,
"step": 365
},
{
"epoch": 0.06,
"grad_norm": 0.8310608863830566,
"learning_rate": 4.9952462083197246e-05,
"loss": 0.8525,
"step": 370
},
{
"epoch": 0.06,
"grad_norm": 0.9814196825027466,
"learning_rate": 4.9951165486938765e-05,
"loss": 0.8522,
"step": 375
},
{
"epoch": 0.06,
"grad_norm": 0.9878122210502625,
"learning_rate": 4.994985146274633e-05,
"loss": 0.6618,
"step": 380
},
{
"epoch": 0.06,
"grad_norm": 1.2652586698532104,
"learning_rate": 4.994852001153777e-05,
"loss": 1.0489,
"step": 385
},
{
"epoch": 0.06,
"grad_norm": 1.2940975427627563,
"learning_rate": 4.994717113424307e-05,
"loss": 1.104,
"step": 390
},
{
"epoch": 0.06,
"grad_norm": 0.9636249542236328,
"learning_rate": 4.99458048318044e-05,
"loss": 0.9228,
"step": 395
},
{
"epoch": 0.06,
"grad_norm": 0.8122813105583191,
"learning_rate": 4.994442110517611e-05,
"loss": 0.9209,
"step": 400
},
{
"epoch": 0.06,
"eval_loss": 0.8184689879417419,
"eval_runtime": 96.4572,
"eval_samples_per_second": 7.226,
"eval_steps_per_second": 7.226,
"step": 400
},
{
"epoch": 0.06,
"grad_norm": 0.8742052912712097,
"learning_rate": 4.99430199553247e-05,
"loss": 0.9608,
"step": 405
},
{
"epoch": 0.07,
"grad_norm": 0.5679522752761841,
"learning_rate": 4.9941601383228835e-05,
"loss": 0.5963,
"step": 410
},
{
"epoch": 0.07,
"grad_norm": 1.0234627723693848,
"learning_rate": 4.994016538987938e-05,
"loss": 0.8642,
"step": 415
},
{
"epoch": 0.07,
"grad_norm": 0.8581897616386414,
"learning_rate": 4.993871197627934e-05,
"loss": 0.8993,
"step": 420
},
{
"epoch": 0.07,
"grad_norm": 1.4666485786437988,
"learning_rate": 4.9937241143443904e-05,
"loss": 0.8565,
"step": 425
},
{
"epoch": 0.07,
"grad_norm": 1.1166578531265259,
"learning_rate": 4.993575289240041e-05,
"loss": 0.881,
"step": 430
},
{
"epoch": 0.07,
"grad_norm": 1.303992748260498,
"learning_rate": 4.9934247224188393e-05,
"loss": 0.9962,
"step": 435
},
{
"epoch": 0.07,
"grad_norm": 0.9011989235877991,
"learning_rate": 4.993272413985952e-05,
"loss": 0.9316,
"step": 440
},
{
"epoch": 0.07,
"grad_norm": 0.8321458101272583,
"learning_rate": 4.993118364047764e-05,
"loss": 0.7889,
"step": 445
},
{
"epoch": 0.07,
"grad_norm": 0.7780352234840393,
"learning_rate": 4.992962572711877e-05,
"loss": 0.8287,
"step": 450
},
{
"epoch": 0.07,
"grad_norm": 0.9090210199356079,
"learning_rate": 4.992805040087108e-05,
"loss": 0.7018,
"step": 455
},
{
"epoch": 0.07,
"grad_norm": 0.8694137334823608,
"learning_rate": 4.9926457662834906e-05,
"loss": 0.8484,
"step": 460
},
{
"epoch": 0.07,
"grad_norm": 0.6327371001243591,
"learning_rate": 4.992484751412274e-05,
"loss": 0.716,
"step": 465
},
{
"epoch": 0.08,
"grad_norm": 1.200668215751648,
"learning_rate": 4.9923219955859254e-05,
"loss": 0.9525,
"step": 470
},
{
"epoch": 0.08,
"grad_norm": 0.8530198931694031,
"learning_rate": 4.9921574989181266e-05,
"loss": 0.744,
"step": 475
},
{
"epoch": 0.08,
"grad_norm": 1.168479323387146,
"learning_rate": 4.991991261523775e-05,
"loss": 0.729,
"step": 480
},
{
"epoch": 0.08,
"grad_norm": 0.9499714970588684,
"learning_rate": 4.9918232835189834e-05,
"loss": 0.7725,
"step": 485
},
{
"epoch": 0.08,
"grad_norm": 0.8434467911720276,
"learning_rate": 4.991653565021084e-05,
"loss": 1.1558,
"step": 490
},
{
"epoch": 0.08,
"grad_norm": 0.7665804624557495,
"learning_rate": 4.99148210614862e-05,
"loss": 1.0208,
"step": 495
},
{
"epoch": 0.08,
"grad_norm": 0.5782546401023865,
"learning_rate": 4.991308907021353e-05,
"loss": 0.8306,
"step": 500
},
{
"epoch": 0.08,
"eval_loss": 0.8132078051567078,
"eval_runtime": 96.433,
"eval_samples_per_second": 7.228,
"eval_steps_per_second": 7.228,
"step": 500
},
{
"epoch": 0.08,
"grad_norm": 1.0821778774261475,
"learning_rate": 4.9911339677602584e-05,
"loss": 0.9503,
"step": 505
},
{
"epoch": 0.08,
"grad_norm": 0.5409029126167297,
"learning_rate": 4.99095728848753e-05,
"loss": 0.8586,
"step": 510
},
{
"epoch": 0.08,
"grad_norm": 0.9011789560317993,
"learning_rate": 4.990778869326575e-05,
"loss": 0.7981,
"step": 515
},
{
"epoch": 0.08,
"grad_norm": 1.0092263221740723,
"learning_rate": 4.990598710402013e-05,
"loss": 1.0174,
"step": 520
},
{
"epoch": 0.08,
"grad_norm": 1.4362307786941528,
"learning_rate": 4.9904168118396844e-05,
"loss": 0.8373,
"step": 525
},
{
"epoch": 0.08,
"grad_norm": 2.1772639751434326,
"learning_rate": 4.9902331737666414e-05,
"loss": 0.9599,
"step": 530
},
{
"epoch": 0.09,
"grad_norm": 0.9610542058944702,
"learning_rate": 4.990047796311151e-05,
"loss": 0.6895,
"step": 535
},
{
"epoch": 0.09,
"grad_norm": 0.9922348260879517,
"learning_rate": 4.989860679602698e-05,
"loss": 0.7315,
"step": 540
},
{
"epoch": 0.09,
"grad_norm": 1.2409151792526245,
"learning_rate": 4.9896718237719785e-05,
"loss": 0.8574,
"step": 545
},
{
"epoch": 0.09,
"grad_norm": 1.016333818435669,
"learning_rate": 4.9894812289509046e-05,
"loss": 1.1248,
"step": 550
},
{
"epoch": 0.09,
"grad_norm": 0.9131489396095276,
"learning_rate": 4.989288895272604e-05,
"loss": 0.9847,
"step": 555
},
{
"epoch": 0.09,
"grad_norm": 1.215469479560852,
"learning_rate": 4.989094822871419e-05,
"loss": 0.912,
"step": 560
},
{
"epoch": 0.09,
"grad_norm": 1.0536105632781982,
"learning_rate": 4.988899011882903e-05,
"loss": 0.8425,
"step": 565
},
{
"epoch": 0.09,
"grad_norm": 1.9705311059951782,
"learning_rate": 4.988701462443829e-05,
"loss": 0.9385,
"step": 570
},
{
"epoch": 0.09,
"grad_norm": 1.2488442659378052,
"learning_rate": 4.98850217469218e-05,
"loss": 0.7865,
"step": 575
},
{
"epoch": 0.09,
"grad_norm": 1.7318600416183472,
"learning_rate": 4.988301148767157e-05,
"loss": 0.8231,
"step": 580
},
{
"epoch": 0.09,
"grad_norm": 0.8247858881950378,
"learning_rate": 4.9880983848091704e-05,
"loss": 0.8553,
"step": 585
},
{
"epoch": 0.09,
"grad_norm": 0.858172595500946,
"learning_rate": 4.987893882959849e-05,
"loss": 1.3952,
"step": 590
},
{
"epoch": 0.09,
"grad_norm": 1.2286418676376343,
"learning_rate": 4.987687643362033e-05,
"loss": 0.837,
"step": 595
},
{
"epoch": 0.1,
"grad_norm": 1.034350872039795,
"learning_rate": 4.9874796661597765e-05,
"loss": 0.9175,
"step": 600
},
{
"epoch": 0.1,
"eval_loss": 0.8063747882843018,
"eval_runtime": 96.4224,
"eval_samples_per_second": 7.229,
"eval_steps_per_second": 7.229,
"step": 600
},
{
"epoch": 0.1,
"grad_norm": 0.7192366123199463,
"learning_rate": 4.987269951498348e-05,
"loss": 0.8563,
"step": 605
},
{
"epoch": 0.1,
"grad_norm": 1.2645854949951172,
"learning_rate": 4.98705849952423e-05,
"loss": 0.6663,
"step": 610
},
{
"epoch": 0.1,
"grad_norm": 1.0610381364822388,
"learning_rate": 4.9868453103851176e-05,
"loss": 0.8452,
"step": 615
},
{
"epoch": 0.1,
"grad_norm": 0.8550002574920654,
"learning_rate": 4.986630384229919e-05,
"loss": 0.8894,
"step": 620
},
{
"epoch": 0.1,
"grad_norm": 0.7490519285202026,
"learning_rate": 4.986413721208757e-05,
"loss": 0.9106,
"step": 625
},
{
"epoch": 0.1,
"grad_norm": 0.557860255241394,
"learning_rate": 4.986195321472965e-05,
"loss": 0.685,
"step": 630
},
{
"epoch": 0.1,
"grad_norm": 0.7450752258300781,
"learning_rate": 4.9859751851750934e-05,
"loss": 0.8472,
"step": 635
},
{
"epoch": 0.1,
"grad_norm": 1.176376461982727,
"learning_rate": 4.985753312468903e-05,
"loss": 1.0197,
"step": 640
},
{
"epoch": 0.1,
"grad_norm": 1.0625300407409668,
"learning_rate": 4.985529703509367e-05,
"loss": 0.9685,
"step": 645
},
{
"epoch": 0.1,
"grad_norm": 0.8808372616767883,
"learning_rate": 4.985304358452672e-05,
"loss": 0.8612,
"step": 650
},
{
"epoch": 0.1,
"grad_norm": 0.8110201954841614,
"learning_rate": 4.985077277456218e-05,
"loss": 0.8401,
"step": 655
},
{
"epoch": 0.11,
"grad_norm": 0.9364888072013855,
"learning_rate": 4.984848460678618e-05,
"loss": 0.6197,
"step": 660
},
{
"epoch": 0.11,
"grad_norm": 1.0113518238067627,
"learning_rate": 4.984617908279694e-05,
"loss": 0.9889,
"step": 665
},
{
"epoch": 0.11,
"grad_norm": 1.1148868799209595,
"learning_rate": 4.984385620420485e-05,
"loss": 0.9558,
"step": 670
},
{
"epoch": 0.11,
"grad_norm": 0.9506175518035889,
"learning_rate": 4.984151597263238e-05,
"loss": 0.7323,
"step": 675
},
{
"epoch": 0.11,
"grad_norm": 1.0044193267822266,
"learning_rate": 4.983915838971415e-05,
"loss": 0.7504,
"step": 680
},
{
"epoch": 0.11,
"grad_norm": 2.2674214839935303,
"learning_rate": 4.9836783457096875e-05,
"loss": 1.032,
"step": 685
},
{
"epoch": 0.11,
"grad_norm": 1.4945333003997803,
"learning_rate": 4.983439117643942e-05,
"loss": 1.0359,
"step": 690
},
{
"epoch": 0.11,
"grad_norm": 0.9860715866088867,
"learning_rate": 4.9831981549412744e-05,
"loss": 1.1152,
"step": 695
},
{
"epoch": 0.11,
"grad_norm": 0.8287227153778076,
"learning_rate": 4.982955457769992e-05,
"loss": 0.8157,
"step": 700
},
{
"epoch": 0.11,
"eval_loss": 0.8022791743278503,
"eval_runtime": 96.5324,
"eval_samples_per_second": 7.22,
"eval_steps_per_second": 7.22,
"step": 700
},
{
"epoch": 0.11,
"grad_norm": 0.9216273427009583,
"learning_rate": 4.9827110262996144e-05,
"loss": 0.8395,
"step": 705
},
{
"epoch": 0.11,
"grad_norm": 0.7642357349395752,
"learning_rate": 4.982464860700874e-05,
"loss": 0.8817,
"step": 710
},
{
"epoch": 0.11,
"grad_norm": 0.8851175308227539,
"learning_rate": 4.982216961145711e-05,
"loss": 0.8558,
"step": 715
},
{
"epoch": 0.11,
"grad_norm": 0.44226109981536865,
"learning_rate": 4.98196732780728e-05,
"loss": 0.882,
"step": 720
},
{
"epoch": 0.12,
"grad_norm": 0.8005027174949646,
"learning_rate": 4.981715960859945e-05,
"loss": 0.8835,
"step": 725
},
{
"epoch": 0.12,
"grad_norm": 0.7451304793357849,
"learning_rate": 4.981462860479281e-05,
"loss": 0.8551,
"step": 730
},
{
"epoch": 0.12,
"grad_norm": 1.1069347858428955,
"learning_rate": 4.9812080268420745e-05,
"loss": 0.999,
"step": 735
},
{
"epoch": 0.12,
"grad_norm": 0.8892244100570679,
"learning_rate": 4.980951460126322e-05,
"loss": 1.012,
"step": 740
},
{
"epoch": 0.12,
"grad_norm": 0.8935977816581726,
"learning_rate": 4.9806931605112305e-05,
"loss": 0.9911,
"step": 745
},
{
"epoch": 0.12,
"grad_norm": 0.8456961512565613,
"learning_rate": 4.9804331281772176e-05,
"loss": 0.7595,
"step": 750
},
{
"epoch": 0.12,
"grad_norm": 0.78443443775177,
"learning_rate": 4.980171363305911e-05,
"loss": 0.8308,
"step": 755
},
{
"epoch": 0.12,
"grad_norm": 1.0028038024902344,
"learning_rate": 4.979907866080149e-05,
"loss": 0.9637,
"step": 760
},
{
"epoch": 0.12,
"grad_norm": 1.1801577806472778,
"learning_rate": 4.9796426366839786e-05,
"loss": 0.6159,
"step": 765
},
{
"epoch": 0.12,
"grad_norm": 0.8370681405067444,
"learning_rate": 4.979375675302659e-05,
"loss": 0.9276,
"step": 770
},
{
"epoch": 0.12,
"grad_norm": 0.8605382442474365,
"learning_rate": 4.979106982122658e-05,
"loss": 1.1077,
"step": 775
},
{
"epoch": 0.12,
"grad_norm": 0.7788259387016296,
"learning_rate": 4.978836557331652e-05,
"loss": 0.8172,
"step": 780
},
{
"epoch": 0.13,
"grad_norm": 1.4312686920166016,
"learning_rate": 4.978564401118528e-05,
"loss": 0.8759,
"step": 785
},
{
"epoch": 0.13,
"grad_norm": 0.9109662175178528,
"learning_rate": 4.978290513673381e-05,
"loss": 0.947,
"step": 790
},
{
"epoch": 0.13,
"grad_norm": 1.1819065809249878,
"learning_rate": 4.9780148951875195e-05,
"loss": 0.7364,
"step": 795
},
{
"epoch": 0.13,
"grad_norm": 0.9400575160980225,
"learning_rate": 4.977737545853455e-05,
"loss": 0.9469,
"step": 800
},
{
"epoch": 0.13,
"eval_loss": 0.7995806932449341,
"eval_runtime": 96.5877,
"eval_samples_per_second": 7.216,
"eval_steps_per_second": 7.216,
"step": 800
},
{
"epoch": 0.13,
"grad_norm": 1.693812370300293,
"learning_rate": 4.9774584658649126e-05,
"loss": 0.9433,
"step": 805
},
{
"epoch": 0.13,
"grad_norm": 1.0892895460128784,
"learning_rate": 4.9771776554168234e-05,
"loss": 0.7027,
"step": 810
},
{
"epoch": 0.13,
"grad_norm": 0.9118362665176392,
"learning_rate": 4.976895114705329e-05,
"loss": 0.9468,
"step": 815
},
{
"epoch": 0.13,
"grad_norm": 0.8032681345939636,
"learning_rate": 4.976610843927779e-05,
"loss": 0.7927,
"step": 820
},
{
"epoch": 0.13,
"grad_norm": 1.168225646018982,
"learning_rate": 4.976324843282732e-05,
"loss": 0.9673,
"step": 825
},
{
"epoch": 0.13,
"grad_norm": 1.077602744102478,
"learning_rate": 4.976037112969953e-05,
"loss": 0.9156,
"step": 830
},
{
"epoch": 0.13,
"grad_norm": 0.8643108606338501,
"learning_rate": 4.9757476531904165e-05,
"loss": 0.6999,
"step": 835
},
{
"epoch": 0.13,
"grad_norm": 0.933397650718689,
"learning_rate": 4.975456464146306e-05,
"loss": 0.8828,
"step": 840
},
{
"epoch": 0.13,
"grad_norm": 0.7036295533180237,
"learning_rate": 4.975163546041011e-05,
"loss": 0.8709,
"step": 845
},
{
"epoch": 0.14,
"grad_norm": 0.5974694490432739,
"learning_rate": 4.974868899079128e-05,
"loss": 0.7594,
"step": 850
},
{
"epoch": 0.14,
"grad_norm": 0.7244943380355835,
"learning_rate": 4.974572523466465e-05,
"loss": 0.8714,
"step": 855
},
{
"epoch": 0.14,
"grad_norm": 0.5783522725105286,
"learning_rate": 4.9742744194100345e-05,
"loss": 0.8941,
"step": 860
},
{
"epoch": 0.14,
"grad_norm": 0.7480617761611938,
"learning_rate": 4.973974587118055e-05,
"loss": 0.9798,
"step": 865
},
{
"epoch": 0.14,
"grad_norm": 0.7548874020576477,
"learning_rate": 4.973673026799956e-05,
"loss": 0.7767,
"step": 870
},
{
"epoch": 0.14,
"grad_norm": 0.7075071930885315,
"learning_rate": 4.97336973866637e-05,
"loss": 0.7779,
"step": 875
},
{
"epoch": 0.14,
"grad_norm": 0.7042987942695618,
"learning_rate": 4.97306472292914e-05,
"loss": 0.8249,
"step": 880
},
{
"epoch": 0.14,
"grad_norm": 1.0242459774017334,
"learning_rate": 4.972757979801313e-05,
"loss": 0.9223,
"step": 885
},
{
"epoch": 0.14,
"grad_norm": 0.6138095259666443,
"learning_rate": 4.9724495094971436e-05,
"loss": 0.9842,
"step": 890
},
{
"epoch": 0.14,
"grad_norm": 0.7905042767524719,
"learning_rate": 4.9721393122320925e-05,
"loss": 0.8738,
"step": 895
},
{
"epoch": 0.14,
"grad_norm": 0.9658048748970032,
"learning_rate": 4.9718273882228265e-05,
"loss": 0.8872,
"step": 900
},
{
"epoch": 0.14,
"eval_loss": 0.7954564690589905,
"eval_runtime": 96.643,
"eval_samples_per_second": 7.212,
"eval_steps_per_second": 7.212,
"step": 900
},
{
"epoch": 0.14,
"grad_norm": 0.8425014019012451,
"learning_rate": 4.97151373768722e-05,
"loss": 0.778,
"step": 905
},
{
"epoch": 0.15,
"grad_norm": 0.5527231693267822,
"learning_rate": 4.971198360844351e-05,
"loss": 0.8332,
"step": 910
},
{
"epoch": 0.15,
"grad_norm": 0.7870334386825562,
"learning_rate": 4.9708812579145056e-05,
"loss": 0.9265,
"step": 915
},
{
"epoch": 0.15,
"grad_norm": 0.9935321807861328,
"learning_rate": 4.970562429119173e-05,
"loss": 0.7243,
"step": 920
},
{
"epoch": 0.15,
"grad_norm": 0.9546892046928406,
"learning_rate": 4.970241874681051e-05,
"loss": 0.9908,
"step": 925
},
{
"epoch": 0.15,
"grad_norm": 0.7340118885040283,
"learning_rate": 4.969919594824039e-05,
"loss": 0.7932,
"step": 930
},
{
"epoch": 0.15,
"grad_norm": 5.1686015129089355,
"learning_rate": 4.9695955897732453e-05,
"loss": 0.9842,
"step": 935
},
{
"epoch": 0.15,
"grad_norm": 0.9721456170082092,
"learning_rate": 4.9692698597549815e-05,
"loss": 0.9271,
"step": 940
},
{
"epoch": 0.15,
"grad_norm": 0.6477334499359131,
"learning_rate": 4.9689424049967623e-05,
"loss": 0.934,
"step": 945
},
{
"epoch": 0.15,
"grad_norm": 1.0759055614471436,
"learning_rate": 4.968613225727311e-05,
"loss": 1.0465,
"step": 950
},
{
"epoch": 0.15,
"grad_norm": 0.7222158908843994,
"learning_rate": 4.968282322176552e-05,
"loss": 0.7732,
"step": 955
},
{
"epoch": 0.15,
"grad_norm": 0.8591343760490417,
"learning_rate": 4.9679496945756155e-05,
"loss": 0.9062,
"step": 960
},
{
"epoch": 0.15,
"grad_norm": 1.8495111465454102,
"learning_rate": 4.967615343156837e-05,
"loss": 0.8861,
"step": 965
},
{
"epoch": 0.15,
"grad_norm": 0.6847331523895264,
"learning_rate": 4.967279268153753e-05,
"loss": 0.8001,
"step": 970
},
{
"epoch": 0.16,
"grad_norm": 0.690113365650177,
"learning_rate": 4.9669414698011074e-05,
"loss": 0.7378,
"step": 975
},
{
"epoch": 0.16,
"grad_norm": 0.8349626064300537,
"learning_rate": 4.9666019483348456e-05,
"loss": 0.7193,
"step": 980
},
{
"epoch": 0.16,
"grad_norm": 0.6444108486175537,
"learning_rate": 4.966260703992116e-05,
"loss": 0.8729,
"step": 985
},
{
"epoch": 0.16,
"grad_norm": 0.9515655040740967,
"learning_rate": 4.965917737011274e-05,
"loss": 0.7532,
"step": 990
},
{
"epoch": 0.16,
"grad_norm": 0.8138986229896545,
"learning_rate": 4.965573047631873e-05,
"loss": 1.0124,
"step": 995
},
{
"epoch": 0.16,
"grad_norm": 1.0182080268859863,
"learning_rate": 4.9652266360946745e-05,
"loss": 0.8842,
"step": 1000
},
{
"epoch": 0.16,
"eval_loss": 0.7912728190422058,
"eval_runtime": 96.5004,
"eval_samples_per_second": 7.223,
"eval_steps_per_second": 7.223,
"step": 1000
},
{
"epoch": 0.16,
"grad_norm": 0.9665297269821167,
"learning_rate": 4.96487850264164e-05,
"loss": 1.0155,
"step": 1005
},
{
"epoch": 0.16,
"grad_norm": 1.1356585025787354,
"learning_rate": 4.964528647515933e-05,
"loss": 0.8705,
"step": 1010
},
{
"epoch": 0.16,
"grad_norm": 0.5548833608627319,
"learning_rate": 4.9641770709619234e-05,
"loss": 0.9634,
"step": 1015
},
{
"epoch": 0.16,
"grad_norm": 0.8028444647789001,
"learning_rate": 4.9638237732251794e-05,
"loss": 0.8722,
"step": 1020
},
{
"epoch": 0.16,
"grad_norm": 0.934234082698822,
"learning_rate": 4.9634687545524724e-05,
"loss": 0.9731,
"step": 1025
},
{
"epoch": 0.16,
"grad_norm": 0.7293463349342346,
"learning_rate": 4.963112015191778e-05,
"loss": 1.0237,
"step": 1030
},
{
"epoch": 0.17,
"grad_norm": 0.6442769169807434,
"learning_rate": 4.962753555392271e-05,
"loss": 1.1331,
"step": 1035
},
{
"epoch": 0.17,
"grad_norm": 0.7877534031867981,
"learning_rate": 4.962393375404331e-05,
"loss": 1.0737,
"step": 1040
},
{
"epoch": 0.17,
"grad_norm": 0.5739997625350952,
"learning_rate": 4.9620314754795343e-05,
"loss": 0.8836,
"step": 1045
},
{
"epoch": 0.17,
"grad_norm": 0.7318402528762817,
"learning_rate": 4.9616678558706634e-05,
"loss": 0.9981,
"step": 1050
},
{
"epoch": 0.17,
"grad_norm": 0.5463365316390991,
"learning_rate": 4.961302516831699e-05,
"loss": 0.7336,
"step": 1055
},
{
"epoch": 0.17,
"grad_norm": 0.7839176654815674,
"learning_rate": 4.960935458617824e-05,
"loss": 1.025,
"step": 1060
},
{
"epoch": 0.17,
"grad_norm": 0.7076404690742493,
"learning_rate": 4.9605666814854225e-05,
"loss": 0.833,
"step": 1065
},
{
"epoch": 0.17,
"grad_norm": 0.732940673828125,
"learning_rate": 4.960196185692077e-05,
"loss": 0.5103,
"step": 1070
},
{
"epoch": 0.17,
"grad_norm": 0.7256388068199158,
"learning_rate": 4.959823971496574e-05,
"loss": 0.8617,
"step": 1075
},
{
"epoch": 0.17,
"grad_norm": 1.1714242696762085,
"learning_rate": 4.959450039158898e-05,
"loss": 1.0345,
"step": 1080
},
{
"epoch": 0.17,
"grad_norm": 0.5849193930625916,
"learning_rate": 4.9590743889402325e-05,
"loss": 0.729,
"step": 1085
},
{
"epoch": 0.17,
"grad_norm": 0.6283109784126282,
"learning_rate": 4.958697021102963e-05,
"loss": 0.8527,
"step": 1090
},
{
"epoch": 0.17,
"grad_norm": 0.6387770175933838,
"learning_rate": 4.9583179359106746e-05,
"loss": 0.7411,
"step": 1095
},
{
"epoch": 0.18,
"grad_norm": 0.5853758454322815,
"learning_rate": 4.957937133628151e-05,
"loss": 0.7909,
"step": 1100
},
{
"epoch": 0.18,
"eval_loss": 0.7863278985023499,
"eval_runtime": 96.3784,
"eval_samples_per_second": 7.232,
"eval_steps_per_second": 7.232,
"step": 1100
},
{
"epoch": 0.18,
"grad_norm": 0.9301708936691284,
"learning_rate": 4.9575546145213755e-05,
"loss": 0.7149,
"step": 1105
},
{
"epoch": 0.18,
"grad_norm": 1.125088095664978,
"learning_rate": 4.9571703788575314e-05,
"loss": 0.8034,
"step": 1110
},
{
"epoch": 0.18,
"grad_norm": 1.0697988271713257,
"learning_rate": 4.956784426905e-05,
"loss": 0.8874,
"step": 1115
},
{
"epoch": 0.18,
"grad_norm": 0.7094873189926147,
"learning_rate": 4.956396758933361e-05,
"loss": 0.6612,
"step": 1120
},
{
"epoch": 0.18,
"grad_norm": 0.8048680424690247,
"learning_rate": 4.956007375213393e-05,
"loss": 0.9558,
"step": 1125
},
{
"epoch": 0.18,
"grad_norm": 0.8820949196815491,
"learning_rate": 4.9556162760170756e-05,
"loss": 0.9442,
"step": 1130
},
{
"epoch": 0.18,
"grad_norm": 0.7214958071708679,
"learning_rate": 4.955223461617583e-05,
"loss": 0.8392,
"step": 1135
},
{
"epoch": 0.18,
"grad_norm": 0.8364250063896179,
"learning_rate": 4.954828932289288e-05,
"loss": 0.9834,
"step": 1140
},
{
"epoch": 0.18,
"grad_norm": 0.8735854625701904,
"learning_rate": 4.954432688307764e-05,
"loss": 0.8817,
"step": 1145
},
{
"epoch": 0.18,
"grad_norm": 0.810013473033905,
"learning_rate": 4.9540347299497805e-05,
"loss": 0.7723,
"step": 1150
},
{
"epoch": 0.18,
"grad_norm": 0.8791002035140991,
"learning_rate": 4.953635057493302e-05,
"loss": 0.706,
"step": 1155
},
{
"epoch": 0.19,
"grad_norm": 0.7556783556938171,
"learning_rate": 4.953233671217493e-05,
"loss": 0.8145,
"step": 1160
},
{
"epoch": 0.19,
"grad_norm": 1.3251086473464966,
"learning_rate": 4.952830571402716e-05,
"loss": 0.8413,
"step": 1165
},
{
"epoch": 0.19,
"grad_norm": 0.8531173467636108,
"learning_rate": 4.952425758330527e-05,
"loss": 0.8236,
"step": 1170
},
{
"epoch": 0.19,
"grad_norm": 1.0738744735717773,
"learning_rate": 4.952019232283681e-05,
"loss": 0.8357,
"step": 1175
},
{
"epoch": 0.19,
"grad_norm": 0.7908213138580322,
"learning_rate": 4.9516109935461306e-05,
"loss": 0.6165,
"step": 1180
},
{
"epoch": 0.19,
"grad_norm": 0.9802565574645996,
"learning_rate": 4.951201042403021e-05,
"loss": 0.7203,
"step": 1185
},
{
"epoch": 0.19,
"grad_norm": 0.7866708636283875,
"learning_rate": 4.9507893791406974e-05,
"loss": 0.8479,
"step": 1190
},
{
"epoch": 0.19,
"grad_norm": 0.6721138954162598,
"learning_rate": 4.950376004046698e-05,
"loss": 0.8871,
"step": 1195
},
{
"epoch": 0.19,
"grad_norm": 1.1981366872787476,
"learning_rate": 4.9499609174097574e-05,
"loss": 0.8196,
"step": 1200
},
{
"epoch": 0.19,
"eval_loss": 0.7843652367591858,
"eval_runtime": 96.5411,
"eval_samples_per_second": 7.22,
"eval_steps_per_second": 7.22,
"step": 1200
},
{
"epoch": 0.19,
"grad_norm": 0.7013841867446899,
"learning_rate": 4.9495441195198064e-05,
"loss": 1.0009,
"step": 1205
},
{
"epoch": 0.19,
"grad_norm": 0.8476290702819824,
"learning_rate": 4.949125610667972e-05,
"loss": 0.5127,
"step": 1210
},
{
"epoch": 0.19,
"grad_norm": 0.7680797576904297,
"learning_rate": 4.9487053911465735e-05,
"loss": 0.7003,
"step": 1215
},
{
"epoch": 0.19,
"grad_norm": 0.9771925806999207,
"learning_rate": 4.948283461249127e-05,
"loss": 1.1135,
"step": 1220
},
{
"epoch": 0.2,
"grad_norm": 1.4247405529022217,
"learning_rate": 4.947859821270342e-05,
"loss": 0.8253,
"step": 1225
},
{
"epoch": 0.2,
"grad_norm": 1.184887409210205,
"learning_rate": 4.947434471506125e-05,
"loss": 1.1208,
"step": 1230
},
{
"epoch": 0.2,
"grad_norm": 0.7579745054244995,
"learning_rate": 4.9470074122535745e-05,
"loss": 1.1363,
"step": 1235
},
{
"epoch": 0.2,
"grad_norm": 0.8529625535011292,
"learning_rate": 4.9465786438109826e-05,
"loss": 0.8699,
"step": 1240
},
{
"epoch": 0.2,
"grad_norm": 1.810576319694519,
"learning_rate": 4.9461481664778374e-05,
"loss": 1.0166,
"step": 1245
},
{
"epoch": 0.2,
"grad_norm": 0.8605110049247742,
"learning_rate": 4.9457159805548187e-05,
"loss": 0.9427,
"step": 1250
},
{
"epoch": 0.2,
"grad_norm": 0.59971684217453,
"learning_rate": 4.945282086343801e-05,
"loss": 0.6536,
"step": 1255
},
{
"epoch": 0.2,
"grad_norm": 1.0233818292617798,
"learning_rate": 4.9448464841478506e-05,
"loss": 0.9505,
"step": 1260
},
{
"epoch": 0.2,
"grad_norm": 0.8945149779319763,
"learning_rate": 4.9444091742712293e-05,
"loss": 0.8416,
"step": 1265
},
{
"epoch": 0.2,
"grad_norm": 0.702805757522583,
"learning_rate": 4.9439701570193886e-05,
"loss": 0.9419,
"step": 1270
},
{
"epoch": 0.2,
"grad_norm": 0.7464181184768677,
"learning_rate": 4.9435294326989745e-05,
"loss": 0.7972,
"step": 1275
},
{
"epoch": 0.2,
"grad_norm": 1.1765002012252808,
"learning_rate": 4.943175624360097e-05,
"loss": 0.9914,
"step": 1280
},
{
"epoch": 0.21,
"grad_norm": 0.6549853682518005,
"learning_rate": 4.9427318280928034e-05,
"loss": 0.8924,
"step": 1285
},
{
"epoch": 0.21,
"grad_norm": 0.5978650450706482,
"learning_rate": 4.942286325621888e-05,
"loss": 0.6224,
"step": 1290
},
{
"epoch": 0.21,
"grad_norm": 0.7752617597579956,
"learning_rate": 4.941839117258523e-05,
"loss": 0.8666,
"step": 1295
},
{
"epoch": 0.21,
"grad_norm": 0.6919072866439819,
"learning_rate": 4.941390203315078e-05,
"loss": 0.9341,
"step": 1300
},
{
"epoch": 0.21,
"eval_loss": 0.7824844717979431,
"eval_runtime": 96.8874,
"eval_samples_per_second": 7.194,
"eval_steps_per_second": 7.194,
"step": 1300
},
{
"epoch": 0.21,
"grad_norm": 0.7222729325294495,
"learning_rate": 4.94093958410511e-05,
"loss": 0.9925,
"step": 1305
},
{
"epoch": 0.21,
"grad_norm": 0.9575716853141785,
"learning_rate": 4.9404872599433686e-05,
"loss": 0.8623,
"step": 1310
},
{
"epoch": 0.21,
"grad_norm": 0.7721400260925293,
"learning_rate": 4.940033231145793e-05,
"loss": 1.0061,
"step": 1315
},
{
"epoch": 0.21,
"grad_norm": 0.7019990682601929,
"learning_rate": 4.9395774980295165e-05,
"loss": 0.8697,
"step": 1320
},
{
"epoch": 0.21,
"grad_norm": 0.7828916907310486,
"learning_rate": 4.939120060912858e-05,
"loss": 1.0066,
"step": 1325
},
{
"epoch": 0.21,
"grad_norm": 1.0238871574401855,
"learning_rate": 4.93866092011533e-05,
"loss": 1.0285,
"step": 1330
},
{
"epoch": 0.21,
"grad_norm": 0.48669734597206116,
"learning_rate": 4.938200075957634e-05,
"loss": 0.7454,
"step": 1335
},
{
"epoch": 0.21,
"grad_norm": 0.8834619522094727,
"learning_rate": 4.93773752876166e-05,
"loss": 0.9998,
"step": 1340
},
{
"epoch": 0.21,
"grad_norm": 0.6462609767913818,
"learning_rate": 4.9372732788504905e-05,
"loss": 0.7278,
"step": 1345
},
{
"epoch": 0.22,
"grad_norm": 0.7309257388114929,
"learning_rate": 4.936807326548395e-05,
"loss": 0.7301,
"step": 1350
},
{
"epoch": 0.22,
"grad_norm": 0.8515027165412903,
"learning_rate": 4.936339672180833e-05,
"loss": 0.8307,
"step": 1355
},
{
"epoch": 0.22,
"grad_norm": 0.913206934928894,
"learning_rate": 4.935870316074451e-05,
"loss": 0.9467,
"step": 1360
},
{
"epoch": 0.22,
"grad_norm": 0.6705841422080994,
"learning_rate": 4.935399258557088e-05,
"loss": 0.7124,
"step": 1365
},
{
"epoch": 0.22,
"grad_norm": 0.676695704460144,
"learning_rate": 4.934926499957767e-05,
"loss": 0.9318,
"step": 1370
},
{
"epoch": 0.22,
"grad_norm": 1.0529104471206665,
"learning_rate": 4.934452040606703e-05,
"loss": 1.0307,
"step": 1375
},
{
"epoch": 0.22,
"grad_norm": 0.7150225639343262,
"learning_rate": 4.933975880835296e-05,
"loss": 0.8718,
"step": 1380
},
{
"epoch": 0.22,
"grad_norm": 0.7180047035217285,
"learning_rate": 4.933498020976135e-05,
"loss": 0.7515,
"step": 1385
},
{
"epoch": 0.22,
"grad_norm": 1.0961759090423584,
"learning_rate": 4.933018461362997e-05,
"loss": 0.8797,
"step": 1390
},
{
"epoch": 0.22,
"grad_norm": 0.830609142780304,
"learning_rate": 4.9325372023308446e-05,
"loss": 0.6927,
"step": 1395
},
{
"epoch": 0.22,
"grad_norm": 0.5277318358421326,
"learning_rate": 4.9320542442158305e-05,
"loss": 0.8801,
"step": 1400
},
{
"epoch": 0.22,
"eval_loss": 0.7787255644798279,
"eval_runtime": 96.8812,
"eval_samples_per_second": 7.194,
"eval_steps_per_second": 7.194,
"step": 1400
},
{
"epoch": 0.22,
"grad_norm": 1.3845161199569702,
"learning_rate": 4.931569587355289e-05,
"loss": 0.8782,
"step": 1405
},
{
"epoch": 0.23,
"grad_norm": 0.8579941987991333,
"learning_rate": 4.9310832320877476e-05,
"loss": 0.713,
"step": 1410
},
{
"epoch": 0.23,
"grad_norm": 0.2643532454967499,
"learning_rate": 4.930595178752914e-05,
"loss": 0.9781,
"step": 1415
},
{
"epoch": 0.23,
"grad_norm": 0.4968445897102356,
"learning_rate": 4.930105427691685e-05,
"loss": 0.93,
"step": 1420
},
{
"epoch": 0.23,
"grad_norm": 0.9254417419433594,
"learning_rate": 4.929613979246144e-05,
"loss": 0.6353,
"step": 1425
},
{
"epoch": 0.23,
"grad_norm": 0.9814417958259583,
"learning_rate": 4.9291208337595574e-05,
"loss": 0.9672,
"step": 1430
},
{
"epoch": 0.23,
"grad_norm": 0.7159338593482971,
"learning_rate": 4.928625991576379e-05,
"loss": 0.9482,
"step": 1435
},
{
"epoch": 0.23,
"grad_norm": 0.623866617679596,
"learning_rate": 4.9281294530422476e-05,
"loss": 0.623,
"step": 1440
},
{
"epoch": 0.23,
"grad_norm": 0.8750379681587219,
"learning_rate": 4.927631218503985e-05,
"loss": 0.772,
"step": 1445
},
{
"epoch": 0.23,
"grad_norm": 0.5593128800392151,
"learning_rate": 4.9271312883096e-05,
"loss": 0.6579,
"step": 1450
},
{
"epoch": 0.23,
"grad_norm": 0.6411569714546204,
"learning_rate": 4.9266296628082834e-05,
"loss": 0.9239,
"step": 1455
},
{
"epoch": 0.23,
"grad_norm": 0.9317705631256104,
"learning_rate": 4.9261263423504135e-05,
"loss": 0.9315,
"step": 1460
},
{
"epoch": 0.23,
"grad_norm": 0.8312699198722839,
"learning_rate": 4.9256213272875486e-05,
"loss": 0.7334,
"step": 1465
},
{
"epoch": 0.23,
"grad_norm": 0.6170663833618164,
"learning_rate": 4.925114617972433e-05,
"loss": 0.8603,
"step": 1470
},
{
"epoch": 0.24,
"grad_norm": 0.7176920771598816,
"learning_rate": 4.924606214758995e-05,
"loss": 0.8738,
"step": 1475
},
{
"epoch": 0.24,
"grad_norm": 0.8957033157348633,
"learning_rate": 4.924096118002343e-05,
"loss": 0.8861,
"step": 1480
},
{
"epoch": 0.24,
"grad_norm": 0.5490685701370239,
"learning_rate": 4.923584328058772e-05,
"loss": 0.712,
"step": 1485
},
{
"epoch": 0.24,
"grad_norm": 0.7401763796806335,
"learning_rate": 4.923070845285757e-05,
"loss": 0.8118,
"step": 1490
},
{
"epoch": 0.24,
"grad_norm": 0.7380841374397278,
"learning_rate": 4.922555670041957e-05,
"loss": 0.8476,
"step": 1495
},
{
"epoch": 0.24,
"grad_norm": 1.0009427070617676,
"learning_rate": 4.922038802687212e-05,
"loss": 0.9109,
"step": 1500
},
{
"epoch": 0.24,
"eval_loss": 0.777683675289154,
"eval_runtime": 96.9147,
"eval_samples_per_second": 7.192,
"eval_steps_per_second": 7.192,
"step": 1500
},
{
"epoch": 0.24,
"grad_norm": 0.7970065474510193,
"learning_rate": 4.921520243582545e-05,
"loss": 0.616,
"step": 1505
},
{
"epoch": 0.24,
"grad_norm": 0.6530303955078125,
"learning_rate": 4.92099999309016e-05,
"loss": 0.9223,
"step": 1510
},
{
"epoch": 0.24,
"grad_norm": 0.48044708371162415,
"learning_rate": 4.9204780515734406e-05,
"loss": 0.6762,
"step": 1515
},
{
"epoch": 0.24,
"grad_norm": 0.7560244798660278,
"learning_rate": 4.919954419396956e-05,
"loss": 0.8726,
"step": 1520
},
{
"epoch": 0.24,
"grad_norm": 0.8580659031867981,
"learning_rate": 4.919429096926453e-05,
"loss": 0.7654,
"step": 1525
},
{
"epoch": 0.24,
"grad_norm": 1.1246473789215088,
"learning_rate": 4.918902084528859e-05,
"loss": 0.9123,
"step": 1530
},
{
"epoch": 0.25,
"grad_norm": 1.0745307207107544,
"learning_rate": 4.918373382572283e-05,
"loss": 0.79,
"step": 1535
},
{
"epoch": 0.25,
"grad_norm": 0.9591856598854065,
"learning_rate": 4.917842991426014e-05,
"loss": 1.1778,
"step": 1540
},
{
"epoch": 0.25,
"grad_norm": 1.0233389139175415,
"learning_rate": 4.91731091146052e-05,
"loss": 0.8827,
"step": 1545
},
{
"epoch": 0.25,
"grad_norm": 0.648965060710907,
"learning_rate": 4.91677714304745e-05,
"loss": 0.8634,
"step": 1550
},
{
"epoch": 0.25,
"grad_norm": 0.6523327231407166,
"learning_rate": 4.91624168655963e-05,
"loss": 0.9916,
"step": 1555
},
{
"epoch": 0.25,
"grad_norm": 0.8029198050498962,
"learning_rate": 4.915704542371068e-05,
"loss": 0.7867,
"step": 1560
},
{
"epoch": 0.25,
"grad_norm": 0.6397082805633545,
"learning_rate": 4.915165710856948e-05,
"loss": 0.7738,
"step": 1565
},
{
"epoch": 0.25,
"grad_norm": 0.5862845778465271,
"learning_rate": 4.914625192393636e-05,
"loss": 0.7026,
"step": 1570
},
{
"epoch": 0.25,
"grad_norm": 0.5333505868911743,
"learning_rate": 4.914082987358673e-05,
"loss": 0.8623,
"step": 1575
},
{
"epoch": 0.25,
"grad_norm": 0.5689602494239807,
"learning_rate": 4.913539096130779e-05,
"loss": 0.7619,
"step": 1580
},
{
"epoch": 0.25,
"grad_norm": 0.7333836555480957,
"learning_rate": 4.912993519089853e-05,
"loss": 0.8116,
"step": 1585
},
{
"epoch": 0.25,
"grad_norm": 0.7610496282577515,
"learning_rate": 4.91244625661697e-05,
"loss": 0.74,
"step": 1590
},
{
"epoch": 0.25,
"grad_norm": 0.6331669092178345,
"learning_rate": 4.9118973090943835e-05,
"loss": 1.0445,
"step": 1595
},
{
"epoch": 0.26,
"grad_norm": 0.7263479828834534,
"learning_rate": 4.911346676905521e-05,
"loss": 0.8964,
"step": 1600
},
{
"epoch": 0.26,
"eval_loss": 0.7759388089179993,
"eval_runtime": 96.8818,
"eval_samples_per_second": 7.194,
"eval_steps_per_second": 7.194,
"step": 1600
},
{
"epoch": 0.26,
"grad_norm": 0.6523721814155579,
"learning_rate": 4.910794360434993e-05,
"loss": 1.0127,
"step": 1605
},
{
"epoch": 0.26,
"grad_norm": 1.055384874343872,
"learning_rate": 4.9102403600685796e-05,
"loss": 0.9855,
"step": 1610
},
{
"epoch": 0.26,
"grad_norm": 0.7640814185142517,
"learning_rate": 4.9096846761932414e-05,
"loss": 0.7963,
"step": 1615
},
{
"epoch": 0.26,
"grad_norm": 0.5843799710273743,
"learning_rate": 4.9091273091971124e-05,
"loss": 0.8854,
"step": 1620
},
{
"epoch": 0.26,
"grad_norm": 0.9825207591056824,
"learning_rate": 4.9085682594695036e-05,
"loss": 0.8086,
"step": 1625
},
{
"epoch": 0.26,
"grad_norm": 0.9490563869476318,
"learning_rate": 4.908007527400901e-05,
"loss": 0.6838,
"step": 1630
},
{
"epoch": 0.26,
"grad_norm": 0.9472922682762146,
"learning_rate": 4.907445113382966e-05,
"loss": 0.8732,
"step": 1635
},
{
"epoch": 0.26,
"grad_norm": 0.6690593957901001,
"learning_rate": 4.9068810178085344e-05,
"loss": 0.8551,
"step": 1640
},
{
"epoch": 0.26,
"grad_norm": 0.7245538830757141,
"learning_rate": 4.906315241071616e-05,
"loss": 0.7639,
"step": 1645
},
{
"epoch": 0.26,
"grad_norm": 0.8342815041542053,
"learning_rate": 4.905747783567397e-05,
"loss": 0.9417,
"step": 1650
},
{
"epoch": 0.26,
"grad_norm": 0.6241989135742188,
"learning_rate": 4.9051786456922354e-05,
"loss": 0.9394,
"step": 1655
},
{
"epoch": 0.26,
"grad_norm": 0.5671687126159668,
"learning_rate": 4.904607827843663e-05,
"loss": 0.6381,
"step": 1660
},
{
"epoch": 0.27,
"grad_norm": 0.795868456363678,
"learning_rate": 4.9040353304203864e-05,
"loss": 0.7676,
"step": 1665
},
{
"epoch": 0.27,
"grad_norm": 0.9995182156562805,
"learning_rate": 4.9034611538222844e-05,
"loss": 1.0327,
"step": 1670
},
{
"epoch": 0.27,
"grad_norm": 0.7473803758621216,
"learning_rate": 4.902885298450409e-05,
"loss": 0.8835,
"step": 1675
},
{
"epoch": 0.27,
"grad_norm": 0.5757468938827515,
"learning_rate": 4.902307764706984e-05,
"loss": 0.7548,
"step": 1680
},
{
"epoch": 0.27,
"grad_norm": 0.8357987403869629,
"learning_rate": 4.901728552995407e-05,
"loss": 0.9184,
"step": 1685
},
{
"epoch": 0.27,
"grad_norm": 0.6664137244224548,
"learning_rate": 4.901147663720247e-05,
"loss": 0.9872,
"step": 1690
},
{
"epoch": 0.27,
"grad_norm": 0.861997663974762,
"learning_rate": 4.900565097287243e-05,
"loss": 0.8541,
"step": 1695
},
{
"epoch": 0.27,
"grad_norm": 0.7566475868225098,
"learning_rate": 4.8999808541033086e-05,
"loss": 0.9265,
"step": 1700
},
{
"epoch": 0.27,
"eval_loss": 0.7741928696632385,
"eval_runtime": 96.9038,
"eval_samples_per_second": 7.193,
"eval_steps_per_second": 7.193,
"step": 1700
},
{
"epoch": 0.27,
"grad_norm": 0.45475611090660095,
"learning_rate": 4.8993949345765266e-05,
"loss": 0.7186,
"step": 1705
},
{
"epoch": 0.27,
"grad_norm": 0.8672823905944824,
"learning_rate": 4.8988073391161515e-05,
"loss": 0.919,
"step": 1710
},
{
"epoch": 0.27,
"grad_norm": 0.7782495617866516,
"learning_rate": 4.8982180681326074e-05,
"loss": 0.6618,
"step": 1715
},
{
"epoch": 0.27,
"grad_norm": 0.6640329957008362,
"learning_rate": 4.897627122037489e-05,
"loss": 0.6662,
"step": 1720
},
{
"epoch": 0.28,
"grad_norm": 0.8019454479217529,
"learning_rate": 4.897034501243561e-05,
"loss": 0.9459,
"step": 1725
},
{
"epoch": 0.28,
"grad_norm": 0.8336368799209595,
"learning_rate": 4.896440206164761e-05,
"loss": 0.8058,
"step": 1730
},
{
"epoch": 0.28,
"grad_norm": 0.6316781044006348,
"learning_rate": 4.8958442372161906e-05,
"loss": 0.9132,
"step": 1735
},
{
"epoch": 0.28,
"grad_norm": 0.7768308520317078,
"learning_rate": 4.895246594814124e-05,
"loss": 0.7512,
"step": 1740
},
{
"epoch": 0.28,
"grad_norm": 0.9891632795333862,
"learning_rate": 4.894647279376002e-05,
"loss": 0.843,
"step": 1745
},
{
"epoch": 0.28,
"grad_norm": 0.6162430047988892,
"learning_rate": 4.894046291320439e-05,
"loss": 0.8233,
"step": 1750
},
{
"epoch": 0.28,
"grad_norm": 0.6184887290000916,
"learning_rate": 4.893443631067211e-05,
"loss": 0.7428,
"step": 1755
},
{
"epoch": 0.28,
"grad_norm": 0.7117312550544739,
"learning_rate": 4.892839299037267e-05,
"loss": 0.8707,
"step": 1760
},
{
"epoch": 0.28,
"grad_norm": 0.7165163159370422,
"learning_rate": 4.892233295652721e-05,
"loss": 1.0485,
"step": 1765
},
{
"epoch": 0.28,
"grad_norm": 0.8377657532691956,
"learning_rate": 4.891625621336855e-05,
"loss": 0.7368,
"step": 1770
},
{
"epoch": 0.28,
"grad_norm": 0.6349939703941345,
"learning_rate": 4.89101627651412e-05,
"loss": 0.7357,
"step": 1775
},
{
"epoch": 0.28,
"grad_norm": 4.969137191772461,
"learning_rate": 4.890405261610131e-05,
"loss": 0.7605,
"step": 1780
},
{
"epoch": 0.28,
"grad_norm": 1.5980018377304077,
"learning_rate": 4.889792577051671e-05,
"loss": 0.9253,
"step": 1785
},
{
"epoch": 0.29,
"grad_norm": 0.681398332118988,
"learning_rate": 4.889178223266688e-05,
"loss": 0.7235,
"step": 1790
},
{
"epoch": 0.29,
"grad_norm": 0.6999421715736389,
"learning_rate": 4.888562200684299e-05,
"loss": 0.8521,
"step": 1795
},
{
"epoch": 0.29,
"grad_norm": 0.7693730592727661,
"learning_rate": 4.887944509734783e-05,
"loss": 0.8632,
"step": 1800
},
{
"epoch": 0.29,
"eval_loss": 0.76987224817276,
"eval_runtime": 96.9052,
"eval_samples_per_second": 7.193,
"eval_steps_per_second": 7.193,
"step": 1800
},
{
"epoch": 0.29,
"grad_norm": 0.7641138434410095,
"learning_rate": 4.8873251508495865e-05,
"loss": 0.7074,
"step": 1805
},
{
"epoch": 0.29,
"grad_norm": 0.732545018196106,
"learning_rate": 4.886704124461321e-05,
"loss": 0.6901,
"step": 1810
},
{
"epoch": 0.29,
"grad_norm": 1.0327179431915283,
"learning_rate": 4.88608143100376e-05,
"loss": 0.8256,
"step": 1815
},
{
"epoch": 0.29,
"grad_norm": 0.7066757082939148,
"learning_rate": 4.885457070911845e-05,
"loss": 0.6635,
"step": 1820
},
{
"epoch": 0.29,
"grad_norm": 0.809877336025238,
"learning_rate": 4.8848310446216806e-05,
"loss": 0.795,
"step": 1825
},
{
"epoch": 0.29,
"grad_norm": 0.738153338432312,
"learning_rate": 4.8842033525705335e-05,
"loss": 0.9089,
"step": 1830
},
{
"epoch": 0.29,
"grad_norm": 0.754896879196167,
"learning_rate": 4.883573995196836e-05,
"loss": 0.7103,
"step": 1835
},
{
"epoch": 0.29,
"grad_norm": 1.0111182928085327,
"learning_rate": 4.8829429729401826e-05,
"loss": 1.046,
"step": 1840
},
{
"epoch": 0.29,
"grad_norm": 0.6233395934104919,
"learning_rate": 4.8823102862413306e-05,
"loss": 0.761,
"step": 1845
},
{
"epoch": 0.3,
"grad_norm": 1.3443419933319092,
"learning_rate": 4.8816759355422e-05,
"loss": 0.8436,
"step": 1850
},
{
"epoch": 0.3,
"grad_norm": 0.6685923337936401,
"learning_rate": 4.8810399212858736e-05,
"loss": 0.8956,
"step": 1855
},
{
"epoch": 0.3,
"grad_norm": 1.0405924320220947,
"learning_rate": 4.880402243916596e-05,
"loss": 1.1458,
"step": 1860
},
{
"epoch": 0.3,
"grad_norm": 0.8413107991218567,
"learning_rate": 4.879762903879772e-05,
"loss": 0.8133,
"step": 1865
},
{
"epoch": 0.3,
"grad_norm": 0.7151504158973694,
"learning_rate": 4.8791219016219705e-05,
"loss": 0.9207,
"step": 1870
},
{
"epoch": 0.3,
"grad_norm": 0.6887856125831604,
"learning_rate": 4.878479237590918e-05,
"loss": 0.8185,
"step": 1875
},
{
"epoch": 0.3,
"grad_norm": 0.5687748193740845,
"learning_rate": 4.877834912235506e-05,
"loss": 0.9035,
"step": 1880
},
{
"epoch": 0.3,
"grad_norm": 0.9966350793838501,
"learning_rate": 4.877188926005782e-05,
"loss": 0.7764,
"step": 1885
},
{
"epoch": 0.3,
"grad_norm": 1.0459462404251099,
"learning_rate": 4.8765412793529574e-05,
"loss": 0.6658,
"step": 1890
},
{
"epoch": 0.3,
"grad_norm": 0.8338847160339355,
"learning_rate": 4.8758919727293995e-05,
"loss": 0.7363,
"step": 1895
},
{
"epoch": 0.3,
"grad_norm": 0.7602768540382385,
"learning_rate": 4.875241006588638e-05,
"loss": 1.0081,
"step": 1900
},
{
"epoch": 0.3,
"eval_loss": 0.7692809700965881,
"eval_runtime": 96.4899,
"eval_samples_per_second": 7.224,
"eval_steps_per_second": 7.224,
"step": 1900
},
{
"epoch": 0.3,
"grad_norm": 0.5455746054649353,
"learning_rate": 4.874588381385362e-05,
"loss": 0.7855,
"step": 1905
},
{
"epoch": 0.3,
"grad_norm": 0.8574795126914978,
"learning_rate": 4.8739340975754165e-05,
"loss": 1.068,
"step": 1910
},
{
"epoch": 0.31,
"grad_norm": 1.0321904420852661,
"learning_rate": 4.873278155615808e-05,
"loss": 0.8239,
"step": 1915
},
{
"epoch": 0.31,
"grad_norm": 1.2484744787216187,
"learning_rate": 4.8726205559646996e-05,
"loss": 0.9307,
"step": 1920
},
{
"epoch": 0.31,
"grad_norm": 0.7140147686004639,
"learning_rate": 4.871961299081412e-05,
"loss": 0.9876,
"step": 1925
},
{
"epoch": 0.31,
"grad_norm": 0.8003590106964111,
"learning_rate": 4.871300385426426e-05,
"loss": 0.8615,
"step": 1930
},
{
"epoch": 0.31,
"grad_norm": 0.7282931208610535,
"learning_rate": 4.870637815461376e-05,
"loss": 0.8734,
"step": 1935
},
{
"epoch": 0.31,
"grad_norm": 0.6800629496574402,
"learning_rate": 4.869973589649055e-05,
"loss": 0.7718,
"step": 1940
},
{
"epoch": 0.31,
"grad_norm": 0.8813210129737854,
"learning_rate": 4.869307708453413e-05,
"loss": 0.7943,
"step": 1945
},
{
"epoch": 0.31,
"grad_norm": 0.6612805724143982,
"learning_rate": 4.868640172339557e-05,
"loss": 0.6807,
"step": 1950
},
{
"epoch": 0.31,
"grad_norm": 0.653191328048706,
"learning_rate": 4.867970981773748e-05,
"loss": 0.8948,
"step": 1955
},
{
"epoch": 0.31,
"grad_norm": 0.7479822635650635,
"learning_rate": 4.8673001372234025e-05,
"loss": 0.8583,
"step": 1960
},
{
"epoch": 0.31,
"grad_norm": NaN,
"learning_rate": 4.8667622710291026e-05,
"loss": 0.7443,
"step": 1965
},
{
"epoch": 0.31,
"grad_norm": 0.5788535475730896,
"learning_rate": 4.866088450488172e-05,
"loss": 0.7249,
"step": 1970
},
{
"epoch": 0.32,
"grad_norm": 0.7408040165901184,
"learning_rate": 4.86541297727762e-05,
"loss": 0.7115,
"step": 1975
},
{
"epoch": 0.32,
"grad_norm": 0.6549968719482422,
"learning_rate": 4.864735851869251e-05,
"loss": 0.9095,
"step": 1980
},
{
"epoch": 0.32,
"grad_norm": 0.4595119059085846,
"learning_rate": 4.864057074736026e-05,
"loss": 1.2808,
"step": 1985
},
{
"epoch": 0.32,
"grad_norm": 0.5746715068817139,
"learning_rate": 4.863376646352058e-05,
"loss": 0.8139,
"step": 1990
},
{
"epoch": 0.32,
"grad_norm": 0.6972643136978149,
"learning_rate": 4.862694567192614e-05,
"loss": 0.9797,
"step": 1995
},
{
"epoch": 0.32,
"grad_norm": 0.6935243010520935,
"learning_rate": 4.8620108377341124e-05,
"loss": 0.7651,
"step": 2000
},
{
"epoch": 0.32,
"eval_loss": 0.766412615776062,
"eval_runtime": 96.4555,
"eval_samples_per_second": 7.226,
"eval_steps_per_second": 7.226,
"step": 2000
},
{
"epoch": 0.32,
"grad_norm": 0.9983006715774536,
"learning_rate": 4.861325458454128e-05,
"loss": 0.8256,
"step": 2005
},
{
"epoch": 0.32,
"grad_norm": 0.6732650995254517,
"learning_rate": 4.860638429831384e-05,
"loss": 0.8136,
"step": 2010
},
{
"epoch": 0.32,
"grad_norm": 0.6780042052268982,
"learning_rate": 4.859949752345758e-05,
"loss": 0.8911,
"step": 2015
},
{
"epoch": 0.32,
"grad_norm": 0.9892123937606812,
"learning_rate": 4.8592594264782794e-05,
"loss": 0.7907,
"step": 2020
},
{
"epoch": 0.32,
"grad_norm": 0.9327254295349121,
"learning_rate": 4.8585674527111266e-05,
"loss": 0.8712,
"step": 2025
},
{
"epoch": 0.32,
"grad_norm": 1.0295612812042236,
"learning_rate": 4.857873831527632e-05,
"loss": 0.9188,
"step": 2030
},
{
"epoch": 0.32,
"grad_norm": 3.3071186542510986,
"learning_rate": 4.8571785634122766e-05,
"loss": 0.8801,
"step": 2035
},
{
"epoch": 0.33,
"grad_norm": 0.9625150561332703,
"learning_rate": 4.856481648850694e-05,
"loss": 0.8333,
"step": 2040
},
{
"epoch": 0.33,
"grad_norm": 0.6674854159355164,
"learning_rate": 4.855783088329664e-05,
"loss": 1.0388,
"step": 2045
},
{
"epoch": 0.33,
"grad_norm": 0.5447000861167908,
"learning_rate": 4.8550828823371196e-05,
"loss": 0.7893,
"step": 2050
},
{
"epoch": 0.33,
"grad_norm": 0.9970148801803589,
"learning_rate": 4.854381031362142e-05,
"loss": 0.8198,
"step": 2055
},
{
"epoch": 0.33,
"grad_norm": 0.7657136917114258,
"learning_rate": 4.853677535894961e-05,
"loss": 0.5977,
"step": 2060
},
{
"epoch": 0.33,
"grad_norm": 0.4694065451622009,
"learning_rate": 4.852972396426956e-05,
"loss": 0.5965,
"step": 2065
},
{
"epoch": 0.33,
"grad_norm": 0.8955700993537903,
"learning_rate": 4.852265613450653e-05,
"loss": 0.6938,
"step": 2070
},
{
"epoch": 0.33,
"grad_norm": 0.9884099960327148,
"learning_rate": 4.851557187459727e-05,
"loss": 0.8946,
"step": 2075
},
{
"epoch": 0.33,
"grad_norm": 0.6793637871742249,
"learning_rate": 4.850847118949002e-05,
"loss": 0.841,
"step": 2080
},
{
"epoch": 0.33,
"grad_norm": 0.7438017725944519,
"learning_rate": 4.850135408414447e-05,
"loss": 0.8843,
"step": 2085
},
{
"epoch": 0.33,
"grad_norm": 0.7632609009742737,
"learning_rate": 4.849422056353178e-05,
"loss": 0.8263,
"step": 2090
},
{
"epoch": 0.33,
"grad_norm": 0.7281492352485657,
"learning_rate": 4.84870706326346e-05,
"loss": 0.8989,
"step": 2095
},
{
"epoch": 0.34,
"grad_norm": 0.6480591893196106,
"learning_rate": 4.847990429644702e-05,
"loss": 1.0037,
"step": 2100
},
{
"epoch": 0.34,
"eval_loss": 0.7653521299362183,
"eval_runtime": 96.4452,
"eval_samples_per_second": 7.227,
"eval_steps_per_second": 7.227,
"step": 2100
},
{
"epoch": 0.34,
"grad_norm": 0.5578673481941223,
"learning_rate": 4.8472721559974584e-05,
"loss": 0.911,
"step": 2105
},
{
"epoch": 0.34,
"grad_norm": 0.5615595579147339,
"learning_rate": 4.846552242823433e-05,
"loss": 0.6938,
"step": 2110
},
{
"epoch": 0.34,
"grad_norm": 0.588246762752533,
"learning_rate": 4.845830690625469e-05,
"loss": 0.7898,
"step": 2115
},
{
"epoch": 0.34,
"grad_norm": 0.8140611052513123,
"learning_rate": 4.8451074999075595e-05,
"loss": 0.7702,
"step": 2120
},
{
"epoch": 0.34,
"grad_norm": 0.9400056600570679,
"learning_rate": 4.8443826711748385e-05,
"loss": 0.7959,
"step": 2125
},
{
"epoch": 0.34,
"grad_norm": 0.7187873721122742,
"learning_rate": 4.8436562049335874e-05,
"loss": 0.7223,
"step": 2130
},
{
"epoch": 0.34,
"grad_norm": 0.7627830505371094,
"learning_rate": 4.8429281016912275e-05,
"loss": 0.793,
"step": 2135
},
{
"epoch": 0.34,
"grad_norm": 0.6755004525184631,
"learning_rate": 4.842198361956328e-05,
"loss": 0.7665,
"step": 2140
},
{
"epoch": 0.34,
"grad_norm": 0.6032254695892334,
"learning_rate": 4.8414669862385966e-05,
"loss": 0.7952,
"step": 2145
},
{
"epoch": 0.34,
"grad_norm": 0.8377916216850281,
"learning_rate": 4.840733975048887e-05,
"loss": 1.0016,
"step": 2150
},
{
"epoch": 0.34,
"grad_norm": 0.7361429929733276,
"learning_rate": 4.839999328899194e-05,
"loss": 0.8773,
"step": 2155
},
{
"epoch": 0.34,
"grad_norm": 0.8006517887115479,
"learning_rate": 4.8392630483026546e-05,
"loss": 0.9334,
"step": 2160
},
{
"epoch": 0.35,
"grad_norm": 0.9716467261314392,
"learning_rate": 4.8385251337735473e-05,
"loss": 1.0359,
"step": 2165
},
{
"epoch": 0.35,
"grad_norm": 0.6826418042182922,
"learning_rate": 4.8377855858272925e-05,
"loss": 0.6841,
"step": 2170
},
{
"epoch": 0.35,
"grad_norm": 0.4519975781440735,
"learning_rate": 4.8370444049804494e-05,
"loss": 0.8326,
"step": 2175
},
{
"epoch": 0.35,
"grad_norm": 0.677891731262207,
"learning_rate": 4.836301591750721e-05,
"loss": 1.0841,
"step": 2180
},
{
"epoch": 0.35,
"grad_norm": 1.5161852836608887,
"learning_rate": 4.835557146656948e-05,
"loss": 0.8701,
"step": 2185
},
{
"epoch": 0.35,
"grad_norm": 0.6586780548095703,
"learning_rate": 4.834811070219112e-05,
"loss": 0.8261,
"step": 2190
},
{
"epoch": 0.35,
"grad_norm": 0.48046165704727173,
"learning_rate": 4.834063362958333e-05,
"loss": 0.6375,
"step": 2195
},
{
"epoch": 0.35,
"grad_norm": 1.0315968990325928,
"learning_rate": 4.833314025396872e-05,
"loss": 0.8768,
"step": 2200
},
{
"epoch": 0.35,
"eval_loss": 0.7641988396644592,
"eval_runtime": 96.3923,
"eval_samples_per_second": 7.231,
"eval_steps_per_second": 7.231,
"step": 2200
},
{
"epoch": 0.35,
"grad_norm": 0.7704123258590698,
"learning_rate": 4.8325630580581263e-05,
"loss": 0.8849,
"step": 2205
},
{
"epoch": 0.35,
"grad_norm": 1.087425708770752,
"learning_rate": 4.831810461466634e-05,
"loss": 0.9828,
"step": 2210
},
{
"epoch": 0.35,
"grad_norm": 0.4766077995300293,
"learning_rate": 4.83105623614807e-05,
"loss": 0.7103,
"step": 2215
},
{
"epoch": 0.35,
"grad_norm": 0.6079148054122925,
"learning_rate": 4.830300382629247e-05,
"loss": 0.7253,
"step": 2220
},
{
"epoch": 0.36,
"grad_norm": 0.6767585873603821,
"learning_rate": 4.829542901438115e-05,
"loss": 0.7852,
"step": 2225
},
{
"epoch": 0.36,
"grad_norm": 0.7065784335136414,
"learning_rate": 4.8287837931037585e-05,
"loss": 0.8047,
"step": 2230
},
{
"epoch": 0.36,
"grad_norm": 0.8305274248123169,
"learning_rate": 4.828023058156404e-05,
"loss": 0.7912,
"step": 2235
},
{
"epoch": 0.36,
"grad_norm": 0.8435990810394287,
"learning_rate": 4.827260697127409e-05,
"loss": 0.826,
"step": 2240
},
{
"epoch": 0.36,
"grad_norm": 0.8484389185905457,
"learning_rate": 4.8264967105492705e-05,
"loss": 0.706,
"step": 2245
},
{
"epoch": 0.36,
"grad_norm": 0.7461299300193787,
"learning_rate": 4.825731098955617e-05,
"loss": 0.763,
"step": 2250
},
{
"epoch": 0.36,
"grad_norm": 0.7928741574287415,
"learning_rate": 4.824963862881216e-05,
"loss": 0.8125,
"step": 2255
},
{
"epoch": 0.36,
"grad_norm": 0.7152695059776306,
"learning_rate": 4.824195002861968e-05,
"loss": 1.129,
"step": 2260
},
{
"epoch": 0.36,
"grad_norm": 0.8594226241111755,
"learning_rate": 4.8234245194349056e-05,
"loss": 0.8873,
"step": 2265
},
{
"epoch": 0.36,
"grad_norm": 0.9760085940361023,
"learning_rate": 4.822652413138199e-05,
"loss": 0.9713,
"step": 2270
},
{
"epoch": 0.36,
"grad_norm": 0.7297483682632446,
"learning_rate": 4.8218786845111505e-05,
"loss": 0.6953,
"step": 2275
},
{
"epoch": 0.36,
"grad_norm": 0.8251492381095886,
"learning_rate": 4.8211033340941956e-05,
"loss": 0.7649,
"step": 2280
},
{
"epoch": 0.36,
"grad_norm": 0.742917537689209,
"learning_rate": 4.820326362428901e-05,
"loss": 0.9756,
"step": 2285
},
{
"epoch": 0.37,
"grad_norm": 0.7784115076065063,
"learning_rate": 4.819547770057969e-05,
"loss": 0.6937,
"step": 2290
},
{
"epoch": 0.37,
"grad_norm": 1.782772183418274,
"learning_rate": 4.8187675575252314e-05,
"loss": 0.9062,
"step": 2295
},
{
"epoch": 0.37,
"grad_norm": 0.7802585363388062,
"learning_rate": 4.8179857253756514e-05,
"loss": 0.8052,
"step": 2300
},
{
"epoch": 0.37,
"eval_loss": 0.7618402242660522,
"eval_runtime": 96.4079,
"eval_samples_per_second": 7.23,
"eval_steps_per_second": 7.23,
"step": 2300
},
{
"epoch": 0.37,
"grad_norm": 0.799985945224762,
"learning_rate": 4.8172022741553255e-05,
"loss": 0.9046,
"step": 2305
},
{
"epoch": 0.37,
"grad_norm": 1.026978850364685,
"learning_rate": 4.816417204411481e-05,
"loss": 0.7195,
"step": 2310
},
{
"epoch": 0.37,
"grad_norm": 0.8067365884780884,
"learning_rate": 4.8156305166924734e-05,
"loss": 0.8193,
"step": 2315
},
{
"epoch": 0.37,
"grad_norm": 1.247164249420166,
"learning_rate": 4.81484221154779e-05,
"loss": 0.6138,
"step": 2320
},
{
"epoch": 0.37,
"grad_norm": 0.8662647604942322,
"learning_rate": 4.814052289528047e-05,
"loss": 0.7763,
"step": 2325
},
{
"epoch": 0.37,
"grad_norm": 0.9020537734031677,
"learning_rate": 4.813260751184992e-05,
"loss": 0.9236,
"step": 2330
},
{
"epoch": 0.37,
"grad_norm": 0.6113781929016113,
"learning_rate": 4.812467597071499e-05,
"loss": 0.8753,
"step": 2335
},
{
"epoch": 0.37,
"grad_norm": 0.6988622546195984,
"learning_rate": 4.811672827741572e-05,
"loss": 0.6747,
"step": 2340
},
{
"epoch": 0.37,
"grad_norm": 0.9095928072929382,
"learning_rate": 4.810876443750344e-05,
"loss": 1.0578,
"step": 2345
},
{
"epoch": 0.38,
"grad_norm": 0.643699049949646,
"learning_rate": 4.8100784456540724e-05,
"loss": 0.8177,
"step": 2350
},
{
"epoch": 0.38,
"grad_norm": 0.7084022760391235,
"learning_rate": 4.809278834010146e-05,
"loss": 0.9345,
"step": 2355
},
{
"epoch": 0.38,
"grad_norm": 0.5328305959701538,
"learning_rate": 4.808477609377078e-05,
"loss": 0.6781,
"step": 2360
},
{
"epoch": 0.38,
"grad_norm": 0.8238436579704285,
"learning_rate": 4.80767477231451e-05,
"loss": 0.7306,
"step": 2365
},
{
"epoch": 0.38,
"grad_norm": 1.0184216499328613,
"learning_rate": 4.806870323383208e-05,
"loss": 1.0288,
"step": 2370
},
{
"epoch": 0.38,
"grad_norm": 0.8620426654815674,
"learning_rate": 4.806064263145066e-05,
"loss": 0.7925,
"step": 2375
},
{
"epoch": 0.38,
"grad_norm": 0.6541377305984497,
"learning_rate": 4.805256592163102e-05,
"loss": 0.8629,
"step": 2380
},
{
"epoch": 0.38,
"grad_norm": 0.8664489984512329,
"learning_rate": 4.8044473110014594e-05,
"loss": 0.8184,
"step": 2385
},
{
"epoch": 0.38,
"grad_norm": 0.7283564209938049,
"learning_rate": 4.803636420225406e-05,
"loss": 0.9444,
"step": 2390
},
{
"epoch": 0.38,
"grad_norm": 0.7168800234794617,
"learning_rate": 4.802823920401335e-05,
"loss": 0.8118,
"step": 2395
},
{
"epoch": 0.38,
"grad_norm": 0.8198531866073608,
"learning_rate": 4.802009812096762e-05,
"loss": 0.7271,
"step": 2400
},
{
"epoch": 0.38,
"eval_loss": 0.7595117688179016,
"eval_runtime": 96.4847,
"eval_samples_per_second": 7.224,
"eval_steps_per_second": 7.224,
"step": 2400
},
{
"epoch": 0.38,
"grad_norm": 0.5693966150283813,
"learning_rate": 4.801194095880327e-05,
"loss": 0.7801,
"step": 2405
},
{
"epoch": 0.38,
"grad_norm": 0.7175332307815552,
"learning_rate": 4.800376772321793e-05,
"loss": 0.7873,
"step": 2410
},
{
"epoch": 0.39,
"grad_norm": 0.7779633402824402,
"learning_rate": 4.799557841992046e-05,
"loss": 0.894,
"step": 2415
},
{
"epoch": 0.39,
"grad_norm": 0.7832231521606445,
"learning_rate": 4.798737305463092e-05,
"loss": 0.8035,
"step": 2420
},
{
"epoch": 0.39,
"grad_norm": 0.5115272998809814,
"learning_rate": 4.797915163308064e-05,
"loss": 0.8885,
"step": 2425
},
{
"epoch": 0.39,
"grad_norm": 0.9534878730773926,
"learning_rate": 4.79709141610121e-05,
"loss": 0.8175,
"step": 2430
},
{
"epoch": 0.39,
"grad_norm": 0.7053850889205933,
"learning_rate": 4.796266064417905e-05,
"loss": 0.6971,
"step": 2435
},
{
"epoch": 0.39,
"grad_norm": 1.236257791519165,
"learning_rate": 4.795439108834641e-05,
"loss": 1.0832,
"step": 2440
},
{
"epoch": 0.39,
"grad_norm": 0.6936543583869934,
"learning_rate": 4.794610549929031e-05,
"loss": 0.858,
"step": 2445
},
{
"epoch": 0.39,
"grad_norm": 0.8064691424369812,
"learning_rate": 4.793780388279809e-05,
"loss": 0.6951,
"step": 2450
},
{
"epoch": 0.39,
"grad_norm": 0.7180449962615967,
"learning_rate": 4.792948624466827e-05,
"loss": 0.6779,
"step": 2455
},
{
"epoch": 0.39,
"grad_norm": 0.6903377175331116,
"learning_rate": 4.792115259071058e-05,
"loss": 0.8281,
"step": 2460
},
{
"epoch": 0.39,
"grad_norm": 0.9112733006477356,
"learning_rate": 4.791280292674591e-05,
"loss": 0.938,
"step": 2465
},
{
"epoch": 0.39,
"grad_norm": 0.8657469153404236,
"learning_rate": 4.790443725860636e-05,
"loss": 0.8063,
"step": 2470
},
{
"epoch": 0.4,
"grad_norm": 0.9260883927345276,
"learning_rate": 4.7896055592135194e-05,
"loss": 1.0093,
"step": 2475
},
{
"epoch": 0.4,
"grad_norm": 0.7651245594024658,
"learning_rate": 4.788765793318685e-05,
"loss": 0.6686,
"step": 2480
},
{
"epoch": 0.4,
"grad_norm": 0.6063816547393799,
"learning_rate": 4.7879244287626945e-05,
"loss": 0.8516,
"step": 2485
},
{
"epoch": 0.4,
"grad_norm": 0.9127621650695801,
"learning_rate": 4.787081466133225e-05,
"loss": 0.7992,
"step": 2490
},
{
"epoch": 0.4,
"grad_norm": 1.061246633529663,
"learning_rate": 4.7862369060190716e-05,
"loss": 0.8232,
"step": 2495
},
{
"epoch": 0.4,
"grad_norm": 0.7100695967674255,
"learning_rate": 4.785390749010143e-05,
"loss": 0.9615,
"step": 2500
},
{
"epoch": 0.4,
"eval_loss": 0.7581596970558167,
"eval_runtime": 96.5797,
"eval_samples_per_second": 7.217,
"eval_steps_per_second": 7.217,
"step": 2500
},
{
"epoch": 0.4,
"grad_norm": 16.361513137817383,
"learning_rate": 4.784542995697464e-05,
"loss": 0.7725,
"step": 2505
},
{
"epoch": 0.4,
"grad_norm": 0.7746205925941467,
"learning_rate": 4.7836936466731764e-05,
"loss": 0.8464,
"step": 2510
},
{
"epoch": 0.4,
"grad_norm": 0.7703484892845154,
"learning_rate": 4.7828427025305345e-05,
"loss": 0.8596,
"step": 2515
},
{
"epoch": 0.4,
"grad_norm": 0.7838412523269653,
"learning_rate": 4.7819901638639066e-05,
"loss": 0.666,
"step": 2520
},
{
"epoch": 0.4,
"grad_norm": 0.5832842588424683,
"learning_rate": 4.781136031268776e-05,
"loss": 0.4995,
"step": 2525
},
{
"epoch": 0.4,
"grad_norm": 0.798271894454956,
"learning_rate": 4.780280305341739e-05,
"loss": 1.0017,
"step": 2530
},
{
"epoch": 0.4,
"grad_norm": 0.463828444480896,
"learning_rate": 4.779422986680503e-05,
"loss": 0.5894,
"step": 2535
},
{
"epoch": 0.41,
"grad_norm": 0.761908233165741,
"learning_rate": 4.7785640758838916e-05,
"loss": 0.9198,
"step": 2540
},
{
"epoch": 0.41,
"grad_norm": 0.8427887558937073,
"learning_rate": 4.777703573551837e-05,
"loss": 0.8572,
"step": 2545
},
{
"epoch": 0.41,
"grad_norm": 0.6188894510269165,
"learning_rate": 4.776841480285384e-05,
"loss": 0.9102,
"step": 2550
},
{
"epoch": 0.41,
"grad_norm": 0.7198623418807983,
"learning_rate": 4.775977796686691e-05,
"loss": 0.8472,
"step": 2555
},
{
"epoch": 0.41,
"grad_norm": 1.0144587755203247,
"learning_rate": 4.775112523359023e-05,
"loss": 0.7059,
"step": 2560
},
{
"epoch": 0.41,
"grad_norm": 0.9784219861030579,
"learning_rate": 4.77424566090676e-05,
"loss": 0.7417,
"step": 2565
},
{
"epoch": 0.41,
"grad_norm": 0.5349156856536865,
"learning_rate": 4.773377209935387e-05,
"loss": 0.7287,
"step": 2570
},
{
"epoch": 0.41,
"grad_norm": 0.7715370655059814,
"learning_rate": 4.772507171051502e-05,
"loss": 0.8393,
"step": 2575
},
{
"epoch": 0.41,
"grad_norm": 0.8483054637908936,
"learning_rate": 4.771635544862813e-05,
"loss": 0.8938,
"step": 2580
},
{
"epoch": 0.41,
"grad_norm": 0.8196272253990173,
"learning_rate": 4.770762331978132e-05,
"loss": 0.8321,
"step": 2585
},
{
"epoch": 0.41,
"grad_norm": 0.6155353784561157,
"learning_rate": 4.769887533007384e-05,
"loss": 0.9291,
"step": 2590
},
{
"epoch": 0.41,
"grad_norm": 0.8897277116775513,
"learning_rate": 4.769011148561601e-05,
"loss": 0.7098,
"step": 2595
},
{
"epoch": 0.42,
"grad_norm": 1.2256160974502563,
"learning_rate": 4.768133179252921e-05,
"loss": 0.8284,
"step": 2600
},
{
"epoch": 0.42,
"eval_loss": 0.7554901838302612,
"eval_runtime": 96.5279,
"eval_samples_per_second": 7.221,
"eval_steps_per_second": 7.221,
"step": 2600
},
{
"epoch": 0.42,
"grad_norm": 0.6943432688713074,
"learning_rate": 4.767253625694588e-05,
"loss": 0.8785,
"step": 2605
},
{
"epoch": 0.42,
"grad_norm": 0.6707726120948792,
"learning_rate": 4.7663724885009556e-05,
"loss": 0.7949,
"step": 2610
},
{
"epoch": 0.42,
"grad_norm": 0.5595915913581848,
"learning_rate": 4.765489768287481e-05,
"loss": 0.8796,
"step": 2615
},
{
"epoch": 0.42,
"grad_norm": 0.9889727234840393,
"learning_rate": 4.7646054656707306e-05,
"loss": 1.0676,
"step": 2620
},
{
"epoch": 0.42,
"grad_norm": 0.8624396324157715,
"learning_rate": 4.763719581268371e-05,
"loss": 0.709,
"step": 2625
},
{
"epoch": 0.42,
"grad_norm": 0.7466241121292114,
"learning_rate": 4.7628321156991767e-05,
"loss": 0.8084,
"step": 2630
},
{
"epoch": 0.42,
"grad_norm": 0.6439360976219177,
"learning_rate": 4.761943069583027e-05,
"loss": 0.8831,
"step": 2635
},
{
"epoch": 0.42,
"grad_norm": 0.9999917149543762,
"learning_rate": 4.761052443540904e-05,
"loss": 0.6372,
"step": 2640
},
{
"epoch": 0.42,
"grad_norm": 0.688369870185852,
"learning_rate": 4.760160238194894e-05,
"loss": 0.7938,
"step": 2645
},
{
"epoch": 0.42,
"grad_norm": 0.6920734643936157,
"learning_rate": 4.759266454168186e-05,
"loss": 0.7378,
"step": 2650
},
{
"epoch": 0.42,
"grad_norm": 0.7592100501060486,
"learning_rate": 4.758371092085073e-05,
"loss": 1.097,
"step": 2655
},
{
"epoch": 0.42,
"grad_norm": 0.9243403077125549,
"learning_rate": 4.757474152570946e-05,
"loss": 1.0404,
"step": 2660
},
{
"epoch": 0.43,
"grad_norm": 0.8212980031967163,
"learning_rate": 4.756575636252304e-05,
"loss": 0.6179,
"step": 2665
},
{
"epoch": 0.43,
"grad_norm": 0.6905696392059326,
"learning_rate": 4.755675543756744e-05,
"loss": 0.8398,
"step": 2670
},
{
"epoch": 0.43,
"grad_norm": 0.8420882821083069,
"learning_rate": 4.754773875712961e-05,
"loss": 0.7552,
"step": 2675
},
{
"epoch": 0.43,
"grad_norm": 0.6216087341308594,
"learning_rate": 4.7538706327507575e-05,
"loss": 0.8345,
"step": 2680
},
{
"epoch": 0.43,
"grad_norm": 0.7430551648139954,
"learning_rate": 4.75296581550103e-05,
"loss": 0.8277,
"step": 2685
},
{
"epoch": 0.43,
"grad_norm": 0.7866222262382507,
"learning_rate": 4.752059424595778e-05,
"loss": 0.9178,
"step": 2690
},
{
"epoch": 0.43,
"grad_norm": 0.6548468470573425,
"learning_rate": 4.7511514606680985e-05,
"loss": 0.745,
"step": 2695
},
{
"epoch": 0.43,
"grad_norm": 0.6956586837768555,
"learning_rate": 4.750241924352187e-05,
"loss": 0.8631,
"step": 2700
},
{
"epoch": 0.43,
"eval_loss": 0.7539612650871277,
"eval_runtime": 96.4433,
"eval_samples_per_second": 7.227,
"eval_steps_per_second": 7.227,
"step": 2700
},
{
"epoch": 0.43,
"grad_norm": 0.6508235335350037,
"learning_rate": 4.7493308162833394e-05,
"loss": 0.9936,
"step": 2705
},
{
"epoch": 0.43,
"grad_norm": 0.8658422827720642,
"learning_rate": 4.7484181370979475e-05,
"loss": 0.8,
"step": 2710
},
{
"epoch": 0.43,
"grad_norm": 0.9571516513824463,
"learning_rate": 4.747503887433501e-05,
"loss": 0.7028,
"step": 2715
},
{
"epoch": 0.43,
"grad_norm": 0.7693742513656616,
"learning_rate": 4.7465880679285866e-05,
"loss": 0.7194,
"step": 2720
},
{
"epoch": 0.43,
"grad_norm": 1.34340238571167,
"learning_rate": 4.745670679222888e-05,
"loss": 1.0445,
"step": 2725
},
{
"epoch": 0.44,
"grad_norm": 2.71327805519104,
"learning_rate": 4.7447517219571834e-05,
"loss": 0.8088,
"step": 2730
},
{
"epoch": 0.44,
"grad_norm": 0.9449920058250427,
"learning_rate": 4.743831196773349e-05,
"loss": 0.7939,
"step": 2735
},
{
"epoch": 0.44,
"grad_norm": 0.8091790676116943,
"learning_rate": 4.742909104314353e-05,
"loss": 0.7816,
"step": 2740
},
{
"epoch": 0.44,
"grad_norm": 0.5790795087814331,
"learning_rate": 4.741985445224263e-05,
"loss": 0.8778,
"step": 2745
},
{
"epoch": 0.44,
"grad_norm": 1.1936956644058228,
"learning_rate": 4.741060220148236e-05,
"loss": 1.0242,
"step": 2750
},
{
"epoch": 0.44,
"grad_norm": 0.5158389806747437,
"learning_rate": 4.7401334297325244e-05,
"loss": 0.7954,
"step": 2755
},
{
"epoch": 0.44,
"grad_norm": 0.8950900435447693,
"learning_rate": 4.7392050746244754e-05,
"loss": 0.7603,
"step": 2760
},
{
"epoch": 0.44,
"grad_norm": 0.7289401888847351,
"learning_rate": 4.738275155472528e-05,
"loss": 0.879,
"step": 2765
},
{
"epoch": 0.44,
"grad_norm": 0.8410510420799255,
"learning_rate": 4.7373436729262145e-05,
"loss": 0.7399,
"step": 2770
},
{
"epoch": 0.44,
"grad_norm": 0.7992503643035889,
"learning_rate": 4.736410627636156e-05,
"loss": 0.6779,
"step": 2775
},
{
"epoch": 0.44,
"grad_norm": 0.6706194281578064,
"learning_rate": 4.73547602025407e-05,
"loss": 0.7878,
"step": 2780
},
{
"epoch": 0.44,
"grad_norm": 0.7177903652191162,
"learning_rate": 4.734539851432763e-05,
"loss": 0.6958,
"step": 2785
},
{
"epoch": 0.45,
"grad_norm": 0.6557692885398865,
"learning_rate": 4.73360212182613e-05,
"loss": 0.6695,
"step": 2790
},
{
"epoch": 0.45,
"grad_norm": 0.6754157543182373,
"learning_rate": 4.7326628320891586e-05,
"loss": 0.9057,
"step": 2795
},
{
"epoch": 0.45,
"grad_norm": 1.1403777599334717,
"learning_rate": 4.731721982877926e-05,
"loss": 1.0507,
"step": 2800
},
{
"epoch": 0.45,
"eval_loss": 0.7518497705459595,
"eval_runtime": 96.4525,
"eval_samples_per_second": 7.226,
"eval_steps_per_second": 7.226,
"step": 2800
},
{
"epoch": 0.45,
"grad_norm": 0.8268899321556091,
"learning_rate": 4.730779574849598e-05,
"loss": 0.7375,
"step": 2805
},
{
"epoch": 0.45,
"grad_norm": 0.5358712673187256,
"learning_rate": 4.72983560866243e-05,
"loss": 0.7839,
"step": 2810
},
{
"epoch": 0.45,
"grad_norm": 1.0761948823928833,
"learning_rate": 4.7288900849757636e-05,
"loss": 0.7936,
"step": 2815
},
{
"epoch": 0.45,
"grad_norm": 0.7037429213523865,
"learning_rate": 4.7279430044500315e-05,
"loss": 0.6875,
"step": 2820
},
{
"epoch": 0.45,
"grad_norm": 0.6378889679908752,
"learning_rate": 4.726994367746751e-05,
"loss": 0.9209,
"step": 2825
},
{
"epoch": 0.45,
"grad_norm": 0.5508277416229248,
"learning_rate": 4.7260441755285284e-05,
"loss": 0.9402,
"step": 2830
},
{
"epoch": 0.45,
"grad_norm": 0.9046247005462646,
"learning_rate": 4.725092428459055e-05,
"loss": 0.6336,
"step": 2835
},
{
"epoch": 0.45,
"grad_norm": 0.8689594864845276,
"learning_rate": 4.7241391272031096e-05,
"loss": 1.1281,
"step": 2840
},
{
"epoch": 0.45,
"grad_norm": 0.8785949945449829,
"learning_rate": 4.723184272426555e-05,
"loss": 0.711,
"step": 2845
},
{
"epoch": 0.45,
"grad_norm": 0.9959015250205994,
"learning_rate": 4.722227864796339e-05,
"loss": 0.7432,
"step": 2850
},
{
"epoch": 0.46,
"grad_norm": 0.6438590884208679,
"learning_rate": 4.721269904980497e-05,
"loss": 0.883,
"step": 2855
},
{
"epoch": 0.46,
"grad_norm": 0.6714455485343933,
"learning_rate": 4.720310393648145e-05,
"loss": 1.065,
"step": 2860
},
{
"epoch": 0.46,
"grad_norm": 0.7378780245780945,
"learning_rate": 4.7193493314694846e-05,
"loss": 0.5352,
"step": 2865
},
{
"epoch": 0.46,
"grad_norm": 0.7698020935058594,
"learning_rate": 4.7183867191158006e-05,
"loss": 0.7016,
"step": 2870
},
{
"epoch": 0.46,
"grad_norm": 0.952795684337616,
"learning_rate": 4.7174225572594586e-05,
"loss": 1.0659,
"step": 2875
},
{
"epoch": 0.46,
"grad_norm": 0.6401458978652954,
"learning_rate": 4.71645684657391e-05,
"loss": 0.7335,
"step": 2880
},
{
"epoch": 0.46,
"grad_norm": 0.8375076055526733,
"learning_rate": 4.715489587733685e-05,
"loss": 0.9264,
"step": 2885
},
{
"epoch": 0.46,
"grad_norm": 0.693505048751831,
"learning_rate": 4.714520781414397e-05,
"loss": 1.0286,
"step": 2890
},
{
"epoch": 0.46,
"grad_norm": 1.0239859819412231,
"learning_rate": 4.7135504282927375e-05,
"loss": 0.6875,
"step": 2895
},
{
"epoch": 0.46,
"grad_norm": 0.602035403251648,
"learning_rate": 4.712578529046483e-05,
"loss": 0.8247,
"step": 2900
},
{
"epoch": 0.46,
"eval_loss": 0.7512397766113281,
"eval_runtime": 96.4745,
"eval_samples_per_second": 7.225,
"eval_steps_per_second": 7.225,
"step": 2900
},
{
"epoch": 0.46,
"grad_norm": 0.6859713196754456,
"learning_rate": 4.711605084354487e-05,
"loss": 0.7521,
"step": 2905
},
{
"epoch": 0.46,
"grad_norm": 0.7126486301422119,
"learning_rate": 4.7106300948966817e-05,
"loss": 0.7656,
"step": 2910
},
{
"epoch": 0.47,
"grad_norm": 0.4363511800765991,
"learning_rate": 4.70965356135408e-05,
"loss": 1.1595,
"step": 2915
},
{
"epoch": 0.47,
"grad_norm": 0.6381859183311462,
"learning_rate": 4.7086754844087724e-05,
"loss": 0.6949,
"step": 2920
},
{
"epoch": 0.47,
"grad_norm": 0.7931796312332153,
"learning_rate": 4.7076958647439284e-05,
"loss": 1.0821,
"step": 2925
},
{
"epoch": 0.47,
"grad_norm": 0.9333865642547607,
"learning_rate": 4.706714703043795e-05,
"loss": 0.7753,
"step": 2930
},
{
"epoch": 0.47,
"grad_norm": 0.8860915899276733,
"learning_rate": 4.705731999993694e-05,
"loss": 0.7257,
"step": 2935
},
{
"epoch": 0.47,
"grad_norm": 0.6868377327919006,
"learning_rate": 4.704747756280027e-05,
"loss": 0.8148,
"step": 2940
},
{
"epoch": 0.47,
"grad_norm": 0.5337914228439331,
"learning_rate": 4.7037619725902706e-05,
"loss": 0.7379,
"step": 2945
},
{
"epoch": 0.47,
"grad_norm": 0.4664730429649353,
"learning_rate": 4.7027746496129745e-05,
"loss": 0.6226,
"step": 2950
},
{
"epoch": 0.47,
"grad_norm": 0.7305762767791748,
"learning_rate": 4.701785788037768e-05,
"loss": 0.9018,
"step": 2955
},
{
"epoch": 0.47,
"grad_norm": 0.6576158404350281,
"learning_rate": 4.7007953885553525e-05,
"loss": 0.7777,
"step": 2960
},
{
"epoch": 0.47,
"grad_norm": 0.9728206396102905,
"learning_rate": 4.699803451857503e-05,
"loss": 0.8004,
"step": 2965
},
{
"epoch": 0.47,
"grad_norm": 0.6211077570915222,
"learning_rate": 4.69880997863707e-05,
"loss": 0.7407,
"step": 2970
},
{
"epoch": 0.47,
"grad_norm": 1.2564159631729126,
"learning_rate": 4.697814969587976e-05,
"loss": 0.7993,
"step": 2975
},
{
"epoch": 0.48,
"grad_norm": 0.927930474281311,
"learning_rate": 4.696818425405217e-05,
"loss": 0.8803,
"step": 2980
},
{
"epoch": 0.48,
"grad_norm": 0.9062425494194031,
"learning_rate": 4.695820346784861e-05,
"loss": 0.8835,
"step": 2985
},
{
"epoch": 0.48,
"grad_norm": 0.6738875508308411,
"learning_rate": 4.694820734424047e-05,
"loss": 0.7817,
"step": 2990
},
{
"epoch": 0.48,
"grad_norm": 1.326353669166565,
"learning_rate": 4.6938195890209866e-05,
"loss": 0.9213,
"step": 2995
},
{
"epoch": 0.48,
"grad_norm": 0.4853856563568115,
"learning_rate": 4.692816911274962e-05,
"loss": 0.9835,
"step": 3000
},
{
"epoch": 0.48,
"eval_loss": 0.7496011257171631,
"eval_runtime": 96.515,
"eval_samples_per_second": 7.222,
"eval_steps_per_second": 7.222,
"step": 3000
},
{
"epoch": 0.48,
"grad_norm": 0.5411309003829956,
"learning_rate": 4.691812701886324e-05,
"loss": 0.7556,
"step": 3005
},
{
"epoch": 0.48,
"grad_norm": 0.7545793652534485,
"learning_rate": 4.6908069615564966e-05,
"loss": 0.8295,
"step": 3010
},
{
"epoch": 0.48,
"grad_norm": 0.850104808807373,
"learning_rate": 4.6897996909879695e-05,
"loss": 1.0194,
"step": 3015
},
{
"epoch": 0.48,
"grad_norm": 0.69708651304245,
"learning_rate": 4.6887908908843026e-05,
"loss": 0.7918,
"step": 3020
},
{
"epoch": 0.48,
"grad_norm": 2.1333253383636475,
"learning_rate": 4.687780561950126e-05,
"loss": 0.7287,
"step": 3025
},
{
"epoch": 0.48,
"grad_norm": 0.9223487973213196,
"learning_rate": 4.686768704891134e-05,
"loss": 0.9592,
"step": 3030
},
{
"epoch": 0.48,
"grad_norm": 0.7700949311256409,
"learning_rate": 4.685755320414091e-05,
"loss": 0.8572,
"step": 3035
},
{
"epoch": 0.49,
"grad_norm": 0.5573208332061768,
"learning_rate": 4.684740409226829e-05,
"loss": 0.9441,
"step": 3040
},
{
"epoch": 0.49,
"grad_norm": 0.6346720457077026,
"learning_rate": 4.6837239720382426e-05,
"loss": 0.8398,
"step": 3045
},
{
"epoch": 0.49,
"grad_norm": 0.8065741062164307,
"learning_rate": 4.682706009558297e-05,
"loss": 0.9325,
"step": 3050
},
{
"epoch": 0.49,
"grad_norm": 0.3001660406589508,
"learning_rate": 4.681686522498018e-05,
"loss": 0.8997,
"step": 3055
},
{
"epoch": 0.49,
"grad_norm": 0.860211968421936,
"learning_rate": 4.680665511569501e-05,
"loss": 0.6883,
"step": 3060
},
{
"epoch": 0.49,
"grad_norm": 0.722518265247345,
"learning_rate": 4.6796429774859015e-05,
"loss": 0.8607,
"step": 3065
},
{
"epoch": 0.49,
"grad_norm": 0.6525880694389343,
"learning_rate": 4.678618920961442e-05,
"loss": 0.9256,
"step": 3070
},
{
"epoch": 0.49,
"grad_norm": 0.7581719756126404,
"learning_rate": 4.6775933427114084e-05,
"loss": 0.662,
"step": 3075
},
{
"epoch": 0.49,
"grad_norm": 0.6604760885238647,
"learning_rate": 4.676566243452146e-05,
"loss": 0.734,
"step": 3080
},
{
"epoch": 0.49,
"grad_norm": 0.7573785781860352,
"learning_rate": 4.6755376239010665e-05,
"loss": 0.7113,
"step": 3085
},
{
"epoch": 0.49,
"grad_norm": 0.8933848738670349,
"learning_rate": 4.674507484776641e-05,
"loss": 0.8523,
"step": 3090
},
{
"epoch": 0.49,
"grad_norm": 0.5941946506500244,
"learning_rate": 4.6734758267984044e-05,
"loss": 0.7907,
"step": 3095
},
{
"epoch": 0.49,
"grad_norm": 0.7756261825561523,
"learning_rate": 4.672442650686949e-05,
"loss": 0.8407,
"step": 3100
},
{
"epoch": 0.49,
"eval_loss": 0.7495761513710022,
"eval_runtime": 96.4482,
"eval_samples_per_second": 7.227,
"eval_steps_per_second": 7.227,
"step": 3100
},
{
"epoch": 0.5,
"grad_norm": 0.6407367587089539,
"learning_rate": 4.671407957163931e-05,
"loss": 0.6413,
"step": 3105
},
{
"epoch": 0.5,
"grad_norm": 1.069754719734192,
"learning_rate": 4.670371746952063e-05,
"loss": 0.8934,
"step": 3110
},
{
"epoch": 0.5,
"grad_norm": 0.9698624610900879,
"learning_rate": 4.669334020775122e-05,
"loss": 0.7261,
"step": 3115
},
{
"epoch": 0.5,
"grad_norm": 0.6487118005752563,
"learning_rate": 4.668294779357938e-05,
"loss": 0.8951,
"step": 3120
},
{
"epoch": 0.5,
"grad_norm": 1.0640240907669067,
"learning_rate": 4.667254023426404e-05,
"loss": 0.8568,
"step": 3125
},
{
"epoch": 0.5,
"grad_norm": 0.5772892236709595,
"learning_rate": 4.666211753707468e-05,
"loss": 0.9798,
"step": 3130
},
{
"epoch": 0.5,
"grad_norm": 0.6915898323059082,
"learning_rate": 4.665167970929137e-05,
"loss": 0.8694,
"step": 3135
},
{
"epoch": 0.5,
"grad_norm": 0.5959879159927368,
"learning_rate": 4.664122675820474e-05,
"loss": 0.6521,
"step": 3140
},
{
"epoch": 0.5,
"grad_norm": 0.833991289138794,
"learning_rate": 4.663075869111597e-05,
"loss": 0.9194,
"step": 3145
},
{
"epoch": 0.5,
"grad_norm": 0.9575549960136414,
"learning_rate": 4.662027551533685e-05,
"loss": 1.0088,
"step": 3150
},
{
"epoch": 0.5,
"grad_norm": 0.5501818656921387,
"learning_rate": 4.660977723818965e-05,
"loss": 0.5997,
"step": 3155
},
{
"epoch": 0.5,
"grad_norm": 0.6001989245414734,
"learning_rate": 4.659926386700725e-05,
"loss": 0.7643,
"step": 3160
},
{
"epoch": 0.51,
"grad_norm": 0.6806654930114746,
"learning_rate": 4.658873540913303e-05,
"loss": 0.899,
"step": 3165
},
{
"epoch": 0.51,
"grad_norm": 0.7098959684371948,
"learning_rate": 4.657819187192094e-05,
"loss": 1.0281,
"step": 3170
},
{
"epoch": 0.51,
"grad_norm": 0.9234817028045654,
"learning_rate": 4.6567633262735446e-05,
"loss": 0.9495,
"step": 3175
},
{
"epoch": 0.51,
"grad_norm": 0.596527636051178,
"learning_rate": 4.655705958895153e-05,
"loss": 0.6352,
"step": 3180
},
{
"epoch": 0.51,
"grad_norm": 1.154539704322815,
"learning_rate": 4.6546470857954736e-05,
"loss": 0.8939,
"step": 3185
},
{
"epoch": 0.51,
"grad_norm": 0.7502239942550659,
"learning_rate": 4.653586707714108e-05,
"loss": 0.692,
"step": 3190
},
{
"epoch": 0.51,
"grad_norm": 0.7868794202804565,
"learning_rate": 4.652524825391711e-05,
"loss": 0.908,
"step": 3195
},
{
"epoch": 0.51,
"grad_norm": 0.6436206102371216,
"learning_rate": 4.6514614395699886e-05,
"loss": 0.7417,
"step": 3200
},
{
"epoch": 0.51,
"eval_loss": 0.7466740012168884,
"eval_runtime": 96.4309,
"eval_samples_per_second": 7.228,
"eval_steps_per_second": 7.228,
"step": 3200
},
{
"epoch": 0.51,
"grad_norm": 0.8566870093345642,
"learning_rate": 4.6503965509916956e-05,
"loss": 0.8041,
"step": 3205
},
{
"epoch": 0.51,
"grad_norm": 0.5982272028923035,
"learning_rate": 4.649330160400639e-05,
"loss": 0.4528,
"step": 3210
},
{
"epoch": 0.51,
"grad_norm": 0.6511960029602051,
"learning_rate": 4.648262268541671e-05,
"loss": 0.877,
"step": 3215
},
{
"epoch": 0.51,
"grad_norm": 0.8476071357727051,
"learning_rate": 4.6471928761606965e-05,
"loss": 0.7145,
"step": 3220
},
{
"epoch": 0.51,
"grad_norm": 1.0408881902694702,
"learning_rate": 4.6461219840046654e-05,
"loss": 0.5539,
"step": 3225
},
{
"epoch": 0.52,
"grad_norm": 0.7445903420448303,
"learning_rate": 4.645049592821577e-05,
"loss": 0.8306,
"step": 3230
},
{
"epoch": 0.52,
"grad_norm": 0.9672279357910156,
"learning_rate": 4.6439757033604756e-05,
"loss": 0.8645,
"step": 3235
},
{
"epoch": 0.52,
"grad_norm": 0.7082134485244751,
"learning_rate": 4.6429003163714556e-05,
"loss": 0.8188,
"step": 3240
},
{
"epoch": 0.52,
"grad_norm": 0.8803107142448425,
"learning_rate": 4.641823432605654e-05,
"loss": 0.7956,
"step": 3245
},
{
"epoch": 0.52,
"grad_norm": 0.7926101088523865,
"learning_rate": 4.640745052815254e-05,
"loss": 0.715,
"step": 3250
},
{
"epoch": 0.52,
"grad_norm": 0.890519380569458,
"learning_rate": 4.639665177753485e-05,
"loss": 0.8825,
"step": 3255
},
{
"epoch": 0.52,
"grad_norm": 0.9909971952438354,
"learning_rate": 4.638583808174619e-05,
"loss": 0.7843,
"step": 3260
},
{
"epoch": 0.52,
"grad_norm": 0.7450726628303528,
"learning_rate": 4.6375009448339743e-05,
"loss": 0.9714,
"step": 3265
},
{
"epoch": 0.52,
"grad_norm": 0.8220781683921814,
"learning_rate": 4.636416588487911e-05,
"loss": 0.8467,
"step": 3270
},
{
"epoch": 0.52,
"grad_norm": 1.025499701499939,
"learning_rate": 4.63533073989383e-05,
"loss": 0.9301,
"step": 3275
},
{
"epoch": 0.52,
"grad_norm": 0.8067827820777893,
"learning_rate": 4.634243399810181e-05,
"loss": 0.7078,
"step": 3280
},
{
"epoch": 0.52,
"grad_norm": 0.8833619952201843,
"learning_rate": 4.6331545689964475e-05,
"loss": 0.699,
"step": 3285
},
{
"epoch": 0.53,
"grad_norm": 1.0600448846817017,
"learning_rate": 4.632064248213159e-05,
"loss": 0.7849,
"step": 3290
},
{
"epoch": 0.53,
"grad_norm": 1.0503095388412476,
"learning_rate": 4.630972438221885e-05,
"loss": 0.6215,
"step": 3295
},
{
"epoch": 0.53,
"grad_norm": 0.5159885287284851,
"learning_rate": 4.629879139785235e-05,
"loss": 0.7449,
"step": 3300
},
{
"epoch": 0.53,
"eval_loss": 0.7472941279411316,
"eval_runtime": 96.4994,
"eval_samples_per_second": 7.223,
"eval_steps_per_second": 7.223,
"step": 3300
},
{
"epoch": 0.53,
"grad_norm": 1.072464108467102,
"learning_rate": 4.6287843536668575e-05,
"loss": 0.8511,
"step": 3305
},
{
"epoch": 0.53,
"grad_norm": 0.9016098976135254,
"learning_rate": 4.62768808063144e-05,
"loss": 0.7373,
"step": 3310
},
{
"epoch": 0.53,
"grad_norm": 1.0161947011947632,
"learning_rate": 4.626590321444712e-05,
"loss": 0.9035,
"step": 3315
},
{
"epoch": 0.53,
"grad_norm": 0.7459146976470947,
"learning_rate": 4.625491076873435e-05,
"loss": 0.6468,
"step": 3320
},
{
"epoch": 0.53,
"grad_norm": 0.950080394744873,
"learning_rate": 4.624390347685413e-05,
"loss": 0.7211,
"step": 3325
},
{
"epoch": 0.53,
"grad_norm": 0.7308927774429321,
"learning_rate": 4.623288134649485e-05,
"loss": 0.9238,
"step": 3330
},
{
"epoch": 0.53,
"grad_norm": 0.7227129340171814,
"learning_rate": 4.622184438535527e-05,
"loss": 0.9773,
"step": 3335
},
{
"epoch": 0.53,
"grad_norm": 0.7054020166397095,
"learning_rate": 4.62107926011445e-05,
"loss": 0.7783,
"step": 3340
},
{
"epoch": 0.53,
"grad_norm": 0.6535981297492981,
"learning_rate": 4.619972600158201e-05,
"loss": 0.6559,
"step": 3345
},
{
"epoch": 0.53,
"grad_norm": 0.7245693206787109,
"learning_rate": 4.618864459439762e-05,
"loss": 0.8352,
"step": 3350
},
{
"epoch": 0.54,
"grad_norm": 0.9683626890182495,
"learning_rate": 4.6177548387331485e-05,
"loss": 0.9397,
"step": 3355
},
{
"epoch": 0.54,
"grad_norm": 1.1847660541534424,
"learning_rate": 4.616643738813411e-05,
"loss": 0.7383,
"step": 3360
},
{
"epoch": 0.54,
"grad_norm": 0.8566804528236389,
"learning_rate": 4.615531160456633e-05,
"loss": 0.8066,
"step": 3365
},
{
"epoch": 0.54,
"grad_norm": 0.7312522530555725,
"learning_rate": 4.61441710443993e-05,
"loss": 0.7974,
"step": 3370
},
{
"epoch": 0.54,
"grad_norm": 0.6620572209358215,
"learning_rate": 4.6133015715414484e-05,
"loss": 0.9136,
"step": 3375
},
{
"epoch": 0.54,
"grad_norm": 0.5405072569847107,
"learning_rate": 4.612184562540369e-05,
"loss": 0.6921,
"step": 3380
},
{
"epoch": 0.54,
"grad_norm": 0.7474086284637451,
"learning_rate": 4.611066078216901e-05,
"loss": 0.8463,
"step": 3385
},
{
"epoch": 0.54,
"grad_norm": 0.9125152230262756,
"learning_rate": 4.609946119352287e-05,
"loss": 0.8508,
"step": 3390
},
{
"epoch": 0.54,
"grad_norm": 0.9998400211334229,
"learning_rate": 4.608824686728797e-05,
"loss": 0.8735,
"step": 3395
},
{
"epoch": 0.54,
"grad_norm": 0.5990025401115417,
"learning_rate": 4.6077017811297304e-05,
"loss": 0.8562,
"step": 3400
},
{
"epoch": 0.54,
"eval_loss": 0.743736743927002,
"eval_runtime": 96.3748,
"eval_samples_per_second": 7.232,
"eval_steps_per_second": 7.232,
"step": 3400
},
{
"epoch": 0.54,
"grad_norm": 0.35676899552345276,
"learning_rate": 4.606577403339418e-05,
"loss": 0.8914,
"step": 3405
},
{
"epoch": 0.54,
"grad_norm": 0.772233784198761,
"learning_rate": 4.605451554143216e-05,
"loss": 0.779,
"step": 3410
},
{
"epoch": 0.55,
"grad_norm": 0.7336989641189575,
"learning_rate": 4.604324234327509e-05,
"loss": 0.7678,
"step": 3415
},
{
"epoch": 0.55,
"grad_norm": 0.7039794325828552,
"learning_rate": 4.603195444679711e-05,
"loss": 0.8783,
"step": 3420
},
{
"epoch": 0.55,
"grad_norm": 0.6955629587173462,
"learning_rate": 4.602065185988259e-05,
"loss": 0.818,
"step": 3425
},
{
"epoch": 0.55,
"grad_norm": 0.7369412779808044,
"learning_rate": 4.60093345904262e-05,
"loss": 0.6942,
"step": 3430
},
{
"epoch": 0.55,
"grad_norm": 0.6824669241905212,
"learning_rate": 4.5998002646332835e-05,
"loss": 0.9274,
"step": 3435
},
{
"epoch": 0.55,
"grad_norm": 1.355720043182373,
"learning_rate": 4.598665603551765e-05,
"loss": 0.7219,
"step": 3440
},
{
"epoch": 0.55,
"grad_norm": 0.8629677295684814,
"learning_rate": 4.597529476590605e-05,
"loss": 0.8023,
"step": 3445
},
{
"epoch": 0.55,
"grad_norm": 2.2956135272979736,
"learning_rate": 4.596391884543368e-05,
"loss": 0.9574,
"step": 3450
},
{
"epoch": 0.55,
"grad_norm": 1.2683722972869873,
"learning_rate": 4.59525282820464e-05,
"loss": 0.6996,
"step": 3455
},
{
"epoch": 0.55,
"grad_norm": 0.7317371368408203,
"learning_rate": 4.594112308370032e-05,
"loss": 1.03,
"step": 3460
},
{
"epoch": 0.55,
"grad_norm": 1.0310641527175903,
"learning_rate": 4.5929703258361756e-05,
"loss": 0.6917,
"step": 3465
},
{
"epoch": 0.55,
"grad_norm": 0.9479489326477051,
"learning_rate": 4.591826881400726e-05,
"loss": 0.9939,
"step": 3470
},
{
"epoch": 0.55,
"grad_norm": 0.9485552310943604,
"learning_rate": 4.5906819758623576e-05,
"loss": 1.0317,
"step": 3475
},
{
"epoch": 0.56,
"grad_norm": 0.724987268447876,
"learning_rate": 4.589535610020765e-05,
"loss": 0.6915,
"step": 3480
},
{
"epoch": 0.56,
"grad_norm": 0.7091718316078186,
"learning_rate": 4.5883877846766654e-05,
"loss": 0.8673,
"step": 3485
},
{
"epoch": 0.56,
"grad_norm": 0.8297457098960876,
"learning_rate": 4.587238500631793e-05,
"loss": 0.8114,
"step": 3490
},
{
"epoch": 0.56,
"grad_norm": 0.7213541269302368,
"learning_rate": 4.586087758688903e-05,
"loss": 0.863,
"step": 3495
},
{
"epoch": 0.56,
"grad_norm": 1.1096009016036987,
"learning_rate": 4.584935559651765e-05,
"loss": 0.9222,
"step": 3500
},
{
"epoch": 0.56,
"eval_loss": 0.7428527474403381,
"eval_runtime": 96.3993,
"eval_samples_per_second": 7.23,
"eval_steps_per_second": 7.23,
"step": 3500
},
{
"epoch": 0.56,
"grad_norm": 0.5834380984306335,
"learning_rate": 4.583781904325172e-05,
"loss": 0.6609,
"step": 3505
},
{
"epoch": 0.56,
"grad_norm": 0.5797068476676941,
"learning_rate": 4.5826267935149285e-05,
"loss": 0.7933,
"step": 3510
},
{
"epoch": 0.56,
"grad_norm": 0.6027450561523438,
"learning_rate": 4.581470228027861e-05,
"loss": 0.7841,
"step": 3515
},
{
"epoch": 0.56,
"grad_norm": 0.5674509406089783,
"learning_rate": 4.5803122086718077e-05,
"loss": 0.7721,
"step": 3520
},
{
"epoch": 0.56,
"grad_norm": 0.7398461103439331,
"learning_rate": 4.5791527362556235e-05,
"loss": 0.7651,
"step": 3525
},
{
"epoch": 0.56,
"grad_norm": 0.6617181301116943,
"learning_rate": 4.577991811589181e-05,
"loss": 0.9359,
"step": 3530
},
{
"epoch": 0.56,
"grad_norm": 0.49279505014419556,
"learning_rate": 4.576829435483362e-05,
"loss": 0.6278,
"step": 3535
},
{
"epoch": 0.57,
"grad_norm": 0.5201964378356934,
"learning_rate": 4.575665608750067e-05,
"loss": 0.853,
"step": 3540
},
{
"epoch": 0.57,
"grad_norm": 0.7188725471496582,
"learning_rate": 4.5745003322022084e-05,
"loss": 0.8338,
"step": 3545
},
{
"epoch": 0.57,
"grad_norm": 1.0798031091690063,
"learning_rate": 4.573333606653708e-05,
"loss": 0.9776,
"step": 3550
},
{
"epoch": 0.57,
"grad_norm": 0.6439509987831116,
"learning_rate": 4.5721654329195046e-05,
"loss": 0.9331,
"step": 3555
},
{
"epoch": 0.57,
"grad_norm": 0.7663920521736145,
"learning_rate": 4.570995811815545e-05,
"loss": 1.0533,
"step": 3560
},
{
"epoch": 0.57,
"grad_norm": 0.7230969071388245,
"learning_rate": 4.569824744158789e-05,
"loss": 0.6966,
"step": 3565
},
{
"epoch": 0.57,
"grad_norm": 1.016112208366394,
"learning_rate": 4.568652230767205e-05,
"loss": 0.8393,
"step": 3570
},
{
"epoch": 0.57,
"grad_norm": 1.0165222883224487,
"learning_rate": 4.567478272459773e-05,
"loss": 1.0218,
"step": 3575
},
{
"epoch": 0.57,
"grad_norm": 0.709685742855072,
"learning_rate": 4.5663028700564826e-05,
"loss": 0.7273,
"step": 3580
},
{
"epoch": 0.57,
"grad_norm": 0.5664321780204773,
"learning_rate": 4.565126024378328e-05,
"loss": 0.9079,
"step": 3585
},
{
"epoch": 0.57,
"grad_norm": 0.7938306927680969,
"learning_rate": 4.5639477362473173e-05,
"loss": 0.976,
"step": 3590
},
{
"epoch": 0.57,
"grad_norm": 0.6710417866706848,
"learning_rate": 4.5627680064864606e-05,
"loss": 1.1969,
"step": 3595
},
{
"epoch": 0.57,
"grad_norm": 0.9886580109596252,
"learning_rate": 4.5615868359197796e-05,
"loss": 0.9242,
"step": 3600
},
{
"epoch": 0.57,
"eval_loss": 0.7412505149841309,
"eval_runtime": 96.4,
"eval_samples_per_second": 7.23,
"eval_steps_per_second": 7.23,
"step": 3600
},
{
"epoch": 0.58,
"grad_norm": 0.8157562613487244,
"learning_rate": 4.5604042253723014e-05,
"loss": 0.8398,
"step": 3605
},
{
"epoch": 0.58,
"grad_norm": 1.2368131875991821,
"learning_rate": 4.559220175670054e-05,
"loss": 0.8742,
"step": 3610
},
{
"epoch": 0.58,
"grad_norm": 0.6060155034065247,
"learning_rate": 4.558034687640078e-05,
"loss": 0.6993,
"step": 3615
},
{
"epoch": 0.58,
"grad_norm": 1.1366558074951172,
"learning_rate": 4.556847762110415e-05,
"loss": 0.9328,
"step": 3620
},
{
"epoch": 0.58,
"grad_norm": 0.7205525636672974,
"learning_rate": 4.555659399910108e-05,
"loss": 0.827,
"step": 3625
},
{
"epoch": 0.58,
"grad_norm": 0.6944175958633423,
"learning_rate": 4.554469601869209e-05,
"loss": 0.7805,
"step": 3630
},
{
"epoch": 0.58,
"grad_norm": 0.6939406394958496,
"learning_rate": 4.55327836881877e-05,
"loss": 0.7996,
"step": 3635
},
{
"epoch": 0.58,
"grad_norm": 0.592650830745697,
"learning_rate": 4.552085701590844e-05,
"loss": 0.6599,
"step": 3640
},
{
"epoch": 0.58,
"grad_norm": 0.5287877321243286,
"learning_rate": 4.5508916010184884e-05,
"loss": 0.6856,
"step": 3645
},
{
"epoch": 0.58,
"grad_norm": 0.6414081454277039,
"learning_rate": 4.549696067935762e-05,
"loss": 0.7622,
"step": 3650
},
{
"epoch": 0.58,
"grad_norm": 1.2272289991378784,
"learning_rate": 4.548499103177719e-05,
"loss": 1.0834,
"step": 3655
},
{
"epoch": 0.58,
"grad_norm": 0.5912505388259888,
"learning_rate": 4.547300707580422e-05,
"loss": 0.8738,
"step": 3660
},
{
"epoch": 0.58,
"grad_norm": 0.6686813235282898,
"learning_rate": 4.5461008819809246e-05,
"loss": 0.6221,
"step": 3665
},
{
"epoch": 0.59,
"grad_norm": 0.891153872013092,
"learning_rate": 4.544899627217286e-05,
"loss": 0.9009,
"step": 3670
},
{
"epoch": 0.59,
"grad_norm": 1.1651557683944702,
"learning_rate": 4.543696944128559e-05,
"loss": 0.8448,
"step": 3675
},
{
"epoch": 0.59,
"grad_norm": 0.7525443434715271,
"learning_rate": 4.5424928335547964e-05,
"loss": 0.6654,
"step": 3680
},
{
"epoch": 0.59,
"grad_norm": 0.6798614859580994,
"learning_rate": 4.541287296337048e-05,
"loss": 0.9244,
"step": 3685
},
{
"epoch": 0.59,
"grad_norm": 0.498735249042511,
"learning_rate": 4.540080333317358e-05,
"loss": 0.6815,
"step": 3690
},
{
"epoch": 0.59,
"grad_norm": 0.6097673773765564,
"learning_rate": 4.5388719453387694e-05,
"loss": 0.8536,
"step": 3695
},
{
"epoch": 0.59,
"grad_norm": 0.6685522198677063,
"learning_rate": 4.537662133245319e-05,
"loss": 0.8092,
"step": 3700
},
{
"epoch": 0.59,
"eval_loss": 0.7402560114860535,
"eval_runtime": 96.4998,
"eval_samples_per_second": 7.223,
"eval_steps_per_second": 7.223,
"step": 3700
},
{
"epoch": 0.59,
"grad_norm": 1.162788987159729,
"learning_rate": 4.5364508978820375e-05,
"loss": 0.6143,
"step": 3705
},
{
"epoch": 0.59,
"grad_norm": 0.8281823992729187,
"learning_rate": 4.5352382400949524e-05,
"loss": 0.8143,
"step": 3710
},
{
"epoch": 0.59,
"grad_norm": 0.6465135812759399,
"learning_rate": 4.534024160731082e-05,
"loss": 0.9152,
"step": 3715
},
{
"epoch": 0.59,
"grad_norm": 0.5903899669647217,
"learning_rate": 4.532808660638438e-05,
"loss": 0.7229,
"step": 3720
},
{
"epoch": 0.59,
"grad_norm": 0.6988681554794312,
"learning_rate": 4.5315917406660265e-05,
"loss": 0.6863,
"step": 3725
},
{
"epoch": 0.6,
"grad_norm": 0.7910459637641907,
"learning_rate": 4.530373401663843e-05,
"loss": 0.8762,
"step": 3730
},
{
"epoch": 0.6,
"grad_norm": 0.7580087184906006,
"learning_rate": 4.529153644482875e-05,
"loss": 0.9896,
"step": 3735
},
{
"epoch": 0.6,
"grad_norm": 0.6871665716171265,
"learning_rate": 4.5279324699751005e-05,
"loss": 0.8831,
"step": 3740
},
{
"epoch": 0.6,
"grad_norm": 1.0093677043914795,
"learning_rate": 4.526709878993488e-05,
"loss": 0.742,
"step": 3745
},
{
"epoch": 0.6,
"grad_norm": 0.9898921847343445,
"learning_rate": 4.525485872391996e-05,
"loss": 0.766,
"step": 3750
},
{
"epoch": 0.6,
"grad_norm": 0.8706837296485901,
"learning_rate": 4.524260451025569e-05,
"loss": 0.7545,
"step": 3755
},
{
"epoch": 0.6,
"grad_norm": 1.1715607643127441,
"learning_rate": 4.523033615750142e-05,
"loss": 0.84,
"step": 3760
},
{
"epoch": 0.6,
"grad_norm": 1.017062783241272,
"learning_rate": 4.521805367422638e-05,
"loss": 0.7477,
"step": 3765
},
{
"epoch": 0.6,
"grad_norm": 0.6071624159812927,
"learning_rate": 4.520575706900965e-05,
"loss": 0.793,
"step": 3770
},
{
"epoch": 0.6,
"grad_norm": 0.5821404457092285,
"learning_rate": 4.519344635044018e-05,
"loss": 0.7514,
"step": 3775
},
{
"epoch": 0.6,
"grad_norm": 0.6849238872528076,
"learning_rate": 4.51811215271168e-05,
"loss": 0.862,
"step": 3780
},
{
"epoch": 0.6,
"grad_norm": 0.8808868527412415,
"learning_rate": 4.5168782607648166e-05,
"loss": 0.7189,
"step": 3785
},
{
"epoch": 0.6,
"grad_norm": 0.7080340385437012,
"learning_rate": 4.5156429600652774e-05,
"loss": 0.6987,
"step": 3790
},
{
"epoch": 0.61,
"grad_norm": 0.705869734287262,
"learning_rate": 4.5144062514759e-05,
"loss": 0.6482,
"step": 3795
},
{
"epoch": 0.61,
"grad_norm": 0.6345694065093994,
"learning_rate": 4.5131681358605007e-05,
"loss": 0.7279,
"step": 3800
},
{
"epoch": 0.61,
"eval_loss": 0.7394095063209534,
"eval_runtime": 96.4977,
"eval_samples_per_second": 7.223,
"eval_steps_per_second": 7.223,
"step": 3800
},
{
"epoch": 0.61,
"grad_norm": 0.748913586139679,
"learning_rate": 4.511928614083881e-05,
"loss": 0.7474,
"step": 3805
},
{
"epoch": 0.61,
"grad_norm": 0.6260043382644653,
"learning_rate": 4.5106876870118255e-05,
"loss": 0.7469,
"step": 3810
},
{
"epoch": 0.61,
"grad_norm": 0.596367597579956,
"learning_rate": 4.509445355511098e-05,
"loss": 0.8437,
"step": 3815
},
{
"epoch": 0.61,
"grad_norm": 1.3925014734268188,
"learning_rate": 4.5082016204494445e-05,
"loss": 1.0928,
"step": 3820
},
{
"epoch": 0.61,
"grad_norm": 1.1370338201522827,
"learning_rate": 4.506956482695592e-05,
"loss": 0.8908,
"step": 3825
},
{
"epoch": 0.61,
"grad_norm": 0.6746950149536133,
"learning_rate": 4.505709943119246e-05,
"loss": 0.7121,
"step": 3830
},
{
"epoch": 0.61,
"grad_norm": 0.6608826518058777,
"learning_rate": 4.504462002591091e-05,
"loss": 0.9397,
"step": 3835
},
{
"epoch": 0.61,
"grad_norm": 0.6542508006095886,
"learning_rate": 4.5032126619827916e-05,
"loss": 0.6942,
"step": 3840
},
{
"epoch": 0.61,
"grad_norm": 0.5825070738792419,
"learning_rate": 4.5019619221669895e-05,
"loss": 0.7083,
"step": 3845
},
{
"epoch": 0.61,
"grad_norm": 0.8596588373184204,
"learning_rate": 4.500709784017303e-05,
"loss": 0.839,
"step": 3850
},
{
"epoch": 0.62,
"grad_norm": 0.641009509563446,
"learning_rate": 4.499456248408328e-05,
"loss": 0.72,
"step": 3855
},
{
"epoch": 0.62,
"grad_norm": 1.213782548904419,
"learning_rate": 4.498201316215635e-05,
"loss": 0.7116,
"step": 3860
},
{
"epoch": 0.62,
"grad_norm": 1.1411411762237549,
"learning_rate": 4.496944988315775e-05,
"loss": 1.0208,
"step": 3865
},
{
"epoch": 0.62,
"grad_norm": 0.8265553712844849,
"learning_rate": 4.495687265586266e-05,
"loss": 0.7664,
"step": 3870
},
{
"epoch": 0.62,
"grad_norm": 0.9309681057929993,
"learning_rate": 4.4944281489056065e-05,
"loss": 0.9126,
"step": 3875
},
{
"epoch": 0.62,
"grad_norm": 0.49171608686447144,
"learning_rate": 4.493167639153266e-05,
"loss": 0.6271,
"step": 3880
},
{
"epoch": 0.62,
"grad_norm": 0.743669867515564,
"learning_rate": 4.491905737209688e-05,
"loss": 0.7965,
"step": 3885
},
{
"epoch": 0.62,
"grad_norm": 0.6191633939743042,
"learning_rate": 4.490642443956287e-05,
"loss": 0.5884,
"step": 3890
},
{
"epoch": 0.62,
"grad_norm": 0.5481441020965576,
"learning_rate": 4.489377760275452e-05,
"loss": 0.6281,
"step": 3895
},
{
"epoch": 0.62,
"grad_norm": 0.7155417203903198,
"learning_rate": 4.488111687050539e-05,
"loss": 0.7774,
"step": 3900
},
{
"epoch": 0.62,
"eval_loss": 0.738506019115448,
"eval_runtime": 96.7667,
"eval_samples_per_second": 7.203,
"eval_steps_per_second": 7.203,
"step": 3900
},
{
"epoch": 0.62,
"grad_norm": 1.032523274421692,
"learning_rate": 4.4868442251658795e-05,
"loss": 0.7621,
"step": 3905
},
{
"epoch": 0.62,
"grad_norm": 0.584082841873169,
"learning_rate": 4.4855753755067703e-05,
"loss": 0.6617,
"step": 3910
},
{
"epoch": 0.62,
"grad_norm": 0.7214722037315369,
"learning_rate": 4.4843051389594814e-05,
"loss": 0.8669,
"step": 3915
},
{
"epoch": 0.63,
"grad_norm": 0.6019904613494873,
"learning_rate": 4.4830335164112504e-05,
"loss": 0.736,
"step": 3920
},
{
"epoch": 0.63,
"grad_norm": 0.8038384318351746,
"learning_rate": 4.48176050875028e-05,
"loss": 0.637,
"step": 3925
},
{
"epoch": 0.63,
"grad_norm": 0.9631878733634949,
"learning_rate": 4.4804861168657455e-05,
"loss": 0.9722,
"step": 3930
},
{
"epoch": 0.63,
"grad_norm": 0.5342935919761658,
"learning_rate": 4.4792103416477836e-05,
"loss": 0.8081,
"step": 3935
},
{
"epoch": 0.63,
"grad_norm": 0.5893488526344299,
"learning_rate": 4.477933183987503e-05,
"loss": 0.61,
"step": 3940
},
{
"epoch": 0.63,
"grad_norm": 1.388850212097168,
"learning_rate": 4.476654644776973e-05,
"loss": 0.8454,
"step": 3945
},
{
"epoch": 0.63,
"grad_norm": 0.6928623914718628,
"learning_rate": 4.4753747249092305e-05,
"loss": 0.7209,
"step": 3950
},
{
"epoch": 0.63,
"grad_norm": 1.2383430004119873,
"learning_rate": 4.4740934252782757e-05,
"loss": 0.8205,
"step": 3955
},
{
"epoch": 0.63,
"grad_norm": 0.6005001664161682,
"learning_rate": 4.472810746779074e-05,
"loss": 0.6083,
"step": 3960
},
{
"epoch": 0.63,
"grad_norm": 0.7928474545478821,
"learning_rate": 4.471526690307552e-05,
"loss": 0.9735,
"step": 3965
},
{
"epoch": 0.63,
"grad_norm": 0.8710891008377075,
"learning_rate": 4.4702412567606014e-05,
"loss": 0.7573,
"step": 3970
},
{
"epoch": 0.63,
"grad_norm": 0.6327987313270569,
"learning_rate": 4.468954447036071e-05,
"loss": 0.8563,
"step": 3975
},
{
"epoch": 0.64,
"grad_norm": 0.7048762440681458,
"learning_rate": 4.467666262032777e-05,
"loss": 0.9176,
"step": 3980
},
{
"epoch": 0.64,
"grad_norm": 0.6058861017227173,
"learning_rate": 4.466376702650492e-05,
"loss": 0.5525,
"step": 3985
},
{
"epoch": 0.64,
"grad_norm": 0.637993574142456,
"learning_rate": 4.465085769789949e-05,
"loss": 0.7256,
"step": 3990
},
{
"epoch": 0.64,
"grad_norm": 0.6992897987365723,
"learning_rate": 4.463793464352842e-05,
"loss": 0.8824,
"step": 3995
},
{
"epoch": 0.64,
"grad_norm": 0.7812734246253967,
"learning_rate": 4.462499787241822e-05,
"loss": 0.8942,
"step": 4000
},
{
"epoch": 0.64,
"eval_loss": 0.7363680601119995,
"eval_runtime": 96.9231,
"eval_samples_per_second": 7.191,
"eval_steps_per_second": 7.191,
"step": 4000
},
{
"epoch": 0.64,
"grad_norm": 0.907598078250885,
"learning_rate": 4.4612047393605e-05,
"loss": 0.867,
"step": 4005
},
{
"epoch": 0.64,
"grad_norm": 0.9081722497940063,
"learning_rate": 4.459908321613442e-05,
"loss": 0.8757,
"step": 4010
},
{
"epoch": 0.64,
"grad_norm": 0.5538048148155212,
"learning_rate": 4.4586105349061726e-05,
"loss": 0.6709,
"step": 4015
},
{
"epoch": 0.64,
"grad_norm": 0.6632833480834961,
"learning_rate": 4.457311380145173e-05,
"loss": 0.8362,
"step": 4020
},
{
"epoch": 0.64,
"grad_norm": 0.8646539449691772,
"learning_rate": 4.4560108582378766e-05,
"loss": 0.8527,
"step": 4025
},
{
"epoch": 0.64,
"grad_norm": 0.6309005618095398,
"learning_rate": 4.454708970092678e-05,
"loss": 0.595,
"step": 4030
},
{
"epoch": 0.64,
"grad_norm": 0.5711541175842285,
"learning_rate": 4.45340571661892e-05,
"loss": 0.8069,
"step": 4035
},
{
"epoch": 0.64,
"grad_norm": 1.1379880905151367,
"learning_rate": 4.4521010987269006e-05,
"loss": 0.8464,
"step": 4040
},
{
"epoch": 0.65,
"grad_norm": 0.6005469560623169,
"learning_rate": 4.450795117327874e-05,
"loss": 0.5801,
"step": 4045
},
{
"epoch": 0.65,
"grad_norm": 0.7842866778373718,
"learning_rate": 4.449487773334042e-05,
"loss": 0.6238,
"step": 4050
},
{
"epoch": 0.65,
"grad_norm": 0.7519890069961548,
"learning_rate": 4.448179067658563e-05,
"loss": 1.1255,
"step": 4055
},
{
"epoch": 0.65,
"grad_norm": 0.5955212712287903,
"learning_rate": 4.446869001215542e-05,
"loss": 0.7738,
"step": 4060
},
{
"epoch": 0.65,
"grad_norm": 0.5085921287536621,
"learning_rate": 4.4455575749200364e-05,
"loss": 0.6239,
"step": 4065
},
{
"epoch": 0.65,
"grad_norm": 0.779778003692627,
"learning_rate": 4.444244789688056e-05,
"loss": 0.9719,
"step": 4070
},
{
"epoch": 0.65,
"grad_norm": 0.7279208898544312,
"learning_rate": 4.442930646436554e-05,
"loss": 0.9854,
"step": 4075
},
{
"epoch": 0.65,
"grad_norm": 0.9218065738677979,
"learning_rate": 4.4416151460834376e-05,
"loss": 0.8096,
"step": 4080
},
{
"epoch": 0.65,
"grad_norm": 0.7595914006233215,
"learning_rate": 4.44029828954756e-05,
"loss": 0.7955,
"step": 4085
},
{
"epoch": 0.65,
"grad_norm": 0.785493016242981,
"learning_rate": 4.43898007774872e-05,
"loss": 0.8598,
"step": 4090
},
{
"epoch": 0.65,
"grad_norm": 0.5540453195571899,
"learning_rate": 4.437660511607666e-05,
"loss": 0.8485,
"step": 4095
},
{
"epoch": 0.65,
"grad_norm": 0.7215760350227356,
"learning_rate": 4.43633959204609e-05,
"loss": 0.9286,
"step": 4100
},
{
"epoch": 0.65,
"eval_loss": 0.7347923517227173,
"eval_runtime": 96.8658,
"eval_samples_per_second": 7.196,
"eval_steps_per_second": 7.196,
"step": 4100
},
{
"epoch": 0.66,
"grad_norm": 0.7934743762016296,
"learning_rate": 4.435017319986631e-05,
"loss": 0.7829,
"step": 4105
},
{
"epoch": 0.66,
"grad_norm": 1.503614068031311,
"learning_rate": 4.43369369635287e-05,
"loss": 0.7203,
"step": 4110
},
{
"epoch": 0.66,
"grad_norm": 0.6292420625686646,
"learning_rate": 4.4323687220693365e-05,
"loss": 0.7556,
"step": 4115
},
{
"epoch": 0.66,
"grad_norm": 0.6981114149093628,
"learning_rate": 4.431042398061499e-05,
"loss": 0.6953,
"step": 4120
},
{
"epoch": 0.66,
"grad_norm": 0.8554514050483704,
"learning_rate": 4.4297147252557715e-05,
"loss": 0.7731,
"step": 4125
},
{
"epoch": 0.66,
"grad_norm": 1.1464003324508667,
"learning_rate": 4.428385704579509e-05,
"loss": 0.7761,
"step": 4130
},
{
"epoch": 0.66,
"grad_norm": 0.6772524118423462,
"learning_rate": 4.427055336961008e-05,
"loss": 0.7529,
"step": 4135
},
{
"epoch": 0.66,
"grad_norm": 0.5949820280075073,
"learning_rate": 4.425723623329507e-05,
"loss": 0.9164,
"step": 4140
},
{
"epoch": 0.66,
"grad_norm": 0.848900318145752,
"learning_rate": 4.4243905646151825e-05,
"loss": 0.8385,
"step": 4145
},
{
"epoch": 0.66,
"grad_norm": 0.7119936943054199,
"learning_rate": 4.4230561617491514e-05,
"loss": 0.6342,
"step": 4150
},
{
"epoch": 0.66,
"grad_norm": 0.4240078628063202,
"learning_rate": 4.421720415663472e-05,
"loss": 0.9921,
"step": 4155
},
{
"epoch": 0.66,
"grad_norm": 1.166399359703064,
"learning_rate": 4.4203833272911355e-05,
"loss": 0.6751,
"step": 4160
},
{
"epoch": 0.66,
"grad_norm": 0.7882303595542908,
"learning_rate": 4.4190448975660756e-05,
"loss": 0.8711,
"step": 4165
},
{
"epoch": 0.67,
"grad_norm": 0.7739405632019043,
"learning_rate": 4.417705127423162e-05,
"loss": 0.7635,
"step": 4170
},
{
"epoch": 0.67,
"grad_norm": 0.6729245781898499,
"learning_rate": 4.416364017798197e-05,
"loss": 1.0083,
"step": 4175
},
{
"epoch": 0.67,
"grad_norm": 0.7291648983955383,
"learning_rate": 4.4150215696279233e-05,
"loss": 0.9355,
"step": 4180
},
{
"epoch": 0.67,
"grad_norm": 0.569436252117157,
"learning_rate": 4.413677783850015e-05,
"loss": 0.5718,
"step": 4185
},
{
"epoch": 0.67,
"grad_norm": 0.7857233285903931,
"learning_rate": 4.412332661403085e-05,
"loss": 0.6356,
"step": 4190
},
{
"epoch": 0.67,
"grad_norm": 1.124894618988037,
"learning_rate": 4.410986203226672e-05,
"loss": 0.9911,
"step": 4195
},
{
"epoch": 0.67,
"grad_norm": 0.7088748216629028,
"learning_rate": 4.409638410261256e-05,
"loss": 0.7703,
"step": 4200
},
{
"epoch": 0.67,
"eval_loss": 0.7353793978691101,
"eval_runtime": 96.9146,
"eval_samples_per_second": 7.192,
"eval_steps_per_second": 7.192,
"step": 4200
},
{
"epoch": 0.67,
"grad_norm": 0.8883334398269653,
"learning_rate": 4.4082892834482456e-05,
"loss": 0.7829,
"step": 4205
},
{
"epoch": 0.67,
"grad_norm": 0.5809643864631653,
"learning_rate": 4.406938823729979e-05,
"loss": 0.79,
"step": 4210
},
{
"epoch": 0.67,
"grad_norm": 2.2371888160705566,
"learning_rate": 4.405587032049731e-05,
"loss": 0.9394,
"step": 4215
},
{
"epoch": 0.67,
"grad_norm": 0.6468964219093323,
"learning_rate": 4.4042339093517e-05,
"loss": 0.7621,
"step": 4220
},
{
"epoch": 0.67,
"grad_norm": 0.8613569736480713,
"learning_rate": 4.4028794565810194e-05,
"loss": 0.9303,
"step": 4225
},
{
"epoch": 0.68,
"grad_norm": 0.8210548162460327,
"learning_rate": 4.4015236746837505e-05,
"loss": 1.04,
"step": 4230
},
{
"epoch": 0.68,
"grad_norm": 0.8066801428794861,
"learning_rate": 4.4001665646068804e-05,
"loss": 0.9942,
"step": 4235
},
{
"epoch": 0.68,
"grad_norm": 0.6841477751731873,
"learning_rate": 4.3988081272983263e-05,
"loss": 0.6893,
"step": 4240
},
{
"epoch": 0.68,
"grad_norm": 0.7812705636024475,
"learning_rate": 4.3974483637069333e-05,
"loss": 0.9125,
"step": 4245
},
{
"epoch": 0.68,
"grad_norm": 0.7913382649421692,
"learning_rate": 4.3960872747824686e-05,
"loss": 0.9298,
"step": 4250
},
{
"epoch": 0.68,
"grad_norm": 0.6053805947303772,
"learning_rate": 4.394724861475631e-05,
"loss": 0.7055,
"step": 4255
},
{
"epoch": 0.68,
"grad_norm": 0.6879487633705139,
"learning_rate": 4.393361124738039e-05,
"loss": 0.605,
"step": 4260
},
{
"epoch": 0.68,
"grad_norm": 0.7929925918579102,
"learning_rate": 4.3919960655222394e-05,
"loss": 0.8569,
"step": 4265
},
{
"epoch": 0.68,
"grad_norm": 0.5888631939888,
"learning_rate": 4.390629684781701e-05,
"loss": 0.6246,
"step": 4270
},
{
"epoch": 0.68,
"grad_norm": 0.9546008706092834,
"learning_rate": 4.389261983470815e-05,
"loss": 0.7964,
"step": 4275
},
{
"epoch": 0.68,
"grad_norm": 0.620267391204834,
"learning_rate": 4.387892962544896e-05,
"loss": 0.7127,
"step": 4280
},
{
"epoch": 0.68,
"grad_norm": 0.7655039429664612,
"learning_rate": 4.3865226229601805e-05,
"loss": 0.6936,
"step": 4285
},
{
"epoch": 0.68,
"grad_norm": 0.5404471158981323,
"learning_rate": 4.3851509656738264e-05,
"loss": 0.6141,
"step": 4290
},
{
"epoch": 0.69,
"grad_norm": 0.9140282273292542,
"learning_rate": 4.38377799164391e-05,
"loss": 1.152,
"step": 4295
},
{
"epoch": 0.69,
"grad_norm": 0.4845621585845947,
"learning_rate": 4.382403701829429e-05,
"loss": 0.8322,
"step": 4300
},
{
"epoch": 0.69,
"eval_loss": 0.733027458190918,
"eval_runtime": 96.886,
"eval_samples_per_second": 7.194,
"eval_steps_per_second": 7.194,
"step": 4300
},
{
"epoch": 0.69,
"grad_norm": 0.598147988319397,
"learning_rate": 4.381028097190299e-05,
"loss": 0.772,
"step": 4305
},
{
"epoch": 0.69,
"grad_norm": 0.5572992563247681,
"learning_rate": 4.3796511786873574e-05,
"loss": 0.7232,
"step": 4310
},
{
"epoch": 0.69,
"grad_norm": 0.7913936376571655,
"learning_rate": 4.378272947282354e-05,
"loss": 0.6972,
"step": 4315
},
{
"epoch": 0.69,
"grad_norm": 0.4532865583896637,
"learning_rate": 4.376893403937959e-05,
"loss": 0.7454,
"step": 4320
},
{
"epoch": 0.69,
"grad_norm": 0.8871356844902039,
"learning_rate": 4.375512549617759e-05,
"loss": 0.6946,
"step": 4325
},
{
"epoch": 0.69,
"grad_norm": 0.7564520835876465,
"learning_rate": 4.374130385286255e-05,
"loss": 0.9257,
"step": 4330
},
{
"epoch": 0.69,
"grad_norm": 0.7280387282371521,
"learning_rate": 4.3727469119088624e-05,
"loss": 0.756,
"step": 4335
},
{
"epoch": 0.69,
"grad_norm": 0.6494055986404419,
"learning_rate": 4.3713621304519144e-05,
"loss": 0.6358,
"step": 4340
},
{
"epoch": 0.69,
"grad_norm": 0.6048948764801025,
"learning_rate": 4.369976041882654e-05,
"loss": 0.6705,
"step": 4345
},
{
"epoch": 0.69,
"grad_norm": 0.6458585858345032,
"learning_rate": 4.36858864716924e-05,
"loss": 0.7999,
"step": 4350
},
{
"epoch": 0.7,
"grad_norm": 0.837872326374054,
"learning_rate": 4.36719994728074e-05,
"loss": 0.7671,
"step": 4355
},
{
"epoch": 0.7,
"grad_norm": 0.6451572775840759,
"learning_rate": 4.365809943187138e-05,
"loss": 0.8672,
"step": 4360
},
{
"epoch": 0.7,
"grad_norm": 0.6438645124435425,
"learning_rate": 4.364418635859326e-05,
"loss": 0.78,
"step": 4365
},
{
"epoch": 0.7,
"grad_norm": 0.7427099347114563,
"learning_rate": 4.363026026269106e-05,
"loss": 0.8977,
"step": 4370
},
{
"epoch": 0.7,
"grad_norm": 0.7844499945640564,
"learning_rate": 4.36163211538919e-05,
"loss": 0.7586,
"step": 4375
},
{
"epoch": 0.7,
"grad_norm": 0.8544999361038208,
"learning_rate": 4.360236904193201e-05,
"loss": 0.7085,
"step": 4380
},
{
"epoch": 0.7,
"grad_norm": 2.431629180908203,
"learning_rate": 4.358840393655668e-05,
"loss": 0.8572,
"step": 4385
},
{
"epoch": 0.7,
"grad_norm": 0.6864097118377686,
"learning_rate": 4.357442584752027e-05,
"loss": 0.6848,
"step": 4390
},
{
"epoch": 0.7,
"grad_norm": 0.7158388495445251,
"learning_rate": 4.356043478458623e-05,
"loss": 1.0071,
"step": 4395
},
{
"epoch": 0.7,
"grad_norm": 0.7883514165878296,
"learning_rate": 4.3546430757527066e-05,
"loss": 0.9851,
"step": 4400
},
{
"epoch": 0.7,
"eval_loss": 0.732368528842926,
"eval_runtime": 96.9109,
"eval_samples_per_second": 7.192,
"eval_steps_per_second": 7.192,
"step": 4400
},
{
"epoch": 0.7,
"grad_norm": 0.673925518989563,
"learning_rate": 4.353241377612433e-05,
"loss": 0.7076,
"step": 4405
},
{
"epoch": 0.7,
"grad_norm": 0.9540270566940308,
"learning_rate": 4.351838385016862e-05,
"loss": 0.8989,
"step": 4410
},
{
"epoch": 0.7,
"grad_norm": 0.8137551546096802,
"learning_rate": 4.35043409894596e-05,
"loss": 0.7633,
"step": 4415
},
{
"epoch": 0.71,
"grad_norm": 0.779330313205719,
"learning_rate": 4.349028520380594e-05,
"loss": 0.7013,
"step": 4420
},
{
"epoch": 0.71,
"grad_norm": 0.7883580327033997,
"learning_rate": 4.347621650302535e-05,
"loss": 0.9788,
"step": 4425
},
{
"epoch": 0.71,
"grad_norm": 0.7106336951255798,
"learning_rate": 4.3462134896944565e-05,
"loss": 0.8399,
"step": 4430
},
{
"epoch": 0.71,
"grad_norm": 0.6303668022155762,
"learning_rate": 4.344804039539933e-05,
"loss": 0.5943,
"step": 4435
},
{
"epoch": 0.71,
"grad_norm": 1.2975471019744873,
"learning_rate": 4.3433933008234395e-05,
"loss": 0.8917,
"step": 4440
},
{
"epoch": 0.71,
"grad_norm": 0.68232661485672,
"learning_rate": 4.341981274530351e-05,
"loss": 0.7756,
"step": 4445
},
{
"epoch": 0.71,
"grad_norm": 0.6689594984054565,
"learning_rate": 4.340567961646943e-05,
"loss": 0.772,
"step": 4450
},
{
"epoch": 0.71,
"grad_norm": 1.102365493774414,
"learning_rate": 4.339153363160388e-05,
"loss": 0.738,
"step": 4455
},
{
"epoch": 0.71,
"grad_norm": 0.6535090804100037,
"learning_rate": 4.337737480058758e-05,
"loss": 0.9096,
"step": 4460
},
{
"epoch": 0.71,
"grad_norm": 0.676058292388916,
"learning_rate": 4.3363203133310206e-05,
"loss": 0.9634,
"step": 4465
},
{
"epoch": 0.71,
"grad_norm": 0.9258711934089661,
"learning_rate": 4.3349018639670415e-05,
"loss": 0.8025,
"step": 4470
},
{
"epoch": 0.71,
"grad_norm": 0.5786353349685669,
"learning_rate": 4.333482132957581e-05,
"loss": 0.7638,
"step": 4475
},
{
"epoch": 0.72,
"grad_norm": 0.7258582711219788,
"learning_rate": 4.332061121294296e-05,
"loss": 1.3538,
"step": 4480
},
{
"epoch": 0.72,
"grad_norm": 0.9384926557540894,
"learning_rate": 4.330638829969738e-05,
"loss": 0.8485,
"step": 4485
},
{
"epoch": 0.72,
"grad_norm": 0.5252525806427002,
"learning_rate": 4.3292152599773494e-05,
"loss": 0.8547,
"step": 4490
},
{
"epoch": 0.72,
"grad_norm": 0.7551200985908508,
"learning_rate": 4.32779041231147e-05,
"loss": 0.7435,
"step": 4495
},
{
"epoch": 0.72,
"grad_norm": 0.7492663264274597,
"learning_rate": 4.3263642879673286e-05,
"loss": 0.8712,
"step": 4500
},
{
"epoch": 0.72,
"eval_loss": 0.7316818237304688,
"eval_runtime": 96.9418,
"eval_samples_per_second": 7.19,
"eval_steps_per_second": 7.19,
"step": 4500
},
{
"epoch": 0.72,
"grad_norm": 0.7490917444229126,
"learning_rate": 4.3249368879410475e-05,
"loss": 0.7598,
"step": 4505
},
{
"epoch": 0.72,
"grad_norm": 0.7305790781974792,
"learning_rate": 4.323508213229639e-05,
"loss": 0.8315,
"step": 4510
},
{
"epoch": 0.72,
"grad_norm": 0.7009093165397644,
"learning_rate": 4.3220782648310075e-05,
"loss": 0.7482,
"step": 4515
},
{
"epoch": 0.72,
"grad_norm": 0.7155885100364685,
"learning_rate": 4.320647043743945e-05,
"loss": 0.8385,
"step": 4520
},
{
"epoch": 0.72,
"grad_norm": 0.6159176826477051,
"learning_rate": 4.319214550968133e-05,
"loss": 0.6507,
"step": 4525
},
{
"epoch": 0.72,
"grad_norm": 0.7776069045066833,
"learning_rate": 4.3177807875041424e-05,
"loss": 0.855,
"step": 4530
},
{
"epoch": 0.72,
"grad_norm": 0.6204195618629456,
"learning_rate": 4.316345754353432e-05,
"loss": 0.7169,
"step": 4535
},
{
"epoch": 0.72,
"grad_norm": 0.7233458757400513,
"learning_rate": 4.3149094525183426e-05,
"loss": 0.5399,
"step": 4540
},
{
"epoch": 0.73,
"grad_norm": 0.7921779155731201,
"learning_rate": 4.313471883002108e-05,
"loss": 0.9124,
"step": 4545
},
{
"epoch": 0.73,
"grad_norm": 0.9145547747612,
"learning_rate": 4.3120330468088435e-05,
"loss": 1.2346,
"step": 4550
},
{
"epoch": 0.73,
"grad_norm": 0.8733106255531311,
"learning_rate": 4.310592944943549e-05,
"loss": 0.6737,
"step": 4555
},
{
"epoch": 0.73,
"grad_norm": 0.6620619297027588,
"learning_rate": 4.3091515784121107e-05,
"loss": 0.8041,
"step": 4560
},
{
"epoch": 0.73,
"grad_norm": 0.7026892900466919,
"learning_rate": 4.307708948221296e-05,
"loss": 0.9422,
"step": 4565
},
{
"epoch": 0.73,
"grad_norm": 0.7953292727470398,
"learning_rate": 4.3062650553787566e-05,
"loss": 0.7398,
"step": 4570
},
{
"epoch": 0.73,
"grad_norm": 1.6465870141983032,
"learning_rate": 4.304819900893024e-05,
"loss": 0.8175,
"step": 4575
},
{
"epoch": 0.73,
"grad_norm": 1.3427163362503052,
"learning_rate": 4.303373485773513e-05,
"loss": 0.7331,
"step": 4580
},
{
"epoch": 0.73,
"grad_norm": 0.6665405035018921,
"learning_rate": 4.3019258110305186e-05,
"loss": 0.7529,
"step": 4585
},
{
"epoch": 0.73,
"grad_norm": 0.796320915222168,
"learning_rate": 4.300476877675215e-05,
"loss": 0.915,
"step": 4590
},
{
"epoch": 0.73,
"grad_norm": 0.551832377910614,
"learning_rate": 4.299026686719655e-05,
"loss": 0.7693,
"step": 4595
},
{
"epoch": 0.73,
"grad_norm": 0.75690096616745,
"learning_rate": 4.297575239176771e-05,
"loss": 0.7871,
"step": 4600
},
{
"epoch": 0.73,
"eval_loss": 0.730965256690979,
"eval_runtime": 96.8803,
"eval_samples_per_second": 7.194,
"eval_steps_per_second": 7.194,
"step": 4600
},
{
"epoch": 0.74,
"grad_norm": 0.7093445062637329,
"learning_rate": 4.296122536060373e-05,
"loss": 0.6279,
"step": 4605
},
{
"epoch": 0.74,
"grad_norm": 0.6522731781005859,
"learning_rate": 4.294668578385147e-05,
"loss": 0.5442,
"step": 4610
},
{
"epoch": 0.74,
"grad_norm": 0.7964634299278259,
"learning_rate": 4.2932133671666565e-05,
"loss": 1.0221,
"step": 4615
},
{
"epoch": 0.74,
"grad_norm": 0.7733820676803589,
"learning_rate": 4.2917569034213395e-05,
"loss": 0.7152,
"step": 4620
},
{
"epoch": 0.74,
"grad_norm": 0.8039364218711853,
"learning_rate": 4.2902991881665097e-05,
"loss": 1.0939,
"step": 4625
},
{
"epoch": 0.74,
"grad_norm": 0.756020724773407,
"learning_rate": 4.2888402224203536e-05,
"loss": 0.7539,
"step": 4630
},
{
"epoch": 0.74,
"grad_norm": 0.5059025287628174,
"learning_rate": 4.2873800072019345e-05,
"loss": 0.8716,
"step": 4635
},
{
"epoch": 0.74,
"grad_norm": 0.8273636102676392,
"learning_rate": 4.285918543531183e-05,
"loss": 0.687,
"step": 4640
},
{
"epoch": 0.74,
"grad_norm": 0.6505921483039856,
"learning_rate": 4.2844558324289076e-05,
"loss": 1.0697,
"step": 4645
},
{
"epoch": 0.74,
"grad_norm": 0.6481053829193115,
"learning_rate": 4.282991874916784e-05,
"loss": 0.884,
"step": 4650
},
{
"epoch": 0.74,
"grad_norm": 0.8193663954734802,
"learning_rate": 4.28152667201736e-05,
"loss": 0.962,
"step": 4655
},
{
"epoch": 0.74,
"grad_norm": 0.7153398990631104,
"learning_rate": 4.280060224754053e-05,
"loss": 0.7705,
"step": 4660
},
{
"epoch": 0.74,
"grad_norm": 0.7197556495666504,
"learning_rate": 4.278592534151149e-05,
"loss": 0.8521,
"step": 4665
},
{
"epoch": 0.75,
"grad_norm": 1.009098768234253,
"learning_rate": 4.2771236012338044e-05,
"loss": 0.8425,
"step": 4670
},
{
"epoch": 0.75,
"grad_norm": 0.598564088344574,
"learning_rate": 4.275653427028041e-05,
"loss": 0.7072,
"step": 4675
},
{
"epoch": 0.75,
"grad_norm": 0.8035867810249329,
"learning_rate": 4.2741820125607504e-05,
"loss": 0.6689,
"step": 4680
},
{
"epoch": 0.75,
"grad_norm": 0.6109891533851624,
"learning_rate": 4.2727093588596866e-05,
"loss": 0.7776,
"step": 4685
},
{
"epoch": 0.75,
"grad_norm": 0.7504151463508606,
"learning_rate": 4.271235466953473e-05,
"loss": 0.7481,
"step": 4690
},
{
"epoch": 0.75,
"grad_norm": 0.9492344260215759,
"learning_rate": 4.269760337871594e-05,
"loss": 0.7261,
"step": 4695
},
{
"epoch": 0.75,
"grad_norm": 0.5792133212089539,
"learning_rate": 4.2682839726444035e-05,
"loss": 0.7156,
"step": 4700
},
{
"epoch": 0.75,
"eval_loss": 0.7283556461334229,
"eval_runtime": 96.8998,
"eval_samples_per_second": 7.193,
"eval_steps_per_second": 7.193,
"step": 4700
},
{
"epoch": 0.75,
"grad_norm": 0.8895491361618042,
"learning_rate": 4.266806372303113e-05,
"loss": 0.8466,
"step": 4705
},
{
"epoch": 0.75,
"grad_norm": 0.7995960712432861,
"learning_rate": 4.2653275378798005e-05,
"loss": 0.7823,
"step": 4710
},
{
"epoch": 0.75,
"grad_norm": 0.6673771739006042,
"learning_rate": 4.263847470407405e-05,
"loss": 0.5461,
"step": 4715
},
{
"epoch": 0.75,
"grad_norm": 0.6228974461555481,
"learning_rate": 4.262366170919726e-05,
"loss": 0.7611,
"step": 4720
},
{
"epoch": 0.75,
"grad_norm": 0.8050612807273865,
"learning_rate": 4.2608836404514255e-05,
"loss": 0.6524,
"step": 4725
},
{
"epoch": 0.75,
"grad_norm": 0.8815121650695801,
"learning_rate": 4.2593998800380216e-05,
"loss": 0.9997,
"step": 4730
},
{
"epoch": 0.76,
"grad_norm": 1.0408731698989868,
"learning_rate": 4.257914890715897e-05,
"loss": 0.7031,
"step": 4735
},
{
"epoch": 0.76,
"grad_norm": 0.6744192838668823,
"learning_rate": 4.256428673522287e-05,
"loss": 0.6587,
"step": 4740
},
{
"epoch": 0.76,
"grad_norm": 1.014369249343872,
"learning_rate": 4.254941229495289e-05,
"loss": 0.7726,
"step": 4745
},
{
"epoch": 0.76,
"grad_norm": 0.7497864365577698,
"learning_rate": 4.2534525596738526e-05,
"loss": 0.7327,
"step": 4750
},
{
"epoch": 0.76,
"grad_norm": 0.6479122042655945,
"learning_rate": 4.2519626650977905e-05,
"loss": 0.7071,
"step": 4755
},
{
"epoch": 0.76,
"grad_norm": 0.6300268173217773,
"learning_rate": 4.250471546807765e-05,
"loss": 0.9479,
"step": 4760
},
{
"epoch": 0.76,
"grad_norm": 0.8272077441215515,
"learning_rate": 4.248979205845294e-05,
"loss": 0.9013,
"step": 4765
},
{
"epoch": 0.76,
"grad_norm": 0.7070410847663879,
"learning_rate": 4.2474856432527524e-05,
"loss": 0.713,
"step": 4770
},
{
"epoch": 0.76,
"grad_norm": 0.7199767231941223,
"learning_rate": 4.2459908600733654e-05,
"loss": 0.9308,
"step": 4775
},
{
"epoch": 0.76,
"grad_norm": 0.6886048316955566,
"learning_rate": 4.244494857351212e-05,
"loss": 0.8008,
"step": 4780
},
{
"epoch": 0.76,
"grad_norm": 0.6097077131271362,
"learning_rate": 4.242997636131222e-05,
"loss": 0.9639,
"step": 4785
},
{
"epoch": 0.76,
"grad_norm": 1.0947343111038208,
"learning_rate": 4.241499197459178e-05,
"loss": 0.9012,
"step": 4790
},
{
"epoch": 0.77,
"grad_norm": 0.6965738534927368,
"learning_rate": 4.239999542381712e-05,
"loss": 0.6745,
"step": 4795
},
{
"epoch": 0.77,
"grad_norm": 0.8290371894836426,
"learning_rate": 4.238498671946306e-05,
"loss": 0.7856,
"step": 4800
},
{
"epoch": 0.77,
"eval_loss": 0.7277354598045349,
"eval_runtime": 96.9165,
"eval_samples_per_second": 7.192,
"eval_steps_per_second": 7.192,
"step": 4800
},
{
"epoch": 0.77,
"grad_norm": 0.8061904907226562,
"learning_rate": 4.2369965872012904e-05,
"loss": 0.7034,
"step": 4805
},
{
"epoch": 0.77,
"grad_norm": 0.6652625799179077,
"learning_rate": 4.2354932891958434e-05,
"loss": 0.5825,
"step": 4810
},
{
"epoch": 0.77,
"grad_norm": 0.6529026627540588,
"learning_rate": 4.2339887789799916e-05,
"loss": 0.7407,
"step": 4815
},
{
"epoch": 0.77,
"grad_norm": 3.7802493572235107,
"learning_rate": 4.232483057604607e-05,
"loss": 0.8906,
"step": 4820
},
{
"epoch": 0.77,
"grad_norm": 0.7709060907363892,
"learning_rate": 4.230976126121411e-05,
"loss": 0.863,
"step": 4825
},
{
"epoch": 0.77,
"grad_norm": 1.2582249641418457,
"learning_rate": 4.229467985582966e-05,
"loss": 1.065,
"step": 4830
},
{
"epoch": 0.77,
"grad_norm": 0.5523508191108704,
"learning_rate": 4.22795863704268e-05,
"loss": 0.5925,
"step": 4835
},
{
"epoch": 0.77,
"grad_norm": 1.3535953760147095,
"learning_rate": 4.2264480815548076e-05,
"loss": 0.7993,
"step": 4840
},
{
"epoch": 0.77,
"grad_norm": 1.067133903503418,
"learning_rate": 4.2249363201744425e-05,
"loss": 0.7921,
"step": 4845
},
{
"epoch": 0.77,
"grad_norm": 0.6478603482246399,
"learning_rate": 4.223423353957523e-05,
"loss": 0.6769,
"step": 4850
},
{
"epoch": 0.77,
"grad_norm": 0.6439855694770813,
"learning_rate": 4.2219091839608276e-05,
"loss": 0.9018,
"step": 4855
},
{
"epoch": 0.78,
"grad_norm": 0.5302556753158569,
"learning_rate": 4.2203938112419786e-05,
"loss": 0.837,
"step": 4860
},
{
"epoch": 0.78,
"grad_norm": 0.8129810690879822,
"learning_rate": 4.218877236859433e-05,
"loss": 0.9195,
"step": 4865
},
{
"epoch": 0.78,
"grad_norm": 0.6531801819801331,
"learning_rate": 4.217359461872493e-05,
"loss": 0.6829,
"step": 4870
},
{
"epoch": 0.78,
"grad_norm": 0.7695423364639282,
"learning_rate": 4.215840487341296e-05,
"loss": 0.7739,
"step": 4875
},
{
"epoch": 0.78,
"grad_norm": 1.3163946866989136,
"learning_rate": 4.2143203143268184e-05,
"loss": 0.9678,
"step": 4880
},
{
"epoch": 0.78,
"grad_norm": 1.1124577522277832,
"learning_rate": 4.212798943890871e-05,
"loss": 0.9327,
"step": 4885
},
{
"epoch": 0.78,
"grad_norm": 0.8979106545448303,
"learning_rate": 4.2112763770961074e-05,
"loss": 0.7043,
"step": 4890
},
{
"epoch": 0.78,
"grad_norm": 0.9879763126373291,
"learning_rate": 4.2097526150060085e-05,
"loss": 0.8129,
"step": 4895
},
{
"epoch": 0.78,
"grad_norm": 0.7016007304191589,
"learning_rate": 4.208227658684898e-05,
"loss": 0.7906,
"step": 4900
},
{
"epoch": 0.78,
"eval_loss": 0.7254941463470459,
"eval_runtime": 96.9328,
"eval_samples_per_second": 7.191,
"eval_steps_per_second": 7.191,
"step": 4900
},
{
"epoch": 0.78,
"grad_norm": 0.5404706597328186,
"learning_rate": 4.206701509197927e-05,
"loss": 0.7769,
"step": 4905
},
{
"epoch": 0.78,
"grad_norm": 0.7096789479255676,
"learning_rate": 4.205174167611085e-05,
"loss": 0.5985,
"step": 4910
},
{
"epoch": 0.78,
"grad_norm": 0.8139373660087585,
"learning_rate": 4.20364563499119e-05,
"loss": 0.75,
"step": 4915
},
{
"epoch": 0.79,
"grad_norm": 1.2196255922317505,
"learning_rate": 4.202115912405897e-05,
"loss": 0.8441,
"step": 4920
},
{
"epoch": 0.79,
"grad_norm": 0.6912347674369812,
"learning_rate": 4.200585000923689e-05,
"loss": 0.8885,
"step": 4925
},
{
"epoch": 0.79,
"grad_norm": 0.7879334688186646,
"learning_rate": 4.199052901613878e-05,
"loss": 0.6353,
"step": 4930
},
{
"epoch": 0.79,
"grad_norm": 1.0302627086639404,
"learning_rate": 4.197519615546608e-05,
"loss": 0.7704,
"step": 4935
},
{
"epoch": 0.79,
"grad_norm": 0.5351320505142212,
"learning_rate": 4.195985143792851e-05,
"loss": 0.8094,
"step": 4940
},
{
"epoch": 0.79,
"grad_norm": 0.7632457613945007,
"learning_rate": 4.194449487424409e-05,
"loss": 0.9625,
"step": 4945
},
{
"epoch": 0.79,
"grad_norm": 0.7706131935119629,
"learning_rate": 4.1929126475139096e-05,
"loss": 0.7007,
"step": 4950
},
{
"epoch": 0.79,
"grad_norm": 0.6262048482894897,
"learning_rate": 4.191374625134806e-05,
"loss": 0.7768,
"step": 4955
},
{
"epoch": 0.79,
"grad_norm": 0.8302519917488098,
"learning_rate": 4.189835421361381e-05,
"loss": 0.8281,
"step": 4960
},
{
"epoch": 0.79,
"grad_norm": 0.5914260149002075,
"learning_rate": 4.188295037268738e-05,
"loss": 0.8554,
"step": 4965
},
{
"epoch": 0.79,
"grad_norm": 0.7599936127662659,
"learning_rate": 4.1867534739328085e-05,
"loss": 0.9547,
"step": 4970
},
{
"epoch": 0.79,
"grad_norm": 0.4832470417022705,
"learning_rate": 4.1852107324303455e-05,
"loss": 0.5212,
"step": 4975
},
{
"epoch": 0.79,
"grad_norm": 0.8040557503700256,
"learning_rate": 4.183666813838927e-05,
"loss": 0.8939,
"step": 4980
},
{
"epoch": 0.8,
"grad_norm": 0.7503822445869446,
"learning_rate": 4.182121719236952e-05,
"loss": 0.9279,
"step": 4985
},
{
"epoch": 0.8,
"grad_norm": 0.7608035206794739,
"learning_rate": 4.180575449703639e-05,
"loss": 0.7965,
"step": 4990
},
{
"epoch": 0.8,
"grad_norm": 0.5479308366775513,
"learning_rate": 4.1790280063190315e-05,
"loss": 0.7478,
"step": 4995
},
{
"epoch": 0.8,
"grad_norm": 0.7714606523513794,
"learning_rate": 4.177479390163989e-05,
"loss": 0.7917,
"step": 5000
},
{
"epoch": 0.8,
"eval_loss": 0.7249829173088074,
"eval_runtime": 96.9791,
"eval_samples_per_second": 7.187,
"eval_steps_per_second": 7.187,
"step": 5000
},
{
"epoch": 0.8,
"grad_norm": 0.7731960415840149,
"learning_rate": 4.175929602320192e-05,
"loss": 0.5968,
"step": 5005
},
{
"epoch": 0.8,
"grad_norm": 0.525674045085907,
"learning_rate": 4.174378643870138e-05,
"loss": 0.6748,
"step": 5010
},
{
"epoch": 0.8,
"grad_norm": 0.8199644088745117,
"learning_rate": 4.172826515897146e-05,
"loss": 0.7478,
"step": 5015
},
{
"epoch": 0.8,
"grad_norm": 0.5651956796646118,
"learning_rate": 4.1712732194853464e-05,
"loss": 0.5784,
"step": 5020
},
{
"epoch": 0.8,
"grad_norm": 0.8027454614639282,
"learning_rate": 4.1697187557196896e-05,
"loss": 0.9719,
"step": 5025
},
{
"epoch": 0.8,
"grad_norm": 0.8870482444763184,
"learning_rate": 4.168163125685939e-05,
"loss": 0.7407,
"step": 5030
},
{
"epoch": 0.8,
"grad_norm": 0.6260176301002502,
"learning_rate": 4.1666063304706756e-05,
"loss": 0.8039,
"step": 5035
},
{
"epoch": 0.8,
"grad_norm": 0.5424874424934387,
"learning_rate": 4.165048371161291e-05,
"loss": 0.8169,
"step": 5040
},
{
"epoch": 0.81,
"grad_norm": 0.8967744708061218,
"learning_rate": 4.163489248845992e-05,
"loss": 0.7546,
"step": 5045
},
{
"epoch": 0.81,
"grad_norm": 0.871800422668457,
"learning_rate": 4.161928964613797e-05,
"loss": 0.7789,
"step": 5050
},
{
"epoch": 0.81,
"grad_norm": 0.9717918634414673,
"learning_rate": 4.1603675195545356e-05,
"loss": 0.7955,
"step": 5055
},
{
"epoch": 0.81,
"grad_norm": 0.6605653762817383,
"learning_rate": 4.15880491475885e-05,
"loss": 0.8416,
"step": 5060
},
{
"epoch": 0.81,
"grad_norm": 0.9123837351799011,
"learning_rate": 4.1572411513181896e-05,
"loss": 0.6594,
"step": 5065
},
{
"epoch": 0.81,
"grad_norm": 0.8939911723136902,
"learning_rate": 4.155676230324816e-05,
"loss": 0.858,
"step": 5070
},
{
"epoch": 0.81,
"grad_norm": 0.6447166204452515,
"learning_rate": 4.154110152871797e-05,
"loss": 0.7425,
"step": 5075
},
{
"epoch": 0.81,
"grad_norm": 0.9701817631721497,
"learning_rate": 4.152542920053009e-05,
"loss": 0.9332,
"step": 5080
},
{
"epoch": 0.81,
"grad_norm": 0.7782546281814575,
"learning_rate": 4.1509745329631365e-05,
"loss": 0.8372,
"step": 5085
},
{
"epoch": 0.81,
"grad_norm": 0.6977235078811646,
"learning_rate": 4.149404992697669e-05,
"loss": 0.6689,
"step": 5090
},
{
"epoch": 0.81,
"grad_norm": 0.7119914293289185,
"learning_rate": 4.147834300352901e-05,
"loss": 0.6255,
"step": 5095
},
{
"epoch": 0.81,
"grad_norm": 0.7148675322532654,
"learning_rate": 4.146262457025933e-05,
"loss": 0.6395,
"step": 5100
},
{
"epoch": 0.81,
"eval_loss": 0.7237269282341003,
"eval_runtime": 96.898,
"eval_samples_per_second": 7.193,
"eval_steps_per_second": 7.193,
"step": 5100
},
{
"epoch": 0.81,
"grad_norm": 0.7843590378761292,
"learning_rate": 4.144689463814669e-05,
"loss": 0.6286,
"step": 5105
},
{
"epoch": 0.82,
"grad_norm": 0.8296339511871338,
"learning_rate": 4.143115321817815e-05,
"loss": 0.8533,
"step": 5110
},
{
"epoch": 0.82,
"grad_norm": 0.6789836287498474,
"learning_rate": 4.1415400321348827e-05,
"loss": 0.7821,
"step": 5115
},
{
"epoch": 0.82,
"grad_norm": 0.9112969040870667,
"learning_rate": 4.139963595866182e-05,
"loss": 0.9232,
"step": 5120
},
{
"epoch": 0.82,
"grad_norm": 0.5057934522628784,
"learning_rate": 4.138386014112824e-05,
"loss": 0.5756,
"step": 5125
},
{
"epoch": 0.82,
"grad_norm": 0.7892030477523804,
"learning_rate": 4.136807287976721e-05,
"loss": 0.6673,
"step": 5130
},
{
"epoch": 0.82,
"grad_norm": 0.6285967826843262,
"learning_rate": 4.135227418560585e-05,
"loss": 0.843,
"step": 5135
},
{
"epoch": 0.82,
"grad_norm": 0.5548838973045349,
"learning_rate": 4.133646406967927e-05,
"loss": 0.8089,
"step": 5140
},
{
"epoch": 0.82,
"grad_norm": 0.6649813055992126,
"learning_rate": 4.132064254303053e-05,
"loss": 0.8778,
"step": 5145
},
{
"epoch": 0.82,
"grad_norm": 0.6559309959411621,
"learning_rate": 4.1304809616710685e-05,
"loss": 0.6655,
"step": 5150
},
{
"epoch": 0.82,
"grad_norm": 0.839788019657135,
"learning_rate": 4.1288965301778745e-05,
"loss": 0.8372,
"step": 5155
},
{
"epoch": 0.82,
"grad_norm": 0.5987736582756042,
"learning_rate": 4.1273109609301686e-05,
"loss": 0.6379,
"step": 5160
},
{
"epoch": 0.82,
"grad_norm": 0.7344145774841309,
"learning_rate": 4.125724255035439e-05,
"loss": 0.7199,
"step": 5165
},
{
"epoch": 0.83,
"grad_norm": 1.2459057569503784,
"learning_rate": 4.124136413601973e-05,
"loss": 0.7656,
"step": 5170
},
{
"epoch": 0.83,
"grad_norm": 1.2029509544372559,
"learning_rate": 4.12254743773885e-05,
"loss": 0.6369,
"step": 5175
},
{
"epoch": 0.83,
"grad_norm": 0.8527836799621582,
"learning_rate": 4.1209573285559376e-05,
"loss": 0.7597,
"step": 5180
},
{
"epoch": 0.83,
"grad_norm": 0.6795602440834045,
"learning_rate": 4.1193660871639006e-05,
"loss": 0.8141,
"step": 5185
},
{
"epoch": 0.83,
"grad_norm": 0.8437166810035706,
"learning_rate": 4.1177737146741904e-05,
"loss": 0.7443,
"step": 5190
},
{
"epoch": 0.83,
"grad_norm": 0.7290692329406738,
"learning_rate": 4.1161802121990513e-05,
"loss": 0.5944,
"step": 5195
},
{
"epoch": 0.83,
"grad_norm": 0.624458372592926,
"learning_rate": 4.114585580851515e-05,
"loss": 0.7567,
"step": 5200
},
{
"epoch": 0.83,
"eval_loss": 0.7232300043106079,
"eval_runtime": 96.871,
"eval_samples_per_second": 7.195,
"eval_steps_per_second": 7.195,
"step": 5200
},
{
"epoch": 0.83,
"grad_norm": 0.9863276481628418,
"learning_rate": 4.1129898217454034e-05,
"loss": 0.7102,
"step": 5205
},
{
"epoch": 0.83,
"grad_norm": 0.8225752115249634,
"learning_rate": 4.111392935995324e-05,
"loss": 0.8148,
"step": 5210
},
{
"epoch": 0.83,
"grad_norm": 0.687274694442749,
"learning_rate": 4.109794924716673e-05,
"loss": 0.7285,
"step": 5215
},
{
"epoch": 0.83,
"grad_norm": 1.306779146194458,
"learning_rate": 4.108195789025632e-05,
"loss": 1.0412,
"step": 5220
},
{
"epoch": 0.83,
"grad_norm": 0.872298002243042,
"learning_rate": 4.1065955300391676e-05,
"loss": 0.8104,
"step": 5225
},
{
"epoch": 0.83,
"grad_norm": 0.41228190064430237,
"learning_rate": 4.104994148875032e-05,
"loss": 0.8437,
"step": 5230
},
{
"epoch": 0.84,
"grad_norm": 1.0037802457809448,
"learning_rate": 4.103391646651761e-05,
"loss": 0.579,
"step": 5235
},
{
"epoch": 0.84,
"grad_norm": 0.6584619879722595,
"learning_rate": 4.101788024488673e-05,
"loss": 0.8698,
"step": 5240
},
{
"epoch": 0.84,
"grad_norm": 0.8180784583091736,
"learning_rate": 4.1001832835058685e-05,
"loss": 0.9177,
"step": 5245
},
{
"epoch": 0.84,
"grad_norm": 0.7600345015525818,
"learning_rate": 4.0985774248242296e-05,
"loss": 0.6935,
"step": 5250
},
{
"epoch": 0.84,
"grad_norm": 0.9433395266532898,
"learning_rate": 4.09697044956542e-05,
"loss": 0.642,
"step": 5255
},
{
"epoch": 0.84,
"grad_norm": 0.625467836856842,
"learning_rate": 4.095362358851883e-05,
"loss": 0.8038,
"step": 5260
},
{
"epoch": 0.84,
"grad_norm": 0.5945066213607788,
"learning_rate": 4.0937531538068387e-05,
"loss": 0.7096,
"step": 5265
},
{
"epoch": 0.84,
"grad_norm": 0.9545826315879822,
"learning_rate": 4.092142835554289e-05,
"loss": 0.6535,
"step": 5270
},
{
"epoch": 0.84,
"grad_norm": 0.8320907354354858,
"learning_rate": 4.090531405219012e-05,
"loss": 0.7521,
"step": 5275
},
{
"epoch": 0.84,
"grad_norm": 0.7830407619476318,
"learning_rate": 4.088918863926562e-05,
"loss": 0.5991,
"step": 5280
},
{
"epoch": 0.84,
"grad_norm": 0.6662408709526062,
"learning_rate": 4.0873052128032684e-05,
"loss": 0.9341,
"step": 5285
},
{
"epoch": 0.84,
"grad_norm": 1.2364461421966553,
"learning_rate": 4.0856904529762394e-05,
"loss": 0.6156,
"step": 5290
},
{
"epoch": 0.85,
"grad_norm": 0.5708659291267395,
"learning_rate": 4.084074585573354e-05,
"loss": 0.6609,
"step": 5295
},
{
"epoch": 0.85,
"grad_norm": 0.6155332326889038,
"learning_rate": 4.082457611723266e-05,
"loss": 0.8551,
"step": 5300
},
{
"epoch": 0.85,
"eval_loss": 0.7219573855400085,
"eval_runtime": 96.9322,
"eval_samples_per_second": 7.191,
"eval_steps_per_second": 7.191,
"step": 5300
},
{
"epoch": 0.85,
"grad_norm": 0.5997724533081055,
"learning_rate": 4.0808395325554023e-05,
"loss": 0.7243,
"step": 5305
},
{
"epoch": 0.85,
"grad_norm": 0.5939134359359741,
"learning_rate": 4.079220349199962e-05,
"loss": 0.7455,
"step": 5310
},
{
"epoch": 0.85,
"grad_norm": 0.5447105765342712,
"learning_rate": 4.077600062787915e-05,
"loss": 0.8424,
"step": 5315
},
{
"epoch": 0.85,
"grad_norm": 0.741060197353363,
"learning_rate": 4.075978674451001e-05,
"loss": 0.6837,
"step": 5320
},
{
"epoch": 0.85,
"grad_norm": 0.8776602745056152,
"learning_rate": 4.074356185321732e-05,
"loss": 0.6541,
"step": 5325
},
{
"epoch": 0.85,
"grad_norm": 0.5684154629707336,
"learning_rate": 4.072732596533385e-05,
"loss": 0.6843,
"step": 5330
},
{
"epoch": 0.85,
"grad_norm": 0.9258201718330383,
"learning_rate": 4.071107909220009e-05,
"loss": 0.7829,
"step": 5335
},
{
"epoch": 0.85,
"grad_norm": 0.6690560579299927,
"learning_rate": 4.0694821245164165e-05,
"loss": 0.7726,
"step": 5340
},
{
"epoch": 0.85,
"grad_norm": 0.9005510807037354,
"learning_rate": 4.0678552435581905e-05,
"loss": 0.845,
"step": 5345
},
{
"epoch": 0.85,
"grad_norm": 0.6624660491943359,
"learning_rate": 4.066227267481676e-05,
"loss": 0.5798,
"step": 5350
},
{
"epoch": 0.85,
"grad_norm": 0.5465379953384399,
"learning_rate": 4.064598197423985e-05,
"loss": 0.6488,
"step": 5355
},
{
"epoch": 0.86,
"grad_norm": 0.7643209099769592,
"learning_rate": 4.062968034522995e-05,
"loss": 0.6886,
"step": 5360
},
{
"epoch": 0.86,
"grad_norm": 0.7964931130409241,
"learning_rate": 4.061336779917342e-05,
"loss": 1.0865,
"step": 5365
},
{
"epoch": 0.86,
"grad_norm": 0.6491959095001221,
"learning_rate": 4.05970443474643e-05,
"loss": 0.5869,
"step": 5370
},
{
"epoch": 0.86,
"grad_norm": 1.0309019088745117,
"learning_rate": 4.058071000150421e-05,
"loss": 0.8846,
"step": 5375
},
{
"epoch": 0.86,
"grad_norm": 0.6856257915496826,
"learning_rate": 4.05643647727024e-05,
"loss": 0.774,
"step": 5380
},
{
"epoch": 0.86,
"grad_norm": 0.6670073866844177,
"learning_rate": 4.0548008672475714e-05,
"loss": 0.6854,
"step": 5385
},
{
"epoch": 0.86,
"grad_norm": 0.9411470890045166,
"learning_rate": 4.053164171224858e-05,
"loss": 0.9237,
"step": 5390
},
{
"epoch": 0.86,
"grad_norm": 0.7178419232368469,
"learning_rate": 4.051526390345303e-05,
"loss": 0.6262,
"step": 5395
},
{
"epoch": 0.86,
"grad_norm": 0.69013512134552,
"learning_rate": 4.049887525752867e-05,
"loss": 0.7392,
"step": 5400
},
{
"epoch": 0.86,
"eval_loss": 0.7226108908653259,
"eval_runtime": 96.9511,
"eval_samples_per_second": 7.189,
"eval_steps_per_second": 7.189,
"step": 5400
},
{
"epoch": 0.86,
"grad_norm": 1.5154650211334229,
"learning_rate": 4.048247578592266e-05,
"loss": 1.0026,
"step": 5405
},
{
"epoch": 0.86,
"grad_norm": 0.6985073685646057,
"learning_rate": 4.046606550008973e-05,
"loss": 0.4776,
"step": 5410
},
{
"epoch": 0.86,
"grad_norm": 0.687282383441925,
"learning_rate": 4.044964441149217e-05,
"loss": 0.7665,
"step": 5415
},
{
"epoch": 0.87,
"grad_norm": 0.7575299143791199,
"learning_rate": 4.043321253159981e-05,
"loss": 0.8202,
"step": 5420
},
{
"epoch": 0.87,
"grad_norm": 1.0001776218414307,
"learning_rate": 4.041676987189003e-05,
"loss": 0.7952,
"step": 5425
},
{
"epoch": 0.87,
"grad_norm": 1.240440845489502,
"learning_rate": 4.040031644384771e-05,
"loss": 0.7086,
"step": 5430
},
{
"epoch": 0.87,
"grad_norm": 0.7027778029441833,
"learning_rate": 4.0383852258965274e-05,
"loss": 0.8268,
"step": 5435
},
{
"epoch": 0.87,
"grad_norm": 0.7208477854728699,
"learning_rate": 4.036737732874266e-05,
"loss": 0.7095,
"step": 5440
},
{
"epoch": 0.87,
"grad_norm": 0.6870366334915161,
"learning_rate": 4.035089166468731e-05,
"loss": 0.6324,
"step": 5445
},
{
"epoch": 0.87,
"grad_norm": 0.6353491544723511,
"learning_rate": 4.033439527831415e-05,
"loss": 0.6463,
"step": 5450
},
{
"epoch": 0.87,
"grad_norm": 0.8659310340881348,
"learning_rate": 4.03178881811456e-05,
"loss": 0.5445,
"step": 5455
},
{
"epoch": 0.87,
"grad_norm": 1.102812647819519,
"learning_rate": 4.030137038471158e-05,
"loss": 1.0779,
"step": 5460
},
{
"epoch": 0.87,
"grad_norm": 0.6689164638519287,
"learning_rate": 4.028484190054947e-05,
"loss": 0.7816,
"step": 5465
},
{
"epoch": 0.87,
"grad_norm": 0.7742553949356079,
"learning_rate": 4.02683027402041e-05,
"loss": 0.7534,
"step": 5470
},
{
"epoch": 0.87,
"grad_norm": 0.6275122761726379,
"learning_rate": 4.025175291522779e-05,
"loss": 0.8127,
"step": 5475
},
{
"epoch": 0.87,
"grad_norm": 0.7048559188842773,
"learning_rate": 4.0235192437180276e-05,
"loss": 0.7884,
"step": 5480
},
{
"epoch": 0.88,
"grad_norm": 0.6857985854148865,
"learning_rate": 4.0218621317628755e-05,
"loss": 0.6286,
"step": 5485
},
{
"epoch": 0.88,
"grad_norm": 0.750784158706665,
"learning_rate": 4.0202039568147857e-05,
"loss": 0.76,
"step": 5490
},
{
"epoch": 0.88,
"grad_norm": 0.906353235244751,
"learning_rate": 4.018544720031962e-05,
"loss": 0.7227,
"step": 5495
},
{
"epoch": 0.88,
"grad_norm": 1.001240611076355,
"learning_rate": 4.0168844225733516e-05,
"loss": 0.9312,
"step": 5500
},
{
"epoch": 0.88,
"eval_loss": 0.7204533815383911,
"eval_runtime": 96.9186,
"eval_samples_per_second": 7.192,
"eval_steps_per_second": 7.192,
"step": 5500
},
{
"epoch": 0.88,
"grad_norm": 0.6060788035392761,
"learning_rate": 4.015223065598642e-05,
"loss": 0.7476,
"step": 5505
},
{
"epoch": 0.88,
"grad_norm": 0.9126180410385132,
"learning_rate": 4.013560650268261e-05,
"loss": 0.9137,
"step": 5510
},
{
"epoch": 0.88,
"grad_norm": 0.7585060000419617,
"learning_rate": 4.0118971777433755e-05,
"loss": 0.7926,
"step": 5515
},
{
"epoch": 0.88,
"grad_norm": 0.7748726606369019,
"learning_rate": 4.01023264918589e-05,
"loss": 0.8049,
"step": 5520
},
{
"epoch": 0.88,
"grad_norm": 0.6712713241577148,
"learning_rate": 4.008567065758449e-05,
"loss": 0.6242,
"step": 5525
},
{
"epoch": 0.88,
"grad_norm": 0.9764819741249084,
"learning_rate": 4.006900428624429e-05,
"loss": 0.7445,
"step": 5530
},
{
"epoch": 0.88,
"grad_norm": 0.6748983263969421,
"learning_rate": 4.005232738947948e-05,
"loss": 0.9529,
"step": 5535
},
{
"epoch": 0.88,
"grad_norm": 0.9980877637863159,
"learning_rate": 4.003563997893857e-05,
"loss": 1.0637,
"step": 5540
},
{
"epoch": 0.89,
"grad_norm": 0.8130014538764954,
"learning_rate": 4.0018942066277406e-05,
"loss": 0.7479,
"step": 5545
},
{
"epoch": 0.89,
"grad_norm": 0.5256005525588989,
"learning_rate": 4.000223366315917e-05,
"loss": 0.5274,
"step": 5550
},
{
"epoch": 0.89,
"grad_norm": 0.6466362476348877,
"learning_rate": 3.9985514781254385e-05,
"loss": 0.7175,
"step": 5555
},
{
"epoch": 0.89,
"grad_norm": 0.8849642276763916,
"learning_rate": 3.9968785432240874e-05,
"loss": 0.7762,
"step": 5560
},
{
"epoch": 0.89,
"grad_norm": 0.7992335557937622,
"learning_rate": 3.9952045627803795e-05,
"loss": 0.6103,
"step": 5565
},
{
"epoch": 0.89,
"grad_norm": 0.6376257538795471,
"learning_rate": 3.993529537963559e-05,
"loss": 0.7441,
"step": 5570
},
{
"epoch": 0.89,
"grad_norm": 0.6867454648017883,
"learning_rate": 3.9918534699436e-05,
"loss": 0.8815,
"step": 5575
},
{
"epoch": 0.89,
"grad_norm": 0.8116294741630554,
"learning_rate": 3.990176359891206e-05,
"loss": 0.9572,
"step": 5580
},
{
"epoch": 0.89,
"grad_norm": 0.8836712837219238,
"learning_rate": 3.988498208977808e-05,
"loss": 0.8463,
"step": 5585
},
{
"epoch": 0.89,
"grad_norm": 0.835573673248291,
"learning_rate": 3.9868190183755624e-05,
"loss": 0.7096,
"step": 5590
},
{
"epoch": 0.89,
"grad_norm": 0.9869277477264404,
"learning_rate": 3.9851387892573545e-05,
"loss": 0.9144,
"step": 5595
},
{
"epoch": 0.89,
"grad_norm": 0.6233755350112915,
"learning_rate": 3.983457522796793e-05,
"loss": 0.8323,
"step": 5600
},
{
"epoch": 0.89,
"eval_loss": 0.719563901424408,
"eval_runtime": 96.851,
"eval_samples_per_second": 7.197,
"eval_steps_per_second": 7.197,
"step": 5600
},
{
"epoch": 0.89,
"grad_norm": 0.6275632977485657,
"learning_rate": 3.981775220168212e-05,
"loss": 0.751,
"step": 5605
},
{
"epoch": 0.9,
"grad_norm": 0.7892312407493591,
"learning_rate": 3.980091882546669e-05,
"loss": 0.932,
"step": 5610
},
{
"epoch": 0.9,
"grad_norm": 0.6969819068908691,
"learning_rate": 3.9784075111079456e-05,
"loss": 0.7515,
"step": 5615
},
{
"epoch": 0.9,
"grad_norm": 0.7002953290939331,
"learning_rate": 3.976722107028544e-05,
"loss": 0.8556,
"step": 5620
},
{
"epoch": 0.9,
"grad_norm": 1.1970090866088867,
"learning_rate": 3.9750356714856876e-05,
"loss": 0.7594,
"step": 5625
},
{
"epoch": 0.9,
"grad_norm": 0.7774887084960938,
"learning_rate": 3.9733482056573214e-05,
"loss": 0.788,
"step": 5630
},
{
"epoch": 0.9,
"grad_norm": 0.5987717509269714,
"learning_rate": 3.971659710722109e-05,
"loss": 0.7458,
"step": 5635
},
{
"epoch": 0.9,
"grad_norm": 0.7544295191764832,
"learning_rate": 3.969970187859435e-05,
"loss": 0.9236,
"step": 5640
},
{
"epoch": 0.9,
"grad_norm": 5.217820644378662,
"learning_rate": 3.968279638249398e-05,
"loss": 0.7796,
"step": 5645
},
{
"epoch": 0.9,
"grad_norm": 3.211364507675171,
"learning_rate": 3.966588063072817e-05,
"loss": 0.7593,
"step": 5650
},
{
"epoch": 0.9,
"grad_norm": 1.2978931665420532,
"learning_rate": 3.964895463511227e-05,
"loss": 0.8755,
"step": 5655
},
{
"epoch": 0.9,
"grad_norm": 0.6411646008491516,
"learning_rate": 3.963201840746877e-05,
"loss": 0.8999,
"step": 5660
},
{
"epoch": 0.9,
"grad_norm": 0.8105179667472839,
"learning_rate": 3.961507195962732e-05,
"loss": 0.758,
"step": 5665
},
{
"epoch": 0.91,
"grad_norm": 0.8283591270446777,
"learning_rate": 3.959811530342471e-05,
"loss": 0.8578,
"step": 5670
},
{
"epoch": 0.91,
"grad_norm": 0.8040729761123657,
"learning_rate": 3.958114845070485e-05,
"loss": 0.7062,
"step": 5675
},
{
"epoch": 0.91,
"grad_norm": 0.696398138999939,
"learning_rate": 3.956417141331878e-05,
"loss": 0.8301,
"step": 5680
},
{
"epoch": 0.91,
"grad_norm": 0.7806740403175354,
"learning_rate": 3.9547184203124636e-05,
"loss": 0.7002,
"step": 5685
},
{
"epoch": 0.91,
"grad_norm": 0.6160992980003357,
"learning_rate": 3.9530186831987695e-05,
"loss": 0.7433,
"step": 5690
},
{
"epoch": 0.91,
"grad_norm": 0.9394809603691101,
"learning_rate": 3.95131793117803e-05,
"loss": 0.8878,
"step": 5695
},
{
"epoch": 0.91,
"grad_norm": 0.7936028838157654,
"learning_rate": 3.94961616543819e-05,
"loss": 0.7312,
"step": 5700
},
{
"epoch": 0.91,
"eval_loss": 0.7196908593177795,
"eval_runtime": 96.9199,
"eval_samples_per_second": 7.192,
"eval_steps_per_second": 7.192,
"step": 5700
},
{
"epoch": 0.91,
"grad_norm": 0.7855328321456909,
"learning_rate": 3.947913387167901e-05,
"loss": 0.8013,
"step": 5705
},
{
"epoch": 0.91,
"grad_norm": 0.7000160217285156,
"learning_rate": 3.9462095975565226e-05,
"loss": 0.7631,
"step": 5710
},
{
"epoch": 0.91,
"grad_norm": 0.5747112035751343,
"learning_rate": 3.944504797794122e-05,
"loss": 0.6641,
"step": 5715
},
{
"epoch": 0.91,
"grad_norm": 0.7016122341156006,
"learning_rate": 3.942798989071469e-05,
"loss": 0.8245,
"step": 5720
},
{
"epoch": 0.91,
"grad_norm": 0.6330326795578003,
"learning_rate": 3.941092172580042e-05,
"loss": 0.6604,
"step": 5725
},
{
"epoch": 0.91,
"grad_norm": 0.6938309669494629,
"learning_rate": 3.939384349512019e-05,
"loss": 0.891,
"step": 5730
},
{
"epoch": 0.92,
"grad_norm": 0.6257404685020447,
"learning_rate": 3.937675521060284e-05,
"loss": 0.6852,
"step": 5735
},
{
"epoch": 0.92,
"grad_norm": 1.243001937866211,
"learning_rate": 3.9359656884184234e-05,
"loss": 0.7764,
"step": 5740
},
{
"epoch": 0.92,
"grad_norm": 0.5932753086090088,
"learning_rate": 3.9342548527807225e-05,
"loss": 0.6175,
"step": 5745
},
{
"epoch": 0.92,
"grad_norm": 1.0795527696609497,
"learning_rate": 3.9325430153421706e-05,
"loss": 0.8249,
"step": 5750
},
{
"epoch": 0.92,
"grad_norm": 0.9690996408462524,
"learning_rate": 3.9308301772984535e-05,
"loss": 0.7415,
"step": 5755
},
{
"epoch": 0.92,
"grad_norm": 2.7389626502990723,
"learning_rate": 3.9291163398459585e-05,
"loss": 0.8866,
"step": 5760
},
{
"epoch": 0.92,
"grad_norm": 0.9238413572311401,
"learning_rate": 3.927401504181768e-05,
"loss": 0.9837,
"step": 5765
},
{
"epoch": 0.92,
"grad_norm": 1.18929922580719,
"learning_rate": 3.925685671503665e-05,
"loss": 0.7166,
"step": 5770
},
{
"epoch": 0.92,
"grad_norm": 0.717707633972168,
"learning_rate": 3.923968843010127e-05,
"loss": 0.8331,
"step": 5775
},
{
"epoch": 0.92,
"grad_norm": 0.5547253489494324,
"learning_rate": 3.9222510199003276e-05,
"loss": 0.5933,
"step": 5780
},
{
"epoch": 0.92,
"grad_norm": 0.73494952917099,
"learning_rate": 3.920532203374134e-05,
"loss": 0.7895,
"step": 5785
},
{
"epoch": 0.92,
"grad_norm": 0.8860884308815002,
"learning_rate": 3.91881239463211e-05,
"loss": 0.757,
"step": 5790
},
{
"epoch": 0.92,
"grad_norm": 0.8135414719581604,
"learning_rate": 3.917091594875508e-05,
"loss": 0.8091,
"step": 5795
},
{
"epoch": 0.93,
"grad_norm": 0.7246249318122864,
"learning_rate": 3.915369805306278e-05,
"loss": 1.0,
"step": 5800
},
{
"epoch": 0.93,
"eval_loss": 0.7181897163391113,
"eval_runtime": 96.9106,
"eval_samples_per_second": 7.192,
"eval_steps_per_second": 7.192,
"step": 5800
},
{
"epoch": 0.93,
"grad_norm": 0.6413865089416504,
"learning_rate": 3.913647027127057e-05,
"loss": 0.8138,
"step": 5805
},
{
"epoch": 0.93,
"grad_norm": 0.7799732089042664,
"learning_rate": 3.9119232615411724e-05,
"loss": 0.74,
"step": 5810
},
{
"epoch": 0.93,
"grad_norm": 0.5068621635437012,
"learning_rate": 3.910198509752647e-05,
"loss": 0.6272,
"step": 5815
},
{
"epoch": 0.93,
"grad_norm": 0.6522389054298401,
"learning_rate": 3.9084727729661864e-05,
"loss": 0.747,
"step": 5820
},
{
"epoch": 0.93,
"grad_norm": 0.615208625793457,
"learning_rate": 3.9067460523871855e-05,
"loss": 1.0008,
"step": 5825
},
{
"epoch": 0.93,
"grad_norm": 0.7739670872688293,
"learning_rate": 3.9050183492217284e-05,
"loss": 0.959,
"step": 5830
},
{
"epoch": 0.93,
"grad_norm": 0.7364481687545776,
"learning_rate": 3.9032896646765837e-05,
"loss": 0.7106,
"step": 5835
},
{
"epoch": 0.93,
"grad_norm": 0.9992987513542175,
"learning_rate": 3.901559999959207e-05,
"loss": 0.6638,
"step": 5840
},
{
"epoch": 0.93,
"grad_norm": 0.9391655921936035,
"learning_rate": 3.8998293562777353e-05,
"loss": 0.8529,
"step": 5845
},
{
"epoch": 0.93,
"grad_norm": 0.5818854570388794,
"learning_rate": 3.8980977348409936e-05,
"loss": 0.8186,
"step": 5850
},
{
"epoch": 0.93,
"grad_norm": 0.7900370359420776,
"learning_rate": 3.8963651368584876e-05,
"loss": 0.6693,
"step": 5855
},
{
"epoch": 0.94,
"grad_norm": 0.6716821193695068,
"learning_rate": 3.8946315635404054e-05,
"loss": 0.7926,
"step": 5860
},
{
"epoch": 0.94,
"grad_norm": 0.6300959587097168,
"learning_rate": 3.892897016097615e-05,
"loss": 0.8177,
"step": 5865
},
{
"epoch": 0.94,
"grad_norm": 0.5477313995361328,
"learning_rate": 3.8911614957416675e-05,
"loss": 0.6463,
"step": 5870
},
{
"epoch": 0.94,
"grad_norm": 0.8673714995384216,
"learning_rate": 3.889425003684793e-05,
"loss": 0.7662,
"step": 5875
},
{
"epoch": 0.94,
"grad_norm": 0.9277421236038208,
"learning_rate": 3.8876875411398975e-05,
"loss": 0.7654,
"step": 5880
},
{
"epoch": 0.94,
"grad_norm": 1.111858606338501,
"learning_rate": 3.885949109320567e-05,
"loss": 0.7782,
"step": 5885
},
{
"epoch": 0.94,
"grad_norm": 0.8440257906913757,
"learning_rate": 3.884209709441066e-05,
"loss": 0.7054,
"step": 5890
},
{
"epoch": 0.94,
"grad_norm": 0.8476880192756653,
"learning_rate": 3.882469342716333e-05,
"loss": 0.7382,
"step": 5895
},
{
"epoch": 0.94,
"grad_norm": 0.5822775959968567,
"learning_rate": 3.8807280103619826e-05,
"loss": 0.6164,
"step": 5900
},
{
"epoch": 0.94,
"eval_loss": 0.7177068591117859,
"eval_runtime": 96.8009,
"eval_samples_per_second": 7.2,
"eval_steps_per_second": 7.2,
"step": 5900
},
{
"epoch": 0.94,
"grad_norm": 0.6871796250343323,
"learning_rate": 3.878985713594302e-05,
"loss": 0.6928,
"step": 5905
},
{
"epoch": 0.94,
"grad_norm": 0.7812449932098389,
"learning_rate": 3.8772424536302564e-05,
"loss": 0.8246,
"step": 5910
},
{
"epoch": 0.94,
"grad_norm": 0.8643015623092651,
"learning_rate": 3.87549823168748e-05,
"loss": 0.6893,
"step": 5915
},
{
"epoch": 0.94,
"grad_norm": 0.6011418104171753,
"learning_rate": 3.873753048984278e-05,
"loss": 0.5402,
"step": 5920
},
{
"epoch": 0.95,
"grad_norm": 0.6913355588912964,
"learning_rate": 3.8720069067396315e-05,
"loss": 0.6997,
"step": 5925
},
{
"epoch": 0.95,
"grad_norm": 0.840412437915802,
"learning_rate": 3.870259806173188e-05,
"loss": 0.7524,
"step": 5930
},
{
"epoch": 0.95,
"grad_norm": 0.7205111980438232,
"learning_rate": 3.8685117485052645e-05,
"loss": 0.7316,
"step": 5935
},
{
"epoch": 0.95,
"grad_norm": 0.5926095843315125,
"learning_rate": 3.8667627349568494e-05,
"loss": 0.6616,
"step": 5940
},
{
"epoch": 0.95,
"grad_norm": 0.8431223630905151,
"learning_rate": 3.865012766749595e-05,
"loss": 0.9906,
"step": 5945
},
{
"epoch": 0.95,
"grad_norm": 0.8787121176719666,
"learning_rate": 3.863261845105822e-05,
"loss": 0.7688,
"step": 5950
},
{
"epoch": 0.95,
"grad_norm": 0.5568470358848572,
"learning_rate": 3.861509971248519e-05,
"loss": 0.7319,
"step": 5955
},
{
"epoch": 0.95,
"grad_norm": 0.6905276775360107,
"learning_rate": 3.859757146401336e-05,
"loss": 0.8506,
"step": 5960
},
{
"epoch": 0.95,
"grad_norm": 0.733594536781311,
"learning_rate": 3.858003371788591e-05,
"loss": 0.65,
"step": 5965
},
{
"epoch": 0.95,
"grad_norm": 0.9627974033355713,
"learning_rate": 3.8562486486352614e-05,
"loss": 1.0121,
"step": 5970
},
{
"epoch": 0.95,
"grad_norm": 0.7356416583061218,
"learning_rate": 3.854492978166991e-05,
"loss": 0.698,
"step": 5975
},
{
"epoch": 0.95,
"grad_norm": 0.6784470081329346,
"learning_rate": 3.8527363616100844e-05,
"loss": 0.6361,
"step": 5980
},
{
"epoch": 0.96,
"grad_norm": 0.8226351141929626,
"learning_rate": 3.8509788001915035e-05,
"loss": 0.7528,
"step": 5985
},
{
"epoch": 0.96,
"grad_norm": 1.026175618171692,
"learning_rate": 3.8492202951388744e-05,
"loss": 1.0142,
"step": 5990
},
{
"epoch": 0.96,
"grad_norm": 0.860149085521698,
"learning_rate": 3.847460847680482e-05,
"loss": 0.6081,
"step": 5995
},
{
"epoch": 0.96,
"grad_norm": 0.7153595089912415,
"learning_rate": 3.845700459045266e-05,
"loss": 0.7484,
"step": 6000
},
{
"epoch": 0.96,
"eval_loss": 0.7146974802017212,
"eval_runtime": 96.5929,
"eval_samples_per_second": 7.216,
"eval_steps_per_second": 7.216,
"step": 6000
},
{
"epoch": 0.96,
"grad_norm": 0.8637629151344299,
"learning_rate": 3.843939130462827e-05,
"loss": 0.9717,
"step": 6005
},
{
"epoch": 0.96,
"grad_norm": 0.39873212575912476,
"learning_rate": 3.84217686316342e-05,
"loss": 0.64,
"step": 6010
},
{
"epoch": 0.96,
"grad_norm": 0.5839079022407532,
"learning_rate": 3.8404136583779585e-05,
"loss": 0.8288,
"step": 6015
},
{
"epoch": 0.96,
"grad_norm": 1.2063088417053223,
"learning_rate": 3.8386495173380064e-05,
"loss": 0.7722,
"step": 6020
},
{
"epoch": 0.96,
"grad_norm": 0.6402775049209595,
"learning_rate": 3.836884441275786e-05,
"loss": 0.7645,
"step": 6025
},
{
"epoch": 0.96,
"grad_norm": 0.8476734161376953,
"learning_rate": 3.83511843142417e-05,
"loss": 0.6249,
"step": 6030
},
{
"epoch": 0.96,
"grad_norm": 0.7386351823806763,
"learning_rate": 3.833351489016684e-05,
"loss": 0.6188,
"step": 6035
},
{
"epoch": 0.96,
"grad_norm": 0.661152184009552,
"learning_rate": 3.831583615287504e-05,
"loss": 0.9277,
"step": 6040
},
{
"epoch": 0.96,
"grad_norm": 0.6577073335647583,
"learning_rate": 3.82981481147146e-05,
"loss": 0.7678,
"step": 6045
},
{
"epoch": 0.97,
"grad_norm": 0.9322777986526489,
"learning_rate": 3.828045078804027e-05,
"loss": 0.6959,
"step": 6050
},
{
"epoch": 0.97,
"grad_norm": 1.2478550672531128,
"learning_rate": 3.826274418521332e-05,
"loss": 0.9087,
"step": 6055
},
{
"epoch": 0.97,
"grad_norm": 0.5843325257301331,
"learning_rate": 3.824502831860148e-05,
"loss": 0.5186,
"step": 6060
},
{
"epoch": 0.97,
"grad_norm": 0.4977940618991852,
"learning_rate": 3.822730320057898e-05,
"loss": 0.7431,
"step": 6065
},
{
"epoch": 0.97,
"grad_norm": 0.7447279691696167,
"learning_rate": 3.8209568843526475e-05,
"loss": 0.7302,
"step": 6070
},
{
"epoch": 0.97,
"grad_norm": 0.5054610967636108,
"learning_rate": 3.8191825259831095e-05,
"loss": 0.7878,
"step": 6075
},
{
"epoch": 0.97,
"grad_norm": 1.2159337997436523,
"learning_rate": 3.8174072461886396e-05,
"loss": 0.663,
"step": 6080
},
{
"epoch": 0.97,
"grad_norm": 0.8180432915687561,
"learning_rate": 3.815631046209241e-05,
"loss": 0.7184,
"step": 6085
},
{
"epoch": 0.97,
"grad_norm": 0.8640487790107727,
"learning_rate": 3.8138539272855544e-05,
"loss": 0.7011,
"step": 6090
},
{
"epoch": 0.97,
"grad_norm": 0.569340705871582,
"learning_rate": 3.812075890658867e-05,
"loss": 0.734,
"step": 6095
},
{
"epoch": 0.97,
"grad_norm": 0.774334728717804,
"learning_rate": 3.810296937571103e-05,
"loss": 0.7924,
"step": 6100
},
{
"epoch": 0.97,
"eval_loss": 0.7144004106521606,
"eval_runtime": 96.5311,
"eval_samples_per_second": 7.22,
"eval_steps_per_second": 7.22,
"step": 6100
},
{
"epoch": 0.97,
"grad_norm": 0.6256681680679321,
"learning_rate": 3.80851706926483e-05,
"loss": 0.6543,
"step": 6105
},
{
"epoch": 0.98,
"grad_norm": 0.6679534912109375,
"learning_rate": 3.806736286983254e-05,
"loss": 0.6674,
"step": 6110
},
{
"epoch": 0.98,
"grad_norm": 0.696412980556488,
"learning_rate": 3.804954591970218e-05,
"loss": 0.8335,
"step": 6115
},
{
"epoch": 0.98,
"grad_norm": 0.8444018363952637,
"learning_rate": 3.803171985470203e-05,
"loss": 0.7172,
"step": 6120
},
{
"epoch": 0.98,
"grad_norm": 0.8965733051300049,
"learning_rate": 3.8013884687283274e-05,
"loss": 1.0342,
"step": 6125
},
{
"epoch": 0.98,
"grad_norm": 0.7367607355117798,
"learning_rate": 3.799604042990347e-05,
"loss": 0.6862,
"step": 6130
},
{
"epoch": 0.98,
"grad_norm": 0.8707432746887207,
"learning_rate": 3.797818709502647e-05,
"loss": 0.8098,
"step": 6135
},
{
"epoch": 0.98,
"grad_norm": 0.5112609267234802,
"learning_rate": 3.7960324695122544e-05,
"loss": 0.5907,
"step": 6140
},
{
"epoch": 0.98,
"grad_norm": 0.39674660563468933,
"learning_rate": 3.794245324266823e-05,
"loss": 0.6887,
"step": 6145
},
{
"epoch": 0.98,
"grad_norm": 0.7135210037231445,
"learning_rate": 3.79245727501464e-05,
"loss": 0.7683,
"step": 6150
},
{
"epoch": 0.98,
"grad_norm": 0.6788827776908875,
"learning_rate": 3.790668323004628e-05,
"loss": 0.9064,
"step": 6155
},
{
"epoch": 0.98,
"grad_norm": 0.5234334468841553,
"learning_rate": 3.7888784694863355e-05,
"loss": 0.7373,
"step": 6160
},
{
"epoch": 0.98,
"grad_norm": 0.6748809218406677,
"learning_rate": 3.7870877157099425e-05,
"loss": 0.6998,
"step": 6165
},
{
"epoch": 0.98,
"grad_norm": 0.7746409773826599,
"learning_rate": 3.785296062926259e-05,
"loss": 0.7789,
"step": 6170
},
{
"epoch": 0.99,
"grad_norm": 0.7480994462966919,
"learning_rate": 3.783503512386721e-05,
"loss": 0.7114,
"step": 6175
},
{
"epoch": 0.99,
"grad_norm": 1.2271349430084229,
"learning_rate": 3.781710065343392e-05,
"loss": 0.7294,
"step": 6180
},
{
"epoch": 0.99,
"grad_norm": 0.7802749872207642,
"learning_rate": 3.779915723048963e-05,
"loss": 0.8348,
"step": 6185
},
{
"epoch": 0.99,
"grad_norm": 0.6718894243240356,
"learning_rate": 3.778120486756749e-05,
"loss": 0.7428,
"step": 6190
},
{
"epoch": 0.99,
"grad_norm": 0.7281386852264404,
"learning_rate": 3.776324357720688e-05,
"loss": 0.8076,
"step": 6195
},
{
"epoch": 0.99,
"grad_norm": 0.7086895704269409,
"learning_rate": 3.7745273371953464e-05,
"loss": 0.9389,
"step": 6200
},
{
"epoch": 0.99,
"eval_loss": 0.7145183682441711,
"eval_runtime": 96.6703,
"eval_samples_per_second": 7.21,
"eval_steps_per_second": 7.21,
"step": 6200
},
{
"epoch": 0.99,
"grad_norm": 0.8768818974494934,
"learning_rate": 3.7727294264359095e-05,
"loss": 0.8837,
"step": 6205
},
{
"epoch": 0.99,
"grad_norm": 0.5558096170425415,
"learning_rate": 3.7709306266981835e-05,
"loss": 0.6894,
"step": 6210
},
{
"epoch": 0.99,
"grad_norm": 0.7635754942893982,
"learning_rate": 3.769130939238598e-05,
"loss": 0.7869,
"step": 6215
},
{
"epoch": 0.99,
"grad_norm": 0.8129850625991821,
"learning_rate": 3.7673303653142036e-05,
"loss": 0.8049,
"step": 6220
},
{
"epoch": 0.99,
"grad_norm": 0.7782718539237976,
"learning_rate": 3.765528906182666e-05,
"loss": 1.1669,
"step": 6225
},
{
"epoch": 0.99,
"grad_norm": 0.8600115180015564,
"learning_rate": 3.763726563102272e-05,
"loss": 0.7453,
"step": 6230
},
{
"epoch": 1.0,
"grad_norm": 0.6311535239219666,
"learning_rate": 3.761923337331926e-05,
"loss": 0.7537,
"step": 6235
},
{
"epoch": 1.0,
"grad_norm": 0.7079928517341614,
"learning_rate": 3.7601192301311486e-05,
"loss": 0.8818,
"step": 6240
},
{
"epoch": 1.0,
"grad_norm": 0.6088127493858337,
"learning_rate": 3.7583142427600735e-05,
"loss": 0.8239,
"step": 6245
},
{
"epoch": 1.0,
"grad_norm": 0.9693393707275391,
"learning_rate": 3.756508376479455e-05,
"loss": 0.4772,
"step": 6250
},
{
"epoch": 1.0,
"grad_norm": 0.9866142868995667,
"learning_rate": 3.754701632550656e-05,
"loss": 0.9703,
"step": 6255
},
{
"epoch": 1.0,
"grad_norm": 0.7974732518196106,
"learning_rate": 3.7528940122356534e-05,
"loss": 0.7359,
"step": 6260
},
{
"epoch": 1.0,
"grad_norm": 0.7444310188293457,
"learning_rate": 3.751085516797038e-05,
"loss": 0.6785,
"step": 6265
},
{
"epoch": 1.0,
"grad_norm": 0.586471438407898,
"learning_rate": 3.7492761474980096e-05,
"loss": 0.6079,
"step": 6270
},
{
"epoch": 1.0,
"grad_norm": 0.7271062731742859,
"learning_rate": 3.747465905602382e-05,
"loss": 0.7284,
"step": 6275
},
{
"epoch": 1.0,
"grad_norm": 0.7004055380821228,
"learning_rate": 3.7456547923745745e-05,
"loss": 0.7364,
"step": 6280
},
{
"epoch": 1.0,
"grad_norm": 0.7666735053062439,
"learning_rate": 3.7438428090796166e-05,
"loss": 1.0261,
"step": 6285
},
{
"epoch": 1.0,
"grad_norm": 0.7594092488288879,
"learning_rate": 3.742029956983146e-05,
"loss": 0.7293,
"step": 6290
},
{
"epoch": 1.0,
"grad_norm": 0.5505672097206116,
"learning_rate": 3.7402162373514074e-05,
"loss": 1.1323,
"step": 6295
},
{
"epoch": 1.01,
"grad_norm": 0.6808300018310547,
"learning_rate": 3.73840165145125e-05,
"loss": 0.7108,
"step": 6300
},
{
"epoch": 1.01,
"eval_loss": 0.7136371731758118,
"eval_runtime": 96.6091,
"eval_samples_per_second": 7.215,
"eval_steps_per_second": 7.215,
"step": 6300
},
{
"epoch": 1.01,
"grad_norm": 1.147871494293213,
"learning_rate": 3.736586200550129e-05,
"loss": 1.0234,
"step": 6305
},
{
"epoch": 1.01,
"grad_norm": 0.7230010628700256,
"learning_rate": 3.734769885916104e-05,
"loss": 0.6727,
"step": 6310
},
{
"epoch": 1.01,
"grad_norm": 0.5923539400100708,
"learning_rate": 3.7329527088178376e-05,
"loss": 0.5398,
"step": 6315
},
{
"epoch": 1.01,
"grad_norm": 1.5129786729812622,
"learning_rate": 3.731134670524594e-05,
"loss": 0.8927,
"step": 6320
},
{
"epoch": 1.01,
"grad_norm": 0.6973879933357239,
"learning_rate": 3.7293157723062425e-05,
"loss": 0.5394,
"step": 6325
},
{
"epoch": 1.01,
"grad_norm": 0.6135405898094177,
"learning_rate": 3.727496015433246e-05,
"loss": 0.6949,
"step": 6330
},
{
"epoch": 1.01,
"grad_norm": 0.7291958928108215,
"learning_rate": 3.725675401176675e-05,
"loss": 0.7333,
"step": 6335
},
{
"epoch": 1.01,
"grad_norm": 0.8721144795417786,
"learning_rate": 3.7238539308081956e-05,
"loss": 0.7178,
"step": 6340
},
{
"epoch": 1.01,
"grad_norm": 0.6904857754707336,
"learning_rate": 3.722031605600069e-05,
"loss": 0.5423,
"step": 6345
},
{
"epoch": 1.01,
"grad_norm": 0.6890765428543091,
"learning_rate": 3.720208426825157e-05,
"loss": 0.6649,
"step": 6350
},
{
"epoch": 1.01,
"grad_norm": 0.7029244899749756,
"learning_rate": 3.7183843957569174e-05,
"loss": 0.6566,
"step": 6355
},
{
"epoch": 1.02,
"grad_norm": 0.7533852458000183,
"learning_rate": 3.716559513669403e-05,
"loss": 0.6503,
"step": 6360
},
{
"epoch": 1.02,
"grad_norm": 0.8862756490707397,
"learning_rate": 3.7147337818372595e-05,
"loss": 0.6742,
"step": 6365
},
{
"epoch": 1.02,
"grad_norm": 0.6946579217910767,
"learning_rate": 3.712907201535728e-05,
"loss": 0.6658,
"step": 6370
},
{
"epoch": 1.02,
"grad_norm": 0.7204101085662842,
"learning_rate": 3.711079774040641e-05,
"loss": 0.7723,
"step": 6375
},
{
"epoch": 1.02,
"grad_norm": 0.5266463756561279,
"learning_rate": 3.7092515006284245e-05,
"loss": 0.5676,
"step": 6380
},
{
"epoch": 1.02,
"grad_norm": 0.7050403952598572,
"learning_rate": 3.707422382576092e-05,
"loss": 0.6753,
"step": 6385
},
{
"epoch": 1.02,
"grad_norm": 0.6561117768287659,
"learning_rate": 3.705592421161251e-05,
"loss": 0.6988,
"step": 6390
},
{
"epoch": 1.02,
"grad_norm": 0.5257726311683655,
"learning_rate": 3.7037616176620955e-05,
"loss": 0.6265,
"step": 6395
},
{
"epoch": 1.02,
"grad_norm": 0.6152855753898621,
"learning_rate": 3.7019299733574074e-05,
"loss": 0.8076,
"step": 6400
},
{
"epoch": 1.02,
"eval_loss": 0.7153558731079102,
"eval_runtime": 96.493,
"eval_samples_per_second": 7.223,
"eval_steps_per_second": 7.223,
"step": 6400
},
{
"epoch": 1.02,
"grad_norm": 0.9646801948547363,
"learning_rate": 3.700097489526559e-05,
"loss": 0.8186,
"step": 6405
},
{
"epoch": 1.02,
"grad_norm": 0.8193671703338623,
"learning_rate": 3.6982641674495046e-05,
"loss": 0.5955,
"step": 6410
},
{
"epoch": 1.02,
"grad_norm": 0.5945185422897339,
"learning_rate": 3.696430008406788e-05,
"loss": 0.7449,
"step": 6415
},
{
"epoch": 1.02,
"grad_norm": 0.8218982219696045,
"learning_rate": 3.6945950136795345e-05,
"loss": 0.6192,
"step": 6420
},
{
"epoch": 1.03,
"grad_norm": 0.723145604133606,
"learning_rate": 3.692759184549455e-05,
"loss": 0.7703,
"step": 6425
},
{
"epoch": 1.03,
"grad_norm": 0.8936192989349365,
"learning_rate": 3.690922522298843e-05,
"loss": 0.6683,
"step": 6430
},
{
"epoch": 1.03,
"grad_norm": 0.6989235281944275,
"learning_rate": 3.6890850282105725e-05,
"loss": 0.5886,
"step": 6435
},
{
"epoch": 1.03,
"grad_norm": 0.9722846150398254,
"learning_rate": 3.687246703568101e-05,
"loss": 0.6229,
"step": 6440
},
{
"epoch": 1.03,
"grad_norm": 0.8043028116226196,
"learning_rate": 3.6854075496554636e-05,
"loss": 0.8865,
"step": 6445
},
{
"epoch": 1.03,
"grad_norm": 0.7527907490730286,
"learning_rate": 3.683567567757276e-05,
"loss": 0.6773,
"step": 6450
},
{
"epoch": 1.03,
"grad_norm": 0.6083170175552368,
"learning_rate": 3.681726759158733e-05,
"loss": 0.6311,
"step": 6455
},
{
"epoch": 1.03,
"grad_norm": 0.7950381636619568,
"learning_rate": 3.679885125145605e-05,
"loss": 0.8398,
"step": 6460
},
{
"epoch": 1.03,
"grad_norm": 0.8302904367446899,
"learning_rate": 3.678042667004239e-05,
"loss": 0.4981,
"step": 6465
},
{
"epoch": 1.03,
"grad_norm": 0.7814335227012634,
"learning_rate": 3.67619938602156e-05,
"loss": 0.8345,
"step": 6470
},
{
"epoch": 1.03,
"grad_norm": 1.1472254991531372,
"learning_rate": 3.6743552834850656e-05,
"loss": 0.7503,
"step": 6475
},
{
"epoch": 1.03,
"grad_norm": 0.7045092582702637,
"learning_rate": 3.672510360682827e-05,
"loss": 0.9974,
"step": 6480
},
{
"epoch": 1.04,
"grad_norm": 0.6515605449676514,
"learning_rate": 3.670664618903489e-05,
"loss": 0.7965,
"step": 6485
},
{
"epoch": 1.04,
"grad_norm": 0.6281563639640808,
"learning_rate": 3.66881805943627e-05,
"loss": 0.6854,
"step": 6490
},
{
"epoch": 1.04,
"grad_norm": 0.6973702907562256,
"learning_rate": 3.666970683570957e-05,
"loss": 0.6939,
"step": 6495
},
{
"epoch": 1.04,
"grad_norm": 0.7053531408309937,
"learning_rate": 3.6651224925979086e-05,
"loss": 0.7232,
"step": 6500
},
{
"epoch": 1.04,
"eval_loss": 0.7147144079208374,
"eval_runtime": 96.4093,
"eval_samples_per_second": 7.23,
"eval_steps_per_second": 7.23,
"step": 6500
}
],
"logging_steps": 5,
"max_steps": 18795,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"total_flos": 2.80509564321792e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}