llama-3-2-1b-sft / trainer_state.json
wassname's picture
Upload folder using huggingface_hub
07f81f5 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 200,
"global_step": 3742,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0002672367717797969,
"grad_norm": 8.823081418058603,
"learning_rate": 5.3333333333333334e-08,
"loss": 1.906,
"step": 1
},
{
"epoch": 0.0013361838588989846,
"grad_norm": 9.375402635266639,
"learning_rate": 2.666666666666667e-07,
"loss": 1.8044,
"step": 5
},
{
"epoch": 0.002672367717797969,
"grad_norm": 8.65210459980887,
"learning_rate": 5.333333333333335e-07,
"loss": 1.8182,
"step": 10
},
{
"epoch": 0.004008551576696953,
"grad_norm": 6.51103320753397,
"learning_rate": 8.000000000000001e-07,
"loss": 1.7509,
"step": 15
},
{
"epoch": 0.005344735435595938,
"grad_norm": 4.268692783385918,
"learning_rate": 1.066666666666667e-06,
"loss": 1.767,
"step": 20
},
{
"epoch": 0.006680919294494923,
"grad_norm": 4.3863193537387355,
"learning_rate": 1.3333333333333334e-06,
"loss": 1.7519,
"step": 25
},
{
"epoch": 0.008017103153393906,
"grad_norm": 3.465625691597637,
"learning_rate": 1.6000000000000001e-06,
"loss": 1.7079,
"step": 30
},
{
"epoch": 0.00935328701229289,
"grad_norm": 2.6102827946343203,
"learning_rate": 1.8666666666666669e-06,
"loss": 1.6753,
"step": 35
},
{
"epoch": 0.010689470871191877,
"grad_norm": 2.3789490657646937,
"learning_rate": 2.133333333333334e-06,
"loss": 1.6337,
"step": 40
},
{
"epoch": 0.012025654730090861,
"grad_norm": 2.0835126121909235,
"learning_rate": 2.4000000000000003e-06,
"loss": 1.6641,
"step": 45
},
{
"epoch": 0.013361838588989846,
"grad_norm": 1.8559179146822158,
"learning_rate": 2.666666666666667e-06,
"loss": 1.6296,
"step": 50
},
{
"epoch": 0.01469802244788883,
"grad_norm": 1.8106018636994041,
"learning_rate": 2.9333333333333338e-06,
"loss": 1.5509,
"step": 55
},
{
"epoch": 0.016034206306787813,
"grad_norm": 1.8755467245728914,
"learning_rate": 3.2000000000000003e-06,
"loss": 1.5601,
"step": 60
},
{
"epoch": 0.017370390165686797,
"grad_norm": 1.7318124276404359,
"learning_rate": 3.4666666666666672e-06,
"loss": 1.5308,
"step": 65
},
{
"epoch": 0.01870657402458578,
"grad_norm": 1.6976770309274594,
"learning_rate": 3.7333333333333337e-06,
"loss": 1.5609,
"step": 70
},
{
"epoch": 0.020042757883484766,
"grad_norm": 1.7565922656582902,
"learning_rate": 4.000000000000001e-06,
"loss": 1.5082,
"step": 75
},
{
"epoch": 0.021378941742383754,
"grad_norm": 1.7393962369808933,
"learning_rate": 4.266666666666668e-06,
"loss": 1.5306,
"step": 80
},
{
"epoch": 0.022715125601282738,
"grad_norm": 1.720884117447266,
"learning_rate": 4.533333333333334e-06,
"loss": 1.5378,
"step": 85
},
{
"epoch": 0.024051309460181722,
"grad_norm": 1.7101397134036984,
"learning_rate": 4.800000000000001e-06,
"loss": 1.4813,
"step": 90
},
{
"epoch": 0.025387493319080707,
"grad_norm": 1.615617842911665,
"learning_rate": 5.0666666666666676e-06,
"loss": 1.4817,
"step": 95
},
{
"epoch": 0.02672367717797969,
"grad_norm": 2.691934178931898,
"learning_rate": 5.333333333333334e-06,
"loss": 1.4931,
"step": 100
},
{
"epoch": 0.028059861036878676,
"grad_norm": 1.5821332066035256,
"learning_rate": 5.600000000000001e-06,
"loss": 1.4285,
"step": 105
},
{
"epoch": 0.02939604489577766,
"grad_norm": 1.7190020107955757,
"learning_rate": 5.8666666666666675e-06,
"loss": 1.4481,
"step": 110
},
{
"epoch": 0.030732228754676644,
"grad_norm": 1.6645689315016792,
"learning_rate": 6.133333333333334e-06,
"loss": 1.465,
"step": 115
},
{
"epoch": 0.032068412613575625,
"grad_norm": 1.6462290185572064,
"learning_rate": 6.4000000000000006e-06,
"loss": 1.4665,
"step": 120
},
{
"epoch": 0.03340459647247461,
"grad_norm": 1.649228040380627,
"learning_rate": 6.666666666666667e-06,
"loss": 1.4209,
"step": 125
},
{
"epoch": 0.034740780331373594,
"grad_norm": 1.6335763283148064,
"learning_rate": 6.9333333333333344e-06,
"loss": 1.4151,
"step": 130
},
{
"epoch": 0.03607696419027258,
"grad_norm": 1.6194942414847415,
"learning_rate": 7.2000000000000005e-06,
"loss": 1.4377,
"step": 135
},
{
"epoch": 0.03741314804917156,
"grad_norm": 1.7105888808318117,
"learning_rate": 7.4666666666666675e-06,
"loss": 1.3934,
"step": 140
},
{
"epoch": 0.03874933190807055,
"grad_norm": 1.4814484480334174,
"learning_rate": 7.733333333333334e-06,
"loss": 1.4015,
"step": 145
},
{
"epoch": 0.04008551576696953,
"grad_norm": 1.5777144338014444,
"learning_rate": 8.000000000000001e-06,
"loss": 1.4021,
"step": 150
},
{
"epoch": 0.041421699625868516,
"grad_norm": 1.5930794844441725,
"learning_rate": 8.266666666666667e-06,
"loss": 1.434,
"step": 155
},
{
"epoch": 0.04275788348476751,
"grad_norm": 1.6091950467097091,
"learning_rate": 8.533333333333335e-06,
"loss": 1.3956,
"step": 160
},
{
"epoch": 0.04409406734366649,
"grad_norm": 1.6975335096598412,
"learning_rate": 8.8e-06,
"loss": 1.3996,
"step": 165
},
{
"epoch": 0.045430251202565476,
"grad_norm": 1.5343847156625996,
"learning_rate": 9.066666666666667e-06,
"loss": 1.3799,
"step": 170
},
{
"epoch": 0.04676643506146446,
"grad_norm": 1.770466644954361,
"learning_rate": 9.333333333333334e-06,
"loss": 1.3545,
"step": 175
},
{
"epoch": 0.048102618920363445,
"grad_norm": 1.5660422280689605,
"learning_rate": 9.600000000000001e-06,
"loss": 1.4292,
"step": 180
},
{
"epoch": 0.04943880277926243,
"grad_norm": 1.536676360877846,
"learning_rate": 9.866666666666668e-06,
"loss": 1.4028,
"step": 185
},
{
"epoch": 0.050774986638161414,
"grad_norm": 1.4804803919916396,
"learning_rate": 1.0133333333333335e-05,
"loss": 1.3899,
"step": 190
},
{
"epoch": 0.0521111704970604,
"grad_norm": 1.6476917950123358,
"learning_rate": 1.04e-05,
"loss": 1.4214,
"step": 195
},
{
"epoch": 0.05344735435595938,
"grad_norm": 1.5933336407770893,
"learning_rate": 1.0666666666666667e-05,
"loss": 1.3663,
"step": 200
},
{
"epoch": 0.05344735435595938,
"eval_loss": 1.3954896926879883,
"eval_runtime": 525.9836,
"eval_samples_per_second": 25.183,
"eval_steps_per_second": 3.148,
"step": 200
},
{
"epoch": 0.05478353821485837,
"grad_norm": 1.55465727793257,
"learning_rate": 1.0933333333333334e-05,
"loss": 1.3566,
"step": 205
},
{
"epoch": 0.05611972207375735,
"grad_norm": 1.6465837885812638,
"learning_rate": 1.1200000000000001e-05,
"loss": 1.3572,
"step": 210
},
{
"epoch": 0.057455905932656336,
"grad_norm": 1.6314586937231732,
"learning_rate": 1.1466666666666668e-05,
"loss": 1.4066,
"step": 215
},
{
"epoch": 0.05879208979155532,
"grad_norm": 1.5940186416338837,
"learning_rate": 1.1733333333333335e-05,
"loss": 1.3727,
"step": 220
},
{
"epoch": 0.060128273650454304,
"grad_norm": 1.587832961506543,
"learning_rate": 1.2e-05,
"loss": 1.3755,
"step": 225
},
{
"epoch": 0.06146445750935329,
"grad_norm": 1.5731554504380694,
"learning_rate": 1.2266666666666667e-05,
"loss": 1.4007,
"step": 230
},
{
"epoch": 0.06280064136825227,
"grad_norm": 1.5722181818634398,
"learning_rate": 1.2533333333333336e-05,
"loss": 1.3783,
"step": 235
},
{
"epoch": 0.06413682522715125,
"grad_norm": 1.467532784965517,
"learning_rate": 1.2800000000000001e-05,
"loss": 1.4386,
"step": 240
},
{
"epoch": 0.06547300908605024,
"grad_norm": 1.5125828644122057,
"learning_rate": 1.3066666666666668e-05,
"loss": 1.3186,
"step": 245
},
{
"epoch": 0.06680919294494922,
"grad_norm": 1.5183168300505157,
"learning_rate": 1.3333333333333333e-05,
"loss": 1.3551,
"step": 250
},
{
"epoch": 0.0681453768038482,
"grad_norm": 1.4941218649910812,
"learning_rate": 1.3600000000000002e-05,
"loss": 1.3664,
"step": 255
},
{
"epoch": 0.06948156066274719,
"grad_norm": 1.4869296436066686,
"learning_rate": 1.3866666666666669e-05,
"loss": 1.3576,
"step": 260
},
{
"epoch": 0.07081774452164617,
"grad_norm": 1.4293831791321558,
"learning_rate": 1.4133333333333334e-05,
"loss": 1.3556,
"step": 265
},
{
"epoch": 0.07215392838054516,
"grad_norm": 1.5191533813334088,
"learning_rate": 1.4400000000000001e-05,
"loss": 1.3366,
"step": 270
},
{
"epoch": 0.07349011223944414,
"grad_norm": 1.4810082339187982,
"learning_rate": 1.4666666666666666e-05,
"loss": 1.3573,
"step": 275
},
{
"epoch": 0.07482629609834313,
"grad_norm": 1.5063612979974181,
"learning_rate": 1.4933333333333335e-05,
"loss": 1.3427,
"step": 280
},
{
"epoch": 0.07616247995724211,
"grad_norm": 1.618950454636469,
"learning_rate": 1.5200000000000002e-05,
"loss": 1.3554,
"step": 285
},
{
"epoch": 0.0774986638161411,
"grad_norm": 1.5626468747954396,
"learning_rate": 1.546666666666667e-05,
"loss": 1.3771,
"step": 290
},
{
"epoch": 0.07883484767504008,
"grad_norm": 1.5292548773347097,
"learning_rate": 1.5733333333333334e-05,
"loss": 1.3918,
"step": 295
},
{
"epoch": 0.08017103153393906,
"grad_norm": 1.5358359721754233,
"learning_rate": 1.6000000000000003e-05,
"loss": 1.3815,
"step": 300
},
{
"epoch": 0.08150721539283805,
"grad_norm": 1.3953477584618263,
"learning_rate": 1.6266666666666668e-05,
"loss": 1.3854,
"step": 305
},
{
"epoch": 0.08284339925173703,
"grad_norm": 1.4468372312833726,
"learning_rate": 1.6533333333333333e-05,
"loss": 1.3727,
"step": 310
},
{
"epoch": 0.08417958311063603,
"grad_norm": 1.488344751961309,
"learning_rate": 1.6800000000000002e-05,
"loss": 1.3706,
"step": 315
},
{
"epoch": 0.08551576696953501,
"grad_norm": 1.4093480949221249,
"learning_rate": 1.706666666666667e-05,
"loss": 1.3491,
"step": 320
},
{
"epoch": 0.086851950828434,
"grad_norm": 1.5432094268322356,
"learning_rate": 1.7333333333333336e-05,
"loss": 1.3586,
"step": 325
},
{
"epoch": 0.08818813468733298,
"grad_norm": 1.4550377653252384,
"learning_rate": 1.76e-05,
"loss": 1.39,
"step": 330
},
{
"epoch": 0.08952431854623197,
"grad_norm": 1.4281640050412778,
"learning_rate": 1.7866666666666666e-05,
"loss": 1.3669,
"step": 335
},
{
"epoch": 0.09086050240513095,
"grad_norm": 1.4426556416465324,
"learning_rate": 1.8133333333333335e-05,
"loss": 1.3876,
"step": 340
},
{
"epoch": 0.09219668626402994,
"grad_norm": 1.406168800248235,
"learning_rate": 1.8400000000000003e-05,
"loss": 1.3686,
"step": 345
},
{
"epoch": 0.09353287012292892,
"grad_norm": 1.4301962338673828,
"learning_rate": 1.866666666666667e-05,
"loss": 1.3689,
"step": 350
},
{
"epoch": 0.0948690539818279,
"grad_norm": 1.6044731560961005,
"learning_rate": 1.8933333333333334e-05,
"loss": 1.4118,
"step": 355
},
{
"epoch": 0.09620523784072689,
"grad_norm": 1.4879888050516272,
"learning_rate": 1.9200000000000003e-05,
"loss": 1.3791,
"step": 360
},
{
"epoch": 0.09754142169962587,
"grad_norm": 1.4698368446133456,
"learning_rate": 1.9466666666666668e-05,
"loss": 1.348,
"step": 365
},
{
"epoch": 0.09887760555852486,
"grad_norm": 1.4186637789128058,
"learning_rate": 1.9733333333333336e-05,
"loss": 1.3325,
"step": 370
},
{
"epoch": 0.10021378941742384,
"grad_norm": 1.5476694754206417,
"learning_rate": 2e-05,
"loss": 1.3642,
"step": 375
},
{
"epoch": 0.10154997327632283,
"grad_norm": 1.4733865828906345,
"learning_rate": 1.9999891176487904e-05,
"loss": 1.3807,
"step": 380
},
{
"epoch": 0.10288615713522181,
"grad_norm": 1.4551876993221122,
"learning_rate": 1.9999564708320124e-05,
"loss": 1.3679,
"step": 385
},
{
"epoch": 0.1042223409941208,
"grad_norm": 1.5257129447187303,
"learning_rate": 1.999902060260214e-05,
"loss": 1.4111,
"step": 390
},
{
"epoch": 0.10555852485301978,
"grad_norm": 1.3923714980063469,
"learning_rate": 1.9998258871176252e-05,
"loss": 1.3729,
"step": 395
},
{
"epoch": 0.10689470871191876,
"grad_norm": 1.4311098965377238,
"learning_rate": 1.999727953062132e-05,
"loss": 1.3413,
"step": 400
},
{
"epoch": 0.10689470871191876,
"eval_loss": 1.3721706867218018,
"eval_runtime": 526.6891,
"eval_samples_per_second": 25.15,
"eval_steps_per_second": 3.144,
"step": 400
},
{
"epoch": 0.10823089257081775,
"grad_norm": 1.3829932271044623,
"learning_rate": 1.99960826022524e-05,
"loss": 1.3466,
"step": 405
},
{
"epoch": 0.10956707642971673,
"grad_norm": 1.3521931502381987,
"learning_rate": 1.9994668112120283e-05,
"loss": 1.3713,
"step": 410
},
{
"epoch": 0.11090326028861572,
"grad_norm": 1.4976307839881127,
"learning_rate": 1.9993036091010924e-05,
"loss": 1.3564,
"step": 415
},
{
"epoch": 0.1122394441475147,
"grad_norm": 1.3701705887791993,
"learning_rate": 1.999118657444477e-05,
"loss": 1.3666,
"step": 420
},
{
"epoch": 0.11357562800641369,
"grad_norm": 1.4279953676541446,
"learning_rate": 1.9989119602676007e-05,
"loss": 1.3498,
"step": 425
},
{
"epoch": 0.11491181186531267,
"grad_norm": 1.4077392862076545,
"learning_rate": 1.9986835220691662e-05,
"loss": 1.3655,
"step": 430
},
{
"epoch": 0.11624799572421166,
"grad_norm": 1.5303947503986481,
"learning_rate": 1.9984333478210622e-05,
"loss": 1.3758,
"step": 435
},
{
"epoch": 0.11758417958311064,
"grad_norm": 1.4002227761633332,
"learning_rate": 1.9981614429682576e-05,
"loss": 1.338,
"step": 440
},
{
"epoch": 0.11892036344200962,
"grad_norm": 1.3305325078036485,
"learning_rate": 1.9978678134286796e-05,
"loss": 1.3502,
"step": 445
},
{
"epoch": 0.12025654730090861,
"grad_norm": 1.4218348465021207,
"learning_rate": 1.9975524655930884e-05,
"loss": 1.3526,
"step": 450
},
{
"epoch": 0.1215927311598076,
"grad_norm": 1.3929817535643747,
"learning_rate": 1.997215406324936e-05,
"loss": 1.3575,
"step": 455
},
{
"epoch": 0.12292891501870658,
"grad_norm": 1.4164344515855345,
"learning_rate": 1.9968566429602166e-05,
"loss": 1.3548,
"step": 460
},
{
"epoch": 0.12426509887760556,
"grad_norm": 1.3940688971110364,
"learning_rate": 1.996476183307308e-05,
"loss": 1.3651,
"step": 465
},
{
"epoch": 0.12560128273650453,
"grad_norm": 1.4429246742185509,
"learning_rate": 1.996074035646802e-05,
"loss": 1.3566,
"step": 470
},
{
"epoch": 0.12693746659540353,
"grad_norm": 1.4139926227619217,
"learning_rate": 1.9956502087313217e-05,
"loss": 1.4029,
"step": 475
},
{
"epoch": 0.1282736504543025,
"grad_norm": 1.36708702924336,
"learning_rate": 1.9952047117853345e-05,
"loss": 1.3558,
"step": 480
},
{
"epoch": 0.1296098343132015,
"grad_norm": 1.4333918113560515,
"learning_rate": 1.994737554504949e-05,
"loss": 1.3735,
"step": 485
},
{
"epoch": 0.13094601817210047,
"grad_norm": 1.3176020208151544,
"learning_rate": 1.994248747057704e-05,
"loss": 1.341,
"step": 490
},
{
"epoch": 0.13228220203099947,
"grad_norm": 1.4308466916521525,
"learning_rate": 1.9937383000823485e-05,
"loss": 1.405,
"step": 495
},
{
"epoch": 0.13361838588989844,
"grad_norm": 1.4229715319544036,
"learning_rate": 1.9932062246886087e-05,
"loss": 1.3655,
"step": 500
},
{
"epoch": 0.13495456974879744,
"grad_norm": 1.3858117778473207,
"learning_rate": 1.992652532456947e-05,
"loss": 1.3533,
"step": 505
},
{
"epoch": 0.1362907536076964,
"grad_norm": 1.3134241462309153,
"learning_rate": 1.992077235438311e-05,
"loss": 1.3152,
"step": 510
},
{
"epoch": 0.1376269374665954,
"grad_norm": 1.365018139049316,
"learning_rate": 1.991480346153868e-05,
"loss": 1.3399,
"step": 515
},
{
"epoch": 0.13896312132549438,
"grad_norm": 1.426928066531938,
"learning_rate": 1.9908618775947364e-05,
"loss": 1.3933,
"step": 520
},
{
"epoch": 0.14029930518439337,
"grad_norm": 1.428258632489371,
"learning_rate": 1.9902218432216996e-05,
"loss": 1.4028,
"step": 525
},
{
"epoch": 0.14163548904329235,
"grad_norm": 1.4241733392483251,
"learning_rate": 1.989560256964916e-05,
"loss": 1.4106,
"step": 530
},
{
"epoch": 0.14297167290219134,
"grad_norm": 1.3754571749169886,
"learning_rate": 1.9888771332236137e-05,
"loss": 1.383,
"step": 535
},
{
"epoch": 0.14430785676109031,
"grad_norm": 1.3253761034149971,
"learning_rate": 1.9881724868657768e-05,
"loss": 1.3846,
"step": 540
},
{
"epoch": 0.1456440406199893,
"grad_norm": 1.3292314064664583,
"learning_rate": 1.9874463332278245e-05,
"loss": 1.3831,
"step": 545
},
{
"epoch": 0.14698022447888828,
"grad_norm": 1.5134929576884686,
"learning_rate": 1.9866986881142737e-05,
"loss": 1.3535,
"step": 550
},
{
"epoch": 0.14831640833778728,
"grad_norm": 1.4571661850393143,
"learning_rate": 1.9859295677973988e-05,
"loss": 1.3482,
"step": 555
},
{
"epoch": 0.14965259219668625,
"grad_norm": 1.3720756399964669,
"learning_rate": 1.9851389890168738e-05,
"loss": 1.3677,
"step": 560
},
{
"epoch": 0.15098877605558525,
"grad_norm": 1.3902503536690078,
"learning_rate": 1.9843269689794114e-05,
"loss": 1.3201,
"step": 565
},
{
"epoch": 0.15232495991448422,
"grad_norm": 1.4845499289340192,
"learning_rate": 1.983493525358385e-05,
"loss": 1.4121,
"step": 570
},
{
"epoch": 0.15366114377338322,
"grad_norm": 1.5074167691063687,
"learning_rate": 1.982638676293448e-05,
"loss": 1.3842,
"step": 575
},
{
"epoch": 0.1549973276322822,
"grad_norm": 1.386731826246787,
"learning_rate": 1.981762440390136e-05,
"loss": 1.3248,
"step": 580
},
{
"epoch": 0.1563335114911812,
"grad_norm": 1.398496405386941,
"learning_rate": 1.9808648367194614e-05,
"loss": 1.4116,
"step": 585
},
{
"epoch": 0.15766969535008016,
"grad_norm": 1.3353006312197904,
"learning_rate": 1.9799458848175023e-05,
"loss": 1.3557,
"step": 590
},
{
"epoch": 0.15900587920897916,
"grad_norm": 1.3229490481790953,
"learning_rate": 1.9790056046849726e-05,
"loss": 1.3425,
"step": 595
},
{
"epoch": 0.16034206306787813,
"grad_norm": 1.3413421989191214,
"learning_rate": 1.97804401678679e-05,
"loss": 1.365,
"step": 600
},
{
"epoch": 0.16034206306787813,
"eval_loss": 1.3631840944290161,
"eval_runtime": 523.8522,
"eval_samples_per_second": 25.286,
"eval_steps_per_second": 3.161,
"step": 600
},
{
"epoch": 0.16167824692677712,
"grad_norm": 1.3574397972892758,
"learning_rate": 1.9770611420516286e-05,
"loss": 1.3677,
"step": 605
},
{
"epoch": 0.1630144307856761,
"grad_norm": 1.3451601369327404,
"learning_rate": 1.9760570018714647e-05,
"loss": 1.3509,
"step": 610
},
{
"epoch": 0.1643506146445751,
"grad_norm": 1.384864628214446,
"learning_rate": 1.975031618101111e-05,
"loss": 1.3258,
"step": 615
},
{
"epoch": 0.16568679850347406,
"grad_norm": 1.3682344543460991,
"learning_rate": 1.9739850130577393e-05,
"loss": 1.3552,
"step": 620
},
{
"epoch": 0.16702298236237306,
"grad_norm": 1.4305243117138575,
"learning_rate": 1.9729172095203977e-05,
"loss": 1.3849,
"step": 625
},
{
"epoch": 0.16835916622127206,
"grad_norm": 1.3723669913018262,
"learning_rate": 1.9718282307295115e-05,
"loss": 1.3659,
"step": 630
},
{
"epoch": 0.16969535008017103,
"grad_norm": 1.3516044026149312,
"learning_rate": 1.970718100386381e-05,
"loss": 1.3736,
"step": 635
},
{
"epoch": 0.17103153393907003,
"grad_norm": 1.436359994097034,
"learning_rate": 1.969586842652662e-05,
"loss": 1.3818,
"step": 640
},
{
"epoch": 0.172367717797969,
"grad_norm": 1.4157888272139716,
"learning_rate": 1.9684344821498432e-05,
"loss": 1.3342,
"step": 645
},
{
"epoch": 0.173703901656868,
"grad_norm": 1.3227193318099222,
"learning_rate": 1.9672610439587073e-05,
"loss": 1.3862,
"step": 650
},
{
"epoch": 0.17504008551576697,
"grad_norm": 1.3534338266553816,
"learning_rate": 1.9660665536187875e-05,
"loss": 1.3617,
"step": 655
},
{
"epoch": 0.17637626937466597,
"grad_norm": 1.3075745113464525,
"learning_rate": 1.9648510371278106e-05,
"loss": 1.351,
"step": 660
},
{
"epoch": 0.17771245323356494,
"grad_norm": 1.3543581869904175,
"learning_rate": 1.9636145209411318e-05,
"loss": 1.3866,
"step": 665
},
{
"epoch": 0.17904863709246394,
"grad_norm": 1.3462890650035784,
"learning_rate": 1.9623570319711574e-05,
"loss": 1.3754,
"step": 670
},
{
"epoch": 0.1803848209513629,
"grad_norm": 1.2991571392936196,
"learning_rate": 1.9610785975867608e-05,
"loss": 1.323,
"step": 675
},
{
"epoch": 0.1817210048102619,
"grad_norm": 1.434385743767408,
"learning_rate": 1.9597792456126855e-05,
"loss": 1.383,
"step": 680
},
{
"epoch": 0.18305718866916088,
"grad_norm": 1.3595317685749773,
"learning_rate": 1.9584590043289416e-05,
"loss": 1.2787,
"step": 685
},
{
"epoch": 0.18439337252805987,
"grad_norm": 1.6484220578907431,
"learning_rate": 1.957117902470187e-05,
"loss": 1.3747,
"step": 690
},
{
"epoch": 0.18572955638695884,
"grad_norm": 1.2945139560173622,
"learning_rate": 1.9557559692251047e-05,
"loss": 1.3621,
"step": 695
},
{
"epoch": 0.18706574024585784,
"grad_norm": 1.343012210724551,
"learning_rate": 1.9543732342357664e-05,
"loss": 1.3205,
"step": 700
},
{
"epoch": 0.1884019241047568,
"grad_norm": 1.3223543378479201,
"learning_rate": 1.9529697275969876e-05,
"loss": 1.3101,
"step": 705
},
{
"epoch": 0.1897381079636558,
"grad_norm": 1.3069901527523717,
"learning_rate": 1.951545479855673e-05,
"loss": 1.3752,
"step": 710
},
{
"epoch": 0.19107429182255478,
"grad_norm": 1.3767101616598136,
"learning_rate": 1.9501005220101507e-05,
"loss": 1.3854,
"step": 715
},
{
"epoch": 0.19241047568145378,
"grad_norm": 1.3595423389248131,
"learning_rate": 1.948634885509498e-05,
"loss": 1.3708,
"step": 720
},
{
"epoch": 0.19374665954035275,
"grad_norm": 1.269545504198035,
"learning_rate": 1.947148602252858e-05,
"loss": 1.3443,
"step": 725
},
{
"epoch": 0.19508284339925175,
"grad_norm": 1.2745742850192638,
"learning_rate": 1.9456417045887423e-05,
"loss": 1.3339,
"step": 730
},
{
"epoch": 0.19641902725815072,
"grad_norm": 1.4182072631536495,
"learning_rate": 1.944114225314331e-05,
"loss": 1.3103,
"step": 735
},
{
"epoch": 0.19775521111704972,
"grad_norm": 1.3233791325230955,
"learning_rate": 1.9425661976747552e-05,
"loss": 1.3625,
"step": 740
},
{
"epoch": 0.1990913949759487,
"grad_norm": 1.4760222437081327,
"learning_rate": 1.9409976553623767e-05,
"loss": 1.3624,
"step": 745
},
{
"epoch": 0.20042757883484769,
"grad_norm": 1.4210858490294298,
"learning_rate": 1.9394086325160515e-05,
"loss": 1.3599,
"step": 750
},
{
"epoch": 0.20176376269374666,
"grad_norm": 1.3258331784580841,
"learning_rate": 1.9377991637203894e-05,
"loss": 1.3411,
"step": 755
},
{
"epoch": 0.20309994655264565,
"grad_norm": 1.5486410395081083,
"learning_rate": 1.9361692840049997e-05,
"loss": 1.4124,
"step": 760
},
{
"epoch": 0.20443613041154463,
"grad_norm": 1.3108858505753092,
"learning_rate": 1.9345190288437292e-05,
"loss": 1.3323,
"step": 765
},
{
"epoch": 0.20577231427044362,
"grad_norm": 1.3401186763256265,
"learning_rate": 1.9328484341538903e-05,
"loss": 1.3589,
"step": 770
},
{
"epoch": 0.2071084981293426,
"grad_norm": 1.3258497911117144,
"learning_rate": 1.93115753629548e-05,
"loss": 1.3263,
"step": 775
},
{
"epoch": 0.2084446819882416,
"grad_norm": 1.279705605242652,
"learning_rate": 1.929446372070386e-05,
"loss": 1.3577,
"step": 780
},
{
"epoch": 0.20978086584714056,
"grad_norm": 1.3688317662377307,
"learning_rate": 1.9277149787215893e-05,
"loss": 1.3325,
"step": 785
},
{
"epoch": 0.21111704970603956,
"grad_norm": 1.3308089715501914,
"learning_rate": 1.9259633939323504e-05,
"loss": 1.3594,
"step": 790
},
{
"epoch": 0.21245323356493853,
"grad_norm": 1.3306187188699843,
"learning_rate": 1.924191655825391e-05,
"loss": 1.3706,
"step": 795
},
{
"epoch": 0.21378941742383753,
"grad_norm": 1.3079863145055592,
"learning_rate": 1.922399802962064e-05,
"loss": 1.33,
"step": 800
},
{
"epoch": 0.21378941742383753,
"eval_loss": 1.3532133102416992,
"eval_runtime": 523.8533,
"eval_samples_per_second": 25.286,
"eval_steps_per_second": 3.161,
"step": 800
},
{
"epoch": 0.2151256012827365,
"grad_norm": 1.2760851031416467,
"learning_rate": 1.9205878743415137e-05,
"loss": 1.321,
"step": 805
},
{
"epoch": 0.2164617851416355,
"grad_norm": 1.278778697243769,
"learning_rate": 1.9187559093998275e-05,
"loss": 1.3632,
"step": 810
},
{
"epoch": 0.21779796900053447,
"grad_norm": 1.3321470791378587,
"learning_rate": 1.916903948009177e-05,
"loss": 1.3079,
"step": 815
},
{
"epoch": 0.21913415285943347,
"grad_norm": 1.3698830796691455,
"learning_rate": 1.915032030476951e-05,
"loss": 1.3835,
"step": 820
},
{
"epoch": 0.22047033671833244,
"grad_norm": 1.314469593484434,
"learning_rate": 1.913140197544877e-05,
"loss": 1.3423,
"step": 825
},
{
"epoch": 0.22180652057723144,
"grad_norm": 1.3585983787357763,
"learning_rate": 1.911228490388136e-05,
"loss": 1.3038,
"step": 830
},
{
"epoch": 0.2231427044361304,
"grad_norm": 1.309351148797077,
"learning_rate": 1.9092969506144653e-05,
"loss": 1.3711,
"step": 835
},
{
"epoch": 0.2244788882950294,
"grad_norm": 1.2522351293958003,
"learning_rate": 1.907345620263254e-05,
"loss": 1.3436,
"step": 840
},
{
"epoch": 0.22581507215392838,
"grad_norm": 1.2176016475207827,
"learning_rate": 1.9053745418046257e-05,
"loss": 1.3544,
"step": 845
},
{
"epoch": 0.22715125601282737,
"grad_norm": 1.3937619861187576,
"learning_rate": 1.903383758138517e-05,
"loss": 1.3563,
"step": 850
},
{
"epoch": 0.22848743987172634,
"grad_norm": 1.3085996735753536,
"learning_rate": 1.9013733125937412e-05,
"loss": 1.3149,
"step": 855
},
{
"epoch": 0.22982362373062534,
"grad_norm": 1.5210379330796175,
"learning_rate": 1.8993432489270484e-05,
"loss": 1.3202,
"step": 860
},
{
"epoch": 0.2311598075895243,
"grad_norm": 1.3435267655155434,
"learning_rate": 1.8972936113221696e-05,
"loss": 1.357,
"step": 865
},
{
"epoch": 0.2324959914484233,
"grad_norm": 1.32208190242327,
"learning_rate": 1.8952244443888573e-05,
"loss": 1.3838,
"step": 870
},
{
"epoch": 0.23383217530732228,
"grad_norm": 1.3123583179163099,
"learning_rate": 1.8931357931619143e-05,
"loss": 1.3576,
"step": 875
},
{
"epoch": 0.23516835916622128,
"grad_norm": 1.2321415331674137,
"learning_rate": 1.8910277031002125e-05,
"loss": 1.3413,
"step": 880
},
{
"epoch": 0.23650454302512025,
"grad_norm": 1.2717606000257338,
"learning_rate": 1.888900220085706e-05,
"loss": 1.3412,
"step": 885
},
{
"epoch": 0.23784072688401925,
"grad_norm": 1.3184808872543174,
"learning_rate": 1.886753390422428e-05,
"loss": 1.3616,
"step": 890
},
{
"epoch": 0.23917691074291822,
"grad_norm": 1.2849347258583887,
"learning_rate": 1.8845872608354877e-05,
"loss": 1.3021,
"step": 895
},
{
"epoch": 0.24051309460181722,
"grad_norm": 1.2973872521270147,
"learning_rate": 1.882401878470052e-05,
"loss": 1.3365,
"step": 900
},
{
"epoch": 0.2418492784607162,
"grad_norm": 1.3250631489181373,
"learning_rate": 1.8801972908903162e-05,
"loss": 1.3433,
"step": 905
},
{
"epoch": 0.2431854623196152,
"grad_norm": 1.263181234454277,
"learning_rate": 1.877973546078474e-05,
"loss": 1.3703,
"step": 910
},
{
"epoch": 0.24452164617851416,
"grad_norm": 1.379416457226466,
"learning_rate": 1.875730692433669e-05,
"loss": 1.353,
"step": 915
},
{
"epoch": 0.24585783003741316,
"grad_norm": 1.293087775842146,
"learning_rate": 1.873468778770944e-05,
"loss": 1.3255,
"step": 920
},
{
"epoch": 0.24719401389631213,
"grad_norm": 1.3395508283607358,
"learning_rate": 1.8711878543201757e-05,
"loss": 1.2999,
"step": 925
},
{
"epoch": 0.24853019775521112,
"grad_norm": 1.368607821812935,
"learning_rate": 1.8688879687250067e-05,
"loss": 1.3305,
"step": 930
},
{
"epoch": 0.2498663816141101,
"grad_norm": 1.323555417532407,
"learning_rate": 1.8665691720417624e-05,
"loss": 1.3662,
"step": 935
},
{
"epoch": 0.25120256547300907,
"grad_norm": 1.367652347855113,
"learning_rate": 1.8642315147383628e-05,
"loss": 1.3469,
"step": 940
},
{
"epoch": 0.25253874933190806,
"grad_norm": 1.2966904825916514,
"learning_rate": 1.8618750476932237e-05,
"loss": 1.3071,
"step": 945
},
{
"epoch": 0.25387493319080706,
"grad_norm": 1.4450003246232057,
"learning_rate": 1.8594998221941482e-05,
"loss": 1.3721,
"step": 950
},
{
"epoch": 0.25521111704970606,
"grad_norm": 1.284360919841786,
"learning_rate": 1.857105889937213e-05,
"loss": 1.3638,
"step": 955
},
{
"epoch": 0.256547300908605,
"grad_norm": 1.288974034075832,
"learning_rate": 1.8546933030256417e-05,
"loss": 1.3349,
"step": 960
},
{
"epoch": 0.257883484767504,
"grad_norm": 1.3405220877515678,
"learning_rate": 1.85226211396867e-05,
"loss": 1.3052,
"step": 965
},
{
"epoch": 0.259219668626403,
"grad_norm": 1.3868830711147844,
"learning_rate": 1.8498123756804038e-05,
"loss": 1.3609,
"step": 970
},
{
"epoch": 0.260555852485302,
"grad_norm": 1.3022776914197354,
"learning_rate": 1.8473441414786692e-05,
"loss": 1.3385,
"step": 975
},
{
"epoch": 0.26189203634420094,
"grad_norm": 1.4058613068210792,
"learning_rate": 1.8448574650838477e-05,
"loss": 1.3234,
"step": 980
},
{
"epoch": 0.26322822020309994,
"grad_norm": 1.3008663043779487,
"learning_rate": 1.842352400617712e-05,
"loss": 1.3201,
"step": 985
},
{
"epoch": 0.26456440406199894,
"grad_norm": 1.274233066139097,
"learning_rate": 1.8398290026022444e-05,
"loss": 1.3372,
"step": 990
},
{
"epoch": 0.26590058792089794,
"grad_norm": 1.2895370237805952,
"learning_rate": 1.8372873259584517e-05,
"loss": 1.3082,
"step": 995
},
{
"epoch": 0.2672367717797969,
"grad_norm": 1.2250435736281928,
"learning_rate": 1.83472742600517e-05,
"loss": 1.3219,
"step": 1000
},
{
"epoch": 0.2672367717797969,
"eval_loss": 1.3463348150253296,
"eval_runtime": 523.6844,
"eval_samples_per_second": 25.294,
"eval_steps_per_second": 3.162,
"step": 1000
},
{
"epoch": 0.2685729556386959,
"grad_norm": 1.335279277755242,
"learning_rate": 1.83214935845786e-05,
"loss": 1.3195,
"step": 1005
},
{
"epoch": 0.2699091394975949,
"grad_norm": 1.3383554327189036,
"learning_rate": 1.8295531794273948e-05,
"loss": 1.3471,
"step": 1010
},
{
"epoch": 0.2712453233564939,
"grad_norm": 1.272300682321573,
"learning_rate": 1.826938945418837e-05,
"loss": 1.3156,
"step": 1015
},
{
"epoch": 0.2725815072153928,
"grad_norm": 1.4984409833622696,
"learning_rate": 1.8243067133302143e-05,
"loss": 1.3528,
"step": 1020
},
{
"epoch": 0.2739176910742918,
"grad_norm": 1.2475449536511878,
"learning_rate": 1.8216565404512732e-05,
"loss": 1.341,
"step": 1025
},
{
"epoch": 0.2752538749331908,
"grad_norm": 1.251734878893183,
"learning_rate": 1.818988484462238e-05,
"loss": 1.3106,
"step": 1030
},
{
"epoch": 0.2765900587920898,
"grad_norm": 1.322345839521324,
"learning_rate": 1.8163026034325532e-05,
"loss": 1.294,
"step": 1035
},
{
"epoch": 0.27792624265098875,
"grad_norm": 1.2838160402093648,
"learning_rate": 1.8135989558196207e-05,
"loss": 1.3484,
"step": 1040
},
{
"epoch": 0.27926242650988775,
"grad_norm": 1.2943916155535866,
"learning_rate": 1.8108776004675255e-05,
"loss": 1.3052,
"step": 1045
},
{
"epoch": 0.28059861036878675,
"grad_norm": 1.2762154467420836,
"learning_rate": 1.808138596605758e-05,
"loss": 1.3284,
"step": 1050
},
{
"epoch": 0.28193479422768575,
"grad_norm": 1.255384319187328,
"learning_rate": 1.8053820038479214e-05,
"loss": 1.3686,
"step": 1055
},
{
"epoch": 0.2832709780865847,
"grad_norm": 1.2673273813526882,
"learning_rate": 1.802607882190437e-05,
"loss": 1.3382,
"step": 1060
},
{
"epoch": 0.2846071619454837,
"grad_norm": 1.2940248666834164,
"learning_rate": 1.799816292011237e-05,
"loss": 1.3498,
"step": 1065
},
{
"epoch": 0.2859433458043827,
"grad_norm": 1.2046683889500032,
"learning_rate": 1.7970072940684514e-05,
"loss": 1.3329,
"step": 1070
},
{
"epoch": 0.2872795296632817,
"grad_norm": 1.366921247819638,
"learning_rate": 1.7941809494990838e-05,
"loss": 1.3152,
"step": 1075
},
{
"epoch": 0.28861571352218063,
"grad_norm": 1.2494608185941236,
"learning_rate": 1.7913373198176832e-05,
"loss": 1.3242,
"step": 1080
},
{
"epoch": 0.2899518973810796,
"grad_norm": 1.2279808317447078,
"learning_rate": 1.7884764669150035e-05,
"loss": 1.3359,
"step": 1085
},
{
"epoch": 0.2912880812399786,
"grad_norm": 1.2317749726693008,
"learning_rate": 1.7855984530566564e-05,
"loss": 1.3186,
"step": 1090
},
{
"epoch": 0.2926242650988776,
"grad_norm": 1.3570393319531886,
"learning_rate": 1.7827033408817573e-05,
"loss": 1.3449,
"step": 1095
},
{
"epoch": 0.29396044895777657,
"grad_norm": 1.247405852772324,
"learning_rate": 1.779791193401561e-05,
"loss": 1.3416,
"step": 1100
},
{
"epoch": 0.29529663281667556,
"grad_norm": 1.2893644804363795,
"learning_rate": 1.776862073998091e-05,
"loss": 1.3674,
"step": 1105
},
{
"epoch": 0.29663281667557456,
"grad_norm": 1.2741258542191471,
"learning_rate": 1.7739160464227593e-05,
"loss": 1.3291,
"step": 1110
},
{
"epoch": 0.29796900053447356,
"grad_norm": 1.4051921967322472,
"learning_rate": 1.7709531747949796e-05,
"loss": 1.3592,
"step": 1115
},
{
"epoch": 0.2993051843933725,
"grad_norm": 1.2562313322258833,
"learning_rate": 1.7679735236007715e-05,
"loss": 1.3259,
"step": 1120
},
{
"epoch": 0.3006413682522715,
"grad_norm": 1.316830880638165,
"learning_rate": 1.7649771576913553e-05,
"loss": 1.3448,
"step": 1125
},
{
"epoch": 0.3019775521111705,
"grad_norm": 1.3518115028546631,
"learning_rate": 1.7619641422817446e-05,
"loss": 1.3291,
"step": 1130
},
{
"epoch": 0.3033137359700695,
"grad_norm": 1.3279266806644043,
"learning_rate": 1.758934542949323e-05,
"loss": 1.3589,
"step": 1135
},
{
"epoch": 0.30464991982896844,
"grad_norm": 1.2655330414005816,
"learning_rate": 1.755888425632418e-05,
"loss": 1.3267,
"step": 1140
},
{
"epoch": 0.30598610368786744,
"grad_norm": 1.2109654835513972,
"learning_rate": 1.7528258566288666e-05,
"loss": 1.3264,
"step": 1145
},
{
"epoch": 0.30732228754676644,
"grad_norm": 1.269808597072721,
"learning_rate": 1.7497469025945722e-05,
"loss": 1.2766,
"step": 1150
},
{
"epoch": 0.30865847140566544,
"grad_norm": 1.2456464717450948,
"learning_rate": 1.7466516305420524e-05,
"loss": 1.3352,
"step": 1155
},
{
"epoch": 0.3099946552645644,
"grad_norm": 1.2359484363944293,
"learning_rate": 1.743540107838983e-05,
"loss": 1.3295,
"step": 1160
},
{
"epoch": 0.3113308391234634,
"grad_norm": 1.2244387854880765,
"learning_rate": 1.74041240220673e-05,
"loss": 1.2754,
"step": 1165
},
{
"epoch": 0.3126670229823624,
"grad_norm": 1.34164790224033,
"learning_rate": 1.7372685817188747e-05,
"loss": 1.3066,
"step": 1170
},
{
"epoch": 0.3140032068412614,
"grad_norm": 1.274512676357063,
"learning_rate": 1.734108714799735e-05,
"loss": 1.337,
"step": 1175
},
{
"epoch": 0.3153393907001603,
"grad_norm": 1.1877061220307676,
"learning_rate": 1.7309328702228742e-05,
"loss": 1.304,
"step": 1180
},
{
"epoch": 0.3166755745590593,
"grad_norm": 1.307991502348843,
"learning_rate": 1.7277411171096042e-05,
"loss": 1.3234,
"step": 1185
},
{
"epoch": 0.3180117584179583,
"grad_norm": 1.2459886266896583,
"learning_rate": 1.7245335249274818e-05,
"loss": 1.344,
"step": 1190
},
{
"epoch": 0.3193479422768573,
"grad_norm": 1.3088574044809078,
"learning_rate": 1.7213101634887968e-05,
"loss": 1.3213,
"step": 1195
},
{
"epoch": 0.32068412613575625,
"grad_norm": 1.401051770726762,
"learning_rate": 1.718071102949051e-05,
"loss": 1.3355,
"step": 1200
},
{
"epoch": 0.32068412613575625,
"eval_loss": 1.3390916585922241,
"eval_runtime": 523.6061,
"eval_samples_per_second": 25.298,
"eval_steps_per_second": 3.163,
"step": 1200
},
{
"epoch": 0.32202030999465525,
"grad_norm": 1.2972562697211982,
"learning_rate": 1.7148164138054333e-05,
"loss": 1.3181,
"step": 1205
},
{
"epoch": 0.32335649385355425,
"grad_norm": 1.2825002290000433,
"learning_rate": 1.7115461668952848e-05,
"loss": 1.3422,
"step": 1210
},
{
"epoch": 0.32469267771245325,
"grad_norm": 1.2161188921788924,
"learning_rate": 1.7082604333945557e-05,
"loss": 1.3505,
"step": 1215
},
{
"epoch": 0.3260288615713522,
"grad_norm": 1.2075785701291617,
"learning_rate": 1.7049592848162583e-05,
"loss": 1.331,
"step": 1220
},
{
"epoch": 0.3273650454302512,
"grad_norm": 1.3030995176122255,
"learning_rate": 1.701642793008909e-05,
"loss": 1.3266,
"step": 1225
},
{
"epoch": 0.3287012292891502,
"grad_norm": 1.2590894105748518,
"learning_rate": 1.6983110301549652e-05,
"loss": 1.3199,
"step": 1230
},
{
"epoch": 0.3300374131480492,
"grad_norm": 1.2284782881775649,
"learning_rate": 1.6949640687692535e-05,
"loss": 1.3243,
"step": 1235
},
{
"epoch": 0.33137359700694813,
"grad_norm": 1.2623721858490347,
"learning_rate": 1.691601981697393e-05,
"loss": 1.3299,
"step": 1240
},
{
"epoch": 0.3327097808658471,
"grad_norm": 1.226343404396135,
"learning_rate": 1.688224842114208e-05,
"loss": 1.3031,
"step": 1245
},
{
"epoch": 0.3340459647247461,
"grad_norm": 1.2453664354622114,
"learning_rate": 1.6848327235221368e-05,
"loss": 1.3047,
"step": 1250
},
{
"epoch": 0.3353821485836451,
"grad_norm": 1.2917212982794357,
"learning_rate": 1.681425699749631e-05,
"loss": 1.3367,
"step": 1255
},
{
"epoch": 0.3367183324425441,
"grad_norm": 1.1431548737930348,
"learning_rate": 1.6780038449495492e-05,
"loss": 1.2894,
"step": 1260
},
{
"epoch": 0.33805451630144306,
"grad_norm": 1.305490289585266,
"learning_rate": 1.674567233597542e-05,
"loss": 1.3303,
"step": 1265
},
{
"epoch": 0.33939070016034206,
"grad_norm": 1.3295423667445367,
"learning_rate": 1.6711159404904346e-05,
"loss": 1.3795,
"step": 1270
},
{
"epoch": 0.34072688401924106,
"grad_norm": 1.2557371261898338,
"learning_rate": 1.667650040744593e-05,
"loss": 1.3923,
"step": 1275
},
{
"epoch": 0.34206306787814006,
"grad_norm": 1.2709385762973946,
"learning_rate": 1.6641696097942937e-05,
"loss": 1.3633,
"step": 1280
},
{
"epoch": 0.343399251737039,
"grad_norm": 1.2297417601415581,
"learning_rate": 1.6606747233900816e-05,
"loss": 1.3475,
"step": 1285
},
{
"epoch": 0.344735435595938,
"grad_norm": 1.270772993810448,
"learning_rate": 1.6571654575971186e-05,
"loss": 1.3215,
"step": 1290
},
{
"epoch": 0.346071619454837,
"grad_norm": 1.243136160915021,
"learning_rate": 1.6536418887935307e-05,
"loss": 1.2909,
"step": 1295
},
{
"epoch": 0.347407803313736,
"grad_norm": 1.2453939190767478,
"learning_rate": 1.6501040936687444e-05,
"loss": 1.299,
"step": 1300
},
{
"epoch": 0.34874398717263494,
"grad_norm": 1.3190132616463197,
"learning_rate": 1.6465521492218175e-05,
"loss": 1.3242,
"step": 1305
},
{
"epoch": 0.35008017103153394,
"grad_norm": 1.2836129069270563,
"learning_rate": 1.6429861327597643e-05,
"loss": 1.3257,
"step": 1310
},
{
"epoch": 0.35141635489043294,
"grad_norm": 1.3046224543331126,
"learning_rate": 1.6394061218958714e-05,
"loss": 1.3735,
"step": 1315
},
{
"epoch": 0.35275253874933193,
"grad_norm": 1.195969330334029,
"learning_rate": 1.63581219454801e-05,
"loss": 1.3547,
"step": 1320
},
{
"epoch": 0.3540887226082309,
"grad_norm": 1.2293037135379614,
"learning_rate": 1.63220442893694e-05,
"loss": 1.3258,
"step": 1325
},
{
"epoch": 0.3554249064671299,
"grad_norm": 1.221432152844297,
"learning_rate": 1.6285829035846057e-05,
"loss": 1.3368,
"step": 1330
},
{
"epoch": 0.3567610903260289,
"grad_norm": 1.2802671063330058,
"learning_rate": 1.624947697312429e-05,
"loss": 1.3497,
"step": 1335
},
{
"epoch": 0.35809727418492787,
"grad_norm": 1.408250486602362,
"learning_rate": 1.621298889239592e-05,
"loss": 1.3856,
"step": 1340
},
{
"epoch": 0.3594334580438268,
"grad_norm": 1.2601768753005949,
"learning_rate": 1.617636558781318e-05,
"loss": 1.3151,
"step": 1345
},
{
"epoch": 0.3607696419027258,
"grad_norm": 1.2756859712312687,
"learning_rate": 1.6139607856471377e-05,
"loss": 1.2932,
"step": 1350
},
{
"epoch": 0.3621058257616248,
"grad_norm": 1.2546446591863418,
"learning_rate": 1.610271649839161e-05,
"loss": 1.3112,
"step": 1355
},
{
"epoch": 0.3634420096205238,
"grad_norm": 1.2883557482491337,
"learning_rate": 1.6065692316503306e-05,
"loss": 1.3155,
"step": 1360
},
{
"epoch": 0.36477819347942275,
"grad_norm": 1.2350958249428101,
"learning_rate": 1.6028536116626763e-05,
"loss": 1.2951,
"step": 1365
},
{
"epoch": 0.36611437733832175,
"grad_norm": 1.209684203322871,
"learning_rate": 1.5991248707455614e-05,
"loss": 1.3159,
"step": 1370
},
{
"epoch": 0.36745056119722075,
"grad_norm": 1.3434534905824467,
"learning_rate": 1.595383090053923e-05,
"loss": 1.3079,
"step": 1375
},
{
"epoch": 0.36878674505611975,
"grad_norm": 1.2129683097633361,
"learning_rate": 1.5916283510265037e-05,
"loss": 1.2949,
"step": 1380
},
{
"epoch": 0.3701229289150187,
"grad_norm": 1.3085166024177262,
"learning_rate": 1.5878607353840814e-05,
"loss": 1.3557,
"step": 1385
},
{
"epoch": 0.3714591127739177,
"grad_norm": 1.2946759032750124,
"learning_rate": 1.5840803251276892e-05,
"loss": 1.3005,
"step": 1390
},
{
"epoch": 0.3727952966328167,
"grad_norm": 1.283448494880763,
"learning_rate": 1.5802872025368316e-05,
"loss": 1.2877,
"step": 1395
},
{
"epoch": 0.3741314804917157,
"grad_norm": 1.2309029767719335,
"learning_rate": 1.576481450167693e-05,
"loss": 1.334,
"step": 1400
},
{
"epoch": 0.3741314804917157,
"eval_loss": 1.330536127090454,
"eval_runtime": 523.7724,
"eval_samples_per_second": 25.29,
"eval_steps_per_second": 3.162,
"step": 1400
},
{
"epoch": 0.3754676643506146,
"grad_norm": 1.2496837334635877,
"learning_rate": 1.5726631508513412e-05,
"loss": 1.3089,
"step": 1405
},
{
"epoch": 0.3768038482095136,
"grad_norm": 1.374132042278566,
"learning_rate": 1.568832387691924e-05,
"loss": 1.3465,
"step": 1410
},
{
"epoch": 0.3781400320684126,
"grad_norm": 1.2412100387519944,
"learning_rate": 1.5649892440648625e-05,
"loss": 1.338,
"step": 1415
},
{
"epoch": 0.3794762159273116,
"grad_norm": 1.2515878984452657,
"learning_rate": 1.5611338036150338e-05,
"loss": 1.302,
"step": 1420
},
{
"epoch": 0.38081239978621056,
"grad_norm": 1.3524756121075554,
"learning_rate": 1.5572661502549514e-05,
"loss": 1.3297,
"step": 1425
},
{
"epoch": 0.38214858364510956,
"grad_norm": 1.3613331896664973,
"learning_rate": 1.5533863681629404e-05,
"loss": 1.3375,
"step": 1430
},
{
"epoch": 0.38348476750400856,
"grad_norm": 1.3234108123677346,
"learning_rate": 1.5494945417813034e-05,
"loss": 1.28,
"step": 1435
},
{
"epoch": 0.38482095136290756,
"grad_norm": 1.2464116209347194,
"learning_rate": 1.545590755814483e-05,
"loss": 1.339,
"step": 1440
},
{
"epoch": 0.3861571352218065,
"grad_norm": 1.2082007561781767,
"learning_rate": 1.5416750952272198e-05,
"loss": 1.2924,
"step": 1445
},
{
"epoch": 0.3874933190807055,
"grad_norm": 1.2460568233344762,
"learning_rate": 1.537747645242701e-05,
"loss": 1.2913,
"step": 1450
},
{
"epoch": 0.3888295029396045,
"grad_norm": 1.2643768624756646,
"learning_rate": 1.5338084913407067e-05,
"loss": 1.3385,
"step": 1455
},
{
"epoch": 0.3901656867985035,
"grad_norm": 1.2393822310314082,
"learning_rate": 1.5298577192557487e-05,
"loss": 1.2918,
"step": 1460
},
{
"epoch": 0.39150187065740244,
"grad_norm": 1.2212527378031937,
"learning_rate": 1.525895414975207e-05,
"loss": 1.3496,
"step": 1465
},
{
"epoch": 0.39283805451630144,
"grad_norm": 1.2539579056942822,
"learning_rate": 1.5219216647374546e-05,
"loss": 1.3285,
"step": 1470
},
{
"epoch": 0.39417423837520044,
"grad_norm": 1.194899743310905,
"learning_rate": 1.5179365550299823e-05,
"loss": 1.2987,
"step": 1475
},
{
"epoch": 0.39551042223409943,
"grad_norm": 1.2361033901459717,
"learning_rate": 1.513940172587518e-05,
"loss": 1.3018,
"step": 1480
},
{
"epoch": 0.3968466060929984,
"grad_norm": 1.2040496820635649,
"learning_rate": 1.5099326043901361e-05,
"loss": 1.3144,
"step": 1485
},
{
"epoch": 0.3981827899518974,
"grad_norm": 1.2436110116124757,
"learning_rate": 1.5059139376613652e-05,
"loss": 1.3147,
"step": 1490
},
{
"epoch": 0.3995189738107964,
"grad_norm": 1.229222024704698,
"learning_rate": 1.5018842598662913e-05,
"loss": 1.3084,
"step": 1495
},
{
"epoch": 0.40085515766969537,
"grad_norm": 1.2280023763221926,
"learning_rate": 1.4978436587096526e-05,
"loss": 1.2993,
"step": 1500
},
{
"epoch": 0.4021913415285943,
"grad_norm": 1.2379509698653712,
"learning_rate": 1.4937922221339303e-05,
"loss": 1.3448,
"step": 1505
},
{
"epoch": 0.4035275253874933,
"grad_norm": 1.1819300792108958,
"learning_rate": 1.4897300383174362e-05,
"loss": 1.3093,
"step": 1510
},
{
"epoch": 0.4048637092463923,
"grad_norm": 1.2646351619539313,
"learning_rate": 1.4856571956723924e-05,
"loss": 1.3292,
"step": 1515
},
{
"epoch": 0.4061998931052913,
"grad_norm": 1.20105031681724,
"learning_rate": 1.4815737828430068e-05,
"loss": 1.3275,
"step": 1520
},
{
"epoch": 0.40753607696419025,
"grad_norm": 1.2322477247241055,
"learning_rate": 1.4774798887035446e-05,
"loss": 1.3273,
"step": 1525
},
{
"epoch": 0.40887226082308925,
"grad_norm": 1.2203562838881301,
"learning_rate": 1.4733756023563932e-05,
"loss": 1.3046,
"step": 1530
},
{
"epoch": 0.41020844468198825,
"grad_norm": 1.3701923681759842,
"learning_rate": 1.4692610131301242e-05,
"loss": 1.3468,
"step": 1535
},
{
"epoch": 0.41154462854088725,
"grad_norm": 1.2104295209543379,
"learning_rate": 1.4651362105775471e-05,
"loss": 1.3292,
"step": 1540
},
{
"epoch": 0.4128808123997862,
"grad_norm": 1.178262832900831,
"learning_rate": 1.4610012844737622e-05,
"loss": 1.3095,
"step": 1545
},
{
"epoch": 0.4142169962586852,
"grad_norm": 1.2034684530693827,
"learning_rate": 1.4568563248142058e-05,
"loss": 1.277,
"step": 1550
},
{
"epoch": 0.4155531801175842,
"grad_norm": 1.2346570794875227,
"learning_rate": 1.4527014218126913e-05,
"loss": 1.3323,
"step": 1555
},
{
"epoch": 0.4168893639764832,
"grad_norm": 1.2924089782717,
"learning_rate": 1.4485366658994463e-05,
"loss": 1.3117,
"step": 1560
},
{
"epoch": 0.4182255478353821,
"grad_norm": 1.222906017451485,
"learning_rate": 1.4443621477191434e-05,
"loss": 1.3185,
"step": 1565
},
{
"epoch": 0.4195617316942811,
"grad_norm": 1.2534027609567953,
"learning_rate": 1.440177958128929e-05,
"loss": 1.2978,
"step": 1570
},
{
"epoch": 0.4208979155531801,
"grad_norm": 1.2140810113206562,
"learning_rate": 1.4359841881964445e-05,
"loss": 1.3162,
"step": 1575
},
{
"epoch": 0.4222340994120791,
"grad_norm": 1.1684011954978797,
"learning_rate": 1.4317809291978442e-05,
"loss": 1.3333,
"step": 1580
},
{
"epoch": 0.42357028327097807,
"grad_norm": 1.214319379461352,
"learning_rate": 1.4275682726158092e-05,
"loss": 1.2867,
"step": 1585
},
{
"epoch": 0.42490646712987706,
"grad_norm": 1.182224144852539,
"learning_rate": 1.4233463101375569e-05,
"loss": 1.3039,
"step": 1590
},
{
"epoch": 0.42624265098877606,
"grad_norm": 1.2314041097519686,
"learning_rate": 1.4191151336528441e-05,
"loss": 1.3488,
"step": 1595
},
{
"epoch": 0.42757883484767506,
"grad_norm": 1.2641228680933205,
"learning_rate": 1.4148748352519677e-05,
"loss": 1.3183,
"step": 1600
},
{
"epoch": 0.42757883484767506,
"eval_loss": 1.3232654333114624,
"eval_runtime": 523.7168,
"eval_samples_per_second": 25.292,
"eval_steps_per_second": 3.162,
"step": 1600
},
{
"epoch": 0.428915018706574,
"grad_norm": 1.2410818763020894,
"learning_rate": 1.4106255072237605e-05,
"loss": 1.2977,
"step": 1605
},
{
"epoch": 0.430251202565473,
"grad_norm": 1.2216210579429683,
"learning_rate": 1.406367242053583e-05,
"loss": 1.3388,
"step": 1610
},
{
"epoch": 0.431587386424372,
"grad_norm": 1.2380524456449333,
"learning_rate": 1.402100132421309e-05,
"loss": 1.3048,
"step": 1615
},
{
"epoch": 0.432923570283271,
"grad_norm": 1.222444425938849,
"learning_rate": 1.39782427119931e-05,
"loss": 1.3129,
"step": 1620
},
{
"epoch": 0.43425975414216994,
"grad_norm": 1.2084516798428055,
"learning_rate": 1.3935397514504332e-05,
"loss": 1.3312,
"step": 1625
},
{
"epoch": 0.43559593800106894,
"grad_norm": 1.3138679343999906,
"learning_rate": 1.3892466664259756e-05,
"loss": 1.3358,
"step": 1630
},
{
"epoch": 0.43693212185996794,
"grad_norm": 1.325108022318488,
"learning_rate": 1.3849451095636555e-05,
"loss": 1.3197,
"step": 1635
},
{
"epoch": 0.43826830571886694,
"grad_norm": 1.239099901058549,
"learning_rate": 1.3806351744855781e-05,
"loss": 1.3276,
"step": 1640
},
{
"epoch": 0.4396044895777659,
"grad_norm": 1.2701728016347087,
"learning_rate": 1.3763169549961976e-05,
"loss": 1.3087,
"step": 1645
},
{
"epoch": 0.4409406734366649,
"grad_norm": 1.3093633737758703,
"learning_rate": 1.371990545080276e-05,
"loss": 1.3334,
"step": 1650
},
{
"epoch": 0.4422768572955639,
"grad_norm": 1.2721317308394575,
"learning_rate": 1.3676560389008378e-05,
"loss": 1.3568,
"step": 1655
},
{
"epoch": 0.4436130411544629,
"grad_norm": 1.2184083694463035,
"learning_rate": 1.3633135307971204e-05,
"loss": 1.3279,
"step": 1660
},
{
"epoch": 0.4449492250133618,
"grad_norm": 1.3001651314508893,
"learning_rate": 1.3589631152825197e-05,
"loss": 1.3176,
"step": 1665
},
{
"epoch": 0.4462854088722608,
"grad_norm": 1.2226816442485073,
"learning_rate": 1.3546048870425356e-05,
"loss": 1.2705,
"step": 1670
},
{
"epoch": 0.4476215927311598,
"grad_norm": 1.4176804415374429,
"learning_rate": 1.3502389409327087e-05,
"loss": 1.2679,
"step": 1675
},
{
"epoch": 0.4489577765900588,
"grad_norm": 1.2388040805053258,
"learning_rate": 1.3458653719765564e-05,
"loss": 1.3046,
"step": 1680
},
{
"epoch": 0.45029396044895775,
"grad_norm": 1.2129122448318554,
"learning_rate": 1.341484275363506e-05,
"loss": 1.3149,
"step": 1685
},
{
"epoch": 0.45163014430785675,
"grad_norm": 1.244763644565545,
"learning_rate": 1.3370957464468213e-05,
"loss": 1.3436,
"step": 1690
},
{
"epoch": 0.45296632816675575,
"grad_norm": 1.2759739357160127,
"learning_rate": 1.332699880741528e-05,
"loss": 1.3569,
"step": 1695
},
{
"epoch": 0.45430251202565475,
"grad_norm": 1.2281614718249263,
"learning_rate": 1.3282967739223357e-05,
"loss": 1.3075,
"step": 1700
},
{
"epoch": 0.4556386958845537,
"grad_norm": 1.2189899125191879,
"learning_rate": 1.3238865218215535e-05,
"loss": 1.3017,
"step": 1705
},
{
"epoch": 0.4569748797434527,
"grad_norm": 1.231210053539351,
"learning_rate": 1.3194692204270063e-05,
"loss": 1.3366,
"step": 1710
},
{
"epoch": 0.4583110636023517,
"grad_norm": 1.2018193616808162,
"learning_rate": 1.3150449658799442e-05,
"loss": 1.3535,
"step": 1715
},
{
"epoch": 0.4596472474612507,
"grad_norm": 1.2128311963623164,
"learning_rate": 1.3106138544729511e-05,
"loss": 1.3495,
"step": 1720
},
{
"epoch": 0.46098343132014963,
"grad_norm": 1.2304855721971648,
"learning_rate": 1.3061759826478477e-05,
"loss": 1.308,
"step": 1725
},
{
"epoch": 0.4623196151790486,
"grad_norm": 1.183612891624708,
"learning_rate": 1.3017314469935942e-05,
"loss": 1.2771,
"step": 1730
},
{
"epoch": 0.4636557990379476,
"grad_norm": 1.2035215718306922,
"learning_rate": 1.2972803442441863e-05,
"loss": 1.3374,
"step": 1735
},
{
"epoch": 0.4649919828968466,
"grad_norm": 1.2078408808291714,
"learning_rate": 1.2928227712765504e-05,
"loss": 1.3155,
"step": 1740
},
{
"epoch": 0.46632816675574557,
"grad_norm": 1.2685296349837316,
"learning_rate": 1.2883588251084362e-05,
"loss": 1.3224,
"step": 1745
},
{
"epoch": 0.46766435061464456,
"grad_norm": 1.2298326258192636,
"learning_rate": 1.2838886028963038e-05,
"loss": 1.3332,
"step": 1750
},
{
"epoch": 0.46900053447354356,
"grad_norm": 1.1803857708163348,
"learning_rate": 1.2794122019332087e-05,
"loss": 1.2889,
"step": 1755
},
{
"epoch": 0.47033671833244256,
"grad_norm": 1.274081083989054,
"learning_rate": 1.2749297196466861e-05,
"loss": 1.289,
"step": 1760
},
{
"epoch": 0.4716729021913415,
"grad_norm": 1.3238957747747422,
"learning_rate": 1.270441253596629e-05,
"loss": 1.3189,
"step": 1765
},
{
"epoch": 0.4730090860502405,
"grad_norm": 1.2097611285289955,
"learning_rate": 1.265946901473166e-05,
"loss": 1.3071,
"step": 1770
},
{
"epoch": 0.4743452699091395,
"grad_norm": 1.1625021555183794,
"learning_rate": 1.2614467610945323e-05,
"loss": 1.2987,
"step": 1775
},
{
"epoch": 0.4756814537680385,
"grad_norm": 1.1887076661153424,
"learning_rate": 1.256940930404945e-05,
"loss": 1.2783,
"step": 1780
},
{
"epoch": 0.47701763762693744,
"grad_norm": 1.2645632943006766,
"learning_rate": 1.2524295074724683e-05,
"loss": 1.3448,
"step": 1785
},
{
"epoch": 0.47835382148583644,
"grad_norm": 1.2361406321440211,
"learning_rate": 1.2479125904868795e-05,
"loss": 1.2726,
"step": 1790
},
{
"epoch": 0.47969000534473544,
"grad_norm": 1.274165207428913,
"learning_rate": 1.2433902777575326e-05,
"loss": 1.2907,
"step": 1795
},
{
"epoch": 0.48102618920363444,
"grad_norm": 1.2418881744061736,
"learning_rate": 1.2388626677112185e-05,
"loss": 1.334,
"step": 1800
},
{
"epoch": 0.48102618920363444,
"eval_loss": 1.316095232963562,
"eval_runtime": 523.7124,
"eval_samples_per_second": 25.293,
"eval_steps_per_second": 3.162,
"step": 1800
},
{
"epoch": 0.4823623730625334,
"grad_norm": 1.4406988030285752,
"learning_rate": 1.2343298588900226e-05,
"loss": 1.3175,
"step": 1805
},
{
"epoch": 0.4836985569214324,
"grad_norm": 1.1708167122953599,
"learning_rate": 1.2297919499491797e-05,
"loss": 1.3048,
"step": 1810
},
{
"epoch": 0.4850347407803314,
"grad_norm": 1.2437310171766487,
"learning_rate": 1.2252490396549282e-05,
"loss": 1.3287,
"step": 1815
},
{
"epoch": 0.4863709246392304,
"grad_norm": 1.5815332852740502,
"learning_rate": 1.220701226882358e-05,
"loss": 1.3344,
"step": 1820
},
{
"epoch": 0.4877071084981293,
"grad_norm": 1.2057093168260966,
"learning_rate": 1.2161486106132612e-05,
"loss": 1.3623,
"step": 1825
},
{
"epoch": 0.4890432923570283,
"grad_norm": 1.1494411879995097,
"learning_rate": 1.2115912899339757e-05,
"loss": 1.283,
"step": 1830
},
{
"epoch": 0.4903794762159273,
"grad_norm": 1.1721905423708292,
"learning_rate": 1.2070293640332306e-05,
"loss": 1.3124,
"step": 1835
},
{
"epoch": 0.4917156600748263,
"grad_norm": 1.165222069970401,
"learning_rate": 1.202462932199985e-05,
"loss": 1.3094,
"step": 1840
},
{
"epoch": 0.49305184393372525,
"grad_norm": 1.250148064441355,
"learning_rate": 1.1978920938212691e-05,
"loss": 1.3363,
"step": 1845
},
{
"epoch": 0.49438802779262425,
"grad_norm": 1.1969490785809054,
"learning_rate": 1.1933169483800203e-05,
"loss": 1.3057,
"step": 1850
},
{
"epoch": 0.49572421165152325,
"grad_norm": 1.2469012012408995,
"learning_rate": 1.1887375954529167e-05,
"loss": 1.3053,
"step": 1855
},
{
"epoch": 0.49706039551042225,
"grad_norm": 1.1816859874185814,
"learning_rate": 1.1841541347082134e-05,
"loss": 1.2982,
"step": 1860
},
{
"epoch": 0.4983965793693212,
"grad_norm": 1.258761404972035,
"learning_rate": 1.1795666659035682e-05,
"loss": 1.2963,
"step": 1865
},
{
"epoch": 0.4997327632282202,
"grad_norm": 1.2450588185319733,
"learning_rate": 1.1749752888838754e-05,
"loss": 1.3272,
"step": 1870
},
{
"epoch": 0.5010689470871192,
"grad_norm": 1.2052446498628289,
"learning_rate": 1.1703801035790897e-05,
"loss": 1.2946,
"step": 1875
},
{
"epoch": 0.5024051309460181,
"grad_norm": 1.2142678023982578,
"learning_rate": 1.1657812100020507e-05,
"loss": 1.302,
"step": 1880
},
{
"epoch": 0.5037413148049171,
"grad_norm": 1.2721562041773744,
"learning_rate": 1.1611787082463095e-05,
"loss": 1.3236,
"step": 1885
},
{
"epoch": 0.5050774986638161,
"grad_norm": 1.3019577911767723,
"learning_rate": 1.1565726984839471e-05,
"loss": 1.3,
"step": 1890
},
{
"epoch": 0.5064136825227151,
"grad_norm": 1.2326550134057832,
"learning_rate": 1.1519632809633954e-05,
"loss": 1.3449,
"step": 1895
},
{
"epoch": 0.5077498663816141,
"grad_norm": 1.2108121941048178,
"learning_rate": 1.1473505560072543e-05,
"loss": 1.3003,
"step": 1900
},
{
"epoch": 0.5090860502405131,
"grad_norm": 1.2351796577016483,
"learning_rate": 1.1427346240101105e-05,
"loss": 1.294,
"step": 1905
},
{
"epoch": 0.5104222340994121,
"grad_norm": 1.210376427103751,
"learning_rate": 1.1381155854363503e-05,
"loss": 1.269,
"step": 1910
},
{
"epoch": 0.5117584179583111,
"grad_norm": 1.1902823799298627,
"learning_rate": 1.1334935408179736e-05,
"loss": 1.2871,
"step": 1915
},
{
"epoch": 0.51309460181721,
"grad_norm": 1.2389975794814998,
"learning_rate": 1.1288685907524057e-05,
"loss": 1.3092,
"step": 1920
},
{
"epoch": 0.514430785676109,
"grad_norm": 1.228008698129948,
"learning_rate": 1.1242408359003091e-05,
"loss": 1.287,
"step": 1925
},
{
"epoch": 0.515766969535008,
"grad_norm": 1.1816697296808294,
"learning_rate": 1.1196103769833908e-05,
"loss": 1.3128,
"step": 1930
},
{
"epoch": 0.517103153393907,
"grad_norm": 1.1835653007689506,
"learning_rate": 1.1149773147822112e-05,
"loss": 1.2925,
"step": 1935
},
{
"epoch": 0.518439337252806,
"grad_norm": 1.2380413372980732,
"learning_rate": 1.1103417501339903e-05,
"loss": 1.3405,
"step": 1940
},
{
"epoch": 0.519775521111705,
"grad_norm": 1.2784541489247507,
"learning_rate": 1.1057037839304135e-05,
"loss": 1.2585,
"step": 1945
},
{
"epoch": 0.521111704970604,
"grad_norm": 1.2402864740991135,
"learning_rate": 1.1010635171154342e-05,
"loss": 1.2867,
"step": 1950
},
{
"epoch": 0.522447888829503,
"grad_norm": 1.2346852083660054,
"learning_rate": 1.0964210506830802e-05,
"loss": 1.2781,
"step": 1955
},
{
"epoch": 0.5237840726884019,
"grad_norm": 1.2195618023918173,
"learning_rate": 1.0917764856752509e-05,
"loss": 1.289,
"step": 1960
},
{
"epoch": 0.5251202565473009,
"grad_norm": 1.1817447386492328,
"learning_rate": 1.0871299231795214e-05,
"loss": 1.2883,
"step": 1965
},
{
"epoch": 0.5264564404061999,
"grad_norm": 1.1727533453851813,
"learning_rate": 1.082481464326942e-05,
"loss": 1.32,
"step": 1970
},
{
"epoch": 0.5277926242650989,
"grad_norm": 1.181864666573128,
"learning_rate": 1.077831210289837e-05,
"loss": 1.3068,
"step": 1975
},
{
"epoch": 0.5291288081239979,
"grad_norm": 1.2249148804887326,
"learning_rate": 1.0731792622796007e-05,
"loss": 1.3044,
"step": 1980
},
{
"epoch": 0.5304649919828969,
"grad_norm": 1.2263040268107137,
"learning_rate": 1.0685257215444975e-05,
"loss": 1.3002,
"step": 1985
},
{
"epoch": 0.5318011758417959,
"grad_norm": 1.2237762133974603,
"learning_rate": 1.0638706893674571e-05,
"loss": 1.3138,
"step": 1990
},
{
"epoch": 0.5331373597006949,
"grad_norm": 1.324928279536264,
"learning_rate": 1.059214267063869e-05,
"loss": 1.3364,
"step": 1995
},
{
"epoch": 0.5344735435595938,
"grad_norm": 1.2493747053579303,
"learning_rate": 1.0545565559793796e-05,
"loss": 1.3013,
"step": 2000
},
{
"epoch": 0.5344735435595938,
"eval_loss": 1.3086645603179932,
"eval_runtime": 523.908,
"eval_samples_per_second": 25.283,
"eval_steps_per_second": 3.161,
"step": 2000
},
{
"epoch": 0.5358097274184928,
"grad_norm": 1.2142470124401028,
"learning_rate": 1.0498976574876838e-05,
"loss": 1.2563,
"step": 2005
},
{
"epoch": 0.5371459112773918,
"grad_norm": 1.1948400261490024,
"learning_rate": 1.0452376729883216e-05,
"loss": 1.3066,
"step": 2010
},
{
"epoch": 0.5384820951362908,
"grad_norm": 1.206399971819588,
"learning_rate": 1.0405767039044688e-05,
"loss": 1.2928,
"step": 2015
},
{
"epoch": 0.5398182789951897,
"grad_norm": 1.2064545160074232,
"learning_rate": 1.0359148516807302e-05,
"loss": 1.3377,
"step": 2020
},
{
"epoch": 0.5411544628540887,
"grad_norm": 1.2352197196893993,
"learning_rate": 1.0312522177809326e-05,
"loss": 1.32,
"step": 2025
},
{
"epoch": 0.5424906467129877,
"grad_norm": 1.1837145257724762,
"learning_rate": 1.0265889036859146e-05,
"loss": 1.2636,
"step": 2030
},
{
"epoch": 0.5438268305718867,
"grad_norm": 1.2342337040554485,
"learning_rate": 1.0219250108913206e-05,
"loss": 1.2928,
"step": 2035
},
{
"epoch": 0.5451630144307856,
"grad_norm": 1.2371903390877288,
"learning_rate": 1.0172606409053887e-05,
"loss": 1.3451,
"step": 2040
},
{
"epoch": 0.5464991982896846,
"grad_norm": 1.2027906126810077,
"learning_rate": 1.0125958952467439e-05,
"loss": 1.3274,
"step": 2045
},
{
"epoch": 0.5478353821485836,
"grad_norm": 1.176177465912769,
"learning_rate": 1.0079308754421877e-05,
"loss": 1.2996,
"step": 2050
},
{
"epoch": 0.5491715660074826,
"grad_norm": 1.2416193442138757,
"learning_rate": 1.003265683024487e-05,
"loss": 1.2798,
"step": 2055
},
{
"epoch": 0.5505077498663816,
"grad_norm": 1.2583071246909772,
"learning_rate": 9.986004195301679e-06,
"loss": 1.2727,
"step": 2060
},
{
"epoch": 0.5518439337252806,
"grad_norm": 1.1935438605799258,
"learning_rate": 9.939351864973006e-06,
"loss": 1.256,
"step": 2065
},
{
"epoch": 0.5531801175841796,
"grad_norm": 1.1633338976111967,
"learning_rate": 9.892700854632945e-06,
"loss": 1.2946,
"step": 2070
},
{
"epoch": 0.5545163014430786,
"grad_norm": 1.2341873796862335,
"learning_rate": 9.846052179626854e-06,
"loss": 1.2695,
"step": 2075
},
{
"epoch": 0.5558524853019775,
"grad_norm": 1.2110901028037866,
"learning_rate": 9.799406855249261e-06,
"loss": 1.2939,
"step": 2080
},
{
"epoch": 0.5571886691608765,
"grad_norm": 1.2285980383554018,
"learning_rate": 9.75276589672177e-06,
"loss": 1.2646,
"step": 2085
},
{
"epoch": 0.5585248530197755,
"grad_norm": 1.2442431465515202,
"learning_rate": 9.706130319170968e-06,
"loss": 1.284,
"step": 2090
},
{
"epoch": 0.5598610368786745,
"grad_norm": 1.2387684585320866,
"learning_rate": 9.659501137606317e-06,
"loss": 1.245,
"step": 2095
},
{
"epoch": 0.5611972207375735,
"grad_norm": 1.1969981629519848,
"learning_rate": 9.61287936689808e-06,
"loss": 1.3233,
"step": 2100
},
{
"epoch": 0.5625334045964725,
"grad_norm": 1.2647984251708813,
"learning_rate": 9.566266021755224e-06,
"loss": 1.3229,
"step": 2105
},
{
"epoch": 0.5638695884553715,
"grad_norm": 1.211725302821127,
"learning_rate": 9.519662116703333e-06,
"loss": 1.2725,
"step": 2110
},
{
"epoch": 0.5652057723142705,
"grad_norm": 1.1879063545622028,
"learning_rate": 9.473068666062535e-06,
"loss": 1.2781,
"step": 2115
},
{
"epoch": 0.5665419561731694,
"grad_norm": 1.3276654575701659,
"learning_rate": 9.426486683925412e-06,
"loss": 1.3065,
"step": 2120
},
{
"epoch": 0.5678781400320684,
"grad_norm": 1.237229888093536,
"learning_rate": 9.379917184134949e-06,
"loss": 1.2736,
"step": 2125
},
{
"epoch": 0.5692143238909674,
"grad_norm": 1.1968148529216311,
"learning_rate": 9.33336118026245e-06,
"loss": 1.3061,
"step": 2130
},
{
"epoch": 0.5705505077498664,
"grad_norm": 1.178529508943571,
"learning_rate": 9.286819685585482e-06,
"loss": 1.285,
"step": 2135
},
{
"epoch": 0.5718866916087654,
"grad_norm": 1.1976436960246757,
"learning_rate": 9.240293713065826e-06,
"loss": 1.3261,
"step": 2140
},
{
"epoch": 0.5732228754676644,
"grad_norm": 1.293767414098443,
"learning_rate": 9.193784275327439e-06,
"loss": 1.3081,
"step": 2145
},
{
"epoch": 0.5745590593265634,
"grad_norm": 1.1715456586189346,
"learning_rate": 9.14729238463438e-06,
"loss": 1.2754,
"step": 2150
},
{
"epoch": 0.5758952431854624,
"grad_norm": 1.2266737106027843,
"learning_rate": 9.100819052868818e-06,
"loss": 1.3039,
"step": 2155
},
{
"epoch": 0.5772314270443613,
"grad_norm": 1.2249867841645912,
"learning_rate": 9.054365291508998e-06,
"loss": 1.2862,
"step": 2160
},
{
"epoch": 0.5785676109032603,
"grad_norm": 1.2137804099395924,
"learning_rate": 9.007932111607202e-06,
"loss": 1.3038,
"step": 2165
},
{
"epoch": 0.5799037947621593,
"grad_norm": 1.1366732855918937,
"learning_rate": 8.961520523767777e-06,
"loss": 1.2751,
"step": 2170
},
{
"epoch": 0.5812399786210583,
"grad_norm": 1.2241697536325484,
"learning_rate": 8.915131538125124e-06,
"loss": 1.3154,
"step": 2175
},
{
"epoch": 0.5825761624799572,
"grad_norm": 1.1917760108841098,
"learning_rate": 8.868766164321704e-06,
"loss": 1.3126,
"step": 2180
},
{
"epoch": 0.5839123463388562,
"grad_norm": 1.1679731332413519,
"learning_rate": 8.822425411486087e-06,
"loss": 1.2923,
"step": 2185
},
{
"epoch": 0.5852485301977552,
"grad_norm": 1.160012615589374,
"learning_rate": 8.776110288210964e-06,
"loss": 1.295,
"step": 2190
},
{
"epoch": 0.5865847140566542,
"grad_norm": 1.2221113014345002,
"learning_rate": 8.729821802531213e-06,
"loss": 1.2981,
"step": 2195
},
{
"epoch": 0.5879208979155531,
"grad_norm": 1.1659553272788548,
"learning_rate": 8.683560961901952e-06,
"loss": 1.3156,
"step": 2200
},
{
"epoch": 0.5879208979155531,
"eval_loss": 1.301594853401184,
"eval_runtime": 526.3671,
"eval_samples_per_second": 25.165,
"eval_steps_per_second": 3.146,
"step": 2200
},
{
"epoch": 0.5892570817744521,
"grad_norm": 1.171579385550598,
"learning_rate": 8.637328773176605e-06,
"loss": 1.3301,
"step": 2205
},
{
"epoch": 0.5905932656333511,
"grad_norm": 1.1943778685731268,
"learning_rate": 8.591126242585007e-06,
"loss": 1.3187,
"step": 2210
},
{
"epoch": 0.5919294494922501,
"grad_norm": 1.1944105035087198,
"learning_rate": 8.544954375711487e-06,
"loss": 1.2836,
"step": 2215
},
{
"epoch": 0.5932656333511491,
"grad_norm": 1.2318242169174625,
"learning_rate": 8.498814177472987e-06,
"loss": 1.3131,
"step": 2220
},
{
"epoch": 0.5946018172100481,
"grad_norm": 1.2254563935614524,
"learning_rate": 8.452706652097187e-06,
"loss": 1.2875,
"step": 2225
},
{
"epoch": 0.5959380010689471,
"grad_norm": 1.1704675679433372,
"learning_rate": 8.406632803100665e-06,
"loss": 1.2853,
"step": 2230
},
{
"epoch": 0.5972741849278461,
"grad_norm": 1.1396029283635678,
"learning_rate": 8.360593633267024e-06,
"loss": 1.29,
"step": 2235
},
{
"epoch": 0.598610368786745,
"grad_norm": 1.1872398900997034,
"learning_rate": 8.314590144625102e-06,
"loss": 1.3049,
"step": 2240
},
{
"epoch": 0.599946552645644,
"grad_norm": 1.182289479761897,
"learning_rate": 8.268623338427139e-06,
"loss": 1.2826,
"step": 2245
},
{
"epoch": 0.601282736504543,
"grad_norm": 1.202378486908649,
"learning_rate": 8.222694215126985e-06,
"loss": 1.2919,
"step": 2250
},
{
"epoch": 0.602618920363442,
"grad_norm": 1.1561336328785552,
"learning_rate": 8.176803774358356e-06,
"loss": 1.2615,
"step": 2255
},
{
"epoch": 0.603955104222341,
"grad_norm": 1.2171931903507043,
"learning_rate": 8.130953014913025e-06,
"loss": 1.2887,
"step": 2260
},
{
"epoch": 0.60529128808124,
"grad_norm": 1.1818497689115386,
"learning_rate": 8.085142934719131e-06,
"loss": 1.2627,
"step": 2265
},
{
"epoch": 0.606627471940139,
"grad_norm": 1.2158971480650371,
"learning_rate": 8.039374530819444e-06,
"loss": 1.2829,
"step": 2270
},
{
"epoch": 0.607963655799038,
"grad_norm": 1.275466968098271,
"learning_rate": 7.993648799349646e-06,
"loss": 1.3204,
"step": 2275
},
{
"epoch": 0.6092998396579369,
"grad_norm": 1.2170453783162276,
"learning_rate": 7.947966735516682e-06,
"loss": 1.3045,
"step": 2280
},
{
"epoch": 0.6106360235168359,
"grad_norm": 1.204876466973939,
"learning_rate": 7.902329333577072e-06,
"loss": 1.285,
"step": 2285
},
{
"epoch": 0.6119722073757349,
"grad_norm": 1.2259462458581631,
"learning_rate": 7.856737586815292e-06,
"loss": 1.2937,
"step": 2290
},
{
"epoch": 0.6133083912346339,
"grad_norm": 1.2108988682821065,
"learning_rate": 7.811192487522141e-06,
"loss": 1.2896,
"step": 2295
},
{
"epoch": 0.6146445750935329,
"grad_norm": 1.1922326929306413,
"learning_rate": 7.765695026973155e-06,
"loss": 1.2888,
"step": 2300
},
{
"epoch": 0.6159807589524319,
"grad_norm": 1.2082232502962738,
"learning_rate": 7.72024619540702e-06,
"loss": 1.3027,
"step": 2305
},
{
"epoch": 0.6173169428113309,
"grad_norm": 1.1935181964760695,
"learning_rate": 7.674846982004033e-06,
"loss": 1.2726,
"step": 2310
},
{
"epoch": 0.6186531266702299,
"grad_norm": 1.1988170924470665,
"learning_rate": 7.62949837486456e-06,
"loss": 1.3105,
"step": 2315
},
{
"epoch": 0.6199893105291288,
"grad_norm": 1.1788468720871008,
"learning_rate": 7.584201360987544e-06,
"loss": 1.2962,
"step": 2320
},
{
"epoch": 0.6213254943880278,
"grad_norm": 1.2164638957688612,
"learning_rate": 7.538956926249013e-06,
"loss": 1.2918,
"step": 2325
},
{
"epoch": 0.6226616782469268,
"grad_norm": 1.2540807555588527,
"learning_rate": 7.493766055380622e-06,
"loss": 1.3032,
"step": 2330
},
{
"epoch": 0.6239978621058258,
"grad_norm": 1.2021644980450277,
"learning_rate": 7.44862973194823e-06,
"loss": 1.2784,
"step": 2335
},
{
"epoch": 0.6253340459647247,
"grad_norm": 1.1374900868431708,
"learning_rate": 7.403548938330487e-06,
"loss": 1.2615,
"step": 2340
},
{
"epoch": 0.6266702298236237,
"grad_norm": 1.223926027057831,
"learning_rate": 7.358524655697445e-06,
"loss": 1.3324,
"step": 2345
},
{
"epoch": 0.6280064136825227,
"grad_norm": 1.1663009699556086,
"learning_rate": 7.31355786398922e-06,
"loss": 1.3073,
"step": 2350
},
{
"epoch": 0.6293425975414217,
"grad_norm": 1.1605825631923092,
"learning_rate": 7.268649541894658e-06,
"loss": 1.3167,
"step": 2355
},
{
"epoch": 0.6306787814003206,
"grad_norm": 1.181802501051011,
"learning_rate": 7.223800666830013e-06,
"loss": 1.272,
"step": 2360
},
{
"epoch": 0.6320149652592196,
"grad_norm": 1.2124656825123712,
"learning_rate": 7.1790122149177135e-06,
"loss": 1.2783,
"step": 2365
},
{
"epoch": 0.6333511491181186,
"grad_norm": 1.159478422052137,
"learning_rate": 7.134285160965091e-06,
"loss": 1.3114,
"step": 2370
},
{
"epoch": 0.6346873329770176,
"grad_norm": 1.2191673865530408,
"learning_rate": 7.089620478443152e-06,
"loss": 1.2873,
"step": 2375
},
{
"epoch": 0.6360235168359166,
"grad_norm": 1.2212422633172386,
"learning_rate": 7.045019139465434e-06,
"loss": 1.2976,
"step": 2380
},
{
"epoch": 0.6373597006948156,
"grad_norm": 1.2849058652407424,
"learning_rate": 7.000482114766798e-06,
"loss": 1.3026,
"step": 2385
},
{
"epoch": 0.6386958845537146,
"grad_norm": 1.1693633671456403,
"learning_rate": 6.956010373682334e-06,
"loss": 1.2827,
"step": 2390
},
{
"epoch": 0.6400320684126136,
"grad_norm": 1.160020552936478,
"learning_rate": 6.911604884126259e-06,
"loss": 1.2515,
"step": 2395
},
{
"epoch": 0.6413682522715125,
"grad_norm": 1.1828314546763101,
"learning_rate": 6.867266612570833e-06,
"loss": 1.3092,
"step": 2400
},
{
"epoch": 0.6413682522715125,
"eval_loss": 1.2952780723571777,
"eval_runtime": 525.8904,
"eval_samples_per_second": 25.188,
"eval_steps_per_second": 3.149,
"step": 2400
},
{
"epoch": 0.6427044361304115,
"grad_norm": 1.137455706013817,
"learning_rate": 6.822996524025343e-06,
"loss": 1.2891,
"step": 2405
},
{
"epoch": 0.6440406199893105,
"grad_norm": 1.3754307767427276,
"learning_rate": 6.778795582015096e-06,
"loss": 1.2754,
"step": 2410
},
{
"epoch": 0.6453768038482095,
"grad_norm": 1.216856677079751,
"learning_rate": 6.734664748560437e-06,
"loss": 1.3014,
"step": 2415
},
{
"epoch": 0.6467129877071085,
"grad_norm": 1.2592962476705394,
"learning_rate": 6.690604984155826e-06,
"loss": 1.3221,
"step": 2420
},
{
"epoch": 0.6480491715660075,
"grad_norm": 1.1929612361301374,
"learning_rate": 6.646617247748926e-06,
"loss": 1.2768,
"step": 2425
},
{
"epoch": 0.6493853554249065,
"grad_norm": 1.1709139264257042,
"learning_rate": 6.602702496719724e-06,
"loss": 1.2549,
"step": 2430
},
{
"epoch": 0.6507215392838055,
"grad_norm": 1.2183672529081806,
"learning_rate": 6.558861686859712e-06,
"loss": 1.2764,
"step": 2435
},
{
"epoch": 0.6520577231427044,
"grad_norm": 1.1975495934510272,
"learning_rate": 6.515095772351072e-06,
"loss": 1.2929,
"step": 2440
},
{
"epoch": 0.6533939070016034,
"grad_norm": 1.1613807938703284,
"learning_rate": 6.471405705745906e-06,
"loss": 1.2684,
"step": 2445
},
{
"epoch": 0.6547300908605024,
"grad_norm": 1.2736187330390891,
"learning_rate": 6.427792437945516e-06,
"loss": 1.262,
"step": 2450
},
{
"epoch": 0.6560662747194014,
"grad_norm": 1.174927022230355,
"learning_rate": 6.384256918179692e-06,
"loss": 1.2958,
"step": 2455
},
{
"epoch": 0.6574024585783004,
"grad_norm": 1.1737222195085288,
"learning_rate": 6.340800093986071e-06,
"loss": 1.2987,
"step": 2460
},
{
"epoch": 0.6587386424371994,
"grad_norm": 1.1988962500744893,
"learning_rate": 6.297422911189499e-06,
"loss": 1.2469,
"step": 2465
},
{
"epoch": 0.6600748262960984,
"grad_norm": 1.138343574152133,
"learning_rate": 6.254126313881448e-06,
"loss": 1.2564,
"step": 2470
},
{
"epoch": 0.6614110101549974,
"grad_norm": 1.161565466877101,
"learning_rate": 6.210911244399477e-06,
"loss": 1.305,
"step": 2475
},
{
"epoch": 0.6627471940138963,
"grad_norm": 1.1944909640364771,
"learning_rate": 6.167778643306717e-06,
"loss": 1.2749,
"step": 2480
},
{
"epoch": 0.6640833778727953,
"grad_norm": 1.2250102533436973,
"learning_rate": 6.1247294493713845e-06,
"loss": 1.2892,
"step": 2485
},
{
"epoch": 0.6654195617316943,
"grad_norm": 1.211997261711134,
"learning_rate": 6.0817645995463845e-06,
"loss": 1.3064,
"step": 2490
},
{
"epoch": 0.6667557455905933,
"grad_norm": 1.2199059761183304,
"learning_rate": 6.038885028948889e-06,
"loss": 1.2591,
"step": 2495
},
{
"epoch": 0.6680919294494923,
"grad_norm": 1.204780340659424,
"learning_rate": 5.996091670839983e-06,
"loss": 1.2981,
"step": 2500
},
{
"epoch": 0.6694281133083912,
"grad_norm": 1.1622142280662515,
"learning_rate": 5.953385456604377e-06,
"loss": 1.2837,
"step": 2505
},
{
"epoch": 0.6707642971672902,
"grad_norm": 1.2595949655572636,
"learning_rate": 5.910767315730119e-06,
"loss": 1.3098,
"step": 2510
},
{
"epoch": 0.6721004810261892,
"grad_norm": 1.1991193943391245,
"learning_rate": 5.868238175788355e-06,
"loss": 1.2844,
"step": 2515
},
{
"epoch": 0.6734366648850882,
"grad_norm": 1.1286538895272766,
"learning_rate": 5.825798962413164e-06,
"loss": 1.2963,
"step": 2520
},
{
"epoch": 0.6747728487439871,
"grad_norm": 1.2065687187980167,
"learning_rate": 5.783450599281395e-06,
"loss": 1.3002,
"step": 2525
},
{
"epoch": 0.6761090326028861,
"grad_norm": 1.167072268529669,
"learning_rate": 5.7411940080925685e-06,
"loss": 1.2479,
"step": 2530
},
{
"epoch": 0.6774452164617851,
"grad_norm": 1.1698208994032941,
"learning_rate": 5.6990301085488235e-06,
"loss": 1.2999,
"step": 2535
},
{
"epoch": 0.6787814003206841,
"grad_norm": 1.1710233370221095,
"learning_rate": 5.656959818334873e-06,
"loss": 1.2595,
"step": 2540
},
{
"epoch": 0.6801175841795831,
"grad_norm": 1.386755510116527,
"learning_rate": 5.614984053098076e-06,
"loss": 1.2879,
"step": 2545
},
{
"epoch": 0.6814537680384821,
"grad_norm": 1.132940281229991,
"learning_rate": 5.5731037264284735e-06,
"loss": 1.2559,
"step": 2550
},
{
"epoch": 0.6827899518973811,
"grad_norm": 1.1926751219485074,
"learning_rate": 5.531319749838903e-06,
"loss": 1.2906,
"step": 2555
},
{
"epoch": 0.6841261357562801,
"grad_norm": 1.2244655010133048,
"learning_rate": 5.489633032745185e-06,
"loss": 1.2589,
"step": 2560
},
{
"epoch": 0.685462319615179,
"grad_norm": 1.177667766929752,
"learning_rate": 5.448044482446317e-06,
"loss": 1.2896,
"step": 2565
},
{
"epoch": 0.686798503474078,
"grad_norm": 1.1792403166242948,
"learning_rate": 5.406555004104712e-06,
"loss": 1.2724,
"step": 2570
},
{
"epoch": 0.688134687332977,
"grad_norm": 1.1752737928249541,
"learning_rate": 5.365165500726518e-06,
"loss": 1.3095,
"step": 2575
},
{
"epoch": 0.689470871191876,
"grad_norm": 1.2006301934171553,
"learning_rate": 5.323876873141973e-06,
"loss": 1.2682,
"step": 2580
},
{
"epoch": 0.690807055050775,
"grad_norm": 1.232283420067014,
"learning_rate": 5.282690019985756e-06,
"loss": 1.2973,
"step": 2585
},
{
"epoch": 0.692143238909674,
"grad_norm": 1.8591318420869134,
"learning_rate": 5.241605837677481e-06,
"loss": 1.2748,
"step": 2590
},
{
"epoch": 0.693479422768573,
"grad_norm": 1.1409257082194615,
"learning_rate": 5.200625220402139e-06,
"loss": 1.2786,
"step": 2595
},
{
"epoch": 0.694815606627472,
"grad_norm": 1.1874265547469547,
"learning_rate": 5.159749060090675e-06,
"loss": 1.2518,
"step": 2600
},
{
"epoch": 0.694815606627472,
"eval_loss": 1.2894538640975952,
"eval_runtime": 524.2545,
"eval_samples_per_second": 25.266,
"eval_steps_per_second": 3.159,
"step": 2600
},
{
"epoch": 0.6961517904863709,
"grad_norm": 1.1436341244774122,
"learning_rate": 5.118978246400555e-06,
"loss": 1.2868,
"step": 2605
},
{
"epoch": 0.6974879743452699,
"grad_norm": 1.1773055221354787,
"learning_rate": 5.078313666696404e-06,
"loss": 1.2865,
"step": 2610
},
{
"epoch": 0.6988241582041689,
"grad_norm": 1.205692219736226,
"learning_rate": 5.0377562060307e-06,
"loss": 1.271,
"step": 2615
},
{
"epoch": 0.7001603420630679,
"grad_norm": 1.1681619335726294,
"learning_rate": 4.997306747124508e-06,
"loss": 1.281,
"step": 2620
},
{
"epoch": 0.7014965259219669,
"grad_norm": 1.1239087952042888,
"learning_rate": 4.9569661703482585e-06,
"loss": 1.2933,
"step": 2625
},
{
"epoch": 0.7028327097808659,
"grad_norm": 1.1618547501473921,
"learning_rate": 4.916735353702603e-06,
"loss": 1.2556,
"step": 2630
},
{
"epoch": 0.7041688936397649,
"grad_norm": 1.1446438877087823,
"learning_rate": 4.876615172799294e-06,
"loss": 1.2964,
"step": 2635
},
{
"epoch": 0.7055050774986639,
"grad_norm": 1.2196306868529612,
"learning_rate": 4.836606500842129e-06,
"loss": 1.2934,
"step": 2640
},
{
"epoch": 0.7068412613575628,
"grad_norm": 1.1966884099651143,
"learning_rate": 4.7967102086079485e-06,
"loss": 1.2432,
"step": 2645
},
{
"epoch": 0.7081774452164618,
"grad_norm": 1.1483128968540317,
"learning_rate": 4.756927164427685e-06,
"loss": 1.2795,
"step": 2650
},
{
"epoch": 0.7095136290753608,
"grad_norm": 1.1539714350859465,
"learning_rate": 4.717258234167448e-06,
"loss": 1.2494,
"step": 2655
},
{
"epoch": 0.7108498129342598,
"grad_norm": 1.1421161621669704,
"learning_rate": 4.677704281209707e-06,
"loss": 1.2721,
"step": 2660
},
{
"epoch": 0.7121859967931587,
"grad_norm": 1.1666778493912067,
"learning_rate": 4.6382661664344665e-06,
"loss": 1.2805,
"step": 2665
},
{
"epoch": 0.7135221806520577,
"grad_norm": 1.2152799023877705,
"learning_rate": 4.59894474820057e-06,
"loss": 1.2781,
"step": 2670
},
{
"epoch": 0.7148583645109567,
"grad_norm": 1.1802067902023374,
"learning_rate": 4.559740882326984e-06,
"loss": 1.2889,
"step": 2675
},
{
"epoch": 0.7161945483698557,
"grad_norm": 1.1615396840856154,
"learning_rate": 4.520655422074176e-06,
"loss": 1.2295,
"step": 2680
},
{
"epoch": 0.7175307322287546,
"grad_norm": 1.1290092637035325,
"learning_rate": 4.481689218125561e-06,
"loss": 1.2812,
"step": 2685
},
{
"epoch": 0.7188669160876536,
"grad_norm": 1.1477861204230764,
"learning_rate": 4.442843118568976e-06,
"loss": 1.2565,
"step": 2690
},
{
"epoch": 0.7202030999465526,
"grad_norm": 1.2347710658904285,
"learning_rate": 4.4041179688782095e-06,
"loss": 1.2959,
"step": 2695
},
{
"epoch": 0.7215392838054516,
"grad_norm": 1.2180531197328832,
"learning_rate": 4.365514611894623e-06,
"loss": 1.2976,
"step": 2700
},
{
"epoch": 0.7228754676643506,
"grad_norm": 1.1757278803891242,
"learning_rate": 4.327033887808802e-06,
"loss": 1.2761,
"step": 2705
},
{
"epoch": 0.7242116515232496,
"grad_norm": 1.1645726922846997,
"learning_rate": 4.288676634142247e-06,
"loss": 1.2595,
"step": 2710
},
{
"epoch": 0.7255478353821486,
"grad_norm": 1.1411088256784303,
"learning_rate": 4.25044368572917e-06,
"loss": 1.2714,
"step": 2715
},
{
"epoch": 0.7268840192410476,
"grad_norm": 1.1671327639274043,
"learning_rate": 4.2123358746983225e-06,
"loss": 1.2876,
"step": 2720
},
{
"epoch": 0.7282202030999465,
"grad_norm": 1.1581886797505927,
"learning_rate": 4.1743540304548615e-06,
"loss": 1.2859,
"step": 2725
},
{
"epoch": 0.7295563869588455,
"grad_norm": 1.1510537093782165,
"learning_rate": 4.13649897966233e-06,
"loss": 1.2437,
"step": 2730
},
{
"epoch": 0.7308925708177445,
"grad_norm": 1.1962597941965873,
"learning_rate": 4.098771546224643e-06,
"loss": 1.2697,
"step": 2735
},
{
"epoch": 0.7322287546766435,
"grad_norm": 1.176978028412522,
"learning_rate": 4.061172551268162e-06,
"loss": 1.3069,
"step": 2740
},
{
"epoch": 0.7335649385355425,
"grad_norm": 1.2344307343981975,
"learning_rate": 4.023702813123828e-06,
"loss": 1.3034,
"step": 2745
},
{
"epoch": 0.7349011223944415,
"grad_norm": 1.1490296827570874,
"learning_rate": 3.986363147309332e-06,
"loss": 1.2934,
"step": 2750
},
{
"epoch": 0.7362373062533405,
"grad_norm": 1.128655104477458,
"learning_rate": 3.949154366511395e-06,
"loss": 1.3227,
"step": 2755
},
{
"epoch": 0.7375734901122395,
"grad_norm": 1.1377152493000582,
"learning_rate": 3.9120772805680575e-06,
"loss": 1.2837,
"step": 2760
},
{
"epoch": 0.7389096739711384,
"grad_norm": 1.1144595543031406,
"learning_rate": 3.8751326964510615e-06,
"loss": 1.2451,
"step": 2765
},
{
"epoch": 0.7402458578300374,
"grad_norm": 1.1528921550358049,
"learning_rate": 3.838321418248288e-06,
"loss": 1.2615,
"step": 2770
},
{
"epoch": 0.7415820416889364,
"grad_norm": 1.2690943955678684,
"learning_rate": 3.8016442471462524e-06,
"loss": 1.2829,
"step": 2775
},
{
"epoch": 0.7429182255478354,
"grad_norm": 1.2308517139392547,
"learning_rate": 3.7651019814126656e-06,
"loss": 1.2617,
"step": 2780
},
{
"epoch": 0.7442544094067344,
"grad_norm": 1.2017097006317634,
"learning_rate": 3.7286954163790734e-06,
"loss": 1.2364,
"step": 2785
},
{
"epoch": 0.7455905932656334,
"grad_norm": 1.1828941772780268,
"learning_rate": 3.6924253444235224e-06,
"loss": 1.3087,
"step": 2790
},
{
"epoch": 0.7469267771245324,
"grad_norm": 1.1861330848081788,
"learning_rate": 3.6562925549533355e-06,
"loss": 1.2789,
"step": 2795
},
{
"epoch": 0.7482629609834314,
"grad_norm": 1.1526594671507095,
"learning_rate": 3.6202978343879337e-06,
"loss": 1.2617,
"step": 2800
},
{
"epoch": 0.7482629609834314,
"eval_loss": 1.2845991849899292,
"eval_runtime": 523.8103,
"eval_samples_per_second": 25.288,
"eval_steps_per_second": 3.161,
"step": 2800
},
{
"epoch": 0.7495991448423303,
"grad_norm": 1.1298081744323016,
"learning_rate": 3.5844419661416886e-06,
"loss": 1.2902,
"step": 2805
},
{
"epoch": 0.7509353287012293,
"grad_norm": 1.1561421830885714,
"learning_rate": 3.5487257306069054e-06,
"loss": 1.29,
"step": 2810
},
{
"epoch": 0.7522715125601283,
"grad_norm": 1.1536531647083885,
"learning_rate": 3.5131499051368254e-06,
"loss": 1.2614,
"step": 2815
},
{
"epoch": 0.7536076964190273,
"grad_norm": 1.1856993866727077,
"learning_rate": 3.4777152640286984e-06,
"loss": 1.2861,
"step": 2820
},
{
"epoch": 0.7549438802779262,
"grad_norm": 1.2042645879668878,
"learning_rate": 3.4424225785069444e-06,
"loss": 1.2658,
"step": 2825
},
{
"epoch": 0.7562800641368252,
"grad_norm": 1.1759218780648395,
"learning_rate": 3.4072726167063617e-06,
"loss": 1.2613,
"step": 2830
},
{
"epoch": 0.7576162479957242,
"grad_norm": 1.1941131261396993,
"learning_rate": 3.3722661436554104e-06,
"loss": 1.2926,
"step": 2835
},
{
"epoch": 0.7589524318546232,
"grad_norm": 1.2349621736436946,
"learning_rate": 3.337403921259559e-06,
"loss": 1.2852,
"step": 2840
},
{
"epoch": 0.7602886157135221,
"grad_norm": 1.243962877866245,
"learning_rate": 3.3026867082847058e-06,
"loss": 1.3087,
"step": 2845
},
{
"epoch": 0.7616247995724211,
"grad_norm": 1.2755840307358004,
"learning_rate": 3.268115260340654e-06,
"loss": 1.2574,
"step": 2850
},
{
"epoch": 0.7629609834313201,
"grad_norm": 1.256694820928362,
"learning_rate": 3.233690329864684e-06,
"loss": 1.2477,
"step": 2855
},
{
"epoch": 0.7642971672902191,
"grad_norm": 1.153504157549798,
"learning_rate": 3.1994126661051628e-06,
"loss": 1.2669,
"step": 2860
},
{
"epoch": 0.7656333511491181,
"grad_norm": 1.2272283092280594,
"learning_rate": 3.1652830151052416e-06,
"loss": 1.2919,
"step": 2865
},
{
"epoch": 0.7669695350080171,
"grad_norm": 1.1512364936411144,
"learning_rate": 3.131302119686621e-06,
"loss": 1.2736,
"step": 2870
},
{
"epoch": 0.7683057188669161,
"grad_norm": 1.1564153021008827,
"learning_rate": 3.097470719433373e-06,
"loss": 1.2556,
"step": 2875
},
{
"epoch": 0.7696419027258151,
"grad_norm": 1.164737683509797,
"learning_rate": 3.0637895506758597e-06,
"loss": 1.2672,
"step": 2880
},
{
"epoch": 0.770978086584714,
"grad_norm": 1.1685661560913843,
"learning_rate": 3.0302593464746944e-06,
"loss": 1.2907,
"step": 2885
},
{
"epoch": 0.772314270443613,
"grad_norm": 1.1699267498621475,
"learning_rate": 2.9968808366047942e-06,
"loss": 1.2699,
"step": 2890
},
{
"epoch": 0.773650454302512,
"grad_norm": 1.211199795955048,
"learning_rate": 2.963654747539494e-06,
"loss": 1.2693,
"step": 2895
},
{
"epoch": 0.774986638161411,
"grad_norm": 1.1816669846421493,
"learning_rate": 2.9305818024347378e-06,
"loss": 1.2651,
"step": 2900
},
{
"epoch": 0.77632282202031,
"grad_norm": 1.1914669242505636,
"learning_rate": 2.897662721113328e-06,
"loss": 1.2737,
"step": 2905
},
{
"epoch": 0.777659005879209,
"grad_norm": 1.1738719241452067,
"learning_rate": 2.864898220049277e-06,
"loss": 1.2564,
"step": 2910
},
{
"epoch": 0.778995189738108,
"grad_norm": 1.1840077986894364,
"learning_rate": 2.832289012352203e-06,
"loss": 1.2522,
"step": 2915
},
{
"epoch": 0.780331373597007,
"grad_norm": 1.233343112626304,
"learning_rate": 2.7998358077517975e-06,
"loss": 1.2448,
"step": 2920
},
{
"epoch": 0.7816675574559059,
"grad_norm": 1.1877039129056828,
"learning_rate": 2.7675393125824144e-06,
"loss": 1.2563,
"step": 2925
},
{
"epoch": 0.7830037413148049,
"grad_norm": 1.1982422068810221,
"learning_rate": 2.735400229767652e-06,
"loss": 1.2403,
"step": 2930
},
{
"epoch": 0.7843399251737039,
"grad_norm": 1.1935298921878634,
"learning_rate": 2.7034192588050845e-06,
"loss": 1.2701,
"step": 2935
},
{
"epoch": 0.7856761090326029,
"grad_norm": 1.151254746126473,
"learning_rate": 2.671597095751033e-06,
"loss": 1.2509,
"step": 2940
},
{
"epoch": 0.7870122928915019,
"grad_norm": 1.190278414424794,
"learning_rate": 2.6399344332054e-06,
"loss": 1.2914,
"step": 2945
},
{
"epoch": 0.7883484767504009,
"grad_norm": 1.168908881155384,
"learning_rate": 2.6084319602966167e-06,
"loss": 1.2739,
"step": 2950
},
{
"epoch": 0.7896846606092999,
"grad_norm": 1.1495974699891471,
"learning_rate": 2.577090362666631e-06,
"loss": 1.2822,
"step": 2955
},
{
"epoch": 0.7910208444681989,
"grad_norm": 1.199562609813528,
"learning_rate": 2.54591032245599e-06,
"loss": 1.2833,
"step": 2960
},
{
"epoch": 0.7923570283270978,
"grad_norm": 1.172598488602994,
"learning_rate": 2.514892518288988e-06,
"loss": 1.2427,
"step": 2965
},
{
"epoch": 0.7936932121859968,
"grad_norm": 1.2023007809260173,
"learning_rate": 2.484037625258908e-06,
"loss": 1.3116,
"step": 2970
},
{
"epoch": 0.7950293960448958,
"grad_norm": 1.1808013233447734,
"learning_rate": 2.4533463149133073e-06,
"loss": 1.2367,
"step": 2975
},
{
"epoch": 0.7963655799037948,
"grad_norm": 1.2065243432367916,
"learning_rate": 2.422819255239427e-06,
"loss": 1.2995,
"step": 2980
},
{
"epoch": 0.7977017637626937,
"grad_norm": 1.1490648428460022,
"learning_rate": 2.392457110649634e-06,
"loss": 1.2676,
"step": 2985
},
{
"epoch": 0.7990379476215927,
"grad_norm": 1.136037346488201,
"learning_rate": 2.362260541966972e-06,
"loss": 1.2816,
"step": 2990
},
{
"epoch": 0.8003741314804917,
"grad_norm": 1.149577173830776,
"learning_rate": 2.3322302064107762e-06,
"loss": 1.2948,
"step": 2995
},
{
"epoch": 0.8017103153393907,
"grad_norm": 1.185488122515381,
"learning_rate": 2.302366757582355e-06,
"loss": 1.3041,
"step": 3000
},
{
"epoch": 0.8017103153393907,
"eval_loss": 1.2808870077133179,
"eval_runtime": 523.4307,
"eval_samples_per_second": 25.306,
"eval_steps_per_second": 3.164,
"step": 3000
},
{
"epoch": 0.8030464991982896,
"grad_norm": 1.1513855940527316,
"learning_rate": 2.272670845450791e-06,
"loss": 1.277,
"step": 3005
},
{
"epoch": 0.8043826830571886,
"grad_norm": 1.1602967857229551,
"learning_rate": 2.243143116338773e-06,
"loss": 1.2653,
"step": 3010
},
{
"epoch": 0.8057188669160876,
"grad_norm": 1.197568564758537,
"learning_rate": 2.2137842129085396e-06,
"loss": 1.3524,
"step": 3015
},
{
"epoch": 0.8070550507749866,
"grad_norm": 1.1421243027571122,
"learning_rate": 2.1845947741478857e-06,
"loss": 1.2651,
"step": 3020
},
{
"epoch": 0.8083912346338856,
"grad_norm": 1.167663307330092,
"learning_rate": 2.1555754353562643e-06,
"loss": 1.2337,
"step": 3025
},
{
"epoch": 0.8097274184927846,
"grad_norm": 1.1570704863531982,
"learning_rate": 2.1267268281309418e-06,
"loss": 1.2671,
"step": 3030
},
{
"epoch": 0.8110636023516836,
"grad_norm": 1.1646116509684428,
"learning_rate": 2.098049580353273e-06,
"loss": 1.2831,
"step": 3035
},
{
"epoch": 0.8123997862105826,
"grad_norm": 1.1929244270961756,
"learning_rate": 2.069544316175025e-06,
"loss": 1.2631,
"step": 3040
},
{
"epoch": 0.8137359700694815,
"grad_norm": 1.1495579309904203,
"learning_rate": 2.041211656004781e-06,
"loss": 1.2891,
"step": 3045
},
{
"epoch": 0.8150721539283805,
"grad_norm": 1.158971184848035,
"learning_rate": 2.0130522164944666e-06,
"loss": 1.2543,
"step": 3050
},
{
"epoch": 0.8164083377872795,
"grad_norm": 1.1993305810419228,
"learning_rate": 1.985066610525904e-06,
"loss": 1.3294,
"step": 3055
},
{
"epoch": 0.8177445216461785,
"grad_norm": 1.1523443888711082,
"learning_rate": 1.9572554471974723e-06,
"loss": 1.2771,
"step": 3060
},
{
"epoch": 0.8190807055050775,
"grad_norm": 1.1695216750777881,
"learning_rate": 1.9296193318108723e-06,
"loss": 1.2578,
"step": 3065
},
{
"epoch": 0.8204168893639765,
"grad_norm": 1.1732642497713914,
"learning_rate": 1.9021588658579249e-06,
"loss": 1.2513,
"step": 3070
},
{
"epoch": 0.8217530732228755,
"grad_norm": 1.1519213257119236,
"learning_rate": 1.8748746470075029e-06,
"loss": 1.2977,
"step": 3075
},
{
"epoch": 0.8230892570817745,
"grad_norm": 1.1516668595048642,
"learning_rate": 1.847767269092511e-06,
"loss": 1.2321,
"step": 3080
},
{
"epoch": 0.8244254409406734,
"grad_norm": 1.1909108579078609,
"learning_rate": 1.820837322096961e-06,
"loss": 1.2484,
"step": 3085
},
{
"epoch": 0.8257616247995724,
"grad_norm": 1.101081523016669,
"learning_rate": 1.7940853921431378e-06,
"loss": 1.2575,
"step": 3090
},
{
"epoch": 0.8270978086584714,
"grad_norm": 1.178074475938639,
"learning_rate": 1.7675120614788367e-06,
"loss": 1.2664,
"step": 3095
},
{
"epoch": 0.8284339925173704,
"grad_norm": 1.1731584765176883,
"learning_rate": 1.7411179084646879e-06,
"loss": 1.2804,
"step": 3100
},
{
"epoch": 0.8297701763762694,
"grad_norm": 1.1586731249421351,
"learning_rate": 1.7149035075615795e-06,
"loss": 1.2905,
"step": 3105
},
{
"epoch": 0.8311063602351684,
"grad_norm": 1.169289922029439,
"learning_rate": 1.6888694293181462e-06,
"loss": 1.2836,
"step": 3110
},
{
"epoch": 0.8324425440940674,
"grad_norm": 1.1719327462563847,
"learning_rate": 1.6630162403583538e-06,
"loss": 1.2513,
"step": 3115
},
{
"epoch": 0.8337787279529664,
"grad_norm": 1.1435061474054447,
"learning_rate": 1.637344503369167e-06,
"loss": 1.2635,
"step": 3120
},
{
"epoch": 0.8351149118118654,
"grad_norm": 1.1521864026183537,
"learning_rate": 1.6118547770883031e-06,
"loss": 1.3022,
"step": 3125
},
{
"epoch": 0.8364510956707643,
"grad_norm": 1.2218598184249856,
"learning_rate": 1.5865476162920658e-06,
"loss": 1.293,
"step": 3130
},
{
"epoch": 0.8377872795296633,
"grad_norm": 1.1461871951868958,
"learning_rate": 1.5614235717832838e-06,
"loss": 1.284,
"step": 3135
},
{
"epoch": 0.8391234633885623,
"grad_norm": 1.2002653441412816,
"learning_rate": 1.536483190379302e-06,
"loss": 1.2608,
"step": 3140
},
{
"epoch": 0.8404596472474613,
"grad_norm": 1.194888981517886,
"learning_rate": 1.5117270149001061e-06,
"loss": 1.2728,
"step": 3145
},
{
"epoch": 0.8417958311063602,
"grad_norm": 1.1669886636546563,
"learning_rate": 1.4871555841564889e-06,
"loss": 1.2936,
"step": 3150
},
{
"epoch": 0.8431320149652592,
"grad_norm": 1.178341085561249,
"learning_rate": 1.4627694329383245e-06,
"loss": 1.2737,
"step": 3155
},
{
"epoch": 0.8444681988241582,
"grad_norm": 1.173822019467738,
"learning_rate": 1.4385690920029394e-06,
"loss": 1.2389,
"step": 3160
},
{
"epoch": 0.8458043826830572,
"grad_norm": 1.172444929862991,
"learning_rate": 1.4145550880635551e-06,
"loss": 1.283,
"step": 3165
},
{
"epoch": 0.8471405665419561,
"grad_norm": 1.157959236772995,
"learning_rate": 1.3907279437778154e-06,
"loss": 1.3089,
"step": 3170
},
{
"epoch": 0.8484767504008551,
"grad_norm": 1.142956839312928,
"learning_rate": 1.3670881777364276e-06,
"loss": 1.2385,
"step": 3175
},
{
"epoch": 0.8498129342597541,
"grad_norm": 1.1394377032150336,
"learning_rate": 1.3436363044518685e-06,
"loss": 1.3002,
"step": 3180
},
{
"epoch": 0.8511491181186531,
"grad_norm": 1.1081112610690127,
"learning_rate": 1.3203728343471766e-06,
"loss": 1.2471,
"step": 3185
},
{
"epoch": 0.8524853019775521,
"grad_norm": 1.1233094984645868,
"learning_rate": 1.2972982737448582e-06,
"loss": 1.2511,
"step": 3190
},
{
"epoch": 0.8538214858364511,
"grad_norm": 1.168807245788649,
"learning_rate": 1.27441312485586e-06,
"loss": 1.2487,
"step": 3195
},
{
"epoch": 0.8551576696953501,
"grad_norm": 1.2213102334394346,
"learning_rate": 1.2517178857686318e-06,
"loss": 1.3102,
"step": 3200
},
{
"epoch": 0.8551576696953501,
"eval_loss": 1.278082013130188,
"eval_runtime": 523.3237,
"eval_samples_per_second": 25.311,
"eval_steps_per_second": 3.164,
"step": 3200
},
{
"epoch": 0.8564938535542491,
"grad_norm": 1.120564972312021,
"learning_rate": 1.2292130504383037e-06,
"loss": 1.2578,
"step": 3205
},
{
"epoch": 0.857830037413148,
"grad_norm": 1.1580496245252512,
"learning_rate": 1.2068991086759175e-06,
"loss": 1.2983,
"step": 3210
},
{
"epoch": 0.859166221272047,
"grad_norm": 1.1198656733420578,
"learning_rate": 1.1847765461377757e-06,
"loss": 1.2245,
"step": 3215
},
{
"epoch": 0.860502405130946,
"grad_norm": 1.2004671948391095,
"learning_rate": 1.1628458443148716e-06,
"loss": 1.2737,
"step": 3220
},
{
"epoch": 0.861838588989845,
"grad_norm": 1.114311359039494,
"learning_rate": 1.1411074805223997e-06,
"loss": 1.2583,
"step": 3225
},
{
"epoch": 0.863174772848744,
"grad_norm": 1.1608634318847668,
"learning_rate": 1.119561927889381e-06,
"loss": 1.2543,
"step": 3230
},
{
"epoch": 0.864510956707643,
"grad_norm": 1.178010917617475,
"learning_rate": 1.0982096553483568e-06,
"loss": 1.3171,
"step": 3235
},
{
"epoch": 0.865847140566542,
"grad_norm": 1.1868881875439512,
"learning_rate": 1.077051127625185e-06,
"loss": 1.2985,
"step": 3240
},
{
"epoch": 0.867183324425441,
"grad_norm": 1.1357672595216541,
"learning_rate": 1.0560868052289253e-06,
"loss": 1.2634,
"step": 3245
},
{
"epoch": 0.8685195082843399,
"grad_norm": 1.171433935969695,
"learning_rate": 1.0353171444418187e-06,
"loss": 1.2759,
"step": 3250
},
{
"epoch": 0.8698556921432389,
"grad_norm": 1.1224025445608299,
"learning_rate": 1.014742597309346e-06,
"loss": 1.267,
"step": 3255
},
{
"epoch": 0.8711918760021379,
"grad_norm": 1.1874141914261405,
"learning_rate": 9.943636116304068e-07,
"loss": 1.2509,
"step": 3260
},
{
"epoch": 0.8725280598610369,
"grad_norm": 1.1508324966317753,
"learning_rate": 9.741806309475588e-07,
"loss": 1.2647,
"step": 3265
},
{
"epoch": 0.8738642437199359,
"grad_norm": 1.1671717510783461,
"learning_rate": 9.541940945373718e-07,
"loss": 1.2357,
"step": 3270
},
{
"epoch": 0.8752004275788349,
"grad_norm": 1.1214376758027105,
"learning_rate": 9.344044374008632e-07,
"loss": 1.2598,
"step": 3275
},
{
"epoch": 0.8765366114377339,
"grad_norm": 1.1745237243989533,
"learning_rate": 9.148120902540281e-07,
"loss": 1.238,
"step": 3280
},
{
"epoch": 0.8778727952966329,
"grad_norm": 1.1566262573536532,
"learning_rate": 8.954174795184756e-07,
"loss": 1.2464,
"step": 3285
},
{
"epoch": 0.8792089791555318,
"grad_norm": 1.158731873705362,
"learning_rate": 8.762210273121363e-07,
"loss": 1.2783,
"step": 3290
},
{
"epoch": 0.8805451630144308,
"grad_norm": 1.1187300362715888,
"learning_rate": 8.572231514400775e-07,
"loss": 1.2692,
"step": 3295
},
{
"epoch": 0.8818813468733298,
"grad_norm": 1.1820432928944034,
"learning_rate": 8.384242653854146e-07,
"loss": 1.3092,
"step": 3300
},
{
"epoch": 0.8832175307322288,
"grad_norm": 1.1315388199619425,
"learning_rate": 8.198247783003133e-07,
"loss": 1.2581,
"step": 3305
},
{
"epoch": 0.8845537145911277,
"grad_norm": 1.135539300633719,
"learning_rate": 8.014250949970704e-07,
"loss": 1.2726,
"step": 3310
},
{
"epoch": 0.8858898984500267,
"grad_norm": 1.147749224885632,
"learning_rate": 7.832256159393181e-07,
"loss": 1.2426,
"step": 3315
},
{
"epoch": 0.8872260823089257,
"grad_norm": 1.1681170491391923,
"learning_rate": 7.652267372333056e-07,
"loss": 1.2487,
"step": 3320
},
{
"epoch": 0.8885622661678247,
"grad_norm": 1.1503659735195155,
"learning_rate": 7.474288506192662e-07,
"loss": 1.2898,
"step": 3325
},
{
"epoch": 0.8898984500267236,
"grad_norm": 1.205235234079246,
"learning_rate": 7.298323434629095e-07,
"loss": 1.3094,
"step": 3330
},
{
"epoch": 0.8912346338856226,
"grad_norm": 1.1557951328425964,
"learning_rate": 7.124375987469767e-07,
"loss": 1.2326,
"step": 3335
},
{
"epoch": 0.8925708177445216,
"grad_norm": 1.1334618687477662,
"learning_rate": 6.952449950629103e-07,
"loss": 1.2679,
"step": 3340
},
{
"epoch": 0.8939070016034206,
"grad_norm": 1.14695924636925,
"learning_rate": 6.782549066026145e-07,
"loss": 1.3017,
"step": 3345
},
{
"epoch": 0.8952431854623196,
"grad_norm": 1.147346259057884,
"learning_rate": 6.614677031503059e-07,
"loss": 1.2705,
"step": 3350
},
{
"epoch": 0.8965793693212186,
"grad_norm": 1.1656452627988076,
"learning_rate": 6.448837500744742e-07,
"loss": 1.2475,
"step": 3355
},
{
"epoch": 0.8979155531801176,
"grad_norm": 1.1415605090418526,
"learning_rate": 6.285034083199216e-07,
"loss": 1.273,
"step": 3360
},
{
"epoch": 0.8992517370390166,
"grad_norm": 1.1324906184307248,
"learning_rate": 6.123270343999132e-07,
"loss": 1.2817,
"step": 3365
},
{
"epoch": 0.9005879208979155,
"grad_norm": 1.1353501329617395,
"learning_rate": 5.963549803884128e-07,
"loss": 1.2734,
"step": 3370
},
{
"epoch": 0.9019241047568145,
"grad_norm": 1.2284321012966415,
"learning_rate": 5.80587593912425e-07,
"loss": 1.276,
"step": 3375
},
{
"epoch": 0.9032602886157135,
"grad_norm": 1.1122913016411495,
"learning_rate": 5.650252181444215e-07,
"loss": 1.2524,
"step": 3380
},
{
"epoch": 0.9045964724746125,
"grad_norm": 1.1797725331714115,
"learning_rate": 5.496681917948809e-07,
"loss": 1.2336,
"step": 3385
},
{
"epoch": 0.9059326563335115,
"grad_norm": 1.159426377293814,
"learning_rate": 5.345168491049124e-07,
"loss": 1.242,
"step": 3390
},
{
"epoch": 0.9072688401924105,
"grad_norm": 1.1645037741741007,
"learning_rate": 5.195715198389784e-07,
"loss": 1.3,
"step": 3395
},
{
"epoch": 0.9086050240513095,
"grad_norm": 1.1480424103028632,
"learning_rate": 5.048325292777279e-07,
"loss": 1.2675,
"step": 3400
},
{
"epoch": 0.9086050240513095,
"eval_loss": 1.2765072584152222,
"eval_runtime": 523.9483,
"eval_samples_per_second": 25.281,
"eval_steps_per_second": 3.161,
"step": 3400
},
{
"epoch": 0.9099412079102085,
"grad_norm": 1.1694724018883438,
"learning_rate": 4.903001982109002e-07,
"loss": 1.2881,
"step": 3405
},
{
"epoch": 0.9112773917691074,
"grad_norm": 1.149736399066079,
"learning_rate": 4.759748429303579e-07,
"loss": 1.2993,
"step": 3410
},
{
"epoch": 0.9126135756280064,
"grad_norm": 1.1394564788621853,
"learning_rate": 4.618567752231962e-07,
"loss": 1.2486,
"step": 3415
},
{
"epoch": 0.9139497594869054,
"grad_norm": 1.1538669258463885,
"learning_rate": 4.479463023649555e-07,
"loss": 1.2618,
"step": 3420
},
{
"epoch": 0.9152859433458044,
"grad_norm": 1.1791184429266262,
"learning_rate": 4.342437271129396e-07,
"loss": 1.2583,
"step": 3425
},
{
"epoch": 0.9166221272047034,
"grad_norm": 1.1326513415617325,
"learning_rate": 4.207493476996205e-07,
"loss": 1.2408,
"step": 3430
},
{
"epoch": 0.9179583110636024,
"grad_norm": 1.1222236287232992,
"learning_rate": 4.074634578261516e-07,
"loss": 1.2379,
"step": 3435
},
{
"epoch": 0.9192944949225014,
"grad_norm": 1.1253440717297474,
"learning_rate": 3.9438634665597165e-07,
"loss": 1.2616,
"step": 3440
},
{
"epoch": 0.9206306787814004,
"grad_norm": 1.1510397092293756,
"learning_rate": 3.815182988085153e-07,
"loss": 1.2685,
"step": 3445
},
{
"epoch": 0.9219668626402993,
"grad_norm": 1.1879818159354434,
"learning_rate": 3.6885959435301156e-07,
"loss": 1.2365,
"step": 3450
},
{
"epoch": 0.9233030464991983,
"grad_norm": 1.1515610393737254,
"learning_rate": 3.564105088023984e-07,
"loss": 1.2501,
"step": 3455
},
{
"epoch": 0.9246392303580973,
"grad_norm": 1.1434666832922344,
"learning_rate": 3.441713131073177e-07,
"loss": 1.2604,
"step": 3460
},
{
"epoch": 0.9259754142169963,
"grad_norm": 1.2537303724295639,
"learning_rate": 3.3214227365022e-07,
"loss": 1.2954,
"step": 3465
},
{
"epoch": 0.9273115980758952,
"grad_norm": 1.1324300056445142,
"learning_rate": 3.2032365223957253e-07,
"loss": 1.2553,
"step": 3470
},
{
"epoch": 0.9286477819347942,
"grad_norm": 1.1623455952898252,
"learning_rate": 3.0871570610415124e-07,
"loss": 1.2578,
"step": 3475
},
{
"epoch": 0.9299839657936932,
"grad_norm": 1.1816451537266295,
"learning_rate": 2.97318687887449e-07,
"loss": 1.2512,
"step": 3480
},
{
"epoch": 0.9313201496525922,
"grad_norm": 1.1900717072648277,
"learning_rate": 2.861328456421775e-07,
"loss": 1.275,
"step": 3485
},
{
"epoch": 0.9326563335114911,
"grad_norm": 1.1275013968941647,
"learning_rate": 2.7515842282486274e-07,
"loss": 1.2653,
"step": 3490
},
{
"epoch": 0.9339925173703901,
"grad_norm": 1.1553817658383019,
"learning_rate": 2.6439565829055267e-07,
"loss": 1.2572,
"step": 3495
},
{
"epoch": 0.9353287012292891,
"grad_norm": 1.1368557006469668,
"learning_rate": 2.5384478628761586e-07,
"loss": 1.263,
"step": 3500
},
{
"epoch": 0.9366648850881881,
"grad_norm": 1.1505999909872684,
"learning_rate": 2.435060364526387e-07,
"loss": 1.2504,
"step": 3505
},
{
"epoch": 0.9380010689470871,
"grad_norm": 1.158821621163466,
"learning_rate": 2.3337963380543726e-07,
"loss": 1.2609,
"step": 3510
},
{
"epoch": 0.9393372528059861,
"grad_norm": 1.1699141096438583,
"learning_rate": 2.234657987441502e-07,
"loss": 1.2396,
"step": 3515
},
{
"epoch": 0.9406734366648851,
"grad_norm": 1.1460631269115045,
"learning_rate": 2.1376474704044693e-07,
"loss": 1.2503,
"step": 3520
},
{
"epoch": 0.9420096205237841,
"grad_norm": 1.1207186839548462,
"learning_rate": 2.0427668983483361e-07,
"loss": 1.246,
"step": 3525
},
{
"epoch": 0.943345804382683,
"grad_norm": 1.1268601862855379,
"learning_rate": 1.9500183363205029e-07,
"loss": 1.2367,
"step": 3530
},
{
"epoch": 0.944681988241582,
"grad_norm": 1.1458026011683493,
"learning_rate": 1.85940380296582e-07,
"loss": 1.2363,
"step": 3535
},
{
"epoch": 0.946018172100481,
"grad_norm": 1.126828597537144,
"learning_rate": 1.7709252704826485e-07,
"loss": 1.2771,
"step": 3540
},
{
"epoch": 0.94735435595938,
"grad_norm": 1.159174309108889,
"learning_rate": 1.6845846645799025e-07,
"loss": 1.2846,
"step": 3545
},
{
"epoch": 0.948690539818279,
"grad_norm": 1.144716451177437,
"learning_rate": 1.6003838644351843e-07,
"loss": 1.2711,
"step": 3550
},
{
"epoch": 0.950026723677178,
"grad_norm": 1.2115861792213716,
"learning_rate": 1.5183247026538505e-07,
"loss": 1.2658,
"step": 3555
},
{
"epoch": 0.951362907536077,
"grad_norm": 1.1183503620411726,
"learning_rate": 1.4384089652291544e-07,
"loss": 1.2687,
"step": 3560
},
{
"epoch": 0.952699091394976,
"grad_norm": 1.1482250830428042,
"learning_rate": 1.3606383915033217e-07,
"loss": 1.2471,
"step": 3565
},
{
"epoch": 0.9540352752538749,
"grad_norm": 1.1465209714509406,
"learning_rate": 1.2850146741297586e-07,
"loss": 1.2776,
"step": 3570
},
{
"epoch": 0.9553714591127739,
"grad_norm": 1.1496013307054171,
"learning_rate": 1.2115394590361595e-07,
"loss": 1.2509,
"step": 3575
},
{
"epoch": 0.9567076429716729,
"grad_norm": 1.1607454473870606,
"learning_rate": 1.1402143453887238e-07,
"loss": 1.2887,
"step": 3580
},
{
"epoch": 0.9580438268305719,
"grad_norm": 1.169980738830098,
"learning_rate": 1.0710408855573173e-07,
"loss": 1.2912,
"step": 3585
},
{
"epoch": 0.9593800106894709,
"grad_norm": 1.131820961749528,
"learning_rate": 1.0040205850817109e-07,
"loss": 1.2405,
"step": 3590
},
{
"epoch": 0.9607161945483699,
"grad_norm": 1.1643105760459518,
"learning_rate": 9.391549026387948e-08,
"loss": 1.2845,
"step": 3595
},
{
"epoch": 0.9620523784072689,
"grad_norm": 1.1545962184900476,
"learning_rate": 8.764452500108711e-08,
"loss": 1.2978,
"step": 3600
},
{
"epoch": 0.9620523784072689,
"eval_loss": 1.275867223739624,
"eval_runtime": 523.9419,
"eval_samples_per_second": 25.281,
"eval_steps_per_second": 3.161,
"step": 3600
},
{
"epoch": 0.9633885622661679,
"grad_norm": 1.1593993216116114,
"learning_rate": 8.158929920548342e-08,
"loss": 1.2391,
"step": 3605
},
{
"epoch": 0.9647247461250668,
"grad_norm": 1.1553562511912872,
"learning_rate": 7.574994466725827e-08,
"loss": 1.2495,
"step": 3610
},
{
"epoch": 0.9660609299839658,
"grad_norm": 1.177204123718693,
"learning_rate": 7.012658847822428e-08,
"loss": 1.2905,
"step": 3615
},
{
"epoch": 0.9673971138428648,
"grad_norm": 1.1296251792868242,
"learning_rate": 6.471935302905574e-08,
"loss": 1.2825,
"step": 3620
},
{
"epoch": 0.9687332977017638,
"grad_norm": 1.1818608486775184,
"learning_rate": 5.952835600662288e-08,
"loss": 1.2327,
"step": 3625
},
{
"epoch": 0.9700694815606627,
"grad_norm": 1.1499671622397054,
"learning_rate": 5.455371039143176e-08,
"loss": 1.2795,
"step": 3630
},
{
"epoch": 0.9714056654195617,
"grad_norm": 1.1349247593874663,
"learning_rate": 4.9795524455164e-08,
"loss": 1.2644,
"step": 3635
},
{
"epoch": 0.9727418492784607,
"grad_norm": 1.1311236324206604,
"learning_rate": 4.52539017583209e-08,
"loss": 1.2385,
"step": 3640
},
{
"epoch": 0.9740780331373597,
"grad_norm": 1.1732868443594786,
"learning_rate": 4.0928941147966306e-08,
"loss": 1.2908,
"step": 3645
},
{
"epoch": 0.9754142169962586,
"grad_norm": 1.1619085838301122,
"learning_rate": 3.682073675558395e-08,
"loss": 1.278,
"step": 3650
},
{
"epoch": 0.9767504008551576,
"grad_norm": 1.1539440287056817,
"learning_rate": 3.2929377995019054e-08,
"loss": 1.2386,
"step": 3655
},
{
"epoch": 0.9780865847140566,
"grad_norm": 1.1170137832632987,
"learning_rate": 2.9254949560535428e-08,
"loss": 1.2589,
"step": 3660
},
{
"epoch": 0.9794227685729556,
"grad_norm": 1.1423040789553527,
"learning_rate": 2.5797531424976983e-08,
"loss": 1.2896,
"step": 3665
},
{
"epoch": 0.9807589524318546,
"grad_norm": 1.1680612690612637,
"learning_rate": 2.2557198838019102e-08,
"loss": 1.2404,
"step": 3670
},
{
"epoch": 0.9820951362907536,
"grad_norm": 1.114027530261495,
"learning_rate": 1.9534022324536606e-08,
"loss": 1.2802,
"step": 3675
},
{
"epoch": 0.9834313201496526,
"grad_norm": 1.1636519127752734,
"learning_rate": 1.6728067683066117e-08,
"loss": 1.2701,
"step": 3680
},
{
"epoch": 0.9847675040085516,
"grad_norm": 1.1842605033839753,
"learning_rate": 1.4139395984377191e-08,
"loss": 1.2858,
"step": 3685
},
{
"epoch": 0.9861036878674505,
"grad_norm": 1.1431541821634774,
"learning_rate": 1.1768063570136712e-08,
"loss": 1.2662,
"step": 3690
},
{
"epoch": 0.9874398717263495,
"grad_norm": 1.1466055572304683,
"learning_rate": 9.614122051689878e-09,
"loss": 1.2773,
"step": 3695
},
{
"epoch": 0.9887760555852485,
"grad_norm": 1.180031565283721,
"learning_rate": 7.67761830893443e-09,
"loss": 1.2449,
"step": 3700
},
{
"epoch": 0.9901122394441475,
"grad_norm": 1.1492882566559037,
"learning_rate": 5.958594489295921e-09,
"loss": 1.2521,
"step": 3705
},
{
"epoch": 0.9914484233030465,
"grad_norm": 1.1745365142675883,
"learning_rate": 4.457088006816213e-09,
"loss": 1.2739,
"step": 3710
},
{
"epoch": 0.9927846071619455,
"grad_norm": 1.1360910701337543,
"learning_rate": 3.173131541338581e-09,
"loss": 1.2512,
"step": 3715
},
{
"epoch": 0.9941207910208445,
"grad_norm": 1.1874182237790705,
"learning_rate": 2.1067530377927305e-09,
"loss": 1.2587,
"step": 3720
},
{
"epoch": 0.9954569748797435,
"grad_norm": 1.1682623889051527,
"learning_rate": 1.2579757055897202e-09,
"loss": 1.2834,
"step": 3725
},
{
"epoch": 0.9967931587386424,
"grad_norm": 1.1413958118113652,
"learning_rate": 6.268180181157047e-10,
"loss": 1.2642,
"step": 3730
},
{
"epoch": 0.9981293425975414,
"grad_norm": 1.1586136947062726,
"learning_rate": 2.1329371232892138e-10,
"loss": 1.211,
"step": 3735
},
{
"epoch": 0.9994655264564404,
"grad_norm": 1.196251896759201,
"learning_rate": 1.7411788463261858e-11,
"loss": 1.2948,
"step": 3740
},
{
"epoch": 1.0,
"step": 3742,
"total_flos": 99431605075968.0,
"train_loss": 1.3181866798281097,
"train_runtime": 27158.0723,
"train_samples_per_second": 4.409,
"train_steps_per_second": 0.138
}
],
"logging_steps": 5,
"max_steps": 3742,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 1000000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 99431605075968.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}