kdcyberdude's picture
Training in progress, step 10500, checkpoint
79084e4 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.2734113930093481,
"eval_steps": 500,
"global_step": 10500,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0002603918028660458,
"grad_norm": 0.4500846266746521,
"learning_rate": 5.194805194805195e-06,
"loss": 1.0381,
"step": 10
},
{
"epoch": 0.0005207836057320916,
"grad_norm": 0.35188010334968567,
"learning_rate": 1.038961038961039e-05,
"loss": 1.0108,
"step": 20
},
{
"epoch": 0.0007811754085981374,
"grad_norm": 0.2300374060869217,
"learning_rate": 1.5584415584415583e-05,
"loss": 0.9668,
"step": 30
},
{
"epoch": 0.0010415672114641832,
"grad_norm": 0.16189467906951904,
"learning_rate": 2.077922077922078e-05,
"loss": 0.918,
"step": 40
},
{
"epoch": 0.001301959014330229,
"grad_norm": 0.18843211233615875,
"learning_rate": 2.5974025974025972e-05,
"loss": 0.9265,
"step": 50
},
{
"epoch": 0.0015623508171962747,
"grad_norm": 0.20334510505199432,
"learning_rate": 3.1168831168831166e-05,
"loss": 0.9234,
"step": 60
},
{
"epoch": 0.0018227426200623205,
"grad_norm": 0.1745327115058899,
"learning_rate": 3.6363636363636364e-05,
"loss": 0.881,
"step": 70
},
{
"epoch": 0.0020831344229283663,
"grad_norm": 0.18667331337928772,
"learning_rate": 4.155844155844156e-05,
"loss": 0.8592,
"step": 80
},
{
"epoch": 0.002343526225794412,
"grad_norm": 0.1848158985376358,
"learning_rate": 4.675324675324675e-05,
"loss": 0.8537,
"step": 90
},
{
"epoch": 0.002603918028660458,
"grad_norm": 0.17589879035949707,
"learning_rate": 5.1948051948051944e-05,
"loss": 0.8518,
"step": 100
},
{
"epoch": 0.0028643098315265037,
"grad_norm": 0.2132624089717865,
"learning_rate": 5.714285714285714e-05,
"loss": 0.8511,
"step": 110
},
{
"epoch": 0.0031247016343925495,
"grad_norm": 0.23070092499256134,
"learning_rate": 6.233766233766233e-05,
"loss": 0.7975,
"step": 120
},
{
"epoch": 0.0033850934372585953,
"grad_norm": 0.25368157029151917,
"learning_rate": 6.753246753246754e-05,
"loss": 0.8134,
"step": 130
},
{
"epoch": 0.003645485240124641,
"grad_norm": 0.22897231578826904,
"learning_rate": 7.272727272727273e-05,
"loss": 0.8322,
"step": 140
},
{
"epoch": 0.003905877042990687,
"grad_norm": 0.19932536780834198,
"learning_rate": 7.792207792207793e-05,
"loss": 0.7959,
"step": 150
},
{
"epoch": 0.004166268845856733,
"grad_norm": 0.21011792123317719,
"learning_rate": 8.311688311688312e-05,
"loss": 0.8102,
"step": 160
},
{
"epoch": 0.004426660648722778,
"grad_norm": 0.20594824850559235,
"learning_rate": 8.831168831168831e-05,
"loss": 0.8128,
"step": 170
},
{
"epoch": 0.004687052451588824,
"grad_norm": 0.20465536415576935,
"learning_rate": 9.35064935064935e-05,
"loss": 0.7989,
"step": 180
},
{
"epoch": 0.00494744425445487,
"grad_norm": 0.4109392762184143,
"learning_rate": 9.870129870129871e-05,
"loss": 0.8108,
"step": 190
},
{
"epoch": 0.005207836057320916,
"grad_norm": 0.4293076694011688,
"learning_rate": 0.00010389610389610389,
"loss": 0.8101,
"step": 200
},
{
"epoch": 0.005468227860186962,
"grad_norm": 0.31628963351249695,
"learning_rate": 0.00010909090909090909,
"loss": 0.7989,
"step": 210
},
{
"epoch": 0.005728619663053007,
"grad_norm": 0.24642810225486755,
"learning_rate": 0.00011428571428571428,
"loss": 0.7751,
"step": 220
},
{
"epoch": 0.005989011465919053,
"grad_norm": 0.3599106967449188,
"learning_rate": 0.00011948051948051949,
"loss": 0.8063,
"step": 230
},
{
"epoch": 0.006249403268785099,
"grad_norm": 0.17053447663784027,
"learning_rate": 0.00012467532467532467,
"loss": 0.7751,
"step": 240
},
{
"epoch": 0.006509795071651145,
"grad_norm": 0.17303769290447235,
"learning_rate": 0.00012987012987012987,
"loss": 0.7883,
"step": 250
},
{
"epoch": 0.0067701868745171905,
"grad_norm": 0.1815861016511917,
"learning_rate": 0.00013506493506493507,
"loss": 0.788,
"step": 260
},
{
"epoch": 0.007030578677383236,
"grad_norm": 0.24125365912914276,
"learning_rate": 0.00014025974025974028,
"loss": 0.8018,
"step": 270
},
{
"epoch": 0.007290970480249282,
"grad_norm": 0.19443446397781372,
"learning_rate": 0.00014545454545454546,
"loss": 0.7908,
"step": 280
},
{
"epoch": 0.007551362283115328,
"grad_norm": 0.17829768359661102,
"learning_rate": 0.00015064935064935066,
"loss": 0.8033,
"step": 290
},
{
"epoch": 0.007811754085981374,
"grad_norm": 0.19535653293132782,
"learning_rate": 0.00015584415584415587,
"loss": 0.7997,
"step": 300
},
{
"epoch": 0.008072145888847419,
"grad_norm": 0.19930541515350342,
"learning_rate": 0.00016103896103896104,
"loss": 0.7945,
"step": 310
},
{
"epoch": 0.008332537691713465,
"grad_norm": 0.2156297266483307,
"learning_rate": 0.00016623376623376625,
"loss": 0.8018,
"step": 320
},
{
"epoch": 0.00859292949457951,
"grad_norm": 0.1924206018447876,
"learning_rate": 0.00017142857142857143,
"loss": 0.7746,
"step": 330
},
{
"epoch": 0.008853321297445557,
"grad_norm": 0.2294880747795105,
"learning_rate": 0.00017662337662337663,
"loss": 0.8152,
"step": 340
},
{
"epoch": 0.009113713100311602,
"grad_norm": 0.16817067563533783,
"learning_rate": 0.00018181818181818183,
"loss": 0.7972,
"step": 350
},
{
"epoch": 0.009374104903177648,
"grad_norm": 0.18544812500476837,
"learning_rate": 0.000187012987012987,
"loss": 0.7801,
"step": 360
},
{
"epoch": 0.009634496706043693,
"grad_norm": 0.19597066938877106,
"learning_rate": 0.00019220779220779222,
"loss": 0.7706,
"step": 370
},
{
"epoch": 0.00989488850890974,
"grad_norm": 0.40291881561279297,
"learning_rate": 0.00019740259740259742,
"loss": 0.7911,
"step": 380
},
{
"epoch": 0.010155280311775785,
"grad_norm": 0.23841074109077454,
"learning_rate": 0.00019999996515752773,
"loss": 0.7861,
"step": 390
},
{
"epoch": 0.010415672114641832,
"grad_norm": 0.1675388514995575,
"learning_rate": 0.00019999968641789507,
"loss": 0.788,
"step": 400
},
{
"epoch": 0.010676063917507876,
"grad_norm": 1.8860758543014526,
"learning_rate": 0.0001999991289394067,
"loss": 0.7632,
"step": 410
},
{
"epoch": 0.010936455720373923,
"grad_norm": 0.17022117972373962,
"learning_rate": 0.00019999829272361654,
"loss": 0.784,
"step": 420
},
{
"epoch": 0.011196847523239968,
"grad_norm": 0.21460269391536713,
"learning_rate": 0.00019999717777285545,
"loss": 0.761,
"step": 430
},
{
"epoch": 0.011457239326106015,
"grad_norm": 0.19413785636425018,
"learning_rate": 0.00019999578409023126,
"loss": 0.7772,
"step": 440
},
{
"epoch": 0.01171763112897206,
"grad_norm": 0.20223405957221985,
"learning_rate": 0.00019999411167962868,
"loss": 0.7811,
"step": 450
},
{
"epoch": 0.011978022931838106,
"grad_norm": 0.15166303515434265,
"learning_rate": 0.00019999216054570942,
"loss": 0.7709,
"step": 460
},
{
"epoch": 0.012238414734704151,
"grad_norm": 0.16307081282138824,
"learning_rate": 0.00019998993069391205,
"loss": 0.7811,
"step": 470
},
{
"epoch": 0.012498806537570198,
"grad_norm": 0.15996049344539642,
"learning_rate": 0.00019998742213045206,
"loss": 0.7599,
"step": 480
},
{
"epoch": 0.012759198340436243,
"grad_norm": 0.17560279369354248,
"learning_rate": 0.00019998463486232179,
"loss": 0.7572,
"step": 490
},
{
"epoch": 0.01301959014330229,
"grad_norm": 0.17571642994880676,
"learning_rate": 0.0001999815688972905,
"loss": 0.7643,
"step": 500
},
{
"epoch": 0.013279981946168334,
"grad_norm": 0.17719799280166626,
"learning_rate": 0.00019997822424390422,
"loss": 0.7923,
"step": 510
},
{
"epoch": 0.013540373749034381,
"grad_norm": 0.19846616685390472,
"learning_rate": 0.00019997460091148586,
"loss": 0.7674,
"step": 520
},
{
"epoch": 0.013800765551900426,
"grad_norm": 0.2715558111667633,
"learning_rate": 0.00019997069891013503,
"loss": 0.7421,
"step": 530
},
{
"epoch": 0.014061157354766473,
"grad_norm": 0.1725197583436966,
"learning_rate": 0.00019996651825072826,
"loss": 0.7663,
"step": 540
},
{
"epoch": 0.014321549157632518,
"grad_norm": 0.15060502290725708,
"learning_rate": 0.00019996205894491856,
"loss": 0.7794,
"step": 550
},
{
"epoch": 0.014581940960498564,
"grad_norm": 0.16645808517932892,
"learning_rate": 0.00019995732100513592,
"loss": 0.752,
"step": 560
},
{
"epoch": 0.014842332763364609,
"grad_norm": 0.1736789345741272,
"learning_rate": 0.00019995230444458682,
"loss": 0.7788,
"step": 570
},
{
"epoch": 0.015102724566230656,
"grad_norm": 0.15416319668293,
"learning_rate": 0.0001999470092772544,
"loss": 0.7656,
"step": 580
},
{
"epoch": 0.0153631163690967,
"grad_norm": 0.16610187292099,
"learning_rate": 0.00019994143551789839,
"loss": 0.7676,
"step": 590
},
{
"epoch": 0.015623508171962747,
"grad_norm": 0.15843011438846588,
"learning_rate": 0.00019993558318205507,
"loss": 0.7746,
"step": 600
},
{
"epoch": 0.015883899974828794,
"grad_norm": 0.26837801933288574,
"learning_rate": 0.00019992945228603724,
"loss": 0.7617,
"step": 610
},
{
"epoch": 0.016144291777694837,
"grad_norm": 0.15099173784255981,
"learning_rate": 0.0001999230428469341,
"loss": 0.7601,
"step": 620
},
{
"epoch": 0.016404683580560884,
"grad_norm": 0.15511856973171234,
"learning_rate": 0.00019991635488261138,
"loss": 0.7647,
"step": 630
},
{
"epoch": 0.01666507538342693,
"grad_norm": 0.14919579029083252,
"learning_rate": 0.00019990938841171104,
"loss": 0.7692,
"step": 640
},
{
"epoch": 0.016925467186292977,
"grad_norm": 0.15838642418384552,
"learning_rate": 0.0001999021434536514,
"loss": 0.7763,
"step": 650
},
{
"epoch": 0.01718585898915902,
"grad_norm": 0.15956635773181915,
"learning_rate": 0.00019989462002862704,
"loss": 0.7598,
"step": 660
},
{
"epoch": 0.017446250792025067,
"grad_norm": 0.1499069333076477,
"learning_rate": 0.0001998868181576088,
"loss": 0.7626,
"step": 670
},
{
"epoch": 0.017706642594891114,
"grad_norm": 0.2170073390007019,
"learning_rate": 0.00019987873786234348,
"loss": 0.7569,
"step": 680
},
{
"epoch": 0.01796703439775716,
"grad_norm": 0.17841948568820953,
"learning_rate": 0.00019987037916535417,
"loss": 0.7494,
"step": 690
},
{
"epoch": 0.018227426200623204,
"grad_norm": 0.2066909372806549,
"learning_rate": 0.0001998617420899398,
"loss": 0.7609,
"step": 700
},
{
"epoch": 0.01848781800348925,
"grad_norm": 0.17015361785888672,
"learning_rate": 0.0001998528266601754,
"loss": 0.7761,
"step": 710
},
{
"epoch": 0.018748209806355297,
"grad_norm": 0.22166290879249573,
"learning_rate": 0.0001998436329009118,
"loss": 0.7573,
"step": 720
},
{
"epoch": 0.01900860160922134,
"grad_norm": 0.15084640681743622,
"learning_rate": 0.00019983416083777563,
"loss": 0.7775,
"step": 730
},
{
"epoch": 0.019268993412087387,
"grad_norm": 0.17800921201705933,
"learning_rate": 0.0001998244104971693,
"loss": 0.7359,
"step": 740
},
{
"epoch": 0.019529385214953433,
"grad_norm": 0.17354707419872284,
"learning_rate": 0.0001998143819062709,
"loss": 0.7415,
"step": 750
},
{
"epoch": 0.01978977701781948,
"grad_norm": 0.16408118605613708,
"learning_rate": 0.00019980407509303413,
"loss": 0.7708,
"step": 760
},
{
"epoch": 0.020050168820685523,
"grad_norm": 0.16820089519023895,
"learning_rate": 0.00019979349008618808,
"loss": 0.791,
"step": 770
},
{
"epoch": 0.02031056062355157,
"grad_norm": 0.15958388149738312,
"learning_rate": 0.00019978262691523743,
"loss": 0.7412,
"step": 780
},
{
"epoch": 0.020570952426417616,
"grad_norm": 0.1646542251110077,
"learning_rate": 0.00019977148561046217,
"loss": 0.7529,
"step": 790
},
{
"epoch": 0.020831344229283663,
"grad_norm": 0.17032025754451752,
"learning_rate": 0.0001997600662029175,
"loss": 0.7656,
"step": 800
},
{
"epoch": 0.021091736032149706,
"grad_norm": 0.17189227044582367,
"learning_rate": 0.00019974836872443388,
"loss": 0.7433,
"step": 810
},
{
"epoch": 0.021352127835015753,
"grad_norm": 0.16334249079227448,
"learning_rate": 0.0001997363932076168,
"loss": 0.7703,
"step": 820
},
{
"epoch": 0.0216125196378818,
"grad_norm": 0.1676424890756607,
"learning_rate": 0.00019972413968584682,
"loss": 0.7603,
"step": 830
},
{
"epoch": 0.021872911440747846,
"grad_norm": 0.16826209425926208,
"learning_rate": 0.0001997116081932793,
"loss": 0.7569,
"step": 840
},
{
"epoch": 0.02213330324361389,
"grad_norm": 0.1876436173915863,
"learning_rate": 0.0001996987987648446,
"loss": 0.7553,
"step": 850
},
{
"epoch": 0.022393695046479936,
"grad_norm": 0.17252250015735626,
"learning_rate": 0.0001996857114362476,
"loss": 0.7644,
"step": 860
},
{
"epoch": 0.022654086849345983,
"grad_norm": 0.1632252335548401,
"learning_rate": 0.00019967234624396793,
"loss": 0.7568,
"step": 870
},
{
"epoch": 0.02291447865221203,
"grad_norm": 0.1818259060382843,
"learning_rate": 0.00019965870322525965,
"loss": 0.7672,
"step": 880
},
{
"epoch": 0.023174870455078073,
"grad_norm": 0.15418195724487305,
"learning_rate": 0.0001996447824181513,
"loss": 0.7642,
"step": 890
},
{
"epoch": 0.02343526225794412,
"grad_norm": 0.17383505403995514,
"learning_rate": 0.0001996305838614457,
"loss": 0.7607,
"step": 900
},
{
"epoch": 0.023695654060810166,
"grad_norm": 0.17794272303581238,
"learning_rate": 0.00019961610759471984,
"loss": 0.7588,
"step": 910
},
{
"epoch": 0.023956045863676213,
"grad_norm": 0.1909121572971344,
"learning_rate": 0.00019960135365832486,
"loss": 0.7438,
"step": 920
},
{
"epoch": 0.024216437666542256,
"grad_norm": 0.17758873105049133,
"learning_rate": 0.00019958632209338587,
"loss": 0.7323,
"step": 930
},
{
"epoch": 0.024476829469408302,
"grad_norm": 0.15553662180900574,
"learning_rate": 0.00019957101294180174,
"loss": 0.7508,
"step": 940
},
{
"epoch": 0.02473722127227435,
"grad_norm": 0.15310749411582947,
"learning_rate": 0.00019955542624624522,
"loss": 0.7563,
"step": 950
},
{
"epoch": 0.024997613075140396,
"grad_norm": 0.1628728210926056,
"learning_rate": 0.00019953956205016256,
"loss": 0.7524,
"step": 960
},
{
"epoch": 0.02525800487800644,
"grad_norm": 0.16211454570293427,
"learning_rate": 0.00019952342039777362,
"loss": 0.7564,
"step": 970
},
{
"epoch": 0.025518396680872486,
"grad_norm": 0.15663012862205505,
"learning_rate": 0.00019950700133407163,
"loss": 0.7395,
"step": 980
},
{
"epoch": 0.025778788483738532,
"grad_norm": 0.1684863567352295,
"learning_rate": 0.00019949030490482296,
"loss": 0.753,
"step": 990
},
{
"epoch": 0.02603918028660458,
"grad_norm": 0.1561436653137207,
"learning_rate": 0.0001994733311565673,
"loss": 0.7409,
"step": 1000
},
{
"epoch": 0.026299572089470622,
"grad_norm": 0.1781485229730606,
"learning_rate": 0.0001994560801366171,
"loss": 0.762,
"step": 1010
},
{
"epoch": 0.02655996389233667,
"grad_norm": 0.15422071516513824,
"learning_rate": 0.00019943855189305792,
"loss": 0.7291,
"step": 1020
},
{
"epoch": 0.026820355695202715,
"grad_norm": 0.17980527877807617,
"learning_rate": 0.00019942074647474786,
"loss": 0.7732,
"step": 1030
},
{
"epoch": 0.027080747498068762,
"grad_norm": 0.15810626745224,
"learning_rate": 0.00019940266393131775,
"loss": 0.7764,
"step": 1040
},
{
"epoch": 0.027341139300934805,
"grad_norm": 0.16385480761528015,
"learning_rate": 0.00019938430431317081,
"loss": 0.7404,
"step": 1050
},
{
"epoch": 0.027601531103800852,
"grad_norm": 0.15134255588054657,
"learning_rate": 0.00019936566767148257,
"loss": 0.7506,
"step": 1060
},
{
"epoch": 0.0278619229066669,
"grad_norm": 0.1592187136411667,
"learning_rate": 0.00019934675405820077,
"loss": 0.73,
"step": 1070
},
{
"epoch": 0.028122314709532945,
"grad_norm": 0.16852422058582306,
"learning_rate": 0.00019932756352604515,
"loss": 0.7443,
"step": 1080
},
{
"epoch": 0.02838270651239899,
"grad_norm": 0.15741507709026337,
"learning_rate": 0.00019930809612850735,
"loss": 0.7377,
"step": 1090
},
{
"epoch": 0.028643098315265035,
"grad_norm": 0.22424879670143127,
"learning_rate": 0.00019928835191985076,
"loss": 0.7544,
"step": 1100
},
{
"epoch": 0.028903490118131082,
"grad_norm": 0.2047310769557953,
"learning_rate": 0.0001992683309551103,
"loss": 0.7441,
"step": 1110
},
{
"epoch": 0.02916388192099713,
"grad_norm": 0.16392463445663452,
"learning_rate": 0.00019924803329009243,
"loss": 0.7606,
"step": 1120
},
{
"epoch": 0.02942427372386317,
"grad_norm": 0.16227149963378906,
"learning_rate": 0.00019922745898137473,
"loss": 0.736,
"step": 1130
},
{
"epoch": 0.029684665526729218,
"grad_norm": 0.15652808547019958,
"learning_rate": 0.00019920660808630598,
"loss": 0.7513,
"step": 1140
},
{
"epoch": 0.029945057329595265,
"grad_norm": 0.15162768959999084,
"learning_rate": 0.00019918548066300592,
"loss": 0.7303,
"step": 1150
},
{
"epoch": 0.03020544913246131,
"grad_norm": 0.17650415003299713,
"learning_rate": 0.0001991640767703651,
"loss": 0.7254,
"step": 1160
},
{
"epoch": 0.030465840935327355,
"grad_norm": 0.1594468355178833,
"learning_rate": 0.00019914239646804462,
"loss": 0.741,
"step": 1170
},
{
"epoch": 0.0307262327381934,
"grad_norm": 0.17928367853164673,
"learning_rate": 0.00019912043981647616,
"loss": 0.7515,
"step": 1180
},
{
"epoch": 0.030986624541059448,
"grad_norm": 0.17009998857975006,
"learning_rate": 0.00019909820687686157,
"loss": 0.7539,
"step": 1190
},
{
"epoch": 0.031247016343925495,
"grad_norm": 0.16556763648986816,
"learning_rate": 0.0001990756977111729,
"loss": 0.7418,
"step": 1200
},
{
"epoch": 0.03150740814679154,
"grad_norm": 0.1561640352010727,
"learning_rate": 0.0001990529123821522,
"loss": 0.7465,
"step": 1210
},
{
"epoch": 0.03176779994965759,
"grad_norm": 0.15182287991046906,
"learning_rate": 0.00019902985095331113,
"loss": 0.7694,
"step": 1220
},
{
"epoch": 0.03202819175252363,
"grad_norm": 0.15173685550689697,
"learning_rate": 0.00019900651348893114,
"loss": 0.7519,
"step": 1230
},
{
"epoch": 0.032288583555389674,
"grad_norm": 0.16535787284374237,
"learning_rate": 0.00019898290005406296,
"loss": 0.7646,
"step": 1240
},
{
"epoch": 0.032548975358255725,
"grad_norm": 0.19272534549236298,
"learning_rate": 0.00019895901071452667,
"loss": 0.7655,
"step": 1250
},
{
"epoch": 0.03280936716112177,
"grad_norm": 0.1672705113887787,
"learning_rate": 0.0001989348455369113,
"loss": 0.7486,
"step": 1260
},
{
"epoch": 0.03306975896398781,
"grad_norm": 0.1525493860244751,
"learning_rate": 0.0001989104045885748,
"loss": 0.7546,
"step": 1270
},
{
"epoch": 0.03333015076685386,
"grad_norm": 0.16333037614822388,
"learning_rate": 0.00019888568793764385,
"loss": 0.7299,
"step": 1280
},
{
"epoch": 0.033590542569719904,
"grad_norm": 0.1590205729007721,
"learning_rate": 0.00019886069565301355,
"loss": 0.762,
"step": 1290
},
{
"epoch": 0.033850934372585954,
"grad_norm": 0.15006420016288757,
"learning_rate": 0.00019883542780434733,
"loss": 0.7531,
"step": 1300
},
{
"epoch": 0.034111326175452,
"grad_norm": 0.18390792608261108,
"learning_rate": 0.0001988098844620767,
"loss": 0.7621,
"step": 1310
},
{
"epoch": 0.03437171797831804,
"grad_norm": 0.17046166956424713,
"learning_rate": 0.0001987840656974011,
"loss": 0.7422,
"step": 1320
},
{
"epoch": 0.03463210978118409,
"grad_norm": 0.15121813118457794,
"learning_rate": 0.00019875797158228775,
"loss": 0.7555,
"step": 1330
},
{
"epoch": 0.034892501584050134,
"grad_norm": 0.16219307482242584,
"learning_rate": 0.00019873160218947125,
"loss": 0.7301,
"step": 1340
},
{
"epoch": 0.03515289338691618,
"grad_norm": 0.1779986321926117,
"learning_rate": 0.00019870495759245362,
"loss": 0.7356,
"step": 1350
},
{
"epoch": 0.03541328518978223,
"grad_norm": 0.16951359808444977,
"learning_rate": 0.0001986780378655039,
"loss": 0.7645,
"step": 1360
},
{
"epoch": 0.03567367699264827,
"grad_norm": 0.16620802879333496,
"learning_rate": 0.0001986508430836581,
"loss": 0.7331,
"step": 1370
},
{
"epoch": 0.03593406879551432,
"grad_norm": 0.1577858328819275,
"learning_rate": 0.0001986233733227188,
"loss": 0.7667,
"step": 1380
},
{
"epoch": 0.036194460598380364,
"grad_norm": 0.1637091338634491,
"learning_rate": 0.00019859562865925525,
"loss": 0.7521,
"step": 1390
},
{
"epoch": 0.03645485240124641,
"grad_norm": 0.15061691403388977,
"learning_rate": 0.00019856760917060277,
"loss": 0.744,
"step": 1400
},
{
"epoch": 0.03671524420411246,
"grad_norm": 0.15373477339744568,
"learning_rate": 0.00019853931493486287,
"loss": 0.7677,
"step": 1410
},
{
"epoch": 0.0369756360069785,
"grad_norm": 0.16468606889247894,
"learning_rate": 0.00019851074603090277,
"loss": 0.7179,
"step": 1420
},
{
"epoch": 0.037236027809844544,
"grad_norm": 0.16084876656532288,
"learning_rate": 0.00019848190253835536,
"loss": 0.749,
"step": 1430
},
{
"epoch": 0.037496419612710594,
"grad_norm": 0.16743004322052002,
"learning_rate": 0.00019845278453761896,
"loss": 0.7483,
"step": 1440
},
{
"epoch": 0.03775681141557664,
"grad_norm": 0.17335088551044464,
"learning_rate": 0.00019842339210985696,
"loss": 0.735,
"step": 1450
},
{
"epoch": 0.03801720321844268,
"grad_norm": 0.1546197235584259,
"learning_rate": 0.00019839372533699774,
"loss": 0.7549,
"step": 1460
},
{
"epoch": 0.03827759502130873,
"grad_norm": 0.16218656301498413,
"learning_rate": 0.00019836378430173438,
"loss": 0.7425,
"step": 1470
},
{
"epoch": 0.03853798682417477,
"grad_norm": 0.1712743639945984,
"learning_rate": 0.0001983335690875245,
"loss": 0.733,
"step": 1480
},
{
"epoch": 0.03879837862704082,
"grad_norm": 0.15490613877773285,
"learning_rate": 0.00019830307977858984,
"loss": 0.7265,
"step": 1490
},
{
"epoch": 0.03905877042990687,
"grad_norm": 0.1646670252084732,
"learning_rate": 0.00019827231645991623,
"loss": 0.7315,
"step": 1500
},
{
"epoch": 0.03931916223277291,
"grad_norm": 0.1599082201719284,
"learning_rate": 0.00019824127921725326,
"loss": 0.7293,
"step": 1510
},
{
"epoch": 0.03957955403563896,
"grad_norm": 0.1565747708082199,
"learning_rate": 0.00019820996813711407,
"loss": 0.7396,
"step": 1520
},
{
"epoch": 0.039839945838505,
"grad_norm": 0.154826357960701,
"learning_rate": 0.0001981783833067751,
"loss": 0.7217,
"step": 1530
},
{
"epoch": 0.040100337641371046,
"grad_norm": 0.16705222427845,
"learning_rate": 0.0001981465248142758,
"loss": 0.761,
"step": 1540
},
{
"epoch": 0.040360729444237096,
"grad_norm": 0.15651623904705048,
"learning_rate": 0.00019811439274841842,
"loss": 0.7565,
"step": 1550
},
{
"epoch": 0.04062112124710314,
"grad_norm": 0.16211090981960297,
"learning_rate": 0.00019808198719876782,
"loss": 0.7555,
"step": 1560
},
{
"epoch": 0.04088151304996919,
"grad_norm": 0.16856881976127625,
"learning_rate": 0.00019804930825565112,
"loss": 0.7567,
"step": 1570
},
{
"epoch": 0.04114190485283523,
"grad_norm": 0.1588718593120575,
"learning_rate": 0.00019801635601015752,
"loss": 0.729,
"step": 1580
},
{
"epoch": 0.041402296655701276,
"grad_norm": 0.17078711092472076,
"learning_rate": 0.00019798313055413808,
"loss": 0.7418,
"step": 1590
},
{
"epoch": 0.041662688458567326,
"grad_norm": 0.16652734577655792,
"learning_rate": 0.00019794963198020525,
"loss": 0.7341,
"step": 1600
},
{
"epoch": 0.04192308026143337,
"grad_norm": 0.15535488724708557,
"learning_rate": 0.00019791586038173296,
"loss": 0.7396,
"step": 1610
},
{
"epoch": 0.04218347206429941,
"grad_norm": 0.3506317734718323,
"learning_rate": 0.00019788181585285602,
"loss": 0.7345,
"step": 1620
},
{
"epoch": 0.04244386386716546,
"grad_norm": 0.16875872015953064,
"learning_rate": 0.00019784749848847003,
"loss": 0.7214,
"step": 1630
},
{
"epoch": 0.042704255670031506,
"grad_norm": 0.17675861716270447,
"learning_rate": 0.0001978129083842312,
"loss": 0.7431,
"step": 1640
},
{
"epoch": 0.042964647472897556,
"grad_norm": 0.15601837635040283,
"learning_rate": 0.00019777804563655583,
"loss": 0.7215,
"step": 1650
},
{
"epoch": 0.0432250392757636,
"grad_norm": 0.1874823123216629,
"learning_rate": 0.00019774291034262026,
"loss": 0.727,
"step": 1660
},
{
"epoch": 0.04348543107862964,
"grad_norm": 0.17005637288093567,
"learning_rate": 0.00019770750260036054,
"loss": 0.7446,
"step": 1670
},
{
"epoch": 0.04374582288149569,
"grad_norm": 0.17069579660892487,
"learning_rate": 0.00019767182250847207,
"loss": 0.7266,
"step": 1680
},
{
"epoch": 0.044006214684361736,
"grad_norm": 0.16133156418800354,
"learning_rate": 0.00019763587016640948,
"loss": 0.7568,
"step": 1690
},
{
"epoch": 0.04426660648722778,
"grad_norm": 0.16229428350925446,
"learning_rate": 0.00019759964567438623,
"loss": 0.7402,
"step": 1700
},
{
"epoch": 0.04452699829009383,
"grad_norm": 0.1622512936592102,
"learning_rate": 0.00019756314913337432,
"loss": 0.7536,
"step": 1710
},
{
"epoch": 0.04478739009295987,
"grad_norm": 0.2161218672990799,
"learning_rate": 0.00019752638064510415,
"loss": 0.723,
"step": 1720
},
{
"epoch": 0.04504778189582592,
"grad_norm": 0.154169961810112,
"learning_rate": 0.00019748934031206414,
"loss": 0.7441,
"step": 1730
},
{
"epoch": 0.045308173698691966,
"grad_norm": 0.15468057990074158,
"learning_rate": 0.00019745202823750034,
"loss": 0.7349,
"step": 1740
},
{
"epoch": 0.04556856550155801,
"grad_norm": 0.2015281468629837,
"learning_rate": 0.0001974144445254164,
"loss": 0.726,
"step": 1750
},
{
"epoch": 0.04582895730442406,
"grad_norm": 0.1931644082069397,
"learning_rate": 0.00019737658928057302,
"loss": 0.7604,
"step": 1760
},
{
"epoch": 0.0460893491072901,
"grad_norm": 0.1528482288122177,
"learning_rate": 0.00019733846260848776,
"loss": 0.7408,
"step": 1770
},
{
"epoch": 0.046349740910156145,
"grad_norm": 0.16370061039924622,
"learning_rate": 0.0001973000646154349,
"loss": 0.7647,
"step": 1780
},
{
"epoch": 0.046610132713022195,
"grad_norm": 0.16271348297595978,
"learning_rate": 0.00019726139540844484,
"loss": 0.7212,
"step": 1790
},
{
"epoch": 0.04687052451588824,
"grad_norm": 0.16218173503875732,
"learning_rate": 0.00019722245509530401,
"loss": 0.735,
"step": 1800
},
{
"epoch": 0.04713091631875429,
"grad_norm": 0.17063820362091064,
"learning_rate": 0.00019718324378455458,
"loss": 0.7311,
"step": 1810
},
{
"epoch": 0.04739130812162033,
"grad_norm": 0.1678459346294403,
"learning_rate": 0.00019714376158549404,
"loss": 0.7486,
"step": 1820
},
{
"epoch": 0.047651699924486375,
"grad_norm": 0.15926459431648254,
"learning_rate": 0.00019710400860817494,
"loss": 0.743,
"step": 1830
},
{
"epoch": 0.047912091727352425,
"grad_norm": 0.1775251179933548,
"learning_rate": 0.00019706398496340463,
"loss": 0.7512,
"step": 1840
},
{
"epoch": 0.04817248353021847,
"grad_norm": 0.1572408229112625,
"learning_rate": 0.00019702369076274494,
"loss": 0.733,
"step": 1850
},
{
"epoch": 0.04843287533308451,
"grad_norm": 0.29658186435699463,
"learning_rate": 0.0001969831261185118,
"loss": 0.7297,
"step": 1860
},
{
"epoch": 0.04869326713595056,
"grad_norm": 0.16520118713378906,
"learning_rate": 0.00019694229114377494,
"loss": 0.721,
"step": 1870
},
{
"epoch": 0.048953658938816605,
"grad_norm": 0.17762574553489685,
"learning_rate": 0.00019690118595235774,
"loss": 0.7304,
"step": 1880
},
{
"epoch": 0.049214050741682655,
"grad_norm": 0.16636615991592407,
"learning_rate": 0.00019685981065883663,
"loss": 0.7257,
"step": 1890
},
{
"epoch": 0.0494744425445487,
"grad_norm": 0.1622323989868164,
"learning_rate": 0.00019681816537854102,
"loss": 0.7353,
"step": 1900
},
{
"epoch": 0.04973483434741474,
"grad_norm": 0.17419832944869995,
"learning_rate": 0.00019677625022755289,
"loss": 0.7452,
"step": 1910
},
{
"epoch": 0.04999522615028079,
"grad_norm": 0.17460434138774872,
"learning_rate": 0.00019673406532270634,
"loss": 0.7391,
"step": 1920
},
{
"epoch": 0.050255617953146835,
"grad_norm": 0.15844550728797913,
"learning_rate": 0.00019669161078158753,
"loss": 0.7327,
"step": 1930
},
{
"epoch": 0.05051600975601288,
"grad_norm": 0.1638839989900589,
"learning_rate": 0.0001966488867225341,
"loss": 0.745,
"step": 1940
},
{
"epoch": 0.05077640155887893,
"grad_norm": 0.1587786227464676,
"learning_rate": 0.00019660589326463498,
"loss": 0.7476,
"step": 1950
},
{
"epoch": 0.05103679336174497,
"grad_norm": 0.15708380937576294,
"learning_rate": 0.00019656263052773002,
"loss": 0.7208,
"step": 1960
},
{
"epoch": 0.05129718516461102,
"grad_norm": 0.15816234052181244,
"learning_rate": 0.00019651909863240965,
"loss": 0.7262,
"step": 1970
},
{
"epoch": 0.051557576967477065,
"grad_norm": 0.16749270260334015,
"learning_rate": 0.00019647529770001456,
"loss": 0.7284,
"step": 1980
},
{
"epoch": 0.05181796877034311,
"grad_norm": 0.16943767666816711,
"learning_rate": 0.00019643122785263536,
"loss": 0.7225,
"step": 1990
},
{
"epoch": 0.05207836057320916,
"grad_norm": 0.42929205298423767,
"learning_rate": 0.00019638688921311224,
"loss": 0.7305,
"step": 2000
},
{
"epoch": 0.0523387523760752,
"grad_norm": 0.15851692855358124,
"learning_rate": 0.00019634228190503454,
"loss": 0.7344,
"step": 2010
},
{
"epoch": 0.052599144178941244,
"grad_norm": 0.16053883731365204,
"learning_rate": 0.00019629740605274062,
"loss": 0.7468,
"step": 2020
},
{
"epoch": 0.052859535981807294,
"grad_norm": 0.16504009068012238,
"learning_rate": 0.00019625226178131728,
"loss": 0.7375,
"step": 2030
},
{
"epoch": 0.05311992778467334,
"grad_norm": 0.1618044674396515,
"learning_rate": 0.00019620684921659953,
"loss": 0.7201,
"step": 2040
},
{
"epoch": 0.05338031958753939,
"grad_norm": 0.15512776374816895,
"learning_rate": 0.00019616116848517027,
"loss": 0.7355,
"step": 2050
},
{
"epoch": 0.05364071139040543,
"grad_norm": 0.17377036809921265,
"learning_rate": 0.00019611521971435979,
"loss": 0.7226,
"step": 2060
},
{
"epoch": 0.053901103193271474,
"grad_norm": 0.1685250997543335,
"learning_rate": 0.0001960690030322456,
"loss": 0.7483,
"step": 2070
},
{
"epoch": 0.054161494996137524,
"grad_norm": 0.18394522368907928,
"learning_rate": 0.00019602251856765194,
"loss": 0.7385,
"step": 2080
},
{
"epoch": 0.05442188679900357,
"grad_norm": 0.1753673106431961,
"learning_rate": 0.0001959757664501495,
"loss": 0.7378,
"step": 2090
},
{
"epoch": 0.05468227860186961,
"grad_norm": 0.1795465052127838,
"learning_rate": 0.000195928746810055,
"loss": 0.748,
"step": 2100
},
{
"epoch": 0.05494267040473566,
"grad_norm": 0.16327305138111115,
"learning_rate": 0.0001958814597784309,
"loss": 0.7306,
"step": 2110
},
{
"epoch": 0.055203062207601704,
"grad_norm": 0.15880291163921356,
"learning_rate": 0.00019583390548708486,
"loss": 0.7281,
"step": 2120
},
{
"epoch": 0.05546345401046775,
"grad_norm": 0.1702323853969574,
"learning_rate": 0.0001957860840685696,
"loss": 0.7407,
"step": 2130
},
{
"epoch": 0.0557238458133338,
"grad_norm": 0.16931670904159546,
"learning_rate": 0.0001957379956561825,
"loss": 0.7272,
"step": 2140
},
{
"epoch": 0.05598423761619984,
"grad_norm": 0.15455976128578186,
"learning_rate": 0.000195689640383965,
"loss": 0.7398,
"step": 2150
},
{
"epoch": 0.05624462941906589,
"grad_norm": 0.16061417758464813,
"learning_rate": 0.0001956410183867024,
"loss": 0.749,
"step": 2160
},
{
"epoch": 0.056505021221931934,
"grad_norm": 0.14933143556118011,
"learning_rate": 0.00019559212979992365,
"loss": 0.7418,
"step": 2170
},
{
"epoch": 0.05676541302479798,
"grad_norm": 0.1592816412448883,
"learning_rate": 0.00019554297475990058,
"loss": 0.7423,
"step": 2180
},
{
"epoch": 0.05702580482766403,
"grad_norm": 0.1677238792181015,
"learning_rate": 0.00019549355340364787,
"loss": 0.7101,
"step": 2190
},
{
"epoch": 0.05728619663053007,
"grad_norm": 0.3558599054813385,
"learning_rate": 0.00019544386586892238,
"loss": 0.725,
"step": 2200
},
{
"epoch": 0.05754658843339611,
"grad_norm": 0.1746376007795334,
"learning_rate": 0.00019539391229422313,
"loss": 0.7479,
"step": 2210
},
{
"epoch": 0.057806980236262163,
"grad_norm": 0.15979182720184326,
"learning_rate": 0.00019534369281879049,
"loss": 0.7352,
"step": 2220
},
{
"epoch": 0.05806737203912821,
"grad_norm": 0.16173166036605835,
"learning_rate": 0.0001952932075826061,
"loss": 0.7364,
"step": 2230
},
{
"epoch": 0.05832776384199426,
"grad_norm": 0.1514744907617569,
"learning_rate": 0.00019524245672639245,
"loss": 0.734,
"step": 2240
},
{
"epoch": 0.0585881556448603,
"grad_norm": 0.16860373318195343,
"learning_rate": 0.00019519144039161222,
"loss": 0.7098,
"step": 2250
},
{
"epoch": 0.05884854744772634,
"grad_norm": 0.16847743093967438,
"learning_rate": 0.00019514015872046833,
"loss": 0.7103,
"step": 2260
},
{
"epoch": 0.05910893925059239,
"grad_norm": 0.16181516647338867,
"learning_rate": 0.00019508861185590307,
"loss": 0.7561,
"step": 2270
},
{
"epoch": 0.059369331053458436,
"grad_norm": 0.16594484448432922,
"learning_rate": 0.0001950367999415981,
"loss": 0.7308,
"step": 2280
},
{
"epoch": 0.05962972285632448,
"grad_norm": 0.166441410779953,
"learning_rate": 0.00019498472312197375,
"loss": 0.735,
"step": 2290
},
{
"epoch": 0.05989011465919053,
"grad_norm": 0.16273920238018036,
"learning_rate": 0.00019493238154218886,
"loss": 0.7458,
"step": 2300
},
{
"epoch": 0.06015050646205657,
"grad_norm": 0.16227276623249054,
"learning_rate": 0.00019487977534814012,
"loss": 0.7143,
"step": 2310
},
{
"epoch": 0.06041089826492262,
"grad_norm": 0.1619606912136078,
"learning_rate": 0.000194826904686462,
"loss": 0.7285,
"step": 2320
},
{
"epoch": 0.060671290067788666,
"grad_norm": 0.1596045345067978,
"learning_rate": 0.00019477376970452603,
"loss": 0.7513,
"step": 2330
},
{
"epoch": 0.06093168187065471,
"grad_norm": 0.17504757642745972,
"learning_rate": 0.00019472037055044044,
"loss": 0.7376,
"step": 2340
},
{
"epoch": 0.06119207367352076,
"grad_norm": 0.1559167355298996,
"learning_rate": 0.00019466670737304992,
"loss": 0.7339,
"step": 2350
},
{
"epoch": 0.0614524654763868,
"grad_norm": 0.1624836027622223,
"learning_rate": 0.0001946127803219351,
"loss": 0.7258,
"step": 2360
},
{
"epoch": 0.061712857279252846,
"grad_norm": 0.17907138168811798,
"learning_rate": 0.00019455858954741206,
"loss": 0.72,
"step": 2370
},
{
"epoch": 0.061973249082118896,
"grad_norm": 0.15922705829143524,
"learning_rate": 0.00019450413520053202,
"loss": 0.7187,
"step": 2380
},
{
"epoch": 0.06223364088498494,
"grad_norm": 0.1552513986825943,
"learning_rate": 0.0001944494174330809,
"loss": 0.7183,
"step": 2390
},
{
"epoch": 0.06249403268785099,
"grad_norm": 0.16838514804840088,
"learning_rate": 0.00019439443639757885,
"loss": 0.7286,
"step": 2400
},
{
"epoch": 0.06275442449071703,
"grad_norm": 0.17352423071861267,
"learning_rate": 0.00019433919224727986,
"loss": 0.7436,
"step": 2410
},
{
"epoch": 0.06301481629358308,
"grad_norm": 0.17366603016853333,
"learning_rate": 0.0001942836851361713,
"loss": 0.7265,
"step": 2420
},
{
"epoch": 0.06327520809644913,
"grad_norm": 0.14833413064479828,
"learning_rate": 0.00019422791521897357,
"loss": 0.7234,
"step": 2430
},
{
"epoch": 0.06353559989931518,
"grad_norm": 0.16602723300457,
"learning_rate": 0.00019417188265113958,
"loss": 0.725,
"step": 2440
},
{
"epoch": 0.06379599170218121,
"grad_norm": 0.17290353775024414,
"learning_rate": 0.00019411558758885438,
"loss": 0.7174,
"step": 2450
},
{
"epoch": 0.06405638350504726,
"grad_norm": 0.16486665606498718,
"learning_rate": 0.0001940590301890346,
"loss": 0.7301,
"step": 2460
},
{
"epoch": 0.06431677530791331,
"grad_norm": 0.16255232691764832,
"learning_rate": 0.00019400221060932827,
"loss": 0.7462,
"step": 2470
},
{
"epoch": 0.06457716711077935,
"grad_norm": 0.16139757633209229,
"learning_rate": 0.0001939451290081141,
"loss": 0.7424,
"step": 2480
},
{
"epoch": 0.0648375589136454,
"grad_norm": 0.165597602725029,
"learning_rate": 0.00019388778554450117,
"loss": 0.7426,
"step": 2490
},
{
"epoch": 0.06509795071651145,
"grad_norm": 0.19819000363349915,
"learning_rate": 0.00019383018037832854,
"loss": 0.7356,
"step": 2500
},
{
"epoch": 0.06535834251937749,
"grad_norm": 0.16469696164131165,
"learning_rate": 0.00019377231367016467,
"loss": 0.718,
"step": 2510
},
{
"epoch": 0.06561873432224354,
"grad_norm": 0.1644965261220932,
"learning_rate": 0.00019371418558130702,
"loss": 0.7253,
"step": 2520
},
{
"epoch": 0.06587912612510959,
"grad_norm": 0.15347526967525482,
"learning_rate": 0.00019365579627378174,
"loss": 0.7214,
"step": 2530
},
{
"epoch": 0.06613951792797562,
"grad_norm": 0.1618672013282776,
"learning_rate": 0.00019359714591034302,
"loss": 0.7204,
"step": 2540
},
{
"epoch": 0.06639990973084167,
"grad_norm": 0.17043665051460266,
"learning_rate": 0.00019353823465447268,
"loss": 0.7278,
"step": 2550
},
{
"epoch": 0.06666030153370772,
"grad_norm": 0.15762579441070557,
"learning_rate": 0.00019347906267037983,
"loss": 0.7283,
"step": 2560
},
{
"epoch": 0.06692069333657376,
"grad_norm": 0.1622801572084427,
"learning_rate": 0.00019341963012300029,
"loss": 0.7193,
"step": 2570
},
{
"epoch": 0.06718108513943981,
"grad_norm": 0.16705769300460815,
"learning_rate": 0.00019335993717799617,
"loss": 0.7414,
"step": 2580
},
{
"epoch": 0.06744147694230586,
"grad_norm": 0.15886452794075012,
"learning_rate": 0.00019329998400175545,
"loss": 0.7242,
"step": 2590
},
{
"epoch": 0.06770186874517191,
"grad_norm": 0.17994090914726257,
"learning_rate": 0.00019323977076139142,
"loss": 0.7017,
"step": 2600
},
{
"epoch": 0.06796226054803794,
"grad_norm": 0.1609068214893341,
"learning_rate": 0.00019317929762474232,
"loss": 0.7352,
"step": 2610
},
{
"epoch": 0.068222652350904,
"grad_norm": 0.15605950355529785,
"learning_rate": 0.0001931185647603708,
"loss": 0.7249,
"step": 2620
},
{
"epoch": 0.06848304415377005,
"grad_norm": 0.16057750582695007,
"learning_rate": 0.00019305757233756352,
"loss": 0.7521,
"step": 2630
},
{
"epoch": 0.06874343595663608,
"grad_norm": 0.1703862100839615,
"learning_rate": 0.00019299632052633054,
"loss": 0.7245,
"step": 2640
},
{
"epoch": 0.06900382775950213,
"grad_norm": 0.16324444115161896,
"learning_rate": 0.00019293480949740505,
"loss": 0.7395,
"step": 2650
},
{
"epoch": 0.06926421956236818,
"grad_norm": 0.15283791720867157,
"learning_rate": 0.00019287303942224266,
"loss": 0.7158,
"step": 2660
},
{
"epoch": 0.06952461136523422,
"grad_norm": 0.1882282942533493,
"learning_rate": 0.00019281101047302114,
"loss": 0.724,
"step": 2670
},
{
"epoch": 0.06978500316810027,
"grad_norm": 0.16147953271865845,
"learning_rate": 0.00019274872282263984,
"loss": 0.7365,
"step": 2680
},
{
"epoch": 0.07004539497096632,
"grad_norm": 0.1614103466272354,
"learning_rate": 0.00019268617664471916,
"loss": 0.7206,
"step": 2690
},
{
"epoch": 0.07030578677383235,
"grad_norm": 0.16784432530403137,
"learning_rate": 0.00019262337211360016,
"loss": 0.7279,
"step": 2700
},
{
"epoch": 0.0705661785766984,
"grad_norm": 0.15966112911701202,
"learning_rate": 0.000192560309404344,
"loss": 0.7274,
"step": 2710
},
{
"epoch": 0.07082657037956445,
"grad_norm": 0.16970521211624146,
"learning_rate": 0.0001924969886927315,
"loss": 0.7038,
"step": 2720
},
{
"epoch": 0.07108696218243049,
"grad_norm": 0.16143856942653656,
"learning_rate": 0.00019243341015526272,
"loss": 0.7097,
"step": 2730
},
{
"epoch": 0.07134735398529654,
"grad_norm": 0.16041269898414612,
"learning_rate": 0.00019236957396915623,
"loss": 0.722,
"step": 2740
},
{
"epoch": 0.07160774578816259,
"grad_norm": 0.15845969319343567,
"learning_rate": 0.00019230548031234882,
"loss": 0.7238,
"step": 2750
},
{
"epoch": 0.07186813759102864,
"grad_norm": 0.14966030418872833,
"learning_rate": 0.00019224112936349502,
"loss": 0.7182,
"step": 2760
},
{
"epoch": 0.07212852939389468,
"grad_norm": 0.16525116562843323,
"learning_rate": 0.00019217652130196653,
"loss": 0.7397,
"step": 2770
},
{
"epoch": 0.07238892119676073,
"grad_norm": 0.18119119107723236,
"learning_rate": 0.0001921116563078516,
"loss": 0.7222,
"step": 2780
},
{
"epoch": 0.07264931299962678,
"grad_norm": 0.1709197610616684,
"learning_rate": 0.00019204653456195478,
"loss": 0.7068,
"step": 2790
},
{
"epoch": 0.07290970480249281,
"grad_norm": 0.16309161484241486,
"learning_rate": 0.00019198115624579625,
"loss": 0.7349,
"step": 2800
},
{
"epoch": 0.07317009660535886,
"grad_norm": 0.1736750453710556,
"learning_rate": 0.00019191552154161135,
"loss": 0.7445,
"step": 2810
},
{
"epoch": 0.07343048840822491,
"grad_norm": 0.15009112656116486,
"learning_rate": 0.00019184963063235006,
"loss": 0.7034,
"step": 2820
},
{
"epoch": 0.07369088021109095,
"grad_norm": 0.17244628071784973,
"learning_rate": 0.0001917834837016766,
"loss": 0.7285,
"step": 2830
},
{
"epoch": 0.073951272013957,
"grad_norm": 0.15991820394992828,
"learning_rate": 0.00019171708093396861,
"loss": 0.7096,
"step": 2840
},
{
"epoch": 0.07421166381682305,
"grad_norm": 0.17037667334079742,
"learning_rate": 0.0001916504225143171,
"loss": 0.7177,
"step": 2850
},
{
"epoch": 0.07447205561968909,
"grad_norm": 0.16700348258018494,
"learning_rate": 0.00019158350862852553,
"loss": 0.7453,
"step": 2860
},
{
"epoch": 0.07473244742255514,
"grad_norm": 0.17683659493923187,
"learning_rate": 0.00019151633946310948,
"loss": 0.7331,
"step": 2870
},
{
"epoch": 0.07499283922542119,
"grad_norm": 0.16364306211471558,
"learning_rate": 0.00019144891520529608,
"loss": 0.7347,
"step": 2880
},
{
"epoch": 0.07525323102828722,
"grad_norm": 0.1781424731016159,
"learning_rate": 0.00019138123604302355,
"loss": 0.7169,
"step": 2890
},
{
"epoch": 0.07551362283115327,
"grad_norm": 0.16007259488105774,
"learning_rate": 0.00019131330216494064,
"loss": 0.7269,
"step": 2900
},
{
"epoch": 0.07577401463401932,
"grad_norm": 0.1604921519756317,
"learning_rate": 0.00019124511376040598,
"loss": 0.7094,
"step": 2910
},
{
"epoch": 0.07603440643688536,
"grad_norm": 0.16649965941905975,
"learning_rate": 0.00019117667101948782,
"loss": 0.7271,
"step": 2920
},
{
"epoch": 0.07629479823975141,
"grad_norm": 0.16084066033363342,
"learning_rate": 0.0001911079741329632,
"loss": 0.7239,
"step": 2930
},
{
"epoch": 0.07655519004261746,
"grad_norm": 0.1651066243648529,
"learning_rate": 0.0001910390232923177,
"loss": 0.7304,
"step": 2940
},
{
"epoch": 0.07681558184548351,
"grad_norm": 0.1528957635164261,
"learning_rate": 0.00019096981868974467,
"loss": 0.7068,
"step": 2950
},
{
"epoch": 0.07707597364834955,
"grad_norm": 0.172830730676651,
"learning_rate": 0.00019090036051814483,
"loss": 0.7277,
"step": 2960
},
{
"epoch": 0.0773363654512156,
"grad_norm": 0.15909147262573242,
"learning_rate": 0.00019083064897112571,
"loss": 0.7135,
"step": 2970
},
{
"epoch": 0.07759675725408165,
"grad_norm": 0.16273066401481628,
"learning_rate": 0.0001907606842430011,
"loss": 0.7346,
"step": 2980
},
{
"epoch": 0.07785714905694768,
"grad_norm": 0.1595291793346405,
"learning_rate": 0.00019069046652879049,
"loss": 0.7377,
"step": 2990
},
{
"epoch": 0.07811754085981373,
"grad_norm": 0.15573470294475555,
"learning_rate": 0.0001906199960242185,
"loss": 0.7026,
"step": 3000
},
{
"epoch": 0.07837793266267978,
"grad_norm": 0.1670667678117752,
"learning_rate": 0.0001905492729257145,
"loss": 0.7231,
"step": 3010
},
{
"epoch": 0.07863832446554582,
"grad_norm": 0.17074571549892426,
"learning_rate": 0.00019047829743041184,
"loss": 0.7003,
"step": 3020
},
{
"epoch": 0.07889871626841187,
"grad_norm": 0.16979442536830902,
"learning_rate": 0.00019040706973614738,
"loss": 0.7217,
"step": 3030
},
{
"epoch": 0.07915910807127792,
"grad_norm": 0.15843816101551056,
"learning_rate": 0.00019033559004146103,
"loss": 0.7334,
"step": 3040
},
{
"epoch": 0.07941949987414396,
"grad_norm": 0.1607016921043396,
"learning_rate": 0.0001902638585455951,
"loss": 0.7271,
"step": 3050
},
{
"epoch": 0.07967989167701,
"grad_norm": 0.1619115173816681,
"learning_rate": 0.0001901918754484938,
"loss": 0.7144,
"step": 3060
},
{
"epoch": 0.07994028347987606,
"grad_norm": 0.1638360172510147,
"learning_rate": 0.00019011964095080254,
"loss": 0.7149,
"step": 3070
},
{
"epoch": 0.08020067528274209,
"grad_norm": 0.16503652930259705,
"learning_rate": 0.00019004715525386764,
"loss": 0.7011,
"step": 3080
},
{
"epoch": 0.08046106708560814,
"grad_norm": 0.16763822734355927,
"learning_rate": 0.00018997441855973552,
"loss": 0.7145,
"step": 3090
},
{
"epoch": 0.08072145888847419,
"grad_norm": 0.1621125340461731,
"learning_rate": 0.0001899014310711522,
"loss": 0.7318,
"step": 3100
},
{
"epoch": 0.08098185069134024,
"grad_norm": 0.16480112075805664,
"learning_rate": 0.0001898281929915629,
"loss": 0.7145,
"step": 3110
},
{
"epoch": 0.08124224249420628,
"grad_norm": 0.1805388629436493,
"learning_rate": 0.00018975470452511112,
"loss": 0.7102,
"step": 3120
},
{
"epoch": 0.08150263429707233,
"grad_norm": 0.1902652531862259,
"learning_rate": 0.00018968096587663853,
"loss": 0.7281,
"step": 3130
},
{
"epoch": 0.08176302609993838,
"grad_norm": 0.1732487976551056,
"learning_rate": 0.00018960697725168397,
"loss": 0.7434,
"step": 3140
},
{
"epoch": 0.08202341790280442,
"grad_norm": 0.1662171334028244,
"learning_rate": 0.00018953273885648314,
"loss": 0.716,
"step": 3150
},
{
"epoch": 0.08228380970567047,
"grad_norm": 0.16129222512245178,
"learning_rate": 0.00018945825089796797,
"loss": 0.7318,
"step": 3160
},
{
"epoch": 0.08254420150853652,
"grad_norm": 0.16837772727012634,
"learning_rate": 0.00018938351358376596,
"loss": 0.7137,
"step": 3170
},
{
"epoch": 0.08280459331140255,
"grad_norm": 0.1618524193763733,
"learning_rate": 0.00018930852712219974,
"loss": 0.7079,
"step": 3180
},
{
"epoch": 0.0830649851142686,
"grad_norm": 0.16333432495594025,
"learning_rate": 0.00018923329172228632,
"loss": 0.7062,
"step": 3190
},
{
"epoch": 0.08332537691713465,
"grad_norm": 0.15985700488090515,
"learning_rate": 0.00018915780759373672,
"loss": 0.7277,
"step": 3200
},
{
"epoch": 0.08358576872000069,
"grad_norm": 0.16181236505508423,
"learning_rate": 0.0001890820749469551,
"loss": 0.7048,
"step": 3210
},
{
"epoch": 0.08384616052286674,
"grad_norm": 0.1679672747850418,
"learning_rate": 0.00018900609399303853,
"loss": 0.73,
"step": 3220
},
{
"epoch": 0.08410655232573279,
"grad_norm": 0.16680286824703217,
"learning_rate": 0.00018892986494377606,
"loss": 0.7169,
"step": 3230
},
{
"epoch": 0.08436694412859883,
"grad_norm": 0.15980315208435059,
"learning_rate": 0.00018885338801164834,
"loss": 0.7346,
"step": 3240
},
{
"epoch": 0.08462733593146488,
"grad_norm": 0.16863352060317993,
"learning_rate": 0.00018877666340982695,
"loss": 0.7256,
"step": 3250
},
{
"epoch": 0.08488772773433093,
"grad_norm": 0.1657836139202118,
"learning_rate": 0.0001886996913521739,
"loss": 0.7219,
"step": 3260
},
{
"epoch": 0.08514811953719698,
"grad_norm": 0.1683470606803894,
"learning_rate": 0.00018862247205324087,
"loss": 0.7279,
"step": 3270
},
{
"epoch": 0.08540851134006301,
"grad_norm": 0.1686122715473175,
"learning_rate": 0.00018854500572826867,
"loss": 0.7178,
"step": 3280
},
{
"epoch": 0.08566890314292906,
"grad_norm": 0.16122782230377197,
"learning_rate": 0.00018846729259318682,
"loss": 0.7289,
"step": 3290
},
{
"epoch": 0.08592929494579511,
"grad_norm": 0.1920589804649353,
"learning_rate": 0.0001883893328646126,
"loss": 0.7264,
"step": 3300
},
{
"epoch": 0.08618968674866115,
"grad_norm": 0.17415335774421692,
"learning_rate": 0.00018831112675985083,
"loss": 0.7378,
"step": 3310
},
{
"epoch": 0.0864500785515272,
"grad_norm": 0.16903561353683472,
"learning_rate": 0.00018823267449689292,
"loss": 0.702,
"step": 3320
},
{
"epoch": 0.08671047035439325,
"grad_norm": 0.17158570885658264,
"learning_rate": 0.00018815397629441658,
"loss": 0.7264,
"step": 3330
},
{
"epoch": 0.08697086215725928,
"grad_norm": 0.17094087600708008,
"learning_rate": 0.0001880750323717849,
"loss": 0.741,
"step": 3340
},
{
"epoch": 0.08723125396012534,
"grad_norm": 0.1618223935365677,
"learning_rate": 0.000187995842949046,
"loss": 0.7084,
"step": 3350
},
{
"epoch": 0.08749164576299139,
"grad_norm": 0.1671626716852188,
"learning_rate": 0.0001879164082469322,
"loss": 0.7236,
"step": 3360
},
{
"epoch": 0.08775203756585742,
"grad_norm": 0.1681569665670395,
"learning_rate": 0.00018783672848685966,
"loss": 0.7382,
"step": 3370
},
{
"epoch": 0.08801242936872347,
"grad_norm": 0.16421955823898315,
"learning_rate": 0.00018775680389092748,
"loss": 0.717,
"step": 3380
},
{
"epoch": 0.08827282117158952,
"grad_norm": 0.174809530377388,
"learning_rate": 0.00018767663468191725,
"loss": 0.7225,
"step": 3390
},
{
"epoch": 0.08853321297445556,
"grad_norm": 0.16181902587413788,
"learning_rate": 0.00018759622108329243,
"loss": 0.7014,
"step": 3400
},
{
"epoch": 0.08879360477732161,
"grad_norm": 0.15579254925251007,
"learning_rate": 0.0001875155633191977,
"loss": 0.7125,
"step": 3410
},
{
"epoch": 0.08905399658018766,
"grad_norm": 0.16342496871948242,
"learning_rate": 0.00018743466161445823,
"loss": 0.7075,
"step": 3420
},
{
"epoch": 0.08931438838305371,
"grad_norm": 0.17215611040592194,
"learning_rate": 0.00018735351619457923,
"loss": 0.7331,
"step": 3430
},
{
"epoch": 0.08957478018591974,
"grad_norm": 0.1682904213666916,
"learning_rate": 0.00018727212728574522,
"loss": 0.734,
"step": 3440
},
{
"epoch": 0.0898351719887858,
"grad_norm": 0.16969889402389526,
"learning_rate": 0.00018719049511481948,
"loss": 0.7224,
"step": 3450
},
{
"epoch": 0.09009556379165184,
"grad_norm": 0.16607950627803802,
"learning_rate": 0.00018710861990934324,
"loss": 0.7218,
"step": 3460
},
{
"epoch": 0.09035595559451788,
"grad_norm": 0.16665585339069366,
"learning_rate": 0.00018702650189753525,
"loss": 0.7152,
"step": 3470
},
{
"epoch": 0.09061634739738393,
"grad_norm": 0.16812992095947266,
"learning_rate": 0.00018694414130829103,
"loss": 0.7097,
"step": 3480
},
{
"epoch": 0.09087673920024998,
"grad_norm": 0.16855508089065552,
"learning_rate": 0.00018686153837118224,
"loss": 0.7268,
"step": 3490
},
{
"epoch": 0.09113713100311602,
"grad_norm": 0.1634734570980072,
"learning_rate": 0.00018677869331645613,
"loss": 0.7485,
"step": 3500
},
{
"epoch": 0.09139752280598207,
"grad_norm": 0.17040792107582092,
"learning_rate": 0.00018669560637503478,
"loss": 0.7529,
"step": 3510
},
{
"epoch": 0.09165791460884812,
"grad_norm": 0.16575287282466888,
"learning_rate": 0.00018661227777851447,
"loss": 0.7053,
"step": 3520
},
{
"epoch": 0.09191830641171415,
"grad_norm": 0.16195647418498993,
"learning_rate": 0.00018652870775916515,
"loss": 0.7498,
"step": 3530
},
{
"epoch": 0.0921786982145802,
"grad_norm": 0.16672903299331665,
"learning_rate": 0.00018644489654992956,
"loss": 0.7168,
"step": 3540
},
{
"epoch": 0.09243909001744625,
"grad_norm": 0.16135342419147491,
"learning_rate": 0.00018636084438442298,
"loss": 0.6868,
"step": 3550
},
{
"epoch": 0.09269948182031229,
"grad_norm": 0.16677817702293396,
"learning_rate": 0.0001862765514969321,
"loss": 0.7128,
"step": 3560
},
{
"epoch": 0.09295987362317834,
"grad_norm": 0.16983947157859802,
"learning_rate": 0.00018619201812241474,
"loss": 0.7189,
"step": 3570
},
{
"epoch": 0.09322026542604439,
"grad_norm": 0.17054444551467896,
"learning_rate": 0.00018610724449649895,
"loss": 0.6916,
"step": 3580
},
{
"epoch": 0.09348065722891043,
"grad_norm": 0.1672518253326416,
"learning_rate": 0.0001860222308554825,
"loss": 0.7002,
"step": 3590
},
{
"epoch": 0.09374104903177648,
"grad_norm": 0.16726090013980865,
"learning_rate": 0.00018593697743633225,
"loss": 0.7153,
"step": 3600
},
{
"epoch": 0.09400144083464253,
"grad_norm": 0.16627268493175507,
"learning_rate": 0.0001858514844766833,
"loss": 0.7121,
"step": 3610
},
{
"epoch": 0.09426183263750858,
"grad_norm": 0.16322413086891174,
"learning_rate": 0.00018576575221483852,
"loss": 0.7018,
"step": 3620
},
{
"epoch": 0.09452222444037461,
"grad_norm": 0.1642254739999771,
"learning_rate": 0.00018567978088976776,
"loss": 0.706,
"step": 3630
},
{
"epoch": 0.09478261624324066,
"grad_norm": 0.17001990973949432,
"learning_rate": 0.00018559357074110727,
"loss": 0.7236,
"step": 3640
},
{
"epoch": 0.09504300804610671,
"grad_norm": 0.166624054312706,
"learning_rate": 0.00018550712200915897,
"loss": 0.7129,
"step": 3650
},
{
"epoch": 0.09530339984897275,
"grad_norm": 0.17228394746780396,
"learning_rate": 0.00018542043493488986,
"loss": 0.7096,
"step": 3660
},
{
"epoch": 0.0955637916518388,
"grad_norm": 0.17133943736553192,
"learning_rate": 0.0001853335097599312,
"loss": 0.7164,
"step": 3670
},
{
"epoch": 0.09582418345470485,
"grad_norm": 0.16100934147834778,
"learning_rate": 0.00018524634672657805,
"loss": 0.711,
"step": 3680
},
{
"epoch": 0.09608457525757089,
"grad_norm": 0.17711074650287628,
"learning_rate": 0.00018515894607778837,
"loss": 0.7153,
"step": 3690
},
{
"epoch": 0.09634496706043694,
"grad_norm": 0.20850953459739685,
"learning_rate": 0.00018507130805718252,
"loss": 0.7309,
"step": 3700
},
{
"epoch": 0.09660535886330299,
"grad_norm": 0.15805912017822266,
"learning_rate": 0.00018498343290904246,
"loss": 0.72,
"step": 3710
},
{
"epoch": 0.09686575066616902,
"grad_norm": 0.17850877344608307,
"learning_rate": 0.00018489532087831114,
"loss": 0.7235,
"step": 3720
},
{
"epoch": 0.09712614246903507,
"grad_norm": 0.16705553233623505,
"learning_rate": 0.0001848069722105918,
"loss": 0.7233,
"step": 3730
},
{
"epoch": 0.09738653427190112,
"grad_norm": 0.17024511098861694,
"learning_rate": 0.0001847183871521473,
"loss": 0.7127,
"step": 3740
},
{
"epoch": 0.09764692607476716,
"grad_norm": 0.16217222809791565,
"learning_rate": 0.0001846295659498994,
"loss": 0.6844,
"step": 3750
},
{
"epoch": 0.09790731787763321,
"grad_norm": 0.20443694293498993,
"learning_rate": 0.00018454050885142803,
"loss": 0.7327,
"step": 3760
},
{
"epoch": 0.09816770968049926,
"grad_norm": 0.16370315849781036,
"learning_rate": 0.00018445121610497072,
"loss": 0.7191,
"step": 3770
},
{
"epoch": 0.09842810148336531,
"grad_norm": 0.16387638449668884,
"learning_rate": 0.00018436168795942187,
"loss": 0.724,
"step": 3780
},
{
"epoch": 0.09868849328623135,
"grad_norm": 0.1664215326309204,
"learning_rate": 0.00018427192466433193,
"loss": 0.6955,
"step": 3790
},
{
"epoch": 0.0989488850890974,
"grad_norm": 0.17285983264446259,
"learning_rate": 0.0001841819264699069,
"loss": 0.7063,
"step": 3800
},
{
"epoch": 0.09920927689196345,
"grad_norm": 0.1637914627790451,
"learning_rate": 0.00018409169362700753,
"loss": 0.7211,
"step": 3810
},
{
"epoch": 0.09946966869482948,
"grad_norm": 0.1695072054862976,
"learning_rate": 0.00018400122638714855,
"loss": 0.7172,
"step": 3820
},
{
"epoch": 0.09973006049769553,
"grad_norm": 0.16097985208034515,
"learning_rate": 0.00018391052500249812,
"loss": 0.7073,
"step": 3830
},
{
"epoch": 0.09999045230056158,
"grad_norm": 0.16112865507602692,
"learning_rate": 0.00018381958972587707,
"loss": 0.7126,
"step": 3840
},
{
"epoch": 0.10025084410342762,
"grad_norm": 0.16681082546710968,
"learning_rate": 0.0001837284208107581,
"loss": 0.7331,
"step": 3850
},
{
"epoch": 0.10051123590629367,
"grad_norm": 0.16337643563747406,
"learning_rate": 0.0001836370185112652,
"loss": 0.7185,
"step": 3860
},
{
"epoch": 0.10077162770915972,
"grad_norm": 0.16530117392539978,
"learning_rate": 0.0001835453830821729,
"loss": 0.7058,
"step": 3870
},
{
"epoch": 0.10103201951202576,
"grad_norm": 0.1671489030122757,
"learning_rate": 0.00018345351477890554,
"loss": 0.7225,
"step": 3880
},
{
"epoch": 0.1012924113148918,
"grad_norm": 0.1631312221288681,
"learning_rate": 0.00018336141385753661,
"loss": 0.7189,
"step": 3890
},
{
"epoch": 0.10155280311775786,
"grad_norm": 0.16685089468955994,
"learning_rate": 0.00018326908057478795,
"loss": 0.7076,
"step": 3900
},
{
"epoch": 0.10181319492062389,
"grad_norm": 0.16232247650623322,
"learning_rate": 0.00018317651518802913,
"loss": 0.7168,
"step": 3910
},
{
"epoch": 0.10207358672348994,
"grad_norm": 0.16136577725410461,
"learning_rate": 0.0001830837179552766,
"loss": 0.7335,
"step": 3920
},
{
"epoch": 0.10233397852635599,
"grad_norm": 0.17820622026920319,
"learning_rate": 0.00018299068913519312,
"loss": 0.7071,
"step": 3930
},
{
"epoch": 0.10259437032922204,
"grad_norm": 0.15879027545452118,
"learning_rate": 0.00018289742898708707,
"loss": 0.7301,
"step": 3940
},
{
"epoch": 0.10285476213208808,
"grad_norm": 0.1569281965494156,
"learning_rate": 0.00018280393777091143,
"loss": 0.7018,
"step": 3950
},
{
"epoch": 0.10311515393495413,
"grad_norm": 0.15729983150959015,
"learning_rate": 0.00018271021574726336,
"loss": 0.717,
"step": 3960
},
{
"epoch": 0.10337554573782018,
"grad_norm": 0.1715734452009201,
"learning_rate": 0.00018261626317738348,
"loss": 0.7008,
"step": 3970
},
{
"epoch": 0.10363593754068622,
"grad_norm": 0.16048088669776917,
"learning_rate": 0.00018252208032315482,
"loss": 0.7108,
"step": 3980
},
{
"epoch": 0.10389632934355227,
"grad_norm": 0.1652391403913498,
"learning_rate": 0.00018242766744710248,
"loss": 0.7164,
"step": 3990
},
{
"epoch": 0.10415672114641832,
"grad_norm": 0.17617185413837433,
"learning_rate": 0.0001823330248123926,
"loss": 0.7152,
"step": 4000
},
{
"epoch": 0.10441711294928435,
"grad_norm": 0.16789917647838593,
"learning_rate": 0.00018223815268283186,
"loss": 0.742,
"step": 4010
},
{
"epoch": 0.1046775047521504,
"grad_norm": 0.15830789506435394,
"learning_rate": 0.00018214305132286656,
"loss": 0.7127,
"step": 4020
},
{
"epoch": 0.10493789655501645,
"grad_norm": 0.16791877150535583,
"learning_rate": 0.00018204772099758194,
"loss": 0.7173,
"step": 4030
},
{
"epoch": 0.10519828835788249,
"grad_norm": 0.1708458811044693,
"learning_rate": 0.00018195216197270156,
"loss": 0.7152,
"step": 4040
},
{
"epoch": 0.10545868016074854,
"grad_norm": 0.16406644880771637,
"learning_rate": 0.0001818563745145863,
"loss": 0.709,
"step": 4050
},
{
"epoch": 0.10571907196361459,
"grad_norm": 0.16491225361824036,
"learning_rate": 0.00018176035889023396,
"loss": 0.6957,
"step": 4060
},
{
"epoch": 0.10597946376648062,
"grad_norm": 0.18035203218460083,
"learning_rate": 0.00018166411536727812,
"loss": 0.7166,
"step": 4070
},
{
"epoch": 0.10623985556934668,
"grad_norm": 0.1827452927827835,
"learning_rate": 0.0001815676442139878,
"loss": 0.7359,
"step": 4080
},
{
"epoch": 0.10650024737221273,
"grad_norm": 0.17651152610778809,
"learning_rate": 0.00018147094569926644,
"loss": 0.7055,
"step": 4090
},
{
"epoch": 0.10676063917507878,
"grad_norm": 0.16017574071884155,
"learning_rate": 0.00018137402009265114,
"loss": 0.7175,
"step": 4100
},
{
"epoch": 0.10702103097794481,
"grad_norm": 0.16183608770370483,
"learning_rate": 0.00018127686766431213,
"loss": 0.7099,
"step": 4110
},
{
"epoch": 0.10728142278081086,
"grad_norm": 0.173137828707695,
"learning_rate": 0.00018117948868505182,
"loss": 0.7139,
"step": 4120
},
{
"epoch": 0.10754181458367691,
"grad_norm": 0.1736387312412262,
"learning_rate": 0.00018108188342630413,
"loss": 0.7219,
"step": 4130
},
{
"epoch": 0.10780220638654295,
"grad_norm": 0.1679358333349228,
"learning_rate": 0.0001809840521601337,
"loss": 0.7057,
"step": 4140
},
{
"epoch": 0.108062598189409,
"grad_norm": 0.16037985682487488,
"learning_rate": 0.00018088599515923503,
"loss": 0.7111,
"step": 4150
},
{
"epoch": 0.10832298999227505,
"grad_norm": 0.15554189682006836,
"learning_rate": 0.00018078771269693209,
"loss": 0.6978,
"step": 4160
},
{
"epoch": 0.10858338179514108,
"grad_norm": 0.1662927269935608,
"learning_rate": 0.00018068920504717704,
"loss": 0.7016,
"step": 4170
},
{
"epoch": 0.10884377359800713,
"grad_norm": 0.168484628200531,
"learning_rate": 0.0001805904724845499,
"loss": 0.695,
"step": 4180
},
{
"epoch": 0.10910416540087318,
"grad_norm": 0.18269601464271545,
"learning_rate": 0.0001804915152842575,
"loss": 0.7194,
"step": 4190
},
{
"epoch": 0.10936455720373922,
"grad_norm": 0.17641094326972961,
"learning_rate": 0.00018039233372213294,
"loss": 0.7209,
"step": 4200
},
{
"epoch": 0.10962494900660527,
"grad_norm": 0.18733248114585876,
"learning_rate": 0.00018029292807463456,
"loss": 0.7089,
"step": 4210
},
{
"epoch": 0.10988534080947132,
"grad_norm": 0.16925212740898132,
"learning_rate": 0.00018019329861884543,
"loss": 0.7097,
"step": 4220
},
{
"epoch": 0.11014573261233736,
"grad_norm": 0.15558676421642303,
"learning_rate": 0.00018009344563247245,
"loss": 0.6766,
"step": 4230
},
{
"epoch": 0.11040612441520341,
"grad_norm": 0.16530027985572815,
"learning_rate": 0.00017999336939384555,
"loss": 0.7141,
"step": 4240
},
{
"epoch": 0.11066651621806946,
"grad_norm": 0.16560573875904083,
"learning_rate": 0.00017989307018191693,
"loss": 0.6989,
"step": 4250
},
{
"epoch": 0.1109269080209355,
"grad_norm": 0.17466457188129425,
"learning_rate": 0.00017979254827626037,
"loss": 0.7206,
"step": 4260
},
{
"epoch": 0.11118729982380154,
"grad_norm": 0.18086989223957062,
"learning_rate": 0.00017969180395707038,
"loss": 0.714,
"step": 4270
},
{
"epoch": 0.1114476916266676,
"grad_norm": 1.1102476119995117,
"learning_rate": 0.00017959083750516137,
"loss": 0.723,
"step": 4280
},
{
"epoch": 0.11170808342953364,
"grad_norm": 0.19353629648685455,
"learning_rate": 0.00017948964920196697,
"loss": 0.718,
"step": 4290
},
{
"epoch": 0.11196847523239968,
"grad_norm": 0.2126789540052414,
"learning_rate": 0.00017938823932953915,
"loss": 0.7171,
"step": 4300
},
{
"epoch": 0.11222886703526573,
"grad_norm": 0.16679035127162933,
"learning_rate": 0.00017928660817054758,
"loss": 0.7113,
"step": 4310
},
{
"epoch": 0.11248925883813178,
"grad_norm": 0.1647113859653473,
"learning_rate": 0.00017918475600827861,
"loss": 0.7184,
"step": 4320
},
{
"epoch": 0.11274965064099782,
"grad_norm": 0.16823258996009827,
"learning_rate": 0.00017908268312663475,
"loss": 0.6876,
"step": 4330
},
{
"epoch": 0.11301004244386387,
"grad_norm": 0.17199194431304932,
"learning_rate": 0.0001789803898101336,
"loss": 0.6954,
"step": 4340
},
{
"epoch": 0.11327043424672992,
"grad_norm": 0.16541388630867004,
"learning_rate": 0.00017887787634390733,
"loss": 0.709,
"step": 4350
},
{
"epoch": 0.11353082604959595,
"grad_norm": 0.17238549888134003,
"learning_rate": 0.00017877514301370168,
"loss": 0.7056,
"step": 4360
},
{
"epoch": 0.113791217852462,
"grad_norm": 0.18374156951904297,
"learning_rate": 0.00017867219010587527,
"loss": 0.6951,
"step": 4370
},
{
"epoch": 0.11405160965532805,
"grad_norm": 0.15940402448177338,
"learning_rate": 0.00017856901790739874,
"loss": 0.6873,
"step": 4380
},
{
"epoch": 0.11431200145819409,
"grad_norm": 0.16984274983406067,
"learning_rate": 0.00017846562670585398,
"loss": 0.7065,
"step": 4390
},
{
"epoch": 0.11457239326106014,
"grad_norm": 0.9227787852287292,
"learning_rate": 0.0001783620167894334,
"loss": 0.7591,
"step": 4400
},
{
"epoch": 0.11483278506392619,
"grad_norm": 0.1881164163351059,
"learning_rate": 0.00017825818844693892,
"loss": 0.7133,
"step": 4410
},
{
"epoch": 0.11509317686679223,
"grad_norm": 0.20252032577991486,
"learning_rate": 0.00017815414196778147,
"loss": 0.7145,
"step": 4420
},
{
"epoch": 0.11535356866965828,
"grad_norm": 0.23380456864833832,
"learning_rate": 0.0001780498776419799,
"loss": 0.7232,
"step": 4430
},
{
"epoch": 0.11561396047252433,
"grad_norm": 0.1666252464056015,
"learning_rate": 0.00017794539576016028,
"loss": 0.7224,
"step": 4440
},
{
"epoch": 0.11587435227539038,
"grad_norm": 0.2421305775642395,
"learning_rate": 0.0001778406966135551,
"loss": 0.7031,
"step": 4450
},
{
"epoch": 0.11613474407825641,
"grad_norm": 0.17607644200325012,
"learning_rate": 0.00017773578049400256,
"loss": 0.7162,
"step": 4460
},
{
"epoch": 0.11639513588112246,
"grad_norm": 0.1905539333820343,
"learning_rate": 0.00017763064769394553,
"loss": 0.7087,
"step": 4470
},
{
"epoch": 0.11665552768398851,
"grad_norm": 0.17976076900959015,
"learning_rate": 0.00017752529850643088,
"loss": 0.7274,
"step": 4480
},
{
"epoch": 0.11691591948685455,
"grad_norm": 0.16786770522594452,
"learning_rate": 0.00017741973322510864,
"loss": 0.7003,
"step": 4490
},
{
"epoch": 0.1171763112897206,
"grad_norm": 0.1559310108423233,
"learning_rate": 0.00017731395214423126,
"loss": 0.7279,
"step": 4500
},
{
"epoch": 0.11743670309258665,
"grad_norm": 0.16374662518501282,
"learning_rate": 0.00017720795555865255,
"loss": 0.6924,
"step": 4510
},
{
"epoch": 0.11769709489545269,
"grad_norm": 0.1647823303937912,
"learning_rate": 0.00017710174376382712,
"loss": 0.7224,
"step": 4520
},
{
"epoch": 0.11795748669831874,
"grad_norm": 0.18070833384990692,
"learning_rate": 0.0001769953170558095,
"loss": 0.7223,
"step": 4530
},
{
"epoch": 0.11821787850118479,
"grad_norm": 0.16518941521644592,
"learning_rate": 0.0001768886757312532,
"loss": 0.7153,
"step": 4540
},
{
"epoch": 0.11847827030405082,
"grad_norm": 0.17081689834594727,
"learning_rate": 0.0001767818200874099,
"loss": 0.7197,
"step": 4550
},
{
"epoch": 0.11873866210691687,
"grad_norm": 0.16503626108169556,
"learning_rate": 0.00017667475042212874,
"loss": 0.7222,
"step": 4560
},
{
"epoch": 0.11899905390978292,
"grad_norm": 0.17310748994350433,
"learning_rate": 0.00017656746703385547,
"loss": 0.7057,
"step": 4570
},
{
"epoch": 0.11925944571264896,
"grad_norm": 0.1613416224718094,
"learning_rate": 0.0001764599702216315,
"loss": 0.7133,
"step": 4580
},
{
"epoch": 0.11951983751551501,
"grad_norm": 0.184098020195961,
"learning_rate": 0.00017635226028509313,
"loss": 0.72,
"step": 4590
},
{
"epoch": 0.11978022931838106,
"grad_norm": 0.1819763481616974,
"learning_rate": 0.00017624433752447076,
"loss": 0.7154,
"step": 4600
},
{
"epoch": 0.12004062112124711,
"grad_norm": 0.17078550159931183,
"learning_rate": 0.000176136202240588,
"loss": 0.7205,
"step": 4610
},
{
"epoch": 0.12030101292411315,
"grad_norm": 0.16961923241615295,
"learning_rate": 0.00017602785473486084,
"loss": 0.6997,
"step": 4620
},
{
"epoch": 0.1205614047269792,
"grad_norm": 0.18945324420928955,
"learning_rate": 0.00017591929530929684,
"loss": 0.7295,
"step": 4630
},
{
"epoch": 0.12082179652984525,
"grad_norm": 0.1690925508737564,
"learning_rate": 0.00017581052426649424,
"loss": 0.714,
"step": 4640
},
{
"epoch": 0.12108218833271128,
"grad_norm": 0.16739274561405182,
"learning_rate": 0.00017570154190964113,
"loss": 0.7018,
"step": 4650
},
{
"epoch": 0.12134258013557733,
"grad_norm": 0.17123165726661682,
"learning_rate": 0.0001755923485425147,
"loss": 0.7288,
"step": 4660
},
{
"epoch": 0.12160297193844338,
"grad_norm": 0.17006556689739227,
"learning_rate": 0.0001754829444694802,
"loss": 0.7185,
"step": 4670
},
{
"epoch": 0.12186336374130942,
"grad_norm": 0.17169633507728577,
"learning_rate": 0.00017537332999549023,
"loss": 0.7169,
"step": 4680
},
{
"epoch": 0.12212375554417547,
"grad_norm": 0.17851942777633667,
"learning_rate": 0.00017526350542608393,
"loss": 0.7138,
"step": 4690
},
{
"epoch": 0.12238414734704152,
"grad_norm": 0.16988542675971985,
"learning_rate": 0.00017515347106738597,
"loss": 0.7188,
"step": 4700
},
{
"epoch": 0.12264453914990756,
"grad_norm": 0.16778436303138733,
"learning_rate": 0.00017504322722610582,
"loss": 0.7034,
"step": 4710
},
{
"epoch": 0.1229049309527736,
"grad_norm": 0.1835828721523285,
"learning_rate": 0.00017493277420953688,
"loss": 0.7084,
"step": 4720
},
{
"epoch": 0.12316532275563966,
"grad_norm": 0.16590015590190887,
"learning_rate": 0.0001748221123255556,
"loss": 0.7023,
"step": 4730
},
{
"epoch": 0.12342571455850569,
"grad_norm": 0.16521744430065155,
"learning_rate": 0.00017471124188262062,
"loss": 0.7224,
"step": 4740
},
{
"epoch": 0.12368610636137174,
"grad_norm": 0.17280088365077972,
"learning_rate": 0.00017460016318977188,
"loss": 0.7087,
"step": 4750
},
{
"epoch": 0.12394649816423779,
"grad_norm": 0.17343877255916595,
"learning_rate": 0.00017448887655662984,
"loss": 0.7212,
"step": 4760
},
{
"epoch": 0.12420688996710384,
"grad_norm": 0.17116770148277283,
"learning_rate": 0.0001743773822933946,
"loss": 0.7289,
"step": 4770
},
{
"epoch": 0.12446728176996988,
"grad_norm": 0.17440655827522278,
"learning_rate": 0.0001742656807108449,
"loss": 0.7353,
"step": 4780
},
{
"epoch": 0.12472767357283593,
"grad_norm": 0.17304064333438873,
"learning_rate": 0.0001741537721203375,
"loss": 0.7322,
"step": 4790
},
{
"epoch": 0.12498806537570198,
"grad_norm": 0.17076118290424347,
"learning_rate": 0.00017404165683380604,
"loss": 0.7199,
"step": 4800
},
{
"epoch": 0.12524845717856803,
"grad_norm": 0.17371006309986115,
"learning_rate": 0.0001739293351637604,
"loss": 0.7035,
"step": 4810
},
{
"epoch": 0.12550884898143405,
"grad_norm": 0.16811051964759827,
"learning_rate": 0.0001738168074232857,
"loss": 0.7113,
"step": 4820
},
{
"epoch": 0.1257692407843001,
"grad_norm": 0.17408473789691925,
"learning_rate": 0.00017370407392604145,
"loss": 0.7151,
"step": 4830
},
{
"epoch": 0.12602963258716615,
"grad_norm": 0.16915513575077057,
"learning_rate": 0.0001735911349862607,
"loss": 0.7089,
"step": 4840
},
{
"epoch": 0.1262900243900322,
"grad_norm": 0.17281986773014069,
"learning_rate": 0.00017347799091874913,
"loss": 0.6935,
"step": 4850
},
{
"epoch": 0.12655041619289825,
"grad_norm": 0.16271603107452393,
"learning_rate": 0.00017336464203888426,
"loss": 0.7131,
"step": 4860
},
{
"epoch": 0.1268108079957643,
"grad_norm": 0.1661551147699356,
"learning_rate": 0.00017325108866261439,
"loss": 0.6899,
"step": 4870
},
{
"epoch": 0.12707119979863035,
"grad_norm": 0.16490161418914795,
"learning_rate": 0.00017313733110645792,
"loss": 0.7042,
"step": 4880
},
{
"epoch": 0.12733159160149637,
"grad_norm": 0.17232102155685425,
"learning_rate": 0.00017302336968750236,
"loss": 0.6965,
"step": 4890
},
{
"epoch": 0.12759198340436242,
"grad_norm": 0.17850111424922943,
"learning_rate": 0.00017290920472340347,
"loss": 0.7185,
"step": 4900
},
{
"epoch": 0.12785237520722847,
"grad_norm": 0.16296492516994476,
"learning_rate": 0.00017279483653238434,
"loss": 0.7179,
"step": 4910
},
{
"epoch": 0.12811276701009452,
"grad_norm": 0.1704237312078476,
"learning_rate": 0.00017268026543323454,
"loss": 0.7103,
"step": 4920
},
{
"epoch": 0.12837315881296057,
"grad_norm": 0.18265944719314575,
"learning_rate": 0.00017256549174530928,
"loss": 0.6988,
"step": 4930
},
{
"epoch": 0.12863355061582663,
"grad_norm": 0.17257574200630188,
"learning_rate": 0.0001724505157885284,
"loss": 0.7231,
"step": 4940
},
{
"epoch": 0.12889394241869265,
"grad_norm": 0.1703302413225174,
"learning_rate": 0.0001723353378833756,
"loss": 0.7192,
"step": 4950
},
{
"epoch": 0.1291543342215587,
"grad_norm": 0.16430538892745972,
"learning_rate": 0.00017221995835089743,
"loss": 0.7007,
"step": 4960
},
{
"epoch": 0.12941472602442475,
"grad_norm": 0.162199467420578,
"learning_rate": 0.0001721043775127025,
"loss": 0.7123,
"step": 4970
},
{
"epoch": 0.1296751178272908,
"grad_norm": 0.1680443435907364,
"learning_rate": 0.00017198859569096057,
"loss": 0.6976,
"step": 4980
},
{
"epoch": 0.12993550963015685,
"grad_norm": 0.16768567264080048,
"learning_rate": 0.0001718726132084015,
"loss": 0.6942,
"step": 4990
},
{
"epoch": 0.1301959014330229,
"grad_norm": 0.16759110987186432,
"learning_rate": 0.00017175643038831463,
"loss": 0.6995,
"step": 5000
},
{
"epoch": 0.13045629323588892,
"grad_norm": 0.17288459837436676,
"learning_rate": 0.00017164004755454754,
"loss": 0.7193,
"step": 5010
},
{
"epoch": 0.13071668503875497,
"grad_norm": 0.18251213431358337,
"learning_rate": 0.00017152346503150552,
"loss": 0.7156,
"step": 5020
},
{
"epoch": 0.13097707684162102,
"grad_norm": 0.17177079617977142,
"learning_rate": 0.0001714066831441503,
"loss": 0.7354,
"step": 5030
},
{
"epoch": 0.13123746864448707,
"grad_norm": 0.18522126972675323,
"learning_rate": 0.00017128970221799947,
"loss": 0.717,
"step": 5040
},
{
"epoch": 0.13149786044735312,
"grad_norm": 0.17539465427398682,
"learning_rate": 0.00017117252257912532,
"loss": 0.7204,
"step": 5050
},
{
"epoch": 0.13175825225021917,
"grad_norm": 0.16526655852794647,
"learning_rate": 0.00017105514455415403,
"loss": 0.7099,
"step": 5060
},
{
"epoch": 0.13201864405308522,
"grad_norm": 0.1709279865026474,
"learning_rate": 0.00017093756847026483,
"loss": 0.7205,
"step": 5070
},
{
"epoch": 0.13227903585595124,
"grad_norm": 0.16695177555084229,
"learning_rate": 0.00017081979465518896,
"loss": 0.7092,
"step": 5080
},
{
"epoch": 0.1325394276588173,
"grad_norm": 0.1795840859413147,
"learning_rate": 0.00017070182343720882,
"loss": 0.6983,
"step": 5090
},
{
"epoch": 0.13279981946168334,
"grad_norm": 0.17300280928611755,
"learning_rate": 0.0001705836551451571,
"loss": 0.7018,
"step": 5100
},
{
"epoch": 0.1330602112645494,
"grad_norm": 0.16966886818408966,
"learning_rate": 0.00017046529010841574,
"loss": 0.708,
"step": 5110
},
{
"epoch": 0.13332060306741544,
"grad_norm": 0.18426869809627533,
"learning_rate": 0.00017034672865691515,
"loss": 0.7118,
"step": 5120
},
{
"epoch": 0.1335809948702815,
"grad_norm": 0.1808805912733078,
"learning_rate": 0.00017022797112113314,
"loss": 0.7041,
"step": 5130
},
{
"epoch": 0.13384138667314752,
"grad_norm": 0.17788094282150269,
"learning_rate": 0.00017010901783209427,
"loss": 0.7032,
"step": 5140
},
{
"epoch": 0.13410177847601357,
"grad_norm": 0.17427201569080353,
"learning_rate": 0.00016998986912136852,
"loss": 0.7194,
"step": 5150
},
{
"epoch": 0.13436217027887962,
"grad_norm": 0.17783771455287933,
"learning_rate": 0.0001698705253210707,
"loss": 0.705,
"step": 5160
},
{
"epoch": 0.13462256208174567,
"grad_norm": 0.16358746588230133,
"learning_rate": 0.0001697509867638594,
"loss": 0.7118,
"step": 5170
},
{
"epoch": 0.13488295388461172,
"grad_norm": 0.17140185832977295,
"learning_rate": 0.0001696312537829361,
"loss": 0.6929,
"step": 5180
},
{
"epoch": 0.13514334568747777,
"grad_norm": 0.1690370887517929,
"learning_rate": 0.0001695113267120441,
"loss": 0.7044,
"step": 5190
},
{
"epoch": 0.13540373749034382,
"grad_norm": 0.16261689364910126,
"learning_rate": 0.0001693912058854679,
"loss": 0.6933,
"step": 5200
},
{
"epoch": 0.13566412929320984,
"grad_norm": 0.16243134438991547,
"learning_rate": 0.00016927089163803192,
"loss": 0.7009,
"step": 5210
},
{
"epoch": 0.1359245210960759,
"grad_norm": 0.17476530373096466,
"learning_rate": 0.00016915038430509978,
"loss": 0.7215,
"step": 5220
},
{
"epoch": 0.13618491289894194,
"grad_norm": 0.167569100856781,
"learning_rate": 0.00016902968422257325,
"loss": 0.6975,
"step": 5230
},
{
"epoch": 0.136445304701808,
"grad_norm": 0.1719943732023239,
"learning_rate": 0.00016890879172689146,
"loss": 0.7324,
"step": 5240
},
{
"epoch": 0.13670569650467404,
"grad_norm": 0.167982816696167,
"learning_rate": 0.00016878770715502984,
"loss": 0.6884,
"step": 5250
},
{
"epoch": 0.1369660883075401,
"grad_norm": 0.1711047738790512,
"learning_rate": 0.00016866643084449914,
"loss": 0.7164,
"step": 5260
},
{
"epoch": 0.1372264801104061,
"grad_norm": 0.17784222960472107,
"learning_rate": 0.00016854496313334466,
"loss": 0.6987,
"step": 5270
},
{
"epoch": 0.13748687191327216,
"grad_norm": 0.16513650119304657,
"learning_rate": 0.00016842330436014517,
"loss": 0.7078,
"step": 5280
},
{
"epoch": 0.1377472637161382,
"grad_norm": 0.17001712322235107,
"learning_rate": 0.000168301454864012,
"loss": 0.7053,
"step": 5290
},
{
"epoch": 0.13800765551900426,
"grad_norm": 0.1930648535490036,
"learning_rate": 0.0001681794149845881,
"loss": 0.6954,
"step": 5300
},
{
"epoch": 0.1382680473218703,
"grad_norm": 0.1823079138994217,
"learning_rate": 0.00016805718506204711,
"loss": 0.7325,
"step": 5310
},
{
"epoch": 0.13852843912473636,
"grad_norm": 0.16231094300746918,
"learning_rate": 0.00016793476543709238,
"loss": 0.7074,
"step": 5320
},
{
"epoch": 0.13878883092760239,
"grad_norm": 0.16010914742946625,
"learning_rate": 0.00016781215645095604,
"loss": 0.6946,
"step": 5330
},
{
"epoch": 0.13904922273046844,
"grad_norm": 0.17128457129001617,
"learning_rate": 0.00016768935844539804,
"loss": 0.7116,
"step": 5340
},
{
"epoch": 0.13930961453333449,
"grad_norm": 0.17700603604316711,
"learning_rate": 0.00016756637176270522,
"loss": 0.7276,
"step": 5350
},
{
"epoch": 0.13957000633620054,
"grad_norm": 0.16596612334251404,
"learning_rate": 0.00016744319674569033,
"loss": 0.6683,
"step": 5360
},
{
"epoch": 0.1398303981390666,
"grad_norm": 0.17127594351768494,
"learning_rate": 0.00016731983373769106,
"loss": 0.7053,
"step": 5370
},
{
"epoch": 0.14009078994193264,
"grad_norm": 0.17723415791988373,
"learning_rate": 0.00016719628308256915,
"loss": 0.7294,
"step": 5380
},
{
"epoch": 0.1403511817447987,
"grad_norm": 0.16194766759872437,
"learning_rate": 0.00016707254512470936,
"loss": 0.7164,
"step": 5390
},
{
"epoch": 0.1406115735476647,
"grad_norm": 0.17294222116470337,
"learning_rate": 0.00016694862020901852,
"loss": 0.7084,
"step": 5400
},
{
"epoch": 0.14087196535053076,
"grad_norm": 0.17684406042099,
"learning_rate": 0.00016682450868092463,
"loss": 0.7141,
"step": 5410
},
{
"epoch": 0.1411323571533968,
"grad_norm": 0.16746504604816437,
"learning_rate": 0.00016670021088637579,
"loss": 0.7336,
"step": 5420
},
{
"epoch": 0.14139274895626286,
"grad_norm": 0.2371237874031067,
"learning_rate": 0.00016657572717183937,
"loss": 0.7001,
"step": 5430
},
{
"epoch": 0.1416531407591289,
"grad_norm": 0.16818630695343018,
"learning_rate": 0.00016645105788430095,
"loss": 0.7124,
"step": 5440
},
{
"epoch": 0.14191353256199496,
"grad_norm": 0.1719949096441269,
"learning_rate": 0.00016632620337126333,
"loss": 0.7169,
"step": 5450
},
{
"epoch": 0.14217392436486098,
"grad_norm": 0.1754136085510254,
"learning_rate": 0.00016620116398074567,
"loss": 0.6966,
"step": 5460
},
{
"epoch": 0.14243431616772703,
"grad_norm": 0.1630551517009735,
"learning_rate": 0.0001660759400612824,
"loss": 0.695,
"step": 5470
},
{
"epoch": 0.14269470797059308,
"grad_norm": 0.16431647539138794,
"learning_rate": 0.00016595053196192234,
"loss": 0.6983,
"step": 5480
},
{
"epoch": 0.14295509977345913,
"grad_norm": 0.1706673949956894,
"learning_rate": 0.00016582494003222772,
"loss": 0.7025,
"step": 5490
},
{
"epoch": 0.14321549157632518,
"grad_norm": 0.1757555902004242,
"learning_rate": 0.00016569916462227312,
"loss": 0.7159,
"step": 5500
},
{
"epoch": 0.14347588337919123,
"grad_norm": 0.1697273999452591,
"learning_rate": 0.0001655732060826446,
"loss": 0.7112,
"step": 5510
},
{
"epoch": 0.14373627518205728,
"grad_norm": 0.17706693708896637,
"learning_rate": 0.00016544706476443862,
"loss": 0.746,
"step": 5520
},
{
"epoch": 0.1439966669849233,
"grad_norm": 0.17265696823596954,
"learning_rate": 0.00016532074101926117,
"loss": 0.7276,
"step": 5530
},
{
"epoch": 0.14425705878778936,
"grad_norm": 0.16773734986782074,
"learning_rate": 0.00016519423519922668,
"loss": 0.6944,
"step": 5540
},
{
"epoch": 0.1445174505906554,
"grad_norm": 0.17664338648319244,
"learning_rate": 0.0001650675476569572,
"loss": 0.6953,
"step": 5550
},
{
"epoch": 0.14477784239352146,
"grad_norm": 0.17331312596797943,
"learning_rate": 0.00016494067874558117,
"loss": 0.7113,
"step": 5560
},
{
"epoch": 0.1450382341963875,
"grad_norm": 0.1759510487318039,
"learning_rate": 0.00016481362881873272,
"loss": 0.71,
"step": 5570
},
{
"epoch": 0.14529862599925356,
"grad_norm": 0.18095999956130981,
"learning_rate": 0.00016468639823055044,
"loss": 0.7255,
"step": 5580
},
{
"epoch": 0.14555901780211958,
"grad_norm": 0.17434170842170715,
"learning_rate": 0.0001645589873356765,
"loss": 0.6989,
"step": 5590
},
{
"epoch": 0.14581940960498563,
"grad_norm": 0.16754081845283508,
"learning_rate": 0.00016443139648925572,
"loss": 0.7196,
"step": 5600
},
{
"epoch": 0.14607980140785168,
"grad_norm": 0.17215152084827423,
"learning_rate": 0.00016430362604693448,
"loss": 0.7104,
"step": 5610
},
{
"epoch": 0.14634019321071773,
"grad_norm": 0.17826251685619354,
"learning_rate": 0.00016417567636485973,
"loss": 0.7142,
"step": 5620
},
{
"epoch": 0.14660058501358378,
"grad_norm": 0.17078974843025208,
"learning_rate": 0.00016404754779967813,
"loss": 0.717,
"step": 5630
},
{
"epoch": 0.14686097681644983,
"grad_norm": 0.18085996806621552,
"learning_rate": 0.0001639192407085348,
"loss": 0.711,
"step": 5640
},
{
"epoch": 0.14712136861931585,
"grad_norm": 0.19176393747329712,
"learning_rate": 0.00016379075544907268,
"loss": 0.699,
"step": 5650
},
{
"epoch": 0.1473817604221819,
"grad_norm": 0.17717042565345764,
"learning_rate": 0.00016366209237943122,
"loss": 0.7069,
"step": 5660
},
{
"epoch": 0.14764215222504795,
"grad_norm": 0.17190402746200562,
"learning_rate": 0.0001635332518582455,
"loss": 0.6988,
"step": 5670
},
{
"epoch": 0.147902544027914,
"grad_norm": 0.18004834651947021,
"learning_rate": 0.00016340423424464516,
"loss": 0.7196,
"step": 5680
},
{
"epoch": 0.14816293583078005,
"grad_norm": 0.17457543313503265,
"learning_rate": 0.00016327503989825364,
"loss": 0.7109,
"step": 5690
},
{
"epoch": 0.1484233276336461,
"grad_norm": 0.1904648393392563,
"learning_rate": 0.00016314566917918693,
"loss": 0.6874,
"step": 5700
},
{
"epoch": 0.14868371943651215,
"grad_norm": 0.18096795678138733,
"learning_rate": 0.00016301612244805252,
"loss": 0.7154,
"step": 5710
},
{
"epoch": 0.14894411123937817,
"grad_norm": 0.17521658539772034,
"learning_rate": 0.00016288640006594874,
"loss": 0.703,
"step": 5720
},
{
"epoch": 0.14920450304224422,
"grad_norm": 0.180609330534935,
"learning_rate": 0.00016275650239446328,
"loss": 0.7033,
"step": 5730
},
{
"epoch": 0.14946489484511027,
"grad_norm": 0.19524458050727844,
"learning_rate": 0.00016262642979567268,
"loss": 0.6925,
"step": 5740
},
{
"epoch": 0.14972528664797632,
"grad_norm": 0.16790035367012024,
"learning_rate": 0.00016249618263214087,
"loss": 0.7104,
"step": 5750
},
{
"epoch": 0.14998567845084237,
"grad_norm": 0.16895075142383575,
"learning_rate": 0.00016236576126691843,
"loss": 0.7126,
"step": 5760
},
{
"epoch": 0.15024607025370842,
"grad_norm": 0.18354611098766327,
"learning_rate": 0.00016223516606354163,
"loss": 0.7143,
"step": 5770
},
{
"epoch": 0.15050646205657445,
"grad_norm": 0.1766786128282547,
"learning_rate": 0.00016210439738603108,
"loss": 0.7164,
"step": 5780
},
{
"epoch": 0.1507668538594405,
"grad_norm": 0.17794294655323029,
"learning_rate": 0.00016197345559889107,
"loss": 0.6919,
"step": 5790
},
{
"epoch": 0.15102724566230655,
"grad_norm": 0.1671450436115265,
"learning_rate": 0.00016184234106710838,
"loss": 0.7048,
"step": 5800
},
{
"epoch": 0.1512876374651726,
"grad_norm": 0.16732144355773926,
"learning_rate": 0.00016171105415615134,
"loss": 0.7168,
"step": 5810
},
{
"epoch": 0.15154802926803865,
"grad_norm": 0.16594409942626953,
"learning_rate": 0.00016157959523196866,
"loss": 0.6942,
"step": 5820
},
{
"epoch": 0.1518084210709047,
"grad_norm": 0.173036128282547,
"learning_rate": 0.00016144796466098865,
"loss": 0.7093,
"step": 5830
},
{
"epoch": 0.15206881287377072,
"grad_norm": 0.17697438597679138,
"learning_rate": 0.00016131616281011798,
"loss": 0.7122,
"step": 5840
},
{
"epoch": 0.15232920467663677,
"grad_norm": 0.1677410751581192,
"learning_rate": 0.0001611841900467408,
"loss": 0.7027,
"step": 5850
},
{
"epoch": 0.15258959647950282,
"grad_norm": 0.16899007558822632,
"learning_rate": 0.0001610520467387176,
"loss": 0.6949,
"step": 5860
},
{
"epoch": 0.15284998828236887,
"grad_norm": 0.17043690383434296,
"learning_rate": 0.00016091973325438428,
"loss": 0.7052,
"step": 5870
},
{
"epoch": 0.15311038008523492,
"grad_norm": 0.17432864010334015,
"learning_rate": 0.00016078724996255114,
"loss": 0.7026,
"step": 5880
},
{
"epoch": 0.15337077188810097,
"grad_norm": 0.17059588432312012,
"learning_rate": 0.00016065459723250164,
"loss": 0.7234,
"step": 5890
},
{
"epoch": 0.15363116369096702,
"grad_norm": 0.16677695512771606,
"learning_rate": 0.0001605217754339918,
"loss": 0.7281,
"step": 5900
},
{
"epoch": 0.15389155549383304,
"grad_norm": 0.16831070184707642,
"learning_rate": 0.00016038878493724858,
"loss": 0.6934,
"step": 5910
},
{
"epoch": 0.1541519472966991,
"grad_norm": 0.17344878613948822,
"learning_rate": 0.00016025562611296946,
"loss": 0.7096,
"step": 5920
},
{
"epoch": 0.15441233909956514,
"grad_norm": 0.16978445649147034,
"learning_rate": 0.0001601222993323209,
"loss": 0.6851,
"step": 5930
},
{
"epoch": 0.1546727309024312,
"grad_norm": 0.17398668825626373,
"learning_rate": 0.00015998880496693766,
"loss": 0.7022,
"step": 5940
},
{
"epoch": 0.15493312270529724,
"grad_norm": 0.16338910162448883,
"learning_rate": 0.00015985514338892154,
"loss": 0.7257,
"step": 5950
},
{
"epoch": 0.1551935145081633,
"grad_norm": 0.1754075437784195,
"learning_rate": 0.0001597213149708405,
"loss": 0.7012,
"step": 5960
},
{
"epoch": 0.15545390631102932,
"grad_norm": 0.16998423635959625,
"learning_rate": 0.00015958732008572744,
"loss": 0.7313,
"step": 5970
},
{
"epoch": 0.15571429811389537,
"grad_norm": 0.18673604726791382,
"learning_rate": 0.00015945315910707945,
"loss": 0.6923,
"step": 5980
},
{
"epoch": 0.15597468991676142,
"grad_norm": 0.17040878534317017,
"learning_rate": 0.0001593188324088564,
"loss": 0.7009,
"step": 5990
},
{
"epoch": 0.15623508171962747,
"grad_norm": 0.16569504141807556,
"learning_rate": 0.00015918434036548017,
"loss": 0.6919,
"step": 6000
},
{
"epoch": 0.15649547352249352,
"grad_norm": 0.16722472012043,
"learning_rate": 0.00015904968335183354,
"loss": 0.7028,
"step": 6010
},
{
"epoch": 0.15675586532535957,
"grad_norm": 0.16823537647724152,
"learning_rate": 0.00015891486174325912,
"loss": 0.6822,
"step": 6020
},
{
"epoch": 0.15701625712822562,
"grad_norm": 0.1750400960445404,
"learning_rate": 0.00015877987591555822,
"loss": 0.7146,
"step": 6030
},
{
"epoch": 0.15727664893109164,
"grad_norm": 0.1703154593706131,
"learning_rate": 0.00015864472624499008,
"loss": 0.7057,
"step": 6040
},
{
"epoch": 0.1575370407339577,
"grad_norm": 0.17317266762256622,
"learning_rate": 0.00015850941310827045,
"loss": 0.692,
"step": 6050
},
{
"epoch": 0.15779743253682374,
"grad_norm": 0.18270671367645264,
"learning_rate": 0.00015837393688257083,
"loss": 0.7279,
"step": 6060
},
{
"epoch": 0.1580578243396898,
"grad_norm": 0.1647195667028427,
"learning_rate": 0.00015823829794551724,
"loss": 0.6999,
"step": 6070
},
{
"epoch": 0.15831821614255584,
"grad_norm": 0.17559175193309784,
"learning_rate": 0.00015810249667518936,
"loss": 0.6952,
"step": 6080
},
{
"epoch": 0.1585786079454219,
"grad_norm": 0.168562114238739,
"learning_rate": 0.00015796653345011928,
"loss": 0.7013,
"step": 6090
},
{
"epoch": 0.1588389997482879,
"grad_norm": 0.18079742789268494,
"learning_rate": 0.00015783040864929043,
"loss": 0.7115,
"step": 6100
},
{
"epoch": 0.15909939155115396,
"grad_norm": 0.17427009344100952,
"learning_rate": 0.00015769412265213684,
"loss": 0.6955,
"step": 6110
},
{
"epoch": 0.15935978335402,
"grad_norm": 0.18136094510555267,
"learning_rate": 0.00015755767583854163,
"loss": 0.7129,
"step": 6120
},
{
"epoch": 0.15962017515688606,
"grad_norm": 0.17283126711845398,
"learning_rate": 0.00015742106858883633,
"loss": 0.6993,
"step": 6130
},
{
"epoch": 0.1598805669597521,
"grad_norm": 0.16747362911701202,
"learning_rate": 0.00015728430128379963,
"loss": 0.718,
"step": 6140
},
{
"epoch": 0.16014095876261816,
"grad_norm": 0.17267820239067078,
"learning_rate": 0.0001571473743046563,
"loss": 0.6946,
"step": 6150
},
{
"epoch": 0.16040135056548419,
"grad_norm": 0.17288938164710999,
"learning_rate": 0.00015701028803307625,
"loss": 0.7031,
"step": 6160
},
{
"epoch": 0.16066174236835024,
"grad_norm": 0.16363868117332458,
"learning_rate": 0.0001568730428511734,
"loss": 0.7038,
"step": 6170
},
{
"epoch": 0.16092213417121629,
"grad_norm": 0.17383398115634918,
"learning_rate": 0.0001567356391415046,
"loss": 0.695,
"step": 6180
},
{
"epoch": 0.16118252597408234,
"grad_norm": 0.15880119800567627,
"learning_rate": 0.00015659807728706854,
"loss": 0.6977,
"step": 6190
},
{
"epoch": 0.16144291777694839,
"grad_norm": 0.1778557300567627,
"learning_rate": 0.0001564603576713048,
"loss": 0.7319,
"step": 6200
},
{
"epoch": 0.16170330957981444,
"grad_norm": 0.1838517040014267,
"learning_rate": 0.00015632248067809265,
"loss": 0.6999,
"step": 6210
},
{
"epoch": 0.16196370138268049,
"grad_norm": 0.1765550673007965,
"learning_rate": 0.00015618444669175004,
"loss": 0.689,
"step": 6220
},
{
"epoch": 0.1622240931855465,
"grad_norm": 0.17367658019065857,
"learning_rate": 0.00015604625609703257,
"loss": 0.7028,
"step": 6230
},
{
"epoch": 0.16248448498841256,
"grad_norm": 0.1650581955909729,
"learning_rate": 0.0001559079092791323,
"loss": 0.6926,
"step": 6240
},
{
"epoch": 0.1627448767912786,
"grad_norm": 0.16974525153636932,
"learning_rate": 0.00015576940662367675,
"loss": 0.711,
"step": 6250
},
{
"epoch": 0.16300526859414466,
"grad_norm": 0.17168691754341125,
"learning_rate": 0.00015563074851672787,
"loss": 0.6949,
"step": 6260
},
{
"epoch": 0.1632656603970107,
"grad_norm": 0.17578692734241486,
"learning_rate": 0.00015549193534478094,
"loss": 0.6858,
"step": 6270
},
{
"epoch": 0.16352605219987676,
"grad_norm": 0.17432482540607452,
"learning_rate": 0.00015535296749476335,
"loss": 0.7025,
"step": 6280
},
{
"epoch": 0.16378644400274278,
"grad_norm": 0.17605715990066528,
"learning_rate": 0.00015521384535403376,
"loss": 0.7021,
"step": 6290
},
{
"epoch": 0.16404683580560883,
"grad_norm": 0.17800843715667725,
"learning_rate": 0.0001550745693103808,
"loss": 0.6998,
"step": 6300
},
{
"epoch": 0.16430722760847488,
"grad_norm": 0.17110416293144226,
"learning_rate": 0.0001549351397520222,
"loss": 0.7116,
"step": 6310
},
{
"epoch": 0.16456761941134093,
"grad_norm": 0.1723811775445938,
"learning_rate": 0.00015479555706760344,
"loss": 0.72,
"step": 6320
},
{
"epoch": 0.16482801121420698,
"grad_norm": 0.1752457171678543,
"learning_rate": 0.00015465582164619703,
"loss": 0.6866,
"step": 6330
},
{
"epoch": 0.16508840301707303,
"grad_norm": 0.1849747896194458,
"learning_rate": 0.00015451593387730105,
"loss": 0.7288,
"step": 6340
},
{
"epoch": 0.16534879481993905,
"grad_norm": 0.16762009263038635,
"learning_rate": 0.00015437589415083828,
"loss": 0.6919,
"step": 6350
},
{
"epoch": 0.1656091866228051,
"grad_norm": 0.17073680460453033,
"learning_rate": 0.00015423570285715506,
"loss": 0.709,
"step": 6360
},
{
"epoch": 0.16586957842567115,
"grad_norm": 0.16971205174922943,
"learning_rate": 0.00015409536038702028,
"loss": 0.7156,
"step": 6370
},
{
"epoch": 0.1661299702285372,
"grad_norm": 0.17233216762542725,
"learning_rate": 0.00015395486713162414,
"loss": 0.7187,
"step": 6380
},
{
"epoch": 0.16639036203140325,
"grad_norm": 0.17737407982349396,
"learning_rate": 0.00015381422348257715,
"loss": 0.71,
"step": 6390
},
{
"epoch": 0.1666507538342693,
"grad_norm": 0.17228469252586365,
"learning_rate": 0.00015367342983190906,
"loss": 0.7224,
"step": 6400
},
{
"epoch": 0.16691114563713536,
"grad_norm": 0.17020554840564728,
"learning_rate": 0.00015353248657206764,
"loss": 0.7052,
"step": 6410
},
{
"epoch": 0.16717153744000138,
"grad_norm": 0.1692919284105301,
"learning_rate": 0.00015339139409591783,
"loss": 0.7035,
"step": 6420
},
{
"epoch": 0.16743192924286743,
"grad_norm": 0.17089873552322388,
"learning_rate": 0.00015325015279674036,
"loss": 0.7043,
"step": 6430
},
{
"epoch": 0.16769232104573348,
"grad_norm": 0.16704905033111572,
"learning_rate": 0.00015310876306823092,
"loss": 0.6924,
"step": 6440
},
{
"epoch": 0.16795271284859953,
"grad_norm": 0.1821179836988449,
"learning_rate": 0.00015296722530449877,
"loss": 0.6996,
"step": 6450
},
{
"epoch": 0.16821310465146558,
"grad_norm": 0.17202268540859222,
"learning_rate": 0.000152825539900066,
"loss": 0.7164,
"step": 6460
},
{
"epoch": 0.16847349645433163,
"grad_norm": 0.16377121210098267,
"learning_rate": 0.00015268370724986601,
"loss": 0.699,
"step": 6470
},
{
"epoch": 0.16873388825719765,
"grad_norm": 0.17642685770988464,
"learning_rate": 0.00015254172774924277,
"loss": 0.7127,
"step": 6480
},
{
"epoch": 0.1689942800600637,
"grad_norm": 0.1717827171087265,
"learning_rate": 0.0001523996017939496,
"loss": 0.6927,
"step": 6490
},
{
"epoch": 0.16925467186292975,
"grad_norm": 0.1642349660396576,
"learning_rate": 0.000152257329780148,
"loss": 0.704,
"step": 6500
},
{
"epoch": 0.1695150636657958,
"grad_norm": 0.17559029161930084,
"learning_rate": 0.00015211491210440653,
"loss": 0.7236,
"step": 6510
},
{
"epoch": 0.16977545546866185,
"grad_norm": 0.1805913746356964,
"learning_rate": 0.00015197234916369996,
"loss": 0.6829,
"step": 6520
},
{
"epoch": 0.1700358472715279,
"grad_norm": 0.17583122849464417,
"learning_rate": 0.00015182964135540772,
"loss": 0.7045,
"step": 6530
},
{
"epoch": 0.17029623907439395,
"grad_norm": 0.1662655770778656,
"learning_rate": 0.00015168678907731323,
"loss": 0.7173,
"step": 6540
},
{
"epoch": 0.17055663087725997,
"grad_norm": 0.1719600409269333,
"learning_rate": 0.00015154379272760252,
"loss": 0.7082,
"step": 6550
},
{
"epoch": 0.17081702268012602,
"grad_norm": 0.18137727677822113,
"learning_rate": 0.00015140065270486324,
"loss": 0.7314,
"step": 6560
},
{
"epoch": 0.17107741448299207,
"grad_norm": 0.17727130651474,
"learning_rate": 0.00015125736940808351,
"loss": 0.7093,
"step": 6570
},
{
"epoch": 0.17133780628585812,
"grad_norm": 0.19269149005413055,
"learning_rate": 0.00015111394323665081,
"loss": 0.6859,
"step": 6580
},
{
"epoch": 0.17159819808872417,
"grad_norm": 0.18232578039169312,
"learning_rate": 0.00015097037459035084,
"loss": 0.726,
"step": 6590
},
{
"epoch": 0.17185858989159022,
"grad_norm": 0.16427166759967804,
"learning_rate": 0.0001508266638693665,
"loss": 0.6896,
"step": 6600
},
{
"epoch": 0.17211898169445625,
"grad_norm": 0.17567217350006104,
"learning_rate": 0.00015068281147427657,
"loss": 0.6968,
"step": 6610
},
{
"epoch": 0.1723793734973223,
"grad_norm": 0.17443399131298065,
"learning_rate": 0.00015053881780605495,
"loss": 0.7034,
"step": 6620
},
{
"epoch": 0.17263976530018835,
"grad_norm": 0.16696178913116455,
"learning_rate": 0.00015039468326606911,
"loss": 0.6897,
"step": 6630
},
{
"epoch": 0.1729001571030544,
"grad_norm": 0.1779368370771408,
"learning_rate": 0.00015025040825607935,
"loss": 0.7036,
"step": 6640
},
{
"epoch": 0.17316054890592045,
"grad_norm": 0.17863190174102783,
"learning_rate": 0.00015010599317823734,
"loss": 0.7032,
"step": 6650
},
{
"epoch": 0.1734209407087865,
"grad_norm": 0.1763063222169876,
"learning_rate": 0.00014996143843508538,
"loss": 0.6825,
"step": 6660
},
{
"epoch": 0.17368133251165252,
"grad_norm": 0.17159639298915863,
"learning_rate": 0.00014981674442955488,
"loss": 0.6988,
"step": 6670
},
{
"epoch": 0.17394172431451857,
"grad_norm": 0.17756308615207672,
"learning_rate": 0.00014967191156496564,
"loss": 0.7018,
"step": 6680
},
{
"epoch": 0.17420211611738462,
"grad_norm": 0.17641721665859222,
"learning_rate": 0.00014952694024502429,
"loss": 0.7321,
"step": 6690
},
{
"epoch": 0.17446250792025067,
"grad_norm": 0.17100387811660767,
"learning_rate": 0.00014938183087382354,
"loss": 0.7099,
"step": 6700
},
{
"epoch": 0.17472289972311672,
"grad_norm": 0.16952122747898102,
"learning_rate": 0.00014923658385584084,
"loss": 0.6802,
"step": 6710
},
{
"epoch": 0.17498329152598277,
"grad_norm": 0.1642334908246994,
"learning_rate": 0.00014909119959593734,
"loss": 0.7083,
"step": 6720
},
{
"epoch": 0.17524368332884882,
"grad_norm": 0.16929323971271515,
"learning_rate": 0.00014894567849935673,
"loss": 0.7144,
"step": 6730
},
{
"epoch": 0.17550407513171484,
"grad_norm": 0.180099755525589,
"learning_rate": 0.00014880002097172418,
"loss": 0.7043,
"step": 6740
},
{
"epoch": 0.1757644669345809,
"grad_norm": 0.17287197709083557,
"learning_rate": 0.000148654227419045,
"loss": 0.6891,
"step": 6750
},
{
"epoch": 0.17602485873744694,
"grad_norm": 0.17458143830299377,
"learning_rate": 0.00014850829824770382,
"loss": 0.6889,
"step": 6760
},
{
"epoch": 0.176285250540313,
"grad_norm": 0.1722840964794159,
"learning_rate": 0.00014836223386446317,
"loss": 0.7058,
"step": 6770
},
{
"epoch": 0.17654564234317904,
"grad_norm": 0.17535988986492157,
"learning_rate": 0.0001482160346764625,
"loss": 0.7148,
"step": 6780
},
{
"epoch": 0.1768060341460451,
"grad_norm": 0.16451425850391388,
"learning_rate": 0.00014806970109121709,
"loss": 0.6975,
"step": 6790
},
{
"epoch": 0.17706642594891112,
"grad_norm": 0.16712118685245514,
"learning_rate": 0.00014792323351661671,
"loss": 0.7148,
"step": 6800
},
{
"epoch": 0.17732681775177717,
"grad_norm": 0.17245186865329742,
"learning_rate": 0.0001477766323609248,
"loss": 0.7216,
"step": 6810
},
{
"epoch": 0.17758720955464322,
"grad_norm": 0.17335280776023865,
"learning_rate": 0.00014762989803277686,
"loss": 0.6949,
"step": 6820
},
{
"epoch": 0.17784760135750927,
"grad_norm": 0.16834279894828796,
"learning_rate": 0.0001474830309411799,
"loss": 0.6884,
"step": 6830
},
{
"epoch": 0.17810799316037532,
"grad_norm": 0.17499734461307526,
"learning_rate": 0.0001473360314955108,
"loss": 0.7065,
"step": 6840
},
{
"epoch": 0.17836838496324137,
"grad_norm": 0.17072665691375732,
"learning_rate": 0.00014718890010551543,
"loss": 0.7067,
"step": 6850
},
{
"epoch": 0.17862877676610742,
"grad_norm": 0.192660853266716,
"learning_rate": 0.00014704163718130735,
"loss": 0.7263,
"step": 6860
},
{
"epoch": 0.17888916856897344,
"grad_norm": 0.18734219670295715,
"learning_rate": 0.00014689424313336696,
"loss": 0.7044,
"step": 6870
},
{
"epoch": 0.1791495603718395,
"grad_norm": 0.17117907106876373,
"learning_rate": 0.00014674671837253998,
"loss": 0.7112,
"step": 6880
},
{
"epoch": 0.17940995217470554,
"grad_norm": 0.18425500392913818,
"learning_rate": 0.00014659906331003652,
"loss": 0.6925,
"step": 6890
},
{
"epoch": 0.1796703439775716,
"grad_norm": 0.1706390231847763,
"learning_rate": 0.00014645127835742988,
"loss": 0.6916,
"step": 6900
},
{
"epoch": 0.17993073578043764,
"grad_norm": 0.17749357223510742,
"learning_rate": 0.00014630336392665552,
"loss": 0.7011,
"step": 6910
},
{
"epoch": 0.1801911275833037,
"grad_norm": 0.16545963287353516,
"learning_rate": 0.00014615532043000965,
"loss": 0.7163,
"step": 6920
},
{
"epoch": 0.1804515193861697,
"grad_norm": 0.1756933480501175,
"learning_rate": 0.00014600714828014833,
"loss": 0.6954,
"step": 6930
},
{
"epoch": 0.18071191118903576,
"grad_norm": 0.16752491891384125,
"learning_rate": 0.0001458588478900862,
"loss": 0.6902,
"step": 6940
},
{
"epoch": 0.1809723029919018,
"grad_norm": 0.16998566687107086,
"learning_rate": 0.00014571041967319535,
"loss": 0.6906,
"step": 6950
},
{
"epoch": 0.18123269479476786,
"grad_norm": 0.1744604855775833,
"learning_rate": 0.00014556186404320428,
"loss": 0.6814,
"step": 6960
},
{
"epoch": 0.1814930865976339,
"grad_norm": 0.17596711218357086,
"learning_rate": 0.00014541318141419646,
"loss": 0.7141,
"step": 6970
},
{
"epoch": 0.18175347840049996,
"grad_norm": 0.16803975403308868,
"learning_rate": 0.00014526437220060948,
"loss": 0.6875,
"step": 6980
},
{
"epoch": 0.18201387020336598,
"grad_norm": 0.1845843344926834,
"learning_rate": 0.00014511543681723377,
"loss": 0.6921,
"step": 6990
},
{
"epoch": 0.18227426200623204,
"grad_norm": 0.1783183515071869,
"learning_rate": 0.00014496637567921135,
"loss": 0.6982,
"step": 7000
},
{
"epoch": 0.18253465380909809,
"grad_norm": 0.17463266849517822,
"learning_rate": 0.00014481718920203492,
"loss": 0.7094,
"step": 7010
},
{
"epoch": 0.18279504561196414,
"grad_norm": 0.17779332399368286,
"learning_rate": 0.00014466787780154636,
"loss": 0.6989,
"step": 7020
},
{
"epoch": 0.18305543741483019,
"grad_norm": 0.17536398768424988,
"learning_rate": 0.0001445184418939359,
"loss": 0.7009,
"step": 7030
},
{
"epoch": 0.18331582921769624,
"grad_norm": 0.17731131613254547,
"learning_rate": 0.00014436888189574084,
"loss": 0.6964,
"step": 7040
},
{
"epoch": 0.18357622102056229,
"grad_norm": 0.18129530549049377,
"learning_rate": 0.00014421919822384428,
"loss": 0.6822,
"step": 7050
},
{
"epoch": 0.1838366128234283,
"grad_norm": 0.16552576422691345,
"learning_rate": 0.00014406939129547406,
"loss": 0.7172,
"step": 7060
},
{
"epoch": 0.18409700462629436,
"grad_norm": 0.17690911889076233,
"learning_rate": 0.00014391946152820167,
"loss": 0.6883,
"step": 7070
},
{
"epoch": 0.1843573964291604,
"grad_norm": 0.1851797252893448,
"learning_rate": 0.00014376940933994082,
"loss": 0.6988,
"step": 7080
},
{
"epoch": 0.18461778823202646,
"grad_norm": 0.18322911858558655,
"learning_rate": 0.0001436192351489467,
"loss": 0.689,
"step": 7090
},
{
"epoch": 0.1848781800348925,
"grad_norm": 0.17606942355632782,
"learning_rate": 0.0001434689393738144,
"loss": 0.7034,
"step": 7100
},
{
"epoch": 0.18513857183775856,
"grad_norm": 0.1715836524963379,
"learning_rate": 0.0001433185224334779,
"loss": 0.6939,
"step": 7110
},
{
"epoch": 0.18539896364062458,
"grad_norm": 0.18052712082862854,
"learning_rate": 0.00014316798474720908,
"loss": 0.6858,
"step": 7120
},
{
"epoch": 0.18565935544349063,
"grad_norm": 0.17608265578746796,
"learning_rate": 0.00014301732673461617,
"loss": 0.7166,
"step": 7130
},
{
"epoch": 0.18591974724635668,
"grad_norm": 0.17561575770378113,
"learning_rate": 0.000142866548815643,
"loss": 0.7208,
"step": 7140
},
{
"epoch": 0.18618013904922273,
"grad_norm": 0.17250195145606995,
"learning_rate": 0.00014271565141056746,
"loss": 0.687,
"step": 7150
},
{
"epoch": 0.18644053085208878,
"grad_norm": 0.1769152581691742,
"learning_rate": 0.00014256463494000064,
"loss": 0.7007,
"step": 7160
},
{
"epoch": 0.18670092265495483,
"grad_norm": 0.16784866154193878,
"learning_rate": 0.0001424134998248854,
"loss": 0.6913,
"step": 7170
},
{
"epoch": 0.18696131445782085,
"grad_norm": 0.18374022841453552,
"learning_rate": 0.00014226224648649532,
"loss": 0.7048,
"step": 7180
},
{
"epoch": 0.1872217062606869,
"grad_norm": 0.173793226480484,
"learning_rate": 0.0001421108753464336,
"loss": 0.6877,
"step": 7190
},
{
"epoch": 0.18748209806355295,
"grad_norm": 0.17282545566558838,
"learning_rate": 0.00014195938682663175,
"loss": 0.7132,
"step": 7200
},
{
"epoch": 0.187742489866419,
"grad_norm": 0.16977694630622864,
"learning_rate": 0.00014180778134934843,
"loss": 0.6813,
"step": 7210
},
{
"epoch": 0.18800288166928505,
"grad_norm": 0.18771855533123016,
"learning_rate": 0.00014165605933716836,
"loss": 0.7037,
"step": 7220
},
{
"epoch": 0.1882632734721511,
"grad_norm": 0.17175500094890594,
"learning_rate": 0.00014150422121300105,
"loss": 0.6926,
"step": 7230
},
{
"epoch": 0.18852366527501715,
"grad_norm": 0.17340902984142303,
"learning_rate": 0.00014135226740007967,
"loss": 0.6987,
"step": 7240
},
{
"epoch": 0.18878405707788318,
"grad_norm": 0.16895218193531036,
"learning_rate": 0.00014120019832195986,
"loss": 0.6771,
"step": 7250
},
{
"epoch": 0.18904444888074923,
"grad_norm": 0.17422990500926971,
"learning_rate": 0.00014104801440251863,
"loss": 0.7027,
"step": 7260
},
{
"epoch": 0.18930484068361528,
"grad_norm": 0.1705746352672577,
"learning_rate": 0.00014089571606595294,
"loss": 0.6907,
"step": 7270
},
{
"epoch": 0.18956523248648133,
"grad_norm": 0.16771703958511353,
"learning_rate": 0.0001407433037367788,
"loss": 0.7105,
"step": 7280
},
{
"epoch": 0.18982562428934738,
"grad_norm": 0.1731031835079193,
"learning_rate": 0.00014059077783982993,
"loss": 0.6683,
"step": 7290
},
{
"epoch": 0.19008601609221343,
"grad_norm": 0.18010441958904266,
"learning_rate": 0.00014043813880025658,
"loss": 0.692,
"step": 7300
},
{
"epoch": 0.19034640789507945,
"grad_norm": 0.17735110223293304,
"learning_rate": 0.0001402853870435244,
"loss": 0.6639,
"step": 7310
},
{
"epoch": 0.1906067996979455,
"grad_norm": 0.16834424436092377,
"learning_rate": 0.00014013252299541323,
"loss": 0.6915,
"step": 7320
},
{
"epoch": 0.19086719150081155,
"grad_norm": 0.17520864307880402,
"learning_rate": 0.00013997954708201593,
"loss": 0.6907,
"step": 7330
},
{
"epoch": 0.1911275833036776,
"grad_norm": 0.18474625051021576,
"learning_rate": 0.00013982645972973715,
"loss": 0.7003,
"step": 7340
},
{
"epoch": 0.19138797510654365,
"grad_norm": 0.1774008870124817,
"learning_rate": 0.00013967326136529208,
"loss": 0.7133,
"step": 7350
},
{
"epoch": 0.1916483669094097,
"grad_norm": 0.17432044446468353,
"learning_rate": 0.00013951995241570552,
"loss": 0.7072,
"step": 7360
},
{
"epoch": 0.19190875871227575,
"grad_norm": 0.1843249648809433,
"learning_rate": 0.0001393665333083104,
"loss": 0.6928,
"step": 7370
},
{
"epoch": 0.19216915051514177,
"grad_norm": 0.17402757704257965,
"learning_rate": 0.00013921300447074667,
"loss": 0.7106,
"step": 7380
},
{
"epoch": 0.19242954231800782,
"grad_norm": 0.17242158949375153,
"learning_rate": 0.00013905936633096025,
"loss": 0.6903,
"step": 7390
},
{
"epoch": 0.19268993412087387,
"grad_norm": 0.18488599359989166,
"learning_rate": 0.00013890561931720167,
"loss": 0.688,
"step": 7400
},
{
"epoch": 0.19295032592373992,
"grad_norm": 0.17597128450870514,
"learning_rate": 0.0001387517638580249,
"loss": 0.7062,
"step": 7410
},
{
"epoch": 0.19321071772660597,
"grad_norm": 0.1805078685283661,
"learning_rate": 0.00013859780038228634,
"loss": 0.6999,
"step": 7420
},
{
"epoch": 0.19347110952947202,
"grad_norm": 0.17452898621559143,
"learning_rate": 0.00013844372931914325,
"loss": 0.6961,
"step": 7430
},
{
"epoch": 0.19373150133233805,
"grad_norm": 0.17718560993671417,
"learning_rate": 0.00013828955109805295,
"loss": 0.6909,
"step": 7440
},
{
"epoch": 0.1939918931352041,
"grad_norm": 0.1740763783454895,
"learning_rate": 0.00013813526614877138,
"loss": 0.7126,
"step": 7450
},
{
"epoch": 0.19425228493807015,
"grad_norm": 0.17878217995166779,
"learning_rate": 0.000137980874901352,
"loss": 0.7,
"step": 7460
},
{
"epoch": 0.1945126767409362,
"grad_norm": 0.17267367243766785,
"learning_rate": 0.0001378263777861445,
"loss": 0.6756,
"step": 7470
},
{
"epoch": 0.19477306854380225,
"grad_norm": 0.174414724111557,
"learning_rate": 0.00013767177523379382,
"loss": 0.6948,
"step": 7480
},
{
"epoch": 0.1950334603466683,
"grad_norm": 0.17386887967586517,
"learning_rate": 0.00013751706767523865,
"loss": 0.6848,
"step": 7490
},
{
"epoch": 0.19529385214953432,
"grad_norm": 0.1732412576675415,
"learning_rate": 0.0001373622555417104,
"loss": 0.6914,
"step": 7500
},
{
"epoch": 0.19555424395240037,
"grad_norm": 0.17556026577949524,
"learning_rate": 0.00013720733926473206,
"loss": 0.6866,
"step": 7510
},
{
"epoch": 0.19581463575526642,
"grad_norm": 0.1789248138666153,
"learning_rate": 0.00013705231927611675,
"loss": 0.7071,
"step": 7520
},
{
"epoch": 0.19607502755813247,
"grad_norm": 0.17923463881015778,
"learning_rate": 0.00013689719600796683,
"loss": 0.684,
"step": 7530
},
{
"epoch": 0.19633541936099852,
"grad_norm": 0.18686242401599884,
"learning_rate": 0.00013674196989267245,
"loss": 0.7003,
"step": 7540
},
{
"epoch": 0.19659581116386457,
"grad_norm": 0.1743983030319214,
"learning_rate": 0.00013658664136291046,
"loss": 0.7067,
"step": 7550
},
{
"epoch": 0.19685620296673062,
"grad_norm": 0.17865756154060364,
"learning_rate": 0.0001364312108516432,
"loss": 0.7005,
"step": 7560
},
{
"epoch": 0.19711659476959664,
"grad_norm": 0.17367780208587646,
"learning_rate": 0.0001362756787921173,
"loss": 0.6852,
"step": 7570
},
{
"epoch": 0.1973769865724627,
"grad_norm": 0.17140452563762665,
"learning_rate": 0.00013612004561786232,
"loss": 0.6904,
"step": 7580
},
{
"epoch": 0.19763737837532874,
"grad_norm": 0.17559275031089783,
"learning_rate": 0.00013596431176268978,
"loss": 0.7174,
"step": 7590
},
{
"epoch": 0.1978977701781948,
"grad_norm": 0.17835845053195953,
"learning_rate": 0.00013580847766069183,
"loss": 0.6827,
"step": 7600
},
{
"epoch": 0.19815816198106084,
"grad_norm": 0.1730433702468872,
"learning_rate": 0.00013565254374624,
"loss": 0.7099,
"step": 7610
},
{
"epoch": 0.1984185537839269,
"grad_norm": 0.1799464374780655,
"learning_rate": 0.00013549651045398407,
"loss": 0.6777,
"step": 7620
},
{
"epoch": 0.19867894558679292,
"grad_norm": 0.1781192421913147,
"learning_rate": 0.00013534037821885084,
"loss": 0.7009,
"step": 7630
},
{
"epoch": 0.19893933738965897,
"grad_norm": 0.17447307705879211,
"learning_rate": 0.00013518414747604285,
"loss": 0.7016,
"step": 7640
},
{
"epoch": 0.19919972919252502,
"grad_norm": 0.17640911042690277,
"learning_rate": 0.00013502781866103726,
"loss": 0.6878,
"step": 7650
},
{
"epoch": 0.19946012099539107,
"grad_norm": 0.17320814728736877,
"learning_rate": 0.00013487139220958458,
"loss": 0.6966,
"step": 7660
},
{
"epoch": 0.19972051279825712,
"grad_norm": 0.17917166650295258,
"learning_rate": 0.0001347148685577075,
"loss": 0.6858,
"step": 7670
},
{
"epoch": 0.19998090460112317,
"grad_norm": 0.1751902550458908,
"learning_rate": 0.00013455824814169958,
"loss": 0.7011,
"step": 7680
},
{
"epoch": 0.20024129640398922,
"grad_norm": 0.18323281407356262,
"learning_rate": 0.00013440153139812418,
"loss": 0.6794,
"step": 7690
},
{
"epoch": 0.20050168820685524,
"grad_norm": 0.1677102893590927,
"learning_rate": 0.00013424471876381311,
"loss": 0.6791,
"step": 7700
},
{
"epoch": 0.2007620800097213,
"grad_norm": 0.17628876864910126,
"learning_rate": 0.0001340878106758655,
"loss": 0.6987,
"step": 7710
},
{
"epoch": 0.20102247181258734,
"grad_norm": 0.1839548945426941,
"learning_rate": 0.00013393080757164653,
"loss": 0.6971,
"step": 7720
},
{
"epoch": 0.2012828636154534,
"grad_norm": 0.1783684343099594,
"learning_rate": 0.00013377370988878624,
"loss": 0.6881,
"step": 7730
},
{
"epoch": 0.20154325541831944,
"grad_norm": 0.17916792631149292,
"learning_rate": 0.00013361651806517824,
"loss": 0.698,
"step": 7740
},
{
"epoch": 0.2018036472211855,
"grad_norm": 0.17960959672927856,
"learning_rate": 0.00013345923253897866,
"loss": 0.6858,
"step": 7750
},
{
"epoch": 0.2020640390240515,
"grad_norm": 0.17584697902202606,
"learning_rate": 0.00013330185374860467,
"loss": 0.6835,
"step": 7760
},
{
"epoch": 0.20232443082691756,
"grad_norm": 0.17761898040771484,
"learning_rate": 0.00013314438213273355,
"loss": 0.701,
"step": 7770
},
{
"epoch": 0.2025848226297836,
"grad_norm": 0.16848579049110413,
"learning_rate": 0.00013298681813030122,
"loss": 0.7162,
"step": 7780
},
{
"epoch": 0.20284521443264966,
"grad_norm": 0.17633990943431854,
"learning_rate": 0.00013282916218050118,
"loss": 0.7054,
"step": 7790
},
{
"epoch": 0.2031056062355157,
"grad_norm": 0.1785901039838791,
"learning_rate": 0.00013267141472278316,
"loss": 0.7137,
"step": 7800
},
{
"epoch": 0.20336599803838176,
"grad_norm": 0.16447508335113525,
"learning_rate": 0.00013251357619685207,
"loss": 0.7167,
"step": 7810
},
{
"epoch": 0.20362638984124778,
"grad_norm": 0.17035554349422455,
"learning_rate": 0.00013235564704266653,
"loss": 0.6956,
"step": 7820
},
{
"epoch": 0.20388678164411383,
"grad_norm": 0.18053407967090607,
"learning_rate": 0.00013219762770043787,
"loss": 0.7016,
"step": 7830
},
{
"epoch": 0.20414717344697988,
"grad_norm": 0.18030588328838348,
"learning_rate": 0.00013203951861062873,
"loss": 0.7082,
"step": 7840
},
{
"epoch": 0.20440756524984594,
"grad_norm": 0.17645570635795593,
"learning_rate": 0.00013188132021395204,
"loss": 0.6894,
"step": 7850
},
{
"epoch": 0.20466795705271199,
"grad_norm": 0.1644545942544937,
"learning_rate": 0.00013172303295136948,
"loss": 0.6944,
"step": 7860
},
{
"epoch": 0.20492834885557804,
"grad_norm": 0.1747378408908844,
"learning_rate": 0.0001315646572640906,
"loss": 0.7009,
"step": 7870
},
{
"epoch": 0.20518874065844409,
"grad_norm": 0.17992566525936127,
"learning_rate": 0.00013140619359357139,
"loss": 0.7139,
"step": 7880
},
{
"epoch": 0.2054491324613101,
"grad_norm": 0.17254693806171417,
"learning_rate": 0.000131247642381513,
"loss": 0.6878,
"step": 7890
},
{
"epoch": 0.20570952426417616,
"grad_norm": 0.19822190701961517,
"learning_rate": 0.00013108900406986066,
"loss": 0.7181,
"step": 7900
},
{
"epoch": 0.2059699160670422,
"grad_norm": 0.18208132684230804,
"learning_rate": 0.0001309302791008024,
"loss": 0.7086,
"step": 7910
},
{
"epoch": 0.20623030786990826,
"grad_norm": 0.1721629798412323,
"learning_rate": 0.00013077146791676776,
"loss": 0.6874,
"step": 7920
},
{
"epoch": 0.2064906996727743,
"grad_norm": 0.1776220202445984,
"learning_rate": 0.00013061257096042654,
"loss": 0.6997,
"step": 7930
},
{
"epoch": 0.20675109147564036,
"grad_norm": 0.1767396479845047,
"learning_rate": 0.00013045358867468782,
"loss": 0.6839,
"step": 7940
},
{
"epoch": 0.20701148327850638,
"grad_norm": 0.17351138591766357,
"learning_rate": 0.00013029452150269827,
"loss": 0.7025,
"step": 7950
},
{
"epoch": 0.20727187508137243,
"grad_norm": 0.18614953756332397,
"learning_rate": 0.00013013536988784138,
"loss": 0.7054,
"step": 7960
},
{
"epoch": 0.20753226688423848,
"grad_norm": 0.17871534824371338,
"learning_rate": 0.00012997613427373586,
"loss": 0.6855,
"step": 7970
},
{
"epoch": 0.20779265868710453,
"grad_norm": 0.17882663011550903,
"learning_rate": 0.00012981681510423472,
"loss": 0.6973,
"step": 7980
},
{
"epoch": 0.20805305048997058,
"grad_norm": 0.17025555670261383,
"learning_rate": 0.0001296574128234237,
"loss": 0.6915,
"step": 7990
},
{
"epoch": 0.20831344229283663,
"grad_norm": 0.18960419297218323,
"learning_rate": 0.0001294979278756203,
"loss": 0.6961,
"step": 8000
},
{
"epoch": 0.20857383409570265,
"grad_norm": 0.18628227710723877,
"learning_rate": 0.00012933836070537247,
"loss": 0.6676,
"step": 8010
},
{
"epoch": 0.2088342258985687,
"grad_norm": 0.17117147147655487,
"learning_rate": 0.00012917871175745726,
"loss": 0.6929,
"step": 8020
},
{
"epoch": 0.20909461770143475,
"grad_norm": 0.17737753689289093,
"learning_rate": 0.00012901898147687975,
"loss": 0.7125,
"step": 8030
},
{
"epoch": 0.2093550095043008,
"grad_norm": 0.17105183005332947,
"learning_rate": 0.00012885917030887174,
"loss": 0.6709,
"step": 8040
},
{
"epoch": 0.20961540130716685,
"grad_norm": 0.17960920929908752,
"learning_rate": 0.00012869927869889032,
"loss": 0.6898,
"step": 8050
},
{
"epoch": 0.2098757931100329,
"grad_norm": 0.1776961088180542,
"learning_rate": 0.00012853930709261707,
"loss": 0.6987,
"step": 8060
},
{
"epoch": 0.21013618491289895,
"grad_norm": 0.16933076083660126,
"learning_rate": 0.00012837925593595626,
"loss": 0.6793,
"step": 8070
},
{
"epoch": 0.21039657671576498,
"grad_norm": 0.17447830736637115,
"learning_rate": 0.0001282191256750342,
"loss": 0.69,
"step": 8080
},
{
"epoch": 0.21065696851863103,
"grad_norm": 0.17908291518688202,
"learning_rate": 0.00012805891675619747,
"loss": 0.6805,
"step": 8090
},
{
"epoch": 0.21091736032149708,
"grad_norm": 0.17689751088619232,
"learning_rate": 0.00012789862962601201,
"loss": 0.6993,
"step": 8100
},
{
"epoch": 0.21117775212436313,
"grad_norm": 0.17400488257408142,
"learning_rate": 0.00012773826473126172,
"loss": 0.6836,
"step": 8110
},
{
"epoch": 0.21143814392722918,
"grad_norm": 0.17445240914821625,
"learning_rate": 0.00012757782251894727,
"loss": 0.7105,
"step": 8120
},
{
"epoch": 0.21169853573009523,
"grad_norm": 0.18754400312900543,
"learning_rate": 0.00012741730343628484,
"loss": 0.7093,
"step": 8130
},
{
"epoch": 0.21195892753296125,
"grad_norm": 0.1769934892654419,
"learning_rate": 0.0001272567079307049,
"loss": 0.6922,
"step": 8140
},
{
"epoch": 0.2122193193358273,
"grad_norm": 0.17582187056541443,
"learning_rate": 0.00012709603644985094,
"loss": 0.6955,
"step": 8150
},
{
"epoch": 0.21247971113869335,
"grad_norm": 0.1896626353263855,
"learning_rate": 0.00012693528944157818,
"loss": 0.706,
"step": 8160
},
{
"epoch": 0.2127401029415594,
"grad_norm": 0.17897091805934906,
"learning_rate": 0.0001267744673539524,
"loss": 0.6903,
"step": 8170
},
{
"epoch": 0.21300049474442545,
"grad_norm": 0.1837705820798874,
"learning_rate": 0.0001266135706352487,
"loss": 0.6972,
"step": 8180
},
{
"epoch": 0.2132608865472915,
"grad_norm": 0.17943723499774933,
"learning_rate": 0.0001264525997339501,
"loss": 0.6788,
"step": 8190
},
{
"epoch": 0.21352127835015755,
"grad_norm": 0.18042807281017303,
"learning_rate": 0.0001262915550987465,
"loss": 0.7139,
"step": 8200
},
{
"epoch": 0.21378167015302357,
"grad_norm": 0.1821472942829132,
"learning_rate": 0.00012613043717853324,
"loss": 0.6902,
"step": 8210
},
{
"epoch": 0.21404206195588962,
"grad_norm": 0.18218253552913666,
"learning_rate": 0.00012596924642241004,
"loss": 0.6861,
"step": 8220
},
{
"epoch": 0.21430245375875567,
"grad_norm": 0.18331663310527802,
"learning_rate": 0.00012580798327967952,
"loss": 0.6746,
"step": 8230
},
{
"epoch": 0.21456284556162172,
"grad_norm": 0.1842353492975235,
"learning_rate": 0.0001256466481998461,
"loss": 0.6932,
"step": 8240
},
{
"epoch": 0.21482323736448777,
"grad_norm": 0.16256186366081238,
"learning_rate": 0.00012548524163261485,
"loss": 0.6843,
"step": 8250
},
{
"epoch": 0.21508362916735382,
"grad_norm": 0.17319098114967346,
"learning_rate": 0.00012532376402788988,
"loss": 0.7022,
"step": 8260
},
{
"epoch": 0.21534402097021985,
"grad_norm": 0.1779172569513321,
"learning_rate": 0.00012516221583577355,
"loss": 0.7039,
"step": 8270
},
{
"epoch": 0.2156044127730859,
"grad_norm": 0.17087218165397644,
"learning_rate": 0.00012500059750656474,
"loss": 0.6917,
"step": 8280
},
{
"epoch": 0.21586480457595195,
"grad_norm": 0.16772834956645966,
"learning_rate": 0.00012483890949075801,
"loss": 0.6848,
"step": 8290
},
{
"epoch": 0.216125196378818,
"grad_norm": 0.17318886518478394,
"learning_rate": 0.0001246771522390421,
"loss": 0.6973,
"step": 8300
},
{
"epoch": 0.21638558818168405,
"grad_norm": 0.16938996315002441,
"learning_rate": 0.00012451532620229872,
"loss": 0.7016,
"step": 8310
},
{
"epoch": 0.2166459799845501,
"grad_norm": 0.17802828550338745,
"learning_rate": 0.0001243534318316013,
"loss": 0.6871,
"step": 8320
},
{
"epoch": 0.21690637178741612,
"grad_norm": 0.17595522105693817,
"learning_rate": 0.00012419146957821383,
"loss": 0.7055,
"step": 8330
},
{
"epoch": 0.21716676359028217,
"grad_norm": 0.1841190904378891,
"learning_rate": 0.0001240294398935894,
"loss": 0.7154,
"step": 8340
},
{
"epoch": 0.21742715539314822,
"grad_norm": 0.16829776763916016,
"learning_rate": 0.0001238673432293692,
"loss": 0.6925,
"step": 8350
},
{
"epoch": 0.21768754719601427,
"grad_norm": 0.1729755699634552,
"learning_rate": 0.00012370518003738091,
"loss": 0.7118,
"step": 8360
},
{
"epoch": 0.21794793899888032,
"grad_norm": 0.16530723869800568,
"learning_rate": 0.00012354295076963786,
"loss": 0.7029,
"step": 8370
},
{
"epoch": 0.21820833080174637,
"grad_norm": 0.17311285436153412,
"learning_rate": 0.00012338065587833748,
"loss": 0.7013,
"step": 8380
},
{
"epoch": 0.21846872260461242,
"grad_norm": 0.1771146059036255,
"learning_rate": 0.00012321829581586003,
"loss": 0.6997,
"step": 8390
},
{
"epoch": 0.21872911440747844,
"grad_norm": 0.1830943077802658,
"learning_rate": 0.0001230558710347676,
"loss": 0.6945,
"step": 8400
},
{
"epoch": 0.2189895062103445,
"grad_norm": 0.1760828197002411,
"learning_rate": 0.00012289338198780253,
"loss": 0.6888,
"step": 8410
},
{
"epoch": 0.21924989801321054,
"grad_norm": 0.1744048148393631,
"learning_rate": 0.00012273082912788637,
"loss": 0.676,
"step": 8420
},
{
"epoch": 0.2195102898160766,
"grad_norm": 0.17182762920856476,
"learning_rate": 0.00012256821290811855,
"loss": 0.707,
"step": 8430
},
{
"epoch": 0.21977068161894264,
"grad_norm": 0.1801503300666809,
"learning_rate": 0.00012240553378177504,
"loss": 0.6907,
"step": 8440
},
{
"epoch": 0.2200310734218087,
"grad_norm": 0.1753336787223816,
"learning_rate": 0.0001222427922023072,
"loss": 0.7034,
"step": 8450
},
{
"epoch": 0.22029146522467472,
"grad_norm": 0.180914506316185,
"learning_rate": 0.00012207998862334048,
"loss": 0.6773,
"step": 8460
},
{
"epoch": 0.22055185702754077,
"grad_norm": 0.17284120619297028,
"learning_rate": 0.00012191712349867315,
"loss": 0.6983,
"step": 8470
},
{
"epoch": 0.22081224883040682,
"grad_norm": 0.18915420770645142,
"learning_rate": 0.00012175419728227498,
"loss": 0.7106,
"step": 8480
},
{
"epoch": 0.22107264063327287,
"grad_norm": 0.178519144654274,
"learning_rate": 0.00012159121042828612,
"loss": 0.7054,
"step": 8490
},
{
"epoch": 0.22133303243613892,
"grad_norm": 0.17323723435401917,
"learning_rate": 0.00012142816339101562,
"loss": 0.6965,
"step": 8500
},
{
"epoch": 0.22159342423900497,
"grad_norm": 0.23265057802200317,
"learning_rate": 0.00012126505662494041,
"loss": 0.6914,
"step": 8510
},
{
"epoch": 0.221853816041871,
"grad_norm": 0.1790781468153,
"learning_rate": 0.00012110189058470382,
"loss": 0.6911,
"step": 8520
},
{
"epoch": 0.22211420784473704,
"grad_norm": 0.17768746614456177,
"learning_rate": 0.00012093866572511443,
"loss": 0.6829,
"step": 8530
},
{
"epoch": 0.2223745996476031,
"grad_norm": 0.17421717941761017,
"learning_rate": 0.00012077538250114478,
"loss": 0.6963,
"step": 8540
},
{
"epoch": 0.22263499145046914,
"grad_norm": 0.1880684494972229,
"learning_rate": 0.0001206120413679301,
"loss": 0.6946,
"step": 8550
},
{
"epoch": 0.2228953832533352,
"grad_norm": 0.17597487568855286,
"learning_rate": 0.00012044864278076698,
"loss": 0.7016,
"step": 8560
},
{
"epoch": 0.22315577505620124,
"grad_norm": 0.17353573441505432,
"learning_rate": 0.00012028518719511225,
"loss": 0.6771,
"step": 8570
},
{
"epoch": 0.2234161668590673,
"grad_norm": 0.18333446979522705,
"learning_rate": 0.00012012167506658156,
"loss": 0.6696,
"step": 8580
},
{
"epoch": 0.2236765586619333,
"grad_norm": 0.1803252398967743,
"learning_rate": 0.00011995810685094813,
"loss": 0.6988,
"step": 8590
},
{
"epoch": 0.22393695046479936,
"grad_norm": 0.17487098276615143,
"learning_rate": 0.0001197944830041416,
"loss": 0.7198,
"step": 8600
},
{
"epoch": 0.2241973422676654,
"grad_norm": 0.174115851521492,
"learning_rate": 0.00011963080398224661,
"loss": 0.6904,
"step": 8610
},
{
"epoch": 0.22445773407053146,
"grad_norm": 0.17489659786224365,
"learning_rate": 0.0001194670702415017,
"loss": 0.6738,
"step": 8620
},
{
"epoch": 0.2247181258733975,
"grad_norm": 0.17910204827785492,
"learning_rate": 0.00011930328223829779,
"loss": 0.7042,
"step": 8630
},
{
"epoch": 0.22497851767626356,
"grad_norm": 0.17602995038032532,
"learning_rate": 0.00011913944042917711,
"loss": 0.7068,
"step": 8640
},
{
"epoch": 0.22523890947912958,
"grad_norm": 0.18906055390834808,
"learning_rate": 0.0001189755452708319,
"loss": 0.7128,
"step": 8650
},
{
"epoch": 0.22549930128199563,
"grad_norm": 0.17224588990211487,
"learning_rate": 0.0001188115972201031,
"loss": 0.7158,
"step": 8660
},
{
"epoch": 0.22575969308486168,
"grad_norm": 0.18284164369106293,
"learning_rate": 0.00011864759673397901,
"loss": 0.6838,
"step": 8670
},
{
"epoch": 0.22602008488772773,
"grad_norm": 0.17522230744361877,
"learning_rate": 0.00011848354426959417,
"loss": 0.6832,
"step": 8680
},
{
"epoch": 0.22628047669059378,
"grad_norm": 0.17182756960391998,
"learning_rate": 0.00011831944028422798,
"loss": 0.6911,
"step": 8690
},
{
"epoch": 0.22654086849345983,
"grad_norm": 0.17457079887390137,
"learning_rate": 0.00011815528523530349,
"loss": 0.692,
"step": 8700
},
{
"epoch": 0.22680126029632589,
"grad_norm": 0.17242825031280518,
"learning_rate": 0.00011799107958038593,
"loss": 0.6949,
"step": 8710
},
{
"epoch": 0.2270616520991919,
"grad_norm": 0.18559910356998444,
"learning_rate": 0.0001178268237771818,
"loss": 0.6802,
"step": 8720
},
{
"epoch": 0.22732204390205796,
"grad_norm": 0.18594273924827576,
"learning_rate": 0.00011766251828353722,
"loss": 0.7008,
"step": 8730
},
{
"epoch": 0.227582435704924,
"grad_norm": 0.17495504021644592,
"learning_rate": 0.00011749816355743694,
"loss": 0.7119,
"step": 8740
},
{
"epoch": 0.22784282750779006,
"grad_norm": 0.17004092037677765,
"learning_rate": 0.00011733376005700283,
"loss": 0.685,
"step": 8750
},
{
"epoch": 0.2281032193106561,
"grad_norm": 0.17227309942245483,
"learning_rate": 0.00011716930824049274,
"loss": 0.6895,
"step": 8760
},
{
"epoch": 0.22836361111352216,
"grad_norm": 0.18226534128189087,
"learning_rate": 0.00011700480856629929,
"loss": 0.6874,
"step": 8770
},
{
"epoch": 0.22862400291638818,
"grad_norm": 0.17005032300949097,
"learning_rate": 0.0001168402614929484,
"loss": 0.693,
"step": 8780
},
{
"epoch": 0.22888439471925423,
"grad_norm": 0.1789228320121765,
"learning_rate": 0.0001166756674790981,
"loss": 0.6952,
"step": 8790
},
{
"epoch": 0.22914478652212028,
"grad_norm": 0.17560814321041107,
"learning_rate": 0.00011651102698353738,
"loss": 0.7052,
"step": 8800
},
{
"epoch": 0.22940517832498633,
"grad_norm": 0.18214547634124756,
"learning_rate": 0.0001163463404651846,
"loss": 0.6864,
"step": 8810
},
{
"epoch": 0.22966557012785238,
"grad_norm": 0.17090913653373718,
"learning_rate": 0.00011618160838308663,
"loss": 0.6847,
"step": 8820
},
{
"epoch": 0.22992596193071843,
"grad_norm": 0.1777815967798233,
"learning_rate": 0.00011601683119641714,
"loss": 0.7035,
"step": 8830
},
{
"epoch": 0.23018635373358445,
"grad_norm": 0.1708465963602066,
"learning_rate": 0.00011585200936447567,
"loss": 0.6939,
"step": 8840
},
{
"epoch": 0.2304467455364505,
"grad_norm": 0.1829666793346405,
"learning_rate": 0.00011568714334668612,
"loss": 0.6898,
"step": 8850
},
{
"epoch": 0.23070713733931655,
"grad_norm": 0.17379315197467804,
"learning_rate": 0.0001155222336025956,
"loss": 0.7003,
"step": 8860
},
{
"epoch": 0.2309675291421826,
"grad_norm": 0.17232435941696167,
"learning_rate": 0.00011535728059187307,
"loss": 0.6967,
"step": 8870
},
{
"epoch": 0.23122792094504865,
"grad_norm": 0.1840342879295349,
"learning_rate": 0.00011519228477430814,
"loss": 0.695,
"step": 8880
},
{
"epoch": 0.2314883127479147,
"grad_norm": 0.17518925666809082,
"learning_rate": 0.00011502724660980968,
"loss": 0.6982,
"step": 8890
},
{
"epoch": 0.23174870455078075,
"grad_norm": 0.18575233221054077,
"learning_rate": 0.00011486216655840463,
"loss": 0.722,
"step": 8900
},
{
"epoch": 0.23200909635364678,
"grad_norm": 0.18776971101760864,
"learning_rate": 0.00011469704508023667,
"loss": 0.7283,
"step": 8910
},
{
"epoch": 0.23226948815651283,
"grad_norm": 0.17691543698310852,
"learning_rate": 0.00011453188263556493,
"loss": 0.6796,
"step": 8920
},
{
"epoch": 0.23252987995937888,
"grad_norm": 0.17249678075313568,
"learning_rate": 0.00011436667968476284,
"loss": 0.6891,
"step": 8930
},
{
"epoch": 0.23279027176224493,
"grad_norm": 0.17315714061260223,
"learning_rate": 0.00011420143668831662,
"loss": 0.6844,
"step": 8940
},
{
"epoch": 0.23305066356511098,
"grad_norm": 0.17669366300106049,
"learning_rate": 0.00011403615410682416,
"loss": 0.6848,
"step": 8950
},
{
"epoch": 0.23331105536797703,
"grad_norm": 0.1768043041229248,
"learning_rate": 0.00011387083240099364,
"loss": 0.6887,
"step": 8960
},
{
"epoch": 0.23357144717084305,
"grad_norm": 0.17968913912773132,
"learning_rate": 0.00011370547203164237,
"loss": 0.7092,
"step": 8970
},
{
"epoch": 0.2338318389737091,
"grad_norm": 0.1745702028274536,
"learning_rate": 0.0001135400734596954,
"loss": 0.6812,
"step": 8980
},
{
"epoch": 0.23409223077657515,
"grad_norm": 0.1707821488380432,
"learning_rate": 0.00011337463714618423,
"loss": 0.6685,
"step": 8990
},
{
"epoch": 0.2343526225794412,
"grad_norm": 0.1789083033800125,
"learning_rate": 0.00011320916355224565,
"loss": 0.6746,
"step": 9000
},
{
"epoch": 0.23461301438230725,
"grad_norm": 0.17847011983394623,
"learning_rate": 0.00011304365313912028,
"loss": 0.695,
"step": 9010
},
{
"epoch": 0.2348734061851733,
"grad_norm": 0.18885819613933563,
"learning_rate": 0.00011287810636815139,
"loss": 0.698,
"step": 9020
},
{
"epoch": 0.23513379798803935,
"grad_norm": 0.17764124274253845,
"learning_rate": 0.00011271252370078363,
"loss": 0.6892,
"step": 9030
},
{
"epoch": 0.23539418979090537,
"grad_norm": 0.17720068991184235,
"learning_rate": 0.0001125469055985616,
"loss": 0.6926,
"step": 9040
},
{
"epoch": 0.23565458159377142,
"grad_norm": 0.17689958214759827,
"learning_rate": 0.00011238125252312888,
"loss": 0.6987,
"step": 9050
},
{
"epoch": 0.23591497339663747,
"grad_norm": 0.19373558461666107,
"learning_rate": 0.00011221556493622627,
"loss": 0.687,
"step": 9060
},
{
"epoch": 0.23617536519950352,
"grad_norm": 0.1779530942440033,
"learning_rate": 0.000112049843299691,
"loss": 0.6735,
"step": 9070
},
{
"epoch": 0.23643575700236957,
"grad_norm": 0.1853463500738144,
"learning_rate": 0.00011188408807545505,
"loss": 0.6765,
"step": 9080
},
{
"epoch": 0.23669614880523562,
"grad_norm": 0.1809043437242508,
"learning_rate": 0.00011171829972554415,
"loss": 0.6775,
"step": 9090
},
{
"epoch": 0.23695654060810165,
"grad_norm": 0.16961175203323364,
"learning_rate": 0.00011155247871207624,
"loss": 0.6947,
"step": 9100
},
{
"epoch": 0.2372169324109677,
"grad_norm": 0.1882382184267044,
"learning_rate": 0.0001113866254972604,
"loss": 0.6855,
"step": 9110
},
{
"epoch": 0.23747732421383375,
"grad_norm": 0.18256156146526337,
"learning_rate": 0.00011122074054339537,
"loss": 0.7131,
"step": 9120
},
{
"epoch": 0.2377377160166998,
"grad_norm": 0.18344970047473907,
"learning_rate": 0.00011105482431286852,
"loss": 0.6596,
"step": 9130
},
{
"epoch": 0.23799810781956585,
"grad_norm": 0.18037189543247223,
"learning_rate": 0.00011088887726815422,
"loss": 0.6853,
"step": 9140
},
{
"epoch": 0.2382584996224319,
"grad_norm": 0.17328627407550812,
"learning_rate": 0.00011072289987181288,
"loss": 0.6833,
"step": 9150
},
{
"epoch": 0.23851889142529792,
"grad_norm": 0.1846616417169571,
"learning_rate": 0.00011055689258648938,
"loss": 0.6809,
"step": 9160
},
{
"epoch": 0.23877928322816397,
"grad_norm": 0.17733514308929443,
"learning_rate": 0.00011039085587491203,
"loss": 0.7013,
"step": 9170
},
{
"epoch": 0.23903967503103002,
"grad_norm": 0.17087826132774353,
"learning_rate": 0.0001102247901998911,
"loss": 0.6808,
"step": 9180
},
{
"epoch": 0.23930006683389607,
"grad_norm": 0.17394234240055084,
"learning_rate": 0.00011005869602431758,
"loss": 0.6955,
"step": 9190
},
{
"epoch": 0.23956045863676212,
"grad_norm": 0.1748206615447998,
"learning_rate": 0.00010989257381116194,
"loss": 0.6944,
"step": 9200
},
{
"epoch": 0.23982085043962817,
"grad_norm": 0.1819637566804886,
"learning_rate": 0.00010972642402347277,
"loss": 0.6875,
"step": 9210
},
{
"epoch": 0.24008124224249422,
"grad_norm": 0.17814664542675018,
"learning_rate": 0.00010956024712437558,
"loss": 0.6864,
"step": 9220
},
{
"epoch": 0.24034163404536024,
"grad_norm": 0.17680172622203827,
"learning_rate": 0.00010939404357707141,
"loss": 0.6727,
"step": 9230
},
{
"epoch": 0.2406020258482263,
"grad_norm": 0.175360307097435,
"learning_rate": 0.00010922781384483553,
"loss": 0.6972,
"step": 9240
},
{
"epoch": 0.24086241765109234,
"grad_norm": 0.1745460033416748,
"learning_rate": 0.0001090615583910163,
"loss": 0.6933,
"step": 9250
},
{
"epoch": 0.2411228094539584,
"grad_norm": 0.1743789166212082,
"learning_rate": 0.00010889527767903366,
"loss": 0.6719,
"step": 9260
},
{
"epoch": 0.24138320125682444,
"grad_norm": 0.17672204971313477,
"learning_rate": 0.00010872897217237812,
"loss": 0.6705,
"step": 9270
},
{
"epoch": 0.2416435930596905,
"grad_norm": 0.16973887383937836,
"learning_rate": 0.00010856264233460907,
"loss": 0.6772,
"step": 9280
},
{
"epoch": 0.24190398486255651,
"grad_norm": 0.17453502118587494,
"learning_rate": 0.00010839628862935392,
"loss": 0.6856,
"step": 9290
},
{
"epoch": 0.24216437666542256,
"grad_norm": 0.17804546654224396,
"learning_rate": 0.00010822991152030655,
"loss": 0.6766,
"step": 9300
},
{
"epoch": 0.24242476846828862,
"grad_norm": 0.17404384911060333,
"learning_rate": 0.00010806351147122605,
"loss": 0.6914,
"step": 9310
},
{
"epoch": 0.24268516027115467,
"grad_norm": 0.17581741511821747,
"learning_rate": 0.00010789708894593547,
"loss": 0.6933,
"step": 9320
},
{
"epoch": 0.24294555207402072,
"grad_norm": 0.18238358199596405,
"learning_rate": 0.00010773064440832047,
"loss": 0.7241,
"step": 9330
},
{
"epoch": 0.24320594387688677,
"grad_norm": 0.17800027132034302,
"learning_rate": 0.00010756417832232819,
"loss": 0.6881,
"step": 9340
},
{
"epoch": 0.2434663356797528,
"grad_norm": 0.17522063851356506,
"learning_rate": 0.00010739769115196566,
"loss": 0.6916,
"step": 9350
},
{
"epoch": 0.24372672748261884,
"grad_norm": 0.17781999707221985,
"learning_rate": 0.00010723118336129878,
"loss": 0.6809,
"step": 9360
},
{
"epoch": 0.2439871192854849,
"grad_norm": 0.1767900288105011,
"learning_rate": 0.00010706465541445095,
"loss": 0.6966,
"step": 9370
},
{
"epoch": 0.24424751108835094,
"grad_norm": 0.19158516824245453,
"learning_rate": 0.00010689810777560174,
"loss": 0.6949,
"step": 9380
},
{
"epoch": 0.244507902891217,
"grad_norm": 0.1787518560886383,
"learning_rate": 0.00010673154090898552,
"loss": 0.6922,
"step": 9390
},
{
"epoch": 0.24476829469408304,
"grad_norm": 0.17122618854045868,
"learning_rate": 0.00010656495527889039,
"loss": 0.6878,
"step": 9400
},
{
"epoch": 0.2450286864969491,
"grad_norm": 0.18457964062690735,
"learning_rate": 0.00010639835134965664,
"loss": 0.6723,
"step": 9410
},
{
"epoch": 0.2452890782998151,
"grad_norm": 0.17373178899288177,
"learning_rate": 0.00010623172958567563,
"loss": 0.6954,
"step": 9420
},
{
"epoch": 0.24554947010268116,
"grad_norm": 0.18350104987621307,
"learning_rate": 0.0001060650904513884,
"loss": 0.6824,
"step": 9430
},
{
"epoch": 0.2458098619055472,
"grad_norm": 0.17673034965991974,
"learning_rate": 0.00010589843441128445,
"loss": 0.6968,
"step": 9440
},
{
"epoch": 0.24607025370841326,
"grad_norm": 0.17909649014472961,
"learning_rate": 0.00010573176192990032,
"loss": 0.7048,
"step": 9450
},
{
"epoch": 0.2463306455112793,
"grad_norm": 0.18223048746585846,
"learning_rate": 0.00010556507347181854,
"loss": 0.6939,
"step": 9460
},
{
"epoch": 0.24659103731414536,
"grad_norm": 0.17868250608444214,
"learning_rate": 0.00010539836950166595,
"loss": 0.7082,
"step": 9470
},
{
"epoch": 0.24685142911701138,
"grad_norm": 0.1783342808485031,
"learning_rate": 0.00010523165048411283,
"loss": 0.6903,
"step": 9480
},
{
"epoch": 0.24711182091987743,
"grad_norm": 0.17360354959964752,
"learning_rate": 0.00010506491688387127,
"loss": 0.6635,
"step": 9490
},
{
"epoch": 0.24737221272274348,
"grad_norm": 0.18421316146850586,
"learning_rate": 0.0001048981691656941,
"loss": 0.6848,
"step": 9500
},
{
"epoch": 0.24763260452560953,
"grad_norm": 0.1938525140285492,
"learning_rate": 0.00010473140779437341,
"loss": 0.7,
"step": 9510
},
{
"epoch": 0.24789299632847558,
"grad_norm": 0.17088712751865387,
"learning_rate": 0.00010456463323473942,
"loss": 0.6827,
"step": 9520
},
{
"epoch": 0.24815338813134163,
"grad_norm": 0.17812180519104004,
"learning_rate": 0.00010439784595165909,
"loss": 0.6662,
"step": 9530
},
{
"epoch": 0.24841377993420768,
"grad_norm": 0.18220531940460205,
"learning_rate": 0.00010423104641003484,
"loss": 0.6878,
"step": 9540
},
{
"epoch": 0.2486741717370737,
"grad_norm": 0.17387458682060242,
"learning_rate": 0.00010406423507480325,
"loss": 0.6896,
"step": 9550
},
{
"epoch": 0.24893456353993976,
"grad_norm": 0.16997171938419342,
"learning_rate": 0.00010389741241093382,
"loss": 0.6982,
"step": 9560
},
{
"epoch": 0.2491949553428058,
"grad_norm": 0.18163439631462097,
"learning_rate": 0.00010373057888342755,
"loss": 0.6769,
"step": 9570
},
{
"epoch": 0.24945534714567186,
"grad_norm": 0.1709483563899994,
"learning_rate": 0.00010356373495731579,
"loss": 0.6935,
"step": 9580
},
{
"epoch": 0.2497157389485379,
"grad_norm": 0.17501015961170197,
"learning_rate": 0.00010339688109765882,
"loss": 0.6961,
"step": 9590
},
{
"epoch": 0.24997613075140396,
"grad_norm": 0.177652046084404,
"learning_rate": 0.00010323001776954466,
"loss": 0.687,
"step": 9600
},
{
"epoch": 0.25023652255427,
"grad_norm": 0.16977092623710632,
"learning_rate": 0.00010306314543808768,
"loss": 0.6758,
"step": 9610
},
{
"epoch": 0.25049691435713606,
"grad_norm": 0.17998236417770386,
"learning_rate": 0.00010289626456842743,
"loss": 0.7022,
"step": 9620
},
{
"epoch": 0.2507573061600021,
"grad_norm": 0.17297130823135376,
"learning_rate": 0.00010272937562572714,
"loss": 0.6614,
"step": 9630
},
{
"epoch": 0.2510176979628681,
"grad_norm": 0.1776668131351471,
"learning_rate": 0.00010256247907517258,
"loss": 0.6723,
"step": 9640
},
{
"epoch": 0.2512780897657342,
"grad_norm": 0.17850397527217865,
"learning_rate": 0.00010239557538197077,
"loss": 0.6856,
"step": 9650
},
{
"epoch": 0.2515384815686002,
"grad_norm": 0.17861175537109375,
"learning_rate": 0.00010222866501134864,
"loss": 0.6912,
"step": 9660
},
{
"epoch": 0.2517988733714663,
"grad_norm": 0.18047182261943817,
"learning_rate": 0.00010206174842855167,
"loss": 0.7047,
"step": 9670
},
{
"epoch": 0.2520592651743323,
"grad_norm": 0.1728067547082901,
"learning_rate": 0.00010189482609884269,
"loss": 0.7061,
"step": 9680
},
{
"epoch": 0.2523196569771984,
"grad_norm": 0.1772705316543579,
"learning_rate": 0.00010172789848750057,
"loss": 0.6817,
"step": 9690
},
{
"epoch": 0.2525800487800644,
"grad_norm": 0.17873169481754303,
"learning_rate": 0.00010156096605981889,
"loss": 0.682,
"step": 9700
},
{
"epoch": 0.2528404405829304,
"grad_norm": 0.1827894151210785,
"learning_rate": 0.00010139402928110463,
"loss": 0.6636,
"step": 9710
},
{
"epoch": 0.2531008323857965,
"grad_norm": 0.16760335862636566,
"learning_rate": 0.00010122708861667689,
"loss": 0.6725,
"step": 9720
},
{
"epoch": 0.2533612241886625,
"grad_norm": 0.18811176717281342,
"learning_rate": 0.00010106014453186568,
"loss": 0.6856,
"step": 9730
},
{
"epoch": 0.2536216159915286,
"grad_norm": 0.1821964532136917,
"learning_rate": 0.00010089319749201043,
"loss": 0.6736,
"step": 9740
},
{
"epoch": 0.2538820077943946,
"grad_norm": 0.1711428463459015,
"learning_rate": 0.00010072624796245895,
"loss": 0.6811,
"step": 9750
},
{
"epoch": 0.2541423995972607,
"grad_norm": 0.1852736473083496,
"learning_rate": 0.00010055929640856583,
"loss": 0.703,
"step": 9760
},
{
"epoch": 0.2544027914001267,
"grad_norm": 0.18267115950584412,
"learning_rate": 0.00010039234329569145,
"loss": 0.6874,
"step": 9770
},
{
"epoch": 0.25466318320299275,
"grad_norm": 0.1899646520614624,
"learning_rate": 0.00010022538908920039,
"loss": 0.6655,
"step": 9780
},
{
"epoch": 0.2549235750058588,
"grad_norm": 0.18259014189243317,
"learning_rate": 0.00010005843425446042,
"loss": 0.6755,
"step": 9790
},
{
"epoch": 0.25518396680872485,
"grad_norm": 0.18706420063972473,
"learning_rate": 9.989147925684094e-05,
"loss": 0.678,
"step": 9800
},
{
"epoch": 0.2554443586115909,
"grad_norm": 0.17754283547401428,
"learning_rate": 9.97245245617119e-05,
"loss": 0.6793,
"step": 9810
},
{
"epoch": 0.25570475041445695,
"grad_norm": 0.16971617937088013,
"learning_rate": 9.955757063444235e-05,
"loss": 0.6864,
"step": 9820
},
{
"epoch": 0.25596514221732297,
"grad_norm": 0.1726657599210739,
"learning_rate": 9.939061794039922e-05,
"loss": 0.6751,
"step": 9830
},
{
"epoch": 0.25622553402018905,
"grad_norm": 0.18278145790100098,
"learning_rate": 9.922366694494602e-05,
"loss": 0.7024,
"step": 9840
},
{
"epoch": 0.25648592582305507,
"grad_norm": 0.1729104220867157,
"learning_rate": 9.905671811344141e-05,
"loss": 0.684,
"step": 9850
},
{
"epoch": 0.25674631762592115,
"grad_norm": 0.18848705291748047,
"learning_rate": 9.888977191123825e-05,
"loss": 0.6933,
"step": 9860
},
{
"epoch": 0.2570067094287872,
"grad_norm": 0.19198398292064667,
"learning_rate": 9.872282880368186e-05,
"loss": 0.6784,
"step": 9870
},
{
"epoch": 0.25726710123165325,
"grad_norm": 0.17954795062541962,
"learning_rate": 9.855588925610898e-05,
"loss": 0.6959,
"step": 9880
},
{
"epoch": 0.2575274930345193,
"grad_norm": 0.18434350192546844,
"learning_rate": 9.838895373384649e-05,
"loss": 0.7033,
"step": 9890
},
{
"epoch": 0.2577878848373853,
"grad_norm": 0.17057251930236816,
"learning_rate": 9.822202270221002e-05,
"loss": 0.674,
"step": 9900
},
{
"epoch": 0.2580482766402514,
"grad_norm": 0.18517877161502838,
"learning_rate": 9.805509662650268e-05,
"loss": 0.6922,
"step": 9910
},
{
"epoch": 0.2583086684431174,
"grad_norm": 0.1838076114654541,
"learning_rate": 9.788817597201373e-05,
"loss": 0.6831,
"step": 9920
},
{
"epoch": 0.2585690602459835,
"grad_norm": 0.17518235743045807,
"learning_rate": 9.772126120401734e-05,
"loss": 0.7101,
"step": 9930
},
{
"epoch": 0.2588294520488495,
"grad_norm": 0.18193618953227997,
"learning_rate": 9.755435278777136e-05,
"loss": 0.6702,
"step": 9940
},
{
"epoch": 0.2590898438517156,
"grad_norm": 0.1757705956697464,
"learning_rate": 9.738745118851581e-05,
"loss": 0.6949,
"step": 9950
},
{
"epoch": 0.2593502356545816,
"grad_norm": 0.1771358698606491,
"learning_rate": 9.722055687147174e-05,
"loss": 0.6654,
"step": 9960
},
{
"epoch": 0.2596106274574476,
"grad_norm": 0.17536123096942902,
"learning_rate": 9.705367030183992e-05,
"loss": 0.6797,
"step": 9970
},
{
"epoch": 0.2598710192603137,
"grad_norm": 0.1831720471382141,
"learning_rate": 9.688679194479955e-05,
"loss": 0.696,
"step": 9980
},
{
"epoch": 0.2601314110631797,
"grad_norm": 0.17759549617767334,
"learning_rate": 9.671992226550692e-05,
"loss": 0.7069,
"step": 9990
},
{
"epoch": 0.2603918028660458,
"grad_norm": 0.17348560690879822,
"learning_rate": 9.65530617290941e-05,
"loss": 0.6808,
"step": 10000
},
{
"epoch": 0.2606521946689118,
"grad_norm": 0.1791519969701767,
"learning_rate": 9.638621080066766e-05,
"loss": 0.6982,
"step": 10010
},
{
"epoch": 0.26091258647177784,
"grad_norm": 0.18374033272266388,
"learning_rate": 9.621936994530751e-05,
"loss": 0.6713,
"step": 10020
},
{
"epoch": 0.2611729782746439,
"grad_norm": 0.18197116255760193,
"learning_rate": 9.605253962806537e-05,
"loss": 0.6924,
"step": 10030
},
{
"epoch": 0.26143337007750994,
"grad_norm": 0.17887791991233826,
"learning_rate": 9.588572031396358e-05,
"loss": 0.6999,
"step": 10040
},
{
"epoch": 0.261693761880376,
"grad_norm": 0.17189468443393707,
"learning_rate": 9.571891246799388e-05,
"loss": 0.6617,
"step": 10050
},
{
"epoch": 0.26195415368324204,
"grad_norm": 0.17954093217849731,
"learning_rate": 9.555211655511596e-05,
"loss": 0.6809,
"step": 10060
},
{
"epoch": 0.2622145454861081,
"grad_norm": 0.1888546645641327,
"learning_rate": 9.538533304025639e-05,
"loss": 0.6989,
"step": 10070
},
{
"epoch": 0.26247493728897414,
"grad_norm": 0.17998704314231873,
"learning_rate": 9.521856238830702e-05,
"loss": 0.6715,
"step": 10080
},
{
"epoch": 0.26273532909184016,
"grad_norm": 0.17122139036655426,
"learning_rate": 9.505180506412395e-05,
"loss": 0.6902,
"step": 10090
},
{
"epoch": 0.26299572089470624,
"grad_norm": 0.17990705370903015,
"learning_rate": 9.488506153252599e-05,
"loss": 0.6799,
"step": 10100
},
{
"epoch": 0.26325611269757226,
"grad_norm": 0.17561420798301697,
"learning_rate": 9.471833225829378e-05,
"loss": 0.681,
"step": 10110
},
{
"epoch": 0.26351650450043834,
"grad_norm": 0.25144341588020325,
"learning_rate": 9.455161770616791e-05,
"loss": 0.6767,
"step": 10120
},
{
"epoch": 0.26377689630330436,
"grad_norm": 0.17986753582954407,
"learning_rate": 9.438491834084815e-05,
"loss": 0.6865,
"step": 10130
},
{
"epoch": 0.26403728810617044,
"grad_norm": 0.18176795542240143,
"learning_rate": 9.42182346269918e-05,
"loss": 0.6711,
"step": 10140
},
{
"epoch": 0.26429767990903646,
"grad_norm": 0.18171070516109467,
"learning_rate": 9.405156702921266e-05,
"loss": 0.6947,
"step": 10150
},
{
"epoch": 0.2645580717119025,
"grad_norm": 0.17113040387630463,
"learning_rate": 9.388491601207951e-05,
"loss": 0.6726,
"step": 10160
},
{
"epoch": 0.26481846351476857,
"grad_norm": 0.16792860627174377,
"learning_rate": 9.371828204011496e-05,
"loss": 0.6856,
"step": 10170
},
{
"epoch": 0.2650788553176346,
"grad_norm": 0.18093881011009216,
"learning_rate": 9.355166557779402e-05,
"loss": 0.674,
"step": 10180
},
{
"epoch": 0.26533924712050067,
"grad_norm": 0.17508314549922943,
"learning_rate": 9.338506708954307e-05,
"loss": 0.6845,
"step": 10190
},
{
"epoch": 0.2655996389233667,
"grad_norm": 0.18526801466941833,
"learning_rate": 9.321848703973823e-05,
"loss": 0.684,
"step": 10200
},
{
"epoch": 0.26586003072623277,
"grad_norm": 0.17245131731033325,
"learning_rate": 9.305192589270428e-05,
"loss": 0.6782,
"step": 10210
},
{
"epoch": 0.2661204225290988,
"grad_norm": 0.17289961874485016,
"learning_rate": 9.28853841127133e-05,
"loss": 0.7016,
"step": 10220
},
{
"epoch": 0.2663808143319648,
"grad_norm": 0.1869751363992691,
"learning_rate": 9.271886216398346e-05,
"loss": 0.6754,
"step": 10230
},
{
"epoch": 0.2666412061348309,
"grad_norm": 0.1776220202445984,
"learning_rate": 9.255236051067756e-05,
"loss": 0.6837,
"step": 10240
},
{
"epoch": 0.2669015979376969,
"grad_norm": 0.17744296789169312,
"learning_rate": 9.238587961690184e-05,
"loss": 0.6877,
"step": 10250
},
{
"epoch": 0.267161989740563,
"grad_norm": 0.17213894426822662,
"learning_rate": 9.221941994670465e-05,
"loss": 0.6653,
"step": 10260
},
{
"epoch": 0.267422381543429,
"grad_norm": 0.17459706962108612,
"learning_rate": 9.205298196407533e-05,
"loss": 0.6764,
"step": 10270
},
{
"epoch": 0.26768277334629503,
"grad_norm": 0.17767804861068726,
"learning_rate": 9.188656613294258e-05,
"loss": 0.663,
"step": 10280
},
{
"epoch": 0.2679431651491611,
"grad_norm": 0.1797824501991272,
"learning_rate": 9.17201729171735e-05,
"loss": 0.6841,
"step": 10290
},
{
"epoch": 0.26820355695202713,
"grad_norm": 0.17501597106456757,
"learning_rate": 9.1553802780572e-05,
"loss": 0.6737,
"step": 10300
},
{
"epoch": 0.2684639487548932,
"grad_norm": 0.1872059851884842,
"learning_rate": 9.138745618687785e-05,
"loss": 0.702,
"step": 10310
},
{
"epoch": 0.26872434055775923,
"grad_norm": 0.18220624327659607,
"learning_rate": 9.122113359976507e-05,
"loss": 0.7076,
"step": 10320
},
{
"epoch": 0.2689847323606253,
"grad_norm": 0.18646632134914398,
"learning_rate": 9.105483548284075e-05,
"loss": 0.6995,
"step": 10330
},
{
"epoch": 0.26924512416349133,
"grad_norm": 0.1900705099105835,
"learning_rate": 9.088856229964379e-05,
"loss": 0.6924,
"step": 10340
},
{
"epoch": 0.26950551596635736,
"grad_norm": 0.1800726354122162,
"learning_rate": 9.072231451364369e-05,
"loss": 0.6957,
"step": 10350
},
{
"epoch": 0.26976590776922343,
"grad_norm": 0.18188440799713135,
"learning_rate": 9.055609258823907e-05,
"loss": 0.6675,
"step": 10360
},
{
"epoch": 0.27002629957208946,
"grad_norm": 0.16966912150382996,
"learning_rate": 9.038989698675644e-05,
"loss": 0.6815,
"step": 10370
},
{
"epoch": 0.27028669137495553,
"grad_norm": 0.18895496428012848,
"learning_rate": 9.022372817244894e-05,
"loss": 0.6897,
"step": 10380
},
{
"epoch": 0.27054708317782156,
"grad_norm": 0.1844681203365326,
"learning_rate": 9.005758660849516e-05,
"loss": 0.6829,
"step": 10390
},
{
"epoch": 0.27080747498068763,
"grad_norm": 0.18955713510513306,
"learning_rate": 8.989147275799761e-05,
"loss": 0.7074,
"step": 10400
},
{
"epoch": 0.27106786678355366,
"grad_norm": 0.1776488870382309,
"learning_rate": 8.972538708398154e-05,
"loss": 0.6795,
"step": 10410
},
{
"epoch": 0.2713282585864197,
"grad_norm": 0.17599767446517944,
"learning_rate": 8.955933004939378e-05,
"loss": 0.7013,
"step": 10420
},
{
"epoch": 0.27158865038928576,
"grad_norm": 0.18769358098506927,
"learning_rate": 8.939330211710118e-05,
"loss": 0.6954,
"step": 10430
},
{
"epoch": 0.2718490421921518,
"grad_norm": 0.16961945593357086,
"learning_rate": 8.922730374988965e-05,
"loss": 0.6828,
"step": 10440
},
{
"epoch": 0.27210943399501786,
"grad_norm": 0.17419418692588806,
"learning_rate": 8.906133541046253e-05,
"loss": 0.6765,
"step": 10450
},
{
"epoch": 0.2723698257978839,
"grad_norm": 0.18160662055015564,
"learning_rate": 8.889539756143953e-05,
"loss": 0.6999,
"step": 10460
},
{
"epoch": 0.2726302176007499,
"grad_norm": 0.17763784527778625,
"learning_rate": 8.872949066535532e-05,
"loss": 0.6933,
"step": 10470
},
{
"epoch": 0.272890609403616,
"grad_norm": 0.18089672923088074,
"learning_rate": 8.856361518465841e-05,
"loss": 0.6903,
"step": 10480
},
{
"epoch": 0.273151001206482,
"grad_norm": 0.18201428651809692,
"learning_rate": 8.839777158170962e-05,
"loss": 0.6746,
"step": 10490
},
{
"epoch": 0.2734113930093481,
"grad_norm": 0.17589519917964935,
"learning_rate": 8.823196031878099e-05,
"loss": 0.6898,
"step": 10500
}
],
"logging_steps": 10,
"max_steps": 19202,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 4.92625704517632e+18,
"train_batch_size": 5,
"trial_name": null,
"trial_params": null
}