starcoder-1b-finetuned-ds-4 / trainer_state.json
rohitc33's picture
Upload folder using huggingface_hub
c614eac verified
raw
history blame contribute delete
No virus
249 kB
{
"best_metric": 0.5966796875,
"best_model_checkpoint": "./results/checkpoint-10662",
"epoch": 4.0,
"eval_steps": 500,
"global_step": 14216,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0028137310073157004,
"grad_norm": 0.0,
"learning_rate": 0.0,
"loss": 2.8827,
"step": 10
},
{
"epoch": 0.005627462014631401,
"grad_norm": 0.0,
"learning_rate": 0.0,
"loss": 2.9839,
"step": 20
},
{
"epoch": 0.008441193021947102,
"grad_norm": 224.25678800007668,
"learning_rate": 4.2e-06,
"loss": 2.9641,
"step": 30
},
{
"epoch": 0.011254924029262802,
"grad_norm": 139.77573172893386,
"learning_rate": 1.02e-05,
"loss": 2.3895,
"step": 40
},
{
"epoch": 0.014068655036578503,
"grad_norm": 110.31193690169778,
"learning_rate": 1.6199999999999997e-05,
"loss": 1.55,
"step": 50
},
{
"epoch": 0.016882386043894203,
"grad_norm": 23.291664316032804,
"learning_rate": 2.2199999999999998e-05,
"loss": 2.1463,
"step": 60
},
{
"epoch": 0.019696117051209903,
"grad_norm": 78.12962991651972,
"learning_rate": 2.8199999999999998e-05,
"loss": 2.6968,
"step": 70
},
{
"epoch": 0.022509848058525603,
"grad_norm": 145.88009990286014,
"learning_rate": 3.42e-05,
"loss": 3.5582,
"step": 80
},
{
"epoch": 0.025323579065841307,
"grad_norm": 126.24036573153992,
"learning_rate": 4.02e-05,
"loss": 2.7401,
"step": 90
},
{
"epoch": 0.028137310073157007,
"grad_norm": 7.932821466299223,
"learning_rate": 4.62e-05,
"loss": 1.4496,
"step": 100
},
{
"epoch": 0.030951041080472707,
"grad_norm": 51.39094381445175,
"learning_rate": 5.2199999999999995e-05,
"loss": 1.2416,
"step": 110
},
{
"epoch": 0.03376477208778841,
"grad_norm": 52.627137222293655,
"learning_rate": 5.82e-05,
"loss": 0.9829,
"step": 120
},
{
"epoch": 0.03657850309510411,
"grad_norm": 83.67118972227,
"learning_rate": 6.419999999999999e-05,
"loss": 1.8664,
"step": 130
},
{
"epoch": 0.03939223410241981,
"grad_norm": 254.8820062247784,
"learning_rate": 7.02e-05,
"loss": 1.3919,
"step": 140
},
{
"epoch": 0.04220596510973551,
"grad_norm": 95.05590819779509,
"learning_rate": 7.62e-05,
"loss": 1.8913,
"step": 150
},
{
"epoch": 0.04501969611705121,
"grad_norm": 96.52932022551686,
"learning_rate": 8.22e-05,
"loss": 2.9941,
"step": 160
},
{
"epoch": 0.04783342712436691,
"grad_norm": 152.73209597084838,
"learning_rate": 8.819999999999999e-05,
"loss": 1.9007,
"step": 170
},
{
"epoch": 0.050647158131682614,
"grad_norm": 38.961577452090374,
"learning_rate": 9.419999999999999e-05,
"loss": 1.3572,
"step": 180
},
{
"epoch": 0.05346088913899831,
"grad_norm": 8.365359526839624,
"learning_rate": 0.0001002,
"loss": 0.8396,
"step": 190
},
{
"epoch": 0.056274620146314014,
"grad_norm": 27.415167430506084,
"learning_rate": 0.00010619999999999998,
"loss": 1.6339,
"step": 200
},
{
"epoch": 0.05908835115362971,
"grad_norm": 4.809093819846915,
"learning_rate": 0.00011219999999999999,
"loss": 0.7721,
"step": 210
},
{
"epoch": 0.061902082160945414,
"grad_norm": 25.63960863733989,
"learning_rate": 0.0001182,
"loss": 0.8201,
"step": 220
},
{
"epoch": 0.06471581316826111,
"grad_norm": 81.6209161855533,
"learning_rate": 0.00012419999999999998,
"loss": 1.3382,
"step": 230
},
{
"epoch": 0.06752954417557681,
"grad_norm": 25.965053380742912,
"learning_rate": 0.0001302,
"loss": 1.2083,
"step": 240
},
{
"epoch": 0.07034327518289252,
"grad_norm": 49.9863607853443,
"learning_rate": 0.0001362,
"loss": 1.3294,
"step": 250
},
{
"epoch": 0.07315700619020822,
"grad_norm": 20.007567654071906,
"learning_rate": 0.0001422,
"loss": 0.9669,
"step": 260
},
{
"epoch": 0.07597073719752391,
"grad_norm": 29.047164184009052,
"learning_rate": 0.0001482,
"loss": 0.8447,
"step": 270
},
{
"epoch": 0.07878446820483961,
"grad_norm": 23.644879858956426,
"learning_rate": 0.00015419999999999998,
"loss": 1.141,
"step": 280
},
{
"epoch": 0.08159819921215532,
"grad_norm": 46.372111281936895,
"learning_rate": 0.0001602,
"loss": 1.2533,
"step": 290
},
{
"epoch": 0.08441193021947102,
"grad_norm": 11.20178962457438,
"learning_rate": 0.0001662,
"loss": 0.825,
"step": 300
},
{
"epoch": 0.08722566122678672,
"grad_norm": 48.90453331100731,
"learning_rate": 0.00017219999999999998,
"loss": 1.1152,
"step": 310
},
{
"epoch": 0.09003939223410241,
"grad_norm": 0.091122566845799,
"learning_rate": 0.00017819999999999997,
"loss": 0.877,
"step": 320
},
{
"epoch": 0.09285312324141812,
"grad_norm": 44.370694350506966,
"learning_rate": 0.00018419999999999998,
"loss": 2.8161,
"step": 330
},
{
"epoch": 0.09566685424873382,
"grad_norm": 26.012340742157125,
"learning_rate": 0.0001902,
"loss": 1.8136,
"step": 340
},
{
"epoch": 0.09848058525604952,
"grad_norm": 4.391781057112832,
"learning_rate": 0.0001962,
"loss": 0.8146,
"step": 350
},
{
"epoch": 0.10129431626336523,
"grad_norm": 9.187263907428804,
"learning_rate": 0.0002022,
"loss": 1.9734,
"step": 360
},
{
"epoch": 0.10410804727068092,
"grad_norm": 146.14669734330562,
"learning_rate": 0.00020819999999999996,
"loss": 2.3596,
"step": 370
},
{
"epoch": 0.10692177827799662,
"grad_norm": 103.03855355929782,
"learning_rate": 0.00021419999999999998,
"loss": 4.1036,
"step": 380
},
{
"epoch": 0.10973550928531232,
"grad_norm": 64.30913047008124,
"learning_rate": 0.00022019999999999999,
"loss": 1.337,
"step": 390
},
{
"epoch": 0.11254924029262803,
"grad_norm": 309.69819080980943,
"learning_rate": 0.00022559999999999998,
"loss": 1.7713,
"step": 400
},
{
"epoch": 0.11536297129994373,
"grad_norm": 1.9356075644481516,
"learning_rate": 0.0002316,
"loss": 1.7203,
"step": 410
},
{
"epoch": 0.11817670230725942,
"grad_norm": 79.09050048639865,
"learning_rate": 0.0002376,
"loss": 3.0218,
"step": 420
},
{
"epoch": 0.12099043331457512,
"grad_norm": 24.669088958893436,
"learning_rate": 0.00024359999999999999,
"loss": 2.8796,
"step": 430
},
{
"epoch": 0.12380416432189083,
"grad_norm": 35.50057015666331,
"learning_rate": 0.00024959999999999994,
"loss": 1.3042,
"step": 440
},
{
"epoch": 0.12661789532920653,
"grad_norm": 63.35643345487432,
"learning_rate": 0.0002556,
"loss": 1.0222,
"step": 450
},
{
"epoch": 0.12943162633652222,
"grad_norm": 44.8309413245288,
"learning_rate": 0.00026159999999999996,
"loss": 2.2135,
"step": 460
},
{
"epoch": 0.13224535734383794,
"grad_norm": 38.21235063708972,
"learning_rate": 0.0002676,
"loss": 1.9759,
"step": 470
},
{
"epoch": 0.13505908835115363,
"grad_norm": 42.502230547826144,
"learning_rate": 0.0002736,
"loss": 1.3122,
"step": 480
},
{
"epoch": 0.13787281935846932,
"grad_norm": 36.78039561983335,
"learning_rate": 0.00027959999999999997,
"loss": 0.9201,
"step": 490
},
{
"epoch": 0.14068655036578503,
"grad_norm": 31.432740474500267,
"learning_rate": 0.00028559999999999995,
"loss": 0.7877,
"step": 500
},
{
"epoch": 0.14350028137310072,
"grad_norm": 89.40921662924484,
"learning_rate": 0.0002916,
"loss": 1.5383,
"step": 510
},
{
"epoch": 0.14631401238041644,
"grad_norm": 52.21924604041036,
"learning_rate": 0.00029759999999999997,
"loss": 2.0496,
"step": 520
},
{
"epoch": 0.14912774338773213,
"grad_norm": 26.377094972604038,
"learning_rate": 0.00029986876640419944,
"loss": 1.0984,
"step": 530
},
{
"epoch": 0.15194147439504782,
"grad_norm": 12.678479550625854,
"learning_rate": 0.0002996500437445319,
"loss": 0.7538,
"step": 540
},
{
"epoch": 0.15475520540236354,
"grad_norm": 39.15053950048559,
"learning_rate": 0.0002994313210848644,
"loss": 0.8266,
"step": 550
},
{
"epoch": 0.15756893640967923,
"grad_norm": 8.686627630645642,
"learning_rate": 0.00029921259842519685,
"loss": 0.6999,
"step": 560
},
{
"epoch": 0.16038266741699495,
"grad_norm": 15.686342705824982,
"learning_rate": 0.00029899387576552927,
"loss": 0.7207,
"step": 570
},
{
"epoch": 0.16319639842431063,
"grad_norm": 5.239119778234183,
"learning_rate": 0.00029877515310586174,
"loss": 0.979,
"step": 580
},
{
"epoch": 0.16601012943162632,
"grad_norm": 3.898189968863333,
"learning_rate": 0.0002985564304461942,
"loss": 0.7798,
"step": 590
},
{
"epoch": 0.16882386043894204,
"grad_norm": 82.36259815094716,
"learning_rate": 0.0002983377077865267,
"loss": 1.3224,
"step": 600
},
{
"epoch": 0.17163759144625773,
"grad_norm": 78.13332901615998,
"learning_rate": 0.0002981189851268591,
"loss": 1.5312,
"step": 610
},
{
"epoch": 0.17445132245357345,
"grad_norm": 19.326937621272116,
"learning_rate": 0.00029790026246719157,
"loss": 2.1933,
"step": 620
},
{
"epoch": 0.17726505346088914,
"grad_norm": 63.14152039967042,
"learning_rate": 0.00029768153980752404,
"loss": 1.356,
"step": 630
},
{
"epoch": 0.18007878446820483,
"grad_norm": 80.26072923404266,
"learning_rate": 0.0002974628171478565,
"loss": 2.3056,
"step": 640
},
{
"epoch": 0.18289251547552055,
"grad_norm": 10.103427618986666,
"learning_rate": 0.0002972440944881889,
"loss": 1.2462,
"step": 650
},
{
"epoch": 0.18570624648283623,
"grad_norm": 51.23631739417244,
"learning_rate": 0.0002970253718285214,
"loss": 1.1639,
"step": 660
},
{
"epoch": 0.18851997749015195,
"grad_norm": 6.157960830396006,
"learning_rate": 0.00029680664916885386,
"loss": 0.7786,
"step": 670
},
{
"epoch": 0.19133370849746764,
"grad_norm": 10.247301558325885,
"learning_rate": 0.00029658792650918633,
"loss": 0.7196,
"step": 680
},
{
"epoch": 0.19414743950478333,
"grad_norm": 43.71975555313302,
"learning_rate": 0.0002963692038495188,
"loss": 1.1545,
"step": 690
},
{
"epoch": 0.19696117051209905,
"grad_norm": 7.9395865721238446,
"learning_rate": 0.0002961504811898512,
"loss": 0.8198,
"step": 700
},
{
"epoch": 0.19977490151941474,
"grad_norm": 10.481113571999062,
"learning_rate": 0.0002959317585301837,
"loss": 0.7158,
"step": 710
},
{
"epoch": 0.20258863252673046,
"grad_norm": 12.378546853564645,
"learning_rate": 0.00029571303587051616,
"loss": 0.7685,
"step": 720
},
{
"epoch": 0.20540236353404615,
"grad_norm": 47.604153129845606,
"learning_rate": 0.00029549431321084863,
"loss": 0.7332,
"step": 730
},
{
"epoch": 0.20821609454136183,
"grad_norm": 25.527299782292648,
"learning_rate": 0.0002952755905511811,
"loss": 0.7475,
"step": 740
},
{
"epoch": 0.21102982554867755,
"grad_norm": 45.25744483228148,
"learning_rate": 0.00029505686789151357,
"loss": 0.743,
"step": 750
},
{
"epoch": 0.21384355655599324,
"grad_norm": 37.835568741114564,
"learning_rate": 0.000294838145231846,
"loss": 0.7685,
"step": 760
},
{
"epoch": 0.21665728756330896,
"grad_norm": 27.87886782853722,
"learning_rate": 0.00029461942257217845,
"loss": 0.9676,
"step": 770
},
{
"epoch": 0.21947101857062465,
"grad_norm": 23.79947674931416,
"learning_rate": 0.0002944006999125109,
"loss": 0.7742,
"step": 780
},
{
"epoch": 0.22228474957794034,
"grad_norm": 13.066512959590527,
"learning_rate": 0.0002941819772528434,
"loss": 0.7118,
"step": 790
},
{
"epoch": 0.22509848058525606,
"grad_norm": 25.847486935286263,
"learning_rate": 0.0002939632545931758,
"loss": 0.7724,
"step": 800
},
{
"epoch": 0.22791221159257175,
"grad_norm": 35.49445839081954,
"learning_rate": 0.0002937445319335083,
"loss": 0.7618,
"step": 810
},
{
"epoch": 0.23072594259988746,
"grad_norm": 49.49529219586411,
"learning_rate": 0.00029352580927384075,
"loss": 0.7084,
"step": 820
},
{
"epoch": 0.23353967360720315,
"grad_norm": 37.565688190537166,
"learning_rate": 0.00029330708661417317,
"loss": 0.8741,
"step": 830
},
{
"epoch": 0.23635340461451884,
"grad_norm": 8.413460164151891,
"learning_rate": 0.00029308836395450564,
"loss": 1.0024,
"step": 840
},
{
"epoch": 0.23916713562183456,
"grad_norm": 32.940098975528464,
"learning_rate": 0.0002928696412948381,
"loss": 0.9008,
"step": 850
},
{
"epoch": 0.24198086662915025,
"grad_norm": 41.72472308400113,
"learning_rate": 0.0002926509186351706,
"loss": 0.6943,
"step": 860
},
{
"epoch": 0.24479459763646597,
"grad_norm": 5.973403122009343,
"learning_rate": 0.00029243219597550305,
"loss": 0.8841,
"step": 870
},
{
"epoch": 0.24760832864378166,
"grad_norm": 19.107388950348,
"learning_rate": 0.0002922134733158355,
"loss": 0.7404,
"step": 880
},
{
"epoch": 0.2504220596510974,
"grad_norm": 66.51613947405258,
"learning_rate": 0.00029199475065616793,
"loss": 1.5743,
"step": 890
},
{
"epoch": 0.25323579065841306,
"grad_norm": 35.550471084963576,
"learning_rate": 0.0002917760279965004,
"loss": 0.9394,
"step": 900
},
{
"epoch": 0.25604952166572875,
"grad_norm": 36.649681372911445,
"learning_rate": 0.00029155730533683287,
"loss": 0.7641,
"step": 910
},
{
"epoch": 0.25886325267304444,
"grad_norm": 12.270656532801103,
"learning_rate": 0.00029133858267716534,
"loss": 0.9354,
"step": 920
},
{
"epoch": 0.26167698368036013,
"grad_norm": 18.4202269481449,
"learning_rate": 0.0002911198600174978,
"loss": 0.8232,
"step": 930
},
{
"epoch": 0.2644907146876759,
"grad_norm": 70.47960911214764,
"learning_rate": 0.0002909011373578303,
"loss": 0.8628,
"step": 940
},
{
"epoch": 0.26730444569499157,
"grad_norm": 25.90531145228859,
"learning_rate": 0.0002906824146981627,
"loss": 0.8182,
"step": 950
},
{
"epoch": 0.27011817670230726,
"grad_norm": 29.63336083779562,
"learning_rate": 0.00029046369203849517,
"loss": 0.7766,
"step": 960
},
{
"epoch": 0.27293190770962295,
"grad_norm": 16.666638960939466,
"learning_rate": 0.00029024496937882764,
"loss": 0.7988,
"step": 970
},
{
"epoch": 0.27574563871693863,
"grad_norm": 20.07806754771967,
"learning_rate": 0.0002900262467191601,
"loss": 0.7784,
"step": 980
},
{
"epoch": 0.2785593697242544,
"grad_norm": 12.36951117312153,
"learning_rate": 0.0002898075240594925,
"loss": 0.8476,
"step": 990
},
{
"epoch": 0.28137310073157007,
"grad_norm": 22.1125299804219,
"learning_rate": 0.000289588801399825,
"loss": 0.719,
"step": 1000
},
{
"epoch": 0.28418683173888576,
"grad_norm": 3.109040322200588,
"learning_rate": 0.00028937007874015746,
"loss": 0.7406,
"step": 1010
},
{
"epoch": 0.28700056274620145,
"grad_norm": 28.351720853512052,
"learning_rate": 0.0002891513560804899,
"loss": 0.8295,
"step": 1020
},
{
"epoch": 0.28981429375351714,
"grad_norm": 8.61987445057803,
"learning_rate": 0.00028893263342082235,
"loss": 0.681,
"step": 1030
},
{
"epoch": 0.2926280247608329,
"grad_norm": 11.532817025226382,
"learning_rate": 0.0002887139107611548,
"loss": 0.7396,
"step": 1040
},
{
"epoch": 0.2954417557681486,
"grad_norm": 6.165669988575859,
"learning_rate": 0.0002884951881014873,
"loss": 0.7601,
"step": 1050
},
{
"epoch": 0.29825548677546426,
"grad_norm": 5.860155976144423,
"learning_rate": 0.00028827646544181976,
"loss": 0.7554,
"step": 1060
},
{
"epoch": 0.30106921778277995,
"grad_norm": 19.345458106174757,
"learning_rate": 0.0002880577427821522,
"loss": 0.7848,
"step": 1070
},
{
"epoch": 0.30388294879009564,
"grad_norm": 19.436003760130507,
"learning_rate": 0.00028783902012248464,
"loss": 0.6703,
"step": 1080
},
{
"epoch": 0.3066966797974114,
"grad_norm": 3.958053947843868,
"learning_rate": 0.0002876202974628171,
"loss": 0.7141,
"step": 1090
},
{
"epoch": 0.3095104108047271,
"grad_norm": 9.441658402935863,
"learning_rate": 0.0002874015748031496,
"loss": 0.6783,
"step": 1100
},
{
"epoch": 0.31232414181204277,
"grad_norm": 18.815776769084255,
"learning_rate": 0.00028718285214348205,
"loss": 0.8264,
"step": 1110
},
{
"epoch": 0.31513787281935846,
"grad_norm": 18.750079475373475,
"learning_rate": 0.0002869641294838145,
"loss": 0.702,
"step": 1120
},
{
"epoch": 0.31795160382667415,
"grad_norm": 40.59299255876535,
"learning_rate": 0.000286745406824147,
"loss": 0.7972,
"step": 1130
},
{
"epoch": 0.3207653348339899,
"grad_norm": 25.46547229422401,
"learning_rate": 0.0002865266841644794,
"loss": 0.7,
"step": 1140
},
{
"epoch": 0.3235790658413056,
"grad_norm": 13.988940584670248,
"learning_rate": 0.0002863079615048119,
"loss": 0.6672,
"step": 1150
},
{
"epoch": 0.32639279684862127,
"grad_norm": 4.209523057541857,
"learning_rate": 0.00028608923884514435,
"loss": 0.6089,
"step": 1160
},
{
"epoch": 0.32920652785593696,
"grad_norm": 21.621479938817654,
"learning_rate": 0.0002858705161854768,
"loss": 0.6678,
"step": 1170
},
{
"epoch": 0.33202025886325265,
"grad_norm": 27.115608784965413,
"learning_rate": 0.00028565179352580924,
"loss": 0.656,
"step": 1180
},
{
"epoch": 0.3348339898705684,
"grad_norm": 9.058371119623647,
"learning_rate": 0.0002854330708661417,
"loss": 0.8967,
"step": 1190
},
{
"epoch": 0.3376477208778841,
"grad_norm": 23.7560047514354,
"learning_rate": 0.0002852143482064742,
"loss": 0.8101,
"step": 1200
},
{
"epoch": 0.3404614518851998,
"grad_norm": 35.20987445808512,
"learning_rate": 0.0002849956255468066,
"loss": 0.7738,
"step": 1210
},
{
"epoch": 0.34327518289251546,
"grad_norm": 11.698215716101412,
"learning_rate": 0.00028477690288713906,
"loss": 0.8432,
"step": 1220
},
{
"epoch": 0.34608891389983115,
"grad_norm": 22.50137501429176,
"learning_rate": 0.00028455818022747153,
"loss": 0.7033,
"step": 1230
},
{
"epoch": 0.3489026449071469,
"grad_norm": 19.821463032004264,
"learning_rate": 0.000284339457567804,
"loss": 0.7943,
"step": 1240
},
{
"epoch": 0.3517163759144626,
"grad_norm": 33.67996416219415,
"learning_rate": 0.00028412073490813647,
"loss": 0.9042,
"step": 1250
},
{
"epoch": 0.3545301069217783,
"grad_norm": 10.3108845588795,
"learning_rate": 0.0002839020122484689,
"loss": 0.6961,
"step": 1260
},
{
"epoch": 0.35734383792909397,
"grad_norm": 13.932358160677255,
"learning_rate": 0.00028368328958880136,
"loss": 0.6584,
"step": 1270
},
{
"epoch": 0.36015756893640966,
"grad_norm": 6.665964382062972,
"learning_rate": 0.00028346456692913383,
"loss": 0.7388,
"step": 1280
},
{
"epoch": 0.3629712999437254,
"grad_norm": 10.05896951436482,
"learning_rate": 0.0002832458442694663,
"loss": 0.7343,
"step": 1290
},
{
"epoch": 0.3657850309510411,
"grad_norm": 11.0866654634556,
"learning_rate": 0.00028302712160979877,
"loss": 0.6406,
"step": 1300
},
{
"epoch": 0.3685987619583568,
"grad_norm": 16.53331909071421,
"learning_rate": 0.00028280839895013124,
"loss": 0.7217,
"step": 1310
},
{
"epoch": 0.37141249296567247,
"grad_norm": 29.64720675209545,
"learning_rate": 0.00028258967629046365,
"loss": 0.6978,
"step": 1320
},
{
"epoch": 0.37422622397298816,
"grad_norm": 28.455785874506656,
"learning_rate": 0.0002823709536307961,
"loss": 0.7681,
"step": 1330
},
{
"epoch": 0.3770399549803039,
"grad_norm": 22.981131743639594,
"learning_rate": 0.0002821522309711286,
"loss": 0.6291,
"step": 1340
},
{
"epoch": 0.3798536859876196,
"grad_norm": 29.59061751823822,
"learning_rate": 0.00028193350831146106,
"loss": 0.8084,
"step": 1350
},
{
"epoch": 0.3826674169949353,
"grad_norm": 5.6091230275760475,
"learning_rate": 0.0002817147856517935,
"loss": 0.7371,
"step": 1360
},
{
"epoch": 0.385481148002251,
"grad_norm": 5.883998486444261,
"learning_rate": 0.00028149606299212595,
"loss": 0.7251,
"step": 1370
},
{
"epoch": 0.38829487900956666,
"grad_norm": 9.040960893323161,
"learning_rate": 0.0002812773403324584,
"loss": 0.6948,
"step": 1380
},
{
"epoch": 0.3911086100168824,
"grad_norm": 16.807114962722785,
"learning_rate": 0.0002810586176727909,
"loss": 0.7303,
"step": 1390
},
{
"epoch": 0.3939223410241981,
"grad_norm": 13.0879100690685,
"learning_rate": 0.0002808398950131233,
"loss": 0.6484,
"step": 1400
},
{
"epoch": 0.3967360720315138,
"grad_norm": 16.044210764032027,
"learning_rate": 0.0002806211723534558,
"loss": 0.6612,
"step": 1410
},
{
"epoch": 0.3995498030388295,
"grad_norm": 37.0843541394152,
"learning_rate": 0.00028040244969378825,
"loss": 0.7218,
"step": 1420
},
{
"epoch": 0.40236353404614517,
"grad_norm": 44.12879697805232,
"learning_rate": 0.0002801837270341207,
"loss": 1.2958,
"step": 1430
},
{
"epoch": 0.4051772650534609,
"grad_norm": 28.017644530703276,
"learning_rate": 0.0002799650043744532,
"loss": 0.989,
"step": 1440
},
{
"epoch": 0.4079909960607766,
"grad_norm": 48.95451227633847,
"learning_rate": 0.0002797462817147856,
"loss": 0.7852,
"step": 1450
},
{
"epoch": 0.4108047270680923,
"grad_norm": 13.750288764403155,
"learning_rate": 0.00027952755905511807,
"loss": 1.0036,
"step": 1460
},
{
"epoch": 0.413618458075408,
"grad_norm": 12.62751471781883,
"learning_rate": 0.00027930883639545054,
"loss": 0.7137,
"step": 1470
},
{
"epoch": 0.41643218908272367,
"grad_norm": 27.20810519301277,
"learning_rate": 0.000279090113735783,
"loss": 0.6905,
"step": 1480
},
{
"epoch": 0.4192459200900394,
"grad_norm": 28.107277824414965,
"learning_rate": 0.0002788713910761155,
"loss": 0.6837,
"step": 1490
},
{
"epoch": 0.4220596510973551,
"grad_norm": 5.416144983374891,
"learning_rate": 0.00027865266841644795,
"loss": 0.7208,
"step": 1500
},
{
"epoch": 0.4248733821046708,
"grad_norm": 11.589744326535003,
"learning_rate": 0.00027843394575678037,
"loss": 0.6227,
"step": 1510
},
{
"epoch": 0.4276871131119865,
"grad_norm": 8.906957133772503,
"learning_rate": 0.00027821522309711284,
"loss": 0.7367,
"step": 1520
},
{
"epoch": 0.4305008441193022,
"grad_norm": 32.87231457665042,
"learning_rate": 0.0002779965004374453,
"loss": 0.7919,
"step": 1530
},
{
"epoch": 0.4333145751266179,
"grad_norm": 26.00340134223748,
"learning_rate": 0.0002777777777777778,
"loss": 0.676,
"step": 1540
},
{
"epoch": 0.4361283061339336,
"grad_norm": 34.53119270440781,
"learning_rate": 0.0002775590551181102,
"loss": 0.8778,
"step": 1550
},
{
"epoch": 0.4389420371412493,
"grad_norm": 6.460463403838006,
"learning_rate": 0.00027734033245844266,
"loss": 0.5865,
"step": 1560
},
{
"epoch": 0.441755768148565,
"grad_norm": 11.404929871459034,
"learning_rate": 0.00027712160979877513,
"loss": 0.6955,
"step": 1570
},
{
"epoch": 0.4445694991558807,
"grad_norm": 16.213324604997545,
"learning_rate": 0.0002769028871391076,
"loss": 0.6287,
"step": 1580
},
{
"epoch": 0.4473832301631964,
"grad_norm": 23.82912364680576,
"learning_rate": 0.00027668416447944,
"loss": 1.0301,
"step": 1590
},
{
"epoch": 0.4501969611705121,
"grad_norm": 13.424863094947291,
"learning_rate": 0.0002764654418197725,
"loss": 0.7109,
"step": 1600
},
{
"epoch": 0.4530106921778278,
"grad_norm": 20.253487976081246,
"learning_rate": 0.00027624671916010496,
"loss": 0.5895,
"step": 1610
},
{
"epoch": 0.4558244231851435,
"grad_norm": 12.145349601064195,
"learning_rate": 0.00027602799650043743,
"loss": 0.6832,
"step": 1620
},
{
"epoch": 0.4586381541924592,
"grad_norm": 15.611833511231971,
"learning_rate": 0.0002758092738407699,
"loss": 0.6678,
"step": 1630
},
{
"epoch": 0.4614518851997749,
"grad_norm": 14.283717293563125,
"learning_rate": 0.0002755905511811023,
"loss": 0.596,
"step": 1640
},
{
"epoch": 0.4642656162070906,
"grad_norm": 20.447345294591308,
"learning_rate": 0.0002753718285214348,
"loss": 0.6593,
"step": 1650
},
{
"epoch": 0.4670793472144063,
"grad_norm": 7.225334907859718,
"learning_rate": 0.00027515310586176726,
"loss": 0.7062,
"step": 1660
},
{
"epoch": 0.469893078221722,
"grad_norm": 16.228475676453073,
"learning_rate": 0.0002749343832020997,
"loss": 0.68,
"step": 1670
},
{
"epoch": 0.4727068092290377,
"grad_norm": 15.345865551250505,
"learning_rate": 0.0002747156605424322,
"loss": 0.6377,
"step": 1680
},
{
"epoch": 0.47552054023635343,
"grad_norm": 13.116990150980092,
"learning_rate": 0.00027449693788276467,
"loss": 0.555,
"step": 1690
},
{
"epoch": 0.4783342712436691,
"grad_norm": 7.523456579032664,
"learning_rate": 0.0002742782152230971,
"loss": 0.7089,
"step": 1700
},
{
"epoch": 0.4811480022509848,
"grad_norm": 15.62034181204981,
"learning_rate": 0.00027405949256342955,
"loss": 0.5955,
"step": 1710
},
{
"epoch": 0.4839617332583005,
"grad_norm": 33.72794816539747,
"learning_rate": 0.000273840769903762,
"loss": 0.7967,
"step": 1720
},
{
"epoch": 0.4867754642656162,
"grad_norm": 8.371501278758954,
"learning_rate": 0.0002736220472440945,
"loss": 0.6281,
"step": 1730
},
{
"epoch": 0.48958919527293193,
"grad_norm": 14.674093397655396,
"learning_rate": 0.0002734033245844269,
"loss": 0.7373,
"step": 1740
},
{
"epoch": 0.4924029262802476,
"grad_norm": 4.743062155600575,
"learning_rate": 0.0002731846019247594,
"loss": 0.6782,
"step": 1750
},
{
"epoch": 0.4952166572875633,
"grad_norm": 45.76318589779893,
"learning_rate": 0.00027296587926509185,
"loss": 0.6851,
"step": 1760
},
{
"epoch": 0.498030388294879,
"grad_norm": 7.008897310409392,
"learning_rate": 0.00027274715660542426,
"loss": 0.7031,
"step": 1770
},
{
"epoch": 0.5008441193021947,
"grad_norm": 12.188963127648035,
"learning_rate": 0.00027252843394575673,
"loss": 0.7553,
"step": 1780
},
{
"epoch": 0.5036578503095104,
"grad_norm": 8.04659950296303,
"learning_rate": 0.0002723097112860892,
"loss": 0.6391,
"step": 1790
},
{
"epoch": 0.5064715813168261,
"grad_norm": 23.403833885375786,
"learning_rate": 0.00027209098862642167,
"loss": 0.6564,
"step": 1800
},
{
"epoch": 0.5092853123241418,
"grad_norm": 19.61248291110157,
"learning_rate": 0.00027187226596675414,
"loss": 0.5736,
"step": 1810
},
{
"epoch": 0.5120990433314575,
"grad_norm": 7.232723854059021,
"learning_rate": 0.00027165354330708656,
"loss": 0.7036,
"step": 1820
},
{
"epoch": 0.5149127743387732,
"grad_norm": 13.467653622805527,
"learning_rate": 0.00027143482064741903,
"loss": 0.8533,
"step": 1830
},
{
"epoch": 0.5177265053460889,
"grad_norm": 24.167342861487832,
"learning_rate": 0.0002712160979877515,
"loss": 0.7531,
"step": 1840
},
{
"epoch": 0.5205402363534046,
"grad_norm": 17.840804581591108,
"learning_rate": 0.00027099737532808397,
"loss": 0.6141,
"step": 1850
},
{
"epoch": 0.5233539673607203,
"grad_norm": 6.905072707920589,
"learning_rate": 0.00027077865266841644,
"loss": 0.6358,
"step": 1860
},
{
"epoch": 0.526167698368036,
"grad_norm": 14.15349419929909,
"learning_rate": 0.0002705599300087489,
"loss": 0.7491,
"step": 1870
},
{
"epoch": 0.5289814293753518,
"grad_norm": 10.6411502042627,
"learning_rate": 0.0002703412073490814,
"loss": 0.6761,
"step": 1880
},
{
"epoch": 0.5317951603826674,
"grad_norm": 12.526381470822352,
"learning_rate": 0.0002701224846894138,
"loss": 0.6767,
"step": 1890
},
{
"epoch": 0.5346088913899831,
"grad_norm": 18.982165857200393,
"learning_rate": 0.00026990376202974626,
"loss": 0.6584,
"step": 1900
},
{
"epoch": 0.5374226223972988,
"grad_norm": 21.858403045881502,
"learning_rate": 0.00026968503937007873,
"loss": 0.7442,
"step": 1910
},
{
"epoch": 0.5402363534046145,
"grad_norm": 17.69501842397575,
"learning_rate": 0.0002694663167104112,
"loss": 0.6945,
"step": 1920
},
{
"epoch": 0.5430500844119303,
"grad_norm": 22.834715228106134,
"learning_rate": 0.0002692475940507436,
"loss": 0.7521,
"step": 1930
},
{
"epoch": 0.5458638154192459,
"grad_norm": 26.467656611768952,
"learning_rate": 0.0002690288713910761,
"loss": 0.6719,
"step": 1940
},
{
"epoch": 0.5486775464265616,
"grad_norm": 29.568622749960294,
"learning_rate": 0.00026881014873140856,
"loss": 1.0309,
"step": 1950
},
{
"epoch": 0.5514912774338773,
"grad_norm": 9.2347635939369,
"learning_rate": 0.000268591426071741,
"loss": 0.7166,
"step": 1960
},
{
"epoch": 0.554305008441193,
"grad_norm": 19.460279560031523,
"learning_rate": 0.00026837270341207345,
"loss": 0.7064,
"step": 1970
},
{
"epoch": 0.5571187394485088,
"grad_norm": 5.903067753944877,
"learning_rate": 0.0002681539807524059,
"loss": 0.7376,
"step": 1980
},
{
"epoch": 0.5599324704558244,
"grad_norm": 17.371144674890022,
"learning_rate": 0.0002679352580927384,
"loss": 0.6147,
"step": 1990
},
{
"epoch": 0.5627462014631401,
"grad_norm": 5.4409380839404875,
"learning_rate": 0.00026771653543307086,
"loss": 0.5593,
"step": 2000
},
{
"epoch": 0.5655599324704558,
"grad_norm": 43.10864314052242,
"learning_rate": 0.00026749781277340327,
"loss": 0.7555,
"step": 2010
},
{
"epoch": 0.5683736634777715,
"grad_norm": 7.489455426282972,
"learning_rate": 0.00026727909011373574,
"loss": 0.7477,
"step": 2020
},
{
"epoch": 0.5711873944850873,
"grad_norm": 6.66426392264251,
"learning_rate": 0.0002670603674540682,
"loss": 0.6787,
"step": 2030
},
{
"epoch": 0.5740011254924029,
"grad_norm": 15.342229369129983,
"learning_rate": 0.0002668416447944007,
"loss": 0.6353,
"step": 2040
},
{
"epoch": 0.5768148564997186,
"grad_norm": 17.180941723078337,
"learning_rate": 0.00026662292213473315,
"loss": 0.6054,
"step": 2050
},
{
"epoch": 0.5796285875070343,
"grad_norm": 35.277901006510085,
"learning_rate": 0.0002664041994750656,
"loss": 0.7408,
"step": 2060
},
{
"epoch": 0.58244231851435,
"grad_norm": 11.27450226311033,
"learning_rate": 0.00026618547681539804,
"loss": 0.6806,
"step": 2070
},
{
"epoch": 0.5852560495216658,
"grad_norm": 58.49843655029433,
"learning_rate": 0.0002659667541557305,
"loss": 0.994,
"step": 2080
},
{
"epoch": 0.5880697805289814,
"grad_norm": 13.809131175890833,
"learning_rate": 0.000265748031496063,
"loss": 0.609,
"step": 2090
},
{
"epoch": 0.5908835115362971,
"grad_norm": 20.009880236869453,
"learning_rate": 0.00026552930883639545,
"loss": 0.588,
"step": 2100
},
{
"epoch": 0.5936972425436128,
"grad_norm": 7.191101525174044,
"learning_rate": 0.0002653105861767279,
"loss": 0.6758,
"step": 2110
},
{
"epoch": 0.5965109735509285,
"grad_norm": 36.929986801209665,
"learning_rate": 0.00026509186351706033,
"loss": 0.7098,
"step": 2120
},
{
"epoch": 0.5993247045582443,
"grad_norm": 35.8301224810954,
"learning_rate": 0.0002648731408573928,
"loss": 0.5812,
"step": 2130
},
{
"epoch": 0.6021384355655599,
"grad_norm": 22.332697021851985,
"learning_rate": 0.0002646544181977253,
"loss": 0.578,
"step": 2140
},
{
"epoch": 0.6049521665728756,
"grad_norm": 7.695082610709639,
"learning_rate": 0.0002644356955380577,
"loss": 0.5778,
"step": 2150
},
{
"epoch": 0.6077658975801913,
"grad_norm": 17.046853464953895,
"learning_rate": 0.00026421697287839016,
"loss": 0.7022,
"step": 2160
},
{
"epoch": 0.610579628587507,
"grad_norm": 41.35650306806981,
"learning_rate": 0.00026399825021872263,
"loss": 1.036,
"step": 2170
},
{
"epoch": 0.6133933595948228,
"grad_norm": 8.7115876639432,
"learning_rate": 0.0002637795275590551,
"loss": 0.7931,
"step": 2180
},
{
"epoch": 0.6162070906021384,
"grad_norm": 34.58819046075799,
"learning_rate": 0.00026356080489938757,
"loss": 0.643,
"step": 2190
},
{
"epoch": 0.6190208216094542,
"grad_norm": 10.263602247101797,
"learning_rate": 0.00026334208223972,
"loss": 0.6436,
"step": 2200
},
{
"epoch": 0.6218345526167698,
"grad_norm": 46.6429192571408,
"learning_rate": 0.00026312335958005246,
"loss": 0.6991,
"step": 2210
},
{
"epoch": 0.6246482836240855,
"grad_norm": 11.878255221017628,
"learning_rate": 0.0002629046369203849,
"loss": 0.6504,
"step": 2220
},
{
"epoch": 0.6274620146314013,
"grad_norm": 6.479822059289286,
"learning_rate": 0.0002626859142607174,
"loss": 0.5869,
"step": 2230
},
{
"epoch": 0.6302757456387169,
"grad_norm": 12.174115126023809,
"learning_rate": 0.00026246719160104987,
"loss": 0.8186,
"step": 2240
},
{
"epoch": 0.6330894766460327,
"grad_norm": 14.12631818540897,
"learning_rate": 0.00026224846894138234,
"loss": 0.7347,
"step": 2250
},
{
"epoch": 0.6359032076533483,
"grad_norm": 32.041943287347785,
"learning_rate": 0.00026202974628171475,
"loss": 0.7152,
"step": 2260
},
{
"epoch": 0.638716938660664,
"grad_norm": 5.4013176531081655,
"learning_rate": 0.0002618110236220472,
"loss": 0.6162,
"step": 2270
},
{
"epoch": 0.6415306696679798,
"grad_norm": 13.504260914807004,
"learning_rate": 0.0002615923009623797,
"loss": 0.7279,
"step": 2280
},
{
"epoch": 0.6443444006752954,
"grad_norm": 9.620376593872086,
"learning_rate": 0.00026137357830271216,
"loss": 0.6628,
"step": 2290
},
{
"epoch": 0.6471581316826112,
"grad_norm": 22.427874242699758,
"learning_rate": 0.0002611548556430446,
"loss": 0.6983,
"step": 2300
},
{
"epoch": 0.6499718626899268,
"grad_norm": 41.554954362999574,
"learning_rate": 0.00026093613298337705,
"loss": 1.0201,
"step": 2310
},
{
"epoch": 0.6527855936972425,
"grad_norm": 7.036254242716845,
"learning_rate": 0.0002607174103237095,
"loss": 0.8125,
"step": 2320
},
{
"epoch": 0.6555993247045583,
"grad_norm": 13.380680950268676,
"learning_rate": 0.000260498687664042,
"loss": 0.6904,
"step": 2330
},
{
"epoch": 0.6584130557118739,
"grad_norm": 22.884181519089868,
"learning_rate": 0.0002602799650043744,
"loss": 1.0037,
"step": 2340
},
{
"epoch": 0.6612267867191897,
"grad_norm": 14.511141210797714,
"learning_rate": 0.00026006124234470687,
"loss": 0.7414,
"step": 2350
},
{
"epoch": 0.6640405177265053,
"grad_norm": 13.431978101097688,
"learning_rate": 0.00025984251968503934,
"loss": 0.7078,
"step": 2360
},
{
"epoch": 0.666854248733821,
"grad_norm": 29.24656191831114,
"learning_rate": 0.0002596237970253718,
"loss": 0.5655,
"step": 2370
},
{
"epoch": 0.6696679797411368,
"grad_norm": 6.150879349284207,
"learning_rate": 0.0002594050743657043,
"loss": 0.64,
"step": 2380
},
{
"epoch": 0.6724817107484524,
"grad_norm": 6.533201611095304,
"learning_rate": 0.0002591863517060367,
"loss": 0.6766,
"step": 2390
},
{
"epoch": 0.6752954417557682,
"grad_norm": 11.440054847870906,
"learning_rate": 0.00025896762904636917,
"loss": 0.5998,
"step": 2400
},
{
"epoch": 0.6781091727630838,
"grad_norm": 31.143749184180702,
"learning_rate": 0.00025874890638670164,
"loss": 0.5311,
"step": 2410
},
{
"epoch": 0.6809229037703995,
"grad_norm": 22.852634622577742,
"learning_rate": 0.0002585301837270341,
"loss": 0.5925,
"step": 2420
},
{
"epoch": 0.6837366347777153,
"grad_norm": 24.588223403495608,
"learning_rate": 0.0002583114610673666,
"loss": 0.6734,
"step": 2430
},
{
"epoch": 0.6865503657850309,
"grad_norm": 28.048648610791382,
"learning_rate": 0.00025809273840769905,
"loss": 0.618,
"step": 2440
},
{
"epoch": 0.6893640967923467,
"grad_norm": 10.406220663733844,
"learning_rate": 0.00025787401574803146,
"loss": 0.5913,
"step": 2450
},
{
"epoch": 0.6921778277996623,
"grad_norm": 6.811024400622029,
"learning_rate": 0.00025765529308836393,
"loss": 0.6074,
"step": 2460
},
{
"epoch": 0.694991558806978,
"grad_norm": 12.155916590641617,
"learning_rate": 0.0002574365704286964,
"loss": 0.5704,
"step": 2470
},
{
"epoch": 0.6978052898142938,
"grad_norm": 9.031515529220442,
"learning_rate": 0.0002572178477690289,
"loss": 0.6423,
"step": 2480
},
{
"epoch": 0.7006190208216094,
"grad_norm": 14.205953959415192,
"learning_rate": 0.0002569991251093613,
"loss": 0.6407,
"step": 2490
},
{
"epoch": 0.7034327518289252,
"grad_norm": 48.32963976504197,
"learning_rate": 0.00025678040244969376,
"loss": 0.6504,
"step": 2500
},
{
"epoch": 0.7062464828362408,
"grad_norm": 23.896138968892455,
"learning_rate": 0.00025656167979002623,
"loss": 0.6352,
"step": 2510
},
{
"epoch": 0.7090602138435566,
"grad_norm": 11.036690380167714,
"learning_rate": 0.00025634295713035865,
"loss": 0.7513,
"step": 2520
},
{
"epoch": 0.7118739448508723,
"grad_norm": 38.96438150155598,
"learning_rate": 0.0002561242344706911,
"loss": 0.8657,
"step": 2530
},
{
"epoch": 0.7146876758581879,
"grad_norm": 42.72643454806835,
"learning_rate": 0.0002559055118110236,
"loss": 0.9091,
"step": 2540
},
{
"epoch": 0.7175014068655037,
"grad_norm": 16.610380140185,
"learning_rate": 0.00025568678915135606,
"loss": 0.7069,
"step": 2550
},
{
"epoch": 0.7203151378728193,
"grad_norm": 22.40904574778732,
"learning_rate": 0.0002554680664916885,
"loss": 0.5857,
"step": 2560
},
{
"epoch": 0.7231288688801351,
"grad_norm": 14.389267092163761,
"learning_rate": 0.000255249343832021,
"loss": 0.6018,
"step": 2570
},
{
"epoch": 0.7259425998874508,
"grad_norm": 13.683335764917064,
"learning_rate": 0.0002550306211723534,
"loss": 0.5358,
"step": 2580
},
{
"epoch": 0.7287563308947664,
"grad_norm": 30.498280377063637,
"learning_rate": 0.0002548118985126859,
"loss": 0.9904,
"step": 2590
},
{
"epoch": 0.7315700619020822,
"grad_norm": 8.683287013020767,
"learning_rate": 0.00025459317585301835,
"loss": 0.71,
"step": 2600
},
{
"epoch": 0.7343837929093978,
"grad_norm": 7.103289048611902,
"learning_rate": 0.0002543744531933508,
"loss": 0.8194,
"step": 2610
},
{
"epoch": 0.7371975239167136,
"grad_norm": 6.281668196582603,
"learning_rate": 0.0002541557305336833,
"loss": 0.7173,
"step": 2620
},
{
"epoch": 0.7400112549240293,
"grad_norm": 5.39806094595311,
"learning_rate": 0.00025393700787401576,
"loss": 0.7826,
"step": 2630
},
{
"epoch": 0.7428249859313449,
"grad_norm": 32.09155901494012,
"learning_rate": 0.0002537182852143482,
"loss": 0.6226,
"step": 2640
},
{
"epoch": 0.7456387169386607,
"grad_norm": 8.519967935941418,
"learning_rate": 0.00025349956255468065,
"loss": 0.7454,
"step": 2650
},
{
"epoch": 0.7484524479459763,
"grad_norm": 42.035625193177715,
"learning_rate": 0.0002532808398950131,
"loss": 0.5172,
"step": 2660
},
{
"epoch": 0.7512661789532921,
"grad_norm": 9.967889876277471,
"learning_rate": 0.0002530621172353456,
"loss": 1.0016,
"step": 2670
},
{
"epoch": 0.7540799099606078,
"grad_norm": 26.068769304112433,
"learning_rate": 0.000252843394575678,
"loss": 0.7935,
"step": 2680
},
{
"epoch": 0.7568936409679234,
"grad_norm": 20.87767388709777,
"learning_rate": 0.0002526246719160105,
"loss": 0.8831,
"step": 2690
},
{
"epoch": 0.7597073719752392,
"grad_norm": 23.861680316972155,
"learning_rate": 0.00025240594925634294,
"loss": 0.7566,
"step": 2700
},
{
"epoch": 0.7625211029825548,
"grad_norm": 19.606881060581557,
"learning_rate": 0.00025218722659667536,
"loss": 0.6166,
"step": 2710
},
{
"epoch": 0.7653348339898706,
"grad_norm": 23.231679663361476,
"learning_rate": 0.00025196850393700783,
"loss": 0.7444,
"step": 2720
},
{
"epoch": 0.7681485649971863,
"grad_norm": 14.475225232701424,
"learning_rate": 0.0002517497812773403,
"loss": 0.6132,
"step": 2730
},
{
"epoch": 0.770962296004502,
"grad_norm": 10.196976505426665,
"learning_rate": 0.00025153105861767277,
"loss": 0.6414,
"step": 2740
},
{
"epoch": 0.7737760270118177,
"grad_norm": 12.26153672415283,
"learning_rate": 0.00025131233595800524,
"loss": 0.6497,
"step": 2750
},
{
"epoch": 0.7765897580191333,
"grad_norm": 11.399029351648554,
"learning_rate": 0.00025109361329833766,
"loss": 0.6397,
"step": 2760
},
{
"epoch": 0.7794034890264491,
"grad_norm": 29.29036307836923,
"learning_rate": 0.0002508748906386701,
"loss": 0.6448,
"step": 2770
},
{
"epoch": 0.7822172200337648,
"grad_norm": 6.5421868930376315,
"learning_rate": 0.0002506561679790026,
"loss": 0.619,
"step": 2780
},
{
"epoch": 0.7850309510410804,
"grad_norm": 28.964608214250898,
"learning_rate": 0.00025043744531933507,
"loss": 0.6457,
"step": 2790
},
{
"epoch": 0.7878446820483962,
"grad_norm": 19.105194261412365,
"learning_rate": 0.00025021872265966754,
"loss": 0.8734,
"step": 2800
},
{
"epoch": 0.7906584130557118,
"grad_norm": 21.86681738792712,
"learning_rate": 0.00025,
"loss": 0.6627,
"step": 2810
},
{
"epoch": 0.7934721440630276,
"grad_norm": 21.420403379202583,
"learning_rate": 0.0002497812773403325,
"loss": 0.6947,
"step": 2820
},
{
"epoch": 0.7962858750703433,
"grad_norm": 51.77141397970572,
"learning_rate": 0.0002495625546806649,
"loss": 0.6478,
"step": 2830
},
{
"epoch": 0.799099606077659,
"grad_norm": 19.60546919995034,
"learning_rate": 0.00024934383202099736,
"loss": 0.6738,
"step": 2840
},
{
"epoch": 0.8019133370849747,
"grad_norm": 5.930058860961108,
"learning_rate": 0.00024912510936132983,
"loss": 0.676,
"step": 2850
},
{
"epoch": 0.8047270680922903,
"grad_norm": 9.793830622150702,
"learning_rate": 0.0002489063867016623,
"loss": 0.5543,
"step": 2860
},
{
"epoch": 0.8075407990996061,
"grad_norm": 30.643668253643902,
"learning_rate": 0.0002486876640419947,
"loss": 0.6535,
"step": 2870
},
{
"epoch": 0.8103545301069218,
"grad_norm": 18.840891754548007,
"learning_rate": 0.0002484689413823272,
"loss": 0.7703,
"step": 2880
},
{
"epoch": 0.8131682611142375,
"grad_norm": 11.630930014907443,
"learning_rate": 0.00024825021872265966,
"loss": 0.6506,
"step": 2890
},
{
"epoch": 0.8159819921215532,
"grad_norm": 11.371539982197872,
"learning_rate": 0.00024803149606299207,
"loss": 0.6467,
"step": 2900
},
{
"epoch": 0.8187957231288688,
"grad_norm": 10.427236030304023,
"learning_rate": 0.00024781277340332454,
"loss": 0.8076,
"step": 2910
},
{
"epoch": 0.8216094541361846,
"grad_norm": 36.87576985444582,
"learning_rate": 0.000247594050743657,
"loss": 0.6055,
"step": 2920
},
{
"epoch": 0.8244231851435003,
"grad_norm": 35.48244372566825,
"learning_rate": 0.0002473753280839895,
"loss": 0.8045,
"step": 2930
},
{
"epoch": 0.827236916150816,
"grad_norm": 24.817299467837056,
"learning_rate": 0.00024715660542432195,
"loss": 0.6533,
"step": 2940
},
{
"epoch": 0.8300506471581317,
"grad_norm": 9.141011529069573,
"learning_rate": 0.00024693788276465437,
"loss": 0.6856,
"step": 2950
},
{
"epoch": 0.8328643781654473,
"grad_norm": 16.064568145247428,
"learning_rate": 0.00024671916010498684,
"loss": 0.6118,
"step": 2960
},
{
"epoch": 0.8356781091727631,
"grad_norm": 14.088534153379833,
"learning_rate": 0.0002465004374453193,
"loss": 0.6359,
"step": 2970
},
{
"epoch": 0.8384918401800788,
"grad_norm": 10.800798513388331,
"learning_rate": 0.0002462817147856518,
"loss": 0.4701,
"step": 2980
},
{
"epoch": 0.8413055711873945,
"grad_norm": 23.57379674968355,
"learning_rate": 0.00024606299212598425,
"loss": 0.8891,
"step": 2990
},
{
"epoch": 0.8441193021947102,
"grad_norm": 19.087599267026963,
"learning_rate": 0.0002458442694663167,
"loss": 0.6812,
"step": 3000
},
{
"epoch": 0.8469330332020258,
"grad_norm": 5.851382871921484,
"learning_rate": 0.00024562554680664913,
"loss": 0.5135,
"step": 3010
},
{
"epoch": 0.8497467642093416,
"grad_norm": 32.51696153998222,
"learning_rate": 0.0002454068241469816,
"loss": 0.6505,
"step": 3020
},
{
"epoch": 0.8525604952166573,
"grad_norm": 19.264326332382478,
"learning_rate": 0.0002451881014873141,
"loss": 0.7129,
"step": 3030
},
{
"epoch": 0.855374226223973,
"grad_norm": 16.974285473343233,
"learning_rate": 0.00024496937882764654,
"loss": 0.7383,
"step": 3040
},
{
"epoch": 0.8581879572312887,
"grad_norm": 26.864035695775446,
"learning_rate": 0.00024475065616797896,
"loss": 0.3988,
"step": 3050
},
{
"epoch": 0.8610016882386043,
"grad_norm": 13.174415289969117,
"learning_rate": 0.00024453193350831143,
"loss": 0.7362,
"step": 3060
},
{
"epoch": 0.8638154192459201,
"grad_norm": 17.500378614825863,
"learning_rate": 0.0002443132108486439,
"loss": 0.7758,
"step": 3070
},
{
"epoch": 0.8666291502532358,
"grad_norm": 9.282760951889967,
"learning_rate": 0.00024409448818897634,
"loss": 0.4896,
"step": 3080
},
{
"epoch": 0.8694428812605515,
"grad_norm": 24.683487403388195,
"learning_rate": 0.0002438757655293088,
"loss": 0.6937,
"step": 3090
},
{
"epoch": 0.8722566122678672,
"grad_norm": 6.130822987006422,
"learning_rate": 0.00024365704286964128,
"loss": 0.6025,
"step": 3100
},
{
"epoch": 0.8750703432751828,
"grad_norm": 41.848473080771385,
"learning_rate": 0.00024343832020997373,
"loss": 0.7405,
"step": 3110
},
{
"epoch": 0.8778840742824986,
"grad_norm": 16.807668144029588,
"learning_rate": 0.0002432195975503062,
"loss": 0.8849,
"step": 3120
},
{
"epoch": 0.8806978052898143,
"grad_norm": 12.875519609787505,
"learning_rate": 0.00024300087489063867,
"loss": 0.6274,
"step": 3130
},
{
"epoch": 0.88351153629713,
"grad_norm": 17.334020913099007,
"learning_rate": 0.00024278215223097108,
"loss": 0.4966,
"step": 3140
},
{
"epoch": 0.8863252673044457,
"grad_norm": 3.003018774866583,
"learning_rate": 0.00024256342957130355,
"loss": 0.6363,
"step": 3150
},
{
"epoch": 0.8891389983117614,
"grad_norm": 21.403704259371647,
"learning_rate": 0.00024234470691163602,
"loss": 0.6202,
"step": 3160
},
{
"epoch": 0.8919527293190771,
"grad_norm": 20.49088520270555,
"learning_rate": 0.0002421259842519685,
"loss": 0.6555,
"step": 3170
},
{
"epoch": 0.8947664603263928,
"grad_norm": 3.2248953164538157,
"learning_rate": 0.00024190726159230096,
"loss": 0.667,
"step": 3180
},
{
"epoch": 0.8975801913337085,
"grad_norm": 20.43806666316274,
"learning_rate": 0.0002416885389326334,
"loss": 0.8157,
"step": 3190
},
{
"epoch": 0.9003939223410242,
"grad_norm": 43.16776399371423,
"learning_rate": 0.00024146981627296585,
"loss": 0.9972,
"step": 3200
},
{
"epoch": 0.9032076533483399,
"grad_norm": 7.49046572212332,
"learning_rate": 0.00024125109361329832,
"loss": 0.5636,
"step": 3210
},
{
"epoch": 0.9060213843556556,
"grad_norm": 8.741984739159578,
"learning_rate": 0.00024103237095363076,
"loss": 0.5747,
"step": 3220
},
{
"epoch": 0.9088351153629713,
"grad_norm": 54.83042946791486,
"learning_rate": 0.00024081364829396323,
"loss": 0.5467,
"step": 3230
},
{
"epoch": 0.911648846370287,
"grad_norm": 5.610010207094356,
"learning_rate": 0.0002405949256342957,
"loss": 0.8028,
"step": 3240
},
{
"epoch": 0.9144625773776027,
"grad_norm": 19.20818992178154,
"learning_rate": 0.00024037620297462817,
"loss": 0.5641,
"step": 3250
},
{
"epoch": 0.9172763083849184,
"grad_norm": 19.18323455494463,
"learning_rate": 0.0002401574803149606,
"loss": 0.8983,
"step": 3260
},
{
"epoch": 0.9200900393922341,
"grad_norm": 7.863834695625984,
"learning_rate": 0.00023993875765529306,
"loss": 0.5333,
"step": 3270
},
{
"epoch": 0.9229037703995498,
"grad_norm": 11.037689804640895,
"learning_rate": 0.00023972003499562553,
"loss": 0.5062,
"step": 3280
},
{
"epoch": 0.9257175014068655,
"grad_norm": 14.170498863161551,
"learning_rate": 0.000239501312335958,
"loss": 0.4058,
"step": 3290
},
{
"epoch": 0.9285312324141812,
"grad_norm": 8.276743998093925,
"learning_rate": 0.00023928258967629044,
"loss": 0.8532,
"step": 3300
},
{
"epoch": 0.9313449634214969,
"grad_norm": 31.54134403213458,
"learning_rate": 0.0002390638670166229,
"loss": 0.7638,
"step": 3310
},
{
"epoch": 0.9341586944288126,
"grad_norm": 21.816033120921777,
"learning_rate": 0.00023884514435695538,
"loss": 0.6246,
"step": 3320
},
{
"epoch": 0.9369724254361284,
"grad_norm": 18.097032537484406,
"learning_rate": 0.0002386264216972878,
"loss": 0.5869,
"step": 3330
},
{
"epoch": 0.939786156443444,
"grad_norm": 30.83082131815264,
"learning_rate": 0.00023840769903762027,
"loss": 0.6802,
"step": 3340
},
{
"epoch": 0.9425998874507597,
"grad_norm": 47.924062613651145,
"learning_rate": 0.00023818897637795274,
"loss": 0.7447,
"step": 3350
},
{
"epoch": 0.9454136184580754,
"grad_norm": 9.599329887116856,
"learning_rate": 0.0002379702537182852,
"loss": 0.4587,
"step": 3360
},
{
"epoch": 0.9482273494653911,
"grad_norm": 24.233656237412927,
"learning_rate": 0.00023775153105861765,
"loss": 1.0523,
"step": 3370
},
{
"epoch": 0.9510410804727069,
"grad_norm": 8.951427533475114,
"learning_rate": 0.00023753280839895012,
"loss": 0.6428,
"step": 3380
},
{
"epoch": 0.9538548114800225,
"grad_norm": 7.42291585951858,
"learning_rate": 0.00023731408573928256,
"loss": 0.5499,
"step": 3390
},
{
"epoch": 0.9566685424873382,
"grad_norm": 15.702658109698685,
"learning_rate": 0.00023709536307961503,
"loss": 0.7035,
"step": 3400
},
{
"epoch": 0.9594822734946539,
"grad_norm": 20.19323341543604,
"learning_rate": 0.00023687664041994747,
"loss": 0.6141,
"step": 3410
},
{
"epoch": 0.9622960045019696,
"grad_norm": 43.77793213323037,
"learning_rate": 0.00023665791776027994,
"loss": 0.6725,
"step": 3420
},
{
"epoch": 0.9651097355092854,
"grad_norm": 32.8128961448371,
"learning_rate": 0.00023643919510061241,
"loss": 0.5943,
"step": 3430
},
{
"epoch": 0.967923466516601,
"grad_norm": 46.81002844829282,
"learning_rate": 0.00023622047244094488,
"loss": 0.7372,
"step": 3440
},
{
"epoch": 0.9707371975239167,
"grad_norm": 11.402187714876383,
"learning_rate": 0.0002360017497812773,
"loss": 0.6388,
"step": 3450
},
{
"epoch": 0.9735509285312324,
"grad_norm": 18.063184970189784,
"learning_rate": 0.00023578302712160977,
"loss": 0.6867,
"step": 3460
},
{
"epoch": 0.9763646595385481,
"grad_norm": 25.879726827027653,
"learning_rate": 0.00023556430446194224,
"loss": 0.5965,
"step": 3470
},
{
"epoch": 0.9791783905458639,
"grad_norm": 20.717474130396493,
"learning_rate": 0.0002353455818022747,
"loss": 0.7171,
"step": 3480
},
{
"epoch": 0.9819921215531795,
"grad_norm": 18.1608450158541,
"learning_rate": 0.00023512685914260715,
"loss": 0.584,
"step": 3490
},
{
"epoch": 0.9848058525604952,
"grad_norm": 10.9801787450404,
"learning_rate": 0.00023490813648293962,
"loss": 0.4611,
"step": 3500
},
{
"epoch": 0.9876195835678109,
"grad_norm": 59.845266656605816,
"learning_rate": 0.00023468941382327207,
"loss": 0.6816,
"step": 3510
},
{
"epoch": 0.9904333145751266,
"grad_norm": 22.584791520562405,
"learning_rate": 0.0002344706911636045,
"loss": 0.6226,
"step": 3520
},
{
"epoch": 0.9932470455824424,
"grad_norm": 25.253863357778947,
"learning_rate": 0.00023425196850393698,
"loss": 0.7587,
"step": 3530
},
{
"epoch": 0.996060776589758,
"grad_norm": 13.205996634467093,
"learning_rate": 0.00023403324584426945,
"loss": 0.5234,
"step": 3540
},
{
"epoch": 0.9988745075970737,
"grad_norm": 12.477784447497413,
"learning_rate": 0.00023381452318460192,
"loss": 0.7967,
"step": 3550
},
{
"epoch": 1.0,
"eval_0_f1": 0.5288677130044843,
"eval_0_precision": 0.39911167512690354,
"eval_0_recall": 0.7836378737541528,
"eval_1_f1": 0.6981504758484468,
"eval_1_precision": 0.8818326151054661,
"eval_1_recall": 0.5777975925100312,
"eval_accuracy": 0.6320455291671226,
"eval_loss": 0.6142578125,
"eval_runtime": 469.6152,
"eval_samples_per_second": 19.456,
"eval_steps_per_second": 3.243,
"step": 3554
},
{
"epoch": 1.0016882386043895,
"grad_norm": 20.289578848031194,
"learning_rate": 0.00023359580052493436,
"loss": 0.5507,
"step": 3560
},
{
"epoch": 1.004501969611705,
"grad_norm": 15.822667712574415,
"learning_rate": 0.00023337707786526683,
"loss": 0.5981,
"step": 3570
},
{
"epoch": 1.0073157006190208,
"grad_norm": 23.851500914988023,
"learning_rate": 0.00023315835520559927,
"loss": 0.5925,
"step": 3580
},
{
"epoch": 1.0101294316263365,
"grad_norm": 28.90154599345354,
"learning_rate": 0.00023293963254593174,
"loss": 0.6938,
"step": 3590
},
{
"epoch": 1.0129431626336522,
"grad_norm": 6.673268351357181,
"learning_rate": 0.0002327209098862642,
"loss": 0.4444,
"step": 3600
},
{
"epoch": 1.015756893640968,
"grad_norm": 24.026678476440093,
"learning_rate": 0.00023250218722659666,
"loss": 0.6206,
"step": 3610
},
{
"epoch": 1.0185706246482835,
"grad_norm": 12.993158134163783,
"learning_rate": 0.00023228346456692913,
"loss": 0.7656,
"step": 3620
},
{
"epoch": 1.0213843556555993,
"grad_norm": 8.661592494763198,
"learning_rate": 0.0002320647419072616,
"loss": 0.5644,
"step": 3630
},
{
"epoch": 1.024198086662915,
"grad_norm": 17.738436058324886,
"learning_rate": 0.000231846019247594,
"loss": 0.6354,
"step": 3640
},
{
"epoch": 1.0270118176702308,
"grad_norm": 8.612642747217874,
"learning_rate": 0.00023162729658792648,
"loss": 0.7017,
"step": 3650
},
{
"epoch": 1.0298255486775465,
"grad_norm": 36.741673298246305,
"learning_rate": 0.00023140857392825895,
"loss": 0.6211,
"step": 3660
},
{
"epoch": 1.032639279684862,
"grad_norm": 12.750982475761448,
"learning_rate": 0.0002311898512685914,
"loss": 0.5828,
"step": 3670
},
{
"epoch": 1.0354530106921778,
"grad_norm": 12.486810558826239,
"learning_rate": 0.00023097112860892387,
"loss": 0.5709,
"step": 3680
},
{
"epoch": 1.0382667416994935,
"grad_norm": 24.920452697969928,
"learning_rate": 0.00023075240594925634,
"loss": 0.5783,
"step": 3690
},
{
"epoch": 1.0410804727068093,
"grad_norm": 6.301604046934106,
"learning_rate": 0.00023053368328958878,
"loss": 0.4701,
"step": 3700
},
{
"epoch": 1.043894203714125,
"grad_norm": 1.554082347905222,
"learning_rate": 0.00023031496062992122,
"loss": 0.5536,
"step": 3710
},
{
"epoch": 1.0467079347214405,
"grad_norm": 22.334353822789094,
"learning_rate": 0.0002300962379702537,
"loss": 0.6868,
"step": 3720
},
{
"epoch": 1.0495216657287563,
"grad_norm": 29.418886082506507,
"learning_rate": 0.00022987751531058616,
"loss": 0.9408,
"step": 3730
},
{
"epoch": 1.052335396736072,
"grad_norm": 20.36838671289186,
"learning_rate": 0.00022965879265091863,
"loss": 0.6802,
"step": 3740
},
{
"epoch": 1.0551491277433878,
"grad_norm": 4.954178463432806,
"learning_rate": 0.00022944006999125107,
"loss": 0.6108,
"step": 3750
},
{
"epoch": 1.0579628587507035,
"grad_norm": 30.711986018263367,
"learning_rate": 0.00022922134733158352,
"loss": 0.5281,
"step": 3760
},
{
"epoch": 1.060776589758019,
"grad_norm": 10.233269531300138,
"learning_rate": 0.000229002624671916,
"loss": 0.5335,
"step": 3770
},
{
"epoch": 1.0635903207653348,
"grad_norm": 63.23746707068614,
"learning_rate": 0.00022878390201224843,
"loss": 0.881,
"step": 3780
},
{
"epoch": 1.0664040517726505,
"grad_norm": 5.131625836660247,
"learning_rate": 0.0002285651793525809,
"loss": 0.9595,
"step": 3790
},
{
"epoch": 1.0692177827799663,
"grad_norm": 13.264342728525087,
"learning_rate": 0.00022834645669291337,
"loss": 0.4729,
"step": 3800
},
{
"epoch": 1.072031513787282,
"grad_norm": 3.92201905465672,
"learning_rate": 0.00022812773403324584,
"loss": 0.6524,
"step": 3810
},
{
"epoch": 1.0748452447945978,
"grad_norm": 22.941903277650525,
"learning_rate": 0.0002279090113735783,
"loss": 0.6657,
"step": 3820
},
{
"epoch": 1.0776589758019133,
"grad_norm": 18.622940780105395,
"learning_rate": 0.00022769028871391073,
"loss": 0.4346,
"step": 3830
},
{
"epoch": 1.080472706809229,
"grad_norm": 10.884115952331454,
"learning_rate": 0.0002274715660542432,
"loss": 0.4555,
"step": 3840
},
{
"epoch": 1.0832864378165448,
"grad_norm": 24.118386588827224,
"learning_rate": 0.00022725284339457567,
"loss": 0.7681,
"step": 3850
},
{
"epoch": 1.0861001688238605,
"grad_norm": 20.031875879534606,
"learning_rate": 0.0002270341207349081,
"loss": 0.7259,
"step": 3860
},
{
"epoch": 1.088913899831176,
"grad_norm": 13.903004192497082,
"learning_rate": 0.00022681539807524058,
"loss": 0.6706,
"step": 3870
},
{
"epoch": 1.0917276308384918,
"grad_norm": 12.51136628868392,
"learning_rate": 0.00022659667541557305,
"loss": 0.6882,
"step": 3880
},
{
"epoch": 1.0945413618458075,
"grad_norm": 15.707389951014497,
"learning_rate": 0.00022637795275590547,
"loss": 0.545,
"step": 3890
},
{
"epoch": 1.0973550928531233,
"grad_norm": 21.377996746325916,
"learning_rate": 0.00022615923009623794,
"loss": 0.5169,
"step": 3900
},
{
"epoch": 1.100168823860439,
"grad_norm": 31.777506839495466,
"learning_rate": 0.0002259405074365704,
"loss": 0.8108,
"step": 3910
},
{
"epoch": 1.1029825548677545,
"grad_norm": 17.123955716971917,
"learning_rate": 0.00022572178477690288,
"loss": 0.646,
"step": 3920
},
{
"epoch": 1.1057962858750703,
"grad_norm": 36.33273283007592,
"learning_rate": 0.00022550306211723535,
"loss": 0.7394,
"step": 3930
},
{
"epoch": 1.108610016882386,
"grad_norm": 18.821091198359895,
"learning_rate": 0.0002252843394575678,
"loss": 0.5749,
"step": 3940
},
{
"epoch": 1.1114237478897018,
"grad_norm": 10.61252309290639,
"learning_rate": 0.00022506561679790023,
"loss": 0.5137,
"step": 3950
},
{
"epoch": 1.1142374788970175,
"grad_norm": 13.982125625637144,
"learning_rate": 0.0002248468941382327,
"loss": 0.6647,
"step": 3960
},
{
"epoch": 1.117051209904333,
"grad_norm": 42.981624236536284,
"learning_rate": 0.00022462817147856514,
"loss": 0.5238,
"step": 3970
},
{
"epoch": 1.1198649409116488,
"grad_norm": 27.54632412072402,
"learning_rate": 0.00022440944881889761,
"loss": 0.6707,
"step": 3980
},
{
"epoch": 1.1226786719189645,
"grad_norm": 23.75601752367408,
"learning_rate": 0.00022419072615923008,
"loss": 0.8241,
"step": 3990
},
{
"epoch": 1.1254924029262803,
"grad_norm": 32.47147919838835,
"learning_rate": 0.00022397200349956255,
"loss": 0.6688,
"step": 4000
},
{
"epoch": 1.128306133933596,
"grad_norm": 18.348598242445995,
"learning_rate": 0.00022375328083989497,
"loss": 0.5989,
"step": 4010
},
{
"epoch": 1.1311198649409118,
"grad_norm": 12.324401214506343,
"learning_rate": 0.00022353455818022744,
"loss": 0.5851,
"step": 4020
},
{
"epoch": 1.1339335959482273,
"grad_norm": 17.172404997190124,
"learning_rate": 0.0002233158355205599,
"loss": 0.5568,
"step": 4030
},
{
"epoch": 1.136747326955543,
"grad_norm": 23.04273982046674,
"learning_rate": 0.00022309711286089238,
"loss": 0.5822,
"step": 4040
},
{
"epoch": 1.1395610579628588,
"grad_norm": 7.802471124937242,
"learning_rate": 0.00022287839020122482,
"loss": 0.4082,
"step": 4050
},
{
"epoch": 1.1423747889701745,
"grad_norm": 11.2591451273763,
"learning_rate": 0.0002226596675415573,
"loss": 0.4935,
"step": 4060
},
{
"epoch": 1.14518851997749,
"grad_norm": 9.837614966281794,
"learning_rate": 0.00022244094488188976,
"loss": 0.7146,
"step": 4070
},
{
"epoch": 1.1480022509848058,
"grad_norm": 10.946277605810202,
"learning_rate": 0.00022222222222222218,
"loss": 0.6566,
"step": 4080
},
{
"epoch": 1.1508159819921215,
"grad_norm": 17.85614689532493,
"learning_rate": 0.00022200349956255465,
"loss": 0.4754,
"step": 4090
},
{
"epoch": 1.1536297129994373,
"grad_norm": 24.06573596597373,
"learning_rate": 0.00022178477690288712,
"loss": 0.6678,
"step": 4100
},
{
"epoch": 1.156443444006753,
"grad_norm": 23.587510010090995,
"learning_rate": 0.0002215660542432196,
"loss": 0.5681,
"step": 4110
},
{
"epoch": 1.1592571750140688,
"grad_norm": 15.082272570896748,
"learning_rate": 0.00022134733158355206,
"loss": 0.4931,
"step": 4120
},
{
"epoch": 1.1620709060213843,
"grad_norm": 34.70870912668241,
"learning_rate": 0.0002211286089238845,
"loss": 0.5661,
"step": 4130
},
{
"epoch": 1.1648846370287,
"grad_norm": 11.548252271349462,
"learning_rate": 0.00022090988626421694,
"loss": 0.4761,
"step": 4140
},
{
"epoch": 1.1676983680360158,
"grad_norm": 28.412960573382996,
"learning_rate": 0.00022069116360454941,
"loss": 0.5887,
"step": 4150
},
{
"epoch": 1.1705120990433315,
"grad_norm": 10.783805303855186,
"learning_rate": 0.00022047244094488186,
"loss": 0.5948,
"step": 4160
},
{
"epoch": 1.173325830050647,
"grad_norm": 24.93639136840195,
"learning_rate": 0.00022025371828521433,
"loss": 0.6058,
"step": 4170
},
{
"epoch": 1.1761395610579628,
"grad_norm": 26.39448202076931,
"learning_rate": 0.0002200349956255468,
"loss": 0.607,
"step": 4180
},
{
"epoch": 1.1789532920652785,
"grad_norm": 14.904096598298732,
"learning_rate": 0.00021981627296587927,
"loss": 0.548,
"step": 4190
},
{
"epoch": 1.1817670230725943,
"grad_norm": 17.73765885454678,
"learning_rate": 0.00021959755030621168,
"loss": 0.7417,
"step": 4200
},
{
"epoch": 1.18458075407991,
"grad_norm": 2.867842719491419,
"learning_rate": 0.00021937882764654415,
"loss": 0.6395,
"step": 4210
},
{
"epoch": 1.1873944850872258,
"grad_norm": 34.56602745611629,
"learning_rate": 0.00021916010498687662,
"loss": 0.5349,
"step": 4220
},
{
"epoch": 1.1902082160945413,
"grad_norm": 22.77844815887848,
"learning_rate": 0.0002189413823272091,
"loss": 0.61,
"step": 4230
},
{
"epoch": 1.193021947101857,
"grad_norm": 42.626450175565964,
"learning_rate": 0.00021872265966754154,
"loss": 0.7502,
"step": 4240
},
{
"epoch": 1.1958356781091728,
"grad_norm": 10.693548169842728,
"learning_rate": 0.000218503937007874,
"loss": 0.4929,
"step": 4250
},
{
"epoch": 1.1986494091164885,
"grad_norm": 34.637471031794966,
"learning_rate": 0.00021828521434820645,
"loss": 0.5426,
"step": 4260
},
{
"epoch": 1.201463140123804,
"grad_norm": 43.976042152968205,
"learning_rate": 0.0002180664916885389,
"loss": 0.4653,
"step": 4270
},
{
"epoch": 1.2042768711311198,
"grad_norm": 6.320882760905354,
"learning_rate": 0.00021784776902887136,
"loss": 0.8992,
"step": 4280
},
{
"epoch": 1.2070906021384356,
"grad_norm": 8.502068954123935,
"learning_rate": 0.00021762904636920383,
"loss": 0.6101,
"step": 4290
},
{
"epoch": 1.2099043331457513,
"grad_norm": 11.429050183991922,
"learning_rate": 0.0002174103237095363,
"loss": 0.5721,
"step": 4300
},
{
"epoch": 1.212718064153067,
"grad_norm": 17.562444133601815,
"learning_rate": 0.00021719160104986874,
"loss": 0.6881,
"step": 4310
},
{
"epoch": 1.2155317951603828,
"grad_norm": 21.19106486102643,
"learning_rate": 0.00021697287839020121,
"loss": 0.5565,
"step": 4320
},
{
"epoch": 1.2183455261676983,
"grad_norm": 12.164118551052857,
"learning_rate": 0.00021675415573053366,
"loss": 0.6187,
"step": 4330
},
{
"epoch": 1.221159257175014,
"grad_norm": 5.033893258856872,
"learning_rate": 0.00021653543307086613,
"loss": 0.2693,
"step": 4340
},
{
"epoch": 1.2239729881823298,
"grad_norm": 1.4732793797472918,
"learning_rate": 0.00021631671041119857,
"loss": 1.0616,
"step": 4350
},
{
"epoch": 1.2267867191896455,
"grad_norm": 8.376633978447819,
"learning_rate": 0.00021609798775153104,
"loss": 0.8353,
"step": 4360
},
{
"epoch": 1.229600450196961,
"grad_norm": 30.38632947822225,
"learning_rate": 0.0002158792650918635,
"loss": 0.7668,
"step": 4370
},
{
"epoch": 1.2324141812042768,
"grad_norm": 20.42829408507086,
"learning_rate": 0.00021566054243219598,
"loss": 0.5452,
"step": 4380
},
{
"epoch": 1.2352279122115926,
"grad_norm": 11.244453757125129,
"learning_rate": 0.0002154418197725284,
"loss": 0.5771,
"step": 4390
},
{
"epoch": 1.2380416432189083,
"grad_norm": 29.01355049880867,
"learning_rate": 0.00021522309711286087,
"loss": 0.7755,
"step": 4400
},
{
"epoch": 1.240855374226224,
"grad_norm": 22.1695788769221,
"learning_rate": 0.00021500437445319334,
"loss": 0.6102,
"step": 4410
},
{
"epoch": 1.2436691052335398,
"grad_norm": 11.814950010874579,
"learning_rate": 0.00021478565179352578,
"loss": 0.5579,
"step": 4420
},
{
"epoch": 1.2464828362408553,
"grad_norm": 25.70339419099322,
"learning_rate": 0.00021456692913385825,
"loss": 0.6162,
"step": 4430
},
{
"epoch": 1.249296567248171,
"grad_norm": 4.252865920700129,
"learning_rate": 0.00021434820647419072,
"loss": 0.4195,
"step": 4440
},
{
"epoch": 1.2521102982554868,
"grad_norm": 38.698082556525144,
"learning_rate": 0.00021412948381452316,
"loss": 0.5526,
"step": 4450
},
{
"epoch": 1.2549240292628026,
"grad_norm": 7.8381650122365025,
"learning_rate": 0.0002139107611548556,
"loss": 0.5447,
"step": 4460
},
{
"epoch": 1.257737760270118,
"grad_norm": 14.386500677754873,
"learning_rate": 0.00021369203849518808,
"loss": 0.5332,
"step": 4470
},
{
"epoch": 1.2605514912774338,
"grad_norm": 10.393563025135272,
"learning_rate": 0.00021347331583552055,
"loss": 0.3905,
"step": 4480
},
{
"epoch": 1.2633652222847496,
"grad_norm": 11.830727306060455,
"learning_rate": 0.00021325459317585302,
"loss": 0.6539,
"step": 4490
},
{
"epoch": 1.2661789532920653,
"grad_norm": 14.042878553076964,
"learning_rate": 0.00021303587051618546,
"loss": 0.7558,
"step": 4500
},
{
"epoch": 1.268992684299381,
"grad_norm": 9.152885609833971,
"learning_rate": 0.0002128171478565179,
"loss": 0.684,
"step": 4510
},
{
"epoch": 1.2718064153066968,
"grad_norm": 29.21214553934626,
"learning_rate": 0.00021262029746281714,
"loss": 0.6722,
"step": 4520
},
{
"epoch": 1.2746201463140123,
"grad_norm": 31.814467115420808,
"learning_rate": 0.00021240157480314958,
"loss": 0.5875,
"step": 4530
},
{
"epoch": 1.277433877321328,
"grad_norm": 11.976587331588098,
"learning_rate": 0.00021218285214348205,
"loss": 0.5788,
"step": 4540
},
{
"epoch": 1.2802476083286438,
"grad_norm": 9.351382128729613,
"learning_rate": 0.00021196412948381452,
"loss": 0.5404,
"step": 4550
},
{
"epoch": 1.2830613393359596,
"grad_norm": 9.993287787206937,
"learning_rate": 0.000211745406824147,
"loss": 0.574,
"step": 4560
},
{
"epoch": 1.285875070343275,
"grad_norm": 0.9168680830567139,
"learning_rate": 0.0002115266841644794,
"loss": 0.812,
"step": 4570
},
{
"epoch": 1.2886888013505908,
"grad_norm": 6.660390815631498,
"learning_rate": 0.00021130796150481187,
"loss": 0.7979,
"step": 4580
},
{
"epoch": 1.2915025323579066,
"grad_norm": 10.67143801901763,
"learning_rate": 0.00021108923884514434,
"loss": 0.5466,
"step": 4590
},
{
"epoch": 1.2943162633652223,
"grad_norm": 17.62423042213442,
"learning_rate": 0.00021087051618547681,
"loss": 0.431,
"step": 4600
},
{
"epoch": 1.297129994372538,
"grad_norm": 15.617571509727133,
"learning_rate": 0.00021065179352580926,
"loss": 0.5231,
"step": 4610
},
{
"epoch": 1.2999437253798538,
"grad_norm": 58.40350186744155,
"learning_rate": 0.00021043307086614173,
"loss": 0.4857,
"step": 4620
},
{
"epoch": 1.3027574563871693,
"grad_norm": 15.519074842077424,
"learning_rate": 0.00021021434820647417,
"loss": 0.4439,
"step": 4630
},
{
"epoch": 1.305571187394485,
"grad_norm": 23.71709936979936,
"learning_rate": 0.0002099956255468066,
"loss": 0.514,
"step": 4640
},
{
"epoch": 1.3083849184018008,
"grad_norm": 14.117780601189649,
"learning_rate": 0.00020977690288713908,
"loss": 0.6811,
"step": 4650
},
{
"epoch": 1.3111986494091166,
"grad_norm": 31.859641559976787,
"learning_rate": 0.00020955818022747155,
"loss": 0.7653,
"step": 4660
},
{
"epoch": 1.314012380416432,
"grad_norm": 4.62858313326057,
"learning_rate": 0.00020933945756780402,
"loss": 0.56,
"step": 4670
},
{
"epoch": 1.3168261114237478,
"grad_norm": 32.35923134160814,
"learning_rate": 0.00020912073490813647,
"loss": 0.581,
"step": 4680
},
{
"epoch": 1.3196398424310636,
"grad_norm": 11.88084068278056,
"learning_rate": 0.0002089020122484689,
"loss": 0.5339,
"step": 4690
},
{
"epoch": 1.3224535734383793,
"grad_norm": 9.520992713167384,
"learning_rate": 0.00020868328958880138,
"loss": 0.782,
"step": 4700
},
{
"epoch": 1.325267304445695,
"grad_norm": 22.853640876872127,
"learning_rate": 0.00020846456692913385,
"loss": 0.6565,
"step": 4710
},
{
"epoch": 1.3280810354530108,
"grad_norm": 3.8605452401376685,
"learning_rate": 0.0002082458442694663,
"loss": 0.5357,
"step": 4720
},
{
"epoch": 1.3308947664603263,
"grad_norm": 39.18854892428108,
"learning_rate": 0.00020802712160979876,
"loss": 0.6124,
"step": 4730
},
{
"epoch": 1.333708497467642,
"grad_norm": 12.900555694355658,
"learning_rate": 0.00020780839895013123,
"loss": 0.5629,
"step": 4740
},
{
"epoch": 1.3365222284749578,
"grad_norm": 30.260254281976717,
"learning_rate": 0.00020758967629046365,
"loss": 0.6766,
"step": 4750
},
{
"epoch": 1.3393359594822736,
"grad_norm": 5.549604555260689,
"learning_rate": 0.00020737095363079612,
"loss": 0.6215,
"step": 4760
},
{
"epoch": 1.342149690489589,
"grad_norm": 9.606804838953646,
"learning_rate": 0.0002071522309711286,
"loss": 0.5282,
"step": 4770
},
{
"epoch": 1.3449634214969048,
"grad_norm": 21.0629637183242,
"learning_rate": 0.00020693350831146106,
"loss": 0.567,
"step": 4780
},
{
"epoch": 1.3477771525042206,
"grad_norm": 35.449435817589865,
"learning_rate": 0.0002067147856517935,
"loss": 0.5674,
"step": 4790
},
{
"epoch": 1.3505908835115363,
"grad_norm": 11.42100394739649,
"learning_rate": 0.00020649606299212597,
"loss": 0.4903,
"step": 4800
},
{
"epoch": 1.353404614518852,
"grad_norm": 46.973771954330346,
"learning_rate": 0.00020627734033245844,
"loss": 0.6514,
"step": 4810
},
{
"epoch": 1.3562183455261678,
"grad_norm": 37.42810605601175,
"learning_rate": 0.00020605861767279088,
"loss": 0.7795,
"step": 4820
},
{
"epoch": 1.3590320765334833,
"grad_norm": 17.86195496240817,
"learning_rate": 0.00020583989501312333,
"loss": 0.6856,
"step": 4830
},
{
"epoch": 1.361845807540799,
"grad_norm": 4.510420970955073,
"learning_rate": 0.0002056211723534558,
"loss": 0.5515,
"step": 4840
},
{
"epoch": 1.3646595385481148,
"grad_norm": 35.14529979476458,
"learning_rate": 0.00020540244969378827,
"loss": 0.8242,
"step": 4850
},
{
"epoch": 1.3674732695554306,
"grad_norm": 12.97076064718709,
"learning_rate": 0.00020518372703412074,
"loss": 0.6027,
"step": 4860
},
{
"epoch": 1.370287000562746,
"grad_norm": 37.34136036844645,
"learning_rate": 0.00020496500437445318,
"loss": 0.9699,
"step": 4870
},
{
"epoch": 1.3731007315700618,
"grad_norm": 10.648747474625555,
"learning_rate": 0.00020474628171478562,
"loss": 0.5742,
"step": 4880
},
{
"epoch": 1.3759144625773776,
"grad_norm": 38.35409796130347,
"learning_rate": 0.0002045275590551181,
"loss": 0.4493,
"step": 4890
},
{
"epoch": 1.3787281935846933,
"grad_norm": 10.275255197278986,
"learning_rate": 0.00020430883639545053,
"loss": 0.5563,
"step": 4900
},
{
"epoch": 1.381541924592009,
"grad_norm": 25.09323166849902,
"learning_rate": 0.000204090113735783,
"loss": 0.6363,
"step": 4910
},
{
"epoch": 1.3843556555993248,
"grad_norm": 13.731524114353487,
"learning_rate": 0.00020387139107611547,
"loss": 0.8071,
"step": 4920
},
{
"epoch": 1.3871693866066404,
"grad_norm": 10.601584528659094,
"learning_rate": 0.00020365266841644794,
"loss": 0.4743,
"step": 4930
},
{
"epoch": 1.389983117613956,
"grad_norm": 20.873005848848994,
"learning_rate": 0.00020343394575678036,
"loss": 0.5681,
"step": 4940
},
{
"epoch": 1.3927968486212718,
"grad_norm": 6.779807648777697,
"learning_rate": 0.00020321522309711283,
"loss": 0.4404,
"step": 4950
},
{
"epoch": 1.3956105796285876,
"grad_norm": 52.49716946373782,
"learning_rate": 0.0002029965004374453,
"loss": 0.5812,
"step": 4960
},
{
"epoch": 1.3984243106359031,
"grad_norm": 10.370627801920397,
"learning_rate": 0.00020277777777777777,
"loss": 0.6843,
"step": 4970
},
{
"epoch": 1.4012380416432189,
"grad_norm": 22.239749529632242,
"learning_rate": 0.0002025590551181102,
"loss": 0.6977,
"step": 4980
},
{
"epoch": 1.4040517726505346,
"grad_norm": 14.1275030867891,
"learning_rate": 0.00020234033245844268,
"loss": 0.4539,
"step": 4990
},
{
"epoch": 1.4068655036578503,
"grad_norm": 8.99633615310471,
"learning_rate": 0.00020212160979877513,
"loss": 0.6492,
"step": 5000
},
{
"epoch": 1.409679234665166,
"grad_norm": 22.805411368810102,
"learning_rate": 0.0002019028871391076,
"loss": 0.6462,
"step": 5010
},
{
"epoch": 1.4124929656724818,
"grad_norm": 6.381125315859451,
"learning_rate": 0.00020168416447944004,
"loss": 0.6981,
"step": 5020
},
{
"epoch": 1.4153066966797974,
"grad_norm": 7.54030708950237,
"learning_rate": 0.0002014654418197725,
"loss": 0.4331,
"step": 5030
},
{
"epoch": 1.418120427687113,
"grad_norm": 12.137237315522457,
"learning_rate": 0.00020124671916010498,
"loss": 0.5546,
"step": 5040
},
{
"epoch": 1.4209341586944289,
"grad_norm": 59.47494361525208,
"learning_rate": 0.00020102799650043745,
"loss": 0.5772,
"step": 5050
},
{
"epoch": 1.4237478897017446,
"grad_norm": 31.495786286643714,
"learning_rate": 0.0002008092738407699,
"loss": 0.6932,
"step": 5060
},
{
"epoch": 1.4265616207090601,
"grad_norm": 18.486318084665708,
"learning_rate": 0.00020059055118110234,
"loss": 0.4995,
"step": 5070
},
{
"epoch": 1.4293753517163759,
"grad_norm": 10.238292416097469,
"learning_rate": 0.0002003718285214348,
"loss": 0.6652,
"step": 5080
},
{
"epoch": 1.4321890827236916,
"grad_norm": 4.579110553754593,
"learning_rate": 0.00020015310586176725,
"loss": 0.5336,
"step": 5090
},
{
"epoch": 1.4350028137310074,
"grad_norm": 34.098434311756876,
"learning_rate": 0.00019993438320209972,
"loss": 0.765,
"step": 5100
},
{
"epoch": 1.437816544738323,
"grad_norm": 35.700128715881476,
"learning_rate": 0.0001997156605424322,
"loss": 0.727,
"step": 5110
},
{
"epoch": 1.4406302757456388,
"grad_norm": 28.138298504559835,
"learning_rate": 0.00019949693788276466,
"loss": 0.695,
"step": 5120
},
{
"epoch": 1.4434440067529544,
"grad_norm": 23.026654117472113,
"learning_rate": 0.00019927821522309707,
"loss": 0.5303,
"step": 5130
},
{
"epoch": 1.4462577377602701,
"grad_norm": 9.046788588294012,
"learning_rate": 0.00019905949256342954,
"loss": 0.4019,
"step": 5140
},
{
"epoch": 1.4490714687675859,
"grad_norm": 21.468136979947417,
"learning_rate": 0.00019884076990376201,
"loss": 0.2962,
"step": 5150
},
{
"epoch": 1.4518851997749016,
"grad_norm": 26.34398373401709,
"learning_rate": 0.00019862204724409448,
"loss": 1.0659,
"step": 5160
},
{
"epoch": 1.4546989307822171,
"grad_norm": 17.669274446566238,
"learning_rate": 0.00019840332458442693,
"loss": 0.5564,
"step": 5170
},
{
"epoch": 1.4575126617895329,
"grad_norm": 3.3651916727576987,
"learning_rate": 0.0001981846019247594,
"loss": 0.4807,
"step": 5180
},
{
"epoch": 1.4603263927968486,
"grad_norm": 11.603663088020909,
"learning_rate": 0.00019796587926509184,
"loss": 0.5503,
"step": 5190
},
{
"epoch": 1.4631401238041644,
"grad_norm": 23.63460879726596,
"learning_rate": 0.00019774715660542428,
"loss": 0.7177,
"step": 5200
},
{
"epoch": 1.46595385481148,
"grad_norm": 2.9074863920622134,
"learning_rate": 0.00019752843394575675,
"loss": 0.3413,
"step": 5210
},
{
"epoch": 1.4687675858187959,
"grad_norm": 31.091134376352294,
"learning_rate": 0.00019730971128608922,
"loss": 0.7108,
"step": 5220
},
{
"epoch": 1.4715813168261114,
"grad_norm": 22.73781393986795,
"learning_rate": 0.0001970909886264217,
"loss": 0.7106,
"step": 5230
},
{
"epoch": 1.4743950478334271,
"grad_norm": 10.733654149323753,
"learning_rate": 0.00019687226596675416,
"loss": 0.3555,
"step": 5240
},
{
"epoch": 1.4772087788407429,
"grad_norm": 43.418587910591356,
"learning_rate": 0.00019665354330708658,
"loss": 0.6709,
"step": 5250
},
{
"epoch": 1.4800225098480584,
"grad_norm": 10.924870366111936,
"learning_rate": 0.00019643482064741905,
"loss": 0.5829,
"step": 5260
},
{
"epoch": 1.4828362408553741,
"grad_norm": 4.141398446252563,
"learning_rate": 0.00019621609798775152,
"loss": 0.4006,
"step": 5270
},
{
"epoch": 1.4856499718626899,
"grad_norm": 25.642802616147556,
"learning_rate": 0.00019599737532808396,
"loss": 0.8251,
"step": 5280
},
{
"epoch": 1.4884637028700056,
"grad_norm": 27.534126408595263,
"learning_rate": 0.00019577865266841643,
"loss": 1.0336,
"step": 5290
},
{
"epoch": 1.4912774338773214,
"grad_norm": 6.160307363496283,
"learning_rate": 0.0001955599300087489,
"loss": 0.7243,
"step": 5300
},
{
"epoch": 1.4940911648846371,
"grad_norm": 13.152914687437683,
"learning_rate": 0.00019534120734908137,
"loss": 0.6427,
"step": 5310
},
{
"epoch": 1.4969048958919529,
"grad_norm": 9.301055295352276,
"learning_rate": 0.0001951224846894138,
"loss": 0.6152,
"step": 5320
},
{
"epoch": 1.4997186268992684,
"grad_norm": 6.722778731476633,
"learning_rate": 0.00019490376202974626,
"loss": 0.5331,
"step": 5330
},
{
"epoch": 1.5025323579065841,
"grad_norm": 8.186600974279003,
"learning_rate": 0.00019468503937007873,
"loss": 0.4851,
"step": 5340
},
{
"epoch": 1.5053460889138999,
"grad_norm": 12.233978539104966,
"learning_rate": 0.0001944663167104112,
"loss": 0.5694,
"step": 5350
},
{
"epoch": 1.5081598199212154,
"grad_norm": 35.97067266871921,
"learning_rate": 0.00019424759405074364,
"loss": 0.7976,
"step": 5360
},
{
"epoch": 1.5109735509285311,
"grad_norm": 16.314895195522084,
"learning_rate": 0.0001940288713910761,
"loss": 0.5439,
"step": 5370
},
{
"epoch": 1.5137872819358469,
"grad_norm": 16.9947029531932,
"learning_rate": 0.00019381014873140855,
"loss": 0.4797,
"step": 5380
},
{
"epoch": 1.5166010129431626,
"grad_norm": 22.886764769826087,
"learning_rate": 0.000193591426071741,
"loss": 0.5191,
"step": 5390
},
{
"epoch": 1.5194147439504784,
"grad_norm": 5.870017090348409,
"learning_rate": 0.00019337270341207347,
"loss": 0.5474,
"step": 5400
},
{
"epoch": 1.5222284749577941,
"grad_norm": 28.06407341443698,
"learning_rate": 0.00019315398075240594,
"loss": 0.69,
"step": 5410
},
{
"epoch": 1.5250422059651099,
"grad_norm": 14.781385104588194,
"learning_rate": 0.0001929352580927384,
"loss": 0.5645,
"step": 5420
},
{
"epoch": 1.5278559369724254,
"grad_norm": 5.855766754825115,
"learning_rate": 0.00019271653543307085,
"loss": 0.4795,
"step": 5430
},
{
"epoch": 1.5306696679797411,
"grad_norm": 22.918135069417044,
"learning_rate": 0.0001924978127734033,
"loss": 0.5786,
"step": 5440
},
{
"epoch": 1.5334833989870569,
"grad_norm": 17.207456518107474,
"learning_rate": 0.00019227909011373576,
"loss": 0.651,
"step": 5450
},
{
"epoch": 1.5362971299943724,
"grad_norm": 5.184301427212219,
"learning_rate": 0.00019206036745406823,
"loss": 0.415,
"step": 5460
},
{
"epoch": 1.5391108610016881,
"grad_norm": 16.094276621452206,
"learning_rate": 0.00019184164479440067,
"loss": 0.5126,
"step": 5470
},
{
"epoch": 1.541924592009004,
"grad_norm": 27.587401306674103,
"learning_rate": 0.00019162292213473314,
"loss": 0.5171,
"step": 5480
},
{
"epoch": 1.5447383230163196,
"grad_norm": 36.812039328705936,
"learning_rate": 0.00019140419947506561,
"loss": 0.573,
"step": 5490
},
{
"epoch": 1.5475520540236354,
"grad_norm": 18.270164053028488,
"learning_rate": 0.00019118547681539803,
"loss": 0.4896,
"step": 5500
},
{
"epoch": 1.5503657850309511,
"grad_norm": 21.200268676017966,
"learning_rate": 0.0001909667541557305,
"loss": 0.5086,
"step": 5510
},
{
"epoch": 1.5531795160382669,
"grad_norm": 2.634500985821002,
"learning_rate": 0.00019074803149606297,
"loss": 0.5153,
"step": 5520
},
{
"epoch": 1.5559932470455824,
"grad_norm": 31.303796019116458,
"learning_rate": 0.00019052930883639544,
"loss": 0.5879,
"step": 5530
},
{
"epoch": 1.5588069780528981,
"grad_norm": 13.767169050681202,
"learning_rate": 0.0001903105861767279,
"loss": 0.6354,
"step": 5540
},
{
"epoch": 1.5616207090602139,
"grad_norm": 20.81439861452622,
"learning_rate": 0.00019009186351706035,
"loss": 0.5908,
"step": 5550
},
{
"epoch": 1.5644344400675294,
"grad_norm": 24.25248324077591,
"learning_rate": 0.00018987314085739282,
"loss": 0.568,
"step": 5560
},
{
"epoch": 1.5672481710748452,
"grad_norm": 20.3474642773289,
"learning_rate": 0.000189676290463692,
"loss": 0.6771,
"step": 5570
},
{
"epoch": 1.570061902082161,
"grad_norm": 27.420648385460662,
"learning_rate": 0.00018945756780402447,
"loss": 0.7918,
"step": 5580
},
{
"epoch": 1.5728756330894766,
"grad_norm": 12.762719969507081,
"learning_rate": 0.00018923884514435694,
"loss": 0.6914,
"step": 5590
},
{
"epoch": 1.5756893640967924,
"grad_norm": 12.90335156682279,
"learning_rate": 0.0001890201224846894,
"loss": 0.6175,
"step": 5600
},
{
"epoch": 1.5785030951041081,
"grad_norm": 21.08111096971608,
"learning_rate": 0.00018880139982502188,
"loss": 0.4938,
"step": 5610
},
{
"epoch": 1.5813168261114239,
"grad_norm": 10.112944768654387,
"learning_rate": 0.0001885826771653543,
"loss": 0.5957,
"step": 5620
},
{
"epoch": 1.5841305571187394,
"grad_norm": 11.817780491387401,
"learning_rate": 0.00018836395450568677,
"loss": 0.616,
"step": 5630
},
{
"epoch": 1.5869442881260551,
"grad_norm": 13.413123522838792,
"learning_rate": 0.00018814523184601924,
"loss": 0.5436,
"step": 5640
},
{
"epoch": 1.589758019133371,
"grad_norm": 8.5007708126369,
"learning_rate": 0.00018792650918635168,
"loss": 0.5707,
"step": 5650
},
{
"epoch": 1.5925717501406864,
"grad_norm": 6.355914825325111,
"learning_rate": 0.00018770778652668415,
"loss": 0.5345,
"step": 5660
},
{
"epoch": 1.5953854811480022,
"grad_norm": 18.871338179625443,
"learning_rate": 0.00018748906386701662,
"loss": 0.5581,
"step": 5670
},
{
"epoch": 1.598199212155318,
"grad_norm": 22.113665794555953,
"learning_rate": 0.00018727034120734904,
"loss": 0.3793,
"step": 5680
},
{
"epoch": 1.6010129431626337,
"grad_norm": 18.640229312889087,
"learning_rate": 0.0001870516185476815,
"loss": 0.6424,
"step": 5690
},
{
"epoch": 1.6038266741699494,
"grad_norm": 9.013049101470614,
"learning_rate": 0.00018683289588801398,
"loss": 0.6761,
"step": 5700
},
{
"epoch": 1.6066404051772651,
"grad_norm": 24.66632774615283,
"learning_rate": 0.00018661417322834645,
"loss": 0.4636,
"step": 5710
},
{
"epoch": 1.6094541361845809,
"grad_norm": 21.359311361155275,
"learning_rate": 0.00018639545056867892,
"loss": 0.3643,
"step": 5720
},
{
"epoch": 1.6122678671918964,
"grad_norm": 17.552845045440993,
"learning_rate": 0.00018617672790901136,
"loss": 0.7695,
"step": 5730
},
{
"epoch": 1.6150815981992122,
"grad_norm": 14.750302688475113,
"learning_rate": 0.0001859580052493438,
"loss": 0.7166,
"step": 5740
},
{
"epoch": 1.617895329206528,
"grad_norm": 14.017729808491103,
"learning_rate": 0.00018573928258967627,
"loss": 0.5689,
"step": 5750
},
{
"epoch": 1.6207090602138434,
"grad_norm": 27.853805455956927,
"learning_rate": 0.00018552055993000872,
"loss": 0.6315,
"step": 5760
},
{
"epoch": 1.6235227912211592,
"grad_norm": 16.4717416832815,
"learning_rate": 0.0001853018372703412,
"loss": 0.4896,
"step": 5770
},
{
"epoch": 1.626336522228475,
"grad_norm": 11.48773947806387,
"learning_rate": 0.00018508311461067366,
"loss": 0.5196,
"step": 5780
},
{
"epoch": 1.6291502532357907,
"grad_norm": 55.37757053824189,
"learning_rate": 0.00018486439195100613,
"loss": 0.6459,
"step": 5790
},
{
"epoch": 1.6319639842431064,
"grad_norm": 22.81165151193899,
"learning_rate": 0.00018464566929133857,
"loss": 0.5926,
"step": 5800
},
{
"epoch": 1.6347777152504221,
"grad_norm": 19.036953260995485,
"learning_rate": 0.000184426946631671,
"loss": 0.9008,
"step": 5810
},
{
"epoch": 1.637591446257738,
"grad_norm": 25.57500606412806,
"learning_rate": 0.00018420822397200348,
"loss": 0.5999,
"step": 5820
},
{
"epoch": 1.6404051772650534,
"grad_norm": 8.891606826403597,
"learning_rate": 0.00018398950131233595,
"loss": 0.5492,
"step": 5830
},
{
"epoch": 1.6432189082723692,
"grad_norm": 7.0897653575377975,
"learning_rate": 0.0001837707786526684,
"loss": 0.4375,
"step": 5840
},
{
"epoch": 1.646032639279685,
"grad_norm": 16.82282152611567,
"learning_rate": 0.00018355205599300087,
"loss": 0.6416,
"step": 5850
},
{
"epoch": 1.6488463702870004,
"grad_norm": 26.076012233816623,
"learning_rate": 0.00018333333333333334,
"loss": 0.7995,
"step": 5860
},
{
"epoch": 1.6516601012943162,
"grad_norm": 6.103373372823494,
"learning_rate": 0.00018311461067366575,
"loss": 0.513,
"step": 5870
},
{
"epoch": 1.654473832301632,
"grad_norm": 7.46141704246519,
"learning_rate": 0.00018289588801399822,
"loss": 0.442,
"step": 5880
},
{
"epoch": 1.6572875633089477,
"grad_norm": 21.657859712145655,
"learning_rate": 0.0001826771653543307,
"loss": 0.6058,
"step": 5890
},
{
"epoch": 1.6601012943162634,
"grad_norm": 23.56206415921756,
"learning_rate": 0.00018245844269466316,
"loss": 0.609,
"step": 5900
},
{
"epoch": 1.6629150253235792,
"grad_norm": 11.96355285804545,
"learning_rate": 0.00018223972003499563,
"loss": 0.4169,
"step": 5910
},
{
"epoch": 1.665728756330895,
"grad_norm": 15.80001057748199,
"learning_rate": 0.00018202099737532807,
"loss": 0.7119,
"step": 5920
},
{
"epoch": 1.6685424873382104,
"grad_norm": 24.01734519933029,
"learning_rate": 0.00018180227471566052,
"loss": 0.6546,
"step": 5930
},
{
"epoch": 1.6713562183455262,
"grad_norm": 12.082586359258165,
"learning_rate": 0.000181583552055993,
"loss": 0.7743,
"step": 5940
},
{
"epoch": 1.674169949352842,
"grad_norm": 16.8076808139855,
"learning_rate": 0.00018136482939632543,
"loss": 0.5819,
"step": 5950
},
{
"epoch": 1.6769836803601574,
"grad_norm": 16.864221341224397,
"learning_rate": 0.0001811461067366579,
"loss": 0.6483,
"step": 5960
},
{
"epoch": 1.6797974113674732,
"grad_norm": 11.102468101320996,
"learning_rate": 0.00018092738407699037,
"loss": 0.5152,
"step": 5970
},
{
"epoch": 1.682611142374789,
"grad_norm": 17.6010512401763,
"learning_rate": 0.00018070866141732284,
"loss": 0.5134,
"step": 5980
},
{
"epoch": 1.6854248733821047,
"grad_norm": 8.25091098683039,
"learning_rate": 0.00018048993875765526,
"loss": 0.49,
"step": 5990
},
{
"epoch": 1.6882386043894204,
"grad_norm": 7.5344372075509884,
"learning_rate": 0.00018027121609798773,
"loss": 0.4619,
"step": 6000
},
{
"epoch": 1.6910523353967362,
"grad_norm": 23.21545471999833,
"learning_rate": 0.0001800524934383202,
"loss": 0.8264,
"step": 6010
},
{
"epoch": 1.693866066404052,
"grad_norm": 15.393641748407818,
"learning_rate": 0.00017983377077865267,
"loss": 0.6024,
"step": 6020
},
{
"epoch": 1.6966797974113674,
"grad_norm": 12.417067525367335,
"learning_rate": 0.0001796150481189851,
"loss": 0.6584,
"step": 6030
},
{
"epoch": 1.6994935284186832,
"grad_norm": 15.042896501382003,
"learning_rate": 0.00017939632545931758,
"loss": 0.4492,
"step": 6040
},
{
"epoch": 1.702307259425999,
"grad_norm": 9.115061298735506,
"learning_rate": 0.00017917760279965005,
"loss": 0.4221,
"step": 6050
},
{
"epoch": 1.7051209904333144,
"grad_norm": 0.6607374478724659,
"learning_rate": 0.00017895888013998246,
"loss": 0.434,
"step": 6060
},
{
"epoch": 1.7079347214406302,
"grad_norm": 33.803698820392704,
"learning_rate": 0.00017878390201224846,
"loss": 1.9177,
"step": 6070
},
{
"epoch": 1.710748452447946,
"grad_norm": 361.56918934904206,
"learning_rate": 0.00017856517935258093,
"loss": 2.013,
"step": 6080
},
{
"epoch": 1.7135621834552617,
"grad_norm": 44.98806827034684,
"learning_rate": 0.00017834645669291338,
"loss": 1.6814,
"step": 6090
},
{
"epoch": 1.7163759144625774,
"grad_norm": 22.283635215772854,
"learning_rate": 0.00017812773403324582,
"loss": 0.795,
"step": 6100
},
{
"epoch": 1.7191896454698932,
"grad_norm": 105.85751003748128,
"learning_rate": 0.0001779090113735783,
"loss": 2.4138,
"step": 6110
},
{
"epoch": 1.722003376477209,
"grad_norm": 85.01552368225332,
"learning_rate": 0.00017769028871391076,
"loss": 3.8929,
"step": 6120
},
{
"epoch": 1.7248171074845244,
"grad_norm": 44.282971732965535,
"learning_rate": 0.0001774715660542432,
"loss": 0.7031,
"step": 6130
},
{
"epoch": 1.7276308384918402,
"grad_norm": 298.69778969364853,
"learning_rate": 0.00017725284339457567,
"loss": 0.5441,
"step": 6140
},
{
"epoch": 1.730444569499156,
"grad_norm": 41.66809813265777,
"learning_rate": 0.00017703412073490814,
"loss": 0.7863,
"step": 6150
},
{
"epoch": 1.7332583005064714,
"grad_norm": 34.03455804697322,
"learning_rate": 0.00017681539807524056,
"loss": 0.9071,
"step": 6160
},
{
"epoch": 1.7360720315137872,
"grad_norm": 137.98952284030946,
"learning_rate": 0.00017659667541557303,
"loss": 1.5913,
"step": 6170
},
{
"epoch": 1.738885762521103,
"grad_norm": 54.35499220435977,
"learning_rate": 0.0001763779527559055,
"loss": 1.4096,
"step": 6180
},
{
"epoch": 1.7416994935284187,
"grad_norm": 72.22077387735027,
"learning_rate": 0.00017615923009623797,
"loss": 0.6111,
"step": 6190
},
{
"epoch": 1.7445132245357344,
"grad_norm": 11.271321807307686,
"learning_rate": 0.0001759405074365704,
"loss": 0.8519,
"step": 6200
},
{
"epoch": 1.7473269555430502,
"grad_norm": 50.02675742399026,
"learning_rate": 0.00017572178477690288,
"loss": 0.9388,
"step": 6210
},
{
"epoch": 1.750140686550366,
"grad_norm": 30.543850975892273,
"learning_rate": 0.00017550306211723532,
"loss": 0.6496,
"step": 6220
},
{
"epoch": 1.7529544175576814,
"grad_norm": 17.096512987881336,
"learning_rate": 0.0001752843394575678,
"loss": 0.8429,
"step": 6230
},
{
"epoch": 1.7557681485649972,
"grad_norm": 10.875495203297126,
"learning_rate": 0.00017506561679790024,
"loss": 0.424,
"step": 6240
},
{
"epoch": 1.758581879572313,
"grad_norm": 28.472277481379553,
"learning_rate": 0.0001748468941382327,
"loss": 0.5346,
"step": 6250
},
{
"epoch": 1.7613956105796285,
"grad_norm": 84.611446382734,
"learning_rate": 0.00017462817147856518,
"loss": 0.7309,
"step": 6260
},
{
"epoch": 1.7642093415869442,
"grad_norm": 83.93394745603818,
"learning_rate": 0.00017440944881889765,
"loss": 0.7314,
"step": 6270
},
{
"epoch": 1.76702307259426,
"grad_norm": 29.72369442712257,
"learning_rate": 0.00017419072615923006,
"loss": 0.4404,
"step": 6280
},
{
"epoch": 1.7698368036015757,
"grad_norm": 33.47447474767712,
"learning_rate": 0.00017397200349956253,
"loss": 0.5264,
"step": 6290
},
{
"epoch": 1.7726505346088914,
"grad_norm": 112.25470154565467,
"learning_rate": 0.000173753280839895,
"loss": 0.5341,
"step": 6300
},
{
"epoch": 1.7754642656162072,
"grad_norm": 5.004631103885064,
"learning_rate": 0.00017353455818022744,
"loss": 0.7944,
"step": 6310
},
{
"epoch": 1.778277996623523,
"grad_norm": 36.206284557597996,
"learning_rate": 0.00017331583552055991,
"loss": 0.6088,
"step": 6320
},
{
"epoch": 1.7810917276308385,
"grad_norm": 114.83303534732538,
"learning_rate": 0.00017309711286089238,
"loss": 0.8535,
"step": 6330
},
{
"epoch": 1.7839054586381542,
"grad_norm": 39.25126961762459,
"learning_rate": 0.00017287839020122485,
"loss": 0.4341,
"step": 6340
},
{
"epoch": 1.78671918964547,
"grad_norm": 38.887489483647045,
"learning_rate": 0.00017265966754155727,
"loss": 0.6262,
"step": 6350
},
{
"epoch": 1.7895329206527855,
"grad_norm": 14.662335403344557,
"learning_rate": 0.00017244094488188974,
"loss": 0.7171,
"step": 6360
},
{
"epoch": 1.7923466516601012,
"grad_norm": 12.888841929949086,
"learning_rate": 0.0001722222222222222,
"loss": 0.5094,
"step": 6370
},
{
"epoch": 1.795160382667417,
"grad_norm": 22.26070054592782,
"learning_rate": 0.00017200349956255468,
"loss": 0.4261,
"step": 6380
},
{
"epoch": 1.7979741136747327,
"grad_norm": 23.038642054175508,
"learning_rate": 0.00017178477690288712,
"loss": 0.4987,
"step": 6390
},
{
"epoch": 1.8007878446820484,
"grad_norm": 9.474105949765265,
"learning_rate": 0.0001715660542432196,
"loss": 0.5878,
"step": 6400
},
{
"epoch": 1.8036015756893642,
"grad_norm": 70.27189577371828,
"learning_rate": 0.00017134733158355204,
"loss": 0.4774,
"step": 6410
},
{
"epoch": 1.80641530669668,
"grad_norm": 26.61930756765317,
"learning_rate": 0.00017112860892388448,
"loss": 0.598,
"step": 6420
},
{
"epoch": 1.8092290377039955,
"grad_norm": 12.533144473520764,
"learning_rate": 0.00017090988626421695,
"loss": 0.6024,
"step": 6430
},
{
"epoch": 1.8120427687113112,
"grad_norm": 48.14804877192819,
"learning_rate": 0.00017069116360454942,
"loss": 0.9674,
"step": 6440
},
{
"epoch": 1.814856499718627,
"grad_norm": 15.22684827666546,
"learning_rate": 0.0001704724409448819,
"loss": 0.7041,
"step": 6450
},
{
"epoch": 1.8176702307259425,
"grad_norm": 43.53579267992454,
"learning_rate": 0.00017025371828521436,
"loss": 0.64,
"step": 6460
},
{
"epoch": 1.8204839617332582,
"grad_norm": 41.19355041508803,
"learning_rate": 0.00017003499562554677,
"loss": 0.4662,
"step": 6470
},
{
"epoch": 1.823297692740574,
"grad_norm": 44.036889353364195,
"learning_rate": 0.00016981627296587924,
"loss": 0.58,
"step": 6480
},
{
"epoch": 1.8261114237478897,
"grad_norm": 3.448573380443346,
"learning_rate": 0.00016959755030621171,
"loss": 0.8725,
"step": 6490
},
{
"epoch": 1.8289251547552055,
"grad_norm": 27.321094827902026,
"learning_rate": 0.00016937882764654416,
"loss": 0.5894,
"step": 6500
},
{
"epoch": 1.8317388857625212,
"grad_norm": 11.550339390724506,
"learning_rate": 0.00016916010498687663,
"loss": 1.0566,
"step": 6510
},
{
"epoch": 1.834552616769837,
"grad_norm": 29.635894647284605,
"learning_rate": 0.0001689413823272091,
"loss": 0.5362,
"step": 6520
},
{
"epoch": 1.8373663477771525,
"grad_norm": 28.87624464201189,
"learning_rate": 0.0001687226596675415,
"loss": 0.4639,
"step": 6530
},
{
"epoch": 1.8401800787844682,
"grad_norm": 20.63490125951859,
"learning_rate": 0.00016850393700787398,
"loss": 0.5236,
"step": 6540
},
{
"epoch": 1.842993809791784,
"grad_norm": 24.50339308909374,
"learning_rate": 0.00016828521434820645,
"loss": 0.6794,
"step": 6550
},
{
"epoch": 1.8458075407990995,
"grad_norm": 22.43711891156182,
"learning_rate": 0.00016806649168853892,
"loss": 0.6718,
"step": 6560
},
{
"epoch": 1.8486212718064152,
"grad_norm": 12.312381142318516,
"learning_rate": 0.0001678477690288714,
"loss": 0.546,
"step": 6570
},
{
"epoch": 1.851435002813731,
"grad_norm": 52.91882297397753,
"learning_rate": 0.00016762904636920384,
"loss": 0.6186,
"step": 6580
},
{
"epoch": 1.8542487338210467,
"grad_norm": 16.248616548482175,
"learning_rate": 0.0001674103237095363,
"loss": 0.4443,
"step": 6590
},
{
"epoch": 1.8570624648283625,
"grad_norm": 46.586186471184554,
"learning_rate": 0.00016719160104986875,
"loss": 0.6858,
"step": 6600
},
{
"epoch": 1.8598761958356782,
"grad_norm": 19.395899136066642,
"learning_rate": 0.0001669728783902012,
"loss": 0.7021,
"step": 6610
},
{
"epoch": 1.862689926842994,
"grad_norm": 18.858764154991857,
"learning_rate": 0.00016675415573053366,
"loss": 0.5572,
"step": 6620
},
{
"epoch": 1.8655036578503095,
"grad_norm": 34.85053822034739,
"learning_rate": 0.00016653543307086613,
"loss": 0.5491,
"step": 6630
},
{
"epoch": 1.8683173888576252,
"grad_norm": 153.58844319020815,
"learning_rate": 0.0001663167104111986,
"loss": 1.0728,
"step": 6640
},
{
"epoch": 1.871131119864941,
"grad_norm": 35.47908415964911,
"learning_rate": 0.00016609798775153105,
"loss": 0.7507,
"step": 6650
},
{
"epoch": 1.8739448508722565,
"grad_norm": 25.27011106317989,
"learning_rate": 0.0001658792650918635,
"loss": 0.4367,
"step": 6660
},
{
"epoch": 1.8767585818795722,
"grad_norm": 44.64115963656757,
"learning_rate": 0.00016566054243219596,
"loss": 0.4281,
"step": 6670
},
{
"epoch": 1.879572312886888,
"grad_norm": 12.745753520758505,
"learning_rate": 0.00016544181977252843,
"loss": 0.459,
"step": 6680
},
{
"epoch": 1.8823860438942037,
"grad_norm": 44.33709000085202,
"learning_rate": 0.00016522309711286087,
"loss": 0.6107,
"step": 6690
},
{
"epoch": 1.8851997749015195,
"grad_norm": 133.11619488605578,
"learning_rate": 0.00016500437445319334,
"loss": 0.7659,
"step": 6700
},
{
"epoch": 1.8880135059088352,
"grad_norm": 31.166131234712104,
"learning_rate": 0.0001647856517935258,
"loss": 0.2729,
"step": 6710
},
{
"epoch": 1.890827236916151,
"grad_norm": 254.4189742635797,
"learning_rate": 0.00016456692913385823,
"loss": 0.8195,
"step": 6720
},
{
"epoch": 1.8936409679234665,
"grad_norm": 39.51199032081583,
"learning_rate": 0.0001643482064741907,
"loss": 0.6009,
"step": 6730
},
{
"epoch": 1.8964546989307822,
"grad_norm": 39.900138315281346,
"learning_rate": 0.00016412948381452317,
"loss": 0.5433,
"step": 6740
},
{
"epoch": 1.899268429938098,
"grad_norm": 29.49439914115921,
"learning_rate": 0.00016391076115485564,
"loss": 0.5698,
"step": 6750
},
{
"epoch": 1.9020821609454135,
"grad_norm": 12.369579350171918,
"learning_rate": 0.0001636920384951881,
"loss": 0.5124,
"step": 6760
},
{
"epoch": 1.9048958919527292,
"grad_norm": 35.497941038034284,
"learning_rate": 0.00016347331583552055,
"loss": 0.5609,
"step": 6770
},
{
"epoch": 1.907709622960045,
"grad_norm": 19.91481826563306,
"learning_rate": 0.000163254593175853,
"loss": 0.5812,
"step": 6780
},
{
"epoch": 1.9105233539673607,
"grad_norm": 14.286913944349674,
"learning_rate": 0.00016303587051618546,
"loss": 0.5505,
"step": 6790
},
{
"epoch": 1.9133370849746765,
"grad_norm": 45.55623816014948,
"learning_rate": 0.0001628171478565179,
"loss": 0.5946,
"step": 6800
},
{
"epoch": 1.9161508159819922,
"grad_norm": 37.12930100826606,
"learning_rate": 0.00016259842519685038,
"loss": 0.6553,
"step": 6810
},
{
"epoch": 1.918964546989308,
"grad_norm": 5.951242571048711,
"learning_rate": 0.00016237970253718285,
"loss": 0.4606,
"step": 6820
},
{
"epoch": 1.9217782779966235,
"grad_norm": 15.667646539342769,
"learning_rate": 0.00016216097987751532,
"loss": 0.5596,
"step": 6830
},
{
"epoch": 1.9245920090039392,
"grad_norm": 31.888221180998954,
"learning_rate": 0.00016194225721784776,
"loss": 0.665,
"step": 6840
},
{
"epoch": 1.927405740011255,
"grad_norm": 20.015849211223937,
"learning_rate": 0.0001617235345581802,
"loss": 0.4471,
"step": 6850
},
{
"epoch": 1.9302194710185705,
"grad_norm": 19.065880675790694,
"learning_rate": 0.00016150481189851267,
"loss": 0.5152,
"step": 6860
},
{
"epoch": 1.9330332020258862,
"grad_norm": 97.42577987957829,
"learning_rate": 0.00016128608923884514,
"loss": 0.719,
"step": 6870
},
{
"epoch": 1.935846933033202,
"grad_norm": 21.461313685288744,
"learning_rate": 0.00016106736657917758,
"loss": 0.4803,
"step": 6880
},
{
"epoch": 1.9386606640405177,
"grad_norm": 31.746099523443103,
"learning_rate": 0.00016084864391951005,
"loss": 0.672,
"step": 6890
},
{
"epoch": 1.9414743950478335,
"grad_norm": 11.063287518374715,
"learning_rate": 0.00016062992125984252,
"loss": 0.6651,
"step": 6900
},
{
"epoch": 1.9442881260551492,
"grad_norm": 1.3011595771705704,
"learning_rate": 0.00016041119860017494,
"loss": 0.3245,
"step": 6910
},
{
"epoch": 1.947101857062465,
"grad_norm": 11.00869364626475,
"learning_rate": 0.0001601924759405074,
"loss": 0.6375,
"step": 6920
},
{
"epoch": 1.9499155880697805,
"grad_norm": 95.25621911518874,
"learning_rate": 0.00015997375328083988,
"loss": 0.6515,
"step": 6930
},
{
"epoch": 1.9527293190770962,
"grad_norm": 980.764196522124,
"learning_rate": 0.00015975503062117235,
"loss": 0.7356,
"step": 6940
},
{
"epoch": 1.955543050084412,
"grad_norm": 37.87200039766416,
"learning_rate": 0.0001595363079615048,
"loss": 0.5817,
"step": 6950
},
{
"epoch": 1.9583567810917275,
"grad_norm": 17.35128744114319,
"learning_rate": 0.00015931758530183726,
"loss": 0.7061,
"step": 6960
},
{
"epoch": 1.9611705120990433,
"grad_norm": 74.26952030970506,
"learning_rate": 0.0001590988626421697,
"loss": 0.6526,
"step": 6970
},
{
"epoch": 1.963984243106359,
"grad_norm": 67.6583202864629,
"learning_rate": 0.00015888013998250218,
"loss": 0.5742,
"step": 6980
},
{
"epoch": 1.9667979741136747,
"grad_norm": 78.23948101480053,
"learning_rate": 0.00015866141732283462,
"loss": 0.5176,
"step": 6990
},
{
"epoch": 1.9696117051209905,
"grad_norm": 20.77366817098103,
"learning_rate": 0.0001584426946631671,
"loss": 0.6506,
"step": 7000
},
{
"epoch": 1.9724254361283062,
"grad_norm": 48.56847115116187,
"learning_rate": 0.00015822397200349956,
"loss": 0.6096,
"step": 7010
},
{
"epoch": 1.975239167135622,
"grad_norm": 162.60631212883658,
"learning_rate": 0.00015800524934383203,
"loss": 0.8333,
"step": 7020
},
{
"epoch": 1.9780528981429375,
"grad_norm": 89.91859486527336,
"learning_rate": 0.00015778652668416444,
"loss": 0.7861,
"step": 7030
},
{
"epoch": 1.9808666291502532,
"grad_norm": 25.62581876051493,
"learning_rate": 0.00015756780402449691,
"loss": 0.7207,
"step": 7040
},
{
"epoch": 1.983680360157569,
"grad_norm": 49.05293632646501,
"learning_rate": 0.00015734908136482938,
"loss": 0.6649,
"step": 7050
},
{
"epoch": 1.9864940911648845,
"grad_norm": 29.01417189468632,
"learning_rate": 0.00015713035870516183,
"loss": 0.7179,
"step": 7060
},
{
"epoch": 1.9893078221722003,
"grad_norm": 39.04202893773362,
"learning_rate": 0.0001569116360454943,
"loss": 0.4848,
"step": 7070
},
{
"epoch": 1.992121553179516,
"grad_norm": 31.47398318556384,
"learning_rate": 0.00015669291338582677,
"loss": 0.4892,
"step": 7080
},
{
"epoch": 1.9949352841868317,
"grad_norm": 12.299132356823057,
"learning_rate": 0.00015647419072615924,
"loss": 0.501,
"step": 7090
},
{
"epoch": 1.9977490151941475,
"grad_norm": 7.597881461387406,
"learning_rate": 0.00015625546806649165,
"loss": 0.4163,
"step": 7100
},
{
"epoch": 2.0,
"eval_0_f1": 0.5831303288672351,
"eval_0_precision": 0.46033653846153844,
"eval_0_recall": 0.795265780730897,
"eval_1_f1": 0.7661028532376558,
"eval_1_precision": 0.9009443439823187,
"eval_1_recall": 0.6663694456828652,
"eval_accuracy": 0.7003392798511546,
"eval_loss": 0.6884765625,
"eval_runtime": 468.4626,
"eval_samples_per_second": 19.504,
"eval_steps_per_second": 3.251,
"step": 7108
},
{
"epoch": 2.0005627462014632,
"grad_norm": 14.396603604626321,
"learning_rate": 0.00015603674540682412,
"loss": 0.5533,
"step": 7110
},
{
"epoch": 2.003376477208779,
"grad_norm": 2.8030111262625117,
"learning_rate": 0.0001558180227471566,
"loss": 0.7694,
"step": 7120
},
{
"epoch": 2.0061902082160947,
"grad_norm": 1.4709503361365215,
"learning_rate": 0.00015559930008748906,
"loss": 0.654,
"step": 7130
},
{
"epoch": 2.00900393922341,
"grad_norm": 89.697298372,
"learning_rate": 0.0001553805774278215,
"loss": 1.0665,
"step": 7140
},
{
"epoch": 2.0118176702307258,
"grad_norm": 33.50349679176105,
"learning_rate": 0.00015516185476815398,
"loss": 0.6788,
"step": 7150
},
{
"epoch": 2.0146314012380415,
"grad_norm": 17.68602683534024,
"learning_rate": 0.00015494313210848642,
"loss": 0.7757,
"step": 7160
},
{
"epoch": 2.0174451322453573,
"grad_norm": 9.545453345027244,
"learning_rate": 0.00015472440944881886,
"loss": 0.6453,
"step": 7170
},
{
"epoch": 2.020258863252673,
"grad_norm": 35.59753729386904,
"learning_rate": 0.00015450568678915133,
"loss": 0.3854,
"step": 7180
},
{
"epoch": 2.0230725942599888,
"grad_norm": 9.072270826783758,
"learning_rate": 0.0001542869641294838,
"loss": 0.6609,
"step": 7190
},
{
"epoch": 2.0258863252673045,
"grad_norm": 47.950374988048274,
"learning_rate": 0.00015406824146981627,
"loss": 1.2295,
"step": 7200
},
{
"epoch": 2.0287000562746202,
"grad_norm": 20.65584618351164,
"learning_rate": 0.00015384951881014874,
"loss": 0.638,
"step": 7210
},
{
"epoch": 2.031513787281936,
"grad_norm": 56.42081760798516,
"learning_rate": 0.00015363079615048116,
"loss": 0.6909,
"step": 7220
},
{
"epoch": 2.0343275182892517,
"grad_norm": 40.50031891450003,
"learning_rate": 0.00015341207349081363,
"loss": 0.5031,
"step": 7230
},
{
"epoch": 2.037141249296567,
"grad_norm": 17.00567280211469,
"learning_rate": 0.0001531933508311461,
"loss": 0.3856,
"step": 7240
},
{
"epoch": 2.039954980303883,
"grad_norm": 4.489135688247525,
"learning_rate": 0.00015297462817147854,
"loss": 0.2859,
"step": 7250
},
{
"epoch": 2.0427687113111985,
"grad_norm": 14.775510153378368,
"learning_rate": 0.000152755905511811,
"loss": 0.4759,
"step": 7260
},
{
"epoch": 2.0455824423185143,
"grad_norm": 76.39163767252063,
"learning_rate": 0.00015253718285214348,
"loss": 0.6392,
"step": 7270
},
{
"epoch": 2.04839617332583,
"grad_norm": 22.821590093980884,
"learning_rate": 0.00015231846019247592,
"loss": 0.7293,
"step": 7280
},
{
"epoch": 2.0512099043331458,
"grad_norm": 9.836491539631897,
"learning_rate": 0.00015209973753280837,
"loss": 0.7348,
"step": 7290
},
{
"epoch": 2.0540236353404615,
"grad_norm": 6.630565485108728,
"learning_rate": 0.00015188101487314084,
"loss": 0.5144,
"step": 7300
},
{
"epoch": 2.0568373663477773,
"grad_norm": 34.675432308942774,
"learning_rate": 0.0001516622922134733,
"loss": 0.4596,
"step": 7310
},
{
"epoch": 2.059651097355093,
"grad_norm": 7.181607771013236,
"learning_rate": 0.00015144356955380578,
"loss": 0.4084,
"step": 7320
},
{
"epoch": 2.0624648283624087,
"grad_norm": 66.82180898278047,
"learning_rate": 0.00015122484689413822,
"loss": 0.5495,
"step": 7330
},
{
"epoch": 2.065278559369724,
"grad_norm": 1.5514777421730275,
"learning_rate": 0.0001510061242344707,
"loss": 1.051,
"step": 7340
},
{
"epoch": 2.06809229037704,
"grad_norm": 283.1577772960893,
"learning_rate": 0.00015078740157480313,
"loss": 0.6516,
"step": 7350
},
{
"epoch": 2.0709060213843555,
"grad_norm": 3.7731035338477232,
"learning_rate": 0.00015056867891513558,
"loss": 0.4129,
"step": 7360
},
{
"epoch": 2.0737197523916713,
"grad_norm": 20.19374459118123,
"learning_rate": 0.00015034995625546805,
"loss": 0.6187,
"step": 7370
},
{
"epoch": 2.076533483398987,
"grad_norm": 78.73050551341473,
"learning_rate": 0.00015013123359580052,
"loss": 1.2969,
"step": 7380
},
{
"epoch": 2.0793472144063028,
"grad_norm": 165.4510143886445,
"learning_rate": 0.00014991251093613296,
"loss": 0.7819,
"step": 7390
},
{
"epoch": 2.0821609454136185,
"grad_norm": 7.206876778950156,
"learning_rate": 0.00014969378827646543,
"loss": 0.5006,
"step": 7400
},
{
"epoch": 2.0849746764209343,
"grad_norm": 30.51391990574,
"learning_rate": 0.0001494750656167979,
"loss": 0.5426,
"step": 7410
},
{
"epoch": 2.08778840742825,
"grad_norm": 82.36403908402391,
"learning_rate": 0.00014925634295713034,
"loss": 0.792,
"step": 7420
},
{
"epoch": 2.0906021384355657,
"grad_norm": 4.553958671951656,
"learning_rate": 0.0001490376202974628,
"loss": 0.5293,
"step": 7430
},
{
"epoch": 2.093415869442881,
"grad_norm": 33.752213518528826,
"learning_rate": 0.00014881889763779525,
"loss": 0.5422,
"step": 7440
},
{
"epoch": 2.096229600450197,
"grad_norm": 113.08178742115065,
"learning_rate": 0.00014860017497812772,
"loss": 0.5121,
"step": 7450
},
{
"epoch": 2.0990433314575125,
"grad_norm": 30.526152576678303,
"learning_rate": 0.00014838145231846017,
"loss": 0.5431,
"step": 7460
},
{
"epoch": 2.1018570624648283,
"grad_norm": 51.53196387319876,
"learning_rate": 0.00014816272965879264,
"loss": 0.5991,
"step": 7470
},
{
"epoch": 2.104670793472144,
"grad_norm": 25.153202967002166,
"learning_rate": 0.0001479440069991251,
"loss": 0.4238,
"step": 7480
},
{
"epoch": 2.1074845244794598,
"grad_norm": 10.632643598102124,
"learning_rate": 0.00014772528433945755,
"loss": 0.4666,
"step": 7490
},
{
"epoch": 2.1102982554867755,
"grad_norm": 25.335057772955818,
"learning_rate": 0.00014750656167979002,
"loss": 0.5208,
"step": 7500
},
{
"epoch": 2.1131119864940913,
"grad_norm": 63.40523065355852,
"learning_rate": 0.0001472878390201225,
"loss": 0.6438,
"step": 7510
},
{
"epoch": 2.115925717501407,
"grad_norm": 23.800767421170576,
"learning_rate": 0.00014706911636045493,
"loss": 0.3728,
"step": 7520
},
{
"epoch": 2.1187394485087228,
"grad_norm": 27.228833850051487,
"learning_rate": 0.00014685039370078738,
"loss": 0.4183,
"step": 7530
},
{
"epoch": 2.121553179516038,
"grad_norm": 27.19548175324042,
"learning_rate": 0.00014663167104111985,
"loss": 0.4994,
"step": 7540
},
{
"epoch": 2.124366910523354,
"grad_norm": 24.460974703930734,
"learning_rate": 0.0001464129483814523,
"loss": 0.5199,
"step": 7550
},
{
"epoch": 2.1271806415306695,
"grad_norm": 38.542816752552284,
"learning_rate": 0.00014619422572178476,
"loss": 0.4282,
"step": 7560
},
{
"epoch": 2.1299943725379853,
"grad_norm": 18.694200296950598,
"learning_rate": 0.00014597550306211723,
"loss": 0.4003,
"step": 7570
},
{
"epoch": 2.132808103545301,
"grad_norm": 47.57626879348759,
"learning_rate": 0.00014575678040244967,
"loss": 0.3651,
"step": 7580
},
{
"epoch": 2.135621834552617,
"grad_norm": 137.96598042768042,
"learning_rate": 0.00014553805774278214,
"loss": 0.362,
"step": 7590
},
{
"epoch": 2.1384355655599325,
"grad_norm": 43.01036785686837,
"learning_rate": 0.0001453193350831146,
"loss": 0.8131,
"step": 7600
},
{
"epoch": 2.1412492965672483,
"grad_norm": 7.717528689632034,
"learning_rate": 0.00014510061242344705,
"loss": 0.4267,
"step": 7610
},
{
"epoch": 2.144063027574564,
"grad_norm": 45.109974058089236,
"learning_rate": 0.00014488188976377952,
"loss": 0.5199,
"step": 7620
},
{
"epoch": 2.1468767585818798,
"grad_norm": 34.364453456078586,
"learning_rate": 0.00014466316710411197,
"loss": 0.5723,
"step": 7630
},
{
"epoch": 2.1496904895891955,
"grad_norm": 38.339990028883506,
"learning_rate": 0.0001444444444444444,
"loss": 1.1828,
"step": 7640
},
{
"epoch": 2.152504220596511,
"grad_norm": 10.007067441938148,
"learning_rate": 0.00014422572178477688,
"loss": 0.749,
"step": 7650
},
{
"epoch": 2.1553179516038266,
"grad_norm": 3.8129734415651444,
"learning_rate": 0.00014400699912510935,
"loss": 0.3419,
"step": 7660
},
{
"epoch": 2.1581316826111423,
"grad_norm": 27.437919735141907,
"learning_rate": 0.00014378827646544182,
"loss": 0.9602,
"step": 7670
},
{
"epoch": 2.160945413618458,
"grad_norm": 16.850020692243806,
"learning_rate": 0.00014356955380577426,
"loss": 0.5824,
"step": 7680
},
{
"epoch": 2.163759144625774,
"grad_norm": 26.123500215108415,
"learning_rate": 0.00014335083114610673,
"loss": 0.5254,
"step": 7690
},
{
"epoch": 2.1665728756330895,
"grad_norm": 7.580469258495237,
"learning_rate": 0.00014313210848643918,
"loss": 0.4178,
"step": 7700
},
{
"epoch": 2.1693866066404053,
"grad_norm": 9.462955308502181,
"learning_rate": 0.00014291338582677165,
"loss": 0.4992,
"step": 7710
},
{
"epoch": 2.172200337647721,
"grad_norm": 23.036715463308244,
"learning_rate": 0.0001426946631671041,
"loss": 0.7331,
"step": 7720
},
{
"epoch": 2.1750140686550368,
"grad_norm": 13.038447468985156,
"learning_rate": 0.00014247594050743656,
"loss": 0.5071,
"step": 7730
},
{
"epoch": 2.177827799662352,
"grad_norm": 47.061880181069775,
"learning_rate": 0.000142257217847769,
"loss": 0.5573,
"step": 7740
},
{
"epoch": 2.180641530669668,
"grad_norm": 22.432526114178756,
"learning_rate": 0.00014203849518810147,
"loss": 0.7032,
"step": 7750
},
{
"epoch": 2.1834552616769836,
"grad_norm": 15.983873087217463,
"learning_rate": 0.00014181977252843394,
"loss": 0.5136,
"step": 7760
},
{
"epoch": 2.1862689926842993,
"grad_norm": 5.098441308324375,
"learning_rate": 0.00014160104986876639,
"loss": 0.3924,
"step": 7770
},
{
"epoch": 2.189082723691615,
"grad_norm": 24.02126615806521,
"learning_rate": 0.00014138232720909886,
"loss": 0.4609,
"step": 7780
},
{
"epoch": 2.191896454698931,
"grad_norm": 22.537388916291963,
"learning_rate": 0.00014116360454943133,
"loss": 0.927,
"step": 7790
},
{
"epoch": 2.1947101857062465,
"grad_norm": 126.37626869176091,
"learning_rate": 0.00014094488188976377,
"loss": 0.5503,
"step": 7800
},
{
"epoch": 2.1975239167135623,
"grad_norm": 16.78141729175572,
"learning_rate": 0.00014072615923009624,
"loss": 0.5167,
"step": 7810
},
{
"epoch": 2.200337647720878,
"grad_norm": 11.596014649927676,
"learning_rate": 0.00014050743657042868,
"loss": 0.3368,
"step": 7820
},
{
"epoch": 2.2031513787281938,
"grad_norm": 52.42749578186254,
"learning_rate": 0.00014028871391076112,
"loss": 0.7298,
"step": 7830
},
{
"epoch": 2.205965109735509,
"grad_norm": 8.422176389308614,
"learning_rate": 0.0001400699912510936,
"loss": 0.6176,
"step": 7840
},
{
"epoch": 2.208778840742825,
"grad_norm": 14.525285965374584,
"learning_rate": 0.00013985126859142606,
"loss": 0.3548,
"step": 7850
},
{
"epoch": 2.2115925717501406,
"grad_norm": 74.49968322584152,
"learning_rate": 0.0001396325459317585,
"loss": 0.4476,
"step": 7860
},
{
"epoch": 2.2144063027574563,
"grad_norm": 15.833711009373205,
"learning_rate": 0.00013941382327209098,
"loss": 0.5139,
"step": 7870
},
{
"epoch": 2.217220033764772,
"grad_norm": 9.502417611259494,
"learning_rate": 0.00013919510061242345,
"loss": 0.3339,
"step": 7880
},
{
"epoch": 2.220033764772088,
"grad_norm": 42.80936499484905,
"learning_rate": 0.0001389763779527559,
"loss": 0.8684,
"step": 7890
},
{
"epoch": 2.2228474957794035,
"grad_norm": 22.65081394619362,
"learning_rate": 0.00013875765529308836,
"loss": 0.4503,
"step": 7900
},
{
"epoch": 2.2256612267867193,
"grad_norm": 53.929376514239436,
"learning_rate": 0.0001385389326334208,
"loss": 0.631,
"step": 7910
},
{
"epoch": 2.228474957794035,
"grad_norm": 22.765753793450298,
"learning_rate": 0.00013832020997375327,
"loss": 0.4681,
"step": 7920
},
{
"epoch": 2.231288688801351,
"grad_norm": 20.083204174681672,
"learning_rate": 0.00013810148731408572,
"loss": 0.4727,
"step": 7930
},
{
"epoch": 2.234102419808666,
"grad_norm": 41.8348829519395,
"learning_rate": 0.00013788276465441819,
"loss": 0.5225,
"step": 7940
},
{
"epoch": 2.236916150815982,
"grad_norm": 24.39047987978454,
"learning_rate": 0.00013766404199475066,
"loss": 0.5678,
"step": 7950
},
{
"epoch": 2.2397298818232976,
"grad_norm": 5.829038461480086,
"learning_rate": 0.0001374453193350831,
"loss": 0.5068,
"step": 7960
},
{
"epoch": 2.2425436128306133,
"grad_norm": 3.39378744630721,
"learning_rate": 0.00013722659667541557,
"loss": 0.6194,
"step": 7970
},
{
"epoch": 2.245357343837929,
"grad_norm": 5.237893149202979,
"learning_rate": 0.00013700787401574804,
"loss": 0.5441,
"step": 7980
},
{
"epoch": 2.248171074845245,
"grad_norm": 35.256946231031435,
"learning_rate": 0.00013678915135608048,
"loss": 0.8805,
"step": 7990
},
{
"epoch": 2.2509848058525606,
"grad_norm": 6.339404320662685,
"learning_rate": 0.00013657042869641292,
"loss": 0.3679,
"step": 8000
},
{
"epoch": 2.2537985368598763,
"grad_norm": 39.61705527700101,
"learning_rate": 0.0001363517060367454,
"loss": 0.6514,
"step": 8010
},
{
"epoch": 2.256612267867192,
"grad_norm": 4.860545258191048,
"learning_rate": 0.00013613298337707784,
"loss": 0.4431,
"step": 8020
},
{
"epoch": 2.259425998874508,
"grad_norm": 23.334033297076132,
"learning_rate": 0.0001359142607174103,
"loss": 0.2908,
"step": 8030
},
{
"epoch": 2.2622397298818235,
"grad_norm": 2.255539515554214,
"learning_rate": 0.00013569553805774278,
"loss": 0.5004,
"step": 8040
},
{
"epoch": 2.265053460889139,
"grad_norm": 4.512388168148573,
"learning_rate": 0.00013547681539807522,
"loss": 0.3133,
"step": 8050
},
{
"epoch": 2.2678671918964546,
"grad_norm": 88.81343772870977,
"learning_rate": 0.0001352580927384077,
"loss": 0.7748,
"step": 8060
},
{
"epoch": 2.2706809229037703,
"grad_norm": 16.622608833874658,
"learning_rate": 0.00013503937007874016,
"loss": 0.4579,
"step": 8070
},
{
"epoch": 2.273494653911086,
"grad_norm": 24.28677195401668,
"learning_rate": 0.00013484251968503937,
"loss": 1.008,
"step": 8080
},
{
"epoch": 2.276308384918402,
"grad_norm": 12.088925730126462,
"learning_rate": 0.0001346237970253718,
"loss": 0.7886,
"step": 8090
},
{
"epoch": 2.2791221159257176,
"grad_norm": 9.060718509259697,
"learning_rate": 0.00013440507436570428,
"loss": 0.395,
"step": 8100
},
{
"epoch": 2.2819358469330333,
"grad_norm": 10.115823041010223,
"learning_rate": 0.00013418635170603672,
"loss": 0.4165,
"step": 8110
},
{
"epoch": 2.284749577940349,
"grad_norm": 30.26934802189062,
"learning_rate": 0.0001339676290463692,
"loss": 0.542,
"step": 8120
},
{
"epoch": 2.287563308947665,
"grad_norm": 25.66327045675839,
"learning_rate": 0.00013374890638670164,
"loss": 0.4935,
"step": 8130
},
{
"epoch": 2.29037703995498,
"grad_norm": 56.781658920637945,
"learning_rate": 0.0001335301837270341,
"loss": 0.8249,
"step": 8140
},
{
"epoch": 2.293190770962296,
"grad_norm": 21.68653409329514,
"learning_rate": 0.00013331146106736658,
"loss": 0.5771,
"step": 8150
},
{
"epoch": 2.2960045019696116,
"grad_norm": 9.068927621383619,
"learning_rate": 0.00013309273840769902,
"loss": 0.4923,
"step": 8160
},
{
"epoch": 2.2988182329769273,
"grad_norm": 54.33234299837627,
"learning_rate": 0.0001328740157480315,
"loss": 0.718,
"step": 8170
},
{
"epoch": 2.301631963984243,
"grad_norm": 8.0851611902692,
"learning_rate": 0.00013265529308836396,
"loss": 0.3015,
"step": 8180
},
{
"epoch": 2.304445694991559,
"grad_norm": 14.93759192354656,
"learning_rate": 0.0001324365704286964,
"loss": 0.4911,
"step": 8190
},
{
"epoch": 2.3072594259988746,
"grad_norm": 67.05525829681581,
"learning_rate": 0.00013221784776902884,
"loss": 0.6472,
"step": 8200
},
{
"epoch": 2.3100731570061903,
"grad_norm": 39.202689322357536,
"learning_rate": 0.00013199912510936131,
"loss": 0.4606,
"step": 8210
},
{
"epoch": 2.312886888013506,
"grad_norm": 178.24379134099266,
"learning_rate": 0.00013178040244969378,
"loss": 0.6354,
"step": 8220
},
{
"epoch": 2.315700619020822,
"grad_norm": 85.25424593199081,
"learning_rate": 0.00013156167979002623,
"loss": 0.4367,
"step": 8230
},
{
"epoch": 2.3185143500281375,
"grad_norm": 53.75533136940712,
"learning_rate": 0.0001313429571303587,
"loss": 0.6239,
"step": 8240
},
{
"epoch": 2.321328081035453,
"grad_norm": 12.586951832356146,
"learning_rate": 0.00013112423447069117,
"loss": 0.5565,
"step": 8250
},
{
"epoch": 2.3241418120427686,
"grad_norm": 16.833339361466688,
"learning_rate": 0.0001309055118110236,
"loss": 0.5959,
"step": 8260
},
{
"epoch": 2.3269555430500843,
"grad_norm": 50.26646442085951,
"learning_rate": 0.00013068678915135608,
"loss": 0.4654,
"step": 8270
},
{
"epoch": 2.3297692740574,
"grad_norm": 129.88253348184315,
"learning_rate": 0.00013046806649168852,
"loss": 0.6027,
"step": 8280
},
{
"epoch": 2.332583005064716,
"grad_norm": 120.6512466258759,
"learning_rate": 0.000130249343832021,
"loss": 0.6416,
"step": 8290
},
{
"epoch": 2.3353967360720316,
"grad_norm": 76.6852745473175,
"learning_rate": 0.00013003062117235344,
"loss": 0.4121,
"step": 8300
},
{
"epoch": 2.3382104670793473,
"grad_norm": 35.47557842382567,
"learning_rate": 0.0001298118985126859,
"loss": 0.5189,
"step": 8310
},
{
"epoch": 2.341024198086663,
"grad_norm": 192.0503118081952,
"learning_rate": 0.00012959317585301835,
"loss": 0.8179,
"step": 8320
},
{
"epoch": 2.3438379290939784,
"grad_norm": 18.588204667690324,
"learning_rate": 0.00012937445319335082,
"loss": 0.3917,
"step": 8330
},
{
"epoch": 2.346651660101294,
"grad_norm": 191.70421960014338,
"learning_rate": 0.0001291557305336833,
"loss": 0.5404,
"step": 8340
},
{
"epoch": 2.34946539110861,
"grad_norm": 238.9625701963259,
"learning_rate": 0.00012893700787401573,
"loss": 0.5827,
"step": 8350
},
{
"epoch": 2.3522791221159256,
"grad_norm": 25.743296204281318,
"learning_rate": 0.0001287182852143482,
"loss": 0.3308,
"step": 8360
},
{
"epoch": 2.3550928531232413,
"grad_norm": 17.13298864313216,
"learning_rate": 0.00012849956255468065,
"loss": 0.4597,
"step": 8370
},
{
"epoch": 2.357906584130557,
"grad_norm": 27.69653969266591,
"learning_rate": 0.00012828083989501312,
"loss": 0.2887,
"step": 8380
},
{
"epoch": 2.360720315137873,
"grad_norm": 18.82886891300214,
"learning_rate": 0.00012808398950131232,
"loss": 0.4994,
"step": 8390
},
{
"epoch": 2.3635340461451886,
"grad_norm": 46.28194584088359,
"learning_rate": 0.00012786526684164476,
"loss": 0.6073,
"step": 8400
},
{
"epoch": 2.3663477771525043,
"grad_norm": 4.807033074829807,
"learning_rate": 0.00012764654418197723,
"loss": 0.5521,
"step": 8410
},
{
"epoch": 2.36916150815982,
"grad_norm": 25.502911733365003,
"learning_rate": 0.0001274278215223097,
"loss": 0.4886,
"step": 8420
},
{
"epoch": 2.371975239167136,
"grad_norm": 37.960031508401514,
"learning_rate": 0.00012720909886264215,
"loss": 0.676,
"step": 8430
},
{
"epoch": 2.3747889701744516,
"grad_norm": 9.396468446980126,
"learning_rate": 0.00012699037620297462,
"loss": 0.3818,
"step": 8440
},
{
"epoch": 2.377602701181767,
"grad_norm": 140.64704015650366,
"learning_rate": 0.0001267716535433071,
"loss": 0.6695,
"step": 8450
},
{
"epoch": 2.3804164321890826,
"grad_norm": 704.8913985778679,
"learning_rate": 0.00012655293088363953,
"loss": 0.4278,
"step": 8460
},
{
"epoch": 2.3832301631963984,
"grad_norm": 24.070673929018703,
"learning_rate": 0.000126334208223972,
"loss": 0.6819,
"step": 8470
},
{
"epoch": 2.386043894203714,
"grad_norm": 49.75637617417452,
"learning_rate": 0.00012611548556430444,
"loss": 0.7224,
"step": 8480
},
{
"epoch": 2.38885762521103,
"grad_norm": 25.377596250206288,
"learning_rate": 0.0001258967629046369,
"loss": 0.5522,
"step": 8490
},
{
"epoch": 2.3916713562183456,
"grad_norm": 44.21196059010374,
"learning_rate": 0.00012567804024496936,
"loss": 0.552,
"step": 8500
},
{
"epoch": 2.3944850872256613,
"grad_norm": 93.86546608453293,
"learning_rate": 0.00012545931758530183,
"loss": 0.4249,
"step": 8510
},
{
"epoch": 2.397298818232977,
"grad_norm": 20.326633023305288,
"learning_rate": 0.0001252405949256343,
"loss": 0.4898,
"step": 8520
},
{
"epoch": 2.4001125492402924,
"grad_norm": 6.861362262817904,
"learning_rate": 0.00012502187226596674,
"loss": 0.4746,
"step": 8530
},
{
"epoch": 2.402926280247608,
"grad_norm": 5.791391274939596,
"learning_rate": 0.0001248031496062992,
"loss": 0.4244,
"step": 8540
},
{
"epoch": 2.405740011254924,
"grad_norm": 31.062706775859727,
"learning_rate": 0.00012460629921259842,
"loss": 0.6831,
"step": 8550
},
{
"epoch": 2.4085537422622396,
"grad_norm": 155.36648204775855,
"learning_rate": 0.00012438757655293089,
"loss": 0.6126,
"step": 8560
},
{
"epoch": 2.4113674732695554,
"grad_norm": 54.962311559010956,
"learning_rate": 0.00012416885389326333,
"loss": 0.5716,
"step": 8570
},
{
"epoch": 2.414181204276871,
"grad_norm": 118.40831467456086,
"learning_rate": 0.00012395013123359577,
"loss": 0.7426,
"step": 8580
},
{
"epoch": 2.416994935284187,
"grad_norm": 810.7603005158664,
"learning_rate": 0.00012373140857392824,
"loss": 0.5642,
"step": 8590
},
{
"epoch": 2.4198086662915026,
"grad_norm": 51.79506077875997,
"learning_rate": 0.0001235126859142607,
"loss": 0.3925,
"step": 8600
},
{
"epoch": 2.4226223972988183,
"grad_norm": 57.44045267412865,
"learning_rate": 0.00012329396325459315,
"loss": 0.4651,
"step": 8610
},
{
"epoch": 2.425436128306134,
"grad_norm": 10.645743056447664,
"learning_rate": 0.00012307524059492562,
"loss": 0.5592,
"step": 8620
},
{
"epoch": 2.42824985931345,
"grad_norm": 39.82888930894237,
"learning_rate": 0.0001228565179352581,
"loss": 0.5235,
"step": 8630
},
{
"epoch": 2.4310635903207656,
"grad_norm": 202.40272841895077,
"learning_rate": 0.00012263779527559054,
"loss": 0.6403,
"step": 8640
},
{
"epoch": 2.433877321328081,
"grad_norm": 8.768929997416398,
"learning_rate": 0.000122419072615923,
"loss": 0.3559,
"step": 8650
},
{
"epoch": 2.4366910523353966,
"grad_norm": 73.24658593563804,
"learning_rate": 0.00012220034995625545,
"loss": 0.8297,
"step": 8660
},
{
"epoch": 2.4395047833427124,
"grad_norm": 9.18198585469821,
"learning_rate": 0.00012198162729658791,
"loss": 0.7252,
"step": 8670
},
{
"epoch": 2.442318514350028,
"grad_norm": 364.57870592301464,
"learning_rate": 0.00012176290463692038,
"loss": 0.5064,
"step": 8680
},
{
"epoch": 2.445132245357344,
"grad_norm": 21.670796127528217,
"learning_rate": 0.00012154418197725283,
"loss": 0.374,
"step": 8690
},
{
"epoch": 2.4479459763646596,
"grad_norm": 17.324860013255872,
"learning_rate": 0.00012132545931758528,
"loss": 0.4137,
"step": 8700
},
{
"epoch": 2.4507597073719753,
"grad_norm": 22.780757919669334,
"learning_rate": 0.00012110673665791775,
"loss": 0.4646,
"step": 8710
},
{
"epoch": 2.453573438379291,
"grad_norm": 39.75471721705483,
"learning_rate": 0.00012088801399825022,
"loss": 0.5432,
"step": 8720
},
{
"epoch": 2.4563871693866064,
"grad_norm": 49.74115368293155,
"learning_rate": 0.00012066929133858267,
"loss": 0.9697,
"step": 8730
},
{
"epoch": 2.459200900393922,
"grad_norm": 15.825907605556525,
"learning_rate": 0.00012045056867891512,
"loss": 0.4463,
"step": 8740
},
{
"epoch": 2.462014631401238,
"grad_norm": 17.73048175980216,
"learning_rate": 0.00012023184601924759,
"loss": 0.4398,
"step": 8750
},
{
"epoch": 2.4648283624085536,
"grad_norm": 30.183102182735098,
"learning_rate": 0.00012001312335958004,
"loss": 0.6332,
"step": 8760
},
{
"epoch": 2.4676420934158694,
"grad_norm": 25.73335808399061,
"learning_rate": 0.0001197944006999125,
"loss": 0.4339,
"step": 8770
},
{
"epoch": 2.470455824423185,
"grad_norm": 18.419862117919163,
"learning_rate": 0.00011957567804024496,
"loss": 0.4134,
"step": 8780
},
{
"epoch": 2.473269555430501,
"grad_norm": 30.616274458695887,
"learning_rate": 0.00011935695538057743,
"loss": 0.2893,
"step": 8790
},
{
"epoch": 2.4760832864378166,
"grad_norm": 1.356291998114622,
"learning_rate": 0.00011913823272090987,
"loss": 0.7339,
"step": 8800
},
{
"epoch": 2.4788970174451324,
"grad_norm": 21.213031498236028,
"learning_rate": 0.00011891951006124234,
"loss": 0.3974,
"step": 8810
},
{
"epoch": 2.481710748452448,
"grad_norm": 16.21040199222578,
"learning_rate": 0.0001187007874015748,
"loss": 0.5217,
"step": 8820
},
{
"epoch": 2.484524479459764,
"grad_norm": 13.860914140582063,
"learning_rate": 0.00011848206474190725,
"loss": 0.4859,
"step": 8830
},
{
"epoch": 2.4873382104670796,
"grad_norm": 48.496668101430515,
"learning_rate": 0.00011826334208223971,
"loss": 0.6578,
"step": 8840
},
{
"epoch": 2.490151941474395,
"grad_norm": 8.89511240692744,
"learning_rate": 0.00011804461942257218,
"loss": 0.4366,
"step": 8850
},
{
"epoch": 2.4929656724817106,
"grad_norm": 22.47093178418468,
"learning_rate": 0.00011782589676290462,
"loss": 0.5115,
"step": 8860
},
{
"epoch": 2.4957794034890264,
"grad_norm": 16.50971197997101,
"learning_rate": 0.00011760717410323709,
"loss": 0.4433,
"step": 8870
},
{
"epoch": 2.498593134496342,
"grad_norm": 6.399589126360222,
"learning_rate": 0.00011738845144356955,
"loss": 0.3354,
"step": 8880
},
{
"epoch": 2.501406865503658,
"grad_norm": 10.640344890543053,
"learning_rate": 0.00011716972878390199,
"loss": 0.9088,
"step": 8890
},
{
"epoch": 2.5042205965109736,
"grad_norm": 17.669174903791003,
"learning_rate": 0.00011695100612423446,
"loss": 0.7106,
"step": 8900
},
{
"epoch": 2.5070343275182894,
"grad_norm": 19.417937682945556,
"learning_rate": 0.00011673228346456692,
"loss": 0.5522,
"step": 8910
},
{
"epoch": 2.509848058525605,
"grad_norm": 21.275739872323875,
"learning_rate": 0.00011651356080489937,
"loss": 0.5965,
"step": 8920
},
{
"epoch": 2.5126617895329204,
"grad_norm": 13.434077583334158,
"learning_rate": 0.00011629483814523183,
"loss": 0.5786,
"step": 8930
},
{
"epoch": 2.515475520540236,
"grad_norm": 25.83592797652058,
"learning_rate": 0.0001160761154855643,
"loss": 0.4222,
"step": 8940
},
{
"epoch": 2.518289251547552,
"grad_norm": 37.20038618687167,
"learning_rate": 0.00011585739282589676,
"loss": 0.4256,
"step": 8950
},
{
"epoch": 2.5211029825548676,
"grad_norm": 53.97304836312147,
"learning_rate": 0.00011563867016622921,
"loss": 0.4834,
"step": 8960
},
{
"epoch": 2.5239167135621834,
"grad_norm": 16.475642718077715,
"learning_rate": 0.00011541994750656167,
"loss": 0.5385,
"step": 8970
},
{
"epoch": 2.526730444569499,
"grad_norm": 35.76870275664621,
"learning_rate": 0.00011520122484689414,
"loss": 0.4718,
"step": 8980
},
{
"epoch": 2.529544175576815,
"grad_norm": 17.680183575624334,
"learning_rate": 0.00011498250218722658,
"loss": 0.549,
"step": 8990
},
{
"epoch": 2.5323579065841306,
"grad_norm": 97.68298591049088,
"learning_rate": 0.00011476377952755905,
"loss": 0.4642,
"step": 9000
},
{
"epoch": 2.5351716375914464,
"grad_norm": 95.91488844225455,
"learning_rate": 0.00011454505686789151,
"loss": 0.4499,
"step": 9010
},
{
"epoch": 2.537985368598762,
"grad_norm": 49.98057434380942,
"learning_rate": 0.00011432633420822395,
"loss": 0.4452,
"step": 9020
},
{
"epoch": 2.540799099606078,
"grad_norm": 86.94738373288978,
"learning_rate": 0.00011410761154855642,
"loss": 0.654,
"step": 9030
},
{
"epoch": 2.5436128306133936,
"grad_norm": 46.27504444954838,
"learning_rate": 0.00011388888888888889,
"loss": 0.8217,
"step": 9040
},
{
"epoch": 2.546426561620709,
"grad_norm": 6.8204282978011195,
"learning_rate": 0.00011367016622922133,
"loss": 0.3788,
"step": 9050
},
{
"epoch": 2.5492402926280247,
"grad_norm": 74.45875969586639,
"learning_rate": 0.00011345144356955379,
"loss": 1.7253,
"step": 9060
},
{
"epoch": 2.5520540236353404,
"grad_norm": 8.860003682861251,
"learning_rate": 0.00011323272090988626,
"loss": 0.4397,
"step": 9070
},
{
"epoch": 2.554867754642656,
"grad_norm": 48.65502795851232,
"learning_rate": 0.0001130139982502187,
"loss": 0.386,
"step": 9080
},
{
"epoch": 2.557681485649972,
"grad_norm": 3.2137363317945287,
"learning_rate": 0.00011279527559055117,
"loss": 0.2666,
"step": 9090
},
{
"epoch": 2.5604952166572876,
"grad_norm": 124.42119058882817,
"learning_rate": 0.00011257655293088363,
"loss": 0.9019,
"step": 9100
},
{
"epoch": 2.5633089476646034,
"grad_norm": 32.9544365134875,
"learning_rate": 0.00011235783027121609,
"loss": 0.9048,
"step": 9110
},
{
"epoch": 2.566122678671919,
"grad_norm": 7.015944851676098,
"learning_rate": 0.00011213910761154854,
"loss": 0.3671,
"step": 9120
},
{
"epoch": 2.5689364096792344,
"grad_norm": 36.00750003943152,
"learning_rate": 0.00011192038495188101,
"loss": 0.8046,
"step": 9130
},
{
"epoch": 2.57175014068655,
"grad_norm": 24.83730040509871,
"learning_rate": 0.00011170166229221346,
"loss": 0.5702,
"step": 9140
},
{
"epoch": 2.574563871693866,
"grad_norm": 5.192263862462742,
"learning_rate": 0.00011148293963254593,
"loss": 0.4801,
"step": 9150
},
{
"epoch": 2.5773776027011817,
"grad_norm": 9.485040925668613,
"learning_rate": 0.00011126421697287838,
"loss": 0.7347,
"step": 9160
},
{
"epoch": 2.5801913337084974,
"grad_norm": 188.0655816314744,
"learning_rate": 0.00011104549431321082,
"loss": 0.5629,
"step": 9170
},
{
"epoch": 2.583005064715813,
"grad_norm": 4.262160421678828,
"learning_rate": 0.0001108267716535433,
"loss": 0.3597,
"step": 9180
},
{
"epoch": 2.585818795723129,
"grad_norm": 48.99676536082116,
"learning_rate": 0.00011060804899387576,
"loss": 0.5096,
"step": 9190
},
{
"epoch": 2.5886325267304446,
"grad_norm": 49.32059612461514,
"learning_rate": 0.00011038932633420822,
"loss": 0.8765,
"step": 9200
},
{
"epoch": 2.5914462577377604,
"grad_norm": 11.590855409332988,
"learning_rate": 0.00011017060367454066,
"loss": 0.6143,
"step": 9210
},
{
"epoch": 2.594259988745076,
"grad_norm": 38.29167922597077,
"learning_rate": 0.00010995188101487313,
"loss": 0.7939,
"step": 9220
},
{
"epoch": 2.597073719752392,
"grad_norm": 32.31895255440478,
"learning_rate": 0.00010973315835520559,
"loss": 0.4451,
"step": 9230
},
{
"epoch": 2.5998874507597076,
"grad_norm": 33.42321188287331,
"learning_rate": 0.00010951443569553805,
"loss": 0.5792,
"step": 9240
},
{
"epoch": 2.602701181767023,
"grad_norm": 468.47470275260395,
"learning_rate": 0.0001092957130358705,
"loss": 0.4165,
"step": 9250
},
{
"epoch": 2.6055149127743387,
"grad_norm": 8.786170853391159,
"learning_rate": 0.00010907699037620297,
"loss": 0.3003,
"step": 9260
},
{
"epoch": 2.6083286437816544,
"grad_norm": 17.262373356558324,
"learning_rate": 0.00010885826771653542,
"loss": 0.3603,
"step": 9270
},
{
"epoch": 2.61114237478897,
"grad_norm": 86.87405146897451,
"learning_rate": 0.00010863954505686789,
"loss": 0.3915,
"step": 9280
},
{
"epoch": 2.613956105796286,
"grad_norm": 70.29811492932059,
"learning_rate": 0.00010842082239720034,
"loss": 0.5594,
"step": 9290
},
{
"epoch": 2.6167698368036016,
"grad_norm": 102.37891262155871,
"learning_rate": 0.0001082020997375328,
"loss": 0.477,
"step": 9300
},
{
"epoch": 2.6195835678109174,
"grad_norm": 116.93161920779444,
"learning_rate": 0.00010798337707786526,
"loss": 0.8184,
"step": 9310
},
{
"epoch": 2.622397298818233,
"grad_norm": 36.178742582782164,
"learning_rate": 0.00010776465441819773,
"loss": 0.7671,
"step": 9320
},
{
"epoch": 2.6252110298255484,
"grad_norm": 354.6394442768764,
"learning_rate": 0.00010754593175853017,
"loss": 0.5188,
"step": 9330
},
{
"epoch": 2.628024760832864,
"grad_norm": 4.175931338154612,
"learning_rate": 0.00010732720909886263,
"loss": 0.4434,
"step": 9340
},
{
"epoch": 2.63083849184018,
"grad_norm": 56.1280866933836,
"learning_rate": 0.0001071084864391951,
"loss": 0.8639,
"step": 9350
},
{
"epoch": 2.6336522228474957,
"grad_norm": 59.20745896569175,
"learning_rate": 0.00010688976377952754,
"loss": 0.3947,
"step": 9360
},
{
"epoch": 2.6364659538548114,
"grad_norm": 55.614426780242646,
"learning_rate": 0.00010667104111986001,
"loss": 0.4354,
"step": 9370
},
{
"epoch": 2.639279684862127,
"grad_norm": 50.81213904295994,
"learning_rate": 0.00010645231846019246,
"loss": 0.5589,
"step": 9380
},
{
"epoch": 2.642093415869443,
"grad_norm": 34.20241547637593,
"learning_rate": 0.00010623359580052492,
"loss": 0.5685,
"step": 9390
},
{
"epoch": 2.6449071468767587,
"grad_norm": 17.555635593890102,
"learning_rate": 0.00010601487314085738,
"loss": 0.4639,
"step": 9400
},
{
"epoch": 2.6477208778840744,
"grad_norm": 16.284132923705343,
"learning_rate": 0.00010579615048118985,
"loss": 0.3629,
"step": 9410
},
{
"epoch": 2.65053460889139,
"grad_norm": 64.32745908031606,
"learning_rate": 0.00010557742782152229,
"loss": 0.488,
"step": 9420
},
{
"epoch": 2.653348339898706,
"grad_norm": 75.65983508131147,
"learning_rate": 0.00010535870516185476,
"loss": 0.4648,
"step": 9430
},
{
"epoch": 2.6561620709060216,
"grad_norm": 12.839163573898897,
"learning_rate": 0.00010513998250218722,
"loss": 0.5513,
"step": 9440
},
{
"epoch": 2.658975801913337,
"grad_norm": 4.358631397049856,
"learning_rate": 0.00010492125984251969,
"loss": 0.4445,
"step": 9450
},
{
"epoch": 2.6617895329206527,
"grad_norm": 26.10775381519202,
"learning_rate": 0.00010470253718285213,
"loss": 0.4745,
"step": 9460
},
{
"epoch": 2.6646032639279684,
"grad_norm": 77.73370762217442,
"learning_rate": 0.0001044838145231846,
"loss": 0.7683,
"step": 9470
},
{
"epoch": 2.667416994935284,
"grad_norm": 35.63066051419088,
"learning_rate": 0.00010426509186351706,
"loss": 0.423,
"step": 9480
},
{
"epoch": 2.6702307259426,
"grad_norm": 15.36437442788385,
"learning_rate": 0.0001040463692038495,
"loss": 0.6104,
"step": 9490
},
{
"epoch": 2.6730444569499157,
"grad_norm": 10.203853452654112,
"learning_rate": 0.00010382764654418197,
"loss": 0.5783,
"step": 9500
},
{
"epoch": 2.6758581879572314,
"grad_norm": 20.8110952561946,
"learning_rate": 0.00010360892388451444,
"loss": 0.3268,
"step": 9510
},
{
"epoch": 2.678671918964547,
"grad_norm": 30.832138697360744,
"learning_rate": 0.00010339020122484688,
"loss": 0.4376,
"step": 9520
},
{
"epoch": 2.6814856499718625,
"grad_norm": 458.34051462177115,
"learning_rate": 0.00010317147856517934,
"loss": 0.3852,
"step": 9530
},
{
"epoch": 2.684299380979178,
"grad_norm": 46.6171709293664,
"learning_rate": 0.00010295275590551181,
"loss": 0.9121,
"step": 9540
},
{
"epoch": 2.687113111986494,
"grad_norm": 272.84120193286054,
"learning_rate": 0.00010273403324584425,
"loss": 0.89,
"step": 9550
},
{
"epoch": 2.6899268429938097,
"grad_norm": 120.77864028855092,
"learning_rate": 0.00010251531058617672,
"loss": 0.3202,
"step": 9560
},
{
"epoch": 2.6927405740011254,
"grad_norm": 15.406524257269288,
"learning_rate": 0.00010229658792650918,
"loss": 0.4391,
"step": 9570
},
{
"epoch": 2.695554305008441,
"grad_norm": 19.60890393546223,
"learning_rate": 0.00010207786526684163,
"loss": 0.4306,
"step": 9580
},
{
"epoch": 2.698368036015757,
"grad_norm": 43.141776219101715,
"learning_rate": 0.00010185914260717409,
"loss": 0.6398,
"step": 9590
},
{
"epoch": 2.7011817670230727,
"grad_norm": 43.13769827492887,
"learning_rate": 0.00010164041994750656,
"loss": 0.3964,
"step": 9600
},
{
"epoch": 2.7039954980303884,
"grad_norm": 3.6195378103594056,
"learning_rate": 0.000101421697287839,
"loss": 0.3933,
"step": 9610
},
{
"epoch": 2.706809229037704,
"grad_norm": 37.37053494271392,
"learning_rate": 0.00010120297462817147,
"loss": 0.5938,
"step": 9620
},
{
"epoch": 2.70962296004502,
"grad_norm": 14.787523531149347,
"learning_rate": 0.00010098425196850393,
"loss": 0.4718,
"step": 9630
},
{
"epoch": 2.7124366910523356,
"grad_norm": 64.91170649580671,
"learning_rate": 0.00010076552930883637,
"loss": 0.479,
"step": 9640
},
{
"epoch": 2.715250422059651,
"grad_norm": 10.027583266140544,
"learning_rate": 0.00010054680664916884,
"loss": 0.3553,
"step": 9650
},
{
"epoch": 2.7180641530669667,
"grad_norm": 37.02256693061005,
"learning_rate": 0.00010032808398950131,
"loss": 0.4199,
"step": 9660
},
{
"epoch": 2.7208778840742824,
"grad_norm": 18.07635825713862,
"learning_rate": 0.00010010936132983376,
"loss": 0.6734,
"step": 9670
},
{
"epoch": 2.723691615081598,
"grad_norm": 13.442989922340166,
"learning_rate": 9.989063867016621e-05,
"loss": 0.4984,
"step": 9680
},
{
"epoch": 2.726505346088914,
"grad_norm": 5.303880680734515,
"learning_rate": 9.967191601049868e-05,
"loss": 0.3973,
"step": 9690
},
{
"epoch": 2.7293190770962297,
"grad_norm": 29.495823996978398,
"learning_rate": 9.945319335083114e-05,
"loss": 0.4915,
"step": 9700
},
{
"epoch": 2.7321328081035454,
"grad_norm": 130.2447313645269,
"learning_rate": 9.92344706911636e-05,
"loss": 0.6082,
"step": 9710
},
{
"epoch": 2.734946539110861,
"grad_norm": 6.49456770331547,
"learning_rate": 9.901574803149605e-05,
"loss": 0.3068,
"step": 9720
},
{
"epoch": 2.7377602701181765,
"grad_norm": 118.57235424251638,
"learning_rate": 9.879702537182852e-05,
"loss": 0.6114,
"step": 9730
},
{
"epoch": 2.740574001125492,
"grad_norm": 231.0023016216336,
"learning_rate": 9.857830271216096e-05,
"loss": 0.8726,
"step": 9740
},
{
"epoch": 2.743387732132808,
"grad_norm": 31.19265942143221,
"learning_rate": 9.835958005249344e-05,
"loss": 0.541,
"step": 9750
},
{
"epoch": 2.7462014631401237,
"grad_norm": 5.225618741939991,
"learning_rate": 9.814085739282589e-05,
"loss": 0.1824,
"step": 9760
},
{
"epoch": 2.7490151941474394,
"grad_norm": 15.212142485160932,
"learning_rate": 9.792213473315835e-05,
"loss": 0.759,
"step": 9770
},
{
"epoch": 2.751828925154755,
"grad_norm": 31.73381245582647,
"learning_rate": 9.77034120734908e-05,
"loss": 0.5458,
"step": 9780
},
{
"epoch": 2.754642656162071,
"grad_norm": 0.3819179114787675,
"learning_rate": 9.748468941382327e-05,
"loss": 0.6246,
"step": 9790
},
{
"epoch": 2.7574563871693867,
"grad_norm": 24.57005039190513,
"learning_rate": 9.726596675415572e-05,
"loss": 0.809,
"step": 9800
},
{
"epoch": 2.7602701181767024,
"grad_norm": 205.8236592890733,
"learning_rate": 9.704724409448817e-05,
"loss": 0.7747,
"step": 9810
},
{
"epoch": 2.763083849184018,
"grad_norm": 13.399260177045598,
"learning_rate": 9.682852143482064e-05,
"loss": 0.5277,
"step": 9820
},
{
"epoch": 2.765897580191334,
"grad_norm": 21.19679766301598,
"learning_rate": 9.660979877515309e-05,
"loss": 0.7542,
"step": 9830
},
{
"epoch": 2.7687113111986497,
"grad_norm": 15.049015807925302,
"learning_rate": 9.639107611548556e-05,
"loss": 0.4367,
"step": 9840
},
{
"epoch": 2.771525042205965,
"grad_norm": 34.32401152713521,
"learning_rate": 9.617235345581801e-05,
"loss": 0.4548,
"step": 9850
},
{
"epoch": 2.7743387732132807,
"grad_norm": 3.273022569610971,
"learning_rate": 9.595363079615047e-05,
"loss": 0.5695,
"step": 9860
},
{
"epoch": 2.7771525042205965,
"grad_norm": 55.423334541815585,
"learning_rate": 9.573490813648293e-05,
"loss": 0.4559,
"step": 9870
},
{
"epoch": 2.779966235227912,
"grad_norm": 28.16145208485805,
"learning_rate": 9.55161854768154e-05,
"loss": 0.5161,
"step": 9880
},
{
"epoch": 2.782779966235228,
"grad_norm": 29.663487490852017,
"learning_rate": 9.529746281714784e-05,
"loss": 0.5348,
"step": 9890
},
{
"epoch": 2.7855936972425437,
"grad_norm": 26.125896625555498,
"learning_rate": 9.507874015748031e-05,
"loss": 0.3743,
"step": 9900
},
{
"epoch": 2.7884074282498594,
"grad_norm": 12.642777363424036,
"learning_rate": 9.486001749781277e-05,
"loss": 0.274,
"step": 9910
},
{
"epoch": 2.791221159257175,
"grad_norm": 55.49624560948837,
"learning_rate": 9.464129483814524e-05,
"loss": 0.7875,
"step": 9920
},
{
"epoch": 2.7940348902644905,
"grad_norm": 49.81266964604724,
"learning_rate": 9.442257217847768e-05,
"loss": 0.5697,
"step": 9930
},
{
"epoch": 2.7968486212718062,
"grad_norm": 19.16263333950446,
"learning_rate": 9.420384951881015e-05,
"loss": 0.5035,
"step": 9940
},
{
"epoch": 2.799662352279122,
"grad_norm": 26.980836018843,
"learning_rate": 9.39851268591426e-05,
"loss": 0.6275,
"step": 9950
},
{
"epoch": 2.8024760832864377,
"grad_norm": 3.6601734211863945,
"learning_rate": 9.376640419947505e-05,
"loss": 0.4986,
"step": 9960
},
{
"epoch": 2.8052898142937535,
"grad_norm": 80.76032184024673,
"learning_rate": 9.354768153980752e-05,
"loss": 0.554,
"step": 9970
},
{
"epoch": 2.808103545301069,
"grad_norm": 12.4762811742511,
"learning_rate": 9.332895888013999e-05,
"loss": 0.4325,
"step": 9980
},
{
"epoch": 2.810917276308385,
"grad_norm": 28.339007190053675,
"learning_rate": 9.311023622047243e-05,
"loss": 0.3323,
"step": 9990
},
{
"epoch": 2.8137310073157007,
"grad_norm": 211.21007128891176,
"learning_rate": 9.289151356080489e-05,
"loss": 0.9021,
"step": 10000
},
{
"epoch": 2.8165447383230164,
"grad_norm": 33.31758409534305,
"learning_rate": 9.267279090113736e-05,
"loss": 0.4361,
"step": 10010
},
{
"epoch": 2.819358469330332,
"grad_norm": 38.93754008185466,
"learning_rate": 9.24540682414698e-05,
"loss": 0.6279,
"step": 10020
},
{
"epoch": 2.822172200337648,
"grad_norm": 25.529883648136195,
"learning_rate": 9.223534558180227e-05,
"loss": 0.4936,
"step": 10030
},
{
"epoch": 2.8249859313449637,
"grad_norm": 54.257511831814064,
"learning_rate": 9.201662292213473e-05,
"loss": 0.5469,
"step": 10040
},
{
"epoch": 2.827799662352279,
"grad_norm": 20.90804542506976,
"learning_rate": 9.179790026246718e-05,
"loss": 0.3663,
"step": 10050
},
{
"epoch": 2.8306133933595947,
"grad_norm": 87.21990526473672,
"learning_rate": 9.157917760279964e-05,
"loss": 0.6471,
"step": 10060
},
{
"epoch": 2.8334271243669105,
"grad_norm": 4.279230208467407,
"learning_rate": 9.136045494313211e-05,
"loss": 0.4365,
"step": 10070
},
{
"epoch": 2.836240855374226,
"grad_norm": 16.216814135083176,
"learning_rate": 9.114173228346455e-05,
"loss": 0.395,
"step": 10080
},
{
"epoch": 2.839054586381542,
"grad_norm": 29.69031936910585,
"learning_rate": 9.092300962379702e-05,
"loss": 0.2356,
"step": 10090
},
{
"epoch": 2.8418683173888577,
"grad_norm": 1.5595613943954527,
"learning_rate": 9.070428696412948e-05,
"loss": 0.2262,
"step": 10100
},
{
"epoch": 2.8446820483961734,
"grad_norm": 4.525293278789326,
"learning_rate": 9.048556430446192e-05,
"loss": 0.9052,
"step": 10110
},
{
"epoch": 2.847495779403489,
"grad_norm": 26.773310344606703,
"learning_rate": 9.026684164479439e-05,
"loss": 0.6461,
"step": 10120
},
{
"epoch": 2.8503095104108045,
"grad_norm": 48.70908526560008,
"learning_rate": 9.004811898512685e-05,
"loss": 0.7359,
"step": 10130
},
{
"epoch": 2.8531232414181202,
"grad_norm": 84.11864704783623,
"learning_rate": 8.98293963254593e-05,
"loss": 0.4548,
"step": 10140
},
{
"epoch": 2.855936972425436,
"grad_norm": 5.524906428491934,
"learning_rate": 8.961067366579176e-05,
"loss": 0.4625,
"step": 10150
},
{
"epoch": 2.8587507034327517,
"grad_norm": 10.319915749419431,
"learning_rate": 8.939195100612423e-05,
"loss": 0.6446,
"step": 10160
},
{
"epoch": 2.8615644344400675,
"grad_norm": 22.781369712630177,
"learning_rate": 8.917322834645669e-05,
"loss": 0.353,
"step": 10170
},
{
"epoch": 2.864378165447383,
"grad_norm": 35.25984458553167,
"learning_rate": 8.895450568678914e-05,
"loss": 0.6551,
"step": 10180
},
{
"epoch": 2.867191896454699,
"grad_norm": 42.157133518741865,
"learning_rate": 8.87357830271216e-05,
"loss": 0.4496,
"step": 10190
},
{
"epoch": 2.8700056274620147,
"grad_norm": 22.81314493600198,
"learning_rate": 8.851706036745407e-05,
"loss": 0.5678,
"step": 10200
},
{
"epoch": 2.8728193584693305,
"grad_norm": 36.04659178861918,
"learning_rate": 8.829833770778651e-05,
"loss": 0.4828,
"step": 10210
},
{
"epoch": 2.875633089476646,
"grad_norm": 56.67857438617218,
"learning_rate": 8.807961504811898e-05,
"loss": 0.5019,
"step": 10220
},
{
"epoch": 2.878446820483962,
"grad_norm": 9.111045425788525,
"learning_rate": 8.786089238845144e-05,
"loss": 0.5607,
"step": 10230
},
{
"epoch": 2.8812605514912777,
"grad_norm": 3.505050666852027,
"learning_rate": 8.76421697287839e-05,
"loss": 0.4513,
"step": 10240
},
{
"epoch": 2.884074282498593,
"grad_norm": 54.490312257720156,
"learning_rate": 8.742344706911635e-05,
"loss": 0.6762,
"step": 10250
},
{
"epoch": 2.8868880135059087,
"grad_norm": 25.476228006992702,
"learning_rate": 8.720472440944882e-05,
"loss": 0.5411,
"step": 10260
},
{
"epoch": 2.8897017445132245,
"grad_norm": 12.990747730995873,
"learning_rate": 8.698600174978127e-05,
"loss": 0.5335,
"step": 10270
},
{
"epoch": 2.8925154755205402,
"grad_norm": 116.48590241626219,
"learning_rate": 8.676727909011372e-05,
"loss": 0.3502,
"step": 10280
},
{
"epoch": 2.895329206527856,
"grad_norm": 28.589708094686127,
"learning_rate": 8.654855643044619e-05,
"loss": 0.5962,
"step": 10290
},
{
"epoch": 2.8981429375351717,
"grad_norm": 48.20502421441493,
"learning_rate": 8.632983377077864e-05,
"loss": 0.5072,
"step": 10300
},
{
"epoch": 2.9009566685424875,
"grad_norm": 29.045361435820396,
"learning_rate": 8.61111111111111e-05,
"loss": 0.5328,
"step": 10310
},
{
"epoch": 2.903770399549803,
"grad_norm": 23.463753067966675,
"learning_rate": 8.589238845144356e-05,
"loss": 0.4669,
"step": 10320
},
{
"epoch": 2.9065841305571185,
"grad_norm": 8.94339841328865,
"learning_rate": 8.567366579177602e-05,
"loss": 0.6852,
"step": 10330
},
{
"epoch": 2.9093978615644343,
"grad_norm": 13.126501900027074,
"learning_rate": 8.545494313210847e-05,
"loss": 0.5224,
"step": 10340
},
{
"epoch": 2.91221159257175,
"grad_norm": 11.322296130692187,
"learning_rate": 8.523622047244094e-05,
"loss": 0.4298,
"step": 10350
},
{
"epoch": 2.9150253235790657,
"grad_norm": 3.9331354922682498,
"learning_rate": 8.501749781277339e-05,
"loss": 0.3009,
"step": 10360
},
{
"epoch": 2.9178390545863815,
"grad_norm": 2.3186540408341734,
"learning_rate": 8.479877515310586e-05,
"loss": 0.4631,
"step": 10370
},
{
"epoch": 2.9206527855936972,
"grad_norm": 33.11162361117131,
"learning_rate": 8.458005249343831e-05,
"loss": 0.3775,
"step": 10380
},
{
"epoch": 2.923466516601013,
"grad_norm": 12.497923893181124,
"learning_rate": 8.436132983377076e-05,
"loss": 0.5401,
"step": 10390
},
{
"epoch": 2.9262802476083287,
"grad_norm": 9.707752939333481,
"learning_rate": 8.414260717410323e-05,
"loss": 0.5099,
"step": 10400
},
{
"epoch": 2.9290939786156445,
"grad_norm": 33.075796904013835,
"learning_rate": 8.39238845144357e-05,
"loss": 0.4239,
"step": 10410
},
{
"epoch": 2.93190770962296,
"grad_norm": 27.030408601399838,
"learning_rate": 8.370516185476815e-05,
"loss": 0.3516,
"step": 10420
},
{
"epoch": 2.934721440630276,
"grad_norm": 40.90648498933933,
"learning_rate": 8.34864391951006e-05,
"loss": 0.4291,
"step": 10430
},
{
"epoch": 2.9375351716375917,
"grad_norm": 43.38996380641155,
"learning_rate": 8.326771653543307e-05,
"loss": 0.7152,
"step": 10440
},
{
"epoch": 2.940348902644907,
"grad_norm": 25.52567647846434,
"learning_rate": 8.304899387576552e-05,
"loss": 0.1973,
"step": 10450
},
{
"epoch": 2.9431626336522227,
"grad_norm": 45.972037886947575,
"learning_rate": 8.283027121609798e-05,
"loss": 0.5599,
"step": 10460
},
{
"epoch": 2.9459763646595385,
"grad_norm": 23.400081384448004,
"learning_rate": 8.261154855643044e-05,
"loss": 0.5785,
"step": 10470
},
{
"epoch": 2.9487900956668542,
"grad_norm": 15.453689858013234,
"learning_rate": 8.23928258967629e-05,
"loss": 0.5648,
"step": 10480
},
{
"epoch": 2.95160382667417,
"grad_norm": 23.99708247332893,
"learning_rate": 8.217410323709535e-05,
"loss": 0.6255,
"step": 10490
},
{
"epoch": 2.9544175576814857,
"grad_norm": 85.44333249815278,
"learning_rate": 8.195538057742782e-05,
"loss": 0.4824,
"step": 10500
},
{
"epoch": 2.9572312886888015,
"grad_norm": 13.197420910549328,
"learning_rate": 8.173665791776027e-05,
"loss": 0.382,
"step": 10510
},
{
"epoch": 2.9600450196961168,
"grad_norm": 24.812200580491094,
"learning_rate": 8.151793525809273e-05,
"loss": 0.4638,
"step": 10520
},
{
"epoch": 2.9628587507034325,
"grad_norm": 23.947322941527855,
"learning_rate": 8.129921259842519e-05,
"loss": 0.2828,
"step": 10530
},
{
"epoch": 2.9656724817107483,
"grad_norm": 26.603437638257738,
"learning_rate": 8.108048993875766e-05,
"loss": 0.4689,
"step": 10540
},
{
"epoch": 2.968486212718064,
"grad_norm": 25.162149919783538,
"learning_rate": 8.08617672790901e-05,
"loss": 0.7301,
"step": 10550
},
{
"epoch": 2.9712999437253798,
"grad_norm": 1390.431135363237,
"learning_rate": 8.064304461942257e-05,
"loss": 0.5298,
"step": 10560
},
{
"epoch": 2.9741136747326955,
"grad_norm": 51.62357269235231,
"learning_rate": 8.042432195975503e-05,
"loss": 0.3141,
"step": 10570
},
{
"epoch": 2.9769274057400112,
"grad_norm": 21.428468158450375,
"learning_rate": 8.020559930008747e-05,
"loss": 0.5803,
"step": 10580
},
{
"epoch": 2.979741136747327,
"grad_norm": 12.693813240141665,
"learning_rate": 7.998687664041994e-05,
"loss": 0.7488,
"step": 10590
},
{
"epoch": 2.9825548677546427,
"grad_norm": 67.35584313661865,
"learning_rate": 7.97681539807524e-05,
"loss": 0.3862,
"step": 10600
},
{
"epoch": 2.9853685987619585,
"grad_norm": 75.47237178728545,
"learning_rate": 7.954943132108485e-05,
"loss": 0.5021,
"step": 10610
},
{
"epoch": 2.9881823297692742,
"grad_norm": 3.6925131359371934,
"learning_rate": 7.933070866141731e-05,
"loss": 0.4324,
"step": 10620
},
{
"epoch": 2.99099606077659,
"grad_norm": 11.919767665974996,
"learning_rate": 7.911198600174978e-05,
"loss": 0.4497,
"step": 10630
},
{
"epoch": 2.9938097917839057,
"grad_norm": 84.62823094746291,
"learning_rate": 7.889326334208222e-05,
"loss": 0.7873,
"step": 10640
},
{
"epoch": 2.996623522791221,
"grad_norm": 47.500675839083364,
"learning_rate": 7.867454068241469e-05,
"loss": 0.4138,
"step": 10650
},
{
"epoch": 2.9994372537985368,
"grad_norm": 91.2796382898892,
"learning_rate": 7.845581802274715e-05,
"loss": 0.8051,
"step": 10660
},
{
"epoch": 3.0,
"eval_0_f1": 0.6165994034041059,
"eval_0_precision": 0.5338802795502887,
"eval_0_recall": 0.7296511627906976,
"eval_1_f1": 0.8262425447316105,
"eval_1_precision": 0.8886418063633253,
"eval_1_recall": 0.772031505424283,
"eval_accuracy": 0.7608624274926125,
"eval_loss": 0.5966796875,
"eval_runtime": 467.0404,
"eval_samples_per_second": 19.564,
"eval_steps_per_second": 3.261,
"step": 10662
},
{
"epoch": 3.0022509848058525,
"grad_norm": 7.707584508187591,
"learning_rate": 7.823709536307962e-05,
"loss": 0.2622,
"step": 10670
},
{
"epoch": 3.0050647158131683,
"grad_norm": 51.54450574912668,
"learning_rate": 7.801837270341206e-05,
"loss": 0.4806,
"step": 10680
},
{
"epoch": 3.007878446820484,
"grad_norm": 48.99869202794937,
"learning_rate": 7.779965004374453e-05,
"loss": 0.5045,
"step": 10690
},
{
"epoch": 3.0106921778277997,
"grad_norm": 18.76026822260351,
"learning_rate": 7.758092738407699e-05,
"loss": 0.5507,
"step": 10700
},
{
"epoch": 3.0135059088351155,
"grad_norm": 53.4574294020042,
"learning_rate": 7.736220472440943e-05,
"loss": 0.3936,
"step": 10710
},
{
"epoch": 3.0163196398424312,
"grad_norm": 126.24362236004032,
"learning_rate": 7.71434820647419e-05,
"loss": 0.3933,
"step": 10720
},
{
"epoch": 3.019133370849747,
"grad_norm": 124.38545215336664,
"learning_rate": 7.692475940507437e-05,
"loss": 0.6389,
"step": 10730
},
{
"epoch": 3.0219471018570623,
"grad_norm": 55.57968201814324,
"learning_rate": 7.670603674540681e-05,
"loss": 0.5718,
"step": 10740
},
{
"epoch": 3.024760832864378,
"grad_norm": 48.963769100707765,
"learning_rate": 7.648731408573927e-05,
"loss": 0.6291,
"step": 10750
},
{
"epoch": 3.0275745638716938,
"grad_norm": 12.119240877657461,
"learning_rate": 7.626859142607174e-05,
"loss": 0.5235,
"step": 10760
},
{
"epoch": 3.0303882948790095,
"grad_norm": 20.915222819776293,
"learning_rate": 7.604986876640418e-05,
"loss": 0.4706,
"step": 10770
},
{
"epoch": 3.0332020258863253,
"grad_norm": 9.102773937759299,
"learning_rate": 7.583114610673665e-05,
"loss": 0.232,
"step": 10780
},
{
"epoch": 3.036015756893641,
"grad_norm": 10.249613894191867,
"learning_rate": 7.561242344706911e-05,
"loss": 0.2995,
"step": 10790
},
{
"epoch": 3.0388294879009567,
"grad_norm": 67.40961792746334,
"learning_rate": 7.539370078740157e-05,
"loss": 0.4895,
"step": 10800
},
{
"epoch": 3.0416432189082725,
"grad_norm": 6.032355759360925,
"learning_rate": 7.517497812773402e-05,
"loss": 0.446,
"step": 10810
},
{
"epoch": 3.0444569499155882,
"grad_norm": 86.16554645668533,
"learning_rate": 7.495625546806648e-05,
"loss": 0.5506,
"step": 10820
},
{
"epoch": 3.047270680922904,
"grad_norm": 41.082998364664704,
"learning_rate": 7.473753280839895e-05,
"loss": 0.4396,
"step": 10830
},
{
"epoch": 3.0500844119302193,
"grad_norm": 22.994047192754973,
"learning_rate": 7.45188101487314e-05,
"loss": 0.4426,
"step": 10840
},
{
"epoch": 3.052898142937535,
"grad_norm": 52.522206777883255,
"learning_rate": 7.430008748906386e-05,
"loss": 0.1551,
"step": 10850
},
{
"epoch": 3.0557118739448508,
"grad_norm": 27.72295078584995,
"learning_rate": 7.408136482939632e-05,
"loss": 0.4259,
"step": 10860
},
{
"epoch": 3.0585256049521665,
"grad_norm": 2.357706254274233,
"learning_rate": 7.386264216972878e-05,
"loss": 0.4119,
"step": 10870
},
{
"epoch": 3.0613393359594823,
"grad_norm": 6.85515022950724,
"learning_rate": 7.364391951006125e-05,
"loss": 0.5523,
"step": 10880
},
{
"epoch": 3.064153066966798,
"grad_norm": 34.30181906133321,
"learning_rate": 7.342519685039369e-05,
"loss": 0.2721,
"step": 10890
},
{
"epoch": 3.0669667979741138,
"grad_norm": 54.82144297390585,
"learning_rate": 7.320647419072614e-05,
"loss": 0.9236,
"step": 10900
},
{
"epoch": 3.0697805289814295,
"grad_norm": 129.72620772003393,
"learning_rate": 7.298775153105861e-05,
"loss": 0.5392,
"step": 10910
},
{
"epoch": 3.0725942599887452,
"grad_norm": 3.2339209805746716,
"learning_rate": 7.276902887139107e-05,
"loss": 0.488,
"step": 10920
},
{
"epoch": 3.0754079909960605,
"grad_norm": 10.02255512138656,
"learning_rate": 7.255030621172353e-05,
"loss": 0.5758,
"step": 10930
},
{
"epoch": 3.0782217220033763,
"grad_norm": 18.329541476019806,
"learning_rate": 7.233158355205598e-05,
"loss": 0.4897,
"step": 10940
},
{
"epoch": 3.081035453010692,
"grad_norm": 24.650839029351474,
"learning_rate": 7.211286089238844e-05,
"loss": 0.6129,
"step": 10950
},
{
"epoch": 3.083849184018008,
"grad_norm": 3.945394920205831,
"learning_rate": 7.189413823272091e-05,
"loss": 0.2788,
"step": 10960
},
{
"epoch": 3.0866629150253235,
"grad_norm": 8.209532211869098,
"learning_rate": 7.167541557305337e-05,
"loss": 0.6177,
"step": 10970
},
{
"epoch": 3.0894766460326393,
"grad_norm": 42.404772064384424,
"learning_rate": 7.145669291338582e-05,
"loss": 0.5027,
"step": 10980
},
{
"epoch": 3.092290377039955,
"grad_norm": 71.08207219724257,
"learning_rate": 7.123797025371828e-05,
"loss": 0.4492,
"step": 10990
},
{
"epoch": 3.0951041080472708,
"grad_norm": 24.630629898005367,
"learning_rate": 7.101924759405074e-05,
"loss": 0.6953,
"step": 11000
},
{
"epoch": 3.0979178390545865,
"grad_norm": 28.59624153496924,
"learning_rate": 7.080052493438319e-05,
"loss": 0.4651,
"step": 11010
},
{
"epoch": 3.1007315700619023,
"grad_norm": 10.750146726227943,
"learning_rate": 7.058180227471566e-05,
"loss": 0.4746,
"step": 11020
},
{
"epoch": 3.103545301069218,
"grad_norm": 1.182063914178294,
"learning_rate": 7.036307961504812e-05,
"loss": 0.3329,
"step": 11030
},
{
"epoch": 3.1063590320765333,
"grad_norm": 60.111489378012585,
"learning_rate": 7.014435695538056e-05,
"loss": 0.5148,
"step": 11040
},
{
"epoch": 3.109172763083849,
"grad_norm": 9.162581119984912,
"learning_rate": 6.992563429571303e-05,
"loss": 0.5404,
"step": 11050
},
{
"epoch": 3.111986494091165,
"grad_norm": 630.9338889817419,
"learning_rate": 6.970691163604549e-05,
"loss": 0.454,
"step": 11060
},
{
"epoch": 3.1148002250984805,
"grad_norm": 423.243815972294,
"learning_rate": 6.948818897637794e-05,
"loss": 0.6842,
"step": 11070
},
{
"epoch": 3.1176139561057963,
"grad_norm": 11.73241000732919,
"learning_rate": 6.92694663167104e-05,
"loss": 0.1829,
"step": 11080
},
{
"epoch": 3.120427687113112,
"grad_norm": 13.659308573313247,
"learning_rate": 6.905074365704286e-05,
"loss": 0.6033,
"step": 11090
},
{
"epoch": 3.1232414181204278,
"grad_norm": 4.2358714754973805,
"learning_rate": 6.883202099737533e-05,
"loss": 0.589,
"step": 11100
},
{
"epoch": 3.1260551491277435,
"grad_norm": 0.783313539057735,
"learning_rate": 6.861329833770778e-05,
"loss": 0.2935,
"step": 11110
},
{
"epoch": 3.1288688801350593,
"grad_norm": 66.7036473639581,
"learning_rate": 6.839457567804024e-05,
"loss": 0.7288,
"step": 11120
},
{
"epoch": 3.1316826111423746,
"grad_norm": 49.44587487081624,
"learning_rate": 6.81758530183727e-05,
"loss": 0.5051,
"step": 11130
},
{
"epoch": 3.1344963421496903,
"grad_norm": 82.08430399312728,
"learning_rate": 6.795713035870515e-05,
"loss": 0.7624,
"step": 11140
},
{
"epoch": 3.137310073157006,
"grad_norm": 41.54423962771268,
"learning_rate": 6.773840769903761e-05,
"loss": 0.6681,
"step": 11150
},
{
"epoch": 3.140123804164322,
"grad_norm": 6.054984593639571,
"learning_rate": 6.751968503937008e-05,
"loss": 0.4486,
"step": 11160
},
{
"epoch": 3.1429375351716375,
"grad_norm": 42.03326946639902,
"learning_rate": 6.730096237970254e-05,
"loss": 0.3373,
"step": 11170
},
{
"epoch": 3.1457512661789533,
"grad_norm": 61.9766842778273,
"learning_rate": 6.708223972003498e-05,
"loss": 0.8549,
"step": 11180
},
{
"epoch": 3.148564997186269,
"grad_norm": 31.696383718846477,
"learning_rate": 6.686351706036745e-05,
"loss": 0.8351,
"step": 11190
},
{
"epoch": 3.1513787281935848,
"grad_norm": 40.35191568698288,
"learning_rate": 6.66447944006999e-05,
"loss": 0.3437,
"step": 11200
},
{
"epoch": 3.1541924592009005,
"grad_norm": 10.679496275076508,
"learning_rate": 6.642607174103236e-05,
"loss": 0.4696,
"step": 11210
},
{
"epoch": 3.1570061902082163,
"grad_norm": 15.537348649384192,
"learning_rate": 6.620734908136482e-05,
"loss": 0.4651,
"step": 11220
},
{
"epoch": 3.159819921215532,
"grad_norm": 1.5590651269600222,
"learning_rate": 6.598862642169728e-05,
"loss": 0.355,
"step": 11230
},
{
"epoch": 3.1626336522228473,
"grad_norm": 46.340907086190306,
"learning_rate": 6.576990376202975e-05,
"loss": 0.3964,
"step": 11240
},
{
"epoch": 3.165447383230163,
"grad_norm": 62.12320109474248,
"learning_rate": 6.55511811023622e-05,
"loss": 0.6063,
"step": 11250
},
{
"epoch": 3.168261114237479,
"grad_norm": 19.69816239773773,
"learning_rate": 6.533245844269466e-05,
"loss": 0.3146,
"step": 11260
},
{
"epoch": 3.1710748452447945,
"grad_norm": 73.39996557832582,
"learning_rate": 6.511373578302711e-05,
"loss": 0.5348,
"step": 11270
},
{
"epoch": 3.1738885762521103,
"grad_norm": 26.160849500666586,
"learning_rate": 6.489501312335957e-05,
"loss": 0.2218,
"step": 11280
},
{
"epoch": 3.176702307259426,
"grad_norm": 11.032717942104254,
"learning_rate": 6.467629046369203e-05,
"loss": 0.4841,
"step": 11290
},
{
"epoch": 3.179516038266742,
"grad_norm": 22.122247659427618,
"learning_rate": 6.44575678040245e-05,
"loss": 0.3955,
"step": 11300
},
{
"epoch": 3.1823297692740575,
"grad_norm": 17.431813666502595,
"learning_rate": 6.423884514435695e-05,
"loss": 0.4135,
"step": 11310
},
{
"epoch": 3.1851435002813733,
"grad_norm": 20.230459201101173,
"learning_rate": 6.402012248468941e-05,
"loss": 0.3017,
"step": 11320
},
{
"epoch": 3.1879572312886886,
"grad_norm": 0.40726803235691345,
"learning_rate": 6.380139982502187e-05,
"loss": 0.5075,
"step": 11330
},
{
"epoch": 3.1907709622960043,
"grad_norm": 15.28283361018702,
"learning_rate": 6.358267716535432e-05,
"loss": 0.6175,
"step": 11340
},
{
"epoch": 3.19358469330332,
"grad_norm": 15.820899507911468,
"learning_rate": 6.33639545056868e-05,
"loss": 0.7891,
"step": 11350
},
{
"epoch": 3.196398424310636,
"grad_norm": 55.444795130680475,
"learning_rate": 6.314523184601924e-05,
"loss": 0.5612,
"step": 11360
},
{
"epoch": 3.1992121553179516,
"grad_norm": 32.00507189372659,
"learning_rate": 6.292650918635169e-05,
"loss": 0.3554,
"step": 11370
},
{
"epoch": 3.2020258863252673,
"grad_norm": 2.0628646280824503,
"learning_rate": 6.270778652668416e-05,
"loss": 0.4491,
"step": 11380
},
{
"epoch": 3.204839617332583,
"grad_norm": 29.489592086070243,
"learning_rate": 6.248906386701662e-05,
"loss": 0.7655,
"step": 11390
},
{
"epoch": 3.207653348339899,
"grad_norm": 1.6061570305998563,
"learning_rate": 6.227034120734908e-05,
"loss": 0.3726,
"step": 11400
},
{
"epoch": 3.2104670793472145,
"grad_norm": 17.153820015143744,
"learning_rate": 6.205161854768153e-05,
"loss": 0.63,
"step": 11410
},
{
"epoch": 3.2132808103545303,
"grad_norm": 12.877483190952468,
"learning_rate": 6.183289588801399e-05,
"loss": 0.2199,
"step": 11420
},
{
"epoch": 3.216094541361846,
"grad_norm": 284.41612018123254,
"learning_rate": 6.161417322834645e-05,
"loss": 0.7118,
"step": 11430
},
{
"epoch": 3.2189082723691613,
"grad_norm": 88.85724386333004,
"learning_rate": 6.139545056867892e-05,
"loss": 0.6572,
"step": 11440
},
{
"epoch": 3.221722003376477,
"grad_norm": 116.43335458089302,
"learning_rate": 6.119860017497812e-05,
"loss": 0.3925,
"step": 11450
},
{
"epoch": 3.224535734383793,
"grad_norm": 9.34218826766074,
"learning_rate": 6.0979877515310585e-05,
"loss": 0.4467,
"step": 11460
},
{
"epoch": 3.2273494653911086,
"grad_norm": 5.47990408045989,
"learning_rate": 6.0761154855643035e-05,
"loss": 0.6459,
"step": 11470
},
{
"epoch": 3.2301631963984243,
"grad_norm": 12.032870993467688,
"learning_rate": 6.0542432195975505e-05,
"loss": 0.4638,
"step": 11480
},
{
"epoch": 3.23297692740574,
"grad_norm": 10.978821779199087,
"learning_rate": 6.0323709536307955e-05,
"loss": 0.4797,
"step": 11490
},
{
"epoch": 3.235790658413056,
"grad_norm": 26.344873524302695,
"learning_rate": 6.010498687664041e-05,
"loss": 0.4618,
"step": 11500
},
{
"epoch": 3.2386043894203715,
"grad_norm": 36.71335853765133,
"learning_rate": 5.9886264216972874e-05,
"loss": 0.4282,
"step": 11510
},
{
"epoch": 3.2414181204276873,
"grad_norm": 64.57144789900413,
"learning_rate": 5.966754155730533e-05,
"loss": 0.3277,
"step": 11520
},
{
"epoch": 3.2442318514350026,
"grad_norm": 1.4007879828021592,
"learning_rate": 5.944881889763779e-05,
"loss": 0.4915,
"step": 11530
},
{
"epoch": 3.2470455824423183,
"grad_norm": 3.33308990516012,
"learning_rate": 5.923009623797025e-05,
"loss": 0.6223,
"step": 11540
},
{
"epoch": 3.249859313449634,
"grad_norm": 13.935577439593432,
"learning_rate": 5.901137357830271e-05,
"loss": 0.2424,
"step": 11550
},
{
"epoch": 3.25267304445695,
"grad_norm": 31.282567929182168,
"learning_rate": 5.879265091863516e-05,
"loss": 0.4164,
"step": 11560
},
{
"epoch": 3.2554867754642656,
"grad_norm": 172.63519093501742,
"learning_rate": 5.8573928258967627e-05,
"loss": 0.6778,
"step": 11570
},
{
"epoch": 3.2583005064715813,
"grad_norm": 97.99842872138487,
"learning_rate": 5.835520559930008e-05,
"loss": 0.8454,
"step": 11580
},
{
"epoch": 3.261114237478897,
"grad_norm": 35.317874766103294,
"learning_rate": 5.8136482939632546e-05,
"loss": 0.695,
"step": 11590
},
{
"epoch": 3.263927968486213,
"grad_norm": 16.91213310108752,
"learning_rate": 5.793963254593175e-05,
"loss": 0.484,
"step": 11600
},
{
"epoch": 3.2667416994935286,
"grad_norm": 128.9963496379245,
"learning_rate": 5.772090988626421e-05,
"loss": 0.4257,
"step": 11610
},
{
"epoch": 3.2695554305008443,
"grad_norm": 193.3853393236727,
"learning_rate": 5.750218722659667e-05,
"loss": 0.5854,
"step": 11620
},
{
"epoch": 3.27236916150816,
"grad_norm": 35.31919731163349,
"learning_rate": 5.728346456692913e-05,
"loss": 0.9169,
"step": 11630
},
{
"epoch": 3.2751828925154753,
"grad_norm": 156.87027960130746,
"learning_rate": 5.706474190726159e-05,
"loss": 0.7004,
"step": 11640
},
{
"epoch": 3.277996623522791,
"grad_norm": 23.682213809912607,
"learning_rate": 5.684601924759405e-05,
"loss": 0.3195,
"step": 11650
},
{
"epoch": 3.280810354530107,
"grad_norm": 132.4746326275145,
"learning_rate": 5.66272965879265e-05,
"loss": 0.6326,
"step": 11660
},
{
"epoch": 3.2836240855374226,
"grad_norm": 63.40352213008167,
"learning_rate": 5.640857392825897e-05,
"loss": 0.347,
"step": 11670
},
{
"epoch": 3.2864378165447383,
"grad_norm": 9.625739657480374,
"learning_rate": 5.618985126859142e-05,
"loss": 0.7811,
"step": 11680
},
{
"epoch": 3.289251547552054,
"grad_norm": 13.280340033380412,
"learning_rate": 5.5971128608923875e-05,
"loss": 0.8523,
"step": 11690
},
{
"epoch": 3.29206527855937,
"grad_norm": 13.726951029125418,
"learning_rate": 5.575240594925634e-05,
"loss": 0.5483,
"step": 11700
},
{
"epoch": 3.2948790095666856,
"grad_norm": 3.853407952070311,
"learning_rate": 5.5533683289588794e-05,
"loss": 0.635,
"step": 11710
},
{
"epoch": 3.2976927405740013,
"grad_norm": 32.63263843171223,
"learning_rate": 5.531496062992125e-05,
"loss": 0.2759,
"step": 11720
},
{
"epoch": 3.3005064715813166,
"grad_norm": 17.37594206746597,
"learning_rate": 5.5096237970253714e-05,
"loss": 0.3167,
"step": 11730
},
{
"epoch": 3.3033202025886323,
"grad_norm": 10.92450992851185,
"learning_rate": 5.487751531058617e-05,
"loss": 0.2056,
"step": 11740
},
{
"epoch": 3.306133933595948,
"grad_norm": 18.621614545321687,
"learning_rate": 5.4658792650918634e-05,
"loss": 0.2418,
"step": 11750
},
{
"epoch": 3.308947664603264,
"grad_norm": 365.3094150018103,
"learning_rate": 5.444006999125109e-05,
"loss": 0.8144,
"step": 11760
},
{
"epoch": 3.3117613956105796,
"grad_norm": 0.31282059170952675,
"learning_rate": 5.422134733158355e-05,
"loss": 0.5051,
"step": 11770
},
{
"epoch": 3.3145751266178953,
"grad_norm": 34.65838459619827,
"learning_rate": 5.400262467191601e-05,
"loss": 1.258,
"step": 11780
},
{
"epoch": 3.317388857625211,
"grad_norm": 62.48964602346488,
"learning_rate": 5.380577427821522e-05,
"loss": 1.1,
"step": 11790
},
{
"epoch": 3.320202588632527,
"grad_norm": 126.25946649997921,
"learning_rate": 5.358705161854768e-05,
"loss": 0.4886,
"step": 11800
},
{
"epoch": 3.3230163196398426,
"grad_norm": 94.42597920438025,
"learning_rate": 5.3368328958880136e-05,
"loss": 0.6764,
"step": 11810
},
{
"epoch": 3.3258300506471583,
"grad_norm": 29.13735782010915,
"learning_rate": 5.314960629921259e-05,
"loss": 1.008,
"step": 11820
},
{
"epoch": 3.328643781654474,
"grad_norm": 8.634773437316992,
"learning_rate": 5.2930883639545056e-05,
"loss": 0.2147,
"step": 11830
},
{
"epoch": 3.3314575126617894,
"grad_norm": 14.39857109864753,
"learning_rate": 5.271216097987751e-05,
"loss": 0.4097,
"step": 11840
},
{
"epoch": 3.334271243669105,
"grad_norm": 23.035243755638188,
"learning_rate": 5.249343832020997e-05,
"loss": 0.6312,
"step": 11850
},
{
"epoch": 3.337084974676421,
"grad_norm": 43.17451612005898,
"learning_rate": 5.227471566054243e-05,
"loss": 0.7843,
"step": 11860
},
{
"epoch": 3.3398987056837366,
"grad_norm": 15.873553878518269,
"learning_rate": 5.205599300087488e-05,
"loss": 0.3799,
"step": 11870
},
{
"epoch": 3.3427124366910523,
"grad_norm": 5.309526682904749,
"learning_rate": 5.183727034120735e-05,
"loss": 0.2318,
"step": 11880
},
{
"epoch": 3.345526167698368,
"grad_norm": 1.3966056606002777,
"learning_rate": 5.16185476815398e-05,
"loss": 0.3297,
"step": 11890
},
{
"epoch": 3.348339898705684,
"grad_norm": 3.8353966809516478,
"learning_rate": 5.139982502187226e-05,
"loss": 0.2181,
"step": 11900
},
{
"epoch": 3.3511536297129996,
"grad_norm": 32.67366702302119,
"learning_rate": 5.118110236220472e-05,
"loss": 0.7795,
"step": 11910
},
{
"epoch": 3.3539673607203153,
"grad_norm": 9.654076908853929,
"learning_rate": 5.096237970253718e-05,
"loss": 0.9812,
"step": 11920
},
{
"epoch": 3.3567810917276306,
"grad_norm": 10.000700109531257,
"learning_rate": 5.0743657042869634e-05,
"loss": 0.4223,
"step": 11930
},
{
"epoch": 3.3595948227349464,
"grad_norm": 34.13418422325413,
"learning_rate": 5.05249343832021e-05,
"loss": 0.4949,
"step": 11940
},
{
"epoch": 3.362408553742262,
"grad_norm": 1.0985807347140457,
"learning_rate": 5.0306211723534554e-05,
"loss": 0.4653,
"step": 11950
},
{
"epoch": 3.365222284749578,
"grad_norm": 168.6850360069934,
"learning_rate": 5.008748906386701e-05,
"loss": 0.6093,
"step": 11960
},
{
"epoch": 3.3680360157568936,
"grad_norm": 12.961068610872767,
"learning_rate": 4.9868766404199474e-05,
"loss": 0.5953,
"step": 11970
},
{
"epoch": 3.3708497467642093,
"grad_norm": 18.333361541841942,
"learning_rate": 4.965004374453193e-05,
"loss": 0.3427,
"step": 11980
},
{
"epoch": 3.373663477771525,
"grad_norm": 15.018142235150822,
"learning_rate": 4.9431321084864386e-05,
"loss": 0.4499,
"step": 11990
},
{
"epoch": 3.376477208778841,
"grad_norm": 58.35352085477518,
"learning_rate": 4.921259842519685e-05,
"loss": 0.2707,
"step": 12000
},
{
"epoch": 3.3792909397861566,
"grad_norm": 27.84061313602778,
"learning_rate": 4.8993875765529306e-05,
"loss": 0.4568,
"step": 12010
},
{
"epoch": 3.3821046707934723,
"grad_norm": 2.221696017278666,
"learning_rate": 4.877515310586177e-05,
"loss": 0.286,
"step": 12020
},
{
"epoch": 3.384918401800788,
"grad_norm": 5.977444422857166,
"learning_rate": 4.855643044619422e-05,
"loss": 0.5759,
"step": 12030
},
{
"epoch": 3.3877321328081034,
"grad_norm": 1.1311358791589952,
"learning_rate": 4.8337707786526676e-05,
"loss": 0.5304,
"step": 12040
},
{
"epoch": 3.390545863815419,
"grad_norm": 16.413270716064826,
"learning_rate": 4.811898512685914e-05,
"loss": 0.608,
"step": 12050
},
{
"epoch": 3.393359594822735,
"grad_norm": 76.93565566008341,
"learning_rate": 4.7900262467191595e-05,
"loss": 0.4058,
"step": 12060
},
{
"epoch": 3.3961733258300506,
"grad_norm": 106.30305951256041,
"learning_rate": 4.768153980752405e-05,
"loss": 0.3392,
"step": 12070
},
{
"epoch": 3.3989870568373663,
"grad_norm": 94.06687107005396,
"learning_rate": 4.7462817147856515e-05,
"loss": 0.5494,
"step": 12080
},
{
"epoch": 3.401800787844682,
"grad_norm": 25.01577867392826,
"learning_rate": 4.724409448818897e-05,
"loss": 0.5204,
"step": 12090
},
{
"epoch": 3.404614518851998,
"grad_norm": 4.99525052635641,
"learning_rate": 4.702537182852143e-05,
"loss": 0.4441,
"step": 12100
},
{
"epoch": 3.4074282498593136,
"grad_norm": 16.108946786215625,
"learning_rate": 4.680664916885389e-05,
"loss": 0.3515,
"step": 12110
},
{
"epoch": 3.4102419808666293,
"grad_norm": 99.7096719318533,
"learning_rate": 4.658792650918635e-05,
"loss": 0.4423,
"step": 12120
},
{
"epoch": 3.4130557118739446,
"grad_norm": 85.07580217985975,
"learning_rate": 4.636920384951881e-05,
"loss": 0.7414,
"step": 12130
},
{
"epoch": 3.4158694428812604,
"grad_norm": 81.73261731795073,
"learning_rate": 4.615048118985127e-05,
"loss": 0.6783,
"step": 12140
},
{
"epoch": 3.418683173888576,
"grad_norm": 1172.4065465322012,
"learning_rate": 4.5931758530183724e-05,
"loss": 0.68,
"step": 12150
},
{
"epoch": 3.421496904895892,
"grad_norm": 57.09784884634665,
"learning_rate": 4.571303587051619e-05,
"loss": 0.6683,
"step": 12160
},
{
"epoch": 3.4243106359032076,
"grad_norm": 30.70562258358342,
"learning_rate": 4.549431321084864e-05,
"loss": 0.5569,
"step": 12170
},
{
"epoch": 3.4271243669105234,
"grad_norm": 3.4088793578308696,
"learning_rate": 4.527559055118109e-05,
"loss": 0.6287,
"step": 12180
},
{
"epoch": 3.429938097917839,
"grad_norm": 4.84948894552461,
"learning_rate": 4.5056867891513556e-05,
"loss": 0.4114,
"step": 12190
},
{
"epoch": 3.432751828925155,
"grad_norm": 1.1825455646656198,
"learning_rate": 4.483814523184601e-05,
"loss": 1.039,
"step": 12200
},
{
"epoch": 3.4355655599324706,
"grad_norm": 16.863474594043463,
"learning_rate": 4.461942257217847e-05,
"loss": 0.6146,
"step": 12210
},
{
"epoch": 3.4383792909397863,
"grad_norm": 3.912209944210823,
"learning_rate": 4.440069991251093e-05,
"loss": 0.2781,
"step": 12220
},
{
"epoch": 3.441193021947102,
"grad_norm": 161.80969730924826,
"learning_rate": 4.418197725284339e-05,
"loss": 0.5749,
"step": 12230
},
{
"epoch": 3.4440067529544174,
"grad_norm": 10.500387209468151,
"learning_rate": 4.3963254593175845e-05,
"loss": 0.5682,
"step": 12240
},
{
"epoch": 3.446820483961733,
"grad_norm": 84.92953757043959,
"learning_rate": 4.374453193350831e-05,
"loss": 0.4613,
"step": 12250
},
{
"epoch": 3.449634214969049,
"grad_norm": 5.437611277992704,
"learning_rate": 4.3525809273840765e-05,
"loss": 0.7144,
"step": 12260
},
{
"epoch": 3.4524479459763646,
"grad_norm": 28.101996466451755,
"learning_rate": 4.330708661417323e-05,
"loss": 0.5975,
"step": 12270
},
{
"epoch": 3.4552616769836804,
"grad_norm": 57.3104056471748,
"learning_rate": 4.3088363954505685e-05,
"loss": 0.6467,
"step": 12280
},
{
"epoch": 3.458075407990996,
"grad_norm": 31.907644355337986,
"learning_rate": 4.286964129483814e-05,
"loss": 0.3168,
"step": 12290
},
{
"epoch": 3.460889138998312,
"grad_norm": 121.91686564959656,
"learning_rate": 4.2650918635170604e-05,
"loss": 0.8983,
"step": 12300
},
{
"epoch": 3.4637028700056276,
"grad_norm": 44.536124621836905,
"learning_rate": 4.243219597550306e-05,
"loss": 0.6855,
"step": 12310
},
{
"epoch": 3.4665166010129433,
"grad_norm": 6.7326531100755505,
"learning_rate": 4.221347331583551e-05,
"loss": 0.2299,
"step": 12320
},
{
"epoch": 3.4693303320202586,
"grad_norm": 15.971843953513892,
"learning_rate": 4.199475065616798e-05,
"loss": 0.5481,
"step": 12330
},
{
"epoch": 3.4721440630275744,
"grad_norm": 17.685542251021793,
"learning_rate": 4.177602799650043e-05,
"loss": 0.3566,
"step": 12340
},
{
"epoch": 3.47495779403489,
"grad_norm": 0.7679243984887517,
"learning_rate": 4.155730533683289e-05,
"loss": 0.929,
"step": 12350
},
{
"epoch": 3.477771525042206,
"grad_norm": 35.31835874044769,
"learning_rate": 4.133858267716535e-05,
"loss": 0.3099,
"step": 12360
},
{
"epoch": 3.4805852560495216,
"grad_norm": 2.1302391276294474,
"learning_rate": 4.1119860017497806e-05,
"loss": 0.7195,
"step": 12370
},
{
"epoch": 3.4833989870568374,
"grad_norm": 30.66376411827359,
"learning_rate": 4.090113735783027e-05,
"loss": 0.5912,
"step": 12380
},
{
"epoch": 3.486212718064153,
"grad_norm": 2.7815220772073475,
"learning_rate": 4.0682414698162726e-05,
"loss": 0.311,
"step": 12390
},
{
"epoch": 3.489026449071469,
"grad_norm": 10.446918971739123,
"learning_rate": 4.046369203849518e-05,
"loss": 0.3615,
"step": 12400
},
{
"epoch": 3.4918401800787846,
"grad_norm": 4.63253449800804,
"learning_rate": 4.0244969378827646e-05,
"loss": 0.6542,
"step": 12410
},
{
"epoch": 3.4946539110861004,
"grad_norm": 29.433364869299208,
"learning_rate": 4.00262467191601e-05,
"loss": 0.4585,
"step": 12420
},
{
"epoch": 3.497467642093416,
"grad_norm": 49.618623837965174,
"learning_rate": 3.980752405949256e-05,
"loss": 0.5086,
"step": 12430
},
{
"epoch": 3.500281373100732,
"grad_norm": 1.2525886950971519,
"learning_rate": 3.958880139982502e-05,
"loss": 0.2,
"step": 12440
},
{
"epoch": 3.503095104108047,
"grad_norm": 35.560120749370476,
"learning_rate": 3.937007874015748e-05,
"loss": 0.9585,
"step": 12450
},
{
"epoch": 3.505908835115363,
"grad_norm": 76.4695471070044,
"learning_rate": 3.9151356080489935e-05,
"loss": 0.6961,
"step": 12460
},
{
"epoch": 3.5087225661226786,
"grad_norm": 17.129864527344232,
"learning_rate": 3.89326334208224e-05,
"loss": 0.3217,
"step": 12470
},
{
"epoch": 3.5115362971299944,
"grad_norm": 5.729130422882109,
"learning_rate": 3.871391076115485e-05,
"loss": 0.7064,
"step": 12480
},
{
"epoch": 3.51435002813731,
"grad_norm": 235.83586490561873,
"learning_rate": 3.849518810148732e-05,
"loss": 0.3649,
"step": 12490
},
{
"epoch": 3.517163759144626,
"grad_norm": 2.2017595700585457,
"learning_rate": 3.827646544181977e-05,
"loss": 0.4434,
"step": 12500
},
{
"epoch": 3.5199774901519416,
"grad_norm": 6.576317534322007,
"learning_rate": 3.8057742782152224e-05,
"loss": 0.5095,
"step": 12510
},
{
"epoch": 3.522791221159257,
"grad_norm": 1.6164548073339011,
"learning_rate": 3.783902012248469e-05,
"loss": 0.6829,
"step": 12520
},
{
"epoch": 3.5256049521665727,
"grad_norm": 3.480112918705946,
"learning_rate": 3.7620297462817144e-05,
"loss": 0.9486,
"step": 12530
},
{
"epoch": 3.5284186831738884,
"grad_norm": 92.6860952958962,
"learning_rate": 3.740157480314961e-05,
"loss": 0.4618,
"step": 12540
},
{
"epoch": 3.531232414181204,
"grad_norm": 1.8127678229329505,
"learning_rate": 3.718285214348206e-05,
"loss": 0.5044,
"step": 12550
},
{
"epoch": 3.53404614518852,
"grad_norm": 11.49210167806751,
"learning_rate": 3.696412948381452e-05,
"loss": 0.7462,
"step": 12560
},
{
"epoch": 3.5368598761958356,
"grad_norm": 22.188651197414536,
"learning_rate": 3.6745406824146976e-05,
"loss": 0.3461,
"step": 12570
},
{
"epoch": 3.5396736072031514,
"grad_norm": 31.842658386387633,
"learning_rate": 3.652668416447944e-05,
"loss": 0.3388,
"step": 12580
},
{
"epoch": 3.542487338210467,
"grad_norm": 203.1349459704412,
"learning_rate": 3.6307961504811896e-05,
"loss": 0.3437,
"step": 12590
},
{
"epoch": 3.545301069217783,
"grad_norm": 10.252277397482308,
"learning_rate": 3.608923884514435e-05,
"loss": 0.1386,
"step": 12600
},
{
"epoch": 3.5481148002250986,
"grad_norm": 355.9201349404873,
"learning_rate": 3.5870516185476816e-05,
"loss": 0.2346,
"step": 12610
},
{
"epoch": 3.5509285312324144,
"grad_norm": 48.7713276890867,
"learning_rate": 3.565179352580927e-05,
"loss": 0.6602,
"step": 12620
},
{
"epoch": 3.55374226223973,
"grad_norm": 0.5124599339952361,
"learning_rate": 3.543307086614173e-05,
"loss": 0.4634,
"step": 12630
},
{
"epoch": 3.556555993247046,
"grad_norm": 75.75380478169932,
"learning_rate": 3.5214348206474185e-05,
"loss": 0.9834,
"step": 12640
},
{
"epoch": 3.559369724254361,
"grad_norm": 5.9723791827847466,
"learning_rate": 3.499562554680665e-05,
"loss": 0.2721,
"step": 12650
},
{
"epoch": 3.562183455261677,
"grad_norm": 1.339210154695292,
"learning_rate": 3.4776902887139105e-05,
"loss": 0.5833,
"step": 12660
},
{
"epoch": 3.5649971862689926,
"grad_norm": 2.340812474612855,
"learning_rate": 3.455818022747156e-05,
"loss": 0.8534,
"step": 12670
},
{
"epoch": 3.5678109172763084,
"grad_norm": 42.020492946977896,
"learning_rate": 3.4339457567804024e-05,
"loss": 0.4952,
"step": 12680
},
{
"epoch": 3.570624648283624,
"grad_norm": 30.231287046941738,
"learning_rate": 3.412073490813648e-05,
"loss": 0.469,
"step": 12690
},
{
"epoch": 3.57343837929094,
"grad_norm": 20.637898535782263,
"learning_rate": 3.390201224846894e-05,
"loss": 0.5748,
"step": 12700
},
{
"epoch": 3.5762521102982556,
"grad_norm": 27.709789371715573,
"learning_rate": 3.36832895888014e-05,
"loss": 0.5487,
"step": 12710
},
{
"epoch": 3.579065841305571,
"grad_norm": 1.4646876271019715,
"learning_rate": 3.346456692913386e-05,
"loss": 0.2549,
"step": 12720
},
{
"epoch": 3.5818795723128867,
"grad_norm": 47.275662835213524,
"learning_rate": 3.3245844269466313e-05,
"loss": 0.331,
"step": 12730
},
{
"epoch": 3.5846933033202024,
"grad_norm": 135.70744957937237,
"learning_rate": 3.302712160979877e-05,
"loss": 0.5608,
"step": 12740
},
{
"epoch": 3.587507034327518,
"grad_norm": 78.57081420410536,
"learning_rate": 3.280839895013123e-05,
"loss": 0.2891,
"step": 12750
},
{
"epoch": 3.590320765334834,
"grad_norm": 1.7707715411426224,
"learning_rate": 3.258967629046369e-05,
"loss": 0.345,
"step": 12760
},
{
"epoch": 3.5931344963421497,
"grad_norm": 25.509662246815907,
"learning_rate": 3.2370953630796146e-05,
"loss": 0.8202,
"step": 12770
},
{
"epoch": 3.5959482273494654,
"grad_norm": 36.869039202453266,
"learning_rate": 3.215223097112861e-05,
"loss": 0.4163,
"step": 12780
},
{
"epoch": 3.598761958356781,
"grad_norm": 2.193698658946938,
"learning_rate": 3.1933508311461066e-05,
"loss": 0.4813,
"step": 12790
},
{
"epoch": 3.601575689364097,
"grad_norm": 86.82251364819027,
"learning_rate": 3.171478565179352e-05,
"loss": 0.2963,
"step": 12800
},
{
"epoch": 3.6043894203714126,
"grad_norm": 175.32090889794566,
"learning_rate": 3.149606299212598e-05,
"loss": 0.4121,
"step": 12810
},
{
"epoch": 3.6072031513787284,
"grad_norm": 77.92419912338526,
"learning_rate": 3.127734033245844e-05,
"loss": 0.6049,
"step": 12820
},
{
"epoch": 3.610016882386044,
"grad_norm": 29.416631197885483,
"learning_rate": 3.10586176727909e-05,
"loss": 0.4979,
"step": 12830
},
{
"epoch": 3.61283061339336,
"grad_norm": 11.168327597970197,
"learning_rate": 3.0839895013123355e-05,
"loss": 0.7981,
"step": 12840
},
{
"epoch": 3.615644344400675,
"grad_norm": 5.295416735323613,
"learning_rate": 3.062117235345582e-05,
"loss": 0.5611,
"step": 12850
},
{
"epoch": 3.618458075407991,
"grad_norm": 20.607563312252314,
"learning_rate": 3.0402449693788275e-05,
"loss": 0.4421,
"step": 12860
},
{
"epoch": 3.6212718064153067,
"grad_norm": 79.3047250915384,
"learning_rate": 3.018372703412073e-05,
"loss": 0.6417,
"step": 12870
},
{
"epoch": 3.6240855374226224,
"grad_norm": 51.360432585204684,
"learning_rate": 2.996500437445319e-05,
"loss": 0.4794,
"step": 12880
},
{
"epoch": 3.626899268429938,
"grad_norm": 18.193700637933883,
"learning_rate": 2.974628171478565e-05,
"loss": 0.6764,
"step": 12890
},
{
"epoch": 3.629712999437254,
"grad_norm": 13.2158616023827,
"learning_rate": 2.952755905511811e-05,
"loss": 0.561,
"step": 12900
},
{
"epoch": 3.6325267304445696,
"grad_norm": 39.6048512902133,
"learning_rate": 2.9308836395450564e-05,
"loss": 0.5969,
"step": 12910
},
{
"epoch": 3.635340461451885,
"grad_norm": 119.25617194048463,
"learning_rate": 2.9090113735783023e-05,
"loss": 0.5537,
"step": 12920
},
{
"epoch": 3.6381541924592007,
"grad_norm": 17.312325283917904,
"learning_rate": 2.8871391076115483e-05,
"loss": 0.2931,
"step": 12930
},
{
"epoch": 3.6409679234665164,
"grad_norm": 30.668034379631603,
"learning_rate": 2.865266841644794e-05,
"loss": 0.7078,
"step": 12940
},
{
"epoch": 3.643781654473832,
"grad_norm": 45.124842339660304,
"learning_rate": 2.84339457567804e-05,
"loss": 0.5132,
"step": 12950
},
{
"epoch": 3.646595385481148,
"grad_norm": 15.881149948027138,
"learning_rate": 2.821522309711286e-05,
"loss": 0.6237,
"step": 12960
},
{
"epoch": 3.6494091164884637,
"grad_norm": 23.94430535891449,
"learning_rate": 2.799650043744532e-05,
"loss": 0.2053,
"step": 12970
},
{
"epoch": 3.6522228474957794,
"grad_norm": 120.3735140731541,
"learning_rate": 2.7777777777777772e-05,
"loss": 0.3646,
"step": 12980
},
{
"epoch": 3.655036578503095,
"grad_norm": 64.266595500627,
"learning_rate": 2.7559055118110232e-05,
"loss": 0.5375,
"step": 12990
},
{
"epoch": 3.657850309510411,
"grad_norm": 6.0750969402208135,
"learning_rate": 2.7340332458442692e-05,
"loss": 1.0836,
"step": 13000
},
{
"epoch": 3.6606640405177266,
"grad_norm": 74.7234722768726,
"learning_rate": 2.7121609798775152e-05,
"loss": 0.5668,
"step": 13010
},
{
"epoch": 3.6634777715250424,
"grad_norm": 39.41229028715555,
"learning_rate": 2.690288713910761e-05,
"loss": 0.2667,
"step": 13020
},
{
"epoch": 3.666291502532358,
"grad_norm": 3.053090186442824,
"learning_rate": 2.6684164479440068e-05,
"loss": 0.4439,
"step": 13030
},
{
"epoch": 3.669105233539674,
"grad_norm": 71.21754074911294,
"learning_rate": 2.6465441819772528e-05,
"loss": 0.6616,
"step": 13040
},
{
"epoch": 3.671918964546989,
"grad_norm": 45.19877536600059,
"learning_rate": 2.6246719160104984e-05,
"loss": 0.4174,
"step": 13050
},
{
"epoch": 3.674732695554305,
"grad_norm": 372.5488030536916,
"learning_rate": 2.602799650043744e-05,
"loss": 0.5175,
"step": 13060
},
{
"epoch": 3.6775464265616207,
"grad_norm": 57.57663292368472,
"learning_rate": 2.58092738407699e-05,
"loss": 0.8581,
"step": 13070
},
{
"epoch": 3.6803601575689364,
"grad_norm": 1.7083829355501452,
"learning_rate": 2.559055118110236e-05,
"loss": 0.681,
"step": 13080
},
{
"epoch": 3.683173888576252,
"grad_norm": 6.539997476072728,
"learning_rate": 2.5371828521434817e-05,
"loss": 0.1846,
"step": 13090
},
{
"epoch": 3.685987619583568,
"grad_norm": 207.30300737141843,
"learning_rate": 2.5153105861767277e-05,
"loss": 0.3291,
"step": 13100
},
{
"epoch": 3.6888013505908837,
"grad_norm": 83.28340761635474,
"learning_rate": 2.4934383202099737e-05,
"loss": 0.3801,
"step": 13110
},
{
"epoch": 3.691615081598199,
"grad_norm": 29.331313165726307,
"learning_rate": 2.4715660542432193e-05,
"loss": 0.8411,
"step": 13120
},
{
"epoch": 3.6944288126055147,
"grad_norm": 40.83094333479217,
"learning_rate": 2.4496937882764653e-05,
"loss": 0.2844,
"step": 13130
},
{
"epoch": 3.6972425436128304,
"grad_norm": 282.66752780327295,
"learning_rate": 2.427821522309711e-05,
"loss": 0.7839,
"step": 13140
},
{
"epoch": 3.700056274620146,
"grad_norm": 79.40867859040115,
"learning_rate": 2.405949256342957e-05,
"loss": 0.4881,
"step": 13150
},
{
"epoch": 3.702870005627462,
"grad_norm": 14.039173501520008,
"learning_rate": 2.3840769903762026e-05,
"loss": 0.3336,
"step": 13160
},
{
"epoch": 3.7056837366347777,
"grad_norm": 2.66570396278435,
"learning_rate": 2.3622047244094486e-05,
"loss": 0.4861,
"step": 13170
},
{
"epoch": 3.7084974676420934,
"grad_norm": 297.6891719203325,
"learning_rate": 2.3403324584426946e-05,
"loss": 0.5156,
"step": 13180
},
{
"epoch": 3.711311198649409,
"grad_norm": 42.44776195786567,
"learning_rate": 2.3184601924759405e-05,
"loss": 0.8828,
"step": 13190
},
{
"epoch": 3.714124929656725,
"grad_norm": 28.455047894378716,
"learning_rate": 2.2965879265091862e-05,
"loss": 0.8758,
"step": 13200
},
{
"epoch": 3.7169386606640407,
"grad_norm": 33.07750573387209,
"learning_rate": 2.274715660542432e-05,
"loss": 0.5996,
"step": 13210
},
{
"epoch": 3.7197523916713564,
"grad_norm": 0.9938889512329678,
"learning_rate": 2.2528433945756778e-05,
"loss": 0.2031,
"step": 13220
},
{
"epoch": 3.722566122678672,
"grad_norm": 23.80061424167977,
"learning_rate": 2.2309711286089235e-05,
"loss": 0.4923,
"step": 13230
},
{
"epoch": 3.725379853685988,
"grad_norm": 1.131933505943834,
"learning_rate": 2.2090988626421694e-05,
"loss": 0.3908,
"step": 13240
},
{
"epoch": 3.728193584693303,
"grad_norm": 446.7276314043673,
"learning_rate": 2.1872265966754154e-05,
"loss": 0.7402,
"step": 13250
},
{
"epoch": 3.731007315700619,
"grad_norm": 1206.2342773424812,
"learning_rate": 2.1653543307086614e-05,
"loss": 0.2547,
"step": 13260
},
{
"epoch": 3.7338210467079347,
"grad_norm": 1.7651821391803046,
"learning_rate": 2.143482064741907e-05,
"loss": 0.4101,
"step": 13270
},
{
"epoch": 3.7366347777152504,
"grad_norm": 94.1699997933736,
"learning_rate": 2.121609798775153e-05,
"loss": 0.614,
"step": 13280
},
{
"epoch": 3.739448508722566,
"grad_norm": 17.845824880937165,
"learning_rate": 2.099737532808399e-05,
"loss": 0.4902,
"step": 13290
},
{
"epoch": 3.742262239729882,
"grad_norm": 26.896154000409293,
"learning_rate": 2.0778652668416443e-05,
"loss": 0.4603,
"step": 13300
},
{
"epoch": 3.7450759707371977,
"grad_norm": 14.565889324285259,
"learning_rate": 2.0559930008748903e-05,
"loss": 0.5054,
"step": 13310
},
{
"epoch": 3.747889701744513,
"grad_norm": 5.844233357741922,
"learning_rate": 2.0341207349081363e-05,
"loss": 0.4795,
"step": 13320
},
{
"epoch": 3.7507034327518287,
"grad_norm": 206.0939101443417,
"learning_rate": 2.0122484689413823e-05,
"loss": 0.5896,
"step": 13330
},
{
"epoch": 3.7535171637591445,
"grad_norm": 65.3190773750316,
"learning_rate": 1.990376202974628e-05,
"loss": 0.292,
"step": 13340
},
{
"epoch": 3.75633089476646,
"grad_norm": 12.321249888093957,
"learning_rate": 1.968503937007874e-05,
"loss": 0.5816,
"step": 13350
},
{
"epoch": 3.759144625773776,
"grad_norm": 20.741649400875783,
"learning_rate": 1.94663167104112e-05,
"loss": 0.2388,
"step": 13360
},
{
"epoch": 3.7619583567810917,
"grad_norm": 2.622559284102868,
"learning_rate": 1.924759405074366e-05,
"loss": 0.6685,
"step": 13370
},
{
"epoch": 3.7647720877884074,
"grad_norm": 23.160980070983843,
"learning_rate": 1.9028871391076112e-05,
"loss": 1.0501,
"step": 13380
},
{
"epoch": 3.767585818795723,
"grad_norm": 87.1185595021463,
"learning_rate": 1.8810148731408572e-05,
"loss": 0.9589,
"step": 13390
},
{
"epoch": 3.770399549803039,
"grad_norm": 30.316621338983605,
"learning_rate": 1.859142607174103e-05,
"loss": 0.2501,
"step": 13400
},
{
"epoch": 3.7732132808103547,
"grad_norm": 23.92001820439982,
"learning_rate": 1.8372703412073488e-05,
"loss": 0.7739,
"step": 13410
},
{
"epoch": 3.7760270118176704,
"grad_norm": 8.480682588233163,
"learning_rate": 1.8153980752405948e-05,
"loss": 0.4317,
"step": 13420
},
{
"epoch": 3.778840742824986,
"grad_norm": 11.901124707133304,
"learning_rate": 1.7935258092738408e-05,
"loss": 0.2882,
"step": 13430
},
{
"epoch": 3.7816544738323015,
"grad_norm": 121.66628180316688,
"learning_rate": 1.7716535433070864e-05,
"loss": 0.4091,
"step": 13440
},
{
"epoch": 3.784468204839617,
"grad_norm": 23.934650131977175,
"learning_rate": 1.7497812773403324e-05,
"loss": 0.6122,
"step": 13450
},
{
"epoch": 3.787281935846933,
"grad_norm": 68.18789104905484,
"learning_rate": 1.727909011373578e-05,
"loss": 0.5914,
"step": 13460
},
{
"epoch": 3.7900956668542487,
"grad_norm": 35.60052500777,
"learning_rate": 1.706036745406824e-05,
"loss": 0.4227,
"step": 13470
},
{
"epoch": 3.7929093978615644,
"grad_norm": 17.85284058651301,
"learning_rate": 1.68416447944007e-05,
"loss": 0.1204,
"step": 13480
},
{
"epoch": 3.79572312886888,
"grad_norm": 2.4161722414635207,
"learning_rate": 1.6622922134733157e-05,
"loss": 0.5022,
"step": 13490
},
{
"epoch": 3.798536859876196,
"grad_norm": 5.928187046017908,
"learning_rate": 1.6404199475065617e-05,
"loss": 0.6874,
"step": 13500
},
{
"epoch": 3.8013505908835117,
"grad_norm": 63.792368045524945,
"learning_rate": 1.6185476815398073e-05,
"loss": 0.6038,
"step": 13510
},
{
"epoch": 3.804164321890827,
"grad_norm": 13.41769308503903,
"learning_rate": 1.5966754155730533e-05,
"loss": 0.6469,
"step": 13520
},
{
"epoch": 3.8069780528981427,
"grad_norm": 57.530424477841166,
"learning_rate": 1.574803149606299e-05,
"loss": 0.9088,
"step": 13530
},
{
"epoch": 3.8097917839054585,
"grad_norm": 3.33259141334449,
"learning_rate": 1.552930883639545e-05,
"loss": 0.4786,
"step": 13540
},
{
"epoch": 3.812605514912774,
"grad_norm": 119.02146057750649,
"learning_rate": 1.531058617672791e-05,
"loss": 0.6791,
"step": 13550
},
{
"epoch": 3.81541924592009,
"grad_norm": 67.38672458073057,
"learning_rate": 1.5091863517060365e-05,
"loss": 0.4585,
"step": 13560
},
{
"epoch": 3.8182329769274057,
"grad_norm": 104.08605683069419,
"learning_rate": 1.4873140857392825e-05,
"loss": 0.4812,
"step": 13570
},
{
"epoch": 3.8210467079347215,
"grad_norm": 192.5384876170663,
"learning_rate": 1.4654418197725282e-05,
"loss": 0.4261,
"step": 13580
},
{
"epoch": 3.823860438942037,
"grad_norm": 10.79902650951375,
"learning_rate": 1.4435695538057742e-05,
"loss": 0.3543,
"step": 13590
},
{
"epoch": 3.826674169949353,
"grad_norm": 422.3326786781863,
"learning_rate": 1.42169728783902e-05,
"loss": 0.3316,
"step": 13600
},
{
"epoch": 3.8294879009566687,
"grad_norm": 28.877601143071427,
"learning_rate": 1.399825021872266e-05,
"loss": 0.4107,
"step": 13610
},
{
"epoch": 3.8323016319639844,
"grad_norm": 72.5714412046621,
"learning_rate": 1.3779527559055116e-05,
"loss": 0.5007,
"step": 13620
},
{
"epoch": 3.8351153629713,
"grad_norm": 25.755963857052215,
"learning_rate": 1.3560804899387576e-05,
"loss": 0.6402,
"step": 13630
},
{
"epoch": 3.8379290939786155,
"grad_norm": 1.533241984306202,
"learning_rate": 1.3342082239720034e-05,
"loss": 0.6787,
"step": 13640
},
{
"epoch": 3.8407428249859312,
"grad_norm": 2.439663157954251,
"learning_rate": 1.3123359580052492e-05,
"loss": 0.3079,
"step": 13650
},
{
"epoch": 3.843556555993247,
"grad_norm": 32.38387188806373,
"learning_rate": 1.290463692038495e-05,
"loss": 0.7875,
"step": 13660
},
{
"epoch": 3.8463702870005627,
"grad_norm": 7.353540981294603,
"learning_rate": 1.2685914260717409e-05,
"loss": 0.5157,
"step": 13670
},
{
"epoch": 3.8491840180078785,
"grad_norm": 11.82603640490557,
"learning_rate": 1.2467191601049868e-05,
"loss": 0.4921,
"step": 13680
},
{
"epoch": 3.851997749015194,
"grad_norm": 2.7248783614576997,
"learning_rate": 1.2248468941382327e-05,
"loss": 0.437,
"step": 13690
},
{
"epoch": 3.85481148002251,
"grad_norm": 342.82499995014996,
"learning_rate": 1.2029746281714785e-05,
"loss": 0.8291,
"step": 13700
},
{
"epoch": 3.8576252110298257,
"grad_norm": 28.762266291398717,
"learning_rate": 1.1811023622047243e-05,
"loss": 0.7494,
"step": 13710
},
{
"epoch": 3.860438942037141,
"grad_norm": 19.46284732459688,
"learning_rate": 1.1592300962379703e-05,
"loss": 0.3428,
"step": 13720
},
{
"epoch": 3.8632526730444567,
"grad_norm": 11.122833006077931,
"learning_rate": 1.137357830271216e-05,
"loss": 0.8361,
"step": 13730
},
{
"epoch": 3.8660664040517725,
"grad_norm": 12.12059180019161,
"learning_rate": 1.1154855643044617e-05,
"loss": 0.3567,
"step": 13740
},
{
"epoch": 3.8688801350590882,
"grad_norm": 31.30428497306991,
"learning_rate": 1.0936132983377077e-05,
"loss": 0.5901,
"step": 13750
},
{
"epoch": 3.871693866066404,
"grad_norm": 0.702699931126514,
"learning_rate": 1.0717410323709535e-05,
"loss": 0.2538,
"step": 13760
},
{
"epoch": 3.8745075970737197,
"grad_norm": 1.0957739187938698,
"learning_rate": 1.0498687664041995e-05,
"loss": 0.3035,
"step": 13770
},
{
"epoch": 3.8773213280810355,
"grad_norm": 26.16763388816232,
"learning_rate": 1.0279965004374452e-05,
"loss": 1.0008,
"step": 13780
},
{
"epoch": 3.880135059088351,
"grad_norm": 7.266284196954616,
"learning_rate": 1.0061242344706911e-05,
"loss": 0.6148,
"step": 13790
},
{
"epoch": 3.882948790095667,
"grad_norm": 1.280410994175035,
"learning_rate": 9.84251968503937e-06,
"loss": 0.31,
"step": 13800
},
{
"epoch": 3.8857625211029827,
"grad_norm": 11.332894345972404,
"learning_rate": 9.62379702537183e-06,
"loss": 0.3175,
"step": 13810
},
{
"epoch": 3.8885762521102984,
"grad_norm": 2.349285991354639,
"learning_rate": 9.405074365704286e-06,
"loss": 0.5999,
"step": 13820
},
{
"epoch": 3.891389983117614,
"grad_norm": 0.3849237703681145,
"learning_rate": 9.186351706036744e-06,
"loss": 0.5271,
"step": 13830
},
{
"epoch": 3.8942037141249295,
"grad_norm": 2.644331997108355,
"learning_rate": 8.967629046369204e-06,
"loss": 0.5702,
"step": 13840
},
{
"epoch": 3.8970174451322452,
"grad_norm": 159.62015783602538,
"learning_rate": 8.748906386701662e-06,
"loss": 0.6812,
"step": 13850
},
{
"epoch": 3.899831176139561,
"grad_norm": 9.77439075886051,
"learning_rate": 8.53018372703412e-06,
"loss": 0.9331,
"step": 13860
},
{
"epoch": 3.9026449071468767,
"grad_norm": 9.517338120917794,
"learning_rate": 8.311461067366578e-06,
"loss": 0.3774,
"step": 13870
},
{
"epoch": 3.9054586381541925,
"grad_norm": 1.2035309598986415,
"learning_rate": 8.092738407699037e-06,
"loss": 0.4958,
"step": 13880
},
{
"epoch": 3.908272369161508,
"grad_norm": 22.577502570103032,
"learning_rate": 7.874015748031495e-06,
"loss": 0.42,
"step": 13890
},
{
"epoch": 3.911086100168824,
"grad_norm": 10.830002884736738,
"learning_rate": 7.655293088363955e-06,
"loss": 0.4453,
"step": 13900
},
{
"epoch": 3.9138998311761397,
"grad_norm": 186.4903752602189,
"learning_rate": 7.436570428696413e-06,
"loss": 0.5316,
"step": 13910
},
{
"epoch": 3.916713562183455,
"grad_norm": 255.53257946868905,
"learning_rate": 7.217847769028871e-06,
"loss": 0.4243,
"step": 13920
},
{
"epoch": 3.9195272931907708,
"grad_norm": 66.1440334798085,
"learning_rate": 6.99912510936133e-06,
"loss": 0.5682,
"step": 13930
},
{
"epoch": 3.9223410241980865,
"grad_norm": 22.65274232742973,
"learning_rate": 6.780402449693788e-06,
"loss": 0.4842,
"step": 13940
},
{
"epoch": 3.9251547552054022,
"grad_norm": 32.73205022090457,
"learning_rate": 6.561679790026246e-06,
"loss": 0.6708,
"step": 13950
},
{
"epoch": 3.927968486212718,
"grad_norm": 88.75097655712725,
"learning_rate": 6.342957130358704e-06,
"loss": 0.4971,
"step": 13960
},
{
"epoch": 3.9307822172200337,
"grad_norm": 124.77660806239149,
"learning_rate": 6.124234470691163e-06,
"loss": 0.8132,
"step": 13970
},
{
"epoch": 3.9335959482273495,
"grad_norm": 52.88677822081236,
"learning_rate": 5.905511811023621e-06,
"loss": 0.498,
"step": 13980
},
{
"epoch": 3.9364096792346652,
"grad_norm": 18.73784553764002,
"learning_rate": 5.68678915135608e-06,
"loss": 0.4054,
"step": 13990
},
{
"epoch": 3.939223410241981,
"grad_norm": 22.54489264494174,
"learning_rate": 5.4680664916885386e-06,
"loss": 0.5623,
"step": 14000
},
{
"epoch": 3.9420371412492967,
"grad_norm": 1.1389150302633293,
"learning_rate": 5.2493438320209976e-06,
"loss": 0.5645,
"step": 14010
},
{
"epoch": 3.9448508722566125,
"grad_norm": 0.7575054525220537,
"learning_rate": 5.030621172353456e-06,
"loss": 0.5291,
"step": 14020
},
{
"epoch": 3.947664603263928,
"grad_norm": 14.757391649488548,
"learning_rate": 4.811898512685915e-06,
"loss": 0.3024,
"step": 14030
},
{
"epoch": 3.9504783342712435,
"grad_norm": 2.224238901964847,
"learning_rate": 4.593175853018372e-06,
"loss": 0.5885,
"step": 14040
},
{
"epoch": 3.9532920652785593,
"grad_norm": 52.85370797076157,
"learning_rate": 4.374453193350831e-06,
"loss": 0.6216,
"step": 14050
},
{
"epoch": 3.956105796285875,
"grad_norm": 4.926366273654347,
"learning_rate": 4.155730533683289e-06,
"loss": 0.8768,
"step": 14060
},
{
"epoch": 3.9589195272931907,
"grad_norm": 66.74702595575317,
"learning_rate": 3.937007874015747e-06,
"loss": 0.8456,
"step": 14070
},
{
"epoch": 3.9617332583005065,
"grad_norm": 108.65195259765787,
"learning_rate": 3.7182852143482063e-06,
"loss": 0.5933,
"step": 14080
},
{
"epoch": 3.9645469893078222,
"grad_norm": 2.067190403694772,
"learning_rate": 3.499562554680665e-06,
"loss": 0.4804,
"step": 14090
},
{
"epoch": 3.967360720315138,
"grad_norm": 49.20433018245054,
"learning_rate": 3.280839895013123e-06,
"loss": 0.8341,
"step": 14100
},
{
"epoch": 3.9701744513224537,
"grad_norm": 18.11378428616437,
"learning_rate": 3.0621172353455816e-06,
"loss": 0.1358,
"step": 14110
},
{
"epoch": 3.972988182329769,
"grad_norm": 20.757156258325487,
"learning_rate": 2.84339457567804e-06,
"loss": 0.4837,
"step": 14120
},
{
"epoch": 3.9758019133370848,
"grad_norm": 82.24151004002375,
"learning_rate": 2.6246719160104988e-06,
"loss": 0.4444,
"step": 14130
},
{
"epoch": 3.9786156443444005,
"grad_norm": 54.81837261410593,
"learning_rate": 2.4059492563429574e-06,
"loss": 0.4779,
"step": 14140
},
{
"epoch": 3.9814293753517163,
"grad_norm": 54.39145269424402,
"learning_rate": 2.1872265966754155e-06,
"loss": 0.5289,
"step": 14150
},
{
"epoch": 3.984243106359032,
"grad_norm": 94.0430998911988,
"learning_rate": 1.9685039370078737e-06,
"loss": 0.4829,
"step": 14160
},
{
"epoch": 3.9870568373663478,
"grad_norm": 7.233363364036652,
"learning_rate": 1.7497812773403325e-06,
"loss": 0.214,
"step": 14170
},
{
"epoch": 3.9898705683736635,
"grad_norm": 9.096609714666887,
"learning_rate": 1.5310586176727908e-06,
"loss": 0.759,
"step": 14180
},
{
"epoch": 3.9926842993809792,
"grad_norm": 738.9111636480206,
"learning_rate": 1.3123359580052494e-06,
"loss": 0.4992,
"step": 14190
},
{
"epoch": 3.995498030388295,
"grad_norm": 38.0307073875718,
"learning_rate": 1.0936132983377078e-06,
"loss": 0.2457,
"step": 14200
},
{
"epoch": 3.9983117613956107,
"grad_norm": 216.92497095466356,
"learning_rate": 8.748906386701662e-07,
"loss": 0.3663,
"step": 14210
},
{
"epoch": 4.0,
"eval_0_f1": 0.6524883028498512,
"eval_0_precision": 0.6687009590235397,
"eval_0_recall": 0.6370431893687708,
"eval_1_f1": 0.8796050692602416,
"eval_1_precision": 0.8722782405377758,
"eval_1_recall": 0.8870560261554465,
"eval_accuracy": 0.8211666849075189,
"eval_loss": 0.7568359375,
"eval_runtime": 468.6647,
"eval_samples_per_second": 19.496,
"eval_steps_per_second": 3.25,
"step": 14216
}
],
"logging_steps": 10,
"max_steps": 14216,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 500,
"total_flos": 2.6717228113906893e+17,
"train_batch_size": 6,
"trial_name": null,
"trial_params": null
}