medner_finetuned_lora / trainer_state.json
emilykang's picture
Model save
58123a1 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 9.98769987699877,
"eval_steps": 500,
"global_step": 6090,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.016400164001640016,
"grad_norm": 0.49903953075408936,
"learning_rate": 0.00019999866944080257,
"loss": 1.1198,
"step": 10
},
{
"epoch": 0.03280032800328003,
"grad_norm": 0.38635629415512085,
"learning_rate": 0.00019999467779861793,
"loss": 0.957,
"step": 20
},
{
"epoch": 0.04920049200492005,
"grad_norm": 0.3574012815952301,
"learning_rate": 0.0001999880251796685,
"loss": 0.8963,
"step": 30
},
{
"epoch": 0.06560065600656007,
"grad_norm": 0.36165040731430054,
"learning_rate": 0.00019997871176098827,
"loss": 0.8893,
"step": 40
},
{
"epoch": 0.08200082000820008,
"grad_norm": 0.3633221983909607,
"learning_rate": 0.0001999667377904184,
"loss": 0.8788,
"step": 50
},
{
"epoch": 0.0984009840098401,
"grad_norm": 0.3308263123035431,
"learning_rate": 0.00019995210358660038,
"loss": 0.8597,
"step": 60
},
{
"epoch": 0.11480114801148011,
"grad_norm": 0.37404152750968933,
"learning_rate": 0.0001999348095389677,
"loss": 0.8394,
"step": 70
},
{
"epoch": 0.13120131201312013,
"grad_norm": 0.36840370297431946,
"learning_rate": 0.0001999148561077355,
"loss": 0.8305,
"step": 80
},
{
"epoch": 0.14760147601476015,
"grad_norm": 0.33561620116233826,
"learning_rate": 0.00019989224382388813,
"loss": 0.836,
"step": 90
},
{
"epoch": 0.16400164001640016,
"grad_norm": 0.34898167848587036,
"learning_rate": 0.00019986697328916526,
"loss": 0.8259,
"step": 100
},
{
"epoch": 0.18040180401804018,
"grad_norm": 0.32518813014030457,
"learning_rate": 0.00019983904517604576,
"loss": 0.8255,
"step": 110
},
{
"epoch": 0.1968019680196802,
"grad_norm": 0.34965017437934875,
"learning_rate": 0.00019980846022772978,
"loss": 0.8192,
"step": 120
},
{
"epoch": 0.2132021320213202,
"grad_norm": 0.3208518326282501,
"learning_rate": 0.00019977521925811903,
"loss": 0.8177,
"step": 130
},
{
"epoch": 0.22960229602296023,
"grad_norm": 0.3215318024158478,
"learning_rate": 0.000199739323151795,
"loss": 0.8046,
"step": 140
},
{
"epoch": 0.24600246002460024,
"grad_norm": 0.3197932541370392,
"learning_rate": 0.0001997007728639956,
"loss": 0.8031,
"step": 150
},
{
"epoch": 0.26240262402624026,
"grad_norm": 0.3127509355545044,
"learning_rate": 0.0001996595694205897,
"loss": 0.7911,
"step": 160
},
{
"epoch": 0.2788027880278803,
"grad_norm": 0.3898150324821472,
"learning_rate": 0.00019961571391804962,
"loss": 0.8013,
"step": 170
},
{
"epoch": 0.2952029520295203,
"grad_norm": 0.32089415192604065,
"learning_rate": 0.00019956920752342225,
"loss": 0.7959,
"step": 180
},
{
"epoch": 0.3116031160311603,
"grad_norm": 0.3503674864768982,
"learning_rate": 0.0001995200514742978,
"loss": 0.7869,
"step": 190
},
{
"epoch": 0.3280032800328003,
"grad_norm": 0.36325424909591675,
"learning_rate": 0.00019946824707877693,
"loss": 0.7813,
"step": 200
},
{
"epoch": 0.34440344403444034,
"grad_norm": 0.3318350613117218,
"learning_rate": 0.00019941379571543596,
"loss": 0.796,
"step": 210
},
{
"epoch": 0.36080360803608036,
"grad_norm": 0.3133643865585327,
"learning_rate": 0.00019935669883329013,
"loss": 0.7832,
"step": 220
},
{
"epoch": 0.3772037720377204,
"grad_norm": 0.3414137363433838,
"learning_rate": 0.00019929695795175507,
"loss": 0.7907,
"step": 230
},
{
"epoch": 0.3936039360393604,
"grad_norm": 0.32717064023017883,
"learning_rate": 0.00019923457466060636,
"loss": 0.7824,
"step": 240
},
{
"epoch": 0.4100041000410004,
"grad_norm": 0.3150332272052765,
"learning_rate": 0.00019916955061993725,
"loss": 0.7992,
"step": 250
},
{
"epoch": 0.4264042640426404,
"grad_norm": 0.3353354334831238,
"learning_rate": 0.00019910188756011446,
"loss": 0.7782,
"step": 260
},
{
"epoch": 0.44280442804428044,
"grad_norm": 0.2986792325973511,
"learning_rate": 0.00019903158728173205,
"loss": 0.7609,
"step": 270
},
{
"epoch": 0.45920459204592046,
"grad_norm": 0.3223719596862793,
"learning_rate": 0.00019895865165556377,
"loss": 0.7785,
"step": 280
},
{
"epoch": 0.4756047560475605,
"grad_norm": 0.3108189105987549,
"learning_rate": 0.00019888308262251285,
"loss": 0.7762,
"step": 290
},
{
"epoch": 0.4920049200492005,
"grad_norm": 0.3153323829174042,
"learning_rate": 0.00019880488219356087,
"loss": 0.7797,
"step": 300
},
{
"epoch": 0.5084050840508405,
"grad_norm": 0.33102768659591675,
"learning_rate": 0.00019872405244971374,
"loss": 0.7731,
"step": 310
},
{
"epoch": 0.5248052480524805,
"grad_norm": 0.29496216773986816,
"learning_rate": 0.00019864059554194669,
"loss": 0.7648,
"step": 320
},
{
"epoch": 0.5412054120541205,
"grad_norm": 0.3221481740474701,
"learning_rate": 0.00019855451369114676,
"loss": 0.7762,
"step": 330
},
{
"epoch": 0.5576055760557606,
"grad_norm": 0.3474864065647125,
"learning_rate": 0.000198465809188054,
"loss": 0.7691,
"step": 340
},
{
"epoch": 0.5740057400574006,
"grad_norm": 0.36715272068977356,
"learning_rate": 0.00019837448439320027,
"loss": 0.7751,
"step": 350
},
{
"epoch": 0.5904059040590406,
"grad_norm": 0.33955511450767517,
"learning_rate": 0.00019828054173684644,
"loss": 0.7791,
"step": 360
},
{
"epoch": 0.6068060680606806,
"grad_norm": 0.3461228311061859,
"learning_rate": 0.00019818398371891784,
"loss": 0.7728,
"step": 370
},
{
"epoch": 0.6232062320623206,
"grad_norm": 0.32821011543273926,
"learning_rate": 0.0001980848129089376,
"loss": 0.7736,
"step": 380
},
{
"epoch": 0.6396063960639606,
"grad_norm": 0.3576536476612091,
"learning_rate": 0.00019798303194595846,
"loss": 0.756,
"step": 390
},
{
"epoch": 0.6560065600656007,
"grad_norm": 0.3349688649177551,
"learning_rate": 0.00019787864353849232,
"loss": 0.765,
"step": 400
},
{
"epoch": 0.6724067240672407,
"grad_norm": 0.2960722744464874,
"learning_rate": 0.00019777165046443824,
"loss": 0.7676,
"step": 410
},
{
"epoch": 0.6888068880688807,
"grad_norm": 0.31184637546539307,
"learning_rate": 0.00019766205557100868,
"loss": 0.7366,
"step": 420
},
{
"epoch": 0.7052070520705207,
"grad_norm": 0.3365538418292999,
"learning_rate": 0.00019754986177465344,
"loss": 0.7647,
"step": 430
},
{
"epoch": 0.7216072160721607,
"grad_norm": 0.3039809763431549,
"learning_rate": 0.00019743507206098233,
"loss": 0.741,
"step": 440
},
{
"epoch": 0.7380073800738007,
"grad_norm": 0.32564470171928406,
"learning_rate": 0.00019731768948468549,
"loss": 0.7576,
"step": 450
},
{
"epoch": 0.7544075440754408,
"grad_norm": 0.33124250173568726,
"learning_rate": 0.00019719771716945227,
"loss": 0.7392,
"step": 460
},
{
"epoch": 0.7708077080770808,
"grad_norm": 0.3306678235530853,
"learning_rate": 0.00019707515830788802,
"loss": 0.7513,
"step": 470
},
{
"epoch": 0.7872078720787208,
"grad_norm": 0.31717410683631897,
"learning_rate": 0.00019695001616142915,
"loss": 0.7419,
"step": 480
},
{
"epoch": 0.8036080360803608,
"grad_norm": 0.3117675483226776,
"learning_rate": 0.00019682229406025635,
"loss": 0.7543,
"step": 490
},
{
"epoch": 0.8200082000820008,
"grad_norm": 0.3408834636211395,
"learning_rate": 0.0001966919954032059,
"loss": 0.7356,
"step": 500
},
{
"epoch": 0.8364083640836408,
"grad_norm": 0.3456558883190155,
"learning_rate": 0.0001965591236576794,
"loss": 0.7509,
"step": 510
},
{
"epoch": 0.8528085280852808,
"grad_norm": 0.30214881896972656,
"learning_rate": 0.00019642368235955125,
"loss": 0.7483,
"step": 520
},
{
"epoch": 0.8692086920869209,
"grad_norm": 0.3334808945655823,
"learning_rate": 0.0001962856751130748,
"loss": 0.7532,
"step": 530
},
{
"epoch": 0.8856088560885609,
"grad_norm": 0.3257792890071869,
"learning_rate": 0.00019614510559078625,
"loss": 0.7465,
"step": 540
},
{
"epoch": 0.9020090200902009,
"grad_norm": 0.4038065969944,
"learning_rate": 0.00019600197753340699,
"loss": 0.7555,
"step": 550
},
{
"epoch": 0.9184091840918409,
"grad_norm": 0.32042616605758667,
"learning_rate": 0.00019585629474974415,
"loss": 0.7395,
"step": 560
},
{
"epoch": 0.9348093480934809,
"grad_norm": 0.3661757707595825,
"learning_rate": 0.00019570806111658898,
"loss": 0.7447,
"step": 570
},
{
"epoch": 0.951209512095121,
"grad_norm": 0.31112104654312134,
"learning_rate": 0.0001955572805786141,
"loss": 0.7333,
"step": 580
},
{
"epoch": 0.967609676096761,
"grad_norm": 0.31758418679237366,
"learning_rate": 0.000195403957148268,
"loss": 0.7243,
"step": 590
},
{
"epoch": 0.984009840098401,
"grad_norm": 0.3203936219215393,
"learning_rate": 0.00019524809490566877,
"loss": 0.744,
"step": 600
},
{
"epoch": 1.000410004100041,
"grad_norm": 0.3630532920360565,
"learning_rate": 0.00019508969799849523,
"loss": 0.7249,
"step": 610
},
{
"epoch": 1.016810168101681,
"grad_norm": 0.31734269857406616,
"learning_rate": 0.00019492877064187654,
"loss": 0.7088,
"step": 620
},
{
"epoch": 1.033210332103321,
"grad_norm": 0.3637007176876068,
"learning_rate": 0.00019476531711828027,
"loss": 0.7139,
"step": 630
},
{
"epoch": 1.049610496104961,
"grad_norm": 0.3162935972213745,
"learning_rate": 0.00019459934177739813,
"loss": 0.7064,
"step": 640
},
{
"epoch": 1.066010660106601,
"grad_norm": 0.32640552520751953,
"learning_rate": 0.0001944308490360305,
"loss": 0.7161,
"step": 650
},
{
"epoch": 1.082410824108241,
"grad_norm": 0.34453850984573364,
"learning_rate": 0.0001942598433779687,
"loss": 0.7095,
"step": 660
},
{
"epoch": 1.098810988109881,
"grad_norm": 0.3382568061351776,
"learning_rate": 0.00019408632935387577,
"loss": 0.7092,
"step": 670
},
{
"epoch": 1.115211152111521,
"grad_norm": 0.3298267722129822,
"learning_rate": 0.0001939103115811653,
"loss": 0.7154,
"step": 680
},
{
"epoch": 1.1316113161131611,
"grad_norm": 0.33924031257629395,
"learning_rate": 0.00019373179474387858,
"loss": 0.6945,
"step": 690
},
{
"epoch": 1.1480114801148011,
"grad_norm": 0.3305610418319702,
"learning_rate": 0.0001935507835925601,
"loss": 0.7156,
"step": 700
},
{
"epoch": 1.1644116441164412,
"grad_norm": 0.3400373160839081,
"learning_rate": 0.00019336728294413083,
"loss": 0.7091,
"step": 710
},
{
"epoch": 1.1808118081180812,
"grad_norm": 0.3354150056838989,
"learning_rate": 0.00019318129768176032,
"loss": 0.7019,
"step": 720
},
{
"epoch": 1.1972119721197212,
"grad_norm": 0.3427719175815582,
"learning_rate": 0.00019299283275473657,
"loss": 0.7157,
"step": 730
},
{
"epoch": 1.2136121361213612,
"grad_norm": 0.3088740408420563,
"learning_rate": 0.00019280189317833445,
"loss": 0.7115,
"step": 740
},
{
"epoch": 1.2300123001230012,
"grad_norm": 0.34402570128440857,
"learning_rate": 0.0001926084840336821,
"loss": 0.7197,
"step": 750
},
{
"epoch": 1.2464124641246412,
"grad_norm": 0.3542528450489044,
"learning_rate": 0.0001924126104676259,
"loss": 0.7103,
"step": 760
},
{
"epoch": 1.2628126281262813,
"grad_norm": 0.3533455431461334,
"learning_rate": 0.00019221427769259333,
"loss": 0.7209,
"step": 770
},
{
"epoch": 1.2792127921279213,
"grad_norm": 0.3470947742462158,
"learning_rate": 0.00019201349098645434,
"loss": 0.7137,
"step": 780
},
{
"epoch": 1.2956129561295613,
"grad_norm": 0.3277159035205841,
"learning_rate": 0.0001918102556923809,
"loss": 0.7036,
"step": 790
},
{
"epoch": 1.3120131201312013,
"grad_norm": 0.35741642117500305,
"learning_rate": 0.0001916045772187048,
"loss": 0.7013,
"step": 800
},
{
"epoch": 1.3284132841328413,
"grad_norm": 0.33222687244415283,
"learning_rate": 0.0001913964610387738,
"loss": 0.7052,
"step": 810
},
{
"epoch": 1.3448134481344813,
"grad_norm": 0.3761126399040222,
"learning_rate": 0.00019118591269080574,
"loss": 0.695,
"step": 820
},
{
"epoch": 1.3612136121361214,
"grad_norm": 0.34687796235084534,
"learning_rate": 0.00019097293777774153,
"loss": 0.7053,
"step": 830
},
{
"epoch": 1.3776137761377614,
"grad_norm": 0.3391115367412567,
"learning_rate": 0.00019075754196709572,
"loss": 0.6842,
"step": 840
},
{
"epoch": 1.3940139401394014,
"grad_norm": 0.35560715198516846,
"learning_rate": 0.00019053973099080585,
"loss": 0.6925,
"step": 850
},
{
"epoch": 1.4104141041410414,
"grad_norm": 0.35144075751304626,
"learning_rate": 0.00019031951064507987,
"loss": 0.7063,
"step": 860
},
{
"epoch": 1.4268142681426814,
"grad_norm": 0.3776865005493164,
"learning_rate": 0.0001900968867902419,
"loss": 0.7157,
"step": 870
},
{
"epoch": 1.4432144321443214,
"grad_norm": 0.3723245859146118,
"learning_rate": 0.00018987186535057634,
"loss": 0.7064,
"step": 880
},
{
"epoch": 1.4596145961459615,
"grad_norm": 0.3623642921447754,
"learning_rate": 0.0001896444523141701,
"loss": 0.7056,
"step": 890
},
{
"epoch": 1.4760147601476015,
"grad_norm": 0.34806951880455017,
"learning_rate": 0.0001894146537327533,
"loss": 0.7086,
"step": 900
},
{
"epoch": 1.4924149241492415,
"grad_norm": 0.36614471673965454,
"learning_rate": 0.00018918247572153823,
"loss": 0.7032,
"step": 910
},
{
"epoch": 1.5088150881508815,
"grad_norm": 0.351053923368454,
"learning_rate": 0.00018894792445905674,
"loss": 0.6988,
"step": 920
},
{
"epoch": 1.5252152521525215,
"grad_norm": 0.37270405888557434,
"learning_rate": 0.00018871100618699554,
"loss": 0.6926,
"step": 930
},
{
"epoch": 1.5416154161541615,
"grad_norm": 0.3521835207939148,
"learning_rate": 0.00018847172721003043,
"loss": 0.7004,
"step": 940
},
{
"epoch": 1.5580155801558015,
"grad_norm": 0.37083443999290466,
"learning_rate": 0.00018823009389565818,
"loss": 0.7141,
"step": 950
},
{
"epoch": 1.5744157441574416,
"grad_norm": 0.36266541481018066,
"learning_rate": 0.00018798611267402746,
"loss": 0.7029,
"step": 960
},
{
"epoch": 1.5908159081590816,
"grad_norm": 0.34403783082962036,
"learning_rate": 0.0001877397900377674,
"loss": 0.7022,
"step": 970
},
{
"epoch": 1.6072160721607216,
"grad_norm": 0.37788769602775574,
"learning_rate": 0.00018749113254181498,
"loss": 0.7049,
"step": 980
},
{
"epoch": 1.6236162361623616,
"grad_norm": 0.35607653856277466,
"learning_rate": 0.00018724014680324057,
"loss": 0.6957,
"step": 990
},
{
"epoch": 1.6400164001640016,
"grad_norm": 0.34468725323677063,
"learning_rate": 0.00018698683950107184,
"loss": 0.7148,
"step": 1000
},
{
"epoch": 1.6564165641656416,
"grad_norm": 0.36872032284736633,
"learning_rate": 0.00018673121737611598,
"loss": 0.6922,
"step": 1010
},
{
"epoch": 1.6728167281672817,
"grad_norm": 0.34786421060562134,
"learning_rate": 0.00018647328723078038,
"loss": 0.7021,
"step": 1020
},
{
"epoch": 1.6892168921689217,
"grad_norm": 0.370241641998291,
"learning_rate": 0.00018621305592889163,
"loss": 0.6789,
"step": 1030
},
{
"epoch": 1.7056170561705617,
"grad_norm": 0.3378838300704956,
"learning_rate": 0.00018595053039551274,
"loss": 0.6741,
"step": 1040
},
{
"epoch": 1.7220172201722017,
"grad_norm": 0.3661263883113861,
"learning_rate": 0.00018568571761675893,
"loss": 0.6808,
"step": 1050
},
{
"epoch": 1.7384173841738417,
"grad_norm": 0.3620483875274658,
"learning_rate": 0.0001854186246396118,
"loss": 0.6895,
"step": 1060
},
{
"epoch": 1.7548175481754817,
"grad_norm": 0.3466210663318634,
"learning_rate": 0.00018514925857173177,
"loss": 0.6876,
"step": 1070
},
{
"epoch": 1.7712177121771218,
"grad_norm": 0.36353030800819397,
"learning_rate": 0.0001848776265812687,
"loss": 0.6876,
"step": 1080
},
{
"epoch": 1.7876178761787618,
"grad_norm": 0.3688850998878479,
"learning_rate": 0.00018460373589667154,
"loss": 0.6881,
"step": 1090
},
{
"epoch": 1.8040180401804018,
"grad_norm": 0.35844072699546814,
"learning_rate": 0.00018432759380649562,
"loss": 0.7038,
"step": 1100
},
{
"epoch": 1.8204182041820418,
"grad_norm": 0.38838833570480347,
"learning_rate": 0.00018404920765920896,
"loss": 0.6961,
"step": 1110
},
{
"epoch": 1.8368183681836818,
"grad_norm": 0.34356170892715454,
"learning_rate": 0.00018376858486299647,
"loss": 0.6849,
"step": 1120
},
{
"epoch": 1.8532185321853218,
"grad_norm": 0.3590626120567322,
"learning_rate": 0.00018348573288556308,
"loss": 0.68,
"step": 1130
},
{
"epoch": 1.8696186961869619,
"grad_norm": 0.3639477491378784,
"learning_rate": 0.00018320065925393468,
"loss": 0.6996,
"step": 1140
},
{
"epoch": 1.8860188601886019,
"grad_norm": 0.39745190739631653,
"learning_rate": 0.00018291337155425821,
"loss": 0.6901,
"step": 1150
},
{
"epoch": 1.902419024190242,
"grad_norm": 0.37302690744400024,
"learning_rate": 0.0001826238774315995,
"loss": 0.6784,
"step": 1160
},
{
"epoch": 1.918819188191882,
"grad_norm": 0.37480470538139343,
"learning_rate": 0.00018233218458973984,
"loss": 0.6806,
"step": 1170
},
{
"epoch": 1.935219352193522,
"grad_norm": 0.37153083086013794,
"learning_rate": 0.00018203830079097113,
"loss": 0.6853,
"step": 1180
},
{
"epoch": 1.951619516195162,
"grad_norm": 0.3530969023704529,
"learning_rate": 0.00018174223385588917,
"loss": 0.6785,
"step": 1190
},
{
"epoch": 1.968019680196802,
"grad_norm": 0.37324872612953186,
"learning_rate": 0.00018144399166318572,
"loss": 0.6733,
"step": 1200
},
{
"epoch": 1.984419844198442,
"grad_norm": 0.3680344223976135,
"learning_rate": 0.00018114358214943853,
"loss": 0.6885,
"step": 1210
},
{
"epoch": 2.000820008200082,
"grad_norm": 0.42146405577659607,
"learning_rate": 0.00018084101330890048,
"loss": 0.6759,
"step": 1220
},
{
"epoch": 2.0172201722017222,
"grad_norm": 0.3808436989784241,
"learning_rate": 0.00018053629319328662,
"loss": 0.655,
"step": 1230
},
{
"epoch": 2.033620336203362,
"grad_norm": 0.39610400795936584,
"learning_rate": 0.00018022942991156,
"loss": 0.6675,
"step": 1240
},
{
"epoch": 2.0500205002050023,
"grad_norm": 0.3906943202018738,
"learning_rate": 0.00017992043162971588,
"loss": 0.6451,
"step": 1250
},
{
"epoch": 2.066420664206642,
"grad_norm": 0.4022831916809082,
"learning_rate": 0.00017960930657056438,
"loss": 0.6699,
"step": 1260
},
{
"epoch": 2.0828208282082823,
"grad_norm": 0.41200950741767883,
"learning_rate": 0.00017929606301351168,
"loss": 0.6559,
"step": 1270
},
{
"epoch": 2.099220992209922,
"grad_norm": 0.37226754426956177,
"learning_rate": 0.00017898070929433965,
"loss": 0.655,
"step": 1280
},
{
"epoch": 2.1156211562115623,
"grad_norm": 0.3640751540660858,
"learning_rate": 0.00017866325380498416,
"loss": 0.6471,
"step": 1290
},
{
"epoch": 2.132021320213202,
"grad_norm": 0.37838858366012573,
"learning_rate": 0.00017834370499331166,
"loss": 0.651,
"step": 1300
},
{
"epoch": 2.1484214842148424,
"grad_norm": 0.3831733167171478,
"learning_rate": 0.0001780220713628943,
"loss": 0.6504,
"step": 1310
},
{
"epoch": 2.164821648216482,
"grad_norm": 0.39556068181991577,
"learning_rate": 0.0001776983614727838,
"loss": 0.6501,
"step": 1320
},
{
"epoch": 2.181221812218122,
"grad_norm": 0.3833105266094208,
"learning_rate": 0.00017737258393728364,
"loss": 0.6598,
"step": 1330
},
{
"epoch": 2.197621976219762,
"grad_norm": 0.40265655517578125,
"learning_rate": 0.00017704474742571969,
"loss": 0.663,
"step": 1340
},
{
"epoch": 2.2140221402214024,
"grad_norm": 0.4098280072212219,
"learning_rate": 0.00017671486066220965,
"loss": 0.6567,
"step": 1350
},
{
"epoch": 2.230422304223042,
"grad_norm": 0.4019940197467804,
"learning_rate": 0.0001763829324254309,
"loss": 0.6717,
"step": 1360
},
{
"epoch": 2.246822468224682,
"grad_norm": 0.3930509388446808,
"learning_rate": 0.00017604897154838685,
"loss": 0.6544,
"step": 1370
},
{
"epoch": 2.2632226322263223,
"grad_norm": 0.40629473328590393,
"learning_rate": 0.00017571298691817177,
"loss": 0.65,
"step": 1380
},
{
"epoch": 2.2796227962279625,
"grad_norm": 0.41943445801734924,
"learning_rate": 0.00017537498747573443,
"loss": 0.6612,
"step": 1390
},
{
"epoch": 2.2960229602296023,
"grad_norm": 0.3832157552242279,
"learning_rate": 0.00017503498221564025,
"loss": 0.6599,
"step": 1400
},
{
"epoch": 2.312423124231242,
"grad_norm": 0.3786883056163788,
"learning_rate": 0.0001746929801858317,
"loss": 0.6636,
"step": 1410
},
{
"epoch": 2.3288232882328823,
"grad_norm": 0.40903323888778687,
"learning_rate": 0.0001743489904873877,
"loss": 0.6462,
"step": 1420
},
{
"epoch": 2.3452234522345226,
"grad_norm": 0.3858684003353119,
"learning_rate": 0.00017400302227428143,
"loss": 0.6556,
"step": 1430
},
{
"epoch": 2.3616236162361623,
"grad_norm": 0.43149909377098083,
"learning_rate": 0.0001736550847531366,
"loss": 0.6591,
"step": 1440
},
{
"epoch": 2.378023780237802,
"grad_norm": 0.43920159339904785,
"learning_rate": 0.00017330518718298264,
"loss": 0.6609,
"step": 1450
},
{
"epoch": 2.3944239442394424,
"grad_norm": 0.39907246828079224,
"learning_rate": 0.0001729533388750081,
"loss": 0.6572,
"step": 1460
},
{
"epoch": 2.4108241082410826,
"grad_norm": 0.3920808434486389,
"learning_rate": 0.0001725995491923131,
"loss": 0.6497,
"step": 1470
},
{
"epoch": 2.4272242722427224,
"grad_norm": 0.39895495772361755,
"learning_rate": 0.00017224382754965989,
"loss": 0.646,
"step": 1480
},
{
"epoch": 2.443624436244362,
"grad_norm": 0.41222187876701355,
"learning_rate": 0.00017188618341322254,
"loss": 0.6458,
"step": 1490
},
{
"epoch": 2.4600246002460024,
"grad_norm": 0.41394343972206116,
"learning_rate": 0.00017152662630033505,
"loss": 0.6528,
"step": 1500
},
{
"epoch": 2.4764247642476427,
"grad_norm": 0.4161871373653412,
"learning_rate": 0.00017116516577923775,
"loss": 0.6414,
"step": 1510
},
{
"epoch": 2.4928249282492825,
"grad_norm": 0.43303370475769043,
"learning_rate": 0.00017080181146882317,
"loss": 0.6573,
"step": 1520
},
{
"epoch": 2.5092250922509223,
"grad_norm": 0.3973419964313507,
"learning_rate": 0.00017043657303837963,
"loss": 0.6541,
"step": 1530
},
{
"epoch": 2.5256252562525625,
"grad_norm": 0.3919370472431183,
"learning_rate": 0.00017006946020733425,
"loss": 0.6626,
"step": 1540
},
{
"epoch": 2.5420254202542027,
"grad_norm": 0.4177446663379669,
"learning_rate": 0.00016970048274499408,
"loss": 0.6478,
"step": 1550
},
{
"epoch": 2.5584255842558425,
"grad_norm": 0.39967086911201477,
"learning_rate": 0.0001693296504702862,
"loss": 0.6679,
"step": 1560
},
{
"epoch": 2.5748257482574823,
"grad_norm": 0.4124162495136261,
"learning_rate": 0.00016895697325149657,
"loss": 0.6597,
"step": 1570
},
{
"epoch": 2.5912259122591226,
"grad_norm": 0.37739571928977966,
"learning_rate": 0.00016858246100600715,
"loss": 0.6578,
"step": 1580
},
{
"epoch": 2.607626076260763,
"grad_norm": 0.4335273802280426,
"learning_rate": 0.00016820612370003221,
"loss": 0.6608,
"step": 1590
},
{
"epoch": 2.6240262402624026,
"grad_norm": 0.3718789517879486,
"learning_rate": 0.00016782797134835305,
"loss": 0.6447,
"step": 1600
},
{
"epoch": 2.6404264042640424,
"grad_norm": 0.41410380601882935,
"learning_rate": 0.0001674480140140514,
"loss": 0.6562,
"step": 1610
},
{
"epoch": 2.6568265682656826,
"grad_norm": 0.4115227460861206,
"learning_rate": 0.00016706626180824186,
"loss": 0.6451,
"step": 1620
},
{
"epoch": 2.673226732267323,
"grad_norm": 0.4145969748497009,
"learning_rate": 0.00016668272488980254,
"loss": 0.6326,
"step": 1630
},
{
"epoch": 2.6896268962689627,
"grad_norm": 0.42359670996665955,
"learning_rate": 0.00016629741346510496,
"loss": 0.6324,
"step": 1640
},
{
"epoch": 2.7060270602706025,
"grad_norm": 0.38336479663848877,
"learning_rate": 0.0001659103377877423,
"loss": 0.6508,
"step": 1650
},
{
"epoch": 2.7224272242722427,
"grad_norm": 0.4128510057926178,
"learning_rate": 0.00016552150815825662,
"loss": 0.6538,
"step": 1660
},
{
"epoch": 2.738827388273883,
"grad_norm": 0.47073811292648315,
"learning_rate": 0.0001651309349238647,
"loss": 0.6574,
"step": 1670
},
{
"epoch": 2.7552275522755227,
"grad_norm": 0.4234805703163147,
"learning_rate": 0.00016473862847818277,
"loss": 0.6643,
"step": 1680
},
{
"epoch": 2.7716277162771625,
"grad_norm": 0.4076490104198456,
"learning_rate": 0.0001643445992609498,
"loss": 0.6509,
"step": 1690
},
{
"epoch": 2.7880278802788028,
"grad_norm": 0.4320470988750458,
"learning_rate": 0.00016394885775774975,
"loss": 0.643,
"step": 1700
},
{
"epoch": 2.804428044280443,
"grad_norm": 0.409150630235672,
"learning_rate": 0.00016355141449973256,
"loss": 0.6452,
"step": 1710
},
{
"epoch": 2.820828208282083,
"grad_norm": 0.46668359637260437,
"learning_rate": 0.0001631522800633339,
"loss": 0.645,
"step": 1720
},
{
"epoch": 2.8372283722837226,
"grad_norm": 0.4314485192298889,
"learning_rate": 0.00016275146506999365,
"loss": 0.6374,
"step": 1730
},
{
"epoch": 2.853628536285363,
"grad_norm": 0.41864487528800964,
"learning_rate": 0.00016234898018587337,
"loss": 0.6343,
"step": 1740
},
{
"epoch": 2.870028700287003,
"grad_norm": 0.4255831241607666,
"learning_rate": 0.0001619448361215723,
"loss": 0.6614,
"step": 1750
},
{
"epoch": 2.886428864288643,
"grad_norm": 0.411807656288147,
"learning_rate": 0.0001615390436318425,
"loss": 0.6372,
"step": 1760
},
{
"epoch": 2.9028290282902827,
"grad_norm": 0.41166943311691284,
"learning_rate": 0.0001611316135153026,
"loss": 0.6505,
"step": 1770
},
{
"epoch": 2.919229192291923,
"grad_norm": 0.41985031962394714,
"learning_rate": 0.0001607225566141503,
"loss": 0.6515,
"step": 1780
},
{
"epoch": 2.935629356293563,
"grad_norm": 0.40023133158683777,
"learning_rate": 0.0001603118838138741,
"loss": 0.6466,
"step": 1790
},
{
"epoch": 2.952029520295203,
"grad_norm": 0.4138244688510895,
"learning_rate": 0.0001598996060429634,
"loss": 0.6475,
"step": 1800
},
{
"epoch": 2.9684296842968427,
"grad_norm": 0.39823102951049805,
"learning_rate": 0.0001594857342726178,
"loss": 0.6419,
"step": 1810
},
{
"epoch": 2.984829848298483,
"grad_norm": 0.3888455033302307,
"learning_rate": 0.0001590702795164551,
"loss": 0.6416,
"step": 1820
},
{
"epoch": 3.001230012300123,
"grad_norm": 0.5585771799087524,
"learning_rate": 0.0001586532528302183,
"loss": 0.6565,
"step": 1830
},
{
"epoch": 3.017630176301763,
"grad_norm": 0.4332481026649475,
"learning_rate": 0.00015823466531148124,
"loss": 0.619,
"step": 1840
},
{
"epoch": 3.0340303403034032,
"grad_norm": 0.44493553042411804,
"learning_rate": 0.0001578145280993533,
"loss": 0.6057,
"step": 1850
},
{
"epoch": 3.050430504305043,
"grad_norm": 0.43917590379714966,
"learning_rate": 0.0001573928523741832,
"loss": 0.6058,
"step": 1860
},
{
"epoch": 3.0668306683066833,
"grad_norm": 0.40770649909973145,
"learning_rate": 0.0001569696493572612,
"loss": 0.6092,
"step": 1870
},
{
"epoch": 3.083230832308323,
"grad_norm": 0.44579923152923584,
"learning_rate": 0.0001565449303105207,
"loss": 0.6272,
"step": 1880
},
{
"epoch": 3.0996309963099633,
"grad_norm": 0.4329695403575897,
"learning_rate": 0.00015611870653623825,
"loss": 0.6238,
"step": 1890
},
{
"epoch": 3.116031160311603,
"grad_norm": 0.44702407717704773,
"learning_rate": 0.00015569098937673318,
"loss": 0.6124,
"step": 1900
},
{
"epoch": 3.1324313243132433,
"grad_norm": 0.43535304069519043,
"learning_rate": 0.00015526179021406553,
"loss": 0.6179,
"step": 1910
},
{
"epoch": 3.148831488314883,
"grad_norm": 0.4820752739906311,
"learning_rate": 0.0001548311204697331,
"loss": 0.6228,
"step": 1920
},
{
"epoch": 3.1652316523165234,
"grad_norm": 0.4629499614238739,
"learning_rate": 0.00015439899160436772,
"loss": 0.6189,
"step": 1930
},
{
"epoch": 3.181631816318163,
"grad_norm": 0.4845089316368103,
"learning_rate": 0.00015396541511743012,
"loss": 0.6229,
"step": 1940
},
{
"epoch": 3.1980319803198034,
"grad_norm": 0.46592435240745544,
"learning_rate": 0.00015353040254690393,
"loss": 0.6178,
"step": 1950
},
{
"epoch": 3.214432144321443,
"grad_norm": 0.44851964712142944,
"learning_rate": 0.0001530939654689887,
"loss": 0.6246,
"step": 1960
},
{
"epoch": 3.2308323083230834,
"grad_norm": 0.4483705461025238,
"learning_rate": 0.00015265611549779176,
"loss": 0.6204,
"step": 1970
},
{
"epoch": 3.2472324723247232,
"grad_norm": 0.45693984627723694,
"learning_rate": 0.00015221686428501928,
"loss": 0.6061,
"step": 1980
},
{
"epoch": 3.2636326363263635,
"grad_norm": 0.46255984902381897,
"learning_rate": 0.0001517762235196661,
"loss": 0.6058,
"step": 1990
},
{
"epoch": 3.2800328003280033,
"grad_norm": 0.471983939409256,
"learning_rate": 0.00015133420492770462,
"loss": 0.6157,
"step": 2000
},
{
"epoch": 3.296432964329643,
"grad_norm": 0.49640896916389465,
"learning_rate": 0.0001508908202717729,
"loss": 0.6237,
"step": 2010
},
{
"epoch": 3.3128331283312833,
"grad_norm": 0.4903806149959564,
"learning_rate": 0.00015044608135086164,
"loss": 0.6339,
"step": 2020
},
{
"epoch": 3.3292332923329235,
"grad_norm": 0.47123804688453674,
"learning_rate": 0.00015000000000000001,
"loss": 0.6319,
"step": 2030
},
{
"epoch": 3.3456334563345633,
"grad_norm": 0.47124335169792175,
"learning_rate": 0.00014955258808994096,
"loss": 0.6119,
"step": 2040
},
{
"epoch": 3.362033620336203,
"grad_norm": 0.4579004645347595,
"learning_rate": 0.00014910385752684506,
"loss": 0.6042,
"step": 2050
},
{
"epoch": 3.3784337843378434,
"grad_norm": 0.4778992235660553,
"learning_rate": 0.0001486538202519639,
"loss": 0.6187,
"step": 2060
},
{
"epoch": 3.3948339483394836,
"grad_norm": 0.4603955149650574,
"learning_rate": 0.0001482024882413222,
"loss": 0.6166,
"step": 2070
},
{
"epoch": 3.4112341123411234,
"grad_norm": 0.4642026126384735,
"learning_rate": 0.00014774987350539911,
"loss": 0.6183,
"step": 2080
},
{
"epoch": 3.427634276342763,
"grad_norm": 0.5194045305252075,
"learning_rate": 0.00014729598808880861,
"loss": 0.617,
"step": 2090
},
{
"epoch": 3.4440344403444034,
"grad_norm": 0.4535214900970459,
"learning_rate": 0.00014684084406997903,
"loss": 0.6301,
"step": 2100
},
{
"epoch": 3.4604346043460437,
"grad_norm": 0.47860169410705566,
"learning_rate": 0.00014638445356083155,
"loss": 0.6113,
"step": 2110
},
{
"epoch": 3.4768347683476835,
"grad_norm": 0.4803871810436249,
"learning_rate": 0.000145926828706458,
"loss": 0.6122,
"step": 2120
},
{
"epoch": 3.4932349323493233,
"grad_norm": 0.47383010387420654,
"learning_rate": 0.00014546798168479756,
"loss": 0.6137,
"step": 2130
},
{
"epoch": 3.5096350963509635,
"grad_norm": 0.47791990637779236,
"learning_rate": 0.0001450079247063127,
"loss": 0.6215,
"step": 2140
},
{
"epoch": 3.5260352603526037,
"grad_norm": 0.4606546461582184,
"learning_rate": 0.0001445466700136643,
"loss": 0.6022,
"step": 2150
},
{
"epoch": 3.5424354243542435,
"grad_norm": 0.4335190951824188,
"learning_rate": 0.00014408422988138584,
"loss": 0.6089,
"step": 2160
},
{
"epoch": 3.5588355883558833,
"grad_norm": 0.45156562328338623,
"learning_rate": 0.00014362061661555675,
"loss": 0.6251,
"step": 2170
},
{
"epoch": 3.5752357523575236,
"grad_norm": 0.43693843483924866,
"learning_rate": 0.00014315584255347494,
"loss": 0.62,
"step": 2180
},
{
"epoch": 3.591635916359164,
"grad_norm": 0.46648937463760376,
"learning_rate": 0.00014268992006332846,
"loss": 0.6336,
"step": 2190
},
{
"epoch": 3.6080360803608036,
"grad_norm": 0.47864094376564026,
"learning_rate": 0.00014222286154386642,
"loss": 0.6079,
"step": 2200
},
{
"epoch": 3.6244362443624434,
"grad_norm": 0.43990063667297363,
"learning_rate": 0.00014175467942406894,
"loss": 0.6321,
"step": 2210
},
{
"epoch": 3.6408364083640836,
"grad_norm": 0.4700396656990051,
"learning_rate": 0.0001412853861628166,
"loss": 0.6195,
"step": 2220
},
{
"epoch": 3.657236572365724,
"grad_norm": 0.45286816358566284,
"learning_rate": 0.00014081499424855863,
"loss": 0.6236,
"step": 2230
},
{
"epoch": 3.6736367363673637,
"grad_norm": 0.4623110294342041,
"learning_rate": 0.00014034351619898088,
"loss": 0.6269,
"step": 2240
},
{
"epoch": 3.6900369003690034,
"grad_norm": 0.5022940039634705,
"learning_rate": 0.00013987096456067236,
"loss": 0.6107,
"step": 2250
},
{
"epoch": 3.7064370643706437,
"grad_norm": 0.49338841438293457,
"learning_rate": 0.00013939735190879172,
"loss": 0.5996,
"step": 2260
},
{
"epoch": 3.722837228372284,
"grad_norm": 0.4769245982170105,
"learning_rate": 0.0001389226908467323,
"loss": 0.6074,
"step": 2270
},
{
"epoch": 3.7392373923739237,
"grad_norm": 0.48450419306755066,
"learning_rate": 0.00013844699400578696,
"loss": 0.6092,
"step": 2280
},
{
"epoch": 3.7556375563755635,
"grad_norm": 0.46382635831832886,
"learning_rate": 0.00013797027404481184,
"loss": 0.6154,
"step": 2290
},
{
"epoch": 3.7720377203772038,
"grad_norm": 0.4822761118412018,
"learning_rate": 0.00013749254364988956,
"loss": 0.612,
"step": 2300
},
{
"epoch": 3.788437884378844,
"grad_norm": 0.4664019048213959,
"learning_rate": 0.00013701381553399145,
"loss": 0.6003,
"step": 2310
},
{
"epoch": 3.804838048380484,
"grad_norm": 0.4592795670032501,
"learning_rate": 0.00013653410243663952,
"loss": 0.6122,
"step": 2320
},
{
"epoch": 3.8212382123821236,
"grad_norm": 0.4780454635620117,
"learning_rate": 0.00013605341712356723,
"loss": 0.6216,
"step": 2330
},
{
"epoch": 3.837638376383764,
"grad_norm": 0.48119214177131653,
"learning_rate": 0.00013557177238637986,
"loss": 0.6119,
"step": 2340
},
{
"epoch": 3.854038540385404,
"grad_norm": 0.44965630769729614,
"learning_rate": 0.00013508918104221412,
"loss": 0.619,
"step": 2350
},
{
"epoch": 3.870438704387044,
"grad_norm": 0.4915274977684021,
"learning_rate": 0.00013460565593339705,
"loss": 0.6101,
"step": 2360
},
{
"epoch": 3.8868388683886836,
"grad_norm": 0.4430049955844879,
"learning_rate": 0.00013412120992710425,
"loss": 0.6125,
"step": 2370
},
{
"epoch": 3.903239032390324,
"grad_norm": 0.4661041796207428,
"learning_rate": 0.0001336358559150175,
"loss": 0.6177,
"step": 2380
},
{
"epoch": 3.919639196391964,
"grad_norm": 0.48300737142562866,
"learning_rate": 0.00013314960681298175,
"loss": 0.6039,
"step": 2390
},
{
"epoch": 3.936039360393604,
"grad_norm": 0.45341914892196655,
"learning_rate": 0.00013266247556066122,
"loss": 0.5989,
"step": 2400
},
{
"epoch": 3.9524395243952437,
"grad_norm": 0.46126899123191833,
"learning_rate": 0.00013217447512119533,
"loss": 0.6188,
"step": 2410
},
{
"epoch": 3.968839688396884,
"grad_norm": 0.4723737835884094,
"learning_rate": 0.0001316856184808535,
"loss": 0.6091,
"step": 2420
},
{
"epoch": 3.985239852398524,
"grad_norm": 0.4723500907421112,
"learning_rate": 0.0001311959186486898,
"loss": 0.6178,
"step": 2430
},
{
"epoch": 4.001640016400164,
"grad_norm": 0.4641455113887787,
"learning_rate": 0.00013070538865619642,
"loss": 0.6144,
"step": 2440
},
{
"epoch": 4.018040180401804,
"grad_norm": 0.5465942621231079,
"learning_rate": 0.00013021404155695725,
"loss": 0.5762,
"step": 2450
},
{
"epoch": 4.0344403444034445,
"grad_norm": 0.5195038318634033,
"learning_rate": 0.00012972189042630044,
"loss": 0.5936,
"step": 2460
},
{
"epoch": 4.050840508405084,
"grad_norm": 0.5215120911598206,
"learning_rate": 0.00012922894836095013,
"loss": 0.5797,
"step": 2470
},
{
"epoch": 4.067240672406724,
"grad_norm": 0.5008783340454102,
"learning_rate": 0.00012873522847867835,
"loss": 0.5887,
"step": 2480
},
{
"epoch": 4.083640836408364,
"grad_norm": 0.47933313250541687,
"learning_rate": 0.0001282407439179557,
"loss": 0.5831,
"step": 2490
},
{
"epoch": 4.1000410004100045,
"grad_norm": 0.5141870379447937,
"learning_rate": 0.00012774550783760182,
"loss": 0.5765,
"step": 2500
},
{
"epoch": 4.116441164411644,
"grad_norm": 0.4889928996562958,
"learning_rate": 0.0001272495334164351,
"loss": 0.5834,
"step": 2510
},
{
"epoch": 4.132841328413284,
"grad_norm": 0.5281891822814941,
"learning_rate": 0.00012675283385292212,
"loss": 0.5778,
"step": 2520
},
{
"epoch": 4.149241492414924,
"grad_norm": 0.5182332396507263,
"learning_rate": 0.00012625542236482628,
"loss": 0.5882,
"step": 2530
},
{
"epoch": 4.165641656416565,
"grad_norm": 0.554658055305481,
"learning_rate": 0.00012575731218885625,
"loss": 0.5871,
"step": 2540
},
{
"epoch": 4.182041820418204,
"grad_norm": 0.5192410349845886,
"learning_rate": 0.00012525851658031352,
"loss": 0.5768,
"step": 2550
},
{
"epoch": 4.198441984419844,
"grad_norm": 0.5177326202392578,
"learning_rate": 0.0001247590488127398,
"loss": 0.5849,
"step": 2560
},
{
"epoch": 4.214842148421484,
"grad_norm": 0.5142077803611755,
"learning_rate": 0.0001242589221775637,
"loss": 0.5744,
"step": 2570
},
{
"epoch": 4.231242312423125,
"grad_norm": 0.48389384150505066,
"learning_rate": 0.00012375814998374712,
"loss": 0.5977,
"step": 2580
},
{
"epoch": 4.2476424764247644,
"grad_norm": 0.5631837248802185,
"learning_rate": 0.00012325674555743106,
"loss": 0.5827,
"step": 2590
},
{
"epoch": 4.264042640426404,
"grad_norm": 0.5213803052902222,
"learning_rate": 0.0001227547222415809,
"loss": 0.5815,
"step": 2600
},
{
"epoch": 4.280442804428044,
"grad_norm": 0.5063258409500122,
"learning_rate": 0.00012225209339563145,
"loss": 0.591,
"step": 2610
},
{
"epoch": 4.296842968429685,
"grad_norm": 0.503495454788208,
"learning_rate": 0.0001217488723951314,
"loss": 0.5872,
"step": 2620
},
{
"epoch": 4.3132431324313245,
"grad_norm": 0.5165495276451111,
"learning_rate": 0.00012124507263138736,
"loss": 0.5739,
"step": 2630
},
{
"epoch": 4.329643296432964,
"grad_norm": 0.4980062246322632,
"learning_rate": 0.00012074070751110751,
"loss": 0.5795,
"step": 2640
},
{
"epoch": 4.346043460434604,
"grad_norm": 0.5419421792030334,
"learning_rate": 0.00012023579045604485,
"loss": 0.5844,
"step": 2650
},
{
"epoch": 4.362443624436244,
"grad_norm": 0.5244302749633789,
"learning_rate": 0.00011973033490264001,
"loss": 0.5777,
"step": 2660
},
{
"epoch": 4.378843788437885,
"grad_norm": 0.5156043171882629,
"learning_rate": 0.0001192243543016637,
"loss": 0.5832,
"step": 2670
},
{
"epoch": 4.395243952439524,
"grad_norm": 0.5119035840034485,
"learning_rate": 0.00011871786211785876,
"loss": 0.59,
"step": 2680
},
{
"epoch": 4.411644116441164,
"grad_norm": 0.5079758167266846,
"learning_rate": 0.00011821087182958186,
"loss": 0.5652,
"step": 2690
},
{
"epoch": 4.428044280442805,
"grad_norm": 0.5077067613601685,
"learning_rate": 0.00011770339692844483,
"loss": 0.6014,
"step": 2700
},
{
"epoch": 4.444444444444445,
"grad_norm": 0.5312401056289673,
"learning_rate": 0.00011719545091895564,
"loss": 0.594,
"step": 2710
},
{
"epoch": 4.460844608446084,
"grad_norm": 0.5242477655410767,
"learning_rate": 0.00011668704731815892,
"loss": 0.5914,
"step": 2720
},
{
"epoch": 4.477244772447724,
"grad_norm": 0.48659422993659973,
"learning_rate": 0.0001161781996552765,
"loss": 0.5927,
"step": 2730
},
{
"epoch": 4.493644936449364,
"grad_norm": 0.5395858287811279,
"learning_rate": 0.00011566892147134705,
"loss": 0.5961,
"step": 2740
},
{
"epoch": 4.510045100451005,
"grad_norm": 0.503237783908844,
"learning_rate": 0.00011515922631886605,
"loss": 0.5763,
"step": 2750
},
{
"epoch": 4.5264452644526445,
"grad_norm": 0.5137144923210144,
"learning_rate": 0.00011464912776142494,
"loss": 0.5912,
"step": 2760
},
{
"epoch": 4.542845428454284,
"grad_norm": 0.5336122512817383,
"learning_rate": 0.00011413863937335028,
"loss": 0.5857,
"step": 2770
},
{
"epoch": 4.559245592455925,
"grad_norm": 0.5271728038787842,
"learning_rate": 0.00011362777473934248,
"loss": 0.5915,
"step": 2780
},
{
"epoch": 4.575645756457565,
"grad_norm": 0.5467897057533264,
"learning_rate": 0.00011311654745411425,
"loss": 0.5871,
"step": 2790
},
{
"epoch": 4.592045920459205,
"grad_norm": 0.5065737962722778,
"learning_rate": 0.00011260497112202895,
"loss": 0.5867,
"step": 2800
},
{
"epoch": 4.608446084460844,
"grad_norm": 0.5274237990379333,
"learning_rate": 0.00011209305935673844,
"loss": 0.5885,
"step": 2810
},
{
"epoch": 4.624846248462484,
"grad_norm": 0.5007864832878113,
"learning_rate": 0.00011158082578082089,
"loss": 0.5965,
"step": 2820
},
{
"epoch": 4.641246412464125,
"grad_norm": 0.48857831954956055,
"learning_rate": 0.0001110682840254182,
"loss": 0.5898,
"step": 2830
},
{
"epoch": 4.657646576465765,
"grad_norm": 0.5346599221229553,
"learning_rate": 0.00011055544772987335,
"loss": 0.5941,
"step": 2840
},
{
"epoch": 4.674046740467404,
"grad_norm": 0.4968061149120331,
"learning_rate": 0.00011004233054136725,
"loss": 0.5852,
"step": 2850
},
{
"epoch": 4.690446904469045,
"grad_norm": 0.5415840744972229,
"learning_rate": 0.00010952894611455584,
"loss": 0.5891,
"step": 2860
},
{
"epoch": 4.706847068470685,
"grad_norm": 0.5450642704963684,
"learning_rate": 0.00010901530811120655,
"loss": 0.5758,
"step": 2870
},
{
"epoch": 4.723247232472325,
"grad_norm": 0.5052106976509094,
"learning_rate": 0.00010850143019983474,
"loss": 0.5736,
"step": 2880
},
{
"epoch": 4.7396473964739645,
"grad_norm": 0.5136608481407166,
"learning_rate": 0.00010798732605534006,
"loss": 0.5769,
"step": 2890
},
{
"epoch": 4.756047560475604,
"grad_norm": 0.5369486808776855,
"learning_rate": 0.00010747300935864243,
"loss": 0.5762,
"step": 2900
},
{
"epoch": 4.772447724477245,
"grad_norm": 0.5404406189918518,
"learning_rate": 0.00010695849379631813,
"loss": 0.5862,
"step": 2910
},
{
"epoch": 4.788847888478885,
"grad_norm": 0.5491173267364502,
"learning_rate": 0.0001064437930602354,
"loss": 0.5783,
"step": 2920
},
{
"epoch": 4.805248052480525,
"grad_norm": 0.5794305801391602,
"learning_rate": 0.00010592892084719024,
"loss": 0.596,
"step": 2930
},
{
"epoch": 4.821648216482165,
"grad_norm": 0.5437578558921814,
"learning_rate": 0.00010541389085854176,
"loss": 0.5892,
"step": 2940
},
{
"epoch": 4.838048380483805,
"grad_norm": 0.5468413233757019,
"learning_rate": 0.00010489871679984777,
"loss": 0.57,
"step": 2950
},
{
"epoch": 4.854448544485445,
"grad_norm": 0.5444806814193726,
"learning_rate": 0.00010438341238049991,
"loss": 0.5754,
"step": 2960
},
{
"epoch": 4.870848708487085,
"grad_norm": 0.5279709696769714,
"learning_rate": 0.00010386799131335889,
"loss": 0.5757,
"step": 2970
},
{
"epoch": 4.887248872488724,
"grad_norm": 0.5229681134223938,
"learning_rate": 0.00010335246731438948,
"loss": 0.5851,
"step": 2980
},
{
"epoch": 4.903649036490365,
"grad_norm": 0.5356857180595398,
"learning_rate": 0.00010283685410229571,
"loss": 0.5846,
"step": 2990
},
{
"epoch": 4.920049200492005,
"grad_norm": 0.5306787490844727,
"learning_rate": 0.00010232116539815558,
"loss": 0.5595,
"step": 3000
},
{
"epoch": 4.936449364493645,
"grad_norm": 0.5103681683540344,
"learning_rate": 0.00010180541492505604,
"loss": 0.5878,
"step": 3010
},
{
"epoch": 4.952849528495285,
"grad_norm": 0.528273344039917,
"learning_rate": 0.00010128961640772785,
"loss": 0.5684,
"step": 3020
},
{
"epoch": 4.969249692496925,
"grad_norm": 0.5309355854988098,
"learning_rate": 0.00010077378357218021,
"loss": 0.5949,
"step": 3030
},
{
"epoch": 4.985649856498565,
"grad_norm": 0.5428647994995117,
"learning_rate": 0.00010025793014533558,
"loss": 0.5893,
"step": 3040
},
{
"epoch": 5.002050020500205,
"grad_norm": 0.519814133644104,
"learning_rate": 9.974206985466442e-05,
"loss": 0.5898,
"step": 3050
},
{
"epoch": 5.018450184501845,
"grad_norm": 0.5455579161643982,
"learning_rate": 9.92262164278198e-05,
"loss": 0.5495,
"step": 3060
},
{
"epoch": 5.034850348503485,
"grad_norm": 0.5572001338005066,
"learning_rate": 9.871038359227214e-05,
"loss": 0.543,
"step": 3070
},
{
"epoch": 5.051250512505125,
"grad_norm": 0.5723180174827576,
"learning_rate": 9.819458507494394e-05,
"loss": 0.5557,
"step": 3080
},
{
"epoch": 5.067650676506765,
"grad_norm": 0.5615038275718689,
"learning_rate": 9.767883460184443e-05,
"loss": 0.5505,
"step": 3090
},
{
"epoch": 5.0840508405084055,
"grad_norm": 0.5597648024559021,
"learning_rate": 9.71631458977043e-05,
"loss": 0.5599,
"step": 3100
},
{
"epoch": 5.100451004510045,
"grad_norm": 0.5660073161125183,
"learning_rate": 9.66475326856105e-05,
"loss": 0.5609,
"step": 3110
},
{
"epoch": 5.116851168511685,
"grad_norm": 0.5740461945533752,
"learning_rate": 9.613200868664112e-05,
"loss": 0.5505,
"step": 3120
},
{
"epoch": 5.133251332513325,
"grad_norm": 0.5641244649887085,
"learning_rate": 9.561658761950007e-05,
"loss": 0.5725,
"step": 3130
},
{
"epoch": 5.149651496514966,
"grad_norm": 0.5780383944511414,
"learning_rate": 9.510128320015224e-05,
"loss": 0.5584,
"step": 3140
},
{
"epoch": 5.166051660516605,
"grad_norm": 0.5865011215209961,
"learning_rate": 9.458610914145826e-05,
"loss": 0.555,
"step": 3150
},
{
"epoch": 5.182451824518245,
"grad_norm": 0.5646878480911255,
"learning_rate": 9.40710791528098e-05,
"loss": 0.5524,
"step": 3160
},
{
"epoch": 5.198851988519885,
"grad_norm": 0.6092641353607178,
"learning_rate": 9.355620693976461e-05,
"loss": 0.5704,
"step": 3170
},
{
"epoch": 5.215252152521526,
"grad_norm": 0.5585269927978516,
"learning_rate": 9.304150620368188e-05,
"loss": 0.5652,
"step": 3180
},
{
"epoch": 5.231652316523165,
"grad_norm": 0.5374350547790527,
"learning_rate": 9.252699064135758e-05,
"loss": 0.5591,
"step": 3190
},
{
"epoch": 5.248052480524805,
"grad_norm": 0.5892452001571655,
"learning_rate": 9.201267394465998e-05,
"loss": 0.568,
"step": 3200
},
{
"epoch": 5.264452644526445,
"grad_norm": 0.5627169609069824,
"learning_rate": 9.149856980016529e-05,
"loss": 0.5573,
"step": 3210
},
{
"epoch": 5.280852808528086,
"grad_norm": 0.6352980732917786,
"learning_rate": 9.098469188879349e-05,
"loss": 0.5548,
"step": 3220
},
{
"epoch": 5.2972529725297255,
"grad_norm": 0.5747233629226685,
"learning_rate": 9.047105388544417e-05,
"loss": 0.5802,
"step": 3230
},
{
"epoch": 5.313653136531365,
"grad_norm": 0.66510009765625,
"learning_rate": 8.995766945863277e-05,
"loss": 0.5583,
"step": 3240
},
{
"epoch": 5.330053300533005,
"grad_norm": 0.6103816628456116,
"learning_rate": 8.944455227012666e-05,
"loss": 0.5622,
"step": 3250
},
{
"epoch": 5.346453464534646,
"grad_norm": 0.53193598985672,
"learning_rate": 8.89317159745818e-05,
"loss": 0.5622,
"step": 3260
},
{
"epoch": 5.3628536285362856,
"grad_norm": 0.5971400737762451,
"learning_rate": 8.841917421917912e-05,
"loss": 0.5407,
"step": 3270
},
{
"epoch": 5.379253792537925,
"grad_norm": 0.5562443733215332,
"learning_rate": 8.790694064326157e-05,
"loss": 0.5544,
"step": 3280
},
{
"epoch": 5.395653956539565,
"grad_norm": 0.6092881560325623,
"learning_rate": 8.739502887797107e-05,
"loss": 0.5543,
"step": 3290
},
{
"epoch": 5.412054120541206,
"grad_norm": 0.5391745567321777,
"learning_rate": 8.688345254588578e-05,
"loss": 0.5553,
"step": 3300
},
{
"epoch": 5.428454284542846,
"grad_norm": 0.558480978012085,
"learning_rate": 8.637222526065756e-05,
"loss": 0.5614,
"step": 3310
},
{
"epoch": 5.444854448544485,
"grad_norm": 0.5935754776000977,
"learning_rate": 8.586136062664974e-05,
"loss": 0.5569,
"step": 3320
},
{
"epoch": 5.461254612546125,
"grad_norm": 0.5984821319580078,
"learning_rate": 8.535087223857508e-05,
"loss": 0.5625,
"step": 3330
},
{
"epoch": 5.477654776547766,
"grad_norm": 0.5659565925598145,
"learning_rate": 8.484077368113399e-05,
"loss": 0.5483,
"step": 3340
},
{
"epoch": 5.494054940549406,
"grad_norm": 0.5952328443527222,
"learning_rate": 8.433107852865298e-05,
"loss": 0.5546,
"step": 3350
},
{
"epoch": 5.5104551045510455,
"grad_norm": 0.5910452604293823,
"learning_rate": 8.382180034472353e-05,
"loss": 0.5371,
"step": 3360
},
{
"epoch": 5.526855268552685,
"grad_norm": 0.5623068809509277,
"learning_rate": 8.33129526818411e-05,
"loss": 0.5463,
"step": 3370
},
{
"epoch": 5.543255432554325,
"grad_norm": 0.557976484298706,
"learning_rate": 8.280454908104439e-05,
"loss": 0.555,
"step": 3380
},
{
"epoch": 5.559655596555966,
"grad_norm": 0.5730746984481812,
"learning_rate": 8.229660307155518e-05,
"loss": 0.5612,
"step": 3390
},
{
"epoch": 5.5760557605576055,
"grad_norm": 0.5673303604125977,
"learning_rate": 8.178912817041817e-05,
"loss": 0.5583,
"step": 3400
},
{
"epoch": 5.592455924559245,
"grad_norm": 0.5584287047386169,
"learning_rate": 8.128213788214126e-05,
"loss": 0.5471,
"step": 3410
},
{
"epoch": 5.608856088560886,
"grad_norm": 0.5875297784805298,
"learning_rate": 8.077564569833632e-05,
"loss": 0.5576,
"step": 3420
},
{
"epoch": 5.625256252562526,
"grad_norm": 0.5899044275283813,
"learning_rate": 8.026966509736001e-05,
"loss": 0.5437,
"step": 3430
},
{
"epoch": 5.641656416564166,
"grad_norm": 0.6080020666122437,
"learning_rate": 7.976420954395518e-05,
"loss": 0.5552,
"step": 3440
},
{
"epoch": 5.658056580565805,
"grad_norm": 0.594315230846405,
"learning_rate": 7.92592924888925e-05,
"loss": 0.5587,
"step": 3450
},
{
"epoch": 5.674456744567445,
"grad_norm": 0.5834687948226929,
"learning_rate": 7.875492736861266e-05,
"loss": 0.57,
"step": 3460
},
{
"epoch": 5.690856908569086,
"grad_norm": 0.5709730982780457,
"learning_rate": 7.825112760486861e-05,
"loss": 0.5737,
"step": 3470
},
{
"epoch": 5.707257072570726,
"grad_norm": 0.5856568217277527,
"learning_rate": 7.774790660436858e-05,
"loss": 0.5484,
"step": 3480
},
{
"epoch": 5.7236572365723655,
"grad_norm": 0.5535131692886353,
"learning_rate": 7.724527775841914e-05,
"loss": 0.5483,
"step": 3490
},
{
"epoch": 5.740057400574006,
"grad_norm": 0.5895147323608398,
"learning_rate": 7.674325444256899e-05,
"loss": 0.5463,
"step": 3500
},
{
"epoch": 5.756457564575646,
"grad_norm": 0.57289719581604,
"learning_rate": 7.624185001625292e-05,
"loss": 0.5569,
"step": 3510
},
{
"epoch": 5.772857728577286,
"grad_norm": 0.5502966046333313,
"learning_rate": 7.574107782243634e-05,
"loss": 0.5696,
"step": 3520
},
{
"epoch": 5.7892578925789255,
"grad_norm": 0.5804471969604492,
"learning_rate": 7.524095118726025e-05,
"loss": 0.5506,
"step": 3530
},
{
"epoch": 5.805658056580565,
"grad_norm": 0.5519583225250244,
"learning_rate": 7.474148341968652e-05,
"loss": 0.5571,
"step": 3540
},
{
"epoch": 5.822058220582206,
"grad_norm": 0.5806939005851746,
"learning_rate": 7.42426878111438e-05,
"loss": 0.5556,
"step": 3550
},
{
"epoch": 5.838458384583846,
"grad_norm": 0.5681191682815552,
"learning_rate": 7.374457763517376e-05,
"loss": 0.5557,
"step": 3560
},
{
"epoch": 5.854858548585486,
"grad_norm": 0.6004670262336731,
"learning_rate": 7.324716614707793e-05,
"loss": 0.5686,
"step": 3570
},
{
"epoch": 5.871258712587126,
"grad_norm": 0.5577644109725952,
"learning_rate": 7.275046658356494e-05,
"loss": 0.5619,
"step": 3580
},
{
"epoch": 5.887658876588766,
"grad_norm": 0.5934208631515503,
"learning_rate": 7.225449216239821e-05,
"loss": 0.5593,
"step": 3590
},
{
"epoch": 5.904059040590406,
"grad_norm": 0.5490043759346008,
"learning_rate": 7.175925608204428e-05,
"loss": 0.5404,
"step": 3600
},
{
"epoch": 5.920459204592046,
"grad_norm": 0.662482500076294,
"learning_rate": 7.126477152132164e-05,
"loss": 0.5525,
"step": 3610
},
{
"epoch": 5.9368593685936855,
"grad_norm": 0.5766283869743347,
"learning_rate": 7.077105163904987e-05,
"loss": 0.5706,
"step": 3620
},
{
"epoch": 5.953259532595326,
"grad_norm": 0.5751728415489197,
"learning_rate": 7.027810957369957e-05,
"loss": 0.5626,
"step": 3630
},
{
"epoch": 5.969659696596966,
"grad_norm": 0.5690221190452576,
"learning_rate": 6.978595844304271e-05,
"loss": 0.558,
"step": 3640
},
{
"epoch": 5.986059860598606,
"grad_norm": 0.5637174248695374,
"learning_rate": 6.92946113438036e-05,
"loss": 0.5665,
"step": 3650
},
{
"epoch": 6.002460024600246,
"grad_norm": 0.5749226808547974,
"learning_rate": 6.880408135131022e-05,
"loss": 0.5687,
"step": 3660
},
{
"epoch": 6.018860188601886,
"grad_norm": 0.581462562084198,
"learning_rate": 6.831438151914649e-05,
"loss": 0.5272,
"step": 3670
},
{
"epoch": 6.035260352603526,
"grad_norm": 0.585290253162384,
"learning_rate": 6.782552487880468e-05,
"loss": 0.5347,
"step": 3680
},
{
"epoch": 6.051660516605166,
"grad_norm": 0.6168552041053772,
"learning_rate": 6.733752443933878e-05,
"loss": 0.5224,
"step": 3690
},
{
"epoch": 6.0680606806068065,
"grad_norm": 0.5952621698379517,
"learning_rate": 6.685039318701826e-05,
"loss": 0.5267,
"step": 3700
},
{
"epoch": 6.084460844608446,
"grad_norm": 0.708752453327179,
"learning_rate": 6.636414408498249e-05,
"loss": 0.5323,
"step": 3710
},
{
"epoch": 6.100861008610086,
"grad_norm": 0.6072937250137329,
"learning_rate": 6.587879007289576e-05,
"loss": 0.5361,
"step": 3720
},
{
"epoch": 6.117261172611726,
"grad_norm": 0.6426242589950562,
"learning_rate": 6.539434406660296e-05,
"loss": 0.5439,
"step": 3730
},
{
"epoch": 6.1336613366133665,
"grad_norm": 0.6066299080848694,
"learning_rate": 6.491081895778588e-05,
"loss": 0.5256,
"step": 3740
},
{
"epoch": 6.150061500615006,
"grad_norm": 0.578899621963501,
"learning_rate": 6.442822761362015e-05,
"loss": 0.5373,
"step": 3750
},
{
"epoch": 6.166461664616646,
"grad_norm": 0.6230922341346741,
"learning_rate": 6.394658287643278e-05,
"loss": 0.5249,
"step": 3760
},
{
"epoch": 6.182861828618286,
"grad_norm": 0.610723614692688,
"learning_rate": 6.34658975633605e-05,
"loss": 0.5417,
"step": 3770
},
{
"epoch": 6.199261992619927,
"grad_norm": 0.6284223198890686,
"learning_rate": 6.298618446600856e-05,
"loss": 0.5361,
"step": 3780
},
{
"epoch": 6.215662156621566,
"grad_norm": 0.6320038437843323,
"learning_rate": 6.250745635011048e-05,
"loss": 0.5344,
"step": 3790
},
{
"epoch": 6.232062320623206,
"grad_norm": 0.5748194456100464,
"learning_rate": 6.202972595518817e-05,
"loss": 0.5285,
"step": 3800
},
{
"epoch": 6.248462484624846,
"grad_norm": 0.6720646023750305,
"learning_rate": 6.155300599421306e-05,
"loss": 0.5399,
"step": 3810
},
{
"epoch": 6.264862648626487,
"grad_norm": 0.6088182330131531,
"learning_rate": 6.107730915326772e-05,
"loss": 0.5387,
"step": 3820
},
{
"epoch": 6.2812628126281265,
"grad_norm": 0.6509405970573425,
"learning_rate": 6.0602648091208324e-05,
"loss": 0.5356,
"step": 3830
},
{
"epoch": 6.297662976629766,
"grad_norm": 0.6377142667770386,
"learning_rate": 6.012903543932766e-05,
"loss": 0.5328,
"step": 3840
},
{
"epoch": 6.314063140631406,
"grad_norm": 0.638660728931427,
"learning_rate": 5.965648380101916e-05,
"loss": 0.5403,
"step": 3850
},
{
"epoch": 6.330463304633047,
"grad_norm": 0.6343023180961609,
"learning_rate": 5.918500575144138e-05,
"loss": 0.5316,
"step": 3860
},
{
"epoch": 6.3468634686346865,
"grad_norm": 0.6120122075080872,
"learning_rate": 5.871461383718344e-05,
"loss": 0.5475,
"step": 3870
},
{
"epoch": 6.363263632636326,
"grad_norm": 0.664526104927063,
"learning_rate": 5.8245320575931085e-05,
"loss": 0.5539,
"step": 3880
},
{
"epoch": 6.379663796637966,
"grad_norm": 0.6195924878120422,
"learning_rate": 5.777713845613364e-05,
"loss": 0.5314,
"step": 3890
},
{
"epoch": 6.396063960639607,
"grad_norm": 0.6664077639579773,
"learning_rate": 5.7310079936671545e-05,
"loss": 0.5309,
"step": 3900
},
{
"epoch": 6.412464124641247,
"grad_norm": 0.6408997178077698,
"learning_rate": 5.684415744652509e-05,
"loss": 0.5346,
"step": 3910
},
{
"epoch": 6.428864288642886,
"grad_norm": 0.596552312374115,
"learning_rate": 5.6379383384443255e-05,
"loss": 0.5424,
"step": 3920
},
{
"epoch": 6.445264452644526,
"grad_norm": 0.6852882504463196,
"learning_rate": 5.59157701186142e-05,
"loss": 0.5367,
"step": 3930
},
{
"epoch": 6.461664616646167,
"grad_norm": 0.6832631826400757,
"learning_rate": 5.545332998633572e-05,
"loss": 0.5311,
"step": 3940
},
{
"epoch": 6.478064780647807,
"grad_norm": 0.6201120018959045,
"learning_rate": 5.499207529368734e-05,
"loss": 0.5398,
"step": 3950
},
{
"epoch": 6.4944649446494465,
"grad_norm": 0.6140998601913452,
"learning_rate": 5.453201831520245e-05,
"loss": 0.5388,
"step": 3960
},
{
"epoch": 6.510865108651086,
"grad_norm": 0.6650799512863159,
"learning_rate": 5.4073171293542016e-05,
"loss": 0.5419,
"step": 3970
},
{
"epoch": 6.527265272652727,
"grad_norm": 0.6716285347938538,
"learning_rate": 5.3615546439168485e-05,
"loss": 0.5234,
"step": 3980
},
{
"epoch": 6.543665436654367,
"grad_norm": 0.6595677733421326,
"learning_rate": 5.3159155930021e-05,
"loss": 0.5194,
"step": 3990
},
{
"epoch": 6.5600656006560065,
"grad_norm": 0.5784692168235779,
"learning_rate": 5.270401191119143e-05,
"loss": 0.5222,
"step": 4000
},
{
"epoch": 6.576465764657646,
"grad_norm": 0.6419820189476013,
"learning_rate": 5.2250126494600916e-05,
"loss": 0.5352,
"step": 4010
},
{
"epoch": 6.592865928659286,
"grad_norm": 0.6142451763153076,
"learning_rate": 5.179751175867784e-05,
"loss": 0.5357,
"step": 4020
},
{
"epoch": 6.609266092660927,
"grad_norm": 0.6860802173614502,
"learning_rate": 5.1346179748036116e-05,
"loss": 0.531,
"step": 4030
},
{
"epoch": 6.625666256662567,
"grad_norm": 0.6106059551239014,
"learning_rate": 5.0896142473154987e-05,
"loss": 0.5333,
"step": 4040
},
{
"epoch": 6.642066420664206,
"grad_norm": 0.6644913554191589,
"learning_rate": 5.044741191005908e-05,
"loss": 0.5403,
"step": 4050
},
{
"epoch": 6.658466584665847,
"grad_norm": 0.5942371487617493,
"learning_rate": 5.000000000000002e-05,
"loss": 0.5467,
"step": 4060
},
{
"epoch": 6.674866748667487,
"grad_norm": 0.617970883846283,
"learning_rate": 4.9553918649138386e-05,
"loss": 0.5325,
"step": 4070
},
{
"epoch": 6.691266912669127,
"grad_norm": 0.6382017135620117,
"learning_rate": 4.910917972822713e-05,
"loss": 0.525,
"step": 4080
},
{
"epoch": 6.7076670766707664,
"grad_norm": 0.6390166282653809,
"learning_rate": 4.866579507229545e-05,
"loss": 0.5356,
"step": 4090
},
{
"epoch": 6.724067240672406,
"grad_norm": 0.6630433201789856,
"learning_rate": 4.822377648033394e-05,
"loss": 0.541,
"step": 4100
},
{
"epoch": 6.740467404674047,
"grad_norm": 0.6468001008033752,
"learning_rate": 4.7783135714980744e-05,
"loss": 0.5399,
"step": 4110
},
{
"epoch": 6.756867568675687,
"grad_norm": 0.620833694934845,
"learning_rate": 4.734388450220825e-05,
"loss": 0.5265,
"step": 4120
},
{
"epoch": 6.7732677326773265,
"grad_norm": 0.61692875623703,
"learning_rate": 4.6906034531011346e-05,
"loss": 0.5303,
"step": 4130
},
{
"epoch": 6.789667896678967,
"grad_norm": 0.6430515646934509,
"learning_rate": 4.646959745309609e-05,
"loss": 0.543,
"step": 4140
},
{
"epoch": 6.806068060680607,
"grad_norm": 0.653596043586731,
"learning_rate": 4.603458488256992e-05,
"loss": 0.5384,
"step": 4150
},
{
"epoch": 6.822468224682247,
"grad_norm": 0.6125525832176208,
"learning_rate": 4.560100839563229e-05,
"loss": 0.5417,
"step": 4160
},
{
"epoch": 6.838868388683887,
"grad_norm": 0.5968248844146729,
"learning_rate": 4.516887953026691e-05,
"loss": 0.5326,
"step": 4170
},
{
"epoch": 6.855268552685526,
"grad_norm": 0.5980390906333923,
"learning_rate": 4.4738209785934505e-05,
"loss": 0.5313,
"step": 4180
},
{
"epoch": 6.871668716687167,
"grad_norm": 0.614548921585083,
"learning_rate": 4.430901062326681e-05,
"loss": 0.5321,
"step": 4190
},
{
"epoch": 6.888068880688807,
"grad_norm": 0.5848116278648376,
"learning_rate": 4.388129346376178e-05,
"loss": 0.542,
"step": 4200
},
{
"epoch": 6.904469044690447,
"grad_norm": 0.6172247529029846,
"learning_rate": 4.345506968947931e-05,
"loss": 0.5439,
"step": 4210
},
{
"epoch": 6.920869208692087,
"grad_norm": 0.6525529026985168,
"learning_rate": 4.303035064273878e-05,
"loss": 0.5319,
"step": 4220
},
{
"epoch": 6.937269372693727,
"grad_norm": 0.614452064037323,
"learning_rate": 4.260714762581677e-05,
"loss": 0.5287,
"step": 4230
},
{
"epoch": 6.953669536695367,
"grad_norm": 0.6268100142478943,
"learning_rate": 4.21854719006467e-05,
"loss": 0.5337,
"step": 4240
},
{
"epoch": 6.970069700697007,
"grad_norm": 0.6285504698753357,
"learning_rate": 4.1765334688518766e-05,
"loss": 0.5402,
"step": 4250
},
{
"epoch": 6.9864698646986465,
"grad_norm": 0.6280767917633057,
"learning_rate": 4.13467471697817e-05,
"loss": 0.533,
"step": 4260
},
{
"epoch": 7.002870028700287,
"grad_norm": 0.5814547538757324,
"learning_rate": 4.092972048354491e-05,
"loss": 0.5346,
"step": 4270
},
{
"epoch": 7.019270192701927,
"grad_norm": 0.6583351492881775,
"learning_rate": 4.0514265727382215e-05,
"loss": 0.5261,
"step": 4280
},
{
"epoch": 7.035670356703567,
"grad_norm": 0.6630005240440369,
"learning_rate": 4.010039395703664e-05,
"loss": 0.5196,
"step": 4290
},
{
"epoch": 7.0520705207052075,
"grad_norm": 0.6557034254074097,
"learning_rate": 3.968811618612592e-05,
"loss": 0.5161,
"step": 4300
},
{
"epoch": 7.068470684706847,
"grad_norm": 0.6600791215896606,
"learning_rate": 3.927744338584972e-05,
"loss": 0.5066,
"step": 4310
},
{
"epoch": 7.084870848708487,
"grad_norm": 0.6123402118682861,
"learning_rate": 3.8868386484697417e-05,
"loss": 0.5195,
"step": 4320
},
{
"epoch": 7.101271012710127,
"grad_norm": 0.6454523801803589,
"learning_rate": 3.84609563681575e-05,
"loss": 0.5113,
"step": 4330
},
{
"epoch": 7.1176711767117675,
"grad_norm": 0.6421491503715515,
"learning_rate": 3.80551638784277e-05,
"loss": 0.5134,
"step": 4340
},
{
"epoch": 7.134071340713407,
"grad_norm": 0.6418097615242004,
"learning_rate": 3.7651019814126654e-05,
"loss": 0.5185,
"step": 4350
},
{
"epoch": 7.150471504715047,
"grad_norm": 0.6425509452819824,
"learning_rate": 3.724853493000635e-05,
"loss": 0.5251,
"step": 4360
},
{
"epoch": 7.166871668716687,
"grad_norm": 0.6572188138961792,
"learning_rate": 3.6847719936666124e-05,
"loss": 0.5124,
"step": 4370
},
{
"epoch": 7.183271832718328,
"grad_norm": 0.6694904565811157,
"learning_rate": 3.6448585500267485e-05,
"loss": 0.5174,
"step": 4380
},
{
"epoch": 7.199671996719967,
"grad_norm": 0.6733431816101074,
"learning_rate": 3.605114224225028e-05,
"loss": 0.5227,
"step": 4390
},
{
"epoch": 7.216072160721607,
"grad_norm": 0.6532800793647766,
"learning_rate": 3.565540073905025e-05,
"loss": 0.5128,
"step": 4400
},
{
"epoch": 7.232472324723247,
"grad_norm": 0.6475611329078674,
"learning_rate": 3.5261371521817244e-05,
"loss": 0.5282,
"step": 4410
},
{
"epoch": 7.248872488724888,
"grad_norm": 0.6134998202323914,
"learning_rate": 3.486906507613531e-05,
"loss": 0.5118,
"step": 4420
},
{
"epoch": 7.2652726527265274,
"grad_norm": 0.648802638053894,
"learning_rate": 3.4478491841743397e-05,
"loss": 0.5253,
"step": 4430
},
{
"epoch": 7.281672816728167,
"grad_norm": 0.6617629528045654,
"learning_rate": 3.408966221225773e-05,
"loss": 0.5177,
"step": 4440
},
{
"epoch": 7.298072980729807,
"grad_norm": 0.6465151906013489,
"learning_rate": 3.370258653489505e-05,
"loss": 0.5133,
"step": 4450
},
{
"epoch": 7.314473144731448,
"grad_norm": 0.6327465176582336,
"learning_rate": 3.331727511019749e-05,
"loss": 0.5155,
"step": 4460
},
{
"epoch": 7.3308733087330875,
"grad_norm": 0.6604763865470886,
"learning_rate": 3.293373819175816e-05,
"loss": 0.519,
"step": 4470
},
{
"epoch": 7.347273472734727,
"grad_norm": 0.681736171245575,
"learning_rate": 3.2551985985948616e-05,
"loss": 0.5243,
"step": 4480
},
{
"epoch": 7.363673636736367,
"grad_norm": 0.6531623601913452,
"learning_rate": 3.217202865164697e-05,
"loss": 0.5201,
"step": 4490
},
{
"epoch": 7.380073800738008,
"grad_norm": 0.674201250076294,
"learning_rate": 3.1793876299967816e-05,
"loss": 0.5252,
"step": 4500
},
{
"epoch": 7.396473964739648,
"grad_norm": 0.6714997291564941,
"learning_rate": 3.141753899399289e-05,
"loss": 0.5253,
"step": 4510
},
{
"epoch": 7.412874128741287,
"grad_norm": 0.6872825622558594,
"learning_rate": 3.104302674850346e-05,
"loss": 0.5311,
"step": 4520
},
{
"epoch": 7.429274292742927,
"grad_norm": 0.7338685989379883,
"learning_rate": 3.0670349529713816e-05,
"loss": 0.515,
"step": 4530
},
{
"epoch": 7.445674456744568,
"grad_norm": 0.6552906036376953,
"learning_rate": 3.0299517255005937e-05,
"loss": 0.5077,
"step": 4540
},
{
"epoch": 7.462074620746208,
"grad_norm": 0.6388695240020752,
"learning_rate": 2.993053979266577e-05,
"loss": 0.5296,
"step": 4550
},
{
"epoch": 7.478474784747847,
"grad_norm": 0.6324784159660339,
"learning_rate": 2.9563426961620367e-05,
"loss": 0.5249,
"step": 4560
},
{
"epoch": 7.494874948749487,
"grad_norm": 0.6526088714599609,
"learning_rate": 2.9198188531176863e-05,
"loss": 0.5164,
"step": 4570
},
{
"epoch": 7.511275112751127,
"grad_norm": 0.6632060408592224,
"learning_rate": 2.883483422076225e-05,
"loss": 0.5185,
"step": 4580
},
{
"epoch": 7.527675276752768,
"grad_norm": 0.643341600894928,
"learning_rate": 2.8473373699664997e-05,
"loss": 0.5162,
"step": 4590
},
{
"epoch": 7.5440754407544075,
"grad_norm": 0.6466885805130005,
"learning_rate": 2.811381658677744e-05,
"loss": 0.5103,
"step": 4600
},
{
"epoch": 7.560475604756047,
"grad_norm": 0.6679097414016724,
"learning_rate": 2.7756172450340134e-05,
"loss": 0.5238,
"step": 4610
},
{
"epoch": 7.576875768757688,
"grad_norm": 0.6607965230941772,
"learning_rate": 2.7400450807686938e-05,
"loss": 0.5104,
"step": 4620
},
{
"epoch": 7.593275932759328,
"grad_norm": 0.64109867811203,
"learning_rate": 2.70466611249919e-05,
"loss": 0.5199,
"step": 4630
},
{
"epoch": 7.609676096760968,
"grad_norm": 0.6699521541595459,
"learning_rate": 2.669481281701739e-05,
"loss": 0.5339,
"step": 4640
},
{
"epoch": 7.626076260762607,
"grad_norm": 0.6698071360588074,
"learning_rate": 2.6344915246863412e-05,
"loss": 0.5128,
"step": 4650
},
{
"epoch": 7.642476424764247,
"grad_norm": 0.6424985527992249,
"learning_rate": 2.5996977725718607e-05,
"loss": 0.5053,
"step": 4660
},
{
"epoch": 7.658876588765888,
"grad_norm": 0.7101566195487976,
"learning_rate": 2.5651009512612312e-05,
"loss": 0.5311,
"step": 4670
},
{
"epoch": 7.675276752767528,
"grad_norm": 0.6432877779006958,
"learning_rate": 2.5307019814168342e-05,
"loss": 0.5216,
"step": 4680
},
{
"epoch": 7.691676916769167,
"grad_norm": 0.6850785613059998,
"learning_rate": 2.496501778435977e-05,
"loss": 0.5108,
"step": 4690
},
{
"epoch": 7.708077080770808,
"grad_norm": 0.6734909415245056,
"learning_rate": 2.462501252426559e-05,
"loss": 0.5186,
"step": 4700
},
{
"epoch": 7.724477244772448,
"grad_norm": 0.7187743782997131,
"learning_rate": 2.4287013081828257e-05,
"loss": 0.5182,
"step": 4710
},
{
"epoch": 7.740877408774088,
"grad_norm": 0.7147987484931946,
"learning_rate": 2.3951028451613144e-05,
"loss": 0.5193,
"step": 4720
},
{
"epoch": 7.7572775727757275,
"grad_norm": 0.7066251635551453,
"learning_rate": 2.3617067574569087e-05,
"loss": 0.5175,
"step": 4730
},
{
"epoch": 7.773677736777367,
"grad_norm": 0.7206938862800598,
"learning_rate": 2.328513933779034e-05,
"loss": 0.5255,
"step": 4740
},
{
"epoch": 7.790077900779008,
"grad_norm": 0.6903772354125977,
"learning_rate": 2.2955252574280328e-05,
"loss": 0.512,
"step": 4750
},
{
"epoch": 7.806478064780648,
"grad_norm": 0.6938795447349548,
"learning_rate": 2.2627416062716366e-05,
"loss": 0.513,
"step": 4760
},
{
"epoch": 7.822878228782288,
"grad_norm": 0.6563030481338501,
"learning_rate": 2.2301638527216194e-05,
"loss": 0.5255,
"step": 4770
},
{
"epoch": 7.839278392783928,
"grad_norm": 0.676189124584198,
"learning_rate": 2.1977928637105692e-05,
"loss": 0.5331,
"step": 4780
},
{
"epoch": 7.855678556785568,
"grad_norm": 0.6740292906761169,
"learning_rate": 2.1656295006688353e-05,
"loss": 0.5161,
"step": 4790
},
{
"epoch": 7.872078720787208,
"grad_norm": 0.6368678212165833,
"learning_rate": 2.1336746195015846e-05,
"loss": 0.5181,
"step": 4800
},
{
"epoch": 7.888478884788848,
"grad_norm": 0.6666802763938904,
"learning_rate": 2.1019290705660356e-05,
"loss": 0.5248,
"step": 4810
},
{
"epoch": 7.904879048790487,
"grad_norm": 0.7021058797836304,
"learning_rate": 2.070393698648836e-05,
"loss": 0.5076,
"step": 4820
},
{
"epoch": 7.921279212792128,
"grad_norm": 0.6414440870285034,
"learning_rate": 2.0390693429435627e-05,
"loss": 0.5091,
"step": 4830
},
{
"epoch": 7.937679376793768,
"grad_norm": 0.6494581699371338,
"learning_rate": 2.0079568370284128e-05,
"loss": 0.5205,
"step": 4840
},
{
"epoch": 7.954079540795408,
"grad_norm": 0.6568921208381653,
"learning_rate": 1.977057008844e-05,
"loss": 0.5161,
"step": 4850
},
{
"epoch": 7.970479704797048,
"grad_norm": 0.6623067855834961,
"learning_rate": 1.946370680671341e-05,
"loss": 0.5134,
"step": 4860
},
{
"epoch": 7.986879868798688,
"grad_norm": 0.6465177536010742,
"learning_rate": 1.9158986691099544e-05,
"loss": 0.5097,
"step": 4870
},
{
"epoch": 8.003280032800328,
"grad_norm": 0.6603899002075195,
"learning_rate": 1.885641785056149e-05,
"loss": 0.5152,
"step": 4880
},
{
"epoch": 8.019680196801968,
"grad_norm": 0.6793326735496521,
"learning_rate": 1.85560083368143e-05,
"loss": 0.5201,
"step": 4890
},
{
"epoch": 8.036080360803608,
"grad_norm": 0.6705737709999084,
"learning_rate": 1.825776614411082e-05,
"loss": 0.4945,
"step": 4900
},
{
"epoch": 8.052480524805247,
"grad_norm": 0.6642212271690369,
"learning_rate": 1.7961699209028905e-05,
"loss": 0.5019,
"step": 4910
},
{
"epoch": 8.068880688806889,
"grad_norm": 0.6646486520767212,
"learning_rate": 1.766781541026018e-05,
"loss": 0.5245,
"step": 4920
},
{
"epoch": 8.085280852808529,
"grad_norm": 0.6514326930046082,
"learning_rate": 1.7376122568400532e-05,
"loss": 0.5083,
"step": 4930
},
{
"epoch": 8.101681016810169,
"grad_norm": 0.6762531995773315,
"learning_rate": 1.708662844574178e-05,
"loss": 0.5144,
"step": 4940
},
{
"epoch": 8.118081180811808,
"grad_norm": 0.6511978507041931,
"learning_rate": 1.679934074606533e-05,
"loss": 0.5045,
"step": 4950
},
{
"epoch": 8.134481344813448,
"grad_norm": 0.6896832585334778,
"learning_rate": 1.6514267114436945e-05,
"loss": 0.5089,
"step": 4960
},
{
"epoch": 8.150881508815088,
"grad_norm": 0.6686916947364807,
"learning_rate": 1.6231415137003537e-05,
"loss": 0.5048,
"step": 4970
},
{
"epoch": 8.167281672816728,
"grad_norm": 0.6836830377578735,
"learning_rate": 1.5950792340791043e-05,
"loss": 0.5072,
"step": 4980
},
{
"epoch": 8.183681836818367,
"grad_norm": 0.6605989336967468,
"learning_rate": 1.5672406193504384e-05,
"loss": 0.5115,
"step": 4990
},
{
"epoch": 8.200082000820009,
"grad_norm": 0.6973133087158203,
"learning_rate": 1.5396264103328474e-05,
"loss": 0.5132,
"step": 5000
},
{
"epoch": 8.216482164821649,
"grad_norm": 0.7265208959579468,
"learning_rate": 1.5122373418731306e-05,
"loss": 0.5036,
"step": 5010
},
{
"epoch": 8.232882328823289,
"grad_norm": 0.6954894661903381,
"learning_rate": 1.4850741428268244e-05,
"loss": 0.5037,
"step": 5020
},
{
"epoch": 8.249282492824928,
"grad_norm": 0.655786395072937,
"learning_rate": 1.4581375360388183e-05,
"loss": 0.515,
"step": 5030
},
{
"epoch": 8.265682656826568,
"grad_norm": 0.674171507358551,
"learning_rate": 1.4314282383241096e-05,
"loss": 0.4967,
"step": 5040
},
{
"epoch": 8.282082820828208,
"grad_norm": 0.7670585513114929,
"learning_rate": 1.4049469604487297e-05,
"loss": 0.4991,
"step": 5050
},
{
"epoch": 8.298482984829848,
"grad_norm": 0.6614052057266235,
"learning_rate": 1.3786944071108398e-05,
"loss": 0.5169,
"step": 5060
},
{
"epoch": 8.314883148831488,
"grad_norm": 0.7165561318397522,
"learning_rate": 1.3526712769219618e-05,
"loss": 0.502,
"step": 5070
},
{
"epoch": 8.33128331283313,
"grad_norm": 0.7127593159675598,
"learning_rate": 1.3268782623884047e-05,
"loss": 0.5091,
"step": 5080
},
{
"epoch": 8.347683476834769,
"grad_norm": 0.6743724346160889,
"learning_rate": 1.301316049892818e-05,
"loss": 0.5019,
"step": 5090
},
{
"epoch": 8.364083640836409,
"grad_norm": 0.6609333157539368,
"learning_rate": 1.2759853196759453e-05,
"loss": 0.5052,
"step": 5100
},
{
"epoch": 8.380483804838049,
"grad_norm": 0.7156481146812439,
"learning_rate": 1.2508867458185037e-05,
"loss": 0.51,
"step": 5110
},
{
"epoch": 8.396883968839688,
"grad_norm": 0.7323744893074036,
"learning_rate": 1.2260209962232628e-05,
"loss": 0.5092,
"step": 5120
},
{
"epoch": 8.413284132841328,
"grad_norm": 0.6540841460227966,
"learning_rate": 1.201388732597255e-05,
"loss": 0.507,
"step": 5130
},
{
"epoch": 8.429684296842968,
"grad_norm": 0.6970362663269043,
"learning_rate": 1.1769906104341832e-05,
"loss": 0.5216,
"step": 5140
},
{
"epoch": 8.446084460844608,
"grad_norm": 0.6555099487304688,
"learning_rate": 1.1528272789969618e-05,
"loss": 0.5001,
"step": 5150
},
{
"epoch": 8.46248462484625,
"grad_norm": 0.6940327286720276,
"learning_rate": 1.1288993813004467e-05,
"loss": 0.5025,
"step": 5160
},
{
"epoch": 8.478884788847889,
"grad_norm": 0.6825557351112366,
"learning_rate": 1.1052075540943296e-05,
"loss": 0.5089,
"step": 5170
},
{
"epoch": 8.495284952849529,
"grad_norm": 0.6553847789764404,
"learning_rate": 1.0817524278461776e-05,
"loss": 0.5052,
"step": 5180
},
{
"epoch": 8.511685116851169,
"grad_norm": 0.6892216205596924,
"learning_rate": 1.0585346267246743e-05,
"loss": 0.5158,
"step": 5190
},
{
"epoch": 8.528085280852808,
"grad_norm": 0.6756864786148071,
"learning_rate": 1.0355547685829926e-05,
"loss": 0.5133,
"step": 5200
},
{
"epoch": 8.544485444854448,
"grad_norm": 0.7741680145263672,
"learning_rate": 1.0128134649423671e-05,
"loss": 0.5167,
"step": 5210
},
{
"epoch": 8.560885608856088,
"grad_norm": 0.6552737355232239,
"learning_rate": 9.903113209758096e-06,
"loss": 0.5094,
"step": 5220
},
{
"epoch": 8.577285772857728,
"grad_norm": 0.6564160585403442,
"learning_rate": 9.680489354920152e-06,
"loss": 0.5048,
"step": 5230
},
{
"epoch": 8.59368593685937,
"grad_norm": 0.7316927313804626,
"learning_rate": 9.460269009194167e-06,
"loss": 0.499,
"step": 5240
},
{
"epoch": 8.61008610086101,
"grad_norm": 0.7389256358146667,
"learning_rate": 9.242458032904311e-06,
"loss": 0.5022,
"step": 5250
},
{
"epoch": 8.626486264862649,
"grad_norm": 0.6633957624435425,
"learning_rate": 9.027062222258487e-06,
"loss": 0.5042,
"step": 5260
},
{
"epoch": 8.642886428864289,
"grad_norm": 0.640006422996521,
"learning_rate": 8.814087309194251e-06,
"loss": 0.5068,
"step": 5270
},
{
"epoch": 8.659286592865929,
"grad_norm": 0.6830674409866333,
"learning_rate": 8.603538961226232e-06,
"loss": 0.502,
"step": 5280
},
{
"epoch": 8.675686756867568,
"grad_norm": 0.6639739871025085,
"learning_rate": 8.395422781295192e-06,
"loss": 0.5177,
"step": 5290
},
{
"epoch": 8.692086920869208,
"grad_norm": 0.6661989092826843,
"learning_rate": 8.189744307619118e-06,
"loss": 0.5014,
"step": 5300
},
{
"epoch": 8.708487084870848,
"grad_norm": 0.6645560264587402,
"learning_rate": 7.986509013545673e-06,
"loss": 0.5017,
"step": 5310
},
{
"epoch": 8.724887248872488,
"grad_norm": 0.7042533159255981,
"learning_rate": 7.785722307406684e-06,
"loss": 0.5163,
"step": 5320
},
{
"epoch": 8.74128741287413,
"grad_norm": 0.698277473449707,
"learning_rate": 7.587389532374123e-06,
"loss": 0.5054,
"step": 5330
},
{
"epoch": 8.75768757687577,
"grad_norm": 0.6191078424453735,
"learning_rate": 7.3915159663179075e-06,
"loss": 0.4992,
"step": 5340
},
{
"epoch": 8.774087740877409,
"grad_norm": 0.6769982576370239,
"learning_rate": 7.198106821665585e-06,
"loss": 0.5147,
"step": 5350
},
{
"epoch": 8.790487904879049,
"grad_norm": 0.6643761396408081,
"learning_rate": 7.007167245263435e-06,
"loss": 0.5105,
"step": 5360
},
{
"epoch": 8.806888068880689,
"grad_norm": 0.6664229035377502,
"learning_rate": 6.818702318239689e-06,
"loss": 0.5021,
"step": 5370
},
{
"epoch": 8.823288232882328,
"grad_norm": 0.6411841511726379,
"learning_rate": 6.632717055869164e-06,
"loss": 0.5076,
"step": 5380
},
{
"epoch": 8.839688396883968,
"grad_norm": 0.6643224954605103,
"learning_rate": 6.4492164074399065e-06,
"loss": 0.5044,
"step": 5390
},
{
"epoch": 8.85608856088561,
"grad_norm": 0.6722341775894165,
"learning_rate": 6.268205256121396e-06,
"loss": 0.5092,
"step": 5400
},
{
"epoch": 8.87248872488725,
"grad_norm": 0.6829484105110168,
"learning_rate": 6.089688418834727e-06,
"loss": 0.5164,
"step": 5410
},
{
"epoch": 8.88888888888889,
"grad_norm": 0.7037432789802551,
"learning_rate": 5.913670646124236e-06,
"loss": 0.5142,
"step": 5420
},
{
"epoch": 8.905289052890529,
"grad_norm": 0.6822744607925415,
"learning_rate": 5.7401566220313005e-06,
"loss": 0.5018,
"step": 5430
},
{
"epoch": 8.921689216892169,
"grad_norm": 0.6738746166229248,
"learning_rate": 5.569150963969494e-06,
"loss": 0.5013,
"step": 5440
},
{
"epoch": 8.938089380893809,
"grad_norm": 0.6965447664260864,
"learning_rate": 5.400658222601873e-06,
"loss": 0.5014,
"step": 5450
},
{
"epoch": 8.954489544895448,
"grad_norm": 0.7140269875526428,
"learning_rate": 5.2346828817197655e-06,
"loss": 0.5119,
"step": 5460
},
{
"epoch": 8.970889708897088,
"grad_norm": 0.6803216934204102,
"learning_rate": 5.071229358123464e-06,
"loss": 0.5021,
"step": 5470
},
{
"epoch": 8.987289872898728,
"grad_norm": 0.6827109456062317,
"learning_rate": 4.910302001504807e-06,
"loss": 0.5157,
"step": 5480
},
{
"epoch": 9.00369003690037,
"grad_norm": 0.6659403443336487,
"learning_rate": 4.7519050943312325e-06,
"loss": 0.4975,
"step": 5490
},
{
"epoch": 9.02009020090201,
"grad_norm": 0.7039642333984375,
"learning_rate": 4.596042851732008e-06,
"loss": 0.4949,
"step": 5500
},
{
"epoch": 9.03649036490365,
"grad_norm": 0.6756150722503662,
"learning_rate": 4.442719421385922e-06,
"loss": 0.4925,
"step": 5510
},
{
"epoch": 9.052890528905289,
"grad_norm": 0.684483528137207,
"learning_rate": 4.291938883411007e-06,
"loss": 0.4997,
"step": 5520
},
{
"epoch": 9.069290692906929,
"grad_norm": 0.6780720353126526,
"learning_rate": 4.143705250255869e-06,
"loss": 0.5104,
"step": 5530
},
{
"epoch": 9.085690856908569,
"grad_norm": 0.6675730347633362,
"learning_rate": 3.99802246659301e-06,
"loss": 0.5181,
"step": 5540
},
{
"epoch": 9.102091020910208,
"grad_norm": 0.6699404120445251,
"learning_rate": 3.85489440921376e-06,
"loss": 0.506,
"step": 5550
},
{
"epoch": 9.11849118491185,
"grad_norm": 0.6987513899803162,
"learning_rate": 3.7143248869252022e-06,
"loss": 0.5058,
"step": 5560
},
{
"epoch": 9.13489134891349,
"grad_norm": 0.6789569854736328,
"learning_rate": 3.5763176404487564e-06,
"loss": 0.4977,
"step": 5570
},
{
"epoch": 9.15129151291513,
"grad_norm": 0.6613333821296692,
"learning_rate": 3.440876342320609e-06,
"loss": 0.5096,
"step": 5580
},
{
"epoch": 9.16769167691677,
"grad_norm": 0.689217746257782,
"learning_rate": 3.308004596794101e-06,
"loss": 0.4949,
"step": 5590
},
{
"epoch": 9.18409184091841,
"grad_norm": 0.7100328803062439,
"learning_rate": 3.1777059397436692e-06,
"loss": 0.5044,
"step": 5600
},
{
"epoch": 9.200492004920049,
"grad_norm": 0.6680495738983154,
"learning_rate": 3.049983838570858e-06,
"loss": 0.5037,
"step": 5610
},
{
"epoch": 9.216892168921689,
"grad_norm": 0.6641035676002502,
"learning_rate": 2.9248416921119794e-06,
"loss": 0.4964,
"step": 5620
},
{
"epoch": 9.233292332923329,
"grad_norm": 0.7248251438140869,
"learning_rate": 2.8022828305477423e-06,
"loss": 0.5017,
"step": 5630
},
{
"epoch": 9.24969249692497,
"grad_norm": 0.7348821759223938,
"learning_rate": 2.682310515314512e-06,
"loss": 0.5077,
"step": 5640
},
{
"epoch": 9.26609266092661,
"grad_norm": 0.6882705688476562,
"learning_rate": 2.5649279390176806e-06,
"loss": 0.4926,
"step": 5650
},
{
"epoch": 9.28249282492825,
"grad_norm": 0.6873205900192261,
"learning_rate": 2.4501382253465543e-06,
"loss": 0.5009,
"step": 5660
},
{
"epoch": 9.29889298892989,
"grad_norm": 0.703940212726593,
"learning_rate": 2.3379444289913342e-06,
"loss": 0.5011,
"step": 5670
},
{
"epoch": 9.31529315293153,
"grad_norm": 0.6873738169670105,
"learning_rate": 2.228349535561769e-06,
"loss": 0.4991,
"step": 5680
},
{
"epoch": 9.331693316933169,
"grad_norm": 0.6344056725502014,
"learning_rate": 2.1213564615077065e-06,
"loss": 0.5099,
"step": 5690
},
{
"epoch": 9.348093480934809,
"grad_norm": 0.6923158168792725,
"learning_rate": 2.016968054041546e-06,
"loss": 0.5068,
"step": 5700
},
{
"epoch": 9.364493644936449,
"grad_norm": 0.6982558965682983,
"learning_rate": 1.915187091062387e-06,
"loss": 0.5052,
"step": 5710
},
{
"epoch": 9.38089380893809,
"grad_norm": 0.6965045928955078,
"learning_rate": 1.816016281082178e-06,
"loss": 0.4963,
"step": 5720
},
{
"epoch": 9.39729397293973,
"grad_norm": 0.704952597618103,
"learning_rate": 1.7194582631535617e-06,
"loss": 0.4922,
"step": 5730
},
{
"epoch": 9.41369413694137,
"grad_norm": 0.673637866973877,
"learning_rate": 1.6255156067997323e-06,
"loss": 0.4943,
"step": 5740
},
{
"epoch": 9.43009430094301,
"grad_norm": 0.6612069010734558,
"learning_rate": 1.5341908119459792e-06,
"loss": 0.4973,
"step": 5750
},
{
"epoch": 9.44649446494465,
"grad_norm": 0.7053012847900391,
"learning_rate": 1.4454863088532388e-06,
"loss": 0.5049,
"step": 5760
},
{
"epoch": 9.46289462894629,
"grad_norm": 0.6800163388252258,
"learning_rate": 1.3594044580533482e-06,
"loss": 0.5009,
"step": 5770
},
{
"epoch": 9.479294792947929,
"grad_norm": 0.7071276307106018,
"learning_rate": 1.2759475502862828e-06,
"loss": 0.5016,
"step": 5780
},
{
"epoch": 9.495694956949569,
"grad_norm": 0.693181037902832,
"learning_rate": 1.19511780643915e-06,
"loss": 0.5006,
"step": 5790
},
{
"epoch": 9.512095120951209,
"grad_norm": 0.7130532264709473,
"learning_rate": 1.1169173774871478e-06,
"loss": 0.5072,
"step": 5800
},
{
"epoch": 9.52849528495285,
"grad_norm": 0.6810155510902405,
"learning_rate": 1.0413483444362771e-06,
"loss": 0.5014,
"step": 5810
},
{
"epoch": 9.54489544895449,
"grad_norm": 0.6402135491371155,
"learning_rate": 9.684127182679526e-07,
"loss": 0.4956,
"step": 5820
},
{
"epoch": 9.56129561295613,
"grad_norm": 0.6540482640266418,
"learning_rate": 8.981124398855678e-07,
"loss": 0.5035,
"step": 5830
},
{
"epoch": 9.57769577695777,
"grad_norm": 0.6759128570556641,
"learning_rate": 8.304493800627589e-07,
"loss": 0.4896,
"step": 5840
},
{
"epoch": 9.59409594095941,
"grad_norm": 0.67359459400177,
"learning_rate": 7.654253393936439e-07,
"loss": 0.5071,
"step": 5850
},
{
"epoch": 9.61049610496105,
"grad_norm": 0.740897536277771,
"learning_rate": 7.030420482449395e-07,
"loss": 0.5013,
"step": 5860
},
{
"epoch": 9.626896268962689,
"grad_norm": 0.687887966632843,
"learning_rate": 6.433011667098754e-07,
"loss": 0.5017,
"step": 5870
},
{
"epoch": 9.64329643296433,
"grad_norm": 0.6390769481658936,
"learning_rate": 5.862042845640403e-07,
"loss": 0.5032,
"step": 5880
},
{
"epoch": 9.65969659696597,
"grad_norm": 0.6661211848258972,
"learning_rate": 5.317529212230721e-07,
"loss": 0.5026,
"step": 5890
},
{
"epoch": 9.67609676096761,
"grad_norm": 0.7113086581230164,
"learning_rate": 4.799485257022118e-07,
"loss": 0.4996,
"step": 5900
},
{
"epoch": 9.69249692496925,
"grad_norm": 0.6506823301315308,
"learning_rate": 4.307924765777682e-07,
"loss": 0.5032,
"step": 5910
},
{
"epoch": 9.70889708897089,
"grad_norm": 0.6711443066596985,
"learning_rate": 3.842860819504024e-07,
"loss": 0.502,
"step": 5920
},
{
"epoch": 9.72529725297253,
"grad_norm": 0.7337656021118164,
"learning_rate": 3.404305794103224e-07,
"loss": 0.5138,
"step": 5930
},
{
"epoch": 9.74169741697417,
"grad_norm": 0.6719791889190674,
"learning_rate": 2.9922713600439854e-07,
"loss": 0.5015,
"step": 5940
},
{
"epoch": 9.758097580975809,
"grad_norm": 0.6952204704284668,
"learning_rate": 2.606768482050215e-07,
"loss": 0.5161,
"step": 5950
},
{
"epoch": 9.774497744977449,
"grad_norm": 0.7182480096817017,
"learning_rate": 2.2478074188099219e-07,
"loss": 0.5056,
"step": 5960
},
{
"epoch": 9.79089790897909,
"grad_norm": 0.7192881107330322,
"learning_rate": 1.915397722702217e-07,
"loss": 0.4972,
"step": 5970
},
{
"epoch": 9.80729807298073,
"grad_norm": 0.6714368462562561,
"learning_rate": 1.609548239542402e-07,
"loss": 0.4958,
"step": 5980
},
{
"epoch": 9.82369823698237,
"grad_norm": 0.6657147407531738,
"learning_rate": 1.3302671083474938e-07,
"loss": 0.4943,
"step": 5990
},
{
"epoch": 9.84009840098401,
"grad_norm": 0.6431368589401245,
"learning_rate": 1.0775617611189503e-07,
"loss": 0.5079,
"step": 6000
},
{
"epoch": 9.85649856498565,
"grad_norm": 0.7010581493377686,
"learning_rate": 8.514389226452757e-08,
"loss": 0.52,
"step": 6010
},
{
"epoch": 9.87289872898729,
"grad_norm": 0.6842231750488281,
"learning_rate": 6.519046103230508e-08,
"loss": 0.5011,
"step": 6020
},
{
"epoch": 9.88929889298893,
"grad_norm": 0.6943760514259338,
"learning_rate": 4.789641339963957e-08,
"loss": 0.5041,
"step": 6030
},
{
"epoch": 9.90569905699057,
"grad_norm": 0.6967430710792542,
"learning_rate": 3.3262209581619297e-08,
"loss": 0.5096,
"step": 6040
},
{
"epoch": 9.92209922099221,
"grad_norm": 0.6971690058708191,
"learning_rate": 2.1288239011729717e-08,
"loss": 0.5076,
"step": 6050
},
{
"epoch": 9.93849938499385,
"grad_norm": 0.668303906917572,
"learning_rate": 1.1974820331517312e-08,
"loss": 0.497,
"step": 6060
},
{
"epoch": 9.95489954899549,
"grad_norm": 0.6784859895706177,
"learning_rate": 5.3222013820741765e-09,
"loss": 0.5058,
"step": 6070
},
{
"epoch": 9.97129971299713,
"grad_norm": 0.6811095476150513,
"learning_rate": 1.3305591974543953e-09,
"loss": 0.501,
"step": 6080
},
{
"epoch": 9.98769987699877,
"grad_norm": 0.6882724165916443,
"learning_rate": 0.0,
"loss": 0.5079,
"step": 6090
},
{
"epoch": 9.98769987699877,
"step": 6090,
"total_flos": 4.6640782508241715e+17,
"train_loss": 0.5962841084046513,
"train_runtime": 32440.4443,
"train_samples_per_second": 2.255,
"train_steps_per_second": 0.188
}
],
"logging_steps": 10,
"max_steps": 6090,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 500,
"total_flos": 4.6640782508241715e+17,
"train_batch_size": 3,
"trial_name": null,
"trial_params": null
}