MeedoSam's picture
Uploaded checkpoint-10000
8b84355 verified
raw
history blame
No virus
161 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.5861574323112784,
"eval_steps": 1000,
"global_step": 10000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0,
"grad_norm": 10.134367942810059,
"learning_rate": 9e-08,
"loss": 1.2511,
"step": 10
},
{
"epoch": 0.01,
"grad_norm": 18.544782638549805,
"learning_rate": 1.8e-07,
"loss": 1.4001,
"step": 20
},
{
"epoch": 0.01,
"grad_norm": 6.087899684906006,
"learning_rate": 2.8e-07,
"loss": 1.1501,
"step": 30
},
{
"epoch": 0.01,
"grad_norm": 6.356815814971924,
"learning_rate": 3.7999999999999996e-07,
"loss": 1.2225,
"step": 40
},
{
"epoch": 0.02,
"grad_norm": 3.9461255073547363,
"learning_rate": 4.8e-07,
"loss": 1.2241,
"step": 50
},
{
"epoch": 0.02,
"grad_norm": 4.9660773277282715,
"learning_rate": 5.8e-07,
"loss": 1.2118,
"step": 60
},
{
"epoch": 0.03,
"grad_norm": 6.57827615737915,
"learning_rate": 6.800000000000001e-07,
"loss": 1.0416,
"step": 70
},
{
"epoch": 0.03,
"grad_norm": 5.070159435272217,
"learning_rate": 7.799999999999999e-07,
"loss": 1.1224,
"step": 80
},
{
"epoch": 0.03,
"grad_norm": 5.661293029785156,
"learning_rate": 8.799999999999999e-07,
"loss": 1.0603,
"step": 90
},
{
"epoch": 0.04,
"grad_norm": 4.765249252319336,
"learning_rate": 9.8e-07,
"loss": 1.0449,
"step": 100
},
{
"epoch": 0.04,
"grad_norm": 4.521543502807617,
"learning_rate": 9.991919191919192e-07,
"loss": 0.9193,
"step": 110
},
{
"epoch": 0.04,
"grad_norm": 4.420640468597412,
"learning_rate": 9.98181818181818e-07,
"loss": 0.876,
"step": 120
},
{
"epoch": 0.05,
"grad_norm": 4.311656951904297,
"learning_rate": 9.97171717171717e-07,
"loss": 0.8012,
"step": 130
},
{
"epoch": 0.05,
"grad_norm": 5.07025146484375,
"learning_rate": 9.961616161616162e-07,
"loss": 0.826,
"step": 140
},
{
"epoch": 0.05,
"grad_norm": 4.859681129455566,
"learning_rate": 9.951515151515151e-07,
"loss": 0.8442,
"step": 150
},
{
"epoch": 0.06,
"grad_norm": 4.192190170288086,
"learning_rate": 9.94141414141414e-07,
"loss": 0.708,
"step": 160
},
{
"epoch": 0.06,
"grad_norm": 5.852901935577393,
"learning_rate": 9.93131313131313e-07,
"loss": 0.8201,
"step": 170
},
{
"epoch": 0.06,
"grad_norm": 5.076972484588623,
"learning_rate": 9.92121212121212e-07,
"loss": 0.7973,
"step": 180
},
{
"epoch": 0.07,
"grad_norm": 4.36331844329834,
"learning_rate": 9.911111111111111e-07,
"loss": 0.753,
"step": 190
},
{
"epoch": 0.07,
"grad_norm": 5.03492546081543,
"learning_rate": 9.9010101010101e-07,
"loss": 0.7909,
"step": 200
},
{
"epoch": 0.08,
"grad_norm": 4.043544769287109,
"learning_rate": 9.89090909090909e-07,
"loss": 0.7118,
"step": 210
},
{
"epoch": 0.08,
"grad_norm": 2.2020692825317383,
"learning_rate": 9.880808080808082e-07,
"loss": 0.6516,
"step": 220
},
{
"epoch": 0.08,
"grad_norm": 4.851771354675293,
"learning_rate": 9.870707070707071e-07,
"loss": 0.7275,
"step": 230
},
{
"epoch": 0.09,
"grad_norm": 3.2986488342285156,
"learning_rate": 9.86060606060606e-07,
"loss": 0.6484,
"step": 240
},
{
"epoch": 0.09,
"grad_norm": 5.700368404388428,
"learning_rate": 9.85050505050505e-07,
"loss": 0.7258,
"step": 250
},
{
"epoch": 0.09,
"grad_norm": 6.46458625793457,
"learning_rate": 9.84040404040404e-07,
"loss": 0.6416,
"step": 260
},
{
"epoch": 0.1,
"grad_norm": 4.012514114379883,
"learning_rate": 9.830303030303029e-07,
"loss": 0.5861,
"step": 270
},
{
"epoch": 0.1,
"grad_norm": 3.328000068664551,
"learning_rate": 9.82020202020202e-07,
"loss": 0.6239,
"step": 280
},
{
"epoch": 0.1,
"grad_norm": 3.837636947631836,
"learning_rate": 9.81010101010101e-07,
"loss": 0.6718,
"step": 290
},
{
"epoch": 0.11,
"grad_norm": 4.840264797210693,
"learning_rate": 9.8e-07,
"loss": 0.6623,
"step": 300
},
{
"epoch": 0.11,
"grad_norm": 6.741188049316406,
"learning_rate": 9.789898989898989e-07,
"loss": 0.6066,
"step": 310
},
{
"epoch": 0.11,
"grad_norm": 5.170821666717529,
"learning_rate": 9.779797979797978e-07,
"loss": 0.6634,
"step": 320
},
{
"epoch": 0.12,
"grad_norm": 8.481480598449707,
"learning_rate": 9.76969696969697e-07,
"loss": 0.6474,
"step": 330
},
{
"epoch": 0.12,
"grad_norm": 3.698042631149292,
"learning_rate": 9.75959595959596e-07,
"loss": 0.6594,
"step": 340
},
{
"epoch": 0.13,
"grad_norm": 4.990598201751709,
"learning_rate": 9.749494949494949e-07,
"loss": 0.6588,
"step": 350
},
{
"epoch": 0.13,
"grad_norm": 3.2625105381011963,
"learning_rate": 9.73939393939394e-07,
"loss": 0.6532,
"step": 360
},
{
"epoch": 0.13,
"grad_norm": 6.604375839233398,
"learning_rate": 9.72929292929293e-07,
"loss": 0.7057,
"step": 370
},
{
"epoch": 0.14,
"grad_norm": 5.352667808532715,
"learning_rate": 9.71919191919192e-07,
"loss": 0.6273,
"step": 380
},
{
"epoch": 0.14,
"grad_norm": 3.8649349212646484,
"learning_rate": 9.709090909090909e-07,
"loss": 0.6366,
"step": 390
},
{
"epoch": 0.14,
"grad_norm": 5.2739667892456055,
"learning_rate": 9.698989898989898e-07,
"loss": 0.6433,
"step": 400
},
{
"epoch": 0.15,
"grad_norm": 4.892832279205322,
"learning_rate": 9.68888888888889e-07,
"loss": 0.6504,
"step": 410
},
{
"epoch": 0.15,
"grad_norm": 5.36677360534668,
"learning_rate": 9.67878787878788e-07,
"loss": 0.6661,
"step": 420
},
{
"epoch": 0.15,
"grad_norm": 2.6905229091644287,
"learning_rate": 9.668686868686868e-07,
"loss": 0.6334,
"step": 430
},
{
"epoch": 0.16,
"grad_norm": 4.99962854385376,
"learning_rate": 9.658585858585858e-07,
"loss": 0.6028,
"step": 440
},
{
"epoch": 0.16,
"grad_norm": 5.522456169128418,
"learning_rate": 9.648484848484847e-07,
"loss": 0.6903,
"step": 450
},
{
"epoch": 0.16,
"grad_norm": 7.456121921539307,
"learning_rate": 9.638383838383839e-07,
"loss": 0.6486,
"step": 460
},
{
"epoch": 0.17,
"grad_norm": 5.989662170410156,
"learning_rate": 9.628282828282828e-07,
"loss": 0.7085,
"step": 470
},
{
"epoch": 0.17,
"grad_norm": 3.5824246406555176,
"learning_rate": 9.618181818181818e-07,
"loss": 0.6652,
"step": 480
},
{
"epoch": 0.18,
"grad_norm": 3.074733018875122,
"learning_rate": 9.608080808080807e-07,
"loss": 0.5733,
"step": 490
},
{
"epoch": 0.18,
"grad_norm": 3.964071750640869,
"learning_rate": 9.597979797979797e-07,
"loss": 0.6319,
"step": 500
},
{
"epoch": 0.18,
"grad_norm": 3.8981716632843018,
"learning_rate": 9.587878787878786e-07,
"loss": 0.5937,
"step": 510
},
{
"epoch": 0.19,
"grad_norm": 7.471535682678223,
"learning_rate": 9.577777777777778e-07,
"loss": 0.6564,
"step": 520
},
{
"epoch": 0.19,
"grad_norm": 5.468303680419922,
"learning_rate": 9.567676767676767e-07,
"loss": 0.6339,
"step": 530
},
{
"epoch": 0.19,
"grad_norm": 3.9055886268615723,
"learning_rate": 9.557575757575759e-07,
"loss": 0.7103,
"step": 540
},
{
"epoch": 0.2,
"grad_norm": 4.115898132324219,
"learning_rate": 9.547474747474748e-07,
"loss": 0.6928,
"step": 550
},
{
"epoch": 0.2,
"grad_norm": 4.248528003692627,
"learning_rate": 9.537373737373737e-07,
"loss": 0.6084,
"step": 560
},
{
"epoch": 0.2,
"grad_norm": 3.727003335952759,
"learning_rate": 9.527272727272727e-07,
"loss": 0.6554,
"step": 570
},
{
"epoch": 0.21,
"grad_norm": 4.747541904449463,
"learning_rate": 9.517171717171717e-07,
"loss": 0.5876,
"step": 580
},
{
"epoch": 0.21,
"grad_norm": 3.913226842880249,
"learning_rate": 9.507070707070707e-07,
"loss": 0.5888,
"step": 590
},
{
"epoch": 0.22,
"grad_norm": 4.400748252868652,
"learning_rate": 9.496969696969696e-07,
"loss": 0.6196,
"step": 600
},
{
"epoch": 0.22,
"grad_norm": 3.053006649017334,
"learning_rate": 9.486868686868687e-07,
"loss": 0.6356,
"step": 610
},
{
"epoch": 0.22,
"grad_norm": 4.207491874694824,
"learning_rate": 9.476767676767676e-07,
"loss": 0.6161,
"step": 620
},
{
"epoch": 0.23,
"grad_norm": 6.389359951019287,
"learning_rate": 9.466666666666666e-07,
"loss": 0.6172,
"step": 630
},
{
"epoch": 0.23,
"grad_norm": 4.967292308807373,
"learning_rate": 9.456565656565656e-07,
"loss": 0.6173,
"step": 640
},
{
"epoch": 0.23,
"grad_norm": 3.855407953262329,
"learning_rate": 9.446464646464646e-07,
"loss": 0.585,
"step": 650
},
{
"epoch": 0.24,
"grad_norm": 3.580738067626953,
"learning_rate": 9.436363636363636e-07,
"loss": 0.6166,
"step": 660
},
{
"epoch": 0.24,
"grad_norm": 3.6277294158935547,
"learning_rate": 9.426262626262626e-07,
"loss": 0.5951,
"step": 670
},
{
"epoch": 0.24,
"grad_norm": 3.7501285076141357,
"learning_rate": 9.416161616161615e-07,
"loss": 0.5807,
"step": 680
},
{
"epoch": 0.25,
"grad_norm": 4.80528450012207,
"learning_rate": 9.406060606060605e-07,
"loss": 0.6053,
"step": 690
},
{
"epoch": 0.25,
"grad_norm": 5.07316780090332,
"learning_rate": 9.395959595959596e-07,
"loss": 0.626,
"step": 700
},
{
"epoch": 0.25,
"grad_norm": 5.338922023773193,
"learning_rate": 9.385858585858585e-07,
"loss": 0.5356,
"step": 710
},
{
"epoch": 0.26,
"grad_norm": 3.264220714569092,
"learning_rate": 9.375757575757576e-07,
"loss": 0.5929,
"step": 720
},
{
"epoch": 0.26,
"grad_norm": 3.4565188884735107,
"learning_rate": 9.365656565656565e-07,
"loss": 0.6379,
"step": 730
},
{
"epoch": 0.27,
"grad_norm": 4.202028751373291,
"learning_rate": 9.355555555555556e-07,
"loss": 0.6143,
"step": 740
},
{
"epoch": 0.27,
"grad_norm": 5.128079891204834,
"learning_rate": 9.345454545454545e-07,
"loss": 0.5531,
"step": 750
},
{
"epoch": 0.27,
"grad_norm": 4.205199718475342,
"learning_rate": 9.335353535353535e-07,
"loss": 0.5842,
"step": 760
},
{
"epoch": 0.28,
"grad_norm": 5.925571918487549,
"learning_rate": 9.325252525252525e-07,
"loss": 0.6169,
"step": 770
},
{
"epoch": 0.28,
"grad_norm": 4.256628036499023,
"learning_rate": 9.315151515151515e-07,
"loss": 0.5899,
"step": 780
},
{
"epoch": 0.28,
"grad_norm": 3.2796261310577393,
"learning_rate": 9.305050505050504e-07,
"loss": 0.6654,
"step": 790
},
{
"epoch": 0.29,
"grad_norm": 3.6715095043182373,
"learning_rate": 9.294949494949495e-07,
"loss": 0.6272,
"step": 800
},
{
"epoch": 0.29,
"grad_norm": 5.043895721435547,
"learning_rate": 9.284848484848484e-07,
"loss": 0.6019,
"step": 810
},
{
"epoch": 0.29,
"grad_norm": 4.5200300216674805,
"learning_rate": 9.274747474747475e-07,
"loss": 0.6992,
"step": 820
},
{
"epoch": 0.3,
"grad_norm": 5.557192325592041,
"learning_rate": 9.264646464646464e-07,
"loss": 0.6657,
"step": 830
},
{
"epoch": 0.3,
"grad_norm": 3.2051031589508057,
"learning_rate": 9.254545454545453e-07,
"loss": 0.6262,
"step": 840
},
{
"epoch": 0.3,
"grad_norm": 3.465202808380127,
"learning_rate": 9.244444444444444e-07,
"loss": 0.6603,
"step": 850
},
{
"epoch": 0.31,
"grad_norm": 5.039762496948242,
"learning_rate": 9.234343434343433e-07,
"loss": 0.6276,
"step": 860
},
{
"epoch": 0.31,
"grad_norm": 5.317465305328369,
"learning_rate": 9.224242424242423e-07,
"loss": 0.521,
"step": 870
},
{
"epoch": 0.32,
"grad_norm": 6.985171318054199,
"learning_rate": 9.214141414141414e-07,
"loss": 0.6768,
"step": 880
},
{
"epoch": 0.32,
"grad_norm": 3.9759483337402344,
"learning_rate": 9.204040404040404e-07,
"loss": 0.5982,
"step": 890
},
{
"epoch": 0.32,
"grad_norm": 4.207100868225098,
"learning_rate": 9.193939393939394e-07,
"loss": 0.6117,
"step": 900
},
{
"epoch": 0.33,
"grad_norm": 4.389718055725098,
"learning_rate": 9.183838383838384e-07,
"loss": 0.5767,
"step": 910
},
{
"epoch": 0.33,
"grad_norm": 3.659482717514038,
"learning_rate": 9.173737373737373e-07,
"loss": 0.5957,
"step": 920
},
{
"epoch": 0.33,
"grad_norm": 6.660622596740723,
"learning_rate": 9.163636363636364e-07,
"loss": 0.6403,
"step": 930
},
{
"epoch": 0.34,
"grad_norm": 6.03493070602417,
"learning_rate": 9.153535353535353e-07,
"loss": 0.5857,
"step": 940
},
{
"epoch": 0.34,
"grad_norm": 2.699523687362671,
"learning_rate": 9.143434343434343e-07,
"loss": 0.5913,
"step": 950
},
{
"epoch": 0.34,
"grad_norm": 3.8991754055023193,
"learning_rate": 9.133333333333333e-07,
"loss": 0.6342,
"step": 960
},
{
"epoch": 0.35,
"grad_norm": 4.422178268432617,
"learning_rate": 9.123232323232323e-07,
"loss": 0.6286,
"step": 970
},
{
"epoch": 0.35,
"grad_norm": 4.381803512573242,
"learning_rate": 9.113131313131313e-07,
"loss": 0.6899,
"step": 980
},
{
"epoch": 0.36,
"grad_norm": 4.75442361831665,
"learning_rate": 9.103030303030302e-07,
"loss": 0.5462,
"step": 990
},
{
"epoch": 0.36,
"grad_norm": 3.6305062770843506,
"learning_rate": 9.092929292929292e-07,
"loss": 0.666,
"step": 1000
},
{
"epoch": 0.36,
"eval_loss": 0.772658109664917,
"eval_runtime": 400.6004,
"eval_samples_per_second": 2.496,
"eval_steps_per_second": 2.496,
"step": 1000
},
{
"epoch": 0.36,
"grad_norm": 4.022682189941406,
"learning_rate": 9.082828282828282e-07,
"loss": 0.5796,
"step": 1010
},
{
"epoch": 0.37,
"grad_norm": 5.148658752441406,
"learning_rate": 9.072727272727272e-07,
"loss": 0.5738,
"step": 1020
},
{
"epoch": 0.37,
"grad_norm": 3.78167462348938,
"learning_rate": 9.062626262626261e-07,
"loss": 0.6572,
"step": 1030
},
{
"epoch": 0.37,
"grad_norm": 4.6793212890625,
"learning_rate": 9.052525252525252e-07,
"loss": 0.6466,
"step": 1040
},
{
"epoch": 0.38,
"grad_norm": 4.607004165649414,
"learning_rate": 9.042424242424242e-07,
"loss": 0.5599,
"step": 1050
},
{
"epoch": 0.38,
"grad_norm": 5.730196952819824,
"learning_rate": 9.032323232323233e-07,
"loss": 0.6559,
"step": 1060
},
{
"epoch": 0.38,
"grad_norm": 4.408863544464111,
"learning_rate": 9.022222222222222e-07,
"loss": 0.641,
"step": 1070
},
{
"epoch": 0.39,
"grad_norm": 3.0303945541381836,
"learning_rate": 9.012121212121212e-07,
"loss": 0.5455,
"step": 1080
},
{
"epoch": 0.39,
"grad_norm": 5.44600248336792,
"learning_rate": 9.002020202020202e-07,
"loss": 0.5672,
"step": 1090
},
{
"epoch": 0.39,
"grad_norm": 4.877390384674072,
"learning_rate": 8.991919191919192e-07,
"loss": 0.5626,
"step": 1100
},
{
"epoch": 0.4,
"grad_norm": 4.949882507324219,
"learning_rate": 8.981818181818181e-07,
"loss": 0.6092,
"step": 1110
},
{
"epoch": 0.4,
"grad_norm": 4.810749530792236,
"learning_rate": 8.971717171717172e-07,
"loss": 0.6486,
"step": 1120
},
{
"epoch": 0.41,
"grad_norm": 4.6897053718566895,
"learning_rate": 8.961616161616161e-07,
"loss": 0.5854,
"step": 1130
},
{
"epoch": 0.41,
"grad_norm": 3.393486261367798,
"learning_rate": 8.951515151515151e-07,
"loss": 0.6071,
"step": 1140
},
{
"epoch": 0.41,
"grad_norm": 3.1970651149749756,
"learning_rate": 8.941414141414141e-07,
"loss": 0.601,
"step": 1150
},
{
"epoch": 0.42,
"grad_norm": 4.972078800201416,
"learning_rate": 8.93131313131313e-07,
"loss": 0.6291,
"step": 1160
},
{
"epoch": 0.42,
"grad_norm": 5.2576003074646,
"learning_rate": 8.921212121212121e-07,
"loss": 0.5856,
"step": 1170
},
{
"epoch": 0.42,
"grad_norm": 3.552098035812378,
"learning_rate": 8.91111111111111e-07,
"loss": 0.5712,
"step": 1180
},
{
"epoch": 0.43,
"grad_norm": 4.2522382736206055,
"learning_rate": 8.9010101010101e-07,
"loss": 0.6128,
"step": 1190
},
{
"epoch": 0.43,
"grad_norm": 3.3408310413360596,
"learning_rate": 8.89090909090909e-07,
"loss": 0.6636,
"step": 1200
},
{
"epoch": 0.43,
"grad_norm": 3.3888745307922363,
"learning_rate": 8.88080808080808e-07,
"loss": 0.5906,
"step": 1210
},
{
"epoch": 0.44,
"grad_norm": 4.374114990234375,
"learning_rate": 8.870707070707071e-07,
"loss": 0.5863,
"step": 1220
},
{
"epoch": 0.44,
"grad_norm": 3.318901538848877,
"learning_rate": 8.860606060606061e-07,
"loss": 0.6232,
"step": 1230
},
{
"epoch": 0.44,
"grad_norm": 3.4461617469787598,
"learning_rate": 8.85050505050505e-07,
"loss": 0.6712,
"step": 1240
},
{
"epoch": 0.45,
"grad_norm": 3.6037635803222656,
"learning_rate": 8.840404040404041e-07,
"loss": 0.5787,
"step": 1250
},
{
"epoch": 0.45,
"grad_norm": 4.3236002922058105,
"learning_rate": 8.83030303030303e-07,
"loss": 0.6723,
"step": 1260
},
{
"epoch": 0.46,
"grad_norm": 3.199084997177124,
"learning_rate": 8.820202020202019e-07,
"loss": 0.5484,
"step": 1270
},
{
"epoch": 0.46,
"grad_norm": 3.675905704498291,
"learning_rate": 8.81010101010101e-07,
"loss": 0.6713,
"step": 1280
},
{
"epoch": 0.46,
"grad_norm": 4.667660713195801,
"learning_rate": 8.799999999999999e-07,
"loss": 0.6161,
"step": 1290
},
{
"epoch": 0.47,
"grad_norm": 5.268735408782959,
"learning_rate": 8.78989898989899e-07,
"loss": 0.6282,
"step": 1300
},
{
"epoch": 0.47,
"grad_norm": 3.6777164936065674,
"learning_rate": 8.779797979797979e-07,
"loss": 0.662,
"step": 1310
},
{
"epoch": 0.47,
"grad_norm": 5.912172317504883,
"learning_rate": 8.769696969696969e-07,
"loss": 0.5951,
"step": 1320
},
{
"epoch": 0.48,
"grad_norm": 4.206521034240723,
"learning_rate": 8.759595959595959e-07,
"loss": 0.704,
"step": 1330
},
{
"epoch": 0.48,
"grad_norm": 4.009490013122559,
"learning_rate": 8.749494949494949e-07,
"loss": 0.652,
"step": 1340
},
{
"epoch": 0.48,
"grad_norm": 3.8058555126190186,
"learning_rate": 8.739393939393938e-07,
"loss": 0.5743,
"step": 1350
},
{
"epoch": 0.49,
"grad_norm": 8.637187957763672,
"learning_rate": 8.729292929292929e-07,
"loss": 0.6198,
"step": 1360
},
{
"epoch": 0.49,
"grad_norm": 5.059712886810303,
"learning_rate": 8.719191919191918e-07,
"loss": 0.6171,
"step": 1370
},
{
"epoch": 0.49,
"grad_norm": 5.403115272521973,
"learning_rate": 8.709090909090909e-07,
"loss": 0.5956,
"step": 1380
},
{
"epoch": 0.5,
"grad_norm": 4.084959983825684,
"learning_rate": 8.698989898989899e-07,
"loss": 0.6023,
"step": 1390
},
{
"epoch": 0.5,
"grad_norm": 4.863948345184326,
"learning_rate": 8.688888888888889e-07,
"loss": 0.6295,
"step": 1400
},
{
"epoch": 0.51,
"grad_norm": 5.363975524902344,
"learning_rate": 8.678787878787879e-07,
"loss": 0.5854,
"step": 1410
},
{
"epoch": 0.51,
"grad_norm": 4.450655460357666,
"learning_rate": 8.668686868686868e-07,
"loss": 0.6579,
"step": 1420
},
{
"epoch": 0.51,
"grad_norm": 5.538711071014404,
"learning_rate": 8.658585858585859e-07,
"loss": 0.726,
"step": 1430
},
{
"epoch": 0.52,
"grad_norm": 3.3089439868927,
"learning_rate": 8.648484848484848e-07,
"loss": 0.6435,
"step": 1440
},
{
"epoch": 0.52,
"grad_norm": 3.613719940185547,
"learning_rate": 8.638383838383838e-07,
"loss": 0.6695,
"step": 1450
},
{
"epoch": 0.52,
"grad_norm": 2.9638564586639404,
"learning_rate": 8.628282828282828e-07,
"loss": 0.5516,
"step": 1460
},
{
"epoch": 0.53,
"grad_norm": 3.8651678562164307,
"learning_rate": 8.618181818181818e-07,
"loss": 0.5883,
"step": 1470
},
{
"epoch": 0.53,
"grad_norm": 8.56436824798584,
"learning_rate": 8.608080808080807e-07,
"loss": 0.6476,
"step": 1480
},
{
"epoch": 0.53,
"grad_norm": 6.782663345336914,
"learning_rate": 8.597979797979798e-07,
"loss": 0.6369,
"step": 1490
},
{
"epoch": 0.54,
"grad_norm": 3.181833267211914,
"learning_rate": 8.587878787878787e-07,
"loss": 0.5885,
"step": 1500
},
{
"epoch": 0.54,
"grad_norm": 5.068531513214111,
"learning_rate": 8.577777777777777e-07,
"loss": 0.6241,
"step": 1510
},
{
"epoch": 0.55,
"grad_norm": 5.7847900390625,
"learning_rate": 8.567676767676767e-07,
"loss": 0.623,
"step": 1520
},
{
"epoch": 0.55,
"grad_norm": 3.325660228729248,
"learning_rate": 8.557575757575757e-07,
"loss": 0.574,
"step": 1530
},
{
"epoch": 0.55,
"grad_norm": 3.452211618423462,
"learning_rate": 8.547474747474747e-07,
"loss": 0.6359,
"step": 1540
},
{
"epoch": 0.56,
"grad_norm": 5.663365364074707,
"learning_rate": 8.537373737373736e-07,
"loss": 0.6132,
"step": 1550
},
{
"epoch": 0.56,
"grad_norm": 3.970423460006714,
"learning_rate": 8.527272727272727e-07,
"loss": 0.5644,
"step": 1560
},
{
"epoch": 0.56,
"grad_norm": 4.623054027557373,
"learning_rate": 8.517171717171717e-07,
"loss": 0.5753,
"step": 1570
},
{
"epoch": 0.57,
"grad_norm": 4.318519592285156,
"learning_rate": 8.507070707070707e-07,
"loss": 0.6335,
"step": 1580
},
{
"epoch": 0.57,
"grad_norm": 3.2903385162353516,
"learning_rate": 8.496969696969697e-07,
"loss": 0.5899,
"step": 1590
},
{
"epoch": 0.57,
"grad_norm": 2.847003221511841,
"learning_rate": 8.486868686868687e-07,
"loss": 0.5834,
"step": 1600
},
{
"epoch": 0.58,
"grad_norm": 5.138712406158447,
"learning_rate": 8.476767676767676e-07,
"loss": 0.5485,
"step": 1610
},
{
"epoch": 0.58,
"grad_norm": 5.145106792449951,
"learning_rate": 8.466666666666667e-07,
"loss": 0.6854,
"step": 1620
},
{
"epoch": 0.58,
"grad_norm": 5.585366725921631,
"learning_rate": 8.456565656565656e-07,
"loss": 0.5595,
"step": 1630
},
{
"epoch": 0.59,
"grad_norm": 3.857548952102661,
"learning_rate": 8.446464646464646e-07,
"loss": 0.5856,
"step": 1640
},
{
"epoch": 0.59,
"grad_norm": 3.835942029953003,
"learning_rate": 8.436363636363636e-07,
"loss": 0.5589,
"step": 1650
},
{
"epoch": 0.6,
"grad_norm": 4.407024383544922,
"learning_rate": 8.426262626262626e-07,
"loss": 0.5358,
"step": 1660
},
{
"epoch": 0.6,
"grad_norm": 4.5201826095581055,
"learning_rate": 8.416161616161616e-07,
"loss": 0.6021,
"step": 1670
},
{
"epoch": 0.6,
"grad_norm": 4.167496681213379,
"learning_rate": 8.406060606060606e-07,
"loss": 0.6007,
"step": 1680
},
{
"epoch": 0.61,
"grad_norm": 5.0080695152282715,
"learning_rate": 8.395959595959595e-07,
"loss": 0.6298,
"step": 1690
},
{
"epoch": 0.61,
"grad_norm": 4.7244977951049805,
"learning_rate": 8.385858585858585e-07,
"loss": 0.5713,
"step": 1700
},
{
"epoch": 0.61,
"grad_norm": 5.417051315307617,
"learning_rate": 8.375757575757575e-07,
"loss": 0.6565,
"step": 1710
},
{
"epoch": 0.62,
"grad_norm": 6.2620391845703125,
"learning_rate": 8.365656565656564e-07,
"loss": 0.6977,
"step": 1720
},
{
"epoch": 0.62,
"grad_norm": 5.629709720611572,
"learning_rate": 8.355555555555556e-07,
"loss": 0.6281,
"step": 1730
},
{
"epoch": 0.62,
"grad_norm": 4.182885646820068,
"learning_rate": 8.345454545454545e-07,
"loss": 0.6431,
"step": 1740
},
{
"epoch": 0.63,
"grad_norm": 3.080979824066162,
"learning_rate": 8.335353535353536e-07,
"loss": 0.5938,
"step": 1750
},
{
"epoch": 0.63,
"grad_norm": 2.8959693908691406,
"learning_rate": 8.325252525252525e-07,
"loss": 0.6677,
"step": 1760
},
{
"epoch": 0.63,
"grad_norm": 2.9930922985076904,
"learning_rate": 8.315151515151515e-07,
"loss": 0.5425,
"step": 1770
},
{
"epoch": 0.64,
"grad_norm": 11.59288501739502,
"learning_rate": 8.305050505050505e-07,
"loss": 0.5816,
"step": 1780
},
{
"epoch": 0.64,
"grad_norm": 6.010501384735107,
"learning_rate": 8.294949494949495e-07,
"loss": 0.5944,
"step": 1790
},
{
"epoch": 0.65,
"grad_norm": 3.7763831615448,
"learning_rate": 8.284848484848484e-07,
"loss": 0.6465,
"step": 1800
},
{
"epoch": 0.65,
"grad_norm": 4.849707126617432,
"learning_rate": 8.274747474747475e-07,
"loss": 0.5679,
"step": 1810
},
{
"epoch": 0.65,
"grad_norm": 3.6395959854125977,
"learning_rate": 8.264646464646464e-07,
"loss": 0.6026,
"step": 1820
},
{
"epoch": 0.66,
"grad_norm": 4.717153549194336,
"learning_rate": 8.254545454545455e-07,
"loss": 0.5637,
"step": 1830
},
{
"epoch": 0.66,
"grad_norm": 3.236311674118042,
"learning_rate": 8.244444444444444e-07,
"loss": 0.5955,
"step": 1840
},
{
"epoch": 0.66,
"grad_norm": 4.533290386199951,
"learning_rate": 8.234343434343433e-07,
"loss": 0.5681,
"step": 1850
},
{
"epoch": 0.67,
"grad_norm": 3.0948798656463623,
"learning_rate": 8.224242424242424e-07,
"loss": 0.6272,
"step": 1860
},
{
"epoch": 0.67,
"grad_norm": 3.719708204269409,
"learning_rate": 8.214141414141413e-07,
"loss": 0.6793,
"step": 1870
},
{
"epoch": 0.67,
"grad_norm": 3.228376865386963,
"learning_rate": 8.204040404040403e-07,
"loss": 0.558,
"step": 1880
},
{
"epoch": 0.68,
"grad_norm": 3.4296934604644775,
"learning_rate": 8.193939393939393e-07,
"loss": 0.5717,
"step": 1890
},
{
"epoch": 0.68,
"grad_norm": 5.708250999450684,
"learning_rate": 8.184848484848484e-07,
"loss": 0.6381,
"step": 1900
},
{
"epoch": 0.68,
"grad_norm": 3.8306288719177246,
"learning_rate": 8.174747474747474e-07,
"loss": 0.6083,
"step": 1910
},
{
"epoch": 0.69,
"grad_norm": 4.5576252937316895,
"learning_rate": 8.164646464646464e-07,
"loss": 0.6676,
"step": 1920
},
{
"epoch": 0.69,
"grad_norm": 3.2234649658203125,
"learning_rate": 8.154545454545454e-07,
"loss": 0.6215,
"step": 1930
},
{
"epoch": 0.7,
"grad_norm": 3.4826815128326416,
"learning_rate": 8.144444444444444e-07,
"loss": 0.6827,
"step": 1940
},
{
"epoch": 0.7,
"grad_norm": 3.9034299850463867,
"learning_rate": 8.134343434343433e-07,
"loss": 0.596,
"step": 1950
},
{
"epoch": 0.7,
"grad_norm": 2.7787563800811768,
"learning_rate": 8.124242424242424e-07,
"loss": 0.6234,
"step": 1960
},
{
"epoch": 0.71,
"grad_norm": 4.442075252532959,
"learning_rate": 8.114141414141413e-07,
"loss": 0.6244,
"step": 1970
},
{
"epoch": 0.71,
"grad_norm": 3.8511252403259277,
"learning_rate": 8.104040404040403e-07,
"loss": 0.5721,
"step": 1980
},
{
"epoch": 0.71,
"grad_norm": 3.925318717956543,
"learning_rate": 8.093939393939394e-07,
"loss": 0.6547,
"step": 1990
},
{
"epoch": 0.72,
"grad_norm": 4.262260437011719,
"learning_rate": 8.083838383838384e-07,
"loss": 0.6155,
"step": 2000
},
{
"epoch": 0.72,
"eval_loss": 0.7234830260276794,
"eval_runtime": 401.6534,
"eval_samples_per_second": 2.49,
"eval_steps_per_second": 2.49,
"step": 2000
},
{
"epoch": 0.72,
"grad_norm": 4.890273571014404,
"learning_rate": 8.073737373737374e-07,
"loss": 0.6392,
"step": 2010
},
{
"epoch": 0.72,
"grad_norm": 4.918805122375488,
"learning_rate": 8.063636363636364e-07,
"loss": 0.5549,
"step": 2020
},
{
"epoch": 0.73,
"grad_norm": 3.605008840560913,
"learning_rate": 8.053535353535353e-07,
"loss": 0.5152,
"step": 2030
},
{
"epoch": 0.73,
"grad_norm": 4.606666088104248,
"learning_rate": 8.043434343434344e-07,
"loss": 0.4798,
"step": 2040
},
{
"epoch": 0.74,
"grad_norm": 4.115630626678467,
"learning_rate": 8.033333333333333e-07,
"loss": 0.5856,
"step": 2050
},
{
"epoch": 0.74,
"grad_norm": 4.37373685836792,
"learning_rate": 8.023232323232322e-07,
"loss": 0.6467,
"step": 2060
},
{
"epoch": 0.74,
"grad_norm": 3.3623104095458984,
"learning_rate": 8.013131313131313e-07,
"loss": 0.6591,
"step": 2070
},
{
"epoch": 0.75,
"grad_norm": 3.1440579891204834,
"learning_rate": 8.003030303030302e-07,
"loss": 0.6003,
"step": 2080
},
{
"epoch": 0.75,
"grad_norm": 3.391533136367798,
"learning_rate": 7.992929292929293e-07,
"loss": 0.5872,
"step": 2090
},
{
"epoch": 0.75,
"grad_norm": 4.677767276763916,
"learning_rate": 7.982828282828282e-07,
"loss": 0.5803,
"step": 2100
},
{
"epoch": 0.76,
"grad_norm": 3.759061336517334,
"learning_rate": 7.972727272727272e-07,
"loss": 0.6155,
"step": 2110
},
{
"epoch": 0.76,
"grad_norm": 3.419386863708496,
"learning_rate": 7.962626262626262e-07,
"loss": 0.6303,
"step": 2120
},
{
"epoch": 0.76,
"grad_norm": 3.4935250282287598,
"learning_rate": 7.952525252525252e-07,
"loss": 0.6348,
"step": 2130
},
{
"epoch": 0.77,
"grad_norm": 4.793978214263916,
"learning_rate": 7.942424242424241e-07,
"loss": 0.5004,
"step": 2140
},
{
"epoch": 0.77,
"grad_norm": 5.078155994415283,
"learning_rate": 7.932323232323232e-07,
"loss": 0.5583,
"step": 2150
},
{
"epoch": 0.77,
"grad_norm": 3.8855032920837402,
"learning_rate": 7.922222222222222e-07,
"loss": 0.626,
"step": 2160
},
{
"epoch": 0.78,
"grad_norm": 4.782063007354736,
"learning_rate": 7.912121212121213e-07,
"loss": 0.4847,
"step": 2170
},
{
"epoch": 0.78,
"grad_norm": 4.760573387145996,
"learning_rate": 7.902020202020202e-07,
"loss": 0.5923,
"step": 2180
},
{
"epoch": 0.79,
"grad_norm": 5.196963310241699,
"learning_rate": 7.891919191919192e-07,
"loss": 0.5216,
"step": 2190
},
{
"epoch": 0.79,
"grad_norm": 5.447847843170166,
"learning_rate": 7.881818181818182e-07,
"loss": 0.6144,
"step": 2200
},
{
"epoch": 0.79,
"grad_norm": 5.5991129875183105,
"learning_rate": 7.871717171717171e-07,
"loss": 0.5247,
"step": 2210
},
{
"epoch": 0.8,
"grad_norm": 4.1100921630859375,
"learning_rate": 7.861616161616161e-07,
"loss": 0.6581,
"step": 2220
},
{
"epoch": 0.8,
"grad_norm": 2.998598098754883,
"learning_rate": 7.851515151515151e-07,
"loss": 0.5111,
"step": 2230
},
{
"epoch": 0.8,
"grad_norm": 3.1241579055786133,
"learning_rate": 7.841414141414141e-07,
"loss": 0.626,
"step": 2240
},
{
"epoch": 0.81,
"grad_norm": 3.804655075073242,
"learning_rate": 7.831313131313131e-07,
"loss": 0.596,
"step": 2250
},
{
"epoch": 0.81,
"grad_norm": 3.636692523956299,
"learning_rate": 7.821212121212121e-07,
"loss": 0.598,
"step": 2260
},
{
"epoch": 0.81,
"grad_norm": 5.374478816986084,
"learning_rate": 7.81111111111111e-07,
"loss": 0.5155,
"step": 2270
},
{
"epoch": 0.82,
"grad_norm": 4.611242294311523,
"learning_rate": 7.801010101010101e-07,
"loss": 0.5482,
"step": 2280
},
{
"epoch": 0.82,
"grad_norm": 6.077275276184082,
"learning_rate": 7.79090909090909e-07,
"loss": 0.5642,
"step": 2290
},
{
"epoch": 0.82,
"grad_norm": 4.110134601593018,
"learning_rate": 7.78080808080808e-07,
"loss": 0.6482,
"step": 2300
},
{
"epoch": 0.83,
"grad_norm": 3.603376865386963,
"learning_rate": 7.77070707070707e-07,
"loss": 0.5755,
"step": 2310
},
{
"epoch": 0.83,
"grad_norm": 4.059467315673828,
"learning_rate": 7.76060606060606e-07,
"loss": 0.5205,
"step": 2320
},
{
"epoch": 0.84,
"grad_norm": 4.656579971313477,
"learning_rate": 7.750505050505051e-07,
"loss": 0.5975,
"step": 2330
},
{
"epoch": 0.84,
"grad_norm": 4.863138675689697,
"learning_rate": 7.74040404040404e-07,
"loss": 0.5375,
"step": 2340
},
{
"epoch": 0.84,
"grad_norm": 4.3313517570495605,
"learning_rate": 7.73030303030303e-07,
"loss": 0.5795,
"step": 2350
},
{
"epoch": 0.85,
"grad_norm": 3.870875120162964,
"learning_rate": 7.72020202020202e-07,
"loss": 0.5249,
"step": 2360
},
{
"epoch": 0.85,
"grad_norm": 3.720283031463623,
"learning_rate": 7.71010101010101e-07,
"loss": 0.6733,
"step": 2370
},
{
"epoch": 0.85,
"grad_norm": 3.779296398162842,
"learning_rate": 7.699999999999999e-07,
"loss": 0.588,
"step": 2380
},
{
"epoch": 0.86,
"grad_norm": 4.588533401489258,
"learning_rate": 7.68989898989899e-07,
"loss": 0.6044,
"step": 2390
},
{
"epoch": 0.86,
"grad_norm": 5.907507419586182,
"learning_rate": 7.679797979797979e-07,
"loss": 0.5715,
"step": 2400
},
{
"epoch": 0.86,
"grad_norm": 5.705809116363525,
"learning_rate": 7.66969696969697e-07,
"loss": 0.5486,
"step": 2410
},
{
"epoch": 0.87,
"grad_norm": 3.8104190826416016,
"learning_rate": 7.659595959595959e-07,
"loss": 0.602,
"step": 2420
},
{
"epoch": 0.87,
"grad_norm": 5.82910680770874,
"learning_rate": 7.649494949494949e-07,
"loss": 0.5111,
"step": 2430
},
{
"epoch": 0.88,
"grad_norm": 5.837489604949951,
"learning_rate": 7.639393939393939e-07,
"loss": 0.6145,
"step": 2440
},
{
"epoch": 0.88,
"grad_norm": 3.5505547523498535,
"learning_rate": 7.629292929292929e-07,
"loss": 0.5834,
"step": 2450
},
{
"epoch": 0.88,
"grad_norm": 4.314844131469727,
"learning_rate": 7.619191919191918e-07,
"loss": 0.6152,
"step": 2460
},
{
"epoch": 0.89,
"grad_norm": 4.205423831939697,
"learning_rate": 7.609090909090909e-07,
"loss": 0.6285,
"step": 2470
},
{
"epoch": 0.89,
"grad_norm": 5.292863368988037,
"learning_rate": 7.598989898989898e-07,
"loss": 0.6531,
"step": 2480
},
{
"epoch": 0.89,
"grad_norm": 5.13388204574585,
"learning_rate": 7.588888888888888e-07,
"loss": 0.628,
"step": 2490
},
{
"epoch": 0.9,
"grad_norm": 2.864743232727051,
"learning_rate": 7.578787878787879e-07,
"loss": 0.5403,
"step": 2500
},
{
"epoch": 0.9,
"grad_norm": 4.284875869750977,
"learning_rate": 7.568686868686868e-07,
"loss": 0.6097,
"step": 2510
},
{
"epoch": 0.9,
"grad_norm": 3.3840548992156982,
"learning_rate": 7.558585858585859e-07,
"loss": 0.5759,
"step": 2520
},
{
"epoch": 0.91,
"grad_norm": 4.531360626220703,
"learning_rate": 7.548484848484848e-07,
"loss": 0.5581,
"step": 2530
},
{
"epoch": 0.91,
"grad_norm": 5.861146926879883,
"learning_rate": 7.538383838383838e-07,
"loss": 0.6329,
"step": 2540
},
{
"epoch": 0.91,
"grad_norm": 5.117300987243652,
"learning_rate": 7.528282828282828e-07,
"loss": 0.6071,
"step": 2550
},
{
"epoch": 0.92,
"grad_norm": 3.9243366718292236,
"learning_rate": 7.518181818181818e-07,
"loss": 0.5476,
"step": 2560
},
{
"epoch": 0.92,
"grad_norm": 6.215851783752441,
"learning_rate": 7.508080808080808e-07,
"loss": 0.5731,
"step": 2570
},
{
"epoch": 0.93,
"grad_norm": 6.189659118652344,
"learning_rate": 7.497979797979798e-07,
"loss": 0.592,
"step": 2580
},
{
"epoch": 0.93,
"grad_norm": 2.877923011779785,
"learning_rate": 7.487878787878787e-07,
"loss": 0.5435,
"step": 2590
},
{
"epoch": 0.93,
"grad_norm": 5.963223934173584,
"learning_rate": 7.477777777777778e-07,
"loss": 0.6104,
"step": 2600
},
{
"epoch": 0.94,
"grad_norm": 4.374766826629639,
"learning_rate": 7.467676767676767e-07,
"loss": 0.6022,
"step": 2610
},
{
"epoch": 0.94,
"grad_norm": 4.16854190826416,
"learning_rate": 7.457575757575756e-07,
"loss": 0.4931,
"step": 2620
},
{
"epoch": 0.94,
"grad_norm": 7.225739479064941,
"learning_rate": 7.447474747474747e-07,
"loss": 0.5659,
"step": 2630
},
{
"epoch": 0.95,
"grad_norm": 4.522436141967773,
"learning_rate": 7.437373737373736e-07,
"loss": 0.5449,
"step": 2640
},
{
"epoch": 0.95,
"grad_norm": 4.935425758361816,
"learning_rate": 7.427272727272727e-07,
"loss": 0.6093,
"step": 2650
},
{
"epoch": 0.95,
"grad_norm": 3.947935104370117,
"learning_rate": 7.417171717171716e-07,
"loss": 0.5976,
"step": 2660
},
{
"epoch": 0.96,
"grad_norm": 5.328210830688477,
"learning_rate": 7.407070707070707e-07,
"loss": 0.6283,
"step": 2670
},
{
"epoch": 0.96,
"grad_norm": 3.0664663314819336,
"learning_rate": 7.396969696969697e-07,
"loss": 0.5853,
"step": 2680
},
{
"epoch": 0.96,
"grad_norm": 5.055959701538086,
"learning_rate": 7.386868686868687e-07,
"loss": 0.5767,
"step": 2690
},
{
"epoch": 0.97,
"grad_norm": 5.318157196044922,
"learning_rate": 7.376767676767676e-07,
"loss": 0.5319,
"step": 2700
},
{
"epoch": 0.97,
"grad_norm": 3.5820364952087402,
"learning_rate": 7.366666666666667e-07,
"loss": 0.5506,
"step": 2710
},
{
"epoch": 0.98,
"grad_norm": 4.367839813232422,
"learning_rate": 7.356565656565656e-07,
"loss": 0.5524,
"step": 2720
},
{
"epoch": 0.98,
"grad_norm": 4.305497169494629,
"learning_rate": 7.346464646464647e-07,
"loss": 0.5639,
"step": 2730
},
{
"epoch": 0.98,
"grad_norm": 6.55611515045166,
"learning_rate": 7.336363636363636e-07,
"loss": 0.6346,
"step": 2740
},
{
"epoch": 0.99,
"grad_norm": 9.09347152709961,
"learning_rate": 7.326262626262626e-07,
"loss": 0.5574,
"step": 2750
},
{
"epoch": 0.99,
"grad_norm": 3.9730687141418457,
"learning_rate": 7.316161616161616e-07,
"loss": 0.5908,
"step": 2760
},
{
"epoch": 0.99,
"grad_norm": 16.01730728149414,
"learning_rate": 7.306060606060605e-07,
"loss": 0.5936,
"step": 2770
},
{
"epoch": 1.0,
"grad_norm": 5.077668190002441,
"learning_rate": 7.295959595959595e-07,
"loss": 0.5556,
"step": 2780
},
{
"epoch": 1.0,
"grad_norm": 4.859917640686035,
"learning_rate": 7.285858585858585e-07,
"loss": 0.5452,
"step": 2790
},
{
"epoch": 1.0,
"grad_norm": 5.687443733215332,
"learning_rate": 7.275757575757575e-07,
"loss": 0.6583,
"step": 2800
},
{
"epoch": 1.01,
"grad_norm": 4.369338035583496,
"learning_rate": 7.265656565656565e-07,
"loss": 0.5623,
"step": 2810
},
{
"epoch": 1.01,
"grad_norm": 4.035587310791016,
"learning_rate": 7.255555555555555e-07,
"loss": 0.571,
"step": 2820
},
{
"epoch": 1.01,
"grad_norm": 4.2013139724731445,
"learning_rate": 7.245454545454544e-07,
"loss": 0.4768,
"step": 2830
},
{
"epoch": 1.02,
"grad_norm": 3.272221565246582,
"learning_rate": 7.235353535353536e-07,
"loss": 0.581,
"step": 2840
},
{
"epoch": 1.02,
"grad_norm": 3.97727370262146,
"learning_rate": 7.225252525252525e-07,
"loss": 0.5654,
"step": 2850
},
{
"epoch": 1.03,
"grad_norm": 4.043779373168945,
"learning_rate": 7.215151515151516e-07,
"loss": 0.5971,
"step": 2860
},
{
"epoch": 1.03,
"grad_norm": 4.015261173248291,
"learning_rate": 7.205050505050505e-07,
"loss": 0.5921,
"step": 2870
},
{
"epoch": 1.03,
"grad_norm": 4.540653705596924,
"learning_rate": 7.194949494949495e-07,
"loss": 0.6103,
"step": 2880
},
{
"epoch": 1.04,
"grad_norm": 4.258879661560059,
"learning_rate": 7.184848484848485e-07,
"loss": 0.5519,
"step": 2890
},
{
"epoch": 1.04,
"grad_norm": 7.0088629722595215,
"learning_rate": 7.174747474747475e-07,
"loss": 0.6054,
"step": 2900
},
{
"epoch": 1.04,
"grad_norm": 3.9803833961486816,
"learning_rate": 7.164646464646464e-07,
"loss": 0.5557,
"step": 2910
},
{
"epoch": 1.05,
"grad_norm": 3.6390326023101807,
"learning_rate": 7.154545454545454e-07,
"loss": 0.5406,
"step": 2920
},
{
"epoch": 1.05,
"grad_norm": 3.998908042907715,
"learning_rate": 7.144444444444444e-07,
"loss": 0.604,
"step": 2930
},
{
"epoch": 1.05,
"grad_norm": 4.304749488830566,
"learning_rate": 7.134343434343433e-07,
"loss": 0.5859,
"step": 2940
},
{
"epoch": 1.06,
"grad_norm": 4.742584228515625,
"learning_rate": 7.124242424242424e-07,
"loss": 0.5184,
"step": 2950
},
{
"epoch": 1.06,
"grad_norm": 5.163829326629639,
"learning_rate": 7.114141414141413e-07,
"loss": 0.5669,
"step": 2960
},
{
"epoch": 1.07,
"grad_norm": 5.633606910705566,
"learning_rate": 7.104040404040404e-07,
"loss": 0.5967,
"step": 2970
},
{
"epoch": 1.07,
"grad_norm": 6.37916898727417,
"learning_rate": 7.093939393939393e-07,
"loss": 0.5971,
"step": 2980
},
{
"epoch": 1.07,
"grad_norm": 2.9268598556518555,
"learning_rate": 7.083838383838383e-07,
"loss": 0.5342,
"step": 2990
},
{
"epoch": 1.08,
"grad_norm": 4.331085205078125,
"learning_rate": 7.073737373737373e-07,
"loss": 0.5779,
"step": 3000
},
{
"epoch": 1.08,
"eval_loss": 0.7162447571754456,
"eval_runtime": 401.0241,
"eval_samples_per_second": 2.494,
"eval_steps_per_second": 2.494,
"step": 3000
},
{
"epoch": 1.08,
"grad_norm": 2.9604339599609375,
"learning_rate": 7.063636363636364e-07,
"loss": 0.5723,
"step": 3010
},
{
"epoch": 1.08,
"grad_norm": 4.6410393714904785,
"learning_rate": 7.053535353535354e-07,
"loss": 0.5986,
"step": 3020
},
{
"epoch": 1.09,
"grad_norm": 4.530472278594971,
"learning_rate": 7.043434343434344e-07,
"loss": 0.705,
"step": 3030
},
{
"epoch": 1.09,
"grad_norm": 4.028800964355469,
"learning_rate": 7.033333333333333e-07,
"loss": 0.5135,
"step": 3040
},
{
"epoch": 1.09,
"grad_norm": 2.882619619369507,
"learning_rate": 7.023232323232324e-07,
"loss": 0.5373,
"step": 3050
},
{
"epoch": 1.1,
"grad_norm": 5.280877590179443,
"learning_rate": 7.013131313131313e-07,
"loss": 0.5457,
"step": 3060
},
{
"epoch": 1.1,
"grad_norm": 3.983971357345581,
"learning_rate": 7.003030303030302e-07,
"loss": 0.5678,
"step": 3070
},
{
"epoch": 1.1,
"grad_norm": 3.4524683952331543,
"learning_rate": 6.992929292929293e-07,
"loss": 0.5794,
"step": 3080
},
{
"epoch": 1.11,
"grad_norm": 4.039379119873047,
"learning_rate": 6.982828282828282e-07,
"loss": 0.4778,
"step": 3090
},
{
"epoch": 1.11,
"grad_norm": 4.454843521118164,
"learning_rate": 6.972727272727273e-07,
"loss": 0.5312,
"step": 3100
},
{
"epoch": 1.12,
"grad_norm": 4.90205192565918,
"learning_rate": 6.962626262626262e-07,
"loss": 0.5675,
"step": 3110
},
{
"epoch": 1.12,
"grad_norm": 3.076164484024048,
"learning_rate": 6.952525252525252e-07,
"loss": 0.5949,
"step": 3120
},
{
"epoch": 1.12,
"grad_norm": 3.707550287246704,
"learning_rate": 6.942424242424242e-07,
"loss": 0.6053,
"step": 3130
},
{
"epoch": 1.13,
"grad_norm": 4.36667537689209,
"learning_rate": 6.932323232323232e-07,
"loss": 0.638,
"step": 3140
},
{
"epoch": 1.13,
"grad_norm": 4.635988235473633,
"learning_rate": 6.922222222222221e-07,
"loss": 0.5698,
"step": 3150
},
{
"epoch": 1.13,
"grad_norm": 3.769178628921509,
"learning_rate": 6.912121212121212e-07,
"loss": 0.5563,
"step": 3160
},
{
"epoch": 1.14,
"grad_norm": 3.096804618835449,
"learning_rate": 6.902020202020201e-07,
"loss": 0.5268,
"step": 3170
},
{
"epoch": 1.14,
"grad_norm": 5.9978461265563965,
"learning_rate": 6.891919191919193e-07,
"loss": 0.4878,
"step": 3180
},
{
"epoch": 1.14,
"grad_norm": 4.963356971740723,
"learning_rate": 6.881818181818182e-07,
"loss": 0.5904,
"step": 3190
},
{
"epoch": 1.15,
"grad_norm": 4.526729583740234,
"learning_rate": 6.871717171717171e-07,
"loss": 0.5957,
"step": 3200
},
{
"epoch": 1.15,
"grad_norm": 3.4197323322296143,
"learning_rate": 6.861616161616162e-07,
"loss": 0.5522,
"step": 3210
},
{
"epoch": 1.15,
"grad_norm": 3.989546060562134,
"learning_rate": 6.851515151515151e-07,
"loss": 0.6122,
"step": 3220
},
{
"epoch": 1.16,
"grad_norm": 5.137784004211426,
"learning_rate": 6.841414141414141e-07,
"loss": 0.58,
"step": 3230
},
{
"epoch": 1.16,
"grad_norm": 4.193089485168457,
"learning_rate": 6.831313131313131e-07,
"loss": 0.4918,
"step": 3240
},
{
"epoch": 1.17,
"grad_norm": 4.1724162101745605,
"learning_rate": 6.821212121212121e-07,
"loss": 0.6044,
"step": 3250
},
{
"epoch": 1.17,
"grad_norm": 4.983777046203613,
"learning_rate": 6.811111111111111e-07,
"loss": 0.533,
"step": 3260
},
{
"epoch": 1.17,
"grad_norm": 7.2310333251953125,
"learning_rate": 6.801010101010101e-07,
"loss": 0.5843,
"step": 3270
},
{
"epoch": 1.18,
"grad_norm": 4.620791912078857,
"learning_rate": 6.79090909090909e-07,
"loss": 0.5242,
"step": 3280
},
{
"epoch": 1.18,
"grad_norm": 4.3235087394714355,
"learning_rate": 6.780808080808081e-07,
"loss": 0.5836,
"step": 3290
},
{
"epoch": 1.18,
"grad_norm": 4.413332462310791,
"learning_rate": 6.77070707070707e-07,
"loss": 0.6532,
"step": 3300
},
{
"epoch": 1.19,
"grad_norm": 4.286377429962158,
"learning_rate": 6.76060606060606e-07,
"loss": 0.5527,
"step": 3310
},
{
"epoch": 1.19,
"grad_norm": 6.150529861450195,
"learning_rate": 6.75050505050505e-07,
"loss": 0.5745,
"step": 3320
},
{
"epoch": 1.19,
"grad_norm": 4.088225841522217,
"learning_rate": 6.74040404040404e-07,
"loss": 0.5882,
"step": 3330
},
{
"epoch": 1.2,
"grad_norm": 4.4364824295043945,
"learning_rate": 6.73030303030303e-07,
"loss": 0.513,
"step": 3340
},
{
"epoch": 1.2,
"grad_norm": 3.8595519065856934,
"learning_rate": 6.72020202020202e-07,
"loss": 0.5621,
"step": 3350
},
{
"epoch": 1.2,
"grad_norm": 6.183053016662598,
"learning_rate": 6.71010101010101e-07,
"loss": 0.5318,
"step": 3360
},
{
"epoch": 1.21,
"grad_norm": 4.677096366882324,
"learning_rate": 6.7e-07,
"loss": 0.4965,
"step": 3370
},
{
"epoch": 1.21,
"grad_norm": 5.507148742675781,
"learning_rate": 6.68989898989899e-07,
"loss": 0.569,
"step": 3380
},
{
"epoch": 1.22,
"grad_norm": 3.813816547393799,
"learning_rate": 6.679797979797979e-07,
"loss": 0.5326,
"step": 3390
},
{
"epoch": 1.22,
"grad_norm": 5.973514080047607,
"learning_rate": 6.66969696969697e-07,
"loss": 0.6733,
"step": 3400
},
{
"epoch": 1.22,
"grad_norm": 5.510775089263916,
"learning_rate": 6.659595959595959e-07,
"loss": 0.5978,
"step": 3410
},
{
"epoch": 1.23,
"grad_norm": 3.8347678184509277,
"learning_rate": 6.64949494949495e-07,
"loss": 0.5141,
"step": 3420
},
{
"epoch": 1.23,
"grad_norm": 4.097416877746582,
"learning_rate": 6.639393939393939e-07,
"loss": 0.5503,
"step": 3430
},
{
"epoch": 1.23,
"grad_norm": 3.52165150642395,
"learning_rate": 6.629292929292929e-07,
"loss": 0.5506,
"step": 3440
},
{
"epoch": 1.24,
"grad_norm": 4.4569573402404785,
"learning_rate": 6.619191919191919e-07,
"loss": 0.6121,
"step": 3450
},
{
"epoch": 1.24,
"grad_norm": 4.094751834869385,
"learning_rate": 6.609090909090909e-07,
"loss": 0.5562,
"step": 3460
},
{
"epoch": 1.24,
"grad_norm": 4.277998447418213,
"learning_rate": 6.598989898989898e-07,
"loss": 0.5432,
"step": 3470
},
{
"epoch": 1.25,
"grad_norm": 5.9544148445129395,
"learning_rate": 6.588888888888889e-07,
"loss": 0.5891,
"step": 3480
},
{
"epoch": 1.25,
"grad_norm": 5.688138008117676,
"learning_rate": 6.578787878787878e-07,
"loss": 0.6172,
"step": 3490
},
{
"epoch": 1.26,
"grad_norm": 5.771819591522217,
"learning_rate": 6.568686868686868e-07,
"loss": 0.5362,
"step": 3500
},
{
"epoch": 1.26,
"grad_norm": 4.87513542175293,
"learning_rate": 6.558585858585858e-07,
"loss": 0.5608,
"step": 3510
},
{
"epoch": 1.26,
"grad_norm": 3.709472179412842,
"learning_rate": 6.548484848484848e-07,
"loss": 0.5361,
"step": 3520
},
{
"epoch": 1.27,
"grad_norm": 4.465284824371338,
"learning_rate": 6.538383838383839e-07,
"loss": 0.5435,
"step": 3530
},
{
"epoch": 1.27,
"grad_norm": 4.30505895614624,
"learning_rate": 6.528282828282828e-07,
"loss": 0.5714,
"step": 3540
},
{
"epoch": 1.27,
"grad_norm": 4.445021152496338,
"learning_rate": 6.518181818181818e-07,
"loss": 0.5773,
"step": 3550
},
{
"epoch": 1.28,
"grad_norm": 4.173781394958496,
"learning_rate": 6.508080808080808e-07,
"loss": 0.6223,
"step": 3560
},
{
"epoch": 1.28,
"grad_norm": 5.676968574523926,
"learning_rate": 6.497979797979798e-07,
"loss": 0.5957,
"step": 3570
},
{
"epoch": 1.28,
"grad_norm": 3.447176694869995,
"learning_rate": 6.487878787878788e-07,
"loss": 0.5896,
"step": 3580
},
{
"epoch": 1.29,
"grad_norm": 4.1883111000061035,
"learning_rate": 6.477777777777778e-07,
"loss": 0.5602,
"step": 3590
},
{
"epoch": 1.29,
"grad_norm": 4.16115140914917,
"learning_rate": 6.467676767676767e-07,
"loss": 0.5752,
"step": 3600
},
{
"epoch": 1.29,
"grad_norm": 4.55715274810791,
"learning_rate": 6.457575757575758e-07,
"loss": 0.5344,
"step": 3610
},
{
"epoch": 1.3,
"grad_norm": 5.83474588394165,
"learning_rate": 6.447474747474747e-07,
"loss": 0.5867,
"step": 3620
},
{
"epoch": 1.3,
"grad_norm": 2.700615644454956,
"learning_rate": 6.437373737373736e-07,
"loss": 0.5395,
"step": 3630
},
{
"epoch": 1.31,
"grad_norm": 4.691040992736816,
"learning_rate": 6.428282828282828e-07,
"loss": 0.5509,
"step": 3640
},
{
"epoch": 1.31,
"grad_norm": 4.966342449188232,
"learning_rate": 6.418181818181818e-07,
"loss": 0.6047,
"step": 3650
},
{
"epoch": 1.31,
"grad_norm": 3.6059834957122803,
"learning_rate": 6.408080808080808e-07,
"loss": 0.5338,
"step": 3660
},
{
"epoch": 1.32,
"grad_norm": 3.471660852432251,
"learning_rate": 6.397979797979798e-07,
"loss": 0.5532,
"step": 3670
},
{
"epoch": 1.32,
"grad_norm": 5.410092353820801,
"learning_rate": 6.387878787878788e-07,
"loss": 0.5739,
"step": 3680
},
{
"epoch": 1.32,
"grad_norm": 4.829250335693359,
"learning_rate": 6.377777777777778e-07,
"loss": 0.6064,
"step": 3690
},
{
"epoch": 1.33,
"grad_norm": 5.880673408508301,
"learning_rate": 6.367676767676767e-07,
"loss": 0.575,
"step": 3700
},
{
"epoch": 1.33,
"grad_norm": 3.647641658782959,
"learning_rate": 6.357575757575757e-07,
"loss": 0.594,
"step": 3710
},
{
"epoch": 1.33,
"grad_norm": 4.267665386199951,
"learning_rate": 6.347474747474747e-07,
"loss": 0.574,
"step": 3720
},
{
"epoch": 1.34,
"grad_norm": 3.5303924083709717,
"learning_rate": 6.337373737373736e-07,
"loss": 0.496,
"step": 3730
},
{
"epoch": 1.34,
"grad_norm": 3.1564202308654785,
"learning_rate": 6.327272727272727e-07,
"loss": 0.5196,
"step": 3740
},
{
"epoch": 1.34,
"grad_norm": 4.458827495574951,
"learning_rate": 6.317171717171716e-07,
"loss": 0.5616,
"step": 3750
},
{
"epoch": 1.35,
"grad_norm": 4.057519435882568,
"learning_rate": 6.307070707070707e-07,
"loss": 0.5293,
"step": 3760
},
{
"epoch": 1.35,
"grad_norm": 5.7592363357543945,
"learning_rate": 6.296969696969696e-07,
"loss": 0.5433,
"step": 3770
},
{
"epoch": 1.36,
"grad_norm": 5.579089641571045,
"learning_rate": 6.286868686868687e-07,
"loss": 0.546,
"step": 3780
},
{
"epoch": 1.36,
"grad_norm": 3.8911049365997314,
"learning_rate": 6.276767676767677e-07,
"loss": 0.5991,
"step": 3790
},
{
"epoch": 1.36,
"grad_norm": 6.364475727081299,
"learning_rate": 6.266666666666667e-07,
"loss": 0.5908,
"step": 3800
},
{
"epoch": 1.37,
"grad_norm": 5.0971760749816895,
"learning_rate": 6.256565656565656e-07,
"loss": 0.5915,
"step": 3810
},
{
"epoch": 1.37,
"grad_norm": 3.897958755493164,
"learning_rate": 6.246464646464647e-07,
"loss": 0.5859,
"step": 3820
},
{
"epoch": 1.37,
"grad_norm": 6.308644771575928,
"learning_rate": 6.236363636363636e-07,
"loss": 0.5991,
"step": 3830
},
{
"epoch": 1.38,
"grad_norm": 3.0512614250183105,
"learning_rate": 6.226262626262627e-07,
"loss": 0.5742,
"step": 3840
},
{
"epoch": 1.38,
"grad_norm": 4.052495956420898,
"learning_rate": 6.216161616161616e-07,
"loss": 0.54,
"step": 3850
},
{
"epoch": 1.38,
"grad_norm": 4.723029613494873,
"learning_rate": 6.206060606060605e-07,
"loss": 0.5524,
"step": 3860
},
{
"epoch": 1.39,
"grad_norm": 4.705147743225098,
"learning_rate": 6.195959595959596e-07,
"loss": 0.5924,
"step": 3870
},
{
"epoch": 1.39,
"grad_norm": 4.098758697509766,
"learning_rate": 6.185858585858585e-07,
"loss": 0.5482,
"step": 3880
},
{
"epoch": 1.4,
"grad_norm": 2.9597623348236084,
"learning_rate": 6.175757575757575e-07,
"loss": 0.5521,
"step": 3890
},
{
"epoch": 1.4,
"grad_norm": 3.7308359146118164,
"learning_rate": 6.165656565656565e-07,
"loss": 0.5645,
"step": 3900
},
{
"epoch": 1.4,
"grad_norm": 4.260051250457764,
"learning_rate": 6.155555555555555e-07,
"loss": 0.5471,
"step": 3910
},
{
"epoch": 1.41,
"grad_norm": 6.539883136749268,
"learning_rate": 6.145454545454545e-07,
"loss": 0.5439,
"step": 3920
},
{
"epoch": 1.41,
"grad_norm": 6.235405921936035,
"learning_rate": 6.135353535353535e-07,
"loss": 0.5435,
"step": 3930
},
{
"epoch": 1.41,
"grad_norm": 3.3652477264404297,
"learning_rate": 6.125252525252524e-07,
"loss": 0.5544,
"step": 3940
},
{
"epoch": 1.42,
"grad_norm": 3.795522928237915,
"learning_rate": 6.115151515151516e-07,
"loss": 0.5794,
"step": 3950
},
{
"epoch": 1.42,
"grad_norm": 5.496920108795166,
"learning_rate": 6.105050505050505e-07,
"loss": 0.5133,
"step": 3960
},
{
"epoch": 1.42,
"grad_norm": 5.946590423583984,
"learning_rate": 6.094949494949495e-07,
"loss": 0.6391,
"step": 3970
},
{
"epoch": 1.43,
"grad_norm": 3.943232774734497,
"learning_rate": 6.084848484848485e-07,
"loss": 0.5574,
"step": 3980
},
{
"epoch": 1.43,
"grad_norm": 4.132944107055664,
"learning_rate": 6.074747474747474e-07,
"loss": 0.5014,
"step": 3990
},
{
"epoch": 1.43,
"grad_norm": 4.284780979156494,
"learning_rate": 6.064646464646465e-07,
"loss": 0.5907,
"step": 4000
},
{
"epoch": 1.43,
"eval_loss": 0.698131263256073,
"eval_runtime": 401.1999,
"eval_samples_per_second": 2.493,
"eval_steps_per_second": 2.493,
"step": 4000
},
{
"epoch": 1.44,
"grad_norm": 5.0691447257995605,
"learning_rate": 6.054545454545454e-07,
"loss": 0.5153,
"step": 4010
},
{
"epoch": 1.44,
"grad_norm": 4.563512325286865,
"learning_rate": 6.044444444444444e-07,
"loss": 0.5323,
"step": 4020
},
{
"epoch": 1.45,
"grad_norm": 8.346761703491211,
"learning_rate": 6.034343434343434e-07,
"loss": 0.6202,
"step": 4030
},
{
"epoch": 1.45,
"grad_norm": 4.459041118621826,
"learning_rate": 6.024242424242424e-07,
"loss": 0.5081,
"step": 4040
},
{
"epoch": 1.45,
"grad_norm": 6.5770158767700195,
"learning_rate": 6.014141414141413e-07,
"loss": 0.568,
"step": 4050
},
{
"epoch": 1.46,
"grad_norm": 5.020130634307861,
"learning_rate": 6.004040404040404e-07,
"loss": 0.5027,
"step": 4060
},
{
"epoch": 1.46,
"grad_norm": 6.139995098114014,
"learning_rate": 5.993939393939393e-07,
"loss": 0.5115,
"step": 4070
},
{
"epoch": 1.46,
"grad_norm": 5.601007461547852,
"learning_rate": 5.983838383838384e-07,
"loss": 0.5331,
"step": 4080
},
{
"epoch": 1.47,
"grad_norm": 4.264581680297852,
"learning_rate": 5.973737373737373e-07,
"loss": 0.5474,
"step": 4090
},
{
"epoch": 1.47,
"grad_norm": 3.254601001739502,
"learning_rate": 5.963636363636363e-07,
"loss": 0.5515,
"step": 4100
},
{
"epoch": 1.47,
"grad_norm": 4.111956596374512,
"learning_rate": 5.953535353535353e-07,
"loss": 0.5486,
"step": 4110
},
{
"epoch": 1.48,
"grad_norm": 3.814635992050171,
"learning_rate": 5.943434343434344e-07,
"loss": 0.5332,
"step": 4120
},
{
"epoch": 1.48,
"grad_norm": 4.161899089813232,
"learning_rate": 5.933333333333334e-07,
"loss": 0.5713,
"step": 4130
},
{
"epoch": 1.48,
"grad_norm": 3.490006446838379,
"learning_rate": 5.923232323232323e-07,
"loss": 0.5522,
"step": 4140
},
{
"epoch": 1.49,
"grad_norm": 4.460766792297363,
"learning_rate": 5.913131313131313e-07,
"loss": 0.515,
"step": 4150
},
{
"epoch": 1.49,
"grad_norm": 5.571935176849365,
"learning_rate": 5.903030303030303e-07,
"loss": 0.5976,
"step": 4160
},
{
"epoch": 1.5,
"grad_norm": 4.256524562835693,
"learning_rate": 5.892929292929293e-07,
"loss": 0.5068,
"step": 4170
},
{
"epoch": 1.5,
"grad_norm": 5.909682273864746,
"learning_rate": 5.882828282828282e-07,
"loss": 0.5473,
"step": 4180
},
{
"epoch": 1.5,
"grad_norm": 6.469602584838867,
"learning_rate": 5.872727272727273e-07,
"loss": 0.5715,
"step": 4190
},
{
"epoch": 1.51,
"grad_norm": 4.836788654327393,
"learning_rate": 5.862626262626262e-07,
"loss": 0.4511,
"step": 4200
},
{
"epoch": 1.51,
"grad_norm": 2.996185302734375,
"learning_rate": 5.852525252525252e-07,
"loss": 0.5782,
"step": 4210
},
{
"epoch": 1.51,
"grad_norm": 8.92146110534668,
"learning_rate": 5.842424242424242e-07,
"loss": 0.5329,
"step": 4220
},
{
"epoch": 1.52,
"grad_norm": 6.146700382232666,
"learning_rate": 5.832323232323232e-07,
"loss": 0.5768,
"step": 4230
},
{
"epoch": 1.52,
"grad_norm": 4.365715026855469,
"learning_rate": 5.822222222222222e-07,
"loss": 0.5804,
"step": 4240
},
{
"epoch": 1.52,
"grad_norm": 4.362324237823486,
"learning_rate": 5.812121212121212e-07,
"loss": 0.5435,
"step": 4250
},
{
"epoch": 1.53,
"grad_norm": 4.051389217376709,
"learning_rate": 5.802020202020201e-07,
"loss": 0.5662,
"step": 4260
},
{
"epoch": 1.53,
"grad_norm": 4.774472236633301,
"learning_rate": 5.791919191919192e-07,
"loss": 0.523,
"step": 4270
},
{
"epoch": 1.53,
"grad_norm": 4.41732120513916,
"learning_rate": 5.781818181818181e-07,
"loss": 0.5926,
"step": 4280
},
{
"epoch": 1.54,
"grad_norm": 5.613069534301758,
"learning_rate": 5.771717171717172e-07,
"loss": 0.5205,
"step": 4290
},
{
"epoch": 1.54,
"grad_norm": 4.26994514465332,
"learning_rate": 5.761616161616162e-07,
"loss": 0.5605,
"step": 4300
},
{
"epoch": 1.55,
"grad_norm": 4.853590488433838,
"learning_rate": 5.751515151515151e-07,
"loss": 0.587,
"step": 4310
},
{
"epoch": 1.55,
"grad_norm": 5.43099308013916,
"learning_rate": 5.741414141414142e-07,
"loss": 0.533,
"step": 4320
},
{
"epoch": 1.55,
"grad_norm": 4.218691349029541,
"learning_rate": 5.731313131313131e-07,
"loss": 0.5373,
"step": 4330
},
{
"epoch": 1.56,
"grad_norm": 6.392608642578125,
"learning_rate": 5.721212121212121e-07,
"loss": 0.5346,
"step": 4340
},
{
"epoch": 1.56,
"grad_norm": 4.781940937042236,
"learning_rate": 5.711111111111111e-07,
"loss": 0.6316,
"step": 4350
},
{
"epoch": 1.56,
"grad_norm": 4.0989990234375,
"learning_rate": 5.702020202020201e-07,
"loss": 0.5863,
"step": 4360
},
{
"epoch": 1.57,
"grad_norm": 4.195807456970215,
"learning_rate": 5.691919191919191e-07,
"loss": 0.6091,
"step": 4370
},
{
"epoch": 1.57,
"grad_norm": 3.2372219562530518,
"learning_rate": 5.681818181818182e-07,
"loss": 0.5939,
"step": 4380
},
{
"epoch": 1.57,
"grad_norm": 3.523815631866455,
"learning_rate": 5.671717171717172e-07,
"loss": 0.5908,
"step": 4390
},
{
"epoch": 1.58,
"grad_norm": 4.94608736038208,
"learning_rate": 5.661616161616162e-07,
"loss": 0.5964,
"step": 4400
},
{
"epoch": 1.58,
"grad_norm": 8.730268478393555,
"learning_rate": 5.651515151515151e-07,
"loss": 0.5355,
"step": 4410
},
{
"epoch": 1.59,
"grad_norm": 4.550642490386963,
"learning_rate": 5.641414141414142e-07,
"loss": 0.5295,
"step": 4420
},
{
"epoch": 1.59,
"grad_norm": 4.5829176902771,
"learning_rate": 5.631313131313131e-07,
"loss": 0.5915,
"step": 4430
},
{
"epoch": 1.59,
"grad_norm": 4.118924617767334,
"learning_rate": 5.621212121212121e-07,
"loss": 0.5589,
"step": 4440
},
{
"epoch": 1.6,
"grad_norm": 4.678675174713135,
"learning_rate": 5.611111111111111e-07,
"loss": 0.6621,
"step": 4450
},
{
"epoch": 1.6,
"grad_norm": 4.09968376159668,
"learning_rate": 5.601010101010101e-07,
"loss": 0.6018,
"step": 4460
},
{
"epoch": 1.6,
"grad_norm": 6.312856674194336,
"learning_rate": 5.590909090909091e-07,
"loss": 0.6402,
"step": 4470
},
{
"epoch": 1.61,
"grad_norm": 5.487626552581787,
"learning_rate": 5.58080808080808e-07,
"loss": 0.5799,
"step": 4480
},
{
"epoch": 1.61,
"grad_norm": 4.387622833251953,
"learning_rate": 5.57070707070707e-07,
"loss": 0.5685,
"step": 4490
},
{
"epoch": 1.61,
"grad_norm": 4.692462921142578,
"learning_rate": 5.56060606060606e-07,
"loss": 0.5935,
"step": 4500
},
{
"epoch": 1.62,
"grad_norm": 3.9311652183532715,
"learning_rate": 5.55050505050505e-07,
"loss": 0.6092,
"step": 4510
},
{
"epoch": 1.62,
"grad_norm": 4.879345417022705,
"learning_rate": 5.540404040404039e-07,
"loss": 0.5754,
"step": 4520
},
{
"epoch": 1.62,
"grad_norm": 4.526495933532715,
"learning_rate": 5.53030303030303e-07,
"loss": 0.6167,
"step": 4530
},
{
"epoch": 1.63,
"grad_norm": 5.269490718841553,
"learning_rate": 5.520202020202019e-07,
"loss": 0.5197,
"step": 4540
},
{
"epoch": 1.63,
"grad_norm": 6.770501613616943,
"learning_rate": 5.510101010101011e-07,
"loss": 0.5389,
"step": 4550
},
{
"epoch": 1.64,
"grad_norm": 5.688599109649658,
"learning_rate": 5.5e-07,
"loss": 0.518,
"step": 4560
},
{
"epoch": 1.64,
"grad_norm": 3.049668073654175,
"learning_rate": 5.48989898989899e-07,
"loss": 0.6034,
"step": 4570
},
{
"epoch": 1.64,
"grad_norm": 5.590402126312256,
"learning_rate": 5.47979797979798e-07,
"loss": 0.6146,
"step": 4580
},
{
"epoch": 1.65,
"grad_norm": 4.051222801208496,
"learning_rate": 5.46969696969697e-07,
"loss": 0.5388,
"step": 4590
},
{
"epoch": 1.65,
"grad_norm": 3.68947434425354,
"learning_rate": 5.459595959595959e-07,
"loss": 0.5049,
"step": 4600
},
{
"epoch": 1.65,
"grad_norm": 5.772908687591553,
"learning_rate": 5.44949494949495e-07,
"loss": 0.5447,
"step": 4610
},
{
"epoch": 1.66,
"grad_norm": 3.907376289367676,
"learning_rate": 5.439393939393939e-07,
"loss": 0.5038,
"step": 4620
},
{
"epoch": 1.66,
"grad_norm": 6.761162281036377,
"learning_rate": 5.42929292929293e-07,
"loss": 0.6141,
"step": 4630
},
{
"epoch": 1.66,
"grad_norm": 3.054441452026367,
"learning_rate": 5.419191919191919e-07,
"loss": 0.6043,
"step": 4640
},
{
"epoch": 1.67,
"grad_norm": 5.1576457023620605,
"learning_rate": 5.409090909090908e-07,
"loss": 0.5588,
"step": 4650
},
{
"epoch": 1.67,
"grad_norm": 4.096390724182129,
"learning_rate": 5.398989898989899e-07,
"loss": 0.6355,
"step": 4660
},
{
"epoch": 1.67,
"grad_norm": 3.5117170810699463,
"learning_rate": 5.388888888888888e-07,
"loss": 0.6043,
"step": 4670
},
{
"epoch": 1.68,
"grad_norm": 3.7968451976776123,
"learning_rate": 5.378787878787878e-07,
"loss": 0.6092,
"step": 4680
},
{
"epoch": 1.68,
"grad_norm": 5.3210320472717285,
"learning_rate": 5.368686868686868e-07,
"loss": 0.5375,
"step": 4690
},
{
"epoch": 1.69,
"grad_norm": 3.6537818908691406,
"learning_rate": 5.358585858585858e-07,
"loss": 0.5013,
"step": 4700
},
{
"epoch": 1.69,
"grad_norm": 3.1777806282043457,
"learning_rate": 5.348484848484848e-07,
"loss": 0.5073,
"step": 4710
},
{
"epoch": 1.69,
"grad_norm": 5.203103065490723,
"learning_rate": 5.338383838383839e-07,
"loss": 0.5694,
"step": 4720
},
{
"epoch": 1.7,
"grad_norm": 3.941244602203369,
"learning_rate": 5.328282828282828e-07,
"loss": 0.5337,
"step": 4730
},
{
"epoch": 1.7,
"grad_norm": 3.7557549476623535,
"learning_rate": 5.318181818181819e-07,
"loss": 0.6047,
"step": 4740
},
{
"epoch": 1.7,
"grad_norm": 5.3054351806640625,
"learning_rate": 5.308080808080808e-07,
"loss": 0.5563,
"step": 4750
},
{
"epoch": 1.71,
"grad_norm": 3.023254871368408,
"learning_rate": 5.297979797979798e-07,
"loss": 0.5877,
"step": 4760
},
{
"epoch": 1.71,
"grad_norm": 3.8223531246185303,
"learning_rate": 5.287878787878788e-07,
"loss": 0.5938,
"step": 4770
},
{
"epoch": 1.71,
"grad_norm": 2.889775276184082,
"learning_rate": 5.277777777777777e-07,
"loss": 0.5628,
"step": 4780
},
{
"epoch": 1.72,
"grad_norm": 4.528804302215576,
"learning_rate": 5.267676767676768e-07,
"loss": 0.5518,
"step": 4790
},
{
"epoch": 1.72,
"grad_norm": 4.486887454986572,
"learning_rate": 5.257575757575757e-07,
"loss": 0.6116,
"step": 4800
},
{
"epoch": 1.72,
"grad_norm": 3.2164041996002197,
"learning_rate": 5.247474747474747e-07,
"loss": 0.5949,
"step": 4810
},
{
"epoch": 1.73,
"grad_norm": 6.722822666168213,
"learning_rate": 5.237373737373737e-07,
"loss": 0.6425,
"step": 4820
},
{
"epoch": 1.73,
"grad_norm": 4.429785251617432,
"learning_rate": 5.227272727272727e-07,
"loss": 0.5447,
"step": 4830
},
{
"epoch": 1.74,
"grad_norm": 4.903067111968994,
"learning_rate": 5.217171717171716e-07,
"loss": 0.5635,
"step": 4840
},
{
"epoch": 1.74,
"grad_norm": 6.777737617492676,
"learning_rate": 5.207070707070707e-07,
"loss": 0.5953,
"step": 4850
},
{
"epoch": 1.74,
"grad_norm": 5.027159214019775,
"learning_rate": 5.196969696969696e-07,
"loss": 0.598,
"step": 4860
},
{
"epoch": 1.75,
"grad_norm": 3.1004555225372314,
"learning_rate": 5.186868686868687e-07,
"loss": 0.5972,
"step": 4870
},
{
"epoch": 1.75,
"grad_norm": 6.468344211578369,
"learning_rate": 5.176767676767676e-07,
"loss": 0.5707,
"step": 4880
},
{
"epoch": 1.75,
"grad_norm": 4.483850955963135,
"learning_rate": 5.166666666666667e-07,
"loss": 0.5488,
"step": 4890
},
{
"epoch": 1.76,
"grad_norm": 4.9485602378845215,
"learning_rate": 5.156565656565657e-07,
"loss": 0.5356,
"step": 4900
},
{
"epoch": 1.76,
"grad_norm": 3.1415200233459473,
"learning_rate": 5.146464646464647e-07,
"loss": 0.5455,
"step": 4910
},
{
"epoch": 1.76,
"grad_norm": 4.05121374130249,
"learning_rate": 5.136363636363636e-07,
"loss": 0.5762,
"step": 4920
},
{
"epoch": 1.77,
"grad_norm": 4.803497791290283,
"learning_rate": 5.126262626262626e-07,
"loss": 0.6569,
"step": 4930
},
{
"epoch": 1.77,
"grad_norm": 3.7656466960906982,
"learning_rate": 5.116161616161616e-07,
"loss": 0.5274,
"step": 4940
},
{
"epoch": 1.78,
"grad_norm": 6.404185771942139,
"learning_rate": 5.106060606060606e-07,
"loss": 0.599,
"step": 4950
},
{
"epoch": 1.78,
"grad_norm": 4.116934299468994,
"learning_rate": 5.095959595959596e-07,
"loss": 0.5658,
"step": 4960
},
{
"epoch": 1.78,
"grad_norm": 3.6211934089660645,
"learning_rate": 5.085858585858585e-07,
"loss": 0.6003,
"step": 4970
},
{
"epoch": 1.79,
"grad_norm": 4.541522979736328,
"learning_rate": 5.075757575757576e-07,
"loss": 0.5854,
"step": 4980
},
{
"epoch": 1.79,
"grad_norm": 4.65029764175415,
"learning_rate": 5.065656565656565e-07,
"loss": 0.5504,
"step": 4990
},
{
"epoch": 1.79,
"grad_norm": 4.597330093383789,
"learning_rate": 5.055555555555555e-07,
"loss": 0.5244,
"step": 5000
},
{
"epoch": 1.79,
"eval_loss": 0.6935617923736572,
"eval_runtime": 401.44,
"eval_samples_per_second": 2.491,
"eval_steps_per_second": 2.491,
"step": 5000
},
{
"epoch": 1.8,
"grad_norm": 4.305992126464844,
"learning_rate": 5.045454545454545e-07,
"loss": 0.5294,
"step": 5010
},
{
"epoch": 1.8,
"grad_norm": 9.865640640258789,
"learning_rate": 5.035353535353535e-07,
"loss": 0.5933,
"step": 5020
},
{
"epoch": 1.8,
"grad_norm": 4.958790302276611,
"learning_rate": 5.025252525252525e-07,
"loss": 0.515,
"step": 5030
},
{
"epoch": 1.81,
"grad_norm": 4.508664131164551,
"learning_rate": 5.015151515151515e-07,
"loss": 0.5725,
"step": 5040
},
{
"epoch": 1.81,
"grad_norm": 5.282695293426514,
"learning_rate": 5.005050505050504e-07,
"loss": 0.5779,
"step": 5050
},
{
"epoch": 1.81,
"grad_norm": 3.441600799560547,
"learning_rate": 4.994949494949494e-07,
"loss": 0.5832,
"step": 5060
},
{
"epoch": 1.82,
"grad_norm": 6.297787666320801,
"learning_rate": 4.984848484848485e-07,
"loss": 0.5923,
"step": 5070
},
{
"epoch": 1.82,
"grad_norm": 4.202699184417725,
"learning_rate": 4.974747474747474e-07,
"loss": 0.5525,
"step": 5080
},
{
"epoch": 1.83,
"grad_norm": 6.49127721786499,
"learning_rate": 4.964646464646464e-07,
"loss": 0.5443,
"step": 5090
},
{
"epoch": 1.83,
"grad_norm": 5.5521345138549805,
"learning_rate": 4.954545454545454e-07,
"loss": 0.5871,
"step": 5100
},
{
"epoch": 1.83,
"grad_norm": 6.344796657562256,
"learning_rate": 4.944444444444445e-07,
"loss": 0.5371,
"step": 5110
},
{
"epoch": 1.84,
"grad_norm": 4.388942718505859,
"learning_rate": 4.934343434343434e-07,
"loss": 0.4874,
"step": 5120
},
{
"epoch": 1.84,
"grad_norm": 5.535308837890625,
"learning_rate": 4.924242424242424e-07,
"loss": 0.5937,
"step": 5130
},
{
"epoch": 1.84,
"grad_norm": 3.587024450302124,
"learning_rate": 4.914141414141414e-07,
"loss": 0.5893,
"step": 5140
},
{
"epoch": 1.85,
"grad_norm": 4.178186416625977,
"learning_rate": 4.904040404040404e-07,
"loss": 0.569,
"step": 5150
},
{
"epoch": 1.85,
"grad_norm": 4.824371337890625,
"learning_rate": 4.893939393939393e-07,
"loss": 0.5923,
"step": 5160
},
{
"epoch": 1.85,
"grad_norm": 3.2746853828430176,
"learning_rate": 4.883838383838384e-07,
"loss": 0.5618,
"step": 5170
},
{
"epoch": 1.86,
"grad_norm": 4.693532943725586,
"learning_rate": 4.873737373737373e-07,
"loss": 0.5875,
"step": 5180
},
{
"epoch": 1.86,
"grad_norm": 4.339977264404297,
"learning_rate": 4.863636363636364e-07,
"loss": 0.5863,
"step": 5190
},
{
"epoch": 1.86,
"grad_norm": 4.782194137573242,
"learning_rate": 4.853535353535353e-07,
"loss": 0.5649,
"step": 5200
},
{
"epoch": 1.87,
"grad_norm": 5.156185626983643,
"learning_rate": 4.843434343434343e-07,
"loss": 0.5377,
"step": 5210
},
{
"epoch": 1.87,
"grad_norm": 4.788945198059082,
"learning_rate": 4.833333333333333e-07,
"loss": 0.4862,
"step": 5220
},
{
"epoch": 1.88,
"grad_norm": 4.9057817459106445,
"learning_rate": 4.823232323232323e-07,
"loss": 0.5838,
"step": 5230
},
{
"epoch": 1.88,
"grad_norm": 5.085701942443848,
"learning_rate": 4.813131313131313e-07,
"loss": 0.6107,
"step": 5240
},
{
"epoch": 1.88,
"grad_norm": 3.601012945175171,
"learning_rate": 4.803030303030302e-07,
"loss": 0.4859,
"step": 5250
},
{
"epoch": 1.89,
"grad_norm": 5.70768404006958,
"learning_rate": 4.792929292929293e-07,
"loss": 0.5206,
"step": 5260
},
{
"epoch": 1.89,
"grad_norm": 5.292908191680908,
"learning_rate": 4.782828282828283e-07,
"loss": 0.5652,
"step": 5270
},
{
"epoch": 1.89,
"grad_norm": 5.359946250915527,
"learning_rate": 4.772727272727273e-07,
"loss": 0.5807,
"step": 5280
},
{
"epoch": 1.9,
"grad_norm": 4.304328441619873,
"learning_rate": 4.7626262626262627e-07,
"loss": 0.5867,
"step": 5290
},
{
"epoch": 1.9,
"grad_norm": 3.993483304977417,
"learning_rate": 4.752525252525252e-07,
"loss": 0.6,
"step": 5300
},
{
"epoch": 1.9,
"grad_norm": 5.294483661651611,
"learning_rate": 4.742424242424242e-07,
"loss": 0.6168,
"step": 5310
},
{
"epoch": 1.91,
"grad_norm": 5.31272554397583,
"learning_rate": 4.732323232323232e-07,
"loss": 0.5645,
"step": 5320
},
{
"epoch": 1.91,
"grad_norm": 3.8329992294311523,
"learning_rate": 4.722222222222222e-07,
"loss": 0.5978,
"step": 5330
},
{
"epoch": 1.92,
"grad_norm": 5.329180717468262,
"learning_rate": 4.7121212121212115e-07,
"loss": 0.562,
"step": 5340
},
{
"epoch": 1.92,
"grad_norm": 9.958372116088867,
"learning_rate": 4.7020202020202015e-07,
"loss": 0.6556,
"step": 5350
},
{
"epoch": 1.92,
"grad_norm": 7.695910453796387,
"learning_rate": 4.691919191919192e-07,
"loss": 0.4964,
"step": 5360
},
{
"epoch": 1.93,
"grad_norm": 3.406792163848877,
"learning_rate": 4.681818181818182e-07,
"loss": 0.5105,
"step": 5370
},
{
"epoch": 1.93,
"grad_norm": 6.779275894165039,
"learning_rate": 4.6717171717171714e-07,
"loss": 0.4962,
"step": 5380
},
{
"epoch": 1.93,
"grad_norm": 6.590269088745117,
"learning_rate": 4.6616161616161614e-07,
"loss": 0.5726,
"step": 5390
},
{
"epoch": 1.94,
"grad_norm": 3.9702072143554688,
"learning_rate": 4.6515151515151513e-07,
"loss": 0.5568,
"step": 5400
},
{
"epoch": 1.94,
"grad_norm": 4.912775039672852,
"learning_rate": 4.6414141414141413e-07,
"loss": 0.5604,
"step": 5410
},
{
"epoch": 1.94,
"grad_norm": 5.167705059051514,
"learning_rate": 4.6313131313131307e-07,
"loss": 0.5364,
"step": 5420
},
{
"epoch": 1.95,
"grad_norm": 5.220252513885498,
"learning_rate": 4.6212121212121207e-07,
"loss": 0.5841,
"step": 5430
},
{
"epoch": 1.95,
"grad_norm": 4.965468883514404,
"learning_rate": 4.611111111111111e-07,
"loss": 0.5934,
"step": 5440
},
{
"epoch": 1.95,
"grad_norm": 6.403870582580566,
"learning_rate": 4.601010101010101e-07,
"loss": 0.5694,
"step": 5450
},
{
"epoch": 1.96,
"grad_norm": 4.846173286437988,
"learning_rate": 4.5909090909090906e-07,
"loss": 0.5431,
"step": 5460
},
{
"epoch": 1.96,
"grad_norm": 4.517751216888428,
"learning_rate": 4.5808080808080806e-07,
"loss": 0.5966,
"step": 5470
},
{
"epoch": 1.97,
"grad_norm": 4.690954208374023,
"learning_rate": 4.5707070707070705e-07,
"loss": 0.5472,
"step": 5480
},
{
"epoch": 1.97,
"grad_norm": 5.523789882659912,
"learning_rate": 4.5606060606060605e-07,
"loss": 0.5442,
"step": 5490
},
{
"epoch": 1.97,
"grad_norm": 3.194629430770874,
"learning_rate": 4.55050505050505e-07,
"loss": 0.6165,
"step": 5500
},
{
"epoch": 1.98,
"grad_norm": 5.22808313369751,
"learning_rate": 4.54040404040404e-07,
"loss": 0.5074,
"step": 5510
},
{
"epoch": 1.98,
"grad_norm": 4.611573219299316,
"learning_rate": 4.53030303030303e-07,
"loss": 0.4977,
"step": 5520
},
{
"epoch": 1.98,
"grad_norm": 3.746086835861206,
"learning_rate": 4.5202020202020204e-07,
"loss": 0.5699,
"step": 5530
},
{
"epoch": 1.99,
"grad_norm": 4.322105884552002,
"learning_rate": 4.51010101010101e-07,
"loss": 0.5817,
"step": 5540
},
{
"epoch": 1.99,
"grad_norm": 4.089654922485352,
"learning_rate": 4.5e-07,
"loss": 0.5593,
"step": 5550
},
{
"epoch": 1.99,
"grad_norm": 4.9758710861206055,
"learning_rate": 4.48989898989899e-07,
"loss": 0.5716,
"step": 5560
},
{
"epoch": 2.0,
"grad_norm": 4.36923885345459,
"learning_rate": 4.4797979797979797e-07,
"loss": 0.5937,
"step": 5570
},
{
"epoch": 2.0,
"grad_norm": 4.592928886413574,
"learning_rate": 4.469696969696969e-07,
"loss": 0.5471,
"step": 5580
},
{
"epoch": 2.0,
"grad_norm": 9.409171104431152,
"learning_rate": 4.459595959595959e-07,
"loss": 0.7263,
"step": 5590
},
{
"epoch": 2.01,
"grad_norm": 8.28042221069336,
"learning_rate": 4.449494949494949e-07,
"loss": 0.5391,
"step": 5600
},
{
"epoch": 2.01,
"grad_norm": 4.366082191467285,
"learning_rate": 4.4393939393939396e-07,
"loss": 0.5164,
"step": 5610
},
{
"epoch": 2.02,
"grad_norm": 3.808896541595459,
"learning_rate": 4.429292929292929e-07,
"loss": 0.66,
"step": 5620
},
{
"epoch": 2.02,
"grad_norm": 4.30920934677124,
"learning_rate": 4.419191919191919e-07,
"loss": 0.5418,
"step": 5630
},
{
"epoch": 2.02,
"grad_norm": 4.720395088195801,
"learning_rate": 4.409090909090909e-07,
"loss": 0.526,
"step": 5640
},
{
"epoch": 2.03,
"grad_norm": 3.699301242828369,
"learning_rate": 4.398989898989899e-07,
"loss": 0.5309,
"step": 5650
},
{
"epoch": 2.03,
"grad_norm": 4.549612522125244,
"learning_rate": 4.3888888888888884e-07,
"loss": 0.6452,
"step": 5660
},
{
"epoch": 2.03,
"grad_norm": 3.252539873123169,
"learning_rate": 4.3787878787878784e-07,
"loss": 0.5111,
"step": 5670
},
{
"epoch": 2.04,
"grad_norm": 4.932587146759033,
"learning_rate": 4.3686868686868683e-07,
"loss": 0.539,
"step": 5680
},
{
"epoch": 2.04,
"grad_norm": 4.650817394256592,
"learning_rate": 4.3585858585858583e-07,
"loss": 0.6203,
"step": 5690
},
{
"epoch": 2.04,
"grad_norm": 6.277070045471191,
"learning_rate": 4.3484848484848483e-07,
"loss": 0.5435,
"step": 5700
},
{
"epoch": 2.05,
"grad_norm": 7.179144859313965,
"learning_rate": 4.338383838383838e-07,
"loss": 0.5062,
"step": 5710
},
{
"epoch": 2.05,
"grad_norm": 3.681550979614258,
"learning_rate": 4.328282828282828e-07,
"loss": 0.5985,
"step": 5720
},
{
"epoch": 2.05,
"grad_norm": 3.818920850753784,
"learning_rate": 4.318181818181818e-07,
"loss": 0.4808,
"step": 5730
},
{
"epoch": 2.06,
"grad_norm": 3.2395308017730713,
"learning_rate": 4.3080808080808076e-07,
"loss": 0.4888,
"step": 5740
},
{
"epoch": 2.06,
"grad_norm": 4.037728786468506,
"learning_rate": 4.2979797979797976e-07,
"loss": 0.6042,
"step": 5750
},
{
"epoch": 2.07,
"grad_norm": 5.200090408325195,
"learning_rate": 4.2878787878787876e-07,
"loss": 0.5801,
"step": 5760
},
{
"epoch": 2.07,
"grad_norm": 6.219593524932861,
"learning_rate": 4.2777777777777775e-07,
"loss": 0.5453,
"step": 5770
},
{
"epoch": 2.07,
"grad_norm": 5.049505233764648,
"learning_rate": 4.267676767676767e-07,
"loss": 0.5342,
"step": 5780
},
{
"epoch": 2.08,
"grad_norm": 4.03887939453125,
"learning_rate": 4.2575757575757575e-07,
"loss": 0.511,
"step": 5790
},
{
"epoch": 2.08,
"grad_norm": 4.79861307144165,
"learning_rate": 4.2474747474747474e-07,
"loss": 0.5106,
"step": 5800
},
{
"epoch": 2.08,
"grad_norm": 4.6482744216918945,
"learning_rate": 4.2373737373737374e-07,
"loss": 0.5887,
"step": 5810
},
{
"epoch": 2.09,
"grad_norm": 4.516157150268555,
"learning_rate": 4.227272727272727e-07,
"loss": 0.5637,
"step": 5820
},
{
"epoch": 2.09,
"grad_norm": 5.392308712005615,
"learning_rate": 4.217171717171717e-07,
"loss": 0.5485,
"step": 5830
},
{
"epoch": 2.09,
"grad_norm": 3.8693811893463135,
"learning_rate": 4.207070707070707e-07,
"loss": 0.496,
"step": 5840
},
{
"epoch": 2.1,
"grad_norm": 4.755632400512695,
"learning_rate": 4.196969696969697e-07,
"loss": 0.5315,
"step": 5850
},
{
"epoch": 2.1,
"grad_norm": 5.732184410095215,
"learning_rate": 4.186868686868686e-07,
"loss": 0.5608,
"step": 5860
},
{
"epoch": 2.11,
"grad_norm": 4.804543495178223,
"learning_rate": 4.1767676767676767e-07,
"loss": 0.5641,
"step": 5870
},
{
"epoch": 2.11,
"grad_norm": 6.783200263977051,
"learning_rate": 4.1666666666666667e-07,
"loss": 0.4667,
"step": 5880
},
{
"epoch": 2.11,
"grad_norm": 4.994887828826904,
"learning_rate": 4.1565656565656566e-07,
"loss": 0.5759,
"step": 5890
},
{
"epoch": 2.12,
"grad_norm": 5.834484100341797,
"learning_rate": 4.1464646464646466e-07,
"loss": 0.5619,
"step": 5900
},
{
"epoch": 2.12,
"grad_norm": 5.325469970703125,
"learning_rate": 4.136363636363636e-07,
"loss": 0.5464,
"step": 5910
},
{
"epoch": 2.12,
"grad_norm": 4.983416557312012,
"learning_rate": 4.126262626262626e-07,
"loss": 0.5974,
"step": 5920
},
{
"epoch": 2.13,
"grad_norm": 4.775305271148682,
"learning_rate": 4.116161616161616e-07,
"loss": 0.5709,
"step": 5930
},
{
"epoch": 2.13,
"grad_norm": 5.547243595123291,
"learning_rate": 4.1060606060606054e-07,
"loss": 0.5474,
"step": 5940
},
{
"epoch": 2.13,
"grad_norm": 3.5384747982025146,
"learning_rate": 4.0959595959595954e-07,
"loss": 0.4197,
"step": 5950
},
{
"epoch": 2.14,
"grad_norm": 5.288349628448486,
"learning_rate": 4.085858585858586e-07,
"loss": 0.6163,
"step": 5960
},
{
"epoch": 2.14,
"grad_norm": 5.531615734100342,
"learning_rate": 4.075757575757576e-07,
"loss": 0.53,
"step": 5970
},
{
"epoch": 2.14,
"grad_norm": 6.349400997161865,
"learning_rate": 4.065656565656566e-07,
"loss": 0.5457,
"step": 5980
},
{
"epoch": 2.15,
"grad_norm": 5.897380828857422,
"learning_rate": 4.055555555555555e-07,
"loss": 0.5782,
"step": 5990
},
{
"epoch": 2.15,
"grad_norm": 5.457867622375488,
"learning_rate": 4.045454545454545e-07,
"loss": 0.5734,
"step": 6000
},
{
"epoch": 2.15,
"eval_loss": 0.726588785648346,
"eval_runtime": 400.7393,
"eval_samples_per_second": 2.495,
"eval_steps_per_second": 2.495,
"step": 6000
},
{
"epoch": 2.16,
"grad_norm": 5.9234538078308105,
"learning_rate": 4.035353535353535e-07,
"loss": 0.5843,
"step": 6010
},
{
"epoch": 2.16,
"grad_norm": 6.228603363037109,
"learning_rate": 4.025252525252525e-07,
"loss": 0.5035,
"step": 6020
},
{
"epoch": 2.16,
"grad_norm": 4.385622024536133,
"learning_rate": 4.0151515151515146e-07,
"loss": 0.5226,
"step": 6030
},
{
"epoch": 2.17,
"grad_norm": 5.10957670211792,
"learning_rate": 4.005050505050505e-07,
"loss": 0.5418,
"step": 6040
},
{
"epoch": 2.17,
"grad_norm": 3.951185703277588,
"learning_rate": 3.994949494949495e-07,
"loss": 0.5448,
"step": 6050
},
{
"epoch": 2.17,
"grad_norm": 4.83636474609375,
"learning_rate": 3.984848484848485e-07,
"loss": 0.5483,
"step": 6060
},
{
"epoch": 2.18,
"grad_norm": 6.703040599822998,
"learning_rate": 3.9747474747474745e-07,
"loss": 0.5568,
"step": 6070
},
{
"epoch": 2.18,
"grad_norm": 6.855874538421631,
"learning_rate": 3.9646464646464644e-07,
"loss": 0.5365,
"step": 6080
},
{
"epoch": 2.18,
"grad_norm": 4.763612270355225,
"learning_rate": 3.9545454545454544e-07,
"loss": 0.5672,
"step": 6090
},
{
"epoch": 2.19,
"grad_norm": 5.316641330718994,
"learning_rate": 3.9444444444444444e-07,
"loss": 0.5943,
"step": 6100
},
{
"epoch": 2.19,
"grad_norm": 7.093216896057129,
"learning_rate": 3.934343434343434e-07,
"loss": 0.5548,
"step": 6110
},
{
"epoch": 2.19,
"grad_norm": 5.546849250793457,
"learning_rate": 3.924242424242424e-07,
"loss": 0.5377,
"step": 6120
},
{
"epoch": 2.2,
"grad_norm": 6.597376823425293,
"learning_rate": 3.9141414141414143e-07,
"loss": 0.5512,
"step": 6130
},
{
"epoch": 2.2,
"grad_norm": 4.313337326049805,
"learning_rate": 3.904040404040404e-07,
"loss": 0.498,
"step": 6140
},
{
"epoch": 2.21,
"grad_norm": 7.560310363769531,
"learning_rate": 3.8939393939393937e-07,
"loss": 0.5511,
"step": 6150
},
{
"epoch": 2.21,
"grad_norm": 4.503920555114746,
"learning_rate": 3.8838383838383837e-07,
"loss": 0.5497,
"step": 6160
},
{
"epoch": 2.21,
"grad_norm": 5.242457866668701,
"learning_rate": 3.8737373737373736e-07,
"loss": 0.6372,
"step": 6170
},
{
"epoch": 2.22,
"grad_norm": 5.803231716156006,
"learning_rate": 3.8636363636363636e-07,
"loss": 0.5811,
"step": 6180
},
{
"epoch": 2.22,
"grad_norm": 4.9778008460998535,
"learning_rate": 3.853535353535353e-07,
"loss": 0.5717,
"step": 6190
},
{
"epoch": 2.22,
"grad_norm": 3.297873020172119,
"learning_rate": 3.843434343434343e-07,
"loss": 0.5374,
"step": 6200
},
{
"epoch": 2.23,
"grad_norm": 4.822756290435791,
"learning_rate": 3.8333333333333335e-07,
"loss": 0.5788,
"step": 6210
},
{
"epoch": 2.23,
"grad_norm": 6.319406509399414,
"learning_rate": 3.8232323232323235e-07,
"loss": 0.4811,
"step": 6220
},
{
"epoch": 2.23,
"grad_norm": 5.257617950439453,
"learning_rate": 3.813131313131313e-07,
"loss": 0.6003,
"step": 6230
},
{
"epoch": 2.24,
"grad_norm": 5.456783771514893,
"learning_rate": 3.803030303030303e-07,
"loss": 0.4827,
"step": 6240
},
{
"epoch": 2.24,
"grad_norm": 5.607643127441406,
"learning_rate": 3.792929292929293e-07,
"loss": 0.4954,
"step": 6250
},
{
"epoch": 2.24,
"grad_norm": 7.377701282501221,
"learning_rate": 3.782828282828283e-07,
"loss": 0.6513,
"step": 6260
},
{
"epoch": 2.25,
"grad_norm": 5.05380392074585,
"learning_rate": 3.7727272727272723e-07,
"loss": 0.5108,
"step": 6270
},
{
"epoch": 2.25,
"grad_norm": 5.01906681060791,
"learning_rate": 3.762626262626262e-07,
"loss": 0.5575,
"step": 6280
},
{
"epoch": 2.26,
"grad_norm": 4.88652229309082,
"learning_rate": 3.752525252525252e-07,
"loss": 0.5732,
"step": 6290
},
{
"epoch": 2.26,
"grad_norm": 4.311870098114014,
"learning_rate": 3.7424242424242427e-07,
"loss": 0.5814,
"step": 6300
},
{
"epoch": 2.26,
"grad_norm": 5.589522361755371,
"learning_rate": 3.732323232323232e-07,
"loss": 0.5317,
"step": 6310
},
{
"epoch": 2.27,
"grad_norm": 7.32663631439209,
"learning_rate": 3.722222222222222e-07,
"loss": 0.5446,
"step": 6320
},
{
"epoch": 2.27,
"grad_norm": 4.902598857879639,
"learning_rate": 3.712121212121212e-07,
"loss": 0.5386,
"step": 6330
},
{
"epoch": 2.27,
"grad_norm": 5.21954870223999,
"learning_rate": 3.702020202020202e-07,
"loss": 0.4813,
"step": 6340
},
{
"epoch": 2.28,
"grad_norm": 3.8048980236053467,
"learning_rate": 3.6919191919191915e-07,
"loss": 0.5571,
"step": 6350
},
{
"epoch": 2.28,
"grad_norm": 5.75179386138916,
"learning_rate": 3.6818181818181815e-07,
"loss": 0.6023,
"step": 6360
},
{
"epoch": 2.28,
"grad_norm": 6.469977855682373,
"learning_rate": 3.6717171717171714e-07,
"loss": 0.65,
"step": 6370
},
{
"epoch": 2.29,
"grad_norm": 5.1519317626953125,
"learning_rate": 3.661616161616162e-07,
"loss": 0.5634,
"step": 6380
},
{
"epoch": 2.29,
"grad_norm": 4.514432907104492,
"learning_rate": 3.6515151515151514e-07,
"loss": 0.5638,
"step": 6390
},
{
"epoch": 2.3,
"grad_norm": 5.5475029945373535,
"learning_rate": 3.6414141414141413e-07,
"loss": 0.5922,
"step": 6400
},
{
"epoch": 2.3,
"grad_norm": 6.521148204803467,
"learning_rate": 3.6313131313131313e-07,
"loss": 0.5516,
"step": 6410
},
{
"epoch": 2.3,
"grad_norm": 4.127744674682617,
"learning_rate": 3.6212121212121213e-07,
"loss": 0.5455,
"step": 6420
},
{
"epoch": 2.31,
"grad_norm": 5.580835342407227,
"learning_rate": 3.6111111111111107e-07,
"loss": 0.4762,
"step": 6430
},
{
"epoch": 2.31,
"grad_norm": 5.267888069152832,
"learning_rate": 3.6010101010101007e-07,
"loss": 0.6285,
"step": 6440
},
{
"epoch": 2.31,
"grad_norm": 5.960037708282471,
"learning_rate": 3.5909090909090907e-07,
"loss": 0.5553,
"step": 6450
},
{
"epoch": 2.32,
"grad_norm": 3.847745656967163,
"learning_rate": 3.5808080808080806e-07,
"loss": 0.5603,
"step": 6460
},
{
"epoch": 2.32,
"grad_norm": 5.693478107452393,
"learning_rate": 3.5707070707070706e-07,
"loss": 0.6077,
"step": 6470
},
{
"epoch": 2.32,
"grad_norm": 4.48386287689209,
"learning_rate": 3.5606060606060606e-07,
"loss": 0.5636,
"step": 6480
},
{
"epoch": 2.33,
"grad_norm": 6.377192497253418,
"learning_rate": 3.5505050505050505e-07,
"loss": 0.5089,
"step": 6490
},
{
"epoch": 2.33,
"grad_norm": 5.240867614746094,
"learning_rate": 3.5404040404040405e-07,
"loss": 0.5547,
"step": 6500
},
{
"epoch": 2.33,
"grad_norm": 7.864344120025635,
"learning_rate": 3.53030303030303e-07,
"loss": 0.5254,
"step": 6510
},
{
"epoch": 2.34,
"grad_norm": 8.867986679077148,
"learning_rate": 3.52020202020202e-07,
"loss": 0.5616,
"step": 6520
},
{
"epoch": 2.34,
"grad_norm": 4.047301292419434,
"learning_rate": 3.51010101010101e-07,
"loss": 0.5123,
"step": 6530
},
{
"epoch": 2.35,
"grad_norm": 5.919891357421875,
"learning_rate": 3.5e-07,
"loss": 0.5126,
"step": 6540
},
{
"epoch": 2.35,
"grad_norm": 4.536365509033203,
"learning_rate": 3.48989898989899e-07,
"loss": 0.495,
"step": 6550
},
{
"epoch": 2.35,
"grad_norm": 5.329809188842773,
"learning_rate": 3.47979797979798e-07,
"loss": 0.5118,
"step": 6560
},
{
"epoch": 2.36,
"grad_norm": 5.0158867835998535,
"learning_rate": 3.46969696969697e-07,
"loss": 0.5487,
"step": 6570
},
{
"epoch": 2.36,
"grad_norm": 7.028294086456299,
"learning_rate": 3.4595959595959597e-07,
"loss": 0.5893,
"step": 6580
},
{
"epoch": 2.36,
"grad_norm": 4.060366153717041,
"learning_rate": 3.449494949494949e-07,
"loss": 0.5562,
"step": 6590
},
{
"epoch": 2.37,
"grad_norm": 5.36353063583374,
"learning_rate": 3.439393939393939e-07,
"loss": 0.6158,
"step": 6600
},
{
"epoch": 2.37,
"grad_norm": 4.725005626678467,
"learning_rate": 3.429292929292929e-07,
"loss": 0.4339,
"step": 6610
},
{
"epoch": 2.37,
"grad_norm": 4.875920295715332,
"learning_rate": 3.419191919191919e-07,
"loss": 0.5047,
"step": 6620
},
{
"epoch": 2.38,
"grad_norm": 6.4596266746521,
"learning_rate": 3.4090909090909085e-07,
"loss": 0.5516,
"step": 6630
},
{
"epoch": 2.38,
"grad_norm": 6.09486722946167,
"learning_rate": 3.398989898989899e-07,
"loss": 0.5863,
"step": 6640
},
{
"epoch": 2.38,
"grad_norm": 4.232431411743164,
"learning_rate": 3.388888888888889e-07,
"loss": 0.5221,
"step": 6650
},
{
"epoch": 2.39,
"grad_norm": 4.1522111892700195,
"learning_rate": 3.378787878787879e-07,
"loss": 0.588,
"step": 6660
},
{
"epoch": 2.39,
"grad_norm": 5.693157196044922,
"learning_rate": 3.3686868686868684e-07,
"loss": 0.5966,
"step": 6670
},
{
"epoch": 2.4,
"grad_norm": 3.900887966156006,
"learning_rate": 3.3585858585858583e-07,
"loss": 0.437,
"step": 6680
},
{
"epoch": 2.4,
"grad_norm": 5.744729995727539,
"learning_rate": 3.3484848484848483e-07,
"loss": 0.5483,
"step": 6690
},
{
"epoch": 2.4,
"grad_norm": 4.26437520980835,
"learning_rate": 3.3383838383838383e-07,
"loss": 0.5159,
"step": 6700
},
{
"epoch": 2.41,
"grad_norm": 5.422147750854492,
"learning_rate": 3.3282828282828277e-07,
"loss": 0.4986,
"step": 6710
},
{
"epoch": 2.41,
"grad_norm": 4.445804119110107,
"learning_rate": 3.318181818181818e-07,
"loss": 0.544,
"step": 6720
},
{
"epoch": 2.41,
"grad_norm": 5.514401435852051,
"learning_rate": 3.308080808080808e-07,
"loss": 0.5422,
"step": 6730
},
{
"epoch": 2.42,
"grad_norm": 7.159079074859619,
"learning_rate": 3.297979797979798e-07,
"loss": 0.503,
"step": 6740
},
{
"epoch": 2.42,
"grad_norm": 6.815921783447266,
"learning_rate": 3.2878787878787876e-07,
"loss": 0.4779,
"step": 6750
},
{
"epoch": 2.42,
"grad_norm": 4.787651538848877,
"learning_rate": 3.2777777777777776e-07,
"loss": 0.484,
"step": 6760
},
{
"epoch": 2.43,
"grad_norm": 5.437996864318848,
"learning_rate": 3.2676767676767675e-07,
"loss": 0.5689,
"step": 6770
},
{
"epoch": 2.43,
"grad_norm": 2.6372170448303223,
"learning_rate": 3.2575757575757575e-07,
"loss": 0.6268,
"step": 6780
},
{
"epoch": 2.44,
"grad_norm": 6.545858383178711,
"learning_rate": 3.247474747474747e-07,
"loss": 0.603,
"step": 6790
},
{
"epoch": 2.44,
"grad_norm": 3.9520442485809326,
"learning_rate": 3.237373737373737e-07,
"loss": 0.6243,
"step": 6800
},
{
"epoch": 2.44,
"grad_norm": 4.638275623321533,
"learning_rate": 3.2272727272727274e-07,
"loss": 0.5433,
"step": 6810
},
{
"epoch": 2.45,
"grad_norm": 4.08539342880249,
"learning_rate": 3.2171717171717174e-07,
"loss": 0.576,
"step": 6820
},
{
"epoch": 2.45,
"grad_norm": 6.024731636047363,
"learning_rate": 3.207070707070707e-07,
"loss": 0.61,
"step": 6830
},
{
"epoch": 2.45,
"grad_norm": 3.649667263031006,
"learning_rate": 3.196969696969697e-07,
"loss": 0.4797,
"step": 6840
},
{
"epoch": 2.46,
"grad_norm": 7.055994987487793,
"learning_rate": 3.186868686868687e-07,
"loss": 0.5525,
"step": 6850
},
{
"epoch": 2.46,
"grad_norm": 4.7312116622924805,
"learning_rate": 3.1767676767676767e-07,
"loss": 0.5069,
"step": 6860
},
{
"epoch": 2.46,
"grad_norm": 4.483476638793945,
"learning_rate": 3.166666666666666e-07,
"loss": 0.4884,
"step": 6870
},
{
"epoch": 2.47,
"grad_norm": 7.678333282470703,
"learning_rate": 3.156565656565656e-07,
"loss": 0.4818,
"step": 6880
},
{
"epoch": 2.47,
"grad_norm": 4.442699909210205,
"learning_rate": 3.1464646464646466e-07,
"loss": 0.5771,
"step": 6890
},
{
"epoch": 2.47,
"grad_norm": 4.732019424438477,
"learning_rate": 3.1363636363636366e-07,
"loss": 0.5294,
"step": 6900
},
{
"epoch": 2.48,
"grad_norm": 5.647908687591553,
"learning_rate": 3.126262626262626e-07,
"loss": 0.5496,
"step": 6910
},
{
"epoch": 2.48,
"grad_norm": 4.71852970123291,
"learning_rate": 3.116161616161616e-07,
"loss": 0.503,
"step": 6920
},
{
"epoch": 2.49,
"grad_norm": 8.519858360290527,
"learning_rate": 3.106060606060606e-07,
"loss": 0.547,
"step": 6930
},
{
"epoch": 2.49,
"grad_norm": 4.513672351837158,
"learning_rate": 3.095959595959596e-07,
"loss": 0.5496,
"step": 6940
},
{
"epoch": 2.49,
"grad_norm": 4.779497146606445,
"learning_rate": 3.0858585858585854e-07,
"loss": 0.5193,
"step": 6950
},
{
"epoch": 2.5,
"grad_norm": 5.14954948425293,
"learning_rate": 3.0757575757575754e-07,
"loss": 0.5738,
"step": 6960
},
{
"epoch": 2.5,
"grad_norm": 4.722686290740967,
"learning_rate": 3.0656565656565653e-07,
"loss": 0.5978,
"step": 6970
},
{
"epoch": 2.5,
"grad_norm": 4.7376837730407715,
"learning_rate": 3.055555555555556e-07,
"loss": 0.5476,
"step": 6980
},
{
"epoch": 2.51,
"grad_norm": 5.883960247039795,
"learning_rate": 3.0454545454545453e-07,
"loss": 0.4953,
"step": 6990
},
{
"epoch": 2.51,
"grad_norm": 5.1609721183776855,
"learning_rate": 3.035353535353535e-07,
"loss": 0.5602,
"step": 7000
},
{
"epoch": 2.51,
"eval_loss": 0.7043054103851318,
"eval_runtime": 404.5936,
"eval_samples_per_second": 2.472,
"eval_steps_per_second": 2.472,
"step": 7000
},
{
"epoch": 2.51,
"grad_norm": 4.467599391937256,
"learning_rate": 3.025252525252525e-07,
"loss": 0.5545,
"step": 7010
},
{
"epoch": 2.52,
"grad_norm": 4.831850528717041,
"learning_rate": 3.015151515151515e-07,
"loss": 0.5765,
"step": 7020
},
{
"epoch": 2.52,
"grad_norm": 4.63700008392334,
"learning_rate": 3.0050505050505046e-07,
"loss": 0.5354,
"step": 7030
},
{
"epoch": 2.52,
"grad_norm": 4.822059631347656,
"learning_rate": 2.9949494949494946e-07,
"loss": 0.5239,
"step": 7040
},
{
"epoch": 2.53,
"grad_norm": 4.098064422607422,
"learning_rate": 2.9848484848484846e-07,
"loss": 0.5775,
"step": 7050
},
{
"epoch": 2.53,
"grad_norm": 5.17026948928833,
"learning_rate": 2.9747474747474745e-07,
"loss": 0.6014,
"step": 7060
},
{
"epoch": 2.54,
"grad_norm": 4.5090789794921875,
"learning_rate": 2.9646464646464645e-07,
"loss": 0.5732,
"step": 7070
},
{
"epoch": 2.54,
"grad_norm": 6.120937347412109,
"learning_rate": 2.9545454545454545e-07,
"loss": 0.4529,
"step": 7080
},
{
"epoch": 2.54,
"grad_norm": 5.249495029449463,
"learning_rate": 2.9444444444444444e-07,
"loss": 0.5804,
"step": 7090
},
{
"epoch": 2.55,
"grad_norm": 7.617781639099121,
"learning_rate": 2.9343434343434344e-07,
"loss": 0.5382,
"step": 7100
},
{
"epoch": 2.55,
"grad_norm": 4.722599506378174,
"learning_rate": 2.924242424242424e-07,
"loss": 0.5419,
"step": 7110
},
{
"epoch": 2.55,
"grad_norm": 4.24169921875,
"learning_rate": 2.914141414141414e-07,
"loss": 0.5464,
"step": 7120
},
{
"epoch": 2.56,
"grad_norm": 4.781604766845703,
"learning_rate": 2.904040404040404e-07,
"loss": 0.4158,
"step": 7130
},
{
"epoch": 2.56,
"grad_norm": 5.138591766357422,
"learning_rate": 2.893939393939394e-07,
"loss": 0.5626,
"step": 7140
},
{
"epoch": 2.56,
"grad_norm": 4.901473045349121,
"learning_rate": 2.8838383838383837e-07,
"loss": 0.5164,
"step": 7150
},
{
"epoch": 2.57,
"grad_norm": 3.736123561859131,
"learning_rate": 2.8737373737373737e-07,
"loss": 0.5159,
"step": 7160
},
{
"epoch": 2.57,
"grad_norm": 3.937110662460327,
"learning_rate": 2.8636363636363637e-07,
"loss": 0.5785,
"step": 7170
},
{
"epoch": 2.57,
"grad_norm": 5.2640581130981445,
"learning_rate": 2.8535353535353536e-07,
"loss": 0.5839,
"step": 7180
},
{
"epoch": 2.58,
"grad_norm": 5.652442455291748,
"learning_rate": 2.843434343434343e-07,
"loss": 0.533,
"step": 7190
},
{
"epoch": 2.58,
"grad_norm": 4.787551403045654,
"learning_rate": 2.833333333333333e-07,
"loss": 0.5114,
"step": 7200
},
{
"epoch": 2.59,
"grad_norm": 5.234408855438232,
"learning_rate": 2.823232323232323e-07,
"loss": 0.5777,
"step": 7210
},
{
"epoch": 2.59,
"grad_norm": 2.3309760093688965,
"learning_rate": 2.813131313131313e-07,
"loss": 0.5085,
"step": 7220
},
{
"epoch": 2.59,
"grad_norm": 4.90077543258667,
"learning_rate": 2.8030303030303024e-07,
"loss": 0.5464,
"step": 7230
},
{
"epoch": 2.6,
"grad_norm": 6.651805400848389,
"learning_rate": 2.792929292929293e-07,
"loss": 0.5143,
"step": 7240
},
{
"epoch": 2.6,
"grad_norm": 4.081473350524902,
"learning_rate": 2.782828282828283e-07,
"loss": 0.5307,
"step": 7250
},
{
"epoch": 2.6,
"grad_norm": 8.496968269348145,
"learning_rate": 2.772727272727273e-07,
"loss": 0.5376,
"step": 7260
},
{
"epoch": 2.61,
"grad_norm": 4.46324348449707,
"learning_rate": 2.7626262626262623e-07,
"loss": 0.5871,
"step": 7270
},
{
"epoch": 2.61,
"grad_norm": 5.344108581542969,
"learning_rate": 2.752525252525252e-07,
"loss": 0.4783,
"step": 7280
},
{
"epoch": 2.61,
"grad_norm": 6.3261613845825195,
"learning_rate": 2.742424242424242e-07,
"loss": 0.4924,
"step": 7290
},
{
"epoch": 2.62,
"grad_norm": 4.785797119140625,
"learning_rate": 2.732323232323232e-07,
"loss": 0.5087,
"step": 7300
},
{
"epoch": 2.62,
"grad_norm": 5.484802722930908,
"learning_rate": 2.7222222222222216e-07,
"loss": 0.5763,
"step": 7310
},
{
"epoch": 2.63,
"grad_norm": 5.250958442687988,
"learning_rate": 2.712121212121212e-07,
"loss": 0.5065,
"step": 7320
},
{
"epoch": 2.63,
"grad_norm": 3.2717227935791016,
"learning_rate": 2.702020202020202e-07,
"loss": 0.4802,
"step": 7330
},
{
"epoch": 2.63,
"grad_norm": 5.094733238220215,
"learning_rate": 2.691919191919192e-07,
"loss": 0.5162,
"step": 7340
},
{
"epoch": 2.64,
"grad_norm": 5.360857009887695,
"learning_rate": 2.6818181818181815e-07,
"loss": 0.5947,
"step": 7350
},
{
"epoch": 2.64,
"grad_norm": 5.611083030700684,
"learning_rate": 2.6717171717171715e-07,
"loss": 0.6174,
"step": 7360
},
{
"epoch": 2.64,
"grad_norm": 4.42379093170166,
"learning_rate": 2.6616161616161614e-07,
"loss": 0.5773,
"step": 7370
},
{
"epoch": 2.65,
"grad_norm": 6.272792816162109,
"learning_rate": 2.6515151515151514e-07,
"loss": 0.6246,
"step": 7380
},
{
"epoch": 2.65,
"grad_norm": 5.295745849609375,
"learning_rate": 2.641414141414141e-07,
"loss": 0.5315,
"step": 7390
},
{
"epoch": 2.65,
"grad_norm": 4.078943252563477,
"learning_rate": 2.631313131313131e-07,
"loss": 0.581,
"step": 7400
},
{
"epoch": 2.66,
"grad_norm": 12.75452709197998,
"learning_rate": 2.6212121212121213e-07,
"loss": 0.5608,
"step": 7410
},
{
"epoch": 2.66,
"grad_norm": 5.873865604400635,
"learning_rate": 2.6111111111111113e-07,
"loss": 0.5588,
"step": 7420
},
{
"epoch": 2.66,
"grad_norm": 4.560327529907227,
"learning_rate": 2.6010101010101007e-07,
"loss": 0.6557,
"step": 7430
},
{
"epoch": 2.67,
"grad_norm": 7.971694469451904,
"learning_rate": 2.5909090909090907e-07,
"loss": 0.5636,
"step": 7440
},
{
"epoch": 2.67,
"grad_norm": 4.814733982086182,
"learning_rate": 2.5808080808080807e-07,
"loss": 0.5851,
"step": 7450
},
{
"epoch": 2.68,
"grad_norm": 3.635782241821289,
"learning_rate": 2.5707070707070706e-07,
"loss": 0.5354,
"step": 7460
},
{
"epoch": 2.68,
"grad_norm": 6.356813907623291,
"learning_rate": 2.56060606060606e-07,
"loss": 0.5934,
"step": 7470
},
{
"epoch": 2.68,
"grad_norm": 6.781961441040039,
"learning_rate": 2.55050505050505e-07,
"loss": 0.5249,
"step": 7480
},
{
"epoch": 2.69,
"grad_norm": 3.655015230178833,
"learning_rate": 2.5404040404040405e-07,
"loss": 0.5198,
"step": 7490
},
{
"epoch": 2.69,
"grad_norm": 6.139344692230225,
"learning_rate": 2.5303030303030305e-07,
"loss": 0.5375,
"step": 7500
},
{
"epoch": 2.69,
"grad_norm": 4.679745197296143,
"learning_rate": 2.5202020202020205e-07,
"loss": 0.5107,
"step": 7510
},
{
"epoch": 2.7,
"grad_norm": 5.100747585296631,
"learning_rate": 2.51010101010101e-07,
"loss": 0.515,
"step": 7520
},
{
"epoch": 2.7,
"grad_norm": 3.9256341457366943,
"learning_rate": 2.5e-07,
"loss": 0.5409,
"step": 7530
},
{
"epoch": 2.7,
"grad_norm": 4.619549751281738,
"learning_rate": 2.48989898989899e-07,
"loss": 0.5024,
"step": 7540
},
{
"epoch": 2.71,
"grad_norm": 4.833261489868164,
"learning_rate": 2.4797979797979793e-07,
"loss": 0.5287,
"step": 7550
},
{
"epoch": 2.71,
"grad_norm": 3.5149707794189453,
"learning_rate": 2.46969696969697e-07,
"loss": 0.5176,
"step": 7560
},
{
"epoch": 2.71,
"grad_norm": 6.924714088439941,
"learning_rate": 2.45959595959596e-07,
"loss": 0.5598,
"step": 7570
},
{
"epoch": 2.72,
"grad_norm": 7.813880443572998,
"learning_rate": 2.449494949494949e-07,
"loss": 0.535,
"step": 7580
},
{
"epoch": 2.72,
"grad_norm": 5.943690299987793,
"learning_rate": 2.439393939393939e-07,
"loss": 0.5327,
"step": 7590
},
{
"epoch": 2.73,
"grad_norm": 5.665824890136719,
"learning_rate": 2.429292929292929e-07,
"loss": 0.501,
"step": 7600
},
{
"epoch": 2.73,
"grad_norm": 3.8446450233459473,
"learning_rate": 2.419191919191919e-07,
"loss": 0.5307,
"step": 7610
},
{
"epoch": 2.73,
"grad_norm": 4.556608200073242,
"learning_rate": 2.409090909090909e-07,
"loss": 0.5244,
"step": 7620
},
{
"epoch": 2.74,
"grad_norm": 4.267519950866699,
"learning_rate": 2.398989898989899e-07,
"loss": 0.512,
"step": 7630
},
{
"epoch": 2.74,
"grad_norm": 6.164044380187988,
"learning_rate": 2.388888888888889e-07,
"loss": 0.4957,
"step": 7640
},
{
"epoch": 2.74,
"grad_norm": 5.647671222686768,
"learning_rate": 2.3787878787878787e-07,
"loss": 0.6485,
"step": 7650
},
{
"epoch": 2.75,
"grad_norm": 3.6924054622650146,
"learning_rate": 2.3686868686868687e-07,
"loss": 0.4993,
"step": 7660
},
{
"epoch": 2.75,
"grad_norm": 5.332742691040039,
"learning_rate": 2.3585858585858584e-07,
"loss": 0.5619,
"step": 7670
},
{
"epoch": 2.75,
"grad_norm": 4.201364517211914,
"learning_rate": 2.3484848484848486e-07,
"loss": 0.5011,
"step": 7680
},
{
"epoch": 2.76,
"grad_norm": 4.477077484130859,
"learning_rate": 2.3383838383838383e-07,
"loss": 0.5369,
"step": 7690
},
{
"epoch": 2.76,
"grad_norm": 4.704535961151123,
"learning_rate": 2.3282828282828283e-07,
"loss": 0.5495,
"step": 7700
},
{
"epoch": 2.76,
"grad_norm": 3.2071988582611084,
"learning_rate": 2.318181818181818e-07,
"loss": 0.5478,
"step": 7710
},
{
"epoch": 2.77,
"grad_norm": 4.613890647888184,
"learning_rate": 2.308080808080808e-07,
"loss": 0.5208,
"step": 7720
},
{
"epoch": 2.77,
"grad_norm": 5.759297847747803,
"learning_rate": 2.297979797979798e-07,
"loss": 0.4989,
"step": 7730
},
{
"epoch": 2.78,
"grad_norm": 3.805091619491577,
"learning_rate": 2.287878787878788e-07,
"loss": 0.521,
"step": 7740
},
{
"epoch": 2.78,
"grad_norm": 5.901796817779541,
"learning_rate": 2.2777777777777776e-07,
"loss": 0.603,
"step": 7750
},
{
"epoch": 2.78,
"grad_norm": 4.832998752593994,
"learning_rate": 2.2676767676767676e-07,
"loss": 0.5698,
"step": 7760
},
{
"epoch": 2.79,
"grad_norm": 4.8905439376831055,
"learning_rate": 2.2575757575757576e-07,
"loss": 0.4658,
"step": 7770
},
{
"epoch": 2.79,
"grad_norm": 7.449604511260986,
"learning_rate": 2.2474747474747475e-07,
"loss": 0.5889,
"step": 7780
},
{
"epoch": 2.79,
"grad_norm": 8.905882835388184,
"learning_rate": 2.2373737373737372e-07,
"loss": 0.5323,
"step": 7790
},
{
"epoch": 2.8,
"grad_norm": 4.761772632598877,
"learning_rate": 2.2272727272727272e-07,
"loss": 0.53,
"step": 7800
},
{
"epoch": 2.8,
"grad_norm": 4.811107158660889,
"learning_rate": 2.2171717171717172e-07,
"loss": 0.5189,
"step": 7810
},
{
"epoch": 2.8,
"grad_norm": 2.9238836765289307,
"learning_rate": 2.2070707070707071e-07,
"loss": 0.5283,
"step": 7820
},
{
"epoch": 2.81,
"grad_norm": 5.508478164672852,
"learning_rate": 2.1969696969696968e-07,
"loss": 0.4561,
"step": 7830
},
{
"epoch": 2.81,
"grad_norm": 5.604992866516113,
"learning_rate": 2.1868686868686868e-07,
"loss": 0.5718,
"step": 7840
},
{
"epoch": 2.82,
"grad_norm": 6.870204925537109,
"learning_rate": 2.1767676767676765e-07,
"loss": 0.4882,
"step": 7850
},
{
"epoch": 2.82,
"grad_norm": 4.332894802093506,
"learning_rate": 2.1666666666666667e-07,
"loss": 0.5755,
"step": 7860
},
{
"epoch": 2.82,
"grad_norm": 5.616927623748779,
"learning_rate": 2.1565656565656564e-07,
"loss": 0.5805,
"step": 7870
},
{
"epoch": 2.83,
"grad_norm": 4.477654933929443,
"learning_rate": 2.1464646464646464e-07,
"loss": 0.582,
"step": 7880
},
{
"epoch": 2.83,
"grad_norm": 6.561676025390625,
"learning_rate": 2.136363636363636e-07,
"loss": 0.4924,
"step": 7890
},
{
"epoch": 2.83,
"grad_norm": 5.206873416900635,
"learning_rate": 2.1262626262626264e-07,
"loss": 0.5878,
"step": 7900
},
{
"epoch": 2.84,
"grad_norm": 4.1171345710754395,
"learning_rate": 2.116161616161616e-07,
"loss": 0.5451,
"step": 7910
},
{
"epoch": 2.84,
"grad_norm": 3.9070441722869873,
"learning_rate": 2.106060606060606e-07,
"loss": 0.5403,
"step": 7920
},
{
"epoch": 2.84,
"grad_norm": 4.80625057220459,
"learning_rate": 2.0959595959595957e-07,
"loss": 0.5552,
"step": 7930
},
{
"epoch": 2.85,
"grad_norm": 4.501746654510498,
"learning_rate": 2.085858585858586e-07,
"loss": 0.4697,
"step": 7940
},
{
"epoch": 2.85,
"grad_norm": 3.92985463142395,
"learning_rate": 2.0757575757575757e-07,
"loss": 0.5565,
"step": 7950
},
{
"epoch": 2.85,
"grad_norm": 4.4594807624816895,
"learning_rate": 2.0656565656565656e-07,
"loss": 0.561,
"step": 7960
},
{
"epoch": 2.86,
"grad_norm": 5.731437683105469,
"learning_rate": 2.0555555555555553e-07,
"loss": 0.5311,
"step": 7970
},
{
"epoch": 2.86,
"grad_norm": 4.295395374298096,
"learning_rate": 2.0454545454545456e-07,
"loss": 0.6082,
"step": 7980
},
{
"epoch": 2.87,
"grad_norm": 4.862415313720703,
"learning_rate": 2.0353535353535353e-07,
"loss": 0.6022,
"step": 7990
},
{
"epoch": 2.87,
"grad_norm": 4.604592323303223,
"learning_rate": 2.0252525252525253e-07,
"loss": 0.5314,
"step": 8000
},
{
"epoch": 2.87,
"eval_loss": 0.7028300762176514,
"eval_runtime": 404.8169,
"eval_samples_per_second": 2.47,
"eval_steps_per_second": 2.47,
"step": 8000
},
{
"epoch": 2.87,
"grad_norm": 4.165442943572998,
"learning_rate": 2.015151515151515e-07,
"loss": 0.5624,
"step": 8010
},
{
"epoch": 2.88,
"grad_norm": 4.097321510314941,
"learning_rate": 2.005050505050505e-07,
"loss": 0.5041,
"step": 8020
},
{
"epoch": 2.88,
"grad_norm": 4.476957321166992,
"learning_rate": 1.994949494949495e-07,
"loss": 0.4556,
"step": 8030
},
{
"epoch": 2.88,
"grad_norm": 4.502745628356934,
"learning_rate": 1.9848484848484849e-07,
"loss": 0.535,
"step": 8040
},
{
"epoch": 2.89,
"grad_norm": 5.133005142211914,
"learning_rate": 1.9747474747474746e-07,
"loss": 0.4998,
"step": 8050
},
{
"epoch": 2.89,
"grad_norm": 5.7199296951293945,
"learning_rate": 1.9646464646464645e-07,
"loss": 0.6113,
"step": 8060
},
{
"epoch": 2.89,
"grad_norm": 4.979962348937988,
"learning_rate": 1.9545454545454545e-07,
"loss": 0.6419,
"step": 8070
},
{
"epoch": 2.9,
"grad_norm": 5.971510410308838,
"learning_rate": 1.9444444444444445e-07,
"loss": 0.5269,
"step": 8080
},
{
"epoch": 2.9,
"grad_norm": 6.548705577850342,
"learning_rate": 1.9343434343434342e-07,
"loss": 0.456,
"step": 8090
},
{
"epoch": 2.9,
"grad_norm": 5.575558185577393,
"learning_rate": 1.9242424242424241e-07,
"loss": 0.5068,
"step": 8100
},
{
"epoch": 2.91,
"grad_norm": 6.25363826751709,
"learning_rate": 1.914141414141414e-07,
"loss": 0.4749,
"step": 8110
},
{
"epoch": 2.91,
"grad_norm": 5.45524787902832,
"learning_rate": 1.904040404040404e-07,
"loss": 0.5556,
"step": 8120
},
{
"epoch": 2.92,
"grad_norm": 5.0804548263549805,
"learning_rate": 1.8939393939393938e-07,
"loss": 0.5058,
"step": 8130
},
{
"epoch": 2.92,
"grad_norm": 6.4303460121154785,
"learning_rate": 1.8838383838383838e-07,
"loss": 0.553,
"step": 8140
},
{
"epoch": 2.92,
"grad_norm": 5.177044868469238,
"learning_rate": 1.8737373737373737e-07,
"loss": 0.5986,
"step": 8150
},
{
"epoch": 2.93,
"grad_norm": 4.609193325042725,
"learning_rate": 1.8636363636363637e-07,
"loss": 0.5456,
"step": 8160
},
{
"epoch": 2.93,
"grad_norm": 5.29266357421875,
"learning_rate": 1.8535353535353534e-07,
"loss": 0.5806,
"step": 8170
},
{
"epoch": 2.93,
"grad_norm": 6.320983409881592,
"learning_rate": 1.8434343434343434e-07,
"loss": 0.5621,
"step": 8180
},
{
"epoch": 2.94,
"grad_norm": 5.851290702819824,
"learning_rate": 1.833333333333333e-07,
"loss": 0.5907,
"step": 8190
},
{
"epoch": 2.94,
"grad_norm": 7.4724555015563965,
"learning_rate": 1.8232323232323233e-07,
"loss": 0.5415,
"step": 8200
},
{
"epoch": 2.94,
"grad_norm": 5.282331943511963,
"learning_rate": 1.813131313131313e-07,
"loss": 0.4809,
"step": 8210
},
{
"epoch": 2.95,
"grad_norm": 5.070183277130127,
"learning_rate": 1.803030303030303e-07,
"loss": 0.5138,
"step": 8220
},
{
"epoch": 2.95,
"grad_norm": 4.092304229736328,
"learning_rate": 1.7929292929292927e-07,
"loss": 0.4772,
"step": 8230
},
{
"epoch": 2.95,
"grad_norm": 5.767254829406738,
"learning_rate": 1.782828282828283e-07,
"loss": 0.5756,
"step": 8240
},
{
"epoch": 2.96,
"grad_norm": 5.086340427398682,
"learning_rate": 1.7727272727272726e-07,
"loss": 0.6313,
"step": 8250
},
{
"epoch": 2.96,
"grad_norm": 7.967188835144043,
"learning_rate": 1.7626262626262626e-07,
"loss": 0.5307,
"step": 8260
},
{
"epoch": 2.97,
"grad_norm": 4.956604480743408,
"learning_rate": 1.7525252525252523e-07,
"loss": 0.4744,
"step": 8270
},
{
"epoch": 2.97,
"grad_norm": 5.134040355682373,
"learning_rate": 1.7424242424242425e-07,
"loss": 0.5217,
"step": 8280
},
{
"epoch": 2.97,
"grad_norm": 4.910244464874268,
"learning_rate": 1.7323232323232322e-07,
"loss": 0.513,
"step": 8290
},
{
"epoch": 2.98,
"grad_norm": 6.457435131072998,
"learning_rate": 1.7222222222222222e-07,
"loss": 0.5786,
"step": 8300
},
{
"epoch": 2.98,
"grad_norm": 5.450909614562988,
"learning_rate": 1.712121212121212e-07,
"loss": 0.4931,
"step": 8310
},
{
"epoch": 2.98,
"grad_norm": 5.564160346984863,
"learning_rate": 1.7020202020202021e-07,
"loss": 0.5263,
"step": 8320
},
{
"epoch": 2.99,
"grad_norm": 6.898198127746582,
"learning_rate": 1.6919191919191918e-07,
"loss": 0.5814,
"step": 8330
},
{
"epoch": 2.99,
"grad_norm": 5.737578392028809,
"learning_rate": 1.6818181818181818e-07,
"loss": 0.498,
"step": 8340
},
{
"epoch": 2.99,
"grad_norm": 4.877408027648926,
"learning_rate": 1.6717171717171715e-07,
"loss": 0.5743,
"step": 8350
},
{
"epoch": 3.0,
"grad_norm": 5.765985012054443,
"learning_rate": 1.6616161616161615e-07,
"loss": 0.5476,
"step": 8360
},
{
"epoch": 3.0,
"grad_norm": 12.480027198791504,
"learning_rate": 1.6515151515151515e-07,
"loss": 0.5693,
"step": 8370
},
{
"epoch": 3.01,
"grad_norm": 8.612441062927246,
"learning_rate": 1.6414141414141414e-07,
"loss": 0.6035,
"step": 8380
},
{
"epoch": 3.01,
"grad_norm": 5.978526592254639,
"learning_rate": 1.631313131313131e-07,
"loss": 0.5711,
"step": 8390
},
{
"epoch": 3.01,
"grad_norm": 6.2041192054748535,
"learning_rate": 1.621212121212121e-07,
"loss": 0.4838,
"step": 8400
},
{
"epoch": 3.02,
"grad_norm": 3.6861565113067627,
"learning_rate": 1.611111111111111e-07,
"loss": 0.6104,
"step": 8410
},
{
"epoch": 3.02,
"grad_norm": 5.080516338348389,
"learning_rate": 1.601010101010101e-07,
"loss": 0.4688,
"step": 8420
},
{
"epoch": 3.02,
"grad_norm": 5.527194976806641,
"learning_rate": 1.5909090909090907e-07,
"loss": 0.5746,
"step": 8430
},
{
"epoch": 3.03,
"grad_norm": 4.398497104644775,
"learning_rate": 1.5808080808080807e-07,
"loss": 0.4895,
"step": 8440
},
{
"epoch": 3.03,
"grad_norm": 5.004568099975586,
"learning_rate": 1.5707070707070707e-07,
"loss": 0.5808,
"step": 8450
},
{
"epoch": 3.03,
"grad_norm": 6.635837554931641,
"learning_rate": 1.5606060606060606e-07,
"loss": 0.5519,
"step": 8460
},
{
"epoch": 3.04,
"grad_norm": 4.912801742553711,
"learning_rate": 1.5505050505050504e-07,
"loss": 0.5873,
"step": 8470
},
{
"epoch": 3.04,
"grad_norm": 5.812663555145264,
"learning_rate": 1.5404040404040403e-07,
"loss": 0.5286,
"step": 8480
},
{
"epoch": 3.04,
"grad_norm": 5.986713409423828,
"learning_rate": 1.53030303030303e-07,
"loss": 0.5305,
"step": 8490
},
{
"epoch": 3.05,
"grad_norm": 5.908451557159424,
"learning_rate": 1.5202020202020203e-07,
"loss": 0.6125,
"step": 8500
},
{
"epoch": 3.05,
"grad_norm": 8.727606773376465,
"learning_rate": 1.51010101010101e-07,
"loss": 0.5765,
"step": 8510
},
{
"epoch": 3.06,
"grad_norm": 5.8582987785339355,
"learning_rate": 1.5e-07,
"loss": 0.5178,
"step": 8520
},
{
"epoch": 3.06,
"grad_norm": 5.690489292144775,
"learning_rate": 1.4898989898989896e-07,
"loss": 0.4978,
"step": 8530
},
{
"epoch": 3.06,
"grad_norm": 5.1287384033203125,
"learning_rate": 1.47979797979798e-07,
"loss": 0.5135,
"step": 8540
},
{
"epoch": 3.07,
"grad_norm": 5.430093765258789,
"learning_rate": 1.4696969696969696e-07,
"loss": 0.4935,
"step": 8550
},
{
"epoch": 3.07,
"grad_norm": 5.307154655456543,
"learning_rate": 1.4595959595959595e-07,
"loss": 0.4839,
"step": 8560
},
{
"epoch": 3.07,
"grad_norm": 4.900834560394287,
"learning_rate": 1.4494949494949492e-07,
"loss": 0.5637,
"step": 8570
},
{
"epoch": 3.08,
"grad_norm": 4.796194553375244,
"learning_rate": 1.4393939393939395e-07,
"loss": 0.519,
"step": 8580
},
{
"epoch": 3.08,
"grad_norm": 6.488938331604004,
"learning_rate": 1.4292929292929292e-07,
"loss": 0.6017,
"step": 8590
},
{
"epoch": 3.08,
"grad_norm": 4.90757417678833,
"learning_rate": 1.4191919191919192e-07,
"loss": 0.4873,
"step": 8600
},
{
"epoch": 3.09,
"grad_norm": 6.44563627243042,
"learning_rate": 1.4090909090909089e-07,
"loss": 0.5919,
"step": 8610
},
{
"epoch": 3.09,
"grad_norm": 3.680788516998291,
"learning_rate": 1.398989898989899e-07,
"loss": 0.5062,
"step": 8620
},
{
"epoch": 3.09,
"grad_norm": 4.880250930786133,
"learning_rate": 1.3888888888888888e-07,
"loss": 0.5751,
"step": 8630
},
{
"epoch": 3.1,
"grad_norm": 4.91787052154541,
"learning_rate": 1.3787878787878788e-07,
"loss": 0.5604,
"step": 8640
},
{
"epoch": 3.1,
"grad_norm": 4.990437984466553,
"learning_rate": 1.3686868686868685e-07,
"loss": 0.4807,
"step": 8650
},
{
"epoch": 3.11,
"grad_norm": 4.082001209259033,
"learning_rate": 1.3585858585858584e-07,
"loss": 0.4771,
"step": 8660
},
{
"epoch": 3.11,
"grad_norm": 3.742047071456909,
"learning_rate": 1.3484848484848484e-07,
"loss": 0.5773,
"step": 8670
},
{
"epoch": 3.11,
"grad_norm": 6.419268608093262,
"learning_rate": 1.3383838383838384e-07,
"loss": 0.5903,
"step": 8680
},
{
"epoch": 3.12,
"grad_norm": 5.8881940841674805,
"learning_rate": 1.328282828282828e-07,
"loss": 0.5279,
"step": 8690
},
{
"epoch": 3.12,
"grad_norm": 4.6417999267578125,
"learning_rate": 1.318181818181818e-07,
"loss": 0.5368,
"step": 8700
},
{
"epoch": 3.12,
"grad_norm": 5.411186218261719,
"learning_rate": 1.308080808080808e-07,
"loss": 0.5545,
"step": 8710
},
{
"epoch": 3.13,
"grad_norm": 4.455116271972656,
"learning_rate": 1.297979797979798e-07,
"loss": 0.5631,
"step": 8720
},
{
"epoch": 3.13,
"grad_norm": 11.706398963928223,
"learning_rate": 1.2878787878787877e-07,
"loss": 0.5218,
"step": 8730
},
{
"epoch": 3.13,
"grad_norm": 6.45891809463501,
"learning_rate": 1.2777777777777777e-07,
"loss": 0.5383,
"step": 8740
},
{
"epoch": 3.14,
"grad_norm": 4.817904949188232,
"learning_rate": 1.2676767676767676e-07,
"loss": 0.5093,
"step": 8750
},
{
"epoch": 3.14,
"grad_norm": 7.290755748748779,
"learning_rate": 1.2575757575757576e-07,
"loss": 0.4842,
"step": 8760
},
{
"epoch": 3.15,
"grad_norm": 3.76725435256958,
"learning_rate": 1.2474747474747473e-07,
"loss": 0.5476,
"step": 8770
},
{
"epoch": 3.15,
"grad_norm": 5.706302642822266,
"learning_rate": 1.2373737373737373e-07,
"loss": 0.4709,
"step": 8780
},
{
"epoch": 3.15,
"grad_norm": 7.164012432098389,
"learning_rate": 1.2272727272727272e-07,
"loss": 0.5088,
"step": 8790
},
{
"epoch": 3.16,
"grad_norm": 5.86743688583374,
"learning_rate": 1.2171717171717172e-07,
"loss": 0.5361,
"step": 8800
},
{
"epoch": 3.16,
"grad_norm": 4.16366720199585,
"learning_rate": 1.207070707070707e-07,
"loss": 0.536,
"step": 8810
},
{
"epoch": 3.16,
"grad_norm": 6.925227642059326,
"learning_rate": 1.196969696969697e-07,
"loss": 0.5001,
"step": 8820
},
{
"epoch": 3.17,
"grad_norm": 3.9916467666625977,
"learning_rate": 1.1868686868686869e-07,
"loss": 0.5662,
"step": 8830
},
{
"epoch": 3.17,
"grad_norm": 7.0623931884765625,
"learning_rate": 1.1767676767676767e-07,
"loss": 0.5813,
"step": 8840
},
{
"epoch": 3.17,
"grad_norm": 5.496065139770508,
"learning_rate": 1.1666666666666667e-07,
"loss": 0.5751,
"step": 8850
},
{
"epoch": 3.18,
"grad_norm": 3.784182071685791,
"learning_rate": 1.1565656565656565e-07,
"loss": 0.5863,
"step": 8860
},
{
"epoch": 3.18,
"grad_norm": 5.068544864654541,
"learning_rate": 1.1464646464646465e-07,
"loss": 0.5098,
"step": 8870
},
{
"epoch": 3.18,
"grad_norm": 5.4960808753967285,
"learning_rate": 1.1363636363636363e-07,
"loss": 0.6326,
"step": 8880
},
{
"epoch": 3.19,
"grad_norm": 4.9256792068481445,
"learning_rate": 1.1262626262626263e-07,
"loss": 0.4801,
"step": 8890
},
{
"epoch": 3.19,
"grad_norm": 4.391407489776611,
"learning_rate": 1.1161616161616161e-07,
"loss": 0.5866,
"step": 8900
},
{
"epoch": 3.2,
"grad_norm": 4.210151195526123,
"learning_rate": 1.1060606060606061e-07,
"loss": 0.5464,
"step": 8910
},
{
"epoch": 3.2,
"grad_norm": 6.131332874298096,
"learning_rate": 1.0959595959595959e-07,
"loss": 0.4419,
"step": 8920
},
{
"epoch": 3.2,
"grad_norm": 5.612468719482422,
"learning_rate": 1.0858585858585859e-07,
"loss": 0.6155,
"step": 8930
},
{
"epoch": 3.21,
"grad_norm": 4.1157097816467285,
"learning_rate": 1.0757575757575757e-07,
"loss": 0.5514,
"step": 8940
},
{
"epoch": 3.21,
"grad_norm": 5.19150972366333,
"learning_rate": 1.0656565656565657e-07,
"loss": 0.548,
"step": 8950
},
{
"epoch": 3.21,
"grad_norm": 4.305140972137451,
"learning_rate": 1.0555555555555555e-07,
"loss": 0.5313,
"step": 8960
},
{
"epoch": 3.22,
"grad_norm": 6.152969837188721,
"learning_rate": 1.0454545454545454e-07,
"loss": 0.4861,
"step": 8970
},
{
"epoch": 3.22,
"grad_norm": 4.60936164855957,
"learning_rate": 1.0353535353535353e-07,
"loss": 0.6055,
"step": 8980
},
{
"epoch": 3.22,
"grad_norm": 9.371330261230469,
"learning_rate": 1.0252525252525252e-07,
"loss": 0.5255,
"step": 8990
},
{
"epoch": 3.23,
"grad_norm": 5.383856296539307,
"learning_rate": 1.0151515151515151e-07,
"loss": 0.4944,
"step": 9000
},
{
"epoch": 3.23,
"eval_loss": 0.6957755088806152,
"eval_runtime": 404.454,
"eval_samples_per_second": 2.472,
"eval_steps_per_second": 2.472,
"step": 9000
},
{
"epoch": 3.23,
"grad_norm": 6.594298839569092,
"learning_rate": 1.005050505050505e-07,
"loss": 0.5506,
"step": 9010
},
{
"epoch": 3.23,
"grad_norm": 5.5882248878479,
"learning_rate": 9.94949494949495e-08,
"loss": 0.4821,
"step": 9020
},
{
"epoch": 3.24,
"grad_norm": 5.565232753753662,
"learning_rate": 9.848484848484848e-08,
"loss": 0.5903,
"step": 9030
},
{
"epoch": 3.24,
"grad_norm": 4.196237564086914,
"learning_rate": 9.747474747474747e-08,
"loss": 0.4784,
"step": 9040
},
{
"epoch": 3.25,
"grad_norm": 4.757313251495361,
"learning_rate": 9.646464646464646e-08,
"loss": 0.5256,
"step": 9050
},
{
"epoch": 3.25,
"grad_norm": 5.351820945739746,
"learning_rate": 9.545454545454546e-08,
"loss": 0.4909,
"step": 9060
},
{
"epoch": 3.25,
"grad_norm": 5.794235706329346,
"learning_rate": 9.444444444444444e-08,
"loss": 0.5444,
"step": 9070
},
{
"epoch": 3.26,
"grad_norm": 6.105288028717041,
"learning_rate": 9.343434343434344e-08,
"loss": 0.5256,
"step": 9080
},
{
"epoch": 3.26,
"grad_norm": 6.002346992492676,
"learning_rate": 9.242424242424242e-08,
"loss": 0.5005,
"step": 9090
},
{
"epoch": 3.26,
"grad_norm": 5.016416072845459,
"learning_rate": 9.141414141414142e-08,
"loss": 0.5931,
"step": 9100
},
{
"epoch": 3.27,
"grad_norm": 4.596400737762451,
"learning_rate": 9.04040404040404e-08,
"loss": 0.4898,
"step": 9110
},
{
"epoch": 3.27,
"grad_norm": 5.950545310974121,
"learning_rate": 8.93939393939394e-08,
"loss": 0.6261,
"step": 9120
},
{
"epoch": 3.27,
"grad_norm": 6.038389205932617,
"learning_rate": 8.838383838383838e-08,
"loss": 0.4946,
"step": 9130
},
{
"epoch": 3.28,
"grad_norm": 6.918920516967773,
"learning_rate": 8.737373737373736e-08,
"loss": 0.5929,
"step": 9140
},
{
"epoch": 3.28,
"grad_norm": 4.859712600708008,
"learning_rate": 8.636363636363636e-08,
"loss": 0.5885,
"step": 9150
},
{
"epoch": 3.28,
"grad_norm": 8.945250511169434,
"learning_rate": 8.535353535353534e-08,
"loss": 0.5589,
"step": 9160
},
{
"epoch": 3.29,
"grad_norm": 4.006341934204102,
"learning_rate": 8.434343434343434e-08,
"loss": 0.4796,
"step": 9170
},
{
"epoch": 3.29,
"grad_norm": 6.528043746948242,
"learning_rate": 8.333333333333333e-08,
"loss": 0.5737,
"step": 9180
},
{
"epoch": 3.3,
"grad_norm": 7.656773090362549,
"learning_rate": 8.232323232323232e-08,
"loss": 0.549,
"step": 9190
},
{
"epoch": 3.3,
"grad_norm": 4.172515869140625,
"learning_rate": 8.13131313131313e-08,
"loss": 0.5112,
"step": 9200
},
{
"epoch": 3.3,
"grad_norm": 6.6936116218566895,
"learning_rate": 8.03030303030303e-08,
"loss": 0.5645,
"step": 9210
},
{
"epoch": 3.31,
"grad_norm": 7.118635177612305,
"learning_rate": 7.929292929292929e-08,
"loss": 0.5502,
"step": 9220
},
{
"epoch": 3.31,
"grad_norm": 4.658480167388916,
"learning_rate": 7.828282828282828e-08,
"loss": 0.4883,
"step": 9230
},
{
"epoch": 3.31,
"grad_norm": 7.470480442047119,
"learning_rate": 7.727272727272727e-08,
"loss": 0.5102,
"step": 9240
},
{
"epoch": 3.32,
"grad_norm": 4.436034679412842,
"learning_rate": 7.626262626262626e-08,
"loss": 0.5939,
"step": 9250
},
{
"epoch": 3.32,
"grad_norm": 4.717689037322998,
"learning_rate": 7.525252525252525e-08,
"loss": 0.5976,
"step": 9260
},
{
"epoch": 3.32,
"grad_norm": 4.957371711730957,
"learning_rate": 7.424242424242424e-08,
"loss": 0.5019,
"step": 9270
},
{
"epoch": 3.33,
"grad_norm": 4.873682022094727,
"learning_rate": 7.323232323232323e-08,
"loss": 0.5755,
"step": 9280
},
{
"epoch": 3.33,
"grad_norm": 4.161667823791504,
"learning_rate": 7.222222222222221e-08,
"loss": 0.6371,
"step": 9290
},
{
"epoch": 3.34,
"grad_norm": 4.539839744567871,
"learning_rate": 7.121212121212121e-08,
"loss": 0.6012,
"step": 9300
},
{
"epoch": 3.34,
"grad_norm": 5.414936542510986,
"learning_rate": 7.020202020202019e-08,
"loss": 0.5449,
"step": 9310
},
{
"epoch": 3.34,
"grad_norm": 10.87019157409668,
"learning_rate": 6.919191919191919e-08,
"loss": 0.5272,
"step": 9320
},
{
"epoch": 3.35,
"grad_norm": 5.128973484039307,
"learning_rate": 6.818181818181817e-08,
"loss": 0.5053,
"step": 9330
},
{
"epoch": 3.35,
"grad_norm": 4.943745136260986,
"learning_rate": 6.717171717171717e-08,
"loss": 0.5823,
"step": 9340
},
{
"epoch": 3.35,
"grad_norm": 5.1109185218811035,
"learning_rate": 6.616161616161615e-08,
"loss": 0.5345,
"step": 9350
},
{
"epoch": 3.36,
"grad_norm": 7.44246768951416,
"learning_rate": 6.515151515151515e-08,
"loss": 0.5677,
"step": 9360
},
{
"epoch": 3.36,
"grad_norm": 4.263915538787842,
"learning_rate": 6.414141414141413e-08,
"loss": 0.522,
"step": 9370
},
{
"epoch": 3.36,
"grad_norm": 5.933948516845703,
"learning_rate": 6.313131313131313e-08,
"loss": 0.4782,
"step": 9380
},
{
"epoch": 3.37,
"grad_norm": 4.103212356567383,
"learning_rate": 6.212121212121211e-08,
"loss": 0.5602,
"step": 9390
},
{
"epoch": 3.37,
"grad_norm": 5.865060806274414,
"learning_rate": 6.111111111111111e-08,
"loss": 0.5654,
"step": 9400
},
{
"epoch": 3.37,
"grad_norm": 3.8256332874298096,
"learning_rate": 6.01010101010101e-08,
"loss": 0.5984,
"step": 9410
},
{
"epoch": 3.38,
"grad_norm": 6.088188648223877,
"learning_rate": 5.9090909090909085e-08,
"loss": 0.5782,
"step": 9420
},
{
"epoch": 3.38,
"grad_norm": 6.299181938171387,
"learning_rate": 5.8080808080808076e-08,
"loss": 0.5725,
"step": 9430
},
{
"epoch": 3.39,
"grad_norm": 5.62983512878418,
"learning_rate": 5.7070707070707066e-08,
"loss": 0.5407,
"step": 9440
},
{
"epoch": 3.39,
"grad_norm": 5.949933052062988,
"learning_rate": 5.6060606060606056e-08,
"loss": 0.4399,
"step": 9450
},
{
"epoch": 3.39,
"grad_norm": 5.324888229370117,
"learning_rate": 5.5050505050505046e-08,
"loss": 0.5178,
"step": 9460
},
{
"epoch": 3.4,
"grad_norm": 7.061882019042969,
"learning_rate": 5.4040404040404037e-08,
"loss": 0.5175,
"step": 9470
},
{
"epoch": 3.4,
"grad_norm": 4.5119404792785645,
"learning_rate": 5.303030303030303e-08,
"loss": 0.5282,
"step": 9480
},
{
"epoch": 3.4,
"grad_norm": 4.700840473175049,
"learning_rate": 5.202020202020202e-08,
"loss": 0.5202,
"step": 9490
},
{
"epoch": 3.41,
"grad_norm": 5.251560211181641,
"learning_rate": 5.101010101010101e-08,
"loss": 0.5013,
"step": 9500
},
{
"epoch": 3.41,
"grad_norm": 6.137795925140381,
"learning_rate": 5e-08,
"loss": 0.4514,
"step": 9510
},
{
"epoch": 3.41,
"grad_norm": 6.373246669769287,
"learning_rate": 4.898989898989899e-08,
"loss": 0.5129,
"step": 9520
},
{
"epoch": 3.42,
"grad_norm": 6.3960418701171875,
"learning_rate": 4.797979797979797e-08,
"loss": 0.5055,
"step": 9530
},
{
"epoch": 3.42,
"grad_norm": 4.761574745178223,
"learning_rate": 4.696969696969696e-08,
"loss": 0.5724,
"step": 9540
},
{
"epoch": 3.42,
"grad_norm": 4.658292770385742,
"learning_rate": 4.595959595959595e-08,
"loss": 0.5617,
"step": 9550
},
{
"epoch": 3.43,
"grad_norm": 6.56446647644043,
"learning_rate": 4.494949494949494e-08,
"loss": 0.5495,
"step": 9560
},
{
"epoch": 3.43,
"grad_norm": 5.917604446411133,
"learning_rate": 4.393939393939393e-08,
"loss": 0.5513,
"step": 9570
},
{
"epoch": 3.44,
"grad_norm": 6.456116676330566,
"learning_rate": 4.292929292929292e-08,
"loss": 0.5198,
"step": 9580
},
{
"epoch": 3.44,
"grad_norm": 5.084980487823486,
"learning_rate": 4.1919191919191913e-08,
"loss": 0.5737,
"step": 9590
},
{
"epoch": 3.44,
"grad_norm": 3.8530325889587402,
"learning_rate": 4.090909090909091e-08,
"loss": 0.4901,
"step": 9600
},
{
"epoch": 3.45,
"grad_norm": 4.385349750518799,
"learning_rate": 3.98989898989899e-08,
"loss": 0.5092,
"step": 9610
},
{
"epoch": 3.45,
"grad_norm": 3.888326406478882,
"learning_rate": 3.888888888888889e-08,
"loss": 0.5536,
"step": 9620
},
{
"epoch": 3.45,
"grad_norm": 4.46201229095459,
"learning_rate": 3.787878787878788e-08,
"loss": 0.5697,
"step": 9630
},
{
"epoch": 3.46,
"grad_norm": 7.194265842437744,
"learning_rate": 3.686868686868687e-08,
"loss": 0.4886,
"step": 9640
},
{
"epoch": 3.46,
"grad_norm": 6.102540016174316,
"learning_rate": 3.585858585858586e-08,
"loss": 0.4981,
"step": 9650
},
{
"epoch": 3.46,
"grad_norm": 7.469662189483643,
"learning_rate": 3.484848484848485e-08,
"loss": 0.5294,
"step": 9660
},
{
"epoch": 3.47,
"grad_norm": 4.181239128112793,
"learning_rate": 3.383838383838384e-08,
"loss": 0.5463,
"step": 9670
},
{
"epoch": 3.47,
"grad_norm": 5.108190536499023,
"learning_rate": 3.282828282828283e-08,
"loss": 0.4633,
"step": 9680
},
{
"epoch": 3.47,
"grad_norm": 5.501590251922607,
"learning_rate": 3.1818181818181816e-08,
"loss": 0.5121,
"step": 9690
},
{
"epoch": 3.48,
"grad_norm": 6.437745571136475,
"learning_rate": 3.0808080808080806e-08,
"loss": 0.475,
"step": 9700
},
{
"epoch": 3.48,
"grad_norm": 7.1137166023254395,
"learning_rate": 2.9797979797979797e-08,
"loss": 0.5536,
"step": 9710
},
{
"epoch": 3.49,
"grad_norm": 5.652755260467529,
"learning_rate": 2.8787878787878787e-08,
"loss": 0.5493,
"step": 9720
},
{
"epoch": 3.49,
"grad_norm": 4.296222686767578,
"learning_rate": 2.7777777777777774e-08,
"loss": 0.5323,
"step": 9730
},
{
"epoch": 3.49,
"grad_norm": 5.550398826599121,
"learning_rate": 2.6767676767676764e-08,
"loss": 0.5555,
"step": 9740
},
{
"epoch": 3.5,
"grad_norm": 6.785501956939697,
"learning_rate": 2.5757575757575755e-08,
"loss": 0.5388,
"step": 9750
},
{
"epoch": 3.5,
"grad_norm": 7.063258171081543,
"learning_rate": 2.4747474747474745e-08,
"loss": 0.5742,
"step": 9760
},
{
"epoch": 3.5,
"grad_norm": 5.894256114959717,
"learning_rate": 2.3737373737373735e-08,
"loss": 0.6248,
"step": 9770
},
{
"epoch": 3.51,
"grad_norm": 5.81352424621582,
"learning_rate": 2.2727272727272725e-08,
"loss": 0.4896,
"step": 9780
},
{
"epoch": 3.51,
"grad_norm": 5.251163482666016,
"learning_rate": 2.1717171717171716e-08,
"loss": 0.5633,
"step": 9790
},
{
"epoch": 3.51,
"grad_norm": 4.963428020477295,
"learning_rate": 2.070707070707071e-08,
"loss": 0.4617,
"step": 9800
},
{
"epoch": 3.52,
"grad_norm": 4.8795928955078125,
"learning_rate": 1.9696969696969696e-08,
"loss": 0.5325,
"step": 9810
},
{
"epoch": 3.52,
"grad_norm": 5.59514856338501,
"learning_rate": 1.8686868686868686e-08,
"loss": 0.5535,
"step": 9820
},
{
"epoch": 3.53,
"grad_norm": 5.933664321899414,
"learning_rate": 1.7676767676767677e-08,
"loss": 0.4904,
"step": 9830
},
{
"epoch": 3.53,
"grad_norm": 5.348793029785156,
"learning_rate": 1.6666666666666667e-08,
"loss": 0.543,
"step": 9840
},
{
"epoch": 3.53,
"grad_norm": 4.590538024902344,
"learning_rate": 1.5656565656565657e-08,
"loss": 0.5443,
"step": 9850
},
{
"epoch": 3.54,
"grad_norm": 7.6259918212890625,
"learning_rate": 1.4646464646464646e-08,
"loss": 0.4999,
"step": 9860
},
{
"epoch": 3.54,
"grad_norm": 5.3782124519348145,
"learning_rate": 1.3636363636363635e-08,
"loss": 0.5064,
"step": 9870
},
{
"epoch": 3.54,
"grad_norm": 6.371192932128906,
"learning_rate": 1.2626262626262625e-08,
"loss": 0.5487,
"step": 9880
},
{
"epoch": 3.55,
"grad_norm": 5.356314659118652,
"learning_rate": 1.1616161616161615e-08,
"loss": 0.506,
"step": 9890
},
{
"epoch": 3.55,
"grad_norm": 5.2767014503479,
"learning_rate": 1.0606060606060607e-08,
"loss": 0.5556,
"step": 9900
},
{
"epoch": 3.55,
"grad_norm": 4.274359226226807,
"learning_rate": 9.595959595959596e-09,
"loss": 0.5074,
"step": 9910
},
{
"epoch": 3.56,
"grad_norm": 5.464534759521484,
"learning_rate": 8.585858585858586e-09,
"loss": 0.4612,
"step": 9920
},
{
"epoch": 3.56,
"grad_norm": 9.630477905273438,
"learning_rate": 7.575757575757576e-09,
"loss": 0.6037,
"step": 9930
},
{
"epoch": 3.56,
"grad_norm": 3.692291259765625,
"learning_rate": 6.565656565656565e-09,
"loss": 0.5116,
"step": 9940
},
{
"epoch": 3.57,
"grad_norm": 5.269135475158691,
"learning_rate": 5.555555555555555e-09,
"loss": 0.591,
"step": 9950
},
{
"epoch": 3.57,
"grad_norm": 5.141294956207275,
"learning_rate": 4.545454545454545e-09,
"loss": 0.5726,
"step": 9960
},
{
"epoch": 3.58,
"grad_norm": 6.49954080581665,
"learning_rate": 3.5353535353535353e-09,
"loss": 0.5509,
"step": 9970
},
{
"epoch": 3.58,
"grad_norm": 5.085234642028809,
"learning_rate": 2.5252525252525255e-09,
"loss": 0.5337,
"step": 9980
},
{
"epoch": 3.58,
"grad_norm": 5.688786506652832,
"learning_rate": 1.515151515151515e-09,
"loss": 0.5624,
"step": 9990
},
{
"epoch": 3.59,
"grad_norm": 6.861104965209961,
"learning_rate": 5.050505050505051e-10,
"loss": 0.5167,
"step": 10000
},
{
"epoch": 3.59,
"eval_loss": 0.711556613445282,
"eval_runtime": 405.1832,
"eval_samples_per_second": 2.468,
"eval_steps_per_second": 2.468,
"step": 10000
}
],
"logging_steps": 10,
"max_steps": 10000,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 2000,
"total_flos": 9.4282098671616e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}