{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.5861574323112784, "eval_steps": 1000, "global_step": 10000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "grad_norm": 10.134367942810059, "learning_rate": 9e-08, "loss": 1.2511, "step": 10 }, { "epoch": 0.01, "grad_norm": 18.544782638549805, "learning_rate": 1.8e-07, "loss": 1.4001, "step": 20 }, { "epoch": 0.01, "grad_norm": 6.087899684906006, "learning_rate": 2.8e-07, "loss": 1.1501, "step": 30 }, { "epoch": 0.01, "grad_norm": 6.356815814971924, "learning_rate": 3.7999999999999996e-07, "loss": 1.2225, "step": 40 }, { "epoch": 0.02, "grad_norm": 3.9461255073547363, "learning_rate": 4.8e-07, "loss": 1.2241, "step": 50 }, { "epoch": 0.02, "grad_norm": 4.9660773277282715, "learning_rate": 5.8e-07, "loss": 1.2118, "step": 60 }, { "epoch": 0.03, "grad_norm": 6.57827615737915, "learning_rate": 6.800000000000001e-07, "loss": 1.0416, "step": 70 }, { "epoch": 0.03, "grad_norm": 5.070159435272217, "learning_rate": 7.799999999999999e-07, "loss": 1.1224, "step": 80 }, { "epoch": 0.03, "grad_norm": 5.661293029785156, "learning_rate": 8.799999999999999e-07, "loss": 1.0603, "step": 90 }, { "epoch": 0.04, "grad_norm": 4.765249252319336, "learning_rate": 9.8e-07, "loss": 1.0449, "step": 100 }, { "epoch": 0.04, "grad_norm": 4.521543502807617, "learning_rate": 9.991919191919192e-07, "loss": 0.9193, "step": 110 }, { "epoch": 0.04, "grad_norm": 4.420640468597412, "learning_rate": 9.98181818181818e-07, "loss": 0.876, "step": 120 }, { "epoch": 0.05, "grad_norm": 4.311656951904297, "learning_rate": 9.97171717171717e-07, "loss": 0.8012, "step": 130 }, { "epoch": 0.05, "grad_norm": 5.07025146484375, "learning_rate": 9.961616161616162e-07, "loss": 0.826, "step": 140 }, { "epoch": 0.05, "grad_norm": 4.859681129455566, "learning_rate": 9.951515151515151e-07, "loss": 0.8442, "step": 150 }, { "epoch": 0.06, "grad_norm": 4.192190170288086, "learning_rate": 9.94141414141414e-07, "loss": 0.708, "step": 160 }, { "epoch": 0.06, "grad_norm": 5.852901935577393, "learning_rate": 9.93131313131313e-07, "loss": 0.8201, "step": 170 }, { "epoch": 0.06, "grad_norm": 5.076972484588623, "learning_rate": 9.92121212121212e-07, "loss": 0.7973, "step": 180 }, { "epoch": 0.07, "grad_norm": 4.36331844329834, "learning_rate": 9.911111111111111e-07, "loss": 0.753, "step": 190 }, { "epoch": 0.07, "grad_norm": 5.03492546081543, "learning_rate": 9.9010101010101e-07, "loss": 0.7909, "step": 200 }, { "epoch": 0.08, "grad_norm": 4.043544769287109, "learning_rate": 9.89090909090909e-07, "loss": 0.7118, "step": 210 }, { "epoch": 0.08, "grad_norm": 2.2020692825317383, "learning_rate": 9.880808080808082e-07, "loss": 0.6516, "step": 220 }, { "epoch": 0.08, "grad_norm": 4.851771354675293, "learning_rate": 9.870707070707071e-07, "loss": 0.7275, "step": 230 }, { "epoch": 0.09, "grad_norm": 3.2986488342285156, "learning_rate": 9.86060606060606e-07, "loss": 0.6484, "step": 240 }, { "epoch": 0.09, "grad_norm": 5.700368404388428, "learning_rate": 9.85050505050505e-07, "loss": 0.7258, "step": 250 }, { "epoch": 0.09, "grad_norm": 6.46458625793457, "learning_rate": 9.84040404040404e-07, "loss": 0.6416, "step": 260 }, { "epoch": 0.1, "grad_norm": 4.012514114379883, "learning_rate": 9.830303030303029e-07, "loss": 0.5861, "step": 270 }, { "epoch": 0.1, "grad_norm": 3.328000068664551, "learning_rate": 9.82020202020202e-07, "loss": 0.6239, "step": 280 }, { "epoch": 0.1, "grad_norm": 3.837636947631836, "learning_rate": 9.81010101010101e-07, "loss": 0.6718, "step": 290 }, { "epoch": 0.11, "grad_norm": 4.840264797210693, "learning_rate": 9.8e-07, "loss": 0.6623, "step": 300 }, { "epoch": 0.11, "grad_norm": 6.741188049316406, "learning_rate": 9.789898989898989e-07, "loss": 0.6066, "step": 310 }, { "epoch": 0.11, "grad_norm": 5.170821666717529, "learning_rate": 9.779797979797978e-07, "loss": 0.6634, "step": 320 }, { "epoch": 0.12, "grad_norm": 8.481480598449707, "learning_rate": 9.76969696969697e-07, "loss": 0.6474, "step": 330 }, { "epoch": 0.12, "grad_norm": 3.698042631149292, "learning_rate": 9.75959595959596e-07, "loss": 0.6594, "step": 340 }, { "epoch": 0.13, "grad_norm": 4.990598201751709, "learning_rate": 9.749494949494949e-07, "loss": 0.6588, "step": 350 }, { "epoch": 0.13, "grad_norm": 3.2625105381011963, "learning_rate": 9.73939393939394e-07, "loss": 0.6532, "step": 360 }, { "epoch": 0.13, "grad_norm": 6.604375839233398, "learning_rate": 9.72929292929293e-07, "loss": 0.7057, "step": 370 }, { "epoch": 0.14, "grad_norm": 5.352667808532715, "learning_rate": 9.71919191919192e-07, "loss": 0.6273, "step": 380 }, { "epoch": 0.14, "grad_norm": 3.8649349212646484, "learning_rate": 9.709090909090909e-07, "loss": 0.6366, "step": 390 }, { "epoch": 0.14, "grad_norm": 5.2739667892456055, "learning_rate": 9.698989898989898e-07, "loss": 0.6433, "step": 400 }, { "epoch": 0.15, "grad_norm": 4.892832279205322, "learning_rate": 9.68888888888889e-07, "loss": 0.6504, "step": 410 }, { "epoch": 0.15, "grad_norm": 5.36677360534668, "learning_rate": 9.67878787878788e-07, "loss": 0.6661, "step": 420 }, { "epoch": 0.15, "grad_norm": 2.6905229091644287, "learning_rate": 9.668686868686868e-07, "loss": 0.6334, "step": 430 }, { "epoch": 0.16, "grad_norm": 4.99962854385376, "learning_rate": 9.658585858585858e-07, "loss": 0.6028, "step": 440 }, { "epoch": 0.16, "grad_norm": 5.522456169128418, "learning_rate": 9.648484848484847e-07, "loss": 0.6903, "step": 450 }, { "epoch": 0.16, "grad_norm": 7.456121921539307, "learning_rate": 9.638383838383839e-07, "loss": 0.6486, "step": 460 }, { "epoch": 0.17, "grad_norm": 5.989662170410156, "learning_rate": 9.628282828282828e-07, "loss": 0.7085, "step": 470 }, { "epoch": 0.17, "grad_norm": 3.5824246406555176, "learning_rate": 9.618181818181818e-07, "loss": 0.6652, "step": 480 }, { "epoch": 0.18, "grad_norm": 3.074733018875122, "learning_rate": 9.608080808080807e-07, "loss": 0.5733, "step": 490 }, { "epoch": 0.18, "grad_norm": 3.964071750640869, "learning_rate": 9.597979797979797e-07, "loss": 0.6319, "step": 500 }, { "epoch": 0.18, "grad_norm": 3.8981716632843018, "learning_rate": 9.587878787878786e-07, "loss": 0.5937, "step": 510 }, { "epoch": 0.19, "grad_norm": 7.471535682678223, "learning_rate": 9.577777777777778e-07, "loss": 0.6564, "step": 520 }, { "epoch": 0.19, "grad_norm": 5.468303680419922, "learning_rate": 9.567676767676767e-07, "loss": 0.6339, "step": 530 }, { "epoch": 0.19, "grad_norm": 3.9055886268615723, "learning_rate": 9.557575757575759e-07, "loss": 0.7103, "step": 540 }, { "epoch": 0.2, "grad_norm": 4.115898132324219, "learning_rate": 9.547474747474748e-07, "loss": 0.6928, "step": 550 }, { "epoch": 0.2, "grad_norm": 4.248528003692627, "learning_rate": 9.537373737373737e-07, "loss": 0.6084, "step": 560 }, { "epoch": 0.2, "grad_norm": 3.727003335952759, "learning_rate": 9.527272727272727e-07, "loss": 0.6554, "step": 570 }, { "epoch": 0.21, "grad_norm": 4.747541904449463, "learning_rate": 9.517171717171717e-07, "loss": 0.5876, "step": 580 }, { "epoch": 0.21, "grad_norm": 3.913226842880249, "learning_rate": 9.507070707070707e-07, "loss": 0.5888, "step": 590 }, { "epoch": 0.22, "grad_norm": 4.400748252868652, "learning_rate": 9.496969696969696e-07, "loss": 0.6196, "step": 600 }, { "epoch": 0.22, "grad_norm": 3.053006649017334, "learning_rate": 9.486868686868687e-07, "loss": 0.6356, "step": 610 }, { "epoch": 0.22, "grad_norm": 4.207491874694824, "learning_rate": 9.476767676767676e-07, "loss": 0.6161, "step": 620 }, { "epoch": 0.23, "grad_norm": 6.389359951019287, "learning_rate": 9.466666666666666e-07, "loss": 0.6172, "step": 630 }, { "epoch": 0.23, "grad_norm": 4.967292308807373, "learning_rate": 9.456565656565656e-07, "loss": 0.6173, "step": 640 }, { "epoch": 0.23, "grad_norm": 3.855407953262329, "learning_rate": 9.446464646464646e-07, "loss": 0.585, "step": 650 }, { "epoch": 0.24, "grad_norm": 3.580738067626953, "learning_rate": 9.436363636363636e-07, "loss": 0.6166, "step": 660 }, { "epoch": 0.24, "grad_norm": 3.6277294158935547, "learning_rate": 9.426262626262626e-07, "loss": 0.5951, "step": 670 }, { "epoch": 0.24, "grad_norm": 3.7501285076141357, "learning_rate": 9.416161616161615e-07, "loss": 0.5807, "step": 680 }, { "epoch": 0.25, "grad_norm": 4.80528450012207, "learning_rate": 9.406060606060605e-07, "loss": 0.6053, "step": 690 }, { "epoch": 0.25, "grad_norm": 5.07316780090332, "learning_rate": 9.395959595959596e-07, "loss": 0.626, "step": 700 }, { "epoch": 0.25, "grad_norm": 5.338922023773193, "learning_rate": 9.385858585858585e-07, "loss": 0.5356, "step": 710 }, { "epoch": 0.26, "grad_norm": 3.264220714569092, "learning_rate": 9.375757575757576e-07, "loss": 0.5929, "step": 720 }, { "epoch": 0.26, "grad_norm": 3.4565188884735107, "learning_rate": 9.365656565656565e-07, "loss": 0.6379, "step": 730 }, { "epoch": 0.27, "grad_norm": 4.202028751373291, "learning_rate": 9.355555555555556e-07, "loss": 0.6143, "step": 740 }, { "epoch": 0.27, "grad_norm": 5.128079891204834, "learning_rate": 9.345454545454545e-07, "loss": 0.5531, "step": 750 }, { "epoch": 0.27, "grad_norm": 4.205199718475342, "learning_rate": 9.335353535353535e-07, "loss": 0.5842, "step": 760 }, { "epoch": 0.28, "grad_norm": 5.925571918487549, "learning_rate": 9.325252525252525e-07, "loss": 0.6169, "step": 770 }, { "epoch": 0.28, "grad_norm": 4.256628036499023, "learning_rate": 9.315151515151515e-07, "loss": 0.5899, "step": 780 }, { "epoch": 0.28, "grad_norm": 3.2796261310577393, "learning_rate": 9.305050505050504e-07, "loss": 0.6654, "step": 790 }, { "epoch": 0.29, "grad_norm": 3.6715095043182373, "learning_rate": 9.294949494949495e-07, "loss": 0.6272, "step": 800 }, { "epoch": 0.29, "grad_norm": 5.043895721435547, "learning_rate": 9.284848484848484e-07, "loss": 0.6019, "step": 810 }, { "epoch": 0.29, "grad_norm": 4.5200300216674805, "learning_rate": 9.274747474747475e-07, "loss": 0.6992, "step": 820 }, { "epoch": 0.3, "grad_norm": 5.557192325592041, "learning_rate": 9.264646464646464e-07, "loss": 0.6657, "step": 830 }, { "epoch": 0.3, "grad_norm": 3.2051031589508057, "learning_rate": 9.254545454545453e-07, "loss": 0.6262, "step": 840 }, { "epoch": 0.3, "grad_norm": 3.465202808380127, "learning_rate": 9.244444444444444e-07, "loss": 0.6603, "step": 850 }, { "epoch": 0.31, "grad_norm": 5.039762496948242, "learning_rate": 9.234343434343433e-07, "loss": 0.6276, "step": 860 }, { "epoch": 0.31, "grad_norm": 5.317465305328369, "learning_rate": 9.224242424242423e-07, "loss": 0.521, "step": 870 }, { "epoch": 0.32, "grad_norm": 6.985171318054199, "learning_rate": 9.214141414141414e-07, "loss": 0.6768, "step": 880 }, { "epoch": 0.32, "grad_norm": 3.9759483337402344, "learning_rate": 9.204040404040404e-07, "loss": 0.5982, "step": 890 }, { "epoch": 0.32, "grad_norm": 4.207100868225098, "learning_rate": 9.193939393939394e-07, "loss": 0.6117, "step": 900 }, { "epoch": 0.33, "grad_norm": 4.389718055725098, "learning_rate": 9.183838383838384e-07, "loss": 0.5767, "step": 910 }, { "epoch": 0.33, "grad_norm": 3.659482717514038, "learning_rate": 9.173737373737373e-07, "loss": 0.5957, "step": 920 }, { "epoch": 0.33, "grad_norm": 6.660622596740723, "learning_rate": 9.163636363636364e-07, "loss": 0.6403, "step": 930 }, { "epoch": 0.34, "grad_norm": 6.03493070602417, "learning_rate": 9.153535353535353e-07, "loss": 0.5857, "step": 940 }, { "epoch": 0.34, "grad_norm": 2.699523687362671, "learning_rate": 9.143434343434343e-07, "loss": 0.5913, "step": 950 }, { "epoch": 0.34, "grad_norm": 3.8991754055023193, "learning_rate": 9.133333333333333e-07, "loss": 0.6342, "step": 960 }, { "epoch": 0.35, "grad_norm": 4.422178268432617, "learning_rate": 9.123232323232323e-07, "loss": 0.6286, "step": 970 }, { "epoch": 0.35, "grad_norm": 4.381803512573242, "learning_rate": 9.113131313131313e-07, "loss": 0.6899, "step": 980 }, { "epoch": 0.36, "grad_norm": 4.75442361831665, "learning_rate": 9.103030303030302e-07, "loss": 0.5462, "step": 990 }, { "epoch": 0.36, "grad_norm": 3.6305062770843506, "learning_rate": 9.092929292929292e-07, "loss": 0.666, "step": 1000 }, { "epoch": 0.36, "eval_loss": 0.772658109664917, "eval_runtime": 400.6004, "eval_samples_per_second": 2.496, "eval_steps_per_second": 2.496, "step": 1000 }, { "epoch": 0.36, "grad_norm": 4.022682189941406, "learning_rate": 9.082828282828282e-07, "loss": 0.5796, "step": 1010 }, { "epoch": 0.37, "grad_norm": 5.148658752441406, "learning_rate": 9.072727272727272e-07, "loss": 0.5738, "step": 1020 }, { "epoch": 0.37, "grad_norm": 3.78167462348938, "learning_rate": 9.062626262626261e-07, "loss": 0.6572, "step": 1030 }, { "epoch": 0.37, "grad_norm": 4.6793212890625, "learning_rate": 9.052525252525252e-07, "loss": 0.6466, "step": 1040 }, { "epoch": 0.38, "grad_norm": 4.607004165649414, "learning_rate": 9.042424242424242e-07, "loss": 0.5599, "step": 1050 }, { "epoch": 0.38, "grad_norm": 5.730196952819824, "learning_rate": 9.032323232323233e-07, "loss": 0.6559, "step": 1060 }, { "epoch": 0.38, "grad_norm": 4.408863544464111, "learning_rate": 9.022222222222222e-07, "loss": 0.641, "step": 1070 }, { "epoch": 0.39, "grad_norm": 3.0303945541381836, "learning_rate": 9.012121212121212e-07, "loss": 0.5455, "step": 1080 }, { "epoch": 0.39, "grad_norm": 5.44600248336792, "learning_rate": 9.002020202020202e-07, "loss": 0.5672, "step": 1090 }, { "epoch": 0.39, "grad_norm": 4.877390384674072, "learning_rate": 8.991919191919192e-07, "loss": 0.5626, "step": 1100 }, { "epoch": 0.4, "grad_norm": 4.949882507324219, "learning_rate": 8.981818181818181e-07, "loss": 0.6092, "step": 1110 }, { "epoch": 0.4, "grad_norm": 4.810749530792236, "learning_rate": 8.971717171717172e-07, "loss": 0.6486, "step": 1120 }, { "epoch": 0.41, "grad_norm": 4.6897053718566895, "learning_rate": 8.961616161616161e-07, "loss": 0.5854, "step": 1130 }, { "epoch": 0.41, "grad_norm": 3.393486261367798, "learning_rate": 8.951515151515151e-07, "loss": 0.6071, "step": 1140 }, { "epoch": 0.41, "grad_norm": 3.1970651149749756, "learning_rate": 8.941414141414141e-07, "loss": 0.601, "step": 1150 }, { "epoch": 0.42, "grad_norm": 4.972078800201416, "learning_rate": 8.93131313131313e-07, "loss": 0.6291, "step": 1160 }, { "epoch": 0.42, "grad_norm": 5.2576003074646, "learning_rate": 8.921212121212121e-07, "loss": 0.5856, "step": 1170 }, { "epoch": 0.42, "grad_norm": 3.552098035812378, "learning_rate": 8.91111111111111e-07, "loss": 0.5712, "step": 1180 }, { "epoch": 0.43, "grad_norm": 4.2522382736206055, "learning_rate": 8.9010101010101e-07, "loss": 0.6128, "step": 1190 }, { "epoch": 0.43, "grad_norm": 3.3408310413360596, "learning_rate": 8.89090909090909e-07, "loss": 0.6636, "step": 1200 }, { "epoch": 0.43, "grad_norm": 3.3888745307922363, "learning_rate": 8.88080808080808e-07, "loss": 0.5906, "step": 1210 }, { "epoch": 0.44, "grad_norm": 4.374114990234375, "learning_rate": 8.870707070707071e-07, "loss": 0.5863, "step": 1220 }, { "epoch": 0.44, "grad_norm": 3.318901538848877, "learning_rate": 8.860606060606061e-07, "loss": 0.6232, "step": 1230 }, { "epoch": 0.44, "grad_norm": 3.4461617469787598, "learning_rate": 8.85050505050505e-07, "loss": 0.6712, "step": 1240 }, { "epoch": 0.45, "grad_norm": 3.6037635803222656, "learning_rate": 8.840404040404041e-07, "loss": 0.5787, "step": 1250 }, { "epoch": 0.45, "grad_norm": 4.3236002922058105, "learning_rate": 8.83030303030303e-07, "loss": 0.6723, "step": 1260 }, { "epoch": 0.46, "grad_norm": 3.199084997177124, "learning_rate": 8.820202020202019e-07, "loss": 0.5484, "step": 1270 }, { "epoch": 0.46, "grad_norm": 3.675905704498291, "learning_rate": 8.81010101010101e-07, "loss": 0.6713, "step": 1280 }, { "epoch": 0.46, "grad_norm": 4.667660713195801, "learning_rate": 8.799999999999999e-07, "loss": 0.6161, "step": 1290 }, { "epoch": 0.47, "grad_norm": 5.268735408782959, "learning_rate": 8.78989898989899e-07, "loss": 0.6282, "step": 1300 }, { "epoch": 0.47, "grad_norm": 3.6777164936065674, "learning_rate": 8.779797979797979e-07, "loss": 0.662, "step": 1310 }, { "epoch": 0.47, "grad_norm": 5.912172317504883, "learning_rate": 8.769696969696969e-07, "loss": 0.5951, "step": 1320 }, { "epoch": 0.48, "grad_norm": 4.206521034240723, "learning_rate": 8.759595959595959e-07, "loss": 0.704, "step": 1330 }, { "epoch": 0.48, "grad_norm": 4.009490013122559, "learning_rate": 8.749494949494949e-07, "loss": 0.652, "step": 1340 }, { "epoch": 0.48, "grad_norm": 3.8058555126190186, "learning_rate": 8.739393939393938e-07, "loss": 0.5743, "step": 1350 }, { "epoch": 0.49, "grad_norm": 8.637187957763672, "learning_rate": 8.729292929292929e-07, "loss": 0.6198, "step": 1360 }, { "epoch": 0.49, "grad_norm": 5.059712886810303, "learning_rate": 8.719191919191918e-07, "loss": 0.6171, "step": 1370 }, { "epoch": 0.49, "grad_norm": 5.403115272521973, "learning_rate": 8.709090909090909e-07, "loss": 0.5956, "step": 1380 }, { "epoch": 0.5, "grad_norm": 4.084959983825684, "learning_rate": 8.698989898989899e-07, "loss": 0.6023, "step": 1390 }, { "epoch": 0.5, "grad_norm": 4.863948345184326, "learning_rate": 8.688888888888889e-07, "loss": 0.6295, "step": 1400 }, { "epoch": 0.51, "grad_norm": 5.363975524902344, "learning_rate": 8.678787878787879e-07, "loss": 0.5854, "step": 1410 }, { "epoch": 0.51, "grad_norm": 4.450655460357666, "learning_rate": 8.668686868686868e-07, "loss": 0.6579, "step": 1420 }, { "epoch": 0.51, "grad_norm": 5.538711071014404, "learning_rate": 8.658585858585859e-07, "loss": 0.726, "step": 1430 }, { "epoch": 0.52, "grad_norm": 3.3089439868927, "learning_rate": 8.648484848484848e-07, "loss": 0.6435, "step": 1440 }, { "epoch": 0.52, "grad_norm": 3.613719940185547, "learning_rate": 8.638383838383838e-07, "loss": 0.6695, "step": 1450 }, { "epoch": 0.52, "grad_norm": 2.9638564586639404, "learning_rate": 8.628282828282828e-07, "loss": 0.5516, "step": 1460 }, { "epoch": 0.53, "grad_norm": 3.8651678562164307, "learning_rate": 8.618181818181818e-07, "loss": 0.5883, "step": 1470 }, { "epoch": 0.53, "grad_norm": 8.56436824798584, "learning_rate": 8.608080808080807e-07, "loss": 0.6476, "step": 1480 }, { "epoch": 0.53, "grad_norm": 6.782663345336914, "learning_rate": 8.597979797979798e-07, "loss": 0.6369, "step": 1490 }, { "epoch": 0.54, "grad_norm": 3.181833267211914, "learning_rate": 8.587878787878787e-07, "loss": 0.5885, "step": 1500 }, { "epoch": 0.54, "grad_norm": 5.068531513214111, "learning_rate": 8.577777777777777e-07, "loss": 0.6241, "step": 1510 }, { "epoch": 0.55, "grad_norm": 5.7847900390625, "learning_rate": 8.567676767676767e-07, "loss": 0.623, "step": 1520 }, { "epoch": 0.55, "grad_norm": 3.325660228729248, "learning_rate": 8.557575757575757e-07, "loss": 0.574, "step": 1530 }, { "epoch": 0.55, "grad_norm": 3.452211618423462, "learning_rate": 8.547474747474747e-07, "loss": 0.6359, "step": 1540 }, { "epoch": 0.56, "grad_norm": 5.663365364074707, "learning_rate": 8.537373737373736e-07, "loss": 0.6132, "step": 1550 }, { "epoch": 0.56, "grad_norm": 3.970423460006714, "learning_rate": 8.527272727272727e-07, "loss": 0.5644, "step": 1560 }, { "epoch": 0.56, "grad_norm": 4.623054027557373, "learning_rate": 8.517171717171717e-07, "loss": 0.5753, "step": 1570 }, { "epoch": 0.57, "grad_norm": 4.318519592285156, "learning_rate": 8.507070707070707e-07, "loss": 0.6335, "step": 1580 }, { "epoch": 0.57, "grad_norm": 3.2903385162353516, "learning_rate": 8.496969696969697e-07, "loss": 0.5899, "step": 1590 }, { "epoch": 0.57, "grad_norm": 2.847003221511841, "learning_rate": 8.486868686868687e-07, "loss": 0.5834, "step": 1600 }, { "epoch": 0.58, "grad_norm": 5.138712406158447, "learning_rate": 8.476767676767676e-07, "loss": 0.5485, "step": 1610 }, { "epoch": 0.58, "grad_norm": 5.145106792449951, "learning_rate": 8.466666666666667e-07, "loss": 0.6854, "step": 1620 }, { "epoch": 0.58, "grad_norm": 5.585366725921631, "learning_rate": 8.456565656565656e-07, "loss": 0.5595, "step": 1630 }, { "epoch": 0.59, "grad_norm": 3.857548952102661, "learning_rate": 8.446464646464646e-07, "loss": 0.5856, "step": 1640 }, { "epoch": 0.59, "grad_norm": 3.835942029953003, "learning_rate": 8.436363636363636e-07, "loss": 0.5589, "step": 1650 }, { "epoch": 0.6, "grad_norm": 4.407024383544922, "learning_rate": 8.426262626262626e-07, "loss": 0.5358, "step": 1660 }, { "epoch": 0.6, "grad_norm": 4.5201826095581055, "learning_rate": 8.416161616161616e-07, "loss": 0.6021, "step": 1670 }, { "epoch": 0.6, "grad_norm": 4.167496681213379, "learning_rate": 8.406060606060606e-07, "loss": 0.6007, "step": 1680 }, { "epoch": 0.61, "grad_norm": 5.0080695152282715, "learning_rate": 8.395959595959595e-07, "loss": 0.6298, "step": 1690 }, { "epoch": 0.61, "grad_norm": 4.7244977951049805, "learning_rate": 8.385858585858585e-07, "loss": 0.5713, "step": 1700 }, { "epoch": 0.61, "grad_norm": 5.417051315307617, "learning_rate": 8.375757575757575e-07, "loss": 0.6565, "step": 1710 }, { "epoch": 0.62, "grad_norm": 6.2620391845703125, "learning_rate": 8.365656565656564e-07, "loss": 0.6977, "step": 1720 }, { "epoch": 0.62, "grad_norm": 5.629709720611572, "learning_rate": 8.355555555555556e-07, "loss": 0.6281, "step": 1730 }, { "epoch": 0.62, "grad_norm": 4.182885646820068, "learning_rate": 8.345454545454545e-07, "loss": 0.6431, "step": 1740 }, { "epoch": 0.63, "grad_norm": 3.080979824066162, "learning_rate": 8.335353535353536e-07, "loss": 0.5938, "step": 1750 }, { "epoch": 0.63, "grad_norm": 2.8959693908691406, "learning_rate": 8.325252525252525e-07, "loss": 0.6677, "step": 1760 }, { "epoch": 0.63, "grad_norm": 2.9930922985076904, "learning_rate": 8.315151515151515e-07, "loss": 0.5425, "step": 1770 }, { "epoch": 0.64, "grad_norm": 11.59288501739502, "learning_rate": 8.305050505050505e-07, "loss": 0.5816, "step": 1780 }, { "epoch": 0.64, "grad_norm": 6.010501384735107, "learning_rate": 8.294949494949495e-07, "loss": 0.5944, "step": 1790 }, { "epoch": 0.65, "grad_norm": 3.7763831615448, "learning_rate": 8.284848484848484e-07, "loss": 0.6465, "step": 1800 }, { "epoch": 0.65, "grad_norm": 4.849707126617432, "learning_rate": 8.274747474747475e-07, "loss": 0.5679, "step": 1810 }, { "epoch": 0.65, "grad_norm": 3.6395959854125977, "learning_rate": 8.264646464646464e-07, "loss": 0.6026, "step": 1820 }, { "epoch": 0.66, "grad_norm": 4.717153549194336, "learning_rate": 8.254545454545455e-07, "loss": 0.5637, "step": 1830 }, { "epoch": 0.66, "grad_norm": 3.236311674118042, "learning_rate": 8.244444444444444e-07, "loss": 0.5955, "step": 1840 }, { "epoch": 0.66, "grad_norm": 4.533290386199951, "learning_rate": 8.234343434343433e-07, "loss": 0.5681, "step": 1850 }, { "epoch": 0.67, "grad_norm": 3.0948798656463623, "learning_rate": 8.224242424242424e-07, "loss": 0.6272, "step": 1860 }, { "epoch": 0.67, "grad_norm": 3.719708204269409, "learning_rate": 8.214141414141413e-07, "loss": 0.6793, "step": 1870 }, { "epoch": 0.67, "grad_norm": 3.228376865386963, "learning_rate": 8.204040404040403e-07, "loss": 0.558, "step": 1880 }, { "epoch": 0.68, "grad_norm": 3.4296934604644775, "learning_rate": 8.193939393939393e-07, "loss": 0.5717, "step": 1890 }, { "epoch": 0.68, "grad_norm": 5.708250999450684, "learning_rate": 8.184848484848484e-07, "loss": 0.6381, "step": 1900 }, { "epoch": 0.68, "grad_norm": 3.8306288719177246, "learning_rate": 8.174747474747474e-07, "loss": 0.6083, "step": 1910 }, { "epoch": 0.69, "grad_norm": 4.5576252937316895, "learning_rate": 8.164646464646464e-07, "loss": 0.6676, "step": 1920 }, { "epoch": 0.69, "grad_norm": 3.2234649658203125, "learning_rate": 8.154545454545454e-07, "loss": 0.6215, "step": 1930 }, { "epoch": 0.7, "grad_norm": 3.4826815128326416, "learning_rate": 8.144444444444444e-07, "loss": 0.6827, "step": 1940 }, { "epoch": 0.7, "grad_norm": 3.9034299850463867, "learning_rate": 8.134343434343433e-07, "loss": 0.596, "step": 1950 }, { "epoch": 0.7, "grad_norm": 2.7787563800811768, "learning_rate": 8.124242424242424e-07, "loss": 0.6234, "step": 1960 }, { "epoch": 0.71, "grad_norm": 4.442075252532959, "learning_rate": 8.114141414141413e-07, "loss": 0.6244, "step": 1970 }, { "epoch": 0.71, "grad_norm": 3.8511252403259277, "learning_rate": 8.104040404040403e-07, "loss": 0.5721, "step": 1980 }, { "epoch": 0.71, "grad_norm": 3.925318717956543, "learning_rate": 8.093939393939394e-07, "loss": 0.6547, "step": 1990 }, { "epoch": 0.72, "grad_norm": 4.262260437011719, "learning_rate": 8.083838383838384e-07, "loss": 0.6155, "step": 2000 }, { "epoch": 0.72, "eval_loss": 0.7234830260276794, "eval_runtime": 401.6534, "eval_samples_per_second": 2.49, "eval_steps_per_second": 2.49, "step": 2000 }, { "epoch": 0.72, "grad_norm": 4.890273571014404, "learning_rate": 8.073737373737374e-07, "loss": 0.6392, "step": 2010 }, { "epoch": 0.72, "grad_norm": 4.918805122375488, "learning_rate": 8.063636363636364e-07, "loss": 0.5549, "step": 2020 }, { "epoch": 0.73, "grad_norm": 3.605008840560913, "learning_rate": 8.053535353535353e-07, "loss": 0.5152, "step": 2030 }, { "epoch": 0.73, "grad_norm": 4.606666088104248, "learning_rate": 8.043434343434344e-07, "loss": 0.4798, "step": 2040 }, { "epoch": 0.74, "grad_norm": 4.115630626678467, "learning_rate": 8.033333333333333e-07, "loss": 0.5856, "step": 2050 }, { "epoch": 0.74, "grad_norm": 4.37373685836792, "learning_rate": 8.023232323232322e-07, "loss": 0.6467, "step": 2060 }, { "epoch": 0.74, "grad_norm": 3.3623104095458984, "learning_rate": 8.013131313131313e-07, "loss": 0.6591, "step": 2070 }, { "epoch": 0.75, "grad_norm": 3.1440579891204834, "learning_rate": 8.003030303030302e-07, "loss": 0.6003, "step": 2080 }, { "epoch": 0.75, "grad_norm": 3.391533136367798, "learning_rate": 7.992929292929293e-07, "loss": 0.5872, "step": 2090 }, { "epoch": 0.75, "grad_norm": 4.677767276763916, "learning_rate": 7.982828282828282e-07, "loss": 0.5803, "step": 2100 }, { "epoch": 0.76, "grad_norm": 3.759061336517334, "learning_rate": 7.972727272727272e-07, "loss": 0.6155, "step": 2110 }, { "epoch": 0.76, "grad_norm": 3.419386863708496, "learning_rate": 7.962626262626262e-07, "loss": 0.6303, "step": 2120 }, { "epoch": 0.76, "grad_norm": 3.4935250282287598, "learning_rate": 7.952525252525252e-07, "loss": 0.6348, "step": 2130 }, { "epoch": 0.77, "grad_norm": 4.793978214263916, "learning_rate": 7.942424242424241e-07, "loss": 0.5004, "step": 2140 }, { "epoch": 0.77, "grad_norm": 5.078155994415283, "learning_rate": 7.932323232323232e-07, "loss": 0.5583, "step": 2150 }, { "epoch": 0.77, "grad_norm": 3.8855032920837402, "learning_rate": 7.922222222222222e-07, "loss": 0.626, "step": 2160 }, { "epoch": 0.78, "grad_norm": 4.782063007354736, "learning_rate": 7.912121212121213e-07, "loss": 0.4847, "step": 2170 }, { "epoch": 0.78, "grad_norm": 4.760573387145996, "learning_rate": 7.902020202020202e-07, "loss": 0.5923, "step": 2180 }, { "epoch": 0.79, "grad_norm": 5.196963310241699, "learning_rate": 7.891919191919192e-07, "loss": 0.5216, "step": 2190 }, { "epoch": 0.79, "grad_norm": 5.447847843170166, "learning_rate": 7.881818181818182e-07, "loss": 0.6144, "step": 2200 }, { "epoch": 0.79, "grad_norm": 5.5991129875183105, "learning_rate": 7.871717171717171e-07, "loss": 0.5247, "step": 2210 }, { "epoch": 0.8, "grad_norm": 4.1100921630859375, "learning_rate": 7.861616161616161e-07, "loss": 0.6581, "step": 2220 }, { "epoch": 0.8, "grad_norm": 2.998598098754883, "learning_rate": 7.851515151515151e-07, "loss": 0.5111, "step": 2230 }, { "epoch": 0.8, "grad_norm": 3.1241579055786133, "learning_rate": 7.841414141414141e-07, "loss": 0.626, "step": 2240 }, { "epoch": 0.81, "grad_norm": 3.804655075073242, "learning_rate": 7.831313131313131e-07, "loss": 0.596, "step": 2250 }, { "epoch": 0.81, "grad_norm": 3.636692523956299, "learning_rate": 7.821212121212121e-07, "loss": 0.598, "step": 2260 }, { "epoch": 0.81, "grad_norm": 5.374478816986084, "learning_rate": 7.81111111111111e-07, "loss": 0.5155, "step": 2270 }, { "epoch": 0.82, "grad_norm": 4.611242294311523, "learning_rate": 7.801010101010101e-07, "loss": 0.5482, "step": 2280 }, { "epoch": 0.82, "grad_norm": 6.077275276184082, "learning_rate": 7.79090909090909e-07, "loss": 0.5642, "step": 2290 }, { "epoch": 0.82, "grad_norm": 4.110134601593018, "learning_rate": 7.78080808080808e-07, "loss": 0.6482, "step": 2300 }, { "epoch": 0.83, "grad_norm": 3.603376865386963, "learning_rate": 7.77070707070707e-07, "loss": 0.5755, "step": 2310 }, { "epoch": 0.83, "grad_norm": 4.059467315673828, "learning_rate": 7.76060606060606e-07, "loss": 0.5205, "step": 2320 }, { "epoch": 0.84, "grad_norm": 4.656579971313477, "learning_rate": 7.750505050505051e-07, "loss": 0.5975, "step": 2330 }, { "epoch": 0.84, "grad_norm": 4.863138675689697, "learning_rate": 7.74040404040404e-07, "loss": 0.5375, "step": 2340 }, { "epoch": 0.84, "grad_norm": 4.3313517570495605, "learning_rate": 7.73030303030303e-07, "loss": 0.5795, "step": 2350 }, { "epoch": 0.85, "grad_norm": 3.870875120162964, "learning_rate": 7.72020202020202e-07, "loss": 0.5249, "step": 2360 }, { "epoch": 0.85, "grad_norm": 3.720283031463623, "learning_rate": 7.71010101010101e-07, "loss": 0.6733, "step": 2370 }, { "epoch": 0.85, "grad_norm": 3.779296398162842, "learning_rate": 7.699999999999999e-07, "loss": 0.588, "step": 2380 }, { "epoch": 0.86, "grad_norm": 4.588533401489258, "learning_rate": 7.68989898989899e-07, "loss": 0.6044, "step": 2390 }, { "epoch": 0.86, "grad_norm": 5.907507419586182, "learning_rate": 7.679797979797979e-07, "loss": 0.5715, "step": 2400 }, { "epoch": 0.86, "grad_norm": 5.705809116363525, "learning_rate": 7.66969696969697e-07, "loss": 0.5486, "step": 2410 }, { "epoch": 0.87, "grad_norm": 3.8104190826416016, "learning_rate": 7.659595959595959e-07, "loss": 0.602, "step": 2420 }, { "epoch": 0.87, "grad_norm": 5.82910680770874, "learning_rate": 7.649494949494949e-07, "loss": 0.5111, "step": 2430 }, { "epoch": 0.88, "grad_norm": 5.837489604949951, "learning_rate": 7.639393939393939e-07, "loss": 0.6145, "step": 2440 }, { "epoch": 0.88, "grad_norm": 3.5505547523498535, "learning_rate": 7.629292929292929e-07, "loss": 0.5834, "step": 2450 }, { "epoch": 0.88, "grad_norm": 4.314844131469727, "learning_rate": 7.619191919191918e-07, "loss": 0.6152, "step": 2460 }, { "epoch": 0.89, "grad_norm": 4.205423831939697, "learning_rate": 7.609090909090909e-07, "loss": 0.6285, "step": 2470 }, { "epoch": 0.89, "grad_norm": 5.292863368988037, "learning_rate": 7.598989898989898e-07, "loss": 0.6531, "step": 2480 }, { "epoch": 0.89, "grad_norm": 5.13388204574585, "learning_rate": 7.588888888888888e-07, "loss": 0.628, "step": 2490 }, { "epoch": 0.9, "grad_norm": 2.864743232727051, "learning_rate": 7.578787878787879e-07, "loss": 0.5403, "step": 2500 }, { "epoch": 0.9, "grad_norm": 4.284875869750977, "learning_rate": 7.568686868686868e-07, "loss": 0.6097, "step": 2510 }, { "epoch": 0.9, "grad_norm": 3.3840548992156982, "learning_rate": 7.558585858585859e-07, "loss": 0.5759, "step": 2520 }, { "epoch": 0.91, "grad_norm": 4.531360626220703, "learning_rate": 7.548484848484848e-07, "loss": 0.5581, "step": 2530 }, { "epoch": 0.91, "grad_norm": 5.861146926879883, "learning_rate": 7.538383838383838e-07, "loss": 0.6329, "step": 2540 }, { "epoch": 0.91, "grad_norm": 5.117300987243652, "learning_rate": 7.528282828282828e-07, "loss": 0.6071, "step": 2550 }, { "epoch": 0.92, "grad_norm": 3.9243366718292236, "learning_rate": 7.518181818181818e-07, "loss": 0.5476, "step": 2560 }, { "epoch": 0.92, "grad_norm": 6.215851783752441, "learning_rate": 7.508080808080808e-07, "loss": 0.5731, "step": 2570 }, { "epoch": 0.93, "grad_norm": 6.189659118652344, "learning_rate": 7.497979797979798e-07, "loss": 0.592, "step": 2580 }, { "epoch": 0.93, "grad_norm": 2.877923011779785, "learning_rate": 7.487878787878787e-07, "loss": 0.5435, "step": 2590 }, { "epoch": 0.93, "grad_norm": 5.963223934173584, "learning_rate": 7.477777777777778e-07, "loss": 0.6104, "step": 2600 }, { "epoch": 0.94, "grad_norm": 4.374766826629639, "learning_rate": 7.467676767676767e-07, "loss": 0.6022, "step": 2610 }, { "epoch": 0.94, "grad_norm": 4.16854190826416, "learning_rate": 7.457575757575756e-07, "loss": 0.4931, "step": 2620 }, { "epoch": 0.94, "grad_norm": 7.225739479064941, "learning_rate": 7.447474747474747e-07, "loss": 0.5659, "step": 2630 }, { "epoch": 0.95, "grad_norm": 4.522436141967773, "learning_rate": 7.437373737373736e-07, "loss": 0.5449, "step": 2640 }, { "epoch": 0.95, "grad_norm": 4.935425758361816, "learning_rate": 7.427272727272727e-07, "loss": 0.6093, "step": 2650 }, { "epoch": 0.95, "grad_norm": 3.947935104370117, "learning_rate": 7.417171717171716e-07, "loss": 0.5976, "step": 2660 }, { "epoch": 0.96, "grad_norm": 5.328210830688477, "learning_rate": 7.407070707070707e-07, "loss": 0.6283, "step": 2670 }, { "epoch": 0.96, "grad_norm": 3.0664663314819336, "learning_rate": 7.396969696969697e-07, "loss": 0.5853, "step": 2680 }, { "epoch": 0.96, "grad_norm": 5.055959701538086, "learning_rate": 7.386868686868687e-07, "loss": 0.5767, "step": 2690 }, { "epoch": 0.97, "grad_norm": 5.318157196044922, "learning_rate": 7.376767676767676e-07, "loss": 0.5319, "step": 2700 }, { "epoch": 0.97, "grad_norm": 3.5820364952087402, "learning_rate": 7.366666666666667e-07, "loss": 0.5506, "step": 2710 }, { "epoch": 0.98, "grad_norm": 4.367839813232422, "learning_rate": 7.356565656565656e-07, "loss": 0.5524, "step": 2720 }, { "epoch": 0.98, "grad_norm": 4.305497169494629, "learning_rate": 7.346464646464647e-07, "loss": 0.5639, "step": 2730 }, { "epoch": 0.98, "grad_norm": 6.55611515045166, "learning_rate": 7.336363636363636e-07, "loss": 0.6346, "step": 2740 }, { "epoch": 0.99, "grad_norm": 9.09347152709961, "learning_rate": 7.326262626262626e-07, "loss": 0.5574, "step": 2750 }, { "epoch": 0.99, "grad_norm": 3.9730687141418457, "learning_rate": 7.316161616161616e-07, "loss": 0.5908, "step": 2760 }, { "epoch": 0.99, "grad_norm": 16.01730728149414, "learning_rate": 7.306060606060605e-07, "loss": 0.5936, "step": 2770 }, { "epoch": 1.0, "grad_norm": 5.077668190002441, "learning_rate": 7.295959595959595e-07, "loss": 0.5556, "step": 2780 }, { "epoch": 1.0, "grad_norm": 4.859917640686035, "learning_rate": 7.285858585858585e-07, "loss": 0.5452, "step": 2790 }, { "epoch": 1.0, "grad_norm": 5.687443733215332, "learning_rate": 7.275757575757575e-07, "loss": 0.6583, "step": 2800 }, { "epoch": 1.01, "grad_norm": 4.369338035583496, "learning_rate": 7.265656565656565e-07, "loss": 0.5623, "step": 2810 }, { "epoch": 1.01, "grad_norm": 4.035587310791016, "learning_rate": 7.255555555555555e-07, "loss": 0.571, "step": 2820 }, { "epoch": 1.01, "grad_norm": 4.2013139724731445, "learning_rate": 7.245454545454544e-07, "loss": 0.4768, "step": 2830 }, { "epoch": 1.02, "grad_norm": 3.272221565246582, "learning_rate": 7.235353535353536e-07, "loss": 0.581, "step": 2840 }, { "epoch": 1.02, "grad_norm": 3.97727370262146, "learning_rate": 7.225252525252525e-07, "loss": 0.5654, "step": 2850 }, { "epoch": 1.03, "grad_norm": 4.043779373168945, "learning_rate": 7.215151515151516e-07, "loss": 0.5971, "step": 2860 }, { "epoch": 1.03, "grad_norm": 4.015261173248291, "learning_rate": 7.205050505050505e-07, "loss": 0.5921, "step": 2870 }, { "epoch": 1.03, "grad_norm": 4.540653705596924, "learning_rate": 7.194949494949495e-07, "loss": 0.6103, "step": 2880 }, { "epoch": 1.04, "grad_norm": 4.258879661560059, "learning_rate": 7.184848484848485e-07, "loss": 0.5519, "step": 2890 }, { "epoch": 1.04, "grad_norm": 7.0088629722595215, "learning_rate": 7.174747474747475e-07, "loss": 0.6054, "step": 2900 }, { "epoch": 1.04, "grad_norm": 3.9803833961486816, "learning_rate": 7.164646464646464e-07, "loss": 0.5557, "step": 2910 }, { "epoch": 1.05, "grad_norm": 3.6390326023101807, "learning_rate": 7.154545454545454e-07, "loss": 0.5406, "step": 2920 }, { "epoch": 1.05, "grad_norm": 3.998908042907715, "learning_rate": 7.144444444444444e-07, "loss": 0.604, "step": 2930 }, { "epoch": 1.05, "grad_norm": 4.304749488830566, "learning_rate": 7.134343434343433e-07, "loss": 0.5859, "step": 2940 }, { "epoch": 1.06, "grad_norm": 4.742584228515625, "learning_rate": 7.124242424242424e-07, "loss": 0.5184, "step": 2950 }, { "epoch": 1.06, "grad_norm": 5.163829326629639, "learning_rate": 7.114141414141413e-07, "loss": 0.5669, "step": 2960 }, { "epoch": 1.07, "grad_norm": 5.633606910705566, "learning_rate": 7.104040404040404e-07, "loss": 0.5967, "step": 2970 }, { "epoch": 1.07, "grad_norm": 6.37916898727417, "learning_rate": 7.093939393939393e-07, "loss": 0.5971, "step": 2980 }, { "epoch": 1.07, "grad_norm": 2.9268598556518555, "learning_rate": 7.083838383838383e-07, "loss": 0.5342, "step": 2990 }, { "epoch": 1.08, "grad_norm": 4.331085205078125, "learning_rate": 7.073737373737373e-07, "loss": 0.5779, "step": 3000 }, { "epoch": 1.08, "eval_loss": 0.7162447571754456, "eval_runtime": 401.0241, "eval_samples_per_second": 2.494, "eval_steps_per_second": 2.494, "step": 3000 }, { "epoch": 1.08, "grad_norm": 2.9604339599609375, "learning_rate": 7.063636363636364e-07, "loss": 0.5723, "step": 3010 }, { "epoch": 1.08, "grad_norm": 4.6410393714904785, "learning_rate": 7.053535353535354e-07, "loss": 0.5986, "step": 3020 }, { "epoch": 1.09, "grad_norm": 4.530472278594971, "learning_rate": 7.043434343434344e-07, "loss": 0.705, "step": 3030 }, { "epoch": 1.09, "grad_norm": 4.028800964355469, "learning_rate": 7.033333333333333e-07, "loss": 0.5135, "step": 3040 }, { "epoch": 1.09, "grad_norm": 2.882619619369507, "learning_rate": 7.023232323232324e-07, "loss": 0.5373, "step": 3050 }, { "epoch": 1.1, "grad_norm": 5.280877590179443, "learning_rate": 7.013131313131313e-07, "loss": 0.5457, "step": 3060 }, { "epoch": 1.1, "grad_norm": 3.983971357345581, "learning_rate": 7.003030303030302e-07, "loss": 0.5678, "step": 3070 }, { "epoch": 1.1, "grad_norm": 3.4524683952331543, "learning_rate": 6.992929292929293e-07, "loss": 0.5794, "step": 3080 }, { "epoch": 1.11, "grad_norm": 4.039379119873047, "learning_rate": 6.982828282828282e-07, "loss": 0.4778, "step": 3090 }, { "epoch": 1.11, "grad_norm": 4.454843521118164, "learning_rate": 6.972727272727273e-07, "loss": 0.5312, "step": 3100 }, { "epoch": 1.12, "grad_norm": 4.90205192565918, "learning_rate": 6.962626262626262e-07, "loss": 0.5675, "step": 3110 }, { "epoch": 1.12, "grad_norm": 3.076164484024048, "learning_rate": 6.952525252525252e-07, "loss": 0.5949, "step": 3120 }, { "epoch": 1.12, "grad_norm": 3.707550287246704, "learning_rate": 6.942424242424242e-07, "loss": 0.6053, "step": 3130 }, { "epoch": 1.13, "grad_norm": 4.36667537689209, "learning_rate": 6.932323232323232e-07, "loss": 0.638, "step": 3140 }, { "epoch": 1.13, "grad_norm": 4.635988235473633, "learning_rate": 6.922222222222221e-07, "loss": 0.5698, "step": 3150 }, { "epoch": 1.13, "grad_norm": 3.769178628921509, "learning_rate": 6.912121212121212e-07, "loss": 0.5563, "step": 3160 }, { "epoch": 1.14, "grad_norm": 3.096804618835449, "learning_rate": 6.902020202020201e-07, "loss": 0.5268, "step": 3170 }, { "epoch": 1.14, "grad_norm": 5.9978461265563965, "learning_rate": 6.891919191919193e-07, "loss": 0.4878, "step": 3180 }, { "epoch": 1.14, "grad_norm": 4.963356971740723, "learning_rate": 6.881818181818182e-07, "loss": 0.5904, "step": 3190 }, { "epoch": 1.15, "grad_norm": 4.526729583740234, "learning_rate": 6.871717171717171e-07, "loss": 0.5957, "step": 3200 }, { "epoch": 1.15, "grad_norm": 3.4197323322296143, "learning_rate": 6.861616161616162e-07, "loss": 0.5522, "step": 3210 }, { "epoch": 1.15, "grad_norm": 3.989546060562134, "learning_rate": 6.851515151515151e-07, "loss": 0.6122, "step": 3220 }, { "epoch": 1.16, "grad_norm": 5.137784004211426, "learning_rate": 6.841414141414141e-07, "loss": 0.58, "step": 3230 }, { "epoch": 1.16, "grad_norm": 4.193089485168457, "learning_rate": 6.831313131313131e-07, "loss": 0.4918, "step": 3240 }, { "epoch": 1.17, "grad_norm": 4.1724162101745605, "learning_rate": 6.821212121212121e-07, "loss": 0.6044, "step": 3250 }, { "epoch": 1.17, "grad_norm": 4.983777046203613, "learning_rate": 6.811111111111111e-07, "loss": 0.533, "step": 3260 }, { "epoch": 1.17, "grad_norm": 7.2310333251953125, "learning_rate": 6.801010101010101e-07, "loss": 0.5843, "step": 3270 }, { "epoch": 1.18, "grad_norm": 4.620791912078857, "learning_rate": 6.79090909090909e-07, "loss": 0.5242, "step": 3280 }, { "epoch": 1.18, "grad_norm": 4.3235087394714355, "learning_rate": 6.780808080808081e-07, "loss": 0.5836, "step": 3290 }, { "epoch": 1.18, "grad_norm": 4.413332462310791, "learning_rate": 6.77070707070707e-07, "loss": 0.6532, "step": 3300 }, { "epoch": 1.19, "grad_norm": 4.286377429962158, "learning_rate": 6.76060606060606e-07, "loss": 0.5527, "step": 3310 }, { "epoch": 1.19, "grad_norm": 6.150529861450195, "learning_rate": 6.75050505050505e-07, "loss": 0.5745, "step": 3320 }, { "epoch": 1.19, "grad_norm": 4.088225841522217, "learning_rate": 6.74040404040404e-07, "loss": 0.5882, "step": 3330 }, { "epoch": 1.2, "grad_norm": 4.4364824295043945, "learning_rate": 6.73030303030303e-07, "loss": 0.513, "step": 3340 }, { "epoch": 1.2, "grad_norm": 3.8595519065856934, "learning_rate": 6.72020202020202e-07, "loss": 0.5621, "step": 3350 }, { "epoch": 1.2, "grad_norm": 6.183053016662598, "learning_rate": 6.71010101010101e-07, "loss": 0.5318, "step": 3360 }, { "epoch": 1.21, "grad_norm": 4.677096366882324, "learning_rate": 6.7e-07, "loss": 0.4965, "step": 3370 }, { "epoch": 1.21, "grad_norm": 5.507148742675781, "learning_rate": 6.68989898989899e-07, "loss": 0.569, "step": 3380 }, { "epoch": 1.22, "grad_norm": 3.813816547393799, "learning_rate": 6.679797979797979e-07, "loss": 0.5326, "step": 3390 }, { "epoch": 1.22, "grad_norm": 5.973514080047607, "learning_rate": 6.66969696969697e-07, "loss": 0.6733, "step": 3400 }, { "epoch": 1.22, "grad_norm": 5.510775089263916, "learning_rate": 6.659595959595959e-07, "loss": 0.5978, "step": 3410 }, { "epoch": 1.23, "grad_norm": 3.8347678184509277, "learning_rate": 6.64949494949495e-07, "loss": 0.5141, "step": 3420 }, { "epoch": 1.23, "grad_norm": 4.097416877746582, "learning_rate": 6.639393939393939e-07, "loss": 0.5503, "step": 3430 }, { "epoch": 1.23, "grad_norm": 3.52165150642395, "learning_rate": 6.629292929292929e-07, "loss": 0.5506, "step": 3440 }, { "epoch": 1.24, "grad_norm": 4.4569573402404785, "learning_rate": 6.619191919191919e-07, "loss": 0.6121, "step": 3450 }, { "epoch": 1.24, "grad_norm": 4.094751834869385, "learning_rate": 6.609090909090909e-07, "loss": 0.5562, "step": 3460 }, { "epoch": 1.24, "grad_norm": 4.277998447418213, "learning_rate": 6.598989898989898e-07, "loss": 0.5432, "step": 3470 }, { "epoch": 1.25, "grad_norm": 5.9544148445129395, "learning_rate": 6.588888888888889e-07, "loss": 0.5891, "step": 3480 }, { "epoch": 1.25, "grad_norm": 5.688138008117676, "learning_rate": 6.578787878787878e-07, "loss": 0.6172, "step": 3490 }, { "epoch": 1.26, "grad_norm": 5.771819591522217, "learning_rate": 6.568686868686868e-07, "loss": 0.5362, "step": 3500 }, { "epoch": 1.26, "grad_norm": 4.87513542175293, "learning_rate": 6.558585858585858e-07, "loss": 0.5608, "step": 3510 }, { "epoch": 1.26, "grad_norm": 3.709472179412842, "learning_rate": 6.548484848484848e-07, "loss": 0.5361, "step": 3520 }, { "epoch": 1.27, "grad_norm": 4.465284824371338, "learning_rate": 6.538383838383839e-07, "loss": 0.5435, "step": 3530 }, { "epoch": 1.27, "grad_norm": 4.30505895614624, "learning_rate": 6.528282828282828e-07, "loss": 0.5714, "step": 3540 }, { "epoch": 1.27, "grad_norm": 4.445021152496338, "learning_rate": 6.518181818181818e-07, "loss": 0.5773, "step": 3550 }, { "epoch": 1.28, "grad_norm": 4.173781394958496, "learning_rate": 6.508080808080808e-07, "loss": 0.6223, "step": 3560 }, { "epoch": 1.28, "grad_norm": 5.676968574523926, "learning_rate": 6.497979797979798e-07, "loss": 0.5957, "step": 3570 }, { "epoch": 1.28, "grad_norm": 3.447176694869995, "learning_rate": 6.487878787878788e-07, "loss": 0.5896, "step": 3580 }, { "epoch": 1.29, "grad_norm": 4.1883111000061035, "learning_rate": 6.477777777777778e-07, "loss": 0.5602, "step": 3590 }, { "epoch": 1.29, "grad_norm": 4.16115140914917, "learning_rate": 6.467676767676767e-07, "loss": 0.5752, "step": 3600 }, { "epoch": 1.29, "grad_norm": 4.55715274810791, "learning_rate": 6.457575757575758e-07, "loss": 0.5344, "step": 3610 }, { "epoch": 1.3, "grad_norm": 5.83474588394165, "learning_rate": 6.447474747474747e-07, "loss": 0.5867, "step": 3620 }, { "epoch": 1.3, "grad_norm": 2.700615644454956, "learning_rate": 6.437373737373736e-07, "loss": 0.5395, "step": 3630 }, { "epoch": 1.31, "grad_norm": 4.691040992736816, "learning_rate": 6.428282828282828e-07, "loss": 0.5509, "step": 3640 }, { "epoch": 1.31, "grad_norm": 4.966342449188232, "learning_rate": 6.418181818181818e-07, "loss": 0.6047, "step": 3650 }, { "epoch": 1.31, "grad_norm": 3.6059834957122803, "learning_rate": 6.408080808080808e-07, "loss": 0.5338, "step": 3660 }, { "epoch": 1.32, "grad_norm": 3.471660852432251, "learning_rate": 6.397979797979798e-07, "loss": 0.5532, "step": 3670 }, { "epoch": 1.32, "grad_norm": 5.410092353820801, "learning_rate": 6.387878787878788e-07, "loss": 0.5739, "step": 3680 }, { "epoch": 1.32, "grad_norm": 4.829250335693359, "learning_rate": 6.377777777777778e-07, "loss": 0.6064, "step": 3690 }, { "epoch": 1.33, "grad_norm": 5.880673408508301, "learning_rate": 6.367676767676767e-07, "loss": 0.575, "step": 3700 }, { "epoch": 1.33, "grad_norm": 3.647641658782959, "learning_rate": 6.357575757575757e-07, "loss": 0.594, "step": 3710 }, { "epoch": 1.33, "grad_norm": 4.267665386199951, "learning_rate": 6.347474747474747e-07, "loss": 0.574, "step": 3720 }, { "epoch": 1.34, "grad_norm": 3.5303924083709717, "learning_rate": 6.337373737373736e-07, "loss": 0.496, "step": 3730 }, { "epoch": 1.34, "grad_norm": 3.1564202308654785, "learning_rate": 6.327272727272727e-07, "loss": 0.5196, "step": 3740 }, { "epoch": 1.34, "grad_norm": 4.458827495574951, "learning_rate": 6.317171717171716e-07, "loss": 0.5616, "step": 3750 }, { "epoch": 1.35, "grad_norm": 4.057519435882568, "learning_rate": 6.307070707070707e-07, "loss": 0.5293, "step": 3760 }, { "epoch": 1.35, "grad_norm": 5.7592363357543945, "learning_rate": 6.296969696969696e-07, "loss": 0.5433, "step": 3770 }, { "epoch": 1.36, "grad_norm": 5.579089641571045, "learning_rate": 6.286868686868687e-07, "loss": 0.546, "step": 3780 }, { "epoch": 1.36, "grad_norm": 3.8911049365997314, "learning_rate": 6.276767676767677e-07, "loss": 0.5991, "step": 3790 }, { "epoch": 1.36, "grad_norm": 6.364475727081299, "learning_rate": 6.266666666666667e-07, "loss": 0.5908, "step": 3800 }, { "epoch": 1.37, "grad_norm": 5.0971760749816895, "learning_rate": 6.256565656565656e-07, "loss": 0.5915, "step": 3810 }, { "epoch": 1.37, "grad_norm": 3.897958755493164, "learning_rate": 6.246464646464647e-07, "loss": 0.5859, "step": 3820 }, { "epoch": 1.37, "grad_norm": 6.308644771575928, "learning_rate": 6.236363636363636e-07, "loss": 0.5991, "step": 3830 }, { "epoch": 1.38, "grad_norm": 3.0512614250183105, "learning_rate": 6.226262626262627e-07, "loss": 0.5742, "step": 3840 }, { "epoch": 1.38, "grad_norm": 4.052495956420898, "learning_rate": 6.216161616161616e-07, "loss": 0.54, "step": 3850 }, { "epoch": 1.38, "grad_norm": 4.723029613494873, "learning_rate": 6.206060606060605e-07, "loss": 0.5524, "step": 3860 }, { "epoch": 1.39, "grad_norm": 4.705147743225098, "learning_rate": 6.195959595959596e-07, "loss": 0.5924, "step": 3870 }, { "epoch": 1.39, "grad_norm": 4.098758697509766, "learning_rate": 6.185858585858585e-07, "loss": 0.5482, "step": 3880 }, { "epoch": 1.4, "grad_norm": 2.9597623348236084, "learning_rate": 6.175757575757575e-07, "loss": 0.5521, "step": 3890 }, { "epoch": 1.4, "grad_norm": 3.7308359146118164, "learning_rate": 6.165656565656565e-07, "loss": 0.5645, "step": 3900 }, { "epoch": 1.4, "grad_norm": 4.260051250457764, "learning_rate": 6.155555555555555e-07, "loss": 0.5471, "step": 3910 }, { "epoch": 1.41, "grad_norm": 6.539883136749268, "learning_rate": 6.145454545454545e-07, "loss": 0.5439, "step": 3920 }, { "epoch": 1.41, "grad_norm": 6.235405921936035, "learning_rate": 6.135353535353535e-07, "loss": 0.5435, "step": 3930 }, { "epoch": 1.41, "grad_norm": 3.3652477264404297, "learning_rate": 6.125252525252524e-07, "loss": 0.5544, "step": 3940 }, { "epoch": 1.42, "grad_norm": 3.795522928237915, "learning_rate": 6.115151515151516e-07, "loss": 0.5794, "step": 3950 }, { "epoch": 1.42, "grad_norm": 5.496920108795166, "learning_rate": 6.105050505050505e-07, "loss": 0.5133, "step": 3960 }, { "epoch": 1.42, "grad_norm": 5.946590423583984, "learning_rate": 6.094949494949495e-07, "loss": 0.6391, "step": 3970 }, { "epoch": 1.43, "grad_norm": 3.943232774734497, "learning_rate": 6.084848484848485e-07, "loss": 0.5574, "step": 3980 }, { "epoch": 1.43, "grad_norm": 4.132944107055664, "learning_rate": 6.074747474747474e-07, "loss": 0.5014, "step": 3990 }, { "epoch": 1.43, "grad_norm": 4.284780979156494, "learning_rate": 6.064646464646465e-07, "loss": 0.5907, "step": 4000 }, { "epoch": 1.43, "eval_loss": 0.698131263256073, "eval_runtime": 401.1999, "eval_samples_per_second": 2.493, "eval_steps_per_second": 2.493, "step": 4000 }, { "epoch": 1.44, "grad_norm": 5.0691447257995605, "learning_rate": 6.054545454545454e-07, "loss": 0.5153, "step": 4010 }, { "epoch": 1.44, "grad_norm": 4.563512325286865, "learning_rate": 6.044444444444444e-07, "loss": 0.5323, "step": 4020 }, { "epoch": 1.45, "grad_norm": 8.346761703491211, "learning_rate": 6.034343434343434e-07, "loss": 0.6202, "step": 4030 }, { "epoch": 1.45, "grad_norm": 4.459041118621826, "learning_rate": 6.024242424242424e-07, "loss": 0.5081, "step": 4040 }, { "epoch": 1.45, "grad_norm": 6.5770158767700195, "learning_rate": 6.014141414141413e-07, "loss": 0.568, "step": 4050 }, { "epoch": 1.46, "grad_norm": 5.020130634307861, "learning_rate": 6.004040404040404e-07, "loss": 0.5027, "step": 4060 }, { "epoch": 1.46, "grad_norm": 6.139995098114014, "learning_rate": 5.993939393939393e-07, "loss": 0.5115, "step": 4070 }, { "epoch": 1.46, "grad_norm": 5.601007461547852, "learning_rate": 5.983838383838384e-07, "loss": 0.5331, "step": 4080 }, { "epoch": 1.47, "grad_norm": 4.264581680297852, "learning_rate": 5.973737373737373e-07, "loss": 0.5474, "step": 4090 }, { "epoch": 1.47, "grad_norm": 3.254601001739502, "learning_rate": 5.963636363636363e-07, "loss": 0.5515, "step": 4100 }, { "epoch": 1.47, "grad_norm": 4.111956596374512, "learning_rate": 5.953535353535353e-07, "loss": 0.5486, "step": 4110 }, { "epoch": 1.48, "grad_norm": 3.814635992050171, "learning_rate": 5.943434343434344e-07, "loss": 0.5332, "step": 4120 }, { "epoch": 1.48, "grad_norm": 4.161899089813232, "learning_rate": 5.933333333333334e-07, "loss": 0.5713, "step": 4130 }, { "epoch": 1.48, "grad_norm": 3.490006446838379, "learning_rate": 5.923232323232323e-07, "loss": 0.5522, "step": 4140 }, { "epoch": 1.49, "grad_norm": 4.460766792297363, "learning_rate": 5.913131313131313e-07, "loss": 0.515, "step": 4150 }, { "epoch": 1.49, "grad_norm": 5.571935176849365, "learning_rate": 5.903030303030303e-07, "loss": 0.5976, "step": 4160 }, { "epoch": 1.5, "grad_norm": 4.256524562835693, "learning_rate": 5.892929292929293e-07, "loss": 0.5068, "step": 4170 }, { "epoch": 1.5, "grad_norm": 5.909682273864746, "learning_rate": 5.882828282828282e-07, "loss": 0.5473, "step": 4180 }, { "epoch": 1.5, "grad_norm": 6.469602584838867, "learning_rate": 5.872727272727273e-07, "loss": 0.5715, "step": 4190 }, { "epoch": 1.51, "grad_norm": 4.836788654327393, "learning_rate": 5.862626262626262e-07, "loss": 0.4511, "step": 4200 }, { "epoch": 1.51, "grad_norm": 2.996185302734375, "learning_rate": 5.852525252525252e-07, "loss": 0.5782, "step": 4210 }, { "epoch": 1.51, "grad_norm": 8.92146110534668, "learning_rate": 5.842424242424242e-07, "loss": 0.5329, "step": 4220 }, { "epoch": 1.52, "grad_norm": 6.146700382232666, "learning_rate": 5.832323232323232e-07, "loss": 0.5768, "step": 4230 }, { "epoch": 1.52, "grad_norm": 4.365715026855469, "learning_rate": 5.822222222222222e-07, "loss": 0.5804, "step": 4240 }, { "epoch": 1.52, "grad_norm": 4.362324237823486, "learning_rate": 5.812121212121212e-07, "loss": 0.5435, "step": 4250 }, { "epoch": 1.53, "grad_norm": 4.051389217376709, "learning_rate": 5.802020202020201e-07, "loss": 0.5662, "step": 4260 }, { "epoch": 1.53, "grad_norm": 4.774472236633301, "learning_rate": 5.791919191919192e-07, "loss": 0.523, "step": 4270 }, { "epoch": 1.53, "grad_norm": 4.41732120513916, "learning_rate": 5.781818181818181e-07, "loss": 0.5926, "step": 4280 }, { "epoch": 1.54, "grad_norm": 5.613069534301758, "learning_rate": 5.771717171717172e-07, "loss": 0.5205, "step": 4290 }, { "epoch": 1.54, "grad_norm": 4.26994514465332, "learning_rate": 5.761616161616162e-07, "loss": 0.5605, "step": 4300 }, { "epoch": 1.55, "grad_norm": 4.853590488433838, "learning_rate": 5.751515151515151e-07, "loss": 0.587, "step": 4310 }, { "epoch": 1.55, "grad_norm": 5.43099308013916, "learning_rate": 5.741414141414142e-07, "loss": 0.533, "step": 4320 }, { "epoch": 1.55, "grad_norm": 4.218691349029541, "learning_rate": 5.731313131313131e-07, "loss": 0.5373, "step": 4330 }, { "epoch": 1.56, "grad_norm": 6.392608642578125, "learning_rate": 5.721212121212121e-07, "loss": 0.5346, "step": 4340 }, { "epoch": 1.56, "grad_norm": 4.781940937042236, "learning_rate": 5.711111111111111e-07, "loss": 0.6316, "step": 4350 }, { "epoch": 1.56, "grad_norm": 4.0989990234375, "learning_rate": 5.702020202020201e-07, "loss": 0.5863, "step": 4360 }, { "epoch": 1.57, "grad_norm": 4.195807456970215, "learning_rate": 5.691919191919191e-07, "loss": 0.6091, "step": 4370 }, { "epoch": 1.57, "grad_norm": 3.2372219562530518, "learning_rate": 5.681818181818182e-07, "loss": 0.5939, "step": 4380 }, { "epoch": 1.57, "grad_norm": 3.523815631866455, "learning_rate": 5.671717171717172e-07, "loss": 0.5908, "step": 4390 }, { "epoch": 1.58, "grad_norm": 4.94608736038208, "learning_rate": 5.661616161616162e-07, "loss": 0.5964, "step": 4400 }, { "epoch": 1.58, "grad_norm": 8.730268478393555, "learning_rate": 5.651515151515151e-07, "loss": 0.5355, "step": 4410 }, { "epoch": 1.59, "grad_norm": 4.550642490386963, "learning_rate": 5.641414141414142e-07, "loss": 0.5295, "step": 4420 }, { "epoch": 1.59, "grad_norm": 4.5829176902771, "learning_rate": 5.631313131313131e-07, "loss": 0.5915, "step": 4430 }, { "epoch": 1.59, "grad_norm": 4.118924617767334, "learning_rate": 5.621212121212121e-07, "loss": 0.5589, "step": 4440 }, { "epoch": 1.6, "grad_norm": 4.678675174713135, "learning_rate": 5.611111111111111e-07, "loss": 0.6621, "step": 4450 }, { "epoch": 1.6, "grad_norm": 4.09968376159668, "learning_rate": 5.601010101010101e-07, "loss": 0.6018, "step": 4460 }, { "epoch": 1.6, "grad_norm": 6.312856674194336, "learning_rate": 5.590909090909091e-07, "loss": 0.6402, "step": 4470 }, { "epoch": 1.61, "grad_norm": 5.487626552581787, "learning_rate": 5.58080808080808e-07, "loss": 0.5799, "step": 4480 }, { "epoch": 1.61, "grad_norm": 4.387622833251953, "learning_rate": 5.57070707070707e-07, "loss": 0.5685, "step": 4490 }, { "epoch": 1.61, "grad_norm": 4.692462921142578, "learning_rate": 5.56060606060606e-07, "loss": 0.5935, "step": 4500 }, { "epoch": 1.62, "grad_norm": 3.9311652183532715, "learning_rate": 5.55050505050505e-07, "loss": 0.6092, "step": 4510 }, { "epoch": 1.62, "grad_norm": 4.879345417022705, "learning_rate": 5.540404040404039e-07, "loss": 0.5754, "step": 4520 }, { "epoch": 1.62, "grad_norm": 4.526495933532715, "learning_rate": 5.53030303030303e-07, "loss": 0.6167, "step": 4530 }, { "epoch": 1.63, "grad_norm": 5.269490718841553, "learning_rate": 5.520202020202019e-07, "loss": 0.5197, "step": 4540 }, { "epoch": 1.63, "grad_norm": 6.770501613616943, "learning_rate": 5.510101010101011e-07, "loss": 0.5389, "step": 4550 }, { "epoch": 1.64, "grad_norm": 5.688599109649658, "learning_rate": 5.5e-07, "loss": 0.518, "step": 4560 }, { "epoch": 1.64, "grad_norm": 3.049668073654175, "learning_rate": 5.48989898989899e-07, "loss": 0.6034, "step": 4570 }, { "epoch": 1.64, "grad_norm": 5.590402126312256, "learning_rate": 5.47979797979798e-07, "loss": 0.6146, "step": 4580 }, { "epoch": 1.65, "grad_norm": 4.051222801208496, "learning_rate": 5.46969696969697e-07, "loss": 0.5388, "step": 4590 }, { "epoch": 1.65, "grad_norm": 3.68947434425354, "learning_rate": 5.459595959595959e-07, "loss": 0.5049, "step": 4600 }, { "epoch": 1.65, "grad_norm": 5.772908687591553, "learning_rate": 5.44949494949495e-07, "loss": 0.5447, "step": 4610 }, { "epoch": 1.66, "grad_norm": 3.907376289367676, "learning_rate": 5.439393939393939e-07, "loss": 0.5038, "step": 4620 }, { "epoch": 1.66, "grad_norm": 6.761162281036377, "learning_rate": 5.42929292929293e-07, "loss": 0.6141, "step": 4630 }, { "epoch": 1.66, "grad_norm": 3.054441452026367, "learning_rate": 5.419191919191919e-07, "loss": 0.6043, "step": 4640 }, { "epoch": 1.67, "grad_norm": 5.1576457023620605, "learning_rate": 5.409090909090908e-07, "loss": 0.5588, "step": 4650 }, { "epoch": 1.67, "grad_norm": 4.096390724182129, "learning_rate": 5.398989898989899e-07, "loss": 0.6355, "step": 4660 }, { "epoch": 1.67, "grad_norm": 3.5117170810699463, "learning_rate": 5.388888888888888e-07, "loss": 0.6043, "step": 4670 }, { "epoch": 1.68, "grad_norm": 3.7968451976776123, "learning_rate": 5.378787878787878e-07, "loss": 0.6092, "step": 4680 }, { "epoch": 1.68, "grad_norm": 5.3210320472717285, "learning_rate": 5.368686868686868e-07, "loss": 0.5375, "step": 4690 }, { "epoch": 1.69, "grad_norm": 3.6537818908691406, "learning_rate": 5.358585858585858e-07, "loss": 0.5013, "step": 4700 }, { "epoch": 1.69, "grad_norm": 3.1777806282043457, "learning_rate": 5.348484848484848e-07, "loss": 0.5073, "step": 4710 }, { "epoch": 1.69, "grad_norm": 5.203103065490723, "learning_rate": 5.338383838383839e-07, "loss": 0.5694, "step": 4720 }, { "epoch": 1.7, "grad_norm": 3.941244602203369, "learning_rate": 5.328282828282828e-07, "loss": 0.5337, "step": 4730 }, { "epoch": 1.7, "grad_norm": 3.7557549476623535, "learning_rate": 5.318181818181819e-07, "loss": 0.6047, "step": 4740 }, { "epoch": 1.7, "grad_norm": 5.3054351806640625, "learning_rate": 5.308080808080808e-07, "loss": 0.5563, "step": 4750 }, { "epoch": 1.71, "grad_norm": 3.023254871368408, "learning_rate": 5.297979797979798e-07, "loss": 0.5877, "step": 4760 }, { "epoch": 1.71, "grad_norm": 3.8223531246185303, "learning_rate": 5.287878787878788e-07, "loss": 0.5938, "step": 4770 }, { "epoch": 1.71, "grad_norm": 2.889775276184082, "learning_rate": 5.277777777777777e-07, "loss": 0.5628, "step": 4780 }, { "epoch": 1.72, "grad_norm": 4.528804302215576, "learning_rate": 5.267676767676768e-07, "loss": 0.5518, "step": 4790 }, { "epoch": 1.72, "grad_norm": 4.486887454986572, "learning_rate": 5.257575757575757e-07, "loss": 0.6116, "step": 4800 }, { "epoch": 1.72, "grad_norm": 3.2164041996002197, "learning_rate": 5.247474747474747e-07, "loss": 0.5949, "step": 4810 }, { "epoch": 1.73, "grad_norm": 6.722822666168213, "learning_rate": 5.237373737373737e-07, "loss": 0.6425, "step": 4820 }, { "epoch": 1.73, "grad_norm": 4.429785251617432, "learning_rate": 5.227272727272727e-07, "loss": 0.5447, "step": 4830 }, { "epoch": 1.74, "grad_norm": 4.903067111968994, "learning_rate": 5.217171717171716e-07, "loss": 0.5635, "step": 4840 }, { "epoch": 1.74, "grad_norm": 6.777737617492676, "learning_rate": 5.207070707070707e-07, "loss": 0.5953, "step": 4850 }, { "epoch": 1.74, "grad_norm": 5.027159214019775, "learning_rate": 5.196969696969696e-07, "loss": 0.598, "step": 4860 }, { "epoch": 1.75, "grad_norm": 3.1004555225372314, "learning_rate": 5.186868686868687e-07, "loss": 0.5972, "step": 4870 }, { "epoch": 1.75, "grad_norm": 6.468344211578369, "learning_rate": 5.176767676767676e-07, "loss": 0.5707, "step": 4880 }, { "epoch": 1.75, "grad_norm": 4.483850955963135, "learning_rate": 5.166666666666667e-07, "loss": 0.5488, "step": 4890 }, { "epoch": 1.76, "grad_norm": 4.9485602378845215, "learning_rate": 5.156565656565657e-07, "loss": 0.5356, "step": 4900 }, { "epoch": 1.76, "grad_norm": 3.1415200233459473, "learning_rate": 5.146464646464647e-07, "loss": 0.5455, "step": 4910 }, { "epoch": 1.76, "grad_norm": 4.05121374130249, "learning_rate": 5.136363636363636e-07, "loss": 0.5762, "step": 4920 }, { "epoch": 1.77, "grad_norm": 4.803497791290283, "learning_rate": 5.126262626262626e-07, "loss": 0.6569, "step": 4930 }, { "epoch": 1.77, "grad_norm": 3.7656466960906982, "learning_rate": 5.116161616161616e-07, "loss": 0.5274, "step": 4940 }, { "epoch": 1.78, "grad_norm": 6.404185771942139, "learning_rate": 5.106060606060606e-07, "loss": 0.599, "step": 4950 }, { "epoch": 1.78, "grad_norm": 4.116934299468994, "learning_rate": 5.095959595959596e-07, "loss": 0.5658, "step": 4960 }, { "epoch": 1.78, "grad_norm": 3.6211934089660645, "learning_rate": 5.085858585858585e-07, "loss": 0.6003, "step": 4970 }, { "epoch": 1.79, "grad_norm": 4.541522979736328, "learning_rate": 5.075757575757576e-07, "loss": 0.5854, "step": 4980 }, { "epoch": 1.79, "grad_norm": 4.65029764175415, "learning_rate": 5.065656565656565e-07, "loss": 0.5504, "step": 4990 }, { "epoch": 1.79, "grad_norm": 4.597330093383789, "learning_rate": 5.055555555555555e-07, "loss": 0.5244, "step": 5000 }, { "epoch": 1.79, "eval_loss": 0.6935617923736572, "eval_runtime": 401.44, "eval_samples_per_second": 2.491, "eval_steps_per_second": 2.491, "step": 5000 }, { "epoch": 1.8, "grad_norm": 4.305992126464844, "learning_rate": 5.045454545454545e-07, "loss": 0.5294, "step": 5010 }, { "epoch": 1.8, "grad_norm": 9.865640640258789, "learning_rate": 5.035353535353535e-07, "loss": 0.5933, "step": 5020 }, { "epoch": 1.8, "grad_norm": 4.958790302276611, "learning_rate": 5.025252525252525e-07, "loss": 0.515, "step": 5030 }, { "epoch": 1.81, "grad_norm": 4.508664131164551, "learning_rate": 5.015151515151515e-07, "loss": 0.5725, "step": 5040 }, { "epoch": 1.81, "grad_norm": 5.282695293426514, "learning_rate": 5.005050505050504e-07, "loss": 0.5779, "step": 5050 }, { "epoch": 1.81, "grad_norm": 3.441600799560547, "learning_rate": 4.994949494949494e-07, "loss": 0.5832, "step": 5060 }, { "epoch": 1.82, "grad_norm": 6.297787666320801, "learning_rate": 4.984848484848485e-07, "loss": 0.5923, "step": 5070 }, { "epoch": 1.82, "grad_norm": 4.202699184417725, "learning_rate": 4.974747474747474e-07, "loss": 0.5525, "step": 5080 }, { "epoch": 1.83, "grad_norm": 6.49127721786499, "learning_rate": 4.964646464646464e-07, "loss": 0.5443, "step": 5090 }, { "epoch": 1.83, "grad_norm": 5.5521345138549805, "learning_rate": 4.954545454545454e-07, "loss": 0.5871, "step": 5100 }, { "epoch": 1.83, "grad_norm": 6.344796657562256, "learning_rate": 4.944444444444445e-07, "loss": 0.5371, "step": 5110 }, { "epoch": 1.84, "grad_norm": 4.388942718505859, "learning_rate": 4.934343434343434e-07, "loss": 0.4874, "step": 5120 }, { "epoch": 1.84, "grad_norm": 5.535308837890625, "learning_rate": 4.924242424242424e-07, "loss": 0.5937, "step": 5130 }, { "epoch": 1.84, "grad_norm": 3.587024450302124, "learning_rate": 4.914141414141414e-07, "loss": 0.5893, "step": 5140 }, { "epoch": 1.85, "grad_norm": 4.178186416625977, "learning_rate": 4.904040404040404e-07, "loss": 0.569, "step": 5150 }, { "epoch": 1.85, "grad_norm": 4.824371337890625, "learning_rate": 4.893939393939393e-07, "loss": 0.5923, "step": 5160 }, { "epoch": 1.85, "grad_norm": 3.2746853828430176, "learning_rate": 4.883838383838384e-07, "loss": 0.5618, "step": 5170 }, { "epoch": 1.86, "grad_norm": 4.693532943725586, "learning_rate": 4.873737373737373e-07, "loss": 0.5875, "step": 5180 }, { "epoch": 1.86, "grad_norm": 4.339977264404297, "learning_rate": 4.863636363636364e-07, "loss": 0.5863, "step": 5190 }, { "epoch": 1.86, "grad_norm": 4.782194137573242, "learning_rate": 4.853535353535353e-07, "loss": 0.5649, "step": 5200 }, { "epoch": 1.87, "grad_norm": 5.156185626983643, "learning_rate": 4.843434343434343e-07, "loss": 0.5377, "step": 5210 }, { "epoch": 1.87, "grad_norm": 4.788945198059082, "learning_rate": 4.833333333333333e-07, "loss": 0.4862, "step": 5220 }, { "epoch": 1.88, "grad_norm": 4.9057817459106445, "learning_rate": 4.823232323232323e-07, "loss": 0.5838, "step": 5230 }, { "epoch": 1.88, "grad_norm": 5.085701942443848, "learning_rate": 4.813131313131313e-07, "loss": 0.6107, "step": 5240 }, { "epoch": 1.88, "grad_norm": 3.601012945175171, "learning_rate": 4.803030303030302e-07, "loss": 0.4859, "step": 5250 }, { "epoch": 1.89, "grad_norm": 5.70768404006958, "learning_rate": 4.792929292929293e-07, "loss": 0.5206, "step": 5260 }, { "epoch": 1.89, "grad_norm": 5.292908191680908, "learning_rate": 4.782828282828283e-07, "loss": 0.5652, "step": 5270 }, { "epoch": 1.89, "grad_norm": 5.359946250915527, "learning_rate": 4.772727272727273e-07, "loss": 0.5807, "step": 5280 }, { "epoch": 1.9, "grad_norm": 4.304328441619873, "learning_rate": 4.7626262626262627e-07, "loss": 0.5867, "step": 5290 }, { "epoch": 1.9, "grad_norm": 3.993483304977417, "learning_rate": 4.752525252525252e-07, "loss": 0.6, "step": 5300 }, { "epoch": 1.9, "grad_norm": 5.294483661651611, "learning_rate": 4.742424242424242e-07, "loss": 0.6168, "step": 5310 }, { "epoch": 1.91, "grad_norm": 5.31272554397583, "learning_rate": 4.732323232323232e-07, "loss": 0.5645, "step": 5320 }, { "epoch": 1.91, "grad_norm": 3.8329992294311523, "learning_rate": 4.722222222222222e-07, "loss": 0.5978, "step": 5330 }, { "epoch": 1.92, "grad_norm": 5.329180717468262, "learning_rate": 4.7121212121212115e-07, "loss": 0.562, "step": 5340 }, { "epoch": 1.92, "grad_norm": 9.958372116088867, "learning_rate": 4.7020202020202015e-07, "loss": 0.6556, "step": 5350 }, { "epoch": 1.92, "grad_norm": 7.695910453796387, "learning_rate": 4.691919191919192e-07, "loss": 0.4964, "step": 5360 }, { "epoch": 1.93, "grad_norm": 3.406792163848877, "learning_rate": 4.681818181818182e-07, "loss": 0.5105, "step": 5370 }, { "epoch": 1.93, "grad_norm": 6.779275894165039, "learning_rate": 4.6717171717171714e-07, "loss": 0.4962, "step": 5380 }, { "epoch": 1.93, "grad_norm": 6.590269088745117, "learning_rate": 4.6616161616161614e-07, "loss": 0.5726, "step": 5390 }, { "epoch": 1.94, "grad_norm": 3.9702072143554688, "learning_rate": 4.6515151515151513e-07, "loss": 0.5568, "step": 5400 }, { "epoch": 1.94, "grad_norm": 4.912775039672852, "learning_rate": 4.6414141414141413e-07, "loss": 0.5604, "step": 5410 }, { "epoch": 1.94, "grad_norm": 5.167705059051514, "learning_rate": 4.6313131313131307e-07, "loss": 0.5364, "step": 5420 }, { "epoch": 1.95, "grad_norm": 5.220252513885498, "learning_rate": 4.6212121212121207e-07, "loss": 0.5841, "step": 5430 }, { "epoch": 1.95, "grad_norm": 4.965468883514404, "learning_rate": 4.611111111111111e-07, "loss": 0.5934, "step": 5440 }, { "epoch": 1.95, "grad_norm": 6.403870582580566, "learning_rate": 4.601010101010101e-07, "loss": 0.5694, "step": 5450 }, { "epoch": 1.96, "grad_norm": 4.846173286437988, "learning_rate": 4.5909090909090906e-07, "loss": 0.5431, "step": 5460 }, { "epoch": 1.96, "grad_norm": 4.517751216888428, "learning_rate": 4.5808080808080806e-07, "loss": 0.5966, "step": 5470 }, { "epoch": 1.97, "grad_norm": 4.690954208374023, "learning_rate": 4.5707070707070705e-07, "loss": 0.5472, "step": 5480 }, { "epoch": 1.97, "grad_norm": 5.523789882659912, "learning_rate": 4.5606060606060605e-07, "loss": 0.5442, "step": 5490 }, { "epoch": 1.97, "grad_norm": 3.194629430770874, "learning_rate": 4.55050505050505e-07, "loss": 0.6165, "step": 5500 }, { "epoch": 1.98, "grad_norm": 5.22808313369751, "learning_rate": 4.54040404040404e-07, "loss": 0.5074, "step": 5510 }, { "epoch": 1.98, "grad_norm": 4.611573219299316, "learning_rate": 4.53030303030303e-07, "loss": 0.4977, "step": 5520 }, { "epoch": 1.98, "grad_norm": 3.746086835861206, "learning_rate": 4.5202020202020204e-07, "loss": 0.5699, "step": 5530 }, { "epoch": 1.99, "grad_norm": 4.322105884552002, "learning_rate": 4.51010101010101e-07, "loss": 0.5817, "step": 5540 }, { "epoch": 1.99, "grad_norm": 4.089654922485352, "learning_rate": 4.5e-07, "loss": 0.5593, "step": 5550 }, { "epoch": 1.99, "grad_norm": 4.9758710861206055, "learning_rate": 4.48989898989899e-07, "loss": 0.5716, "step": 5560 }, { "epoch": 2.0, "grad_norm": 4.36923885345459, "learning_rate": 4.4797979797979797e-07, "loss": 0.5937, "step": 5570 }, { "epoch": 2.0, "grad_norm": 4.592928886413574, "learning_rate": 4.469696969696969e-07, "loss": 0.5471, "step": 5580 }, { "epoch": 2.0, "grad_norm": 9.409171104431152, "learning_rate": 4.459595959595959e-07, "loss": 0.7263, "step": 5590 }, { "epoch": 2.01, "grad_norm": 8.28042221069336, "learning_rate": 4.449494949494949e-07, "loss": 0.5391, "step": 5600 }, { "epoch": 2.01, "grad_norm": 4.366082191467285, "learning_rate": 4.4393939393939396e-07, "loss": 0.5164, "step": 5610 }, { "epoch": 2.02, "grad_norm": 3.808896541595459, "learning_rate": 4.429292929292929e-07, "loss": 0.66, "step": 5620 }, { "epoch": 2.02, "grad_norm": 4.30920934677124, "learning_rate": 4.419191919191919e-07, "loss": 0.5418, "step": 5630 }, { "epoch": 2.02, "grad_norm": 4.720395088195801, "learning_rate": 4.409090909090909e-07, "loss": 0.526, "step": 5640 }, { "epoch": 2.03, "grad_norm": 3.699301242828369, "learning_rate": 4.398989898989899e-07, "loss": 0.5309, "step": 5650 }, { "epoch": 2.03, "grad_norm": 4.549612522125244, "learning_rate": 4.3888888888888884e-07, "loss": 0.6452, "step": 5660 }, { "epoch": 2.03, "grad_norm": 3.252539873123169, "learning_rate": 4.3787878787878784e-07, "loss": 0.5111, "step": 5670 }, { "epoch": 2.04, "grad_norm": 4.932587146759033, "learning_rate": 4.3686868686868683e-07, "loss": 0.539, "step": 5680 }, { "epoch": 2.04, "grad_norm": 4.650817394256592, "learning_rate": 4.3585858585858583e-07, "loss": 0.6203, "step": 5690 }, { "epoch": 2.04, "grad_norm": 6.277070045471191, "learning_rate": 4.3484848484848483e-07, "loss": 0.5435, "step": 5700 }, { "epoch": 2.05, "grad_norm": 7.179144859313965, "learning_rate": 4.338383838383838e-07, "loss": 0.5062, "step": 5710 }, { "epoch": 2.05, "grad_norm": 3.681550979614258, "learning_rate": 4.328282828282828e-07, "loss": 0.5985, "step": 5720 }, { "epoch": 2.05, "grad_norm": 3.818920850753784, "learning_rate": 4.318181818181818e-07, "loss": 0.4808, "step": 5730 }, { "epoch": 2.06, "grad_norm": 3.2395308017730713, "learning_rate": 4.3080808080808076e-07, "loss": 0.4888, "step": 5740 }, { "epoch": 2.06, "grad_norm": 4.037728786468506, "learning_rate": 4.2979797979797976e-07, "loss": 0.6042, "step": 5750 }, { "epoch": 2.07, "grad_norm": 5.200090408325195, "learning_rate": 4.2878787878787876e-07, "loss": 0.5801, "step": 5760 }, { "epoch": 2.07, "grad_norm": 6.219593524932861, "learning_rate": 4.2777777777777775e-07, "loss": 0.5453, "step": 5770 }, { "epoch": 2.07, "grad_norm": 5.049505233764648, "learning_rate": 4.267676767676767e-07, "loss": 0.5342, "step": 5780 }, { "epoch": 2.08, "grad_norm": 4.03887939453125, "learning_rate": 4.2575757575757575e-07, "loss": 0.511, "step": 5790 }, { "epoch": 2.08, "grad_norm": 4.79861307144165, "learning_rate": 4.2474747474747474e-07, "loss": 0.5106, "step": 5800 }, { "epoch": 2.08, "grad_norm": 4.6482744216918945, "learning_rate": 4.2373737373737374e-07, "loss": 0.5887, "step": 5810 }, { "epoch": 2.09, "grad_norm": 4.516157150268555, "learning_rate": 4.227272727272727e-07, "loss": 0.5637, "step": 5820 }, { "epoch": 2.09, "grad_norm": 5.392308712005615, "learning_rate": 4.217171717171717e-07, "loss": 0.5485, "step": 5830 }, { "epoch": 2.09, "grad_norm": 3.8693811893463135, "learning_rate": 4.207070707070707e-07, "loss": 0.496, "step": 5840 }, { "epoch": 2.1, "grad_norm": 4.755632400512695, "learning_rate": 4.196969696969697e-07, "loss": 0.5315, "step": 5850 }, { "epoch": 2.1, "grad_norm": 5.732184410095215, "learning_rate": 4.186868686868686e-07, "loss": 0.5608, "step": 5860 }, { "epoch": 2.11, "grad_norm": 4.804543495178223, "learning_rate": 4.1767676767676767e-07, "loss": 0.5641, "step": 5870 }, { "epoch": 2.11, "grad_norm": 6.783200263977051, "learning_rate": 4.1666666666666667e-07, "loss": 0.4667, "step": 5880 }, { "epoch": 2.11, "grad_norm": 4.994887828826904, "learning_rate": 4.1565656565656566e-07, "loss": 0.5759, "step": 5890 }, { "epoch": 2.12, "grad_norm": 5.834484100341797, "learning_rate": 4.1464646464646466e-07, "loss": 0.5619, "step": 5900 }, { "epoch": 2.12, "grad_norm": 5.325469970703125, "learning_rate": 4.136363636363636e-07, "loss": 0.5464, "step": 5910 }, { "epoch": 2.12, "grad_norm": 4.983416557312012, "learning_rate": 4.126262626262626e-07, "loss": 0.5974, "step": 5920 }, { "epoch": 2.13, "grad_norm": 4.775305271148682, "learning_rate": 4.116161616161616e-07, "loss": 0.5709, "step": 5930 }, { "epoch": 2.13, "grad_norm": 5.547243595123291, "learning_rate": 4.1060606060606054e-07, "loss": 0.5474, "step": 5940 }, { "epoch": 2.13, "grad_norm": 3.5384747982025146, "learning_rate": 4.0959595959595954e-07, "loss": 0.4197, "step": 5950 }, { "epoch": 2.14, "grad_norm": 5.288349628448486, "learning_rate": 4.085858585858586e-07, "loss": 0.6163, "step": 5960 }, { "epoch": 2.14, "grad_norm": 5.531615734100342, "learning_rate": 4.075757575757576e-07, "loss": 0.53, "step": 5970 }, { "epoch": 2.14, "grad_norm": 6.349400997161865, "learning_rate": 4.065656565656566e-07, "loss": 0.5457, "step": 5980 }, { "epoch": 2.15, "grad_norm": 5.897380828857422, "learning_rate": 4.055555555555555e-07, "loss": 0.5782, "step": 5990 }, { "epoch": 2.15, "grad_norm": 5.457867622375488, "learning_rate": 4.045454545454545e-07, "loss": 0.5734, "step": 6000 }, { "epoch": 2.15, "eval_loss": 0.726588785648346, "eval_runtime": 400.7393, "eval_samples_per_second": 2.495, "eval_steps_per_second": 2.495, "step": 6000 }, { "epoch": 2.16, "grad_norm": 5.9234538078308105, "learning_rate": 4.035353535353535e-07, "loss": 0.5843, "step": 6010 }, { "epoch": 2.16, "grad_norm": 6.228603363037109, "learning_rate": 4.025252525252525e-07, "loss": 0.5035, "step": 6020 }, { "epoch": 2.16, "grad_norm": 4.385622024536133, "learning_rate": 4.0151515151515146e-07, "loss": 0.5226, "step": 6030 }, { "epoch": 2.17, "grad_norm": 5.10957670211792, "learning_rate": 4.005050505050505e-07, "loss": 0.5418, "step": 6040 }, { "epoch": 2.17, "grad_norm": 3.951185703277588, "learning_rate": 3.994949494949495e-07, "loss": 0.5448, "step": 6050 }, { "epoch": 2.17, "grad_norm": 4.83636474609375, "learning_rate": 3.984848484848485e-07, "loss": 0.5483, "step": 6060 }, { "epoch": 2.18, "grad_norm": 6.703040599822998, "learning_rate": 3.9747474747474745e-07, "loss": 0.5568, "step": 6070 }, { "epoch": 2.18, "grad_norm": 6.855874538421631, "learning_rate": 3.9646464646464644e-07, "loss": 0.5365, "step": 6080 }, { "epoch": 2.18, "grad_norm": 4.763612270355225, "learning_rate": 3.9545454545454544e-07, "loss": 0.5672, "step": 6090 }, { "epoch": 2.19, "grad_norm": 5.316641330718994, "learning_rate": 3.9444444444444444e-07, "loss": 0.5943, "step": 6100 }, { "epoch": 2.19, "grad_norm": 7.093216896057129, "learning_rate": 3.934343434343434e-07, "loss": 0.5548, "step": 6110 }, { "epoch": 2.19, "grad_norm": 5.546849250793457, "learning_rate": 3.924242424242424e-07, "loss": 0.5377, "step": 6120 }, { "epoch": 2.2, "grad_norm": 6.597376823425293, "learning_rate": 3.9141414141414143e-07, "loss": 0.5512, "step": 6130 }, { "epoch": 2.2, "grad_norm": 4.313337326049805, "learning_rate": 3.904040404040404e-07, "loss": 0.498, "step": 6140 }, { "epoch": 2.21, "grad_norm": 7.560310363769531, "learning_rate": 3.8939393939393937e-07, "loss": 0.5511, "step": 6150 }, { "epoch": 2.21, "grad_norm": 4.503920555114746, "learning_rate": 3.8838383838383837e-07, "loss": 0.5497, "step": 6160 }, { "epoch": 2.21, "grad_norm": 5.242457866668701, "learning_rate": 3.8737373737373736e-07, "loss": 0.6372, "step": 6170 }, { "epoch": 2.22, "grad_norm": 5.803231716156006, "learning_rate": 3.8636363636363636e-07, "loss": 0.5811, "step": 6180 }, { "epoch": 2.22, "grad_norm": 4.9778008460998535, "learning_rate": 3.853535353535353e-07, "loss": 0.5717, "step": 6190 }, { "epoch": 2.22, "grad_norm": 3.297873020172119, "learning_rate": 3.843434343434343e-07, "loss": 0.5374, "step": 6200 }, { "epoch": 2.23, "grad_norm": 4.822756290435791, "learning_rate": 3.8333333333333335e-07, "loss": 0.5788, "step": 6210 }, { "epoch": 2.23, "grad_norm": 6.319406509399414, "learning_rate": 3.8232323232323235e-07, "loss": 0.4811, "step": 6220 }, { "epoch": 2.23, "grad_norm": 5.257617950439453, "learning_rate": 3.813131313131313e-07, "loss": 0.6003, "step": 6230 }, { "epoch": 2.24, "grad_norm": 5.456783771514893, "learning_rate": 3.803030303030303e-07, "loss": 0.4827, "step": 6240 }, { "epoch": 2.24, "grad_norm": 5.607643127441406, "learning_rate": 3.792929292929293e-07, "loss": 0.4954, "step": 6250 }, { "epoch": 2.24, "grad_norm": 7.377701282501221, "learning_rate": 3.782828282828283e-07, "loss": 0.6513, "step": 6260 }, { "epoch": 2.25, "grad_norm": 5.05380392074585, "learning_rate": 3.7727272727272723e-07, "loss": 0.5108, "step": 6270 }, { "epoch": 2.25, "grad_norm": 5.01906681060791, "learning_rate": 3.762626262626262e-07, "loss": 0.5575, "step": 6280 }, { "epoch": 2.26, "grad_norm": 4.88652229309082, "learning_rate": 3.752525252525252e-07, "loss": 0.5732, "step": 6290 }, { "epoch": 2.26, "grad_norm": 4.311870098114014, "learning_rate": 3.7424242424242427e-07, "loss": 0.5814, "step": 6300 }, { "epoch": 2.26, "grad_norm": 5.589522361755371, "learning_rate": 3.732323232323232e-07, "loss": 0.5317, "step": 6310 }, { "epoch": 2.27, "grad_norm": 7.32663631439209, "learning_rate": 3.722222222222222e-07, "loss": 0.5446, "step": 6320 }, { "epoch": 2.27, "grad_norm": 4.902598857879639, "learning_rate": 3.712121212121212e-07, "loss": 0.5386, "step": 6330 }, { "epoch": 2.27, "grad_norm": 5.21954870223999, "learning_rate": 3.702020202020202e-07, "loss": 0.4813, "step": 6340 }, { "epoch": 2.28, "grad_norm": 3.8048980236053467, "learning_rate": 3.6919191919191915e-07, "loss": 0.5571, "step": 6350 }, { "epoch": 2.28, "grad_norm": 5.75179386138916, "learning_rate": 3.6818181818181815e-07, "loss": 0.6023, "step": 6360 }, { "epoch": 2.28, "grad_norm": 6.469977855682373, "learning_rate": 3.6717171717171714e-07, "loss": 0.65, "step": 6370 }, { "epoch": 2.29, "grad_norm": 5.1519317626953125, "learning_rate": 3.661616161616162e-07, "loss": 0.5634, "step": 6380 }, { "epoch": 2.29, "grad_norm": 4.514432907104492, "learning_rate": 3.6515151515151514e-07, "loss": 0.5638, "step": 6390 }, { "epoch": 2.3, "grad_norm": 5.5475029945373535, "learning_rate": 3.6414141414141413e-07, "loss": 0.5922, "step": 6400 }, { "epoch": 2.3, "grad_norm": 6.521148204803467, "learning_rate": 3.6313131313131313e-07, "loss": 0.5516, "step": 6410 }, { "epoch": 2.3, "grad_norm": 4.127744674682617, "learning_rate": 3.6212121212121213e-07, "loss": 0.5455, "step": 6420 }, { "epoch": 2.31, "grad_norm": 5.580835342407227, "learning_rate": 3.6111111111111107e-07, "loss": 0.4762, "step": 6430 }, { "epoch": 2.31, "grad_norm": 5.267888069152832, "learning_rate": 3.6010101010101007e-07, "loss": 0.6285, "step": 6440 }, { "epoch": 2.31, "grad_norm": 5.960037708282471, "learning_rate": 3.5909090909090907e-07, "loss": 0.5553, "step": 6450 }, { "epoch": 2.32, "grad_norm": 3.847745656967163, "learning_rate": 3.5808080808080806e-07, "loss": 0.5603, "step": 6460 }, { "epoch": 2.32, "grad_norm": 5.693478107452393, "learning_rate": 3.5707070707070706e-07, "loss": 0.6077, "step": 6470 }, { "epoch": 2.32, "grad_norm": 4.48386287689209, "learning_rate": 3.5606060606060606e-07, "loss": 0.5636, "step": 6480 }, { "epoch": 2.33, "grad_norm": 6.377192497253418, "learning_rate": 3.5505050505050505e-07, "loss": 0.5089, "step": 6490 }, { "epoch": 2.33, "grad_norm": 5.240867614746094, "learning_rate": 3.5404040404040405e-07, "loss": 0.5547, "step": 6500 }, { "epoch": 2.33, "grad_norm": 7.864344120025635, "learning_rate": 3.53030303030303e-07, "loss": 0.5254, "step": 6510 }, { "epoch": 2.34, "grad_norm": 8.867986679077148, "learning_rate": 3.52020202020202e-07, "loss": 0.5616, "step": 6520 }, { "epoch": 2.34, "grad_norm": 4.047301292419434, "learning_rate": 3.51010101010101e-07, "loss": 0.5123, "step": 6530 }, { "epoch": 2.35, "grad_norm": 5.919891357421875, "learning_rate": 3.5e-07, "loss": 0.5126, "step": 6540 }, { "epoch": 2.35, "grad_norm": 4.536365509033203, "learning_rate": 3.48989898989899e-07, "loss": 0.495, "step": 6550 }, { "epoch": 2.35, "grad_norm": 5.329809188842773, "learning_rate": 3.47979797979798e-07, "loss": 0.5118, "step": 6560 }, { "epoch": 2.36, "grad_norm": 5.0158867835998535, "learning_rate": 3.46969696969697e-07, "loss": 0.5487, "step": 6570 }, { "epoch": 2.36, "grad_norm": 7.028294086456299, "learning_rate": 3.4595959595959597e-07, "loss": 0.5893, "step": 6580 }, { "epoch": 2.36, "grad_norm": 4.060366153717041, "learning_rate": 3.449494949494949e-07, "loss": 0.5562, "step": 6590 }, { "epoch": 2.37, "grad_norm": 5.36353063583374, "learning_rate": 3.439393939393939e-07, "loss": 0.6158, "step": 6600 }, { "epoch": 2.37, "grad_norm": 4.725005626678467, "learning_rate": 3.429292929292929e-07, "loss": 0.4339, "step": 6610 }, { "epoch": 2.37, "grad_norm": 4.875920295715332, "learning_rate": 3.419191919191919e-07, "loss": 0.5047, "step": 6620 }, { "epoch": 2.38, "grad_norm": 6.4596266746521, "learning_rate": 3.4090909090909085e-07, "loss": 0.5516, "step": 6630 }, { "epoch": 2.38, "grad_norm": 6.09486722946167, "learning_rate": 3.398989898989899e-07, "loss": 0.5863, "step": 6640 }, { "epoch": 2.38, "grad_norm": 4.232431411743164, "learning_rate": 3.388888888888889e-07, "loss": 0.5221, "step": 6650 }, { "epoch": 2.39, "grad_norm": 4.1522111892700195, "learning_rate": 3.378787878787879e-07, "loss": 0.588, "step": 6660 }, { "epoch": 2.39, "grad_norm": 5.693157196044922, "learning_rate": 3.3686868686868684e-07, "loss": 0.5966, "step": 6670 }, { "epoch": 2.4, "grad_norm": 3.900887966156006, "learning_rate": 3.3585858585858583e-07, "loss": 0.437, "step": 6680 }, { "epoch": 2.4, "grad_norm": 5.744729995727539, "learning_rate": 3.3484848484848483e-07, "loss": 0.5483, "step": 6690 }, { "epoch": 2.4, "grad_norm": 4.26437520980835, "learning_rate": 3.3383838383838383e-07, "loss": 0.5159, "step": 6700 }, { "epoch": 2.41, "grad_norm": 5.422147750854492, "learning_rate": 3.3282828282828277e-07, "loss": 0.4986, "step": 6710 }, { "epoch": 2.41, "grad_norm": 4.445804119110107, "learning_rate": 3.318181818181818e-07, "loss": 0.544, "step": 6720 }, { "epoch": 2.41, "grad_norm": 5.514401435852051, "learning_rate": 3.308080808080808e-07, "loss": 0.5422, "step": 6730 }, { "epoch": 2.42, "grad_norm": 7.159079074859619, "learning_rate": 3.297979797979798e-07, "loss": 0.503, "step": 6740 }, { "epoch": 2.42, "grad_norm": 6.815921783447266, "learning_rate": 3.2878787878787876e-07, "loss": 0.4779, "step": 6750 }, { "epoch": 2.42, "grad_norm": 4.787651538848877, "learning_rate": 3.2777777777777776e-07, "loss": 0.484, "step": 6760 }, { "epoch": 2.43, "grad_norm": 5.437996864318848, "learning_rate": 3.2676767676767675e-07, "loss": 0.5689, "step": 6770 }, { "epoch": 2.43, "grad_norm": 2.6372170448303223, "learning_rate": 3.2575757575757575e-07, "loss": 0.6268, "step": 6780 }, { "epoch": 2.44, "grad_norm": 6.545858383178711, "learning_rate": 3.247474747474747e-07, "loss": 0.603, "step": 6790 }, { "epoch": 2.44, "grad_norm": 3.9520442485809326, "learning_rate": 3.237373737373737e-07, "loss": 0.6243, "step": 6800 }, { "epoch": 2.44, "grad_norm": 4.638275623321533, "learning_rate": 3.2272727272727274e-07, "loss": 0.5433, "step": 6810 }, { "epoch": 2.45, "grad_norm": 4.08539342880249, "learning_rate": 3.2171717171717174e-07, "loss": 0.576, "step": 6820 }, { "epoch": 2.45, "grad_norm": 6.024731636047363, "learning_rate": 3.207070707070707e-07, "loss": 0.61, "step": 6830 }, { "epoch": 2.45, "grad_norm": 3.649667263031006, "learning_rate": 3.196969696969697e-07, "loss": 0.4797, "step": 6840 }, { "epoch": 2.46, "grad_norm": 7.055994987487793, "learning_rate": 3.186868686868687e-07, "loss": 0.5525, "step": 6850 }, { "epoch": 2.46, "grad_norm": 4.7312116622924805, "learning_rate": 3.1767676767676767e-07, "loss": 0.5069, "step": 6860 }, { "epoch": 2.46, "grad_norm": 4.483476638793945, "learning_rate": 3.166666666666666e-07, "loss": 0.4884, "step": 6870 }, { "epoch": 2.47, "grad_norm": 7.678333282470703, "learning_rate": 3.156565656565656e-07, "loss": 0.4818, "step": 6880 }, { "epoch": 2.47, "grad_norm": 4.442699909210205, "learning_rate": 3.1464646464646466e-07, "loss": 0.5771, "step": 6890 }, { "epoch": 2.47, "grad_norm": 4.732019424438477, "learning_rate": 3.1363636363636366e-07, "loss": 0.5294, "step": 6900 }, { "epoch": 2.48, "grad_norm": 5.647908687591553, "learning_rate": 3.126262626262626e-07, "loss": 0.5496, "step": 6910 }, { "epoch": 2.48, "grad_norm": 4.71852970123291, "learning_rate": 3.116161616161616e-07, "loss": 0.503, "step": 6920 }, { "epoch": 2.49, "grad_norm": 8.519858360290527, "learning_rate": 3.106060606060606e-07, "loss": 0.547, "step": 6930 }, { "epoch": 2.49, "grad_norm": 4.513672351837158, "learning_rate": 3.095959595959596e-07, "loss": 0.5496, "step": 6940 }, { "epoch": 2.49, "grad_norm": 4.779497146606445, "learning_rate": 3.0858585858585854e-07, "loss": 0.5193, "step": 6950 }, { "epoch": 2.5, "grad_norm": 5.14954948425293, "learning_rate": 3.0757575757575754e-07, "loss": 0.5738, "step": 6960 }, { "epoch": 2.5, "grad_norm": 4.722686290740967, "learning_rate": 3.0656565656565653e-07, "loss": 0.5978, "step": 6970 }, { "epoch": 2.5, "grad_norm": 4.7376837730407715, "learning_rate": 3.055555555555556e-07, "loss": 0.5476, "step": 6980 }, { "epoch": 2.51, "grad_norm": 5.883960247039795, "learning_rate": 3.0454545454545453e-07, "loss": 0.4953, "step": 6990 }, { "epoch": 2.51, "grad_norm": 5.1609721183776855, "learning_rate": 3.035353535353535e-07, "loss": 0.5602, "step": 7000 }, { "epoch": 2.51, "eval_loss": 0.7043054103851318, "eval_runtime": 404.5936, "eval_samples_per_second": 2.472, "eval_steps_per_second": 2.472, "step": 7000 }, { "epoch": 2.51, "grad_norm": 4.467599391937256, "learning_rate": 3.025252525252525e-07, "loss": 0.5545, "step": 7010 }, { "epoch": 2.52, "grad_norm": 4.831850528717041, "learning_rate": 3.015151515151515e-07, "loss": 0.5765, "step": 7020 }, { "epoch": 2.52, "grad_norm": 4.63700008392334, "learning_rate": 3.0050505050505046e-07, "loss": 0.5354, "step": 7030 }, { "epoch": 2.52, "grad_norm": 4.822059631347656, "learning_rate": 2.9949494949494946e-07, "loss": 0.5239, "step": 7040 }, { "epoch": 2.53, "grad_norm": 4.098064422607422, "learning_rate": 2.9848484848484846e-07, "loss": 0.5775, "step": 7050 }, { "epoch": 2.53, "grad_norm": 5.17026948928833, "learning_rate": 2.9747474747474745e-07, "loss": 0.6014, "step": 7060 }, { "epoch": 2.54, "grad_norm": 4.5090789794921875, "learning_rate": 2.9646464646464645e-07, "loss": 0.5732, "step": 7070 }, { "epoch": 2.54, "grad_norm": 6.120937347412109, "learning_rate": 2.9545454545454545e-07, "loss": 0.4529, "step": 7080 }, { "epoch": 2.54, "grad_norm": 5.249495029449463, "learning_rate": 2.9444444444444444e-07, "loss": 0.5804, "step": 7090 }, { "epoch": 2.55, "grad_norm": 7.617781639099121, "learning_rate": 2.9343434343434344e-07, "loss": 0.5382, "step": 7100 }, { "epoch": 2.55, "grad_norm": 4.722599506378174, "learning_rate": 2.924242424242424e-07, "loss": 0.5419, "step": 7110 }, { "epoch": 2.55, "grad_norm": 4.24169921875, "learning_rate": 2.914141414141414e-07, "loss": 0.5464, "step": 7120 }, { "epoch": 2.56, "grad_norm": 4.781604766845703, "learning_rate": 2.904040404040404e-07, "loss": 0.4158, "step": 7130 }, { "epoch": 2.56, "grad_norm": 5.138591766357422, "learning_rate": 2.893939393939394e-07, "loss": 0.5626, "step": 7140 }, { "epoch": 2.56, "grad_norm": 4.901473045349121, "learning_rate": 2.8838383838383837e-07, "loss": 0.5164, "step": 7150 }, { "epoch": 2.57, "grad_norm": 3.736123561859131, "learning_rate": 2.8737373737373737e-07, "loss": 0.5159, "step": 7160 }, { "epoch": 2.57, "grad_norm": 3.937110662460327, "learning_rate": 2.8636363636363637e-07, "loss": 0.5785, "step": 7170 }, { "epoch": 2.57, "grad_norm": 5.2640581130981445, "learning_rate": 2.8535353535353536e-07, "loss": 0.5839, "step": 7180 }, { "epoch": 2.58, "grad_norm": 5.652442455291748, "learning_rate": 2.843434343434343e-07, "loss": 0.533, "step": 7190 }, { "epoch": 2.58, "grad_norm": 4.787551403045654, "learning_rate": 2.833333333333333e-07, "loss": 0.5114, "step": 7200 }, { "epoch": 2.59, "grad_norm": 5.234408855438232, "learning_rate": 2.823232323232323e-07, "loss": 0.5777, "step": 7210 }, { "epoch": 2.59, "grad_norm": 2.3309760093688965, "learning_rate": 2.813131313131313e-07, "loss": 0.5085, "step": 7220 }, { "epoch": 2.59, "grad_norm": 4.90077543258667, "learning_rate": 2.8030303030303024e-07, "loss": 0.5464, "step": 7230 }, { "epoch": 2.6, "grad_norm": 6.651805400848389, "learning_rate": 2.792929292929293e-07, "loss": 0.5143, "step": 7240 }, { "epoch": 2.6, "grad_norm": 4.081473350524902, "learning_rate": 2.782828282828283e-07, "loss": 0.5307, "step": 7250 }, { "epoch": 2.6, "grad_norm": 8.496968269348145, "learning_rate": 2.772727272727273e-07, "loss": 0.5376, "step": 7260 }, { "epoch": 2.61, "grad_norm": 4.46324348449707, "learning_rate": 2.7626262626262623e-07, "loss": 0.5871, "step": 7270 }, { "epoch": 2.61, "grad_norm": 5.344108581542969, "learning_rate": 2.752525252525252e-07, "loss": 0.4783, "step": 7280 }, { "epoch": 2.61, "grad_norm": 6.3261613845825195, "learning_rate": 2.742424242424242e-07, "loss": 0.4924, "step": 7290 }, { "epoch": 2.62, "grad_norm": 4.785797119140625, "learning_rate": 2.732323232323232e-07, "loss": 0.5087, "step": 7300 }, { "epoch": 2.62, "grad_norm": 5.484802722930908, "learning_rate": 2.7222222222222216e-07, "loss": 0.5763, "step": 7310 }, { "epoch": 2.63, "grad_norm": 5.250958442687988, "learning_rate": 2.712121212121212e-07, "loss": 0.5065, "step": 7320 }, { "epoch": 2.63, "grad_norm": 3.2717227935791016, "learning_rate": 2.702020202020202e-07, "loss": 0.4802, "step": 7330 }, { "epoch": 2.63, "grad_norm": 5.094733238220215, "learning_rate": 2.691919191919192e-07, "loss": 0.5162, "step": 7340 }, { "epoch": 2.64, "grad_norm": 5.360857009887695, "learning_rate": 2.6818181818181815e-07, "loss": 0.5947, "step": 7350 }, { "epoch": 2.64, "grad_norm": 5.611083030700684, "learning_rate": 2.6717171717171715e-07, "loss": 0.6174, "step": 7360 }, { "epoch": 2.64, "grad_norm": 4.42379093170166, "learning_rate": 2.6616161616161614e-07, "loss": 0.5773, "step": 7370 }, { "epoch": 2.65, "grad_norm": 6.272792816162109, "learning_rate": 2.6515151515151514e-07, "loss": 0.6246, "step": 7380 }, { "epoch": 2.65, "grad_norm": 5.295745849609375, "learning_rate": 2.641414141414141e-07, "loss": 0.5315, "step": 7390 }, { "epoch": 2.65, "grad_norm": 4.078943252563477, "learning_rate": 2.631313131313131e-07, "loss": 0.581, "step": 7400 }, { "epoch": 2.66, "grad_norm": 12.75452709197998, "learning_rate": 2.6212121212121213e-07, "loss": 0.5608, "step": 7410 }, { "epoch": 2.66, "grad_norm": 5.873865604400635, "learning_rate": 2.6111111111111113e-07, "loss": 0.5588, "step": 7420 }, { "epoch": 2.66, "grad_norm": 4.560327529907227, "learning_rate": 2.6010101010101007e-07, "loss": 0.6557, "step": 7430 }, { "epoch": 2.67, "grad_norm": 7.971694469451904, "learning_rate": 2.5909090909090907e-07, "loss": 0.5636, "step": 7440 }, { "epoch": 2.67, "grad_norm": 4.814733982086182, "learning_rate": 2.5808080808080807e-07, "loss": 0.5851, "step": 7450 }, { "epoch": 2.68, "grad_norm": 3.635782241821289, "learning_rate": 2.5707070707070706e-07, "loss": 0.5354, "step": 7460 }, { "epoch": 2.68, "grad_norm": 6.356813907623291, "learning_rate": 2.56060606060606e-07, "loss": 0.5934, "step": 7470 }, { "epoch": 2.68, "grad_norm": 6.781961441040039, "learning_rate": 2.55050505050505e-07, "loss": 0.5249, "step": 7480 }, { "epoch": 2.69, "grad_norm": 3.655015230178833, "learning_rate": 2.5404040404040405e-07, "loss": 0.5198, "step": 7490 }, { "epoch": 2.69, "grad_norm": 6.139344692230225, "learning_rate": 2.5303030303030305e-07, "loss": 0.5375, "step": 7500 }, { "epoch": 2.69, "grad_norm": 4.679745197296143, "learning_rate": 2.5202020202020205e-07, "loss": 0.5107, "step": 7510 }, { "epoch": 2.7, "grad_norm": 5.100747585296631, "learning_rate": 2.51010101010101e-07, "loss": 0.515, "step": 7520 }, { "epoch": 2.7, "grad_norm": 3.9256341457366943, "learning_rate": 2.5e-07, "loss": 0.5409, "step": 7530 }, { "epoch": 2.7, "grad_norm": 4.619549751281738, "learning_rate": 2.48989898989899e-07, "loss": 0.5024, "step": 7540 }, { "epoch": 2.71, "grad_norm": 4.833261489868164, "learning_rate": 2.4797979797979793e-07, "loss": 0.5287, "step": 7550 }, { "epoch": 2.71, "grad_norm": 3.5149707794189453, "learning_rate": 2.46969696969697e-07, "loss": 0.5176, "step": 7560 }, { "epoch": 2.71, "grad_norm": 6.924714088439941, "learning_rate": 2.45959595959596e-07, "loss": 0.5598, "step": 7570 }, { "epoch": 2.72, "grad_norm": 7.813880443572998, "learning_rate": 2.449494949494949e-07, "loss": 0.535, "step": 7580 }, { "epoch": 2.72, "grad_norm": 5.943690299987793, "learning_rate": 2.439393939393939e-07, "loss": 0.5327, "step": 7590 }, { "epoch": 2.73, "grad_norm": 5.665824890136719, "learning_rate": 2.429292929292929e-07, "loss": 0.501, "step": 7600 }, { "epoch": 2.73, "grad_norm": 3.8446450233459473, "learning_rate": 2.419191919191919e-07, "loss": 0.5307, "step": 7610 }, { "epoch": 2.73, "grad_norm": 4.556608200073242, "learning_rate": 2.409090909090909e-07, "loss": 0.5244, "step": 7620 }, { "epoch": 2.74, "grad_norm": 4.267519950866699, "learning_rate": 2.398989898989899e-07, "loss": 0.512, "step": 7630 }, { "epoch": 2.74, "grad_norm": 6.164044380187988, "learning_rate": 2.388888888888889e-07, "loss": 0.4957, "step": 7640 }, { "epoch": 2.74, "grad_norm": 5.647671222686768, "learning_rate": 2.3787878787878787e-07, "loss": 0.6485, "step": 7650 }, { "epoch": 2.75, "grad_norm": 3.6924054622650146, "learning_rate": 2.3686868686868687e-07, "loss": 0.4993, "step": 7660 }, { "epoch": 2.75, "grad_norm": 5.332742691040039, "learning_rate": 2.3585858585858584e-07, "loss": 0.5619, "step": 7670 }, { "epoch": 2.75, "grad_norm": 4.201364517211914, "learning_rate": 2.3484848484848486e-07, "loss": 0.5011, "step": 7680 }, { "epoch": 2.76, "grad_norm": 4.477077484130859, "learning_rate": 2.3383838383838383e-07, "loss": 0.5369, "step": 7690 }, { "epoch": 2.76, "grad_norm": 4.704535961151123, "learning_rate": 2.3282828282828283e-07, "loss": 0.5495, "step": 7700 }, { "epoch": 2.76, "grad_norm": 3.2071988582611084, "learning_rate": 2.318181818181818e-07, "loss": 0.5478, "step": 7710 }, { "epoch": 2.77, "grad_norm": 4.613890647888184, "learning_rate": 2.308080808080808e-07, "loss": 0.5208, "step": 7720 }, { "epoch": 2.77, "grad_norm": 5.759297847747803, "learning_rate": 2.297979797979798e-07, "loss": 0.4989, "step": 7730 }, { "epoch": 2.78, "grad_norm": 3.805091619491577, "learning_rate": 2.287878787878788e-07, "loss": 0.521, "step": 7740 }, { "epoch": 2.78, "grad_norm": 5.901796817779541, "learning_rate": 2.2777777777777776e-07, "loss": 0.603, "step": 7750 }, { "epoch": 2.78, "grad_norm": 4.832998752593994, "learning_rate": 2.2676767676767676e-07, "loss": 0.5698, "step": 7760 }, { "epoch": 2.79, "grad_norm": 4.8905439376831055, "learning_rate": 2.2575757575757576e-07, "loss": 0.4658, "step": 7770 }, { "epoch": 2.79, "grad_norm": 7.449604511260986, "learning_rate": 2.2474747474747475e-07, "loss": 0.5889, "step": 7780 }, { "epoch": 2.79, "grad_norm": 8.905882835388184, "learning_rate": 2.2373737373737372e-07, "loss": 0.5323, "step": 7790 }, { "epoch": 2.8, "grad_norm": 4.761772632598877, "learning_rate": 2.2272727272727272e-07, "loss": 0.53, "step": 7800 }, { "epoch": 2.8, "grad_norm": 4.811107158660889, "learning_rate": 2.2171717171717172e-07, "loss": 0.5189, "step": 7810 }, { "epoch": 2.8, "grad_norm": 2.9238836765289307, "learning_rate": 2.2070707070707071e-07, "loss": 0.5283, "step": 7820 }, { "epoch": 2.81, "grad_norm": 5.508478164672852, "learning_rate": 2.1969696969696968e-07, "loss": 0.4561, "step": 7830 }, { "epoch": 2.81, "grad_norm": 5.604992866516113, "learning_rate": 2.1868686868686868e-07, "loss": 0.5718, "step": 7840 }, { "epoch": 2.82, "grad_norm": 6.870204925537109, "learning_rate": 2.1767676767676765e-07, "loss": 0.4882, "step": 7850 }, { "epoch": 2.82, "grad_norm": 4.332894802093506, "learning_rate": 2.1666666666666667e-07, "loss": 0.5755, "step": 7860 }, { "epoch": 2.82, "grad_norm": 5.616927623748779, "learning_rate": 2.1565656565656564e-07, "loss": 0.5805, "step": 7870 }, { "epoch": 2.83, "grad_norm": 4.477654933929443, "learning_rate": 2.1464646464646464e-07, "loss": 0.582, "step": 7880 }, { "epoch": 2.83, "grad_norm": 6.561676025390625, "learning_rate": 2.136363636363636e-07, "loss": 0.4924, "step": 7890 }, { "epoch": 2.83, "grad_norm": 5.206873416900635, "learning_rate": 2.1262626262626264e-07, "loss": 0.5878, "step": 7900 }, { "epoch": 2.84, "grad_norm": 4.1171345710754395, "learning_rate": 2.116161616161616e-07, "loss": 0.5451, "step": 7910 }, { "epoch": 2.84, "grad_norm": 3.9070441722869873, "learning_rate": 2.106060606060606e-07, "loss": 0.5403, "step": 7920 }, { "epoch": 2.84, "grad_norm": 4.80625057220459, "learning_rate": 2.0959595959595957e-07, "loss": 0.5552, "step": 7930 }, { "epoch": 2.85, "grad_norm": 4.501746654510498, "learning_rate": 2.085858585858586e-07, "loss": 0.4697, "step": 7940 }, { "epoch": 2.85, "grad_norm": 3.92985463142395, "learning_rate": 2.0757575757575757e-07, "loss": 0.5565, "step": 7950 }, { "epoch": 2.85, "grad_norm": 4.4594807624816895, "learning_rate": 2.0656565656565656e-07, "loss": 0.561, "step": 7960 }, { "epoch": 2.86, "grad_norm": 5.731437683105469, "learning_rate": 2.0555555555555553e-07, "loss": 0.5311, "step": 7970 }, { "epoch": 2.86, "grad_norm": 4.295395374298096, "learning_rate": 2.0454545454545456e-07, "loss": 0.6082, "step": 7980 }, { "epoch": 2.87, "grad_norm": 4.862415313720703, "learning_rate": 2.0353535353535353e-07, "loss": 0.6022, "step": 7990 }, { "epoch": 2.87, "grad_norm": 4.604592323303223, "learning_rate": 2.0252525252525253e-07, "loss": 0.5314, "step": 8000 }, { "epoch": 2.87, "eval_loss": 0.7028300762176514, "eval_runtime": 404.8169, "eval_samples_per_second": 2.47, "eval_steps_per_second": 2.47, "step": 8000 }, { "epoch": 2.87, "grad_norm": 4.165442943572998, "learning_rate": 2.015151515151515e-07, "loss": 0.5624, "step": 8010 }, { "epoch": 2.88, "grad_norm": 4.097321510314941, "learning_rate": 2.005050505050505e-07, "loss": 0.5041, "step": 8020 }, { "epoch": 2.88, "grad_norm": 4.476957321166992, "learning_rate": 1.994949494949495e-07, "loss": 0.4556, "step": 8030 }, { "epoch": 2.88, "grad_norm": 4.502745628356934, "learning_rate": 1.9848484848484849e-07, "loss": 0.535, "step": 8040 }, { "epoch": 2.89, "grad_norm": 5.133005142211914, "learning_rate": 1.9747474747474746e-07, "loss": 0.4998, "step": 8050 }, { "epoch": 2.89, "grad_norm": 5.7199296951293945, "learning_rate": 1.9646464646464645e-07, "loss": 0.6113, "step": 8060 }, { "epoch": 2.89, "grad_norm": 4.979962348937988, "learning_rate": 1.9545454545454545e-07, "loss": 0.6419, "step": 8070 }, { "epoch": 2.9, "grad_norm": 5.971510410308838, "learning_rate": 1.9444444444444445e-07, "loss": 0.5269, "step": 8080 }, { "epoch": 2.9, "grad_norm": 6.548705577850342, "learning_rate": 1.9343434343434342e-07, "loss": 0.456, "step": 8090 }, { "epoch": 2.9, "grad_norm": 5.575558185577393, "learning_rate": 1.9242424242424241e-07, "loss": 0.5068, "step": 8100 }, { "epoch": 2.91, "grad_norm": 6.25363826751709, "learning_rate": 1.914141414141414e-07, "loss": 0.4749, "step": 8110 }, { "epoch": 2.91, "grad_norm": 5.45524787902832, "learning_rate": 1.904040404040404e-07, "loss": 0.5556, "step": 8120 }, { "epoch": 2.92, "grad_norm": 5.0804548263549805, "learning_rate": 1.8939393939393938e-07, "loss": 0.5058, "step": 8130 }, { "epoch": 2.92, "grad_norm": 6.4303460121154785, "learning_rate": 1.8838383838383838e-07, "loss": 0.553, "step": 8140 }, { "epoch": 2.92, "grad_norm": 5.177044868469238, "learning_rate": 1.8737373737373737e-07, "loss": 0.5986, "step": 8150 }, { "epoch": 2.93, "grad_norm": 4.609193325042725, "learning_rate": 1.8636363636363637e-07, "loss": 0.5456, "step": 8160 }, { "epoch": 2.93, "grad_norm": 5.29266357421875, "learning_rate": 1.8535353535353534e-07, "loss": 0.5806, "step": 8170 }, { "epoch": 2.93, "grad_norm": 6.320983409881592, "learning_rate": 1.8434343434343434e-07, "loss": 0.5621, "step": 8180 }, { "epoch": 2.94, "grad_norm": 5.851290702819824, "learning_rate": 1.833333333333333e-07, "loss": 0.5907, "step": 8190 }, { "epoch": 2.94, "grad_norm": 7.4724555015563965, "learning_rate": 1.8232323232323233e-07, "loss": 0.5415, "step": 8200 }, { "epoch": 2.94, "grad_norm": 5.282331943511963, "learning_rate": 1.813131313131313e-07, "loss": 0.4809, "step": 8210 }, { "epoch": 2.95, "grad_norm": 5.070183277130127, "learning_rate": 1.803030303030303e-07, "loss": 0.5138, "step": 8220 }, { "epoch": 2.95, "grad_norm": 4.092304229736328, "learning_rate": 1.7929292929292927e-07, "loss": 0.4772, "step": 8230 }, { "epoch": 2.95, "grad_norm": 5.767254829406738, "learning_rate": 1.782828282828283e-07, "loss": 0.5756, "step": 8240 }, { "epoch": 2.96, "grad_norm": 5.086340427398682, "learning_rate": 1.7727272727272726e-07, "loss": 0.6313, "step": 8250 }, { "epoch": 2.96, "grad_norm": 7.967188835144043, "learning_rate": 1.7626262626262626e-07, "loss": 0.5307, "step": 8260 }, { "epoch": 2.97, "grad_norm": 4.956604480743408, "learning_rate": 1.7525252525252523e-07, "loss": 0.4744, "step": 8270 }, { "epoch": 2.97, "grad_norm": 5.134040355682373, "learning_rate": 1.7424242424242425e-07, "loss": 0.5217, "step": 8280 }, { "epoch": 2.97, "grad_norm": 4.910244464874268, "learning_rate": 1.7323232323232322e-07, "loss": 0.513, "step": 8290 }, { "epoch": 2.98, "grad_norm": 6.457435131072998, "learning_rate": 1.7222222222222222e-07, "loss": 0.5786, "step": 8300 }, { "epoch": 2.98, "grad_norm": 5.450909614562988, "learning_rate": 1.712121212121212e-07, "loss": 0.4931, "step": 8310 }, { "epoch": 2.98, "grad_norm": 5.564160346984863, "learning_rate": 1.7020202020202021e-07, "loss": 0.5263, "step": 8320 }, { "epoch": 2.99, "grad_norm": 6.898198127746582, "learning_rate": 1.6919191919191918e-07, "loss": 0.5814, "step": 8330 }, { "epoch": 2.99, "grad_norm": 5.737578392028809, "learning_rate": 1.6818181818181818e-07, "loss": 0.498, "step": 8340 }, { "epoch": 2.99, "grad_norm": 4.877408027648926, "learning_rate": 1.6717171717171715e-07, "loss": 0.5743, "step": 8350 }, { "epoch": 3.0, "grad_norm": 5.765985012054443, "learning_rate": 1.6616161616161615e-07, "loss": 0.5476, "step": 8360 }, { "epoch": 3.0, "grad_norm": 12.480027198791504, "learning_rate": 1.6515151515151515e-07, "loss": 0.5693, "step": 8370 }, { "epoch": 3.01, "grad_norm": 8.612441062927246, "learning_rate": 1.6414141414141414e-07, "loss": 0.6035, "step": 8380 }, { "epoch": 3.01, "grad_norm": 5.978526592254639, "learning_rate": 1.631313131313131e-07, "loss": 0.5711, "step": 8390 }, { "epoch": 3.01, "grad_norm": 6.2041192054748535, "learning_rate": 1.621212121212121e-07, "loss": 0.4838, "step": 8400 }, { "epoch": 3.02, "grad_norm": 3.6861565113067627, "learning_rate": 1.611111111111111e-07, "loss": 0.6104, "step": 8410 }, { "epoch": 3.02, "grad_norm": 5.080516338348389, "learning_rate": 1.601010101010101e-07, "loss": 0.4688, "step": 8420 }, { "epoch": 3.02, "grad_norm": 5.527194976806641, "learning_rate": 1.5909090909090907e-07, "loss": 0.5746, "step": 8430 }, { "epoch": 3.03, "grad_norm": 4.398497104644775, "learning_rate": 1.5808080808080807e-07, "loss": 0.4895, "step": 8440 }, { "epoch": 3.03, "grad_norm": 5.004568099975586, "learning_rate": 1.5707070707070707e-07, "loss": 0.5808, "step": 8450 }, { "epoch": 3.03, "grad_norm": 6.635837554931641, "learning_rate": 1.5606060606060606e-07, "loss": 0.5519, "step": 8460 }, { "epoch": 3.04, "grad_norm": 4.912801742553711, "learning_rate": 1.5505050505050504e-07, "loss": 0.5873, "step": 8470 }, { "epoch": 3.04, "grad_norm": 5.812663555145264, "learning_rate": 1.5404040404040403e-07, "loss": 0.5286, "step": 8480 }, { "epoch": 3.04, "grad_norm": 5.986713409423828, "learning_rate": 1.53030303030303e-07, "loss": 0.5305, "step": 8490 }, { "epoch": 3.05, "grad_norm": 5.908451557159424, "learning_rate": 1.5202020202020203e-07, "loss": 0.6125, "step": 8500 }, { "epoch": 3.05, "grad_norm": 8.727606773376465, "learning_rate": 1.51010101010101e-07, "loss": 0.5765, "step": 8510 }, { "epoch": 3.06, "grad_norm": 5.8582987785339355, "learning_rate": 1.5e-07, "loss": 0.5178, "step": 8520 }, { "epoch": 3.06, "grad_norm": 5.690489292144775, "learning_rate": 1.4898989898989896e-07, "loss": 0.4978, "step": 8530 }, { "epoch": 3.06, "grad_norm": 5.1287384033203125, "learning_rate": 1.47979797979798e-07, "loss": 0.5135, "step": 8540 }, { "epoch": 3.07, "grad_norm": 5.430093765258789, "learning_rate": 1.4696969696969696e-07, "loss": 0.4935, "step": 8550 }, { "epoch": 3.07, "grad_norm": 5.307154655456543, "learning_rate": 1.4595959595959595e-07, "loss": 0.4839, "step": 8560 }, { "epoch": 3.07, "grad_norm": 4.900834560394287, "learning_rate": 1.4494949494949492e-07, "loss": 0.5637, "step": 8570 }, { "epoch": 3.08, "grad_norm": 4.796194553375244, "learning_rate": 1.4393939393939395e-07, "loss": 0.519, "step": 8580 }, { "epoch": 3.08, "grad_norm": 6.488938331604004, "learning_rate": 1.4292929292929292e-07, "loss": 0.6017, "step": 8590 }, { "epoch": 3.08, "grad_norm": 4.90757417678833, "learning_rate": 1.4191919191919192e-07, "loss": 0.4873, "step": 8600 }, { "epoch": 3.09, "grad_norm": 6.44563627243042, "learning_rate": 1.4090909090909089e-07, "loss": 0.5919, "step": 8610 }, { "epoch": 3.09, "grad_norm": 3.680788516998291, "learning_rate": 1.398989898989899e-07, "loss": 0.5062, "step": 8620 }, { "epoch": 3.09, "grad_norm": 4.880250930786133, "learning_rate": 1.3888888888888888e-07, "loss": 0.5751, "step": 8630 }, { "epoch": 3.1, "grad_norm": 4.91787052154541, "learning_rate": 1.3787878787878788e-07, "loss": 0.5604, "step": 8640 }, { "epoch": 3.1, "grad_norm": 4.990437984466553, "learning_rate": 1.3686868686868685e-07, "loss": 0.4807, "step": 8650 }, { "epoch": 3.11, "grad_norm": 4.082001209259033, "learning_rate": 1.3585858585858584e-07, "loss": 0.4771, "step": 8660 }, { "epoch": 3.11, "grad_norm": 3.742047071456909, "learning_rate": 1.3484848484848484e-07, "loss": 0.5773, "step": 8670 }, { "epoch": 3.11, "grad_norm": 6.419268608093262, "learning_rate": 1.3383838383838384e-07, "loss": 0.5903, "step": 8680 }, { "epoch": 3.12, "grad_norm": 5.8881940841674805, "learning_rate": 1.328282828282828e-07, "loss": 0.5279, "step": 8690 }, { "epoch": 3.12, "grad_norm": 4.6417999267578125, "learning_rate": 1.318181818181818e-07, "loss": 0.5368, "step": 8700 }, { "epoch": 3.12, "grad_norm": 5.411186218261719, "learning_rate": 1.308080808080808e-07, "loss": 0.5545, "step": 8710 }, { "epoch": 3.13, "grad_norm": 4.455116271972656, "learning_rate": 1.297979797979798e-07, "loss": 0.5631, "step": 8720 }, { "epoch": 3.13, "grad_norm": 11.706398963928223, "learning_rate": 1.2878787878787877e-07, "loss": 0.5218, "step": 8730 }, { "epoch": 3.13, "grad_norm": 6.45891809463501, "learning_rate": 1.2777777777777777e-07, "loss": 0.5383, "step": 8740 }, { "epoch": 3.14, "grad_norm": 4.817904949188232, "learning_rate": 1.2676767676767676e-07, "loss": 0.5093, "step": 8750 }, { "epoch": 3.14, "grad_norm": 7.290755748748779, "learning_rate": 1.2575757575757576e-07, "loss": 0.4842, "step": 8760 }, { "epoch": 3.15, "grad_norm": 3.76725435256958, "learning_rate": 1.2474747474747473e-07, "loss": 0.5476, "step": 8770 }, { "epoch": 3.15, "grad_norm": 5.706302642822266, "learning_rate": 1.2373737373737373e-07, "loss": 0.4709, "step": 8780 }, { "epoch": 3.15, "grad_norm": 7.164012432098389, "learning_rate": 1.2272727272727272e-07, "loss": 0.5088, "step": 8790 }, { "epoch": 3.16, "grad_norm": 5.86743688583374, "learning_rate": 1.2171717171717172e-07, "loss": 0.5361, "step": 8800 }, { "epoch": 3.16, "grad_norm": 4.16366720199585, "learning_rate": 1.207070707070707e-07, "loss": 0.536, "step": 8810 }, { "epoch": 3.16, "grad_norm": 6.925227642059326, "learning_rate": 1.196969696969697e-07, "loss": 0.5001, "step": 8820 }, { "epoch": 3.17, "grad_norm": 3.9916467666625977, "learning_rate": 1.1868686868686869e-07, "loss": 0.5662, "step": 8830 }, { "epoch": 3.17, "grad_norm": 7.0623931884765625, "learning_rate": 1.1767676767676767e-07, "loss": 0.5813, "step": 8840 }, { "epoch": 3.17, "grad_norm": 5.496065139770508, "learning_rate": 1.1666666666666667e-07, "loss": 0.5751, "step": 8850 }, { "epoch": 3.18, "grad_norm": 3.784182071685791, "learning_rate": 1.1565656565656565e-07, "loss": 0.5863, "step": 8860 }, { "epoch": 3.18, "grad_norm": 5.068544864654541, "learning_rate": 1.1464646464646465e-07, "loss": 0.5098, "step": 8870 }, { "epoch": 3.18, "grad_norm": 5.4960808753967285, "learning_rate": 1.1363636363636363e-07, "loss": 0.6326, "step": 8880 }, { "epoch": 3.19, "grad_norm": 4.9256792068481445, "learning_rate": 1.1262626262626263e-07, "loss": 0.4801, "step": 8890 }, { "epoch": 3.19, "grad_norm": 4.391407489776611, "learning_rate": 1.1161616161616161e-07, "loss": 0.5866, "step": 8900 }, { "epoch": 3.2, "grad_norm": 4.210151195526123, "learning_rate": 1.1060606060606061e-07, "loss": 0.5464, "step": 8910 }, { "epoch": 3.2, "grad_norm": 6.131332874298096, "learning_rate": 1.0959595959595959e-07, "loss": 0.4419, "step": 8920 }, { "epoch": 3.2, "grad_norm": 5.612468719482422, "learning_rate": 1.0858585858585859e-07, "loss": 0.6155, "step": 8930 }, { "epoch": 3.21, "grad_norm": 4.1157097816467285, "learning_rate": 1.0757575757575757e-07, "loss": 0.5514, "step": 8940 }, { "epoch": 3.21, "grad_norm": 5.19150972366333, "learning_rate": 1.0656565656565657e-07, "loss": 0.548, "step": 8950 }, { "epoch": 3.21, "grad_norm": 4.305140972137451, "learning_rate": 1.0555555555555555e-07, "loss": 0.5313, "step": 8960 }, { "epoch": 3.22, "grad_norm": 6.152969837188721, "learning_rate": 1.0454545454545454e-07, "loss": 0.4861, "step": 8970 }, { "epoch": 3.22, "grad_norm": 4.60936164855957, "learning_rate": 1.0353535353535353e-07, "loss": 0.6055, "step": 8980 }, { "epoch": 3.22, "grad_norm": 9.371330261230469, "learning_rate": 1.0252525252525252e-07, "loss": 0.5255, "step": 8990 }, { "epoch": 3.23, "grad_norm": 5.383856296539307, "learning_rate": 1.0151515151515151e-07, "loss": 0.4944, "step": 9000 }, { "epoch": 3.23, "eval_loss": 0.6957755088806152, "eval_runtime": 404.454, "eval_samples_per_second": 2.472, "eval_steps_per_second": 2.472, "step": 9000 }, { "epoch": 3.23, "grad_norm": 6.594298839569092, "learning_rate": 1.005050505050505e-07, "loss": 0.5506, "step": 9010 }, { "epoch": 3.23, "grad_norm": 5.5882248878479, "learning_rate": 9.94949494949495e-08, "loss": 0.4821, "step": 9020 }, { "epoch": 3.24, "grad_norm": 5.565232753753662, "learning_rate": 9.848484848484848e-08, "loss": 0.5903, "step": 9030 }, { "epoch": 3.24, "grad_norm": 4.196237564086914, "learning_rate": 9.747474747474747e-08, "loss": 0.4784, "step": 9040 }, { "epoch": 3.25, "grad_norm": 4.757313251495361, "learning_rate": 9.646464646464646e-08, "loss": 0.5256, "step": 9050 }, { "epoch": 3.25, "grad_norm": 5.351820945739746, "learning_rate": 9.545454545454546e-08, "loss": 0.4909, "step": 9060 }, { "epoch": 3.25, "grad_norm": 5.794235706329346, "learning_rate": 9.444444444444444e-08, "loss": 0.5444, "step": 9070 }, { "epoch": 3.26, "grad_norm": 6.105288028717041, "learning_rate": 9.343434343434344e-08, "loss": 0.5256, "step": 9080 }, { "epoch": 3.26, "grad_norm": 6.002346992492676, "learning_rate": 9.242424242424242e-08, "loss": 0.5005, "step": 9090 }, { "epoch": 3.26, "grad_norm": 5.016416072845459, "learning_rate": 9.141414141414142e-08, "loss": 0.5931, "step": 9100 }, { "epoch": 3.27, "grad_norm": 4.596400737762451, "learning_rate": 9.04040404040404e-08, "loss": 0.4898, "step": 9110 }, { "epoch": 3.27, "grad_norm": 5.950545310974121, "learning_rate": 8.93939393939394e-08, "loss": 0.6261, "step": 9120 }, { "epoch": 3.27, "grad_norm": 6.038389205932617, "learning_rate": 8.838383838383838e-08, "loss": 0.4946, "step": 9130 }, { "epoch": 3.28, "grad_norm": 6.918920516967773, "learning_rate": 8.737373737373736e-08, "loss": 0.5929, "step": 9140 }, { "epoch": 3.28, "grad_norm": 4.859712600708008, "learning_rate": 8.636363636363636e-08, "loss": 0.5885, "step": 9150 }, { "epoch": 3.28, "grad_norm": 8.945250511169434, "learning_rate": 8.535353535353534e-08, "loss": 0.5589, "step": 9160 }, { "epoch": 3.29, "grad_norm": 4.006341934204102, "learning_rate": 8.434343434343434e-08, "loss": 0.4796, "step": 9170 }, { "epoch": 3.29, "grad_norm": 6.528043746948242, "learning_rate": 8.333333333333333e-08, "loss": 0.5737, "step": 9180 }, { "epoch": 3.3, "grad_norm": 7.656773090362549, "learning_rate": 8.232323232323232e-08, "loss": 0.549, "step": 9190 }, { "epoch": 3.3, "grad_norm": 4.172515869140625, "learning_rate": 8.13131313131313e-08, "loss": 0.5112, "step": 9200 }, { "epoch": 3.3, "grad_norm": 6.6936116218566895, "learning_rate": 8.03030303030303e-08, "loss": 0.5645, "step": 9210 }, { "epoch": 3.31, "grad_norm": 7.118635177612305, "learning_rate": 7.929292929292929e-08, "loss": 0.5502, "step": 9220 }, { "epoch": 3.31, "grad_norm": 4.658480167388916, "learning_rate": 7.828282828282828e-08, "loss": 0.4883, "step": 9230 }, { "epoch": 3.31, "grad_norm": 7.470480442047119, "learning_rate": 7.727272727272727e-08, "loss": 0.5102, "step": 9240 }, { "epoch": 3.32, "grad_norm": 4.436034679412842, "learning_rate": 7.626262626262626e-08, "loss": 0.5939, "step": 9250 }, { "epoch": 3.32, "grad_norm": 4.717689037322998, "learning_rate": 7.525252525252525e-08, "loss": 0.5976, "step": 9260 }, { "epoch": 3.32, "grad_norm": 4.957371711730957, "learning_rate": 7.424242424242424e-08, "loss": 0.5019, "step": 9270 }, { "epoch": 3.33, "grad_norm": 4.873682022094727, "learning_rate": 7.323232323232323e-08, "loss": 0.5755, "step": 9280 }, { "epoch": 3.33, "grad_norm": 4.161667823791504, "learning_rate": 7.222222222222221e-08, "loss": 0.6371, "step": 9290 }, { "epoch": 3.34, "grad_norm": 4.539839744567871, "learning_rate": 7.121212121212121e-08, "loss": 0.6012, "step": 9300 }, { "epoch": 3.34, "grad_norm": 5.414936542510986, "learning_rate": 7.020202020202019e-08, "loss": 0.5449, "step": 9310 }, { "epoch": 3.34, "grad_norm": 10.87019157409668, "learning_rate": 6.919191919191919e-08, "loss": 0.5272, "step": 9320 }, { "epoch": 3.35, "grad_norm": 5.128973484039307, "learning_rate": 6.818181818181817e-08, "loss": 0.5053, "step": 9330 }, { "epoch": 3.35, "grad_norm": 4.943745136260986, "learning_rate": 6.717171717171717e-08, "loss": 0.5823, "step": 9340 }, { "epoch": 3.35, "grad_norm": 5.1109185218811035, "learning_rate": 6.616161616161615e-08, "loss": 0.5345, "step": 9350 }, { "epoch": 3.36, "grad_norm": 7.44246768951416, "learning_rate": 6.515151515151515e-08, "loss": 0.5677, "step": 9360 }, { "epoch": 3.36, "grad_norm": 4.263915538787842, "learning_rate": 6.414141414141413e-08, "loss": 0.522, "step": 9370 }, { "epoch": 3.36, "grad_norm": 5.933948516845703, "learning_rate": 6.313131313131313e-08, "loss": 0.4782, "step": 9380 }, { "epoch": 3.37, "grad_norm": 4.103212356567383, "learning_rate": 6.212121212121211e-08, "loss": 0.5602, "step": 9390 }, { "epoch": 3.37, "grad_norm": 5.865060806274414, "learning_rate": 6.111111111111111e-08, "loss": 0.5654, "step": 9400 }, { "epoch": 3.37, "grad_norm": 3.8256332874298096, "learning_rate": 6.01010101010101e-08, "loss": 0.5984, "step": 9410 }, { "epoch": 3.38, "grad_norm": 6.088188648223877, "learning_rate": 5.9090909090909085e-08, "loss": 0.5782, "step": 9420 }, { "epoch": 3.38, "grad_norm": 6.299181938171387, "learning_rate": 5.8080808080808076e-08, "loss": 0.5725, "step": 9430 }, { "epoch": 3.39, "grad_norm": 5.62983512878418, "learning_rate": 5.7070707070707066e-08, "loss": 0.5407, "step": 9440 }, { "epoch": 3.39, "grad_norm": 5.949933052062988, "learning_rate": 5.6060606060606056e-08, "loss": 0.4399, "step": 9450 }, { "epoch": 3.39, "grad_norm": 5.324888229370117, "learning_rate": 5.5050505050505046e-08, "loss": 0.5178, "step": 9460 }, { "epoch": 3.4, "grad_norm": 7.061882019042969, "learning_rate": 5.4040404040404037e-08, "loss": 0.5175, "step": 9470 }, { "epoch": 3.4, "grad_norm": 4.5119404792785645, "learning_rate": 5.303030303030303e-08, "loss": 0.5282, "step": 9480 }, { "epoch": 3.4, "grad_norm": 4.700840473175049, "learning_rate": 5.202020202020202e-08, "loss": 0.5202, "step": 9490 }, { "epoch": 3.41, "grad_norm": 5.251560211181641, "learning_rate": 5.101010101010101e-08, "loss": 0.5013, "step": 9500 }, { "epoch": 3.41, "grad_norm": 6.137795925140381, "learning_rate": 5e-08, "loss": 0.4514, "step": 9510 }, { "epoch": 3.41, "grad_norm": 6.373246669769287, "learning_rate": 4.898989898989899e-08, "loss": 0.5129, "step": 9520 }, { "epoch": 3.42, "grad_norm": 6.3960418701171875, "learning_rate": 4.797979797979797e-08, "loss": 0.5055, "step": 9530 }, { "epoch": 3.42, "grad_norm": 4.761574745178223, "learning_rate": 4.696969696969696e-08, "loss": 0.5724, "step": 9540 }, { "epoch": 3.42, "grad_norm": 4.658292770385742, "learning_rate": 4.595959595959595e-08, "loss": 0.5617, "step": 9550 }, { "epoch": 3.43, "grad_norm": 6.56446647644043, "learning_rate": 4.494949494949494e-08, "loss": 0.5495, "step": 9560 }, { "epoch": 3.43, "grad_norm": 5.917604446411133, "learning_rate": 4.393939393939393e-08, "loss": 0.5513, "step": 9570 }, { "epoch": 3.44, "grad_norm": 6.456116676330566, "learning_rate": 4.292929292929292e-08, "loss": 0.5198, "step": 9580 }, { "epoch": 3.44, "grad_norm": 5.084980487823486, "learning_rate": 4.1919191919191913e-08, "loss": 0.5737, "step": 9590 }, { "epoch": 3.44, "grad_norm": 3.8530325889587402, "learning_rate": 4.090909090909091e-08, "loss": 0.4901, "step": 9600 }, { "epoch": 3.45, "grad_norm": 4.385349750518799, "learning_rate": 3.98989898989899e-08, "loss": 0.5092, "step": 9610 }, { "epoch": 3.45, "grad_norm": 3.888326406478882, "learning_rate": 3.888888888888889e-08, "loss": 0.5536, "step": 9620 }, { "epoch": 3.45, "grad_norm": 4.46201229095459, "learning_rate": 3.787878787878788e-08, "loss": 0.5697, "step": 9630 }, { "epoch": 3.46, "grad_norm": 7.194265842437744, "learning_rate": 3.686868686868687e-08, "loss": 0.4886, "step": 9640 }, { "epoch": 3.46, "grad_norm": 6.102540016174316, "learning_rate": 3.585858585858586e-08, "loss": 0.4981, "step": 9650 }, { "epoch": 3.46, "grad_norm": 7.469662189483643, "learning_rate": 3.484848484848485e-08, "loss": 0.5294, "step": 9660 }, { "epoch": 3.47, "grad_norm": 4.181239128112793, "learning_rate": 3.383838383838384e-08, "loss": 0.5463, "step": 9670 }, { "epoch": 3.47, "grad_norm": 5.108190536499023, "learning_rate": 3.282828282828283e-08, "loss": 0.4633, "step": 9680 }, { "epoch": 3.47, "grad_norm": 5.501590251922607, "learning_rate": 3.1818181818181816e-08, "loss": 0.5121, "step": 9690 }, { "epoch": 3.48, "grad_norm": 6.437745571136475, "learning_rate": 3.0808080808080806e-08, "loss": 0.475, "step": 9700 }, { "epoch": 3.48, "grad_norm": 7.1137166023254395, "learning_rate": 2.9797979797979797e-08, "loss": 0.5536, "step": 9710 }, { "epoch": 3.49, "grad_norm": 5.652755260467529, "learning_rate": 2.8787878787878787e-08, "loss": 0.5493, "step": 9720 }, { "epoch": 3.49, "grad_norm": 4.296222686767578, "learning_rate": 2.7777777777777774e-08, "loss": 0.5323, "step": 9730 }, { "epoch": 3.49, "grad_norm": 5.550398826599121, "learning_rate": 2.6767676767676764e-08, "loss": 0.5555, "step": 9740 }, { "epoch": 3.5, "grad_norm": 6.785501956939697, "learning_rate": 2.5757575757575755e-08, "loss": 0.5388, "step": 9750 }, { "epoch": 3.5, "grad_norm": 7.063258171081543, "learning_rate": 2.4747474747474745e-08, "loss": 0.5742, "step": 9760 }, { "epoch": 3.5, "grad_norm": 5.894256114959717, "learning_rate": 2.3737373737373735e-08, "loss": 0.6248, "step": 9770 }, { "epoch": 3.51, "grad_norm": 5.81352424621582, "learning_rate": 2.2727272727272725e-08, "loss": 0.4896, "step": 9780 }, { "epoch": 3.51, "grad_norm": 5.251163482666016, "learning_rate": 2.1717171717171716e-08, "loss": 0.5633, "step": 9790 }, { "epoch": 3.51, "grad_norm": 4.963428020477295, "learning_rate": 2.070707070707071e-08, "loss": 0.4617, "step": 9800 }, { "epoch": 3.52, "grad_norm": 4.8795928955078125, "learning_rate": 1.9696969696969696e-08, "loss": 0.5325, "step": 9810 }, { "epoch": 3.52, "grad_norm": 5.59514856338501, "learning_rate": 1.8686868686868686e-08, "loss": 0.5535, "step": 9820 }, { "epoch": 3.53, "grad_norm": 5.933664321899414, "learning_rate": 1.7676767676767677e-08, "loss": 0.4904, "step": 9830 }, { "epoch": 3.53, "grad_norm": 5.348793029785156, "learning_rate": 1.6666666666666667e-08, "loss": 0.543, "step": 9840 }, { "epoch": 3.53, "grad_norm": 4.590538024902344, "learning_rate": 1.5656565656565657e-08, "loss": 0.5443, "step": 9850 }, { "epoch": 3.54, "grad_norm": 7.6259918212890625, "learning_rate": 1.4646464646464646e-08, "loss": 0.4999, "step": 9860 }, { "epoch": 3.54, "grad_norm": 5.3782124519348145, "learning_rate": 1.3636363636363635e-08, "loss": 0.5064, "step": 9870 }, { "epoch": 3.54, "grad_norm": 6.371192932128906, "learning_rate": 1.2626262626262625e-08, "loss": 0.5487, "step": 9880 }, { "epoch": 3.55, "grad_norm": 5.356314659118652, "learning_rate": 1.1616161616161615e-08, "loss": 0.506, "step": 9890 }, { "epoch": 3.55, "grad_norm": 5.2767014503479, "learning_rate": 1.0606060606060607e-08, "loss": 0.5556, "step": 9900 }, { "epoch": 3.55, "grad_norm": 4.274359226226807, "learning_rate": 9.595959595959596e-09, "loss": 0.5074, "step": 9910 }, { "epoch": 3.56, "grad_norm": 5.464534759521484, "learning_rate": 8.585858585858586e-09, "loss": 0.4612, "step": 9920 }, { "epoch": 3.56, "grad_norm": 9.630477905273438, "learning_rate": 7.575757575757576e-09, "loss": 0.6037, "step": 9930 }, { "epoch": 3.56, "grad_norm": 3.692291259765625, "learning_rate": 6.565656565656565e-09, "loss": 0.5116, "step": 9940 }, { "epoch": 3.57, "grad_norm": 5.269135475158691, "learning_rate": 5.555555555555555e-09, "loss": 0.591, "step": 9950 }, { "epoch": 3.57, "grad_norm": 5.141294956207275, "learning_rate": 4.545454545454545e-09, "loss": 0.5726, "step": 9960 }, { "epoch": 3.58, "grad_norm": 6.49954080581665, "learning_rate": 3.5353535353535353e-09, "loss": 0.5509, "step": 9970 }, { "epoch": 3.58, "grad_norm": 5.085234642028809, "learning_rate": 2.5252525252525255e-09, "loss": 0.5337, "step": 9980 }, { "epoch": 3.58, "grad_norm": 5.688786506652832, "learning_rate": 1.515151515151515e-09, "loss": 0.5624, "step": 9990 }, { "epoch": 3.59, "grad_norm": 6.861104965209961, "learning_rate": 5.050505050505051e-10, "loss": 0.5167, "step": 10000 }, { "epoch": 3.59, "eval_loss": 0.711556613445282, "eval_runtime": 405.1832, "eval_samples_per_second": 2.468, "eval_steps_per_second": 2.468, "step": 10000 } ], "logging_steps": 10, "max_steps": 10000, "num_input_tokens_seen": 0, "num_train_epochs": 4, "save_steps": 2000, "total_flos": 9.4282098671616e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }