{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 3108, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0009652509652509653, "grad_norm": 37.57219314575195, "learning_rate": 2.1276595744680852e-07, "loss": 2.9127, "step": 1 }, { "epoch": 0.0019305019305019305, "grad_norm": 34.530269622802734, "learning_rate": 4.2553191489361704e-07, "loss": 2.9575, "step": 2 }, { "epoch": 0.0028957528957528956, "grad_norm": 31.692672729492188, "learning_rate": 6.382978723404255e-07, "loss": 2.8706, "step": 3 }, { "epoch": 0.003861003861003861, "grad_norm": 31.64986801147461, "learning_rate": 8.510638297872341e-07, "loss": 2.9322, "step": 4 }, { "epoch": 0.004826254826254826, "grad_norm": 27.894390106201172, "learning_rate": 1.0638297872340427e-06, "loss": 2.7927, "step": 5 }, { "epoch": 0.005791505791505791, "grad_norm": 24.181001663208008, "learning_rate": 1.276595744680851e-06, "loss": 2.8115, "step": 6 }, { "epoch": 0.006756756756756757, "grad_norm": 20.626800537109375, "learning_rate": 1.4893617021276596e-06, "loss": 2.6941, "step": 7 }, { "epoch": 0.007722007722007722, "grad_norm": 13.248446464538574, "learning_rate": 1.7021276595744682e-06, "loss": 2.5758, "step": 8 }, { "epoch": 0.008687258687258687, "grad_norm": 15.200055122375488, "learning_rate": 1.9148936170212767e-06, "loss": 2.5315, "step": 9 }, { "epoch": 0.009652509652509652, "grad_norm": 13.172714233398438, "learning_rate": 2.1276595744680853e-06, "loss": 2.5467, "step": 10 }, { "epoch": 0.010617760617760617, "grad_norm": 16.15408706665039, "learning_rate": 2.340425531914894e-06, "loss": 2.4509, "step": 11 }, { "epoch": 0.011583011583011582, "grad_norm": 17.439682006835938, "learning_rate": 2.553191489361702e-06, "loss": 2.5752, "step": 12 }, { "epoch": 0.012548262548262547, "grad_norm": 15.113956451416016, "learning_rate": 2.765957446808511e-06, "loss": 2.4805, "step": 13 }, { "epoch": 0.013513513513513514, "grad_norm": 13.25255012512207, "learning_rate": 2.978723404255319e-06, "loss": 2.4514, "step": 14 }, { "epoch": 0.01447876447876448, "grad_norm": 10.589703559875488, "learning_rate": 3.191489361702128e-06, "loss": 2.3685, "step": 15 }, { "epoch": 0.015444015444015444, "grad_norm": 8.39791488647461, "learning_rate": 3.4042553191489363e-06, "loss": 2.2548, "step": 16 }, { "epoch": 0.016409266409266408, "grad_norm": 11.414656639099121, "learning_rate": 3.6170212765957453e-06, "loss": 2.2782, "step": 17 }, { "epoch": 0.017374517374517374, "grad_norm": 17.181034088134766, "learning_rate": 3.8297872340425535e-06, "loss": 2.2162, "step": 18 }, { "epoch": 0.01833976833976834, "grad_norm": 9.750289916992188, "learning_rate": 4.042553191489362e-06, "loss": 2.2657, "step": 19 }, { "epoch": 0.019305019305019305, "grad_norm": 6.164453506469727, "learning_rate": 4.255319148936171e-06, "loss": 2.1871, "step": 20 }, { "epoch": 0.02027027027027027, "grad_norm": 7.542515754699707, "learning_rate": 4.468085106382979e-06, "loss": 2.2052, "step": 21 }, { "epoch": 0.021235521235521235, "grad_norm": 9.258981704711914, "learning_rate": 4.680851063829788e-06, "loss": 2.1841, "step": 22 }, { "epoch": 0.0222007722007722, "grad_norm": 8.692361831665039, "learning_rate": 4.893617021276596e-06, "loss": 2.1692, "step": 23 }, { "epoch": 0.023166023166023165, "grad_norm": 8.432202339172363, "learning_rate": 5.106382978723404e-06, "loss": 2.1326, "step": 24 }, { "epoch": 0.02413127413127413, "grad_norm": 6.449735164642334, "learning_rate": 5.319148936170213e-06, "loss": 2.1137, "step": 25 }, { "epoch": 0.025096525096525095, "grad_norm": 6.142323017120361, "learning_rate": 5.531914893617022e-06, "loss": 2.0913, "step": 26 }, { "epoch": 0.026061776061776062, "grad_norm": 7.893868923187256, "learning_rate": 5.744680851063831e-06, "loss": 2.0388, "step": 27 }, { "epoch": 0.02702702702702703, "grad_norm": 7.176388263702393, "learning_rate": 5.957446808510638e-06, "loss": 2.0789, "step": 28 }, { "epoch": 0.027992277992277992, "grad_norm": 6.881705284118652, "learning_rate": 6.170212765957447e-06, "loss": 2.128, "step": 29 }, { "epoch": 0.02895752895752896, "grad_norm": 5.771139621734619, "learning_rate": 6.382978723404256e-06, "loss": 2.0329, "step": 30 }, { "epoch": 0.029922779922779922, "grad_norm": 5.3250627517700195, "learning_rate": 6.595744680851064e-06, "loss": 2.0776, "step": 31 }, { "epoch": 0.03088803088803089, "grad_norm": 5.021093845367432, "learning_rate": 6.808510638297873e-06, "loss": 2.0388, "step": 32 }, { "epoch": 0.03185328185328185, "grad_norm": 4.986810684204102, "learning_rate": 7.021276595744682e-06, "loss": 2.0429, "step": 33 }, { "epoch": 0.032818532818532815, "grad_norm": 5.753068447113037, "learning_rate": 7.234042553191491e-06, "loss": 2.0471, "step": 34 }, { "epoch": 0.033783783783783786, "grad_norm": 6.062497615814209, "learning_rate": 7.446808510638298e-06, "loss": 2.0988, "step": 35 }, { "epoch": 0.03474903474903475, "grad_norm": 5.042311191558838, "learning_rate": 7.659574468085107e-06, "loss": 2.0457, "step": 36 }, { "epoch": 0.03571428571428571, "grad_norm": 5.41193151473999, "learning_rate": 7.872340425531916e-06, "loss": 1.972, "step": 37 }, { "epoch": 0.03667953667953668, "grad_norm": 4.624916076660156, "learning_rate": 8.085106382978723e-06, "loss": 1.9425, "step": 38 }, { "epoch": 0.037644787644787646, "grad_norm": 5.057003498077393, "learning_rate": 8.297872340425532e-06, "loss": 1.9985, "step": 39 }, { "epoch": 0.03861003861003861, "grad_norm": 5.198980331420898, "learning_rate": 8.510638297872341e-06, "loss": 1.9534, "step": 40 }, { "epoch": 0.03957528957528957, "grad_norm": 5.4440836906433105, "learning_rate": 8.72340425531915e-06, "loss": 1.9787, "step": 41 }, { "epoch": 0.04054054054054054, "grad_norm": 4.812963008880615, "learning_rate": 8.936170212765958e-06, "loss": 1.9238, "step": 42 }, { "epoch": 0.041505791505791506, "grad_norm": 5.2607831954956055, "learning_rate": 9.148936170212767e-06, "loss": 2.0111, "step": 43 }, { "epoch": 0.04247104247104247, "grad_norm": 4.675182819366455, "learning_rate": 9.361702127659576e-06, "loss": 1.9967, "step": 44 }, { "epoch": 0.04343629343629344, "grad_norm": 5.039648532867432, "learning_rate": 9.574468085106385e-06, "loss": 1.8747, "step": 45 }, { "epoch": 0.0444015444015444, "grad_norm": 5.738279819488525, "learning_rate": 9.787234042553192e-06, "loss": 1.9386, "step": 46 }, { "epoch": 0.045366795366795366, "grad_norm": 5.404884338378906, "learning_rate": 1e-05, "loss": 1.9091, "step": 47 }, { "epoch": 0.04633204633204633, "grad_norm": 5.5826735496521, "learning_rate": 1.0212765957446808e-05, "loss": 1.9017, "step": 48 }, { "epoch": 0.0472972972972973, "grad_norm": 4.8203020095825195, "learning_rate": 1.0425531914893619e-05, "loss": 1.951, "step": 49 }, { "epoch": 0.04826254826254826, "grad_norm": 5.30726957321167, "learning_rate": 1.0638297872340426e-05, "loss": 1.9703, "step": 50 }, { "epoch": 0.04922779922779923, "grad_norm": 4.507512092590332, "learning_rate": 1.0851063829787233e-05, "loss": 1.857, "step": 51 }, { "epoch": 0.05019305019305019, "grad_norm": 5.3983869552612305, "learning_rate": 1.1063829787234044e-05, "loss": 1.967, "step": 52 }, { "epoch": 0.05115830115830116, "grad_norm": 5.134077072143555, "learning_rate": 1.1276595744680851e-05, "loss": 1.8909, "step": 53 }, { "epoch": 0.052123552123552123, "grad_norm": 5.454338073730469, "learning_rate": 1.1489361702127662e-05, "loss": 1.8437, "step": 54 }, { "epoch": 0.05308880308880309, "grad_norm": 5.207176685333252, "learning_rate": 1.170212765957447e-05, "loss": 1.9095, "step": 55 }, { "epoch": 0.05405405405405406, "grad_norm": 5.950668811798096, "learning_rate": 1.1914893617021277e-05, "loss": 1.9094, "step": 56 }, { "epoch": 0.05501930501930502, "grad_norm": 4.508608341217041, "learning_rate": 1.2127659574468087e-05, "loss": 1.8993, "step": 57 }, { "epoch": 0.055984555984555984, "grad_norm": 5.785097122192383, "learning_rate": 1.2340425531914895e-05, "loss": 1.8405, "step": 58 }, { "epoch": 0.05694980694980695, "grad_norm": 4.732987880706787, "learning_rate": 1.2553191489361702e-05, "loss": 1.9891, "step": 59 }, { "epoch": 0.05791505791505792, "grad_norm": 4.726958751678467, "learning_rate": 1.2765957446808513e-05, "loss": 1.9444, "step": 60 }, { "epoch": 0.05888030888030888, "grad_norm": 5.205662727355957, "learning_rate": 1.297872340425532e-05, "loss": 1.838, "step": 61 }, { "epoch": 0.059845559845559844, "grad_norm": 5.053771495819092, "learning_rate": 1.3191489361702127e-05, "loss": 1.9335, "step": 62 }, { "epoch": 0.060810810810810814, "grad_norm": 4.497619152069092, "learning_rate": 1.3404255319148938e-05, "loss": 1.8424, "step": 63 }, { "epoch": 0.06177606177606178, "grad_norm": 4.811068534851074, "learning_rate": 1.3617021276595745e-05, "loss": 1.8365, "step": 64 }, { "epoch": 0.06274131274131274, "grad_norm": 4.842855930328369, "learning_rate": 1.3829787234042556e-05, "loss": 1.825, "step": 65 }, { "epoch": 0.0637065637065637, "grad_norm": 5.438333988189697, "learning_rate": 1.4042553191489363e-05, "loss": 1.848, "step": 66 }, { "epoch": 0.06467181467181467, "grad_norm": 5.783076286315918, "learning_rate": 1.425531914893617e-05, "loss": 1.8859, "step": 67 }, { "epoch": 0.06563706563706563, "grad_norm": 6.11845588684082, "learning_rate": 1.4468085106382981e-05, "loss": 1.8958, "step": 68 }, { "epoch": 0.06660231660231661, "grad_norm": 6.755260467529297, "learning_rate": 1.4680851063829789e-05, "loss": 1.8746, "step": 69 }, { "epoch": 0.06756756756756757, "grad_norm": 4.893709659576416, "learning_rate": 1.4893617021276596e-05, "loss": 1.872, "step": 70 }, { "epoch": 0.06853281853281853, "grad_norm": 5.599435806274414, "learning_rate": 1.5106382978723407e-05, "loss": 1.8361, "step": 71 }, { "epoch": 0.0694980694980695, "grad_norm": 6.107633113861084, "learning_rate": 1.5319148936170214e-05, "loss": 1.8572, "step": 72 }, { "epoch": 0.07046332046332046, "grad_norm": 4.610060691833496, "learning_rate": 1.5531914893617023e-05, "loss": 1.9096, "step": 73 }, { "epoch": 0.07142857142857142, "grad_norm": 5.756293296813965, "learning_rate": 1.5744680851063832e-05, "loss": 1.9358, "step": 74 }, { "epoch": 0.07239382239382239, "grad_norm": 4.8217573165893555, "learning_rate": 1.595744680851064e-05, "loss": 1.839, "step": 75 }, { "epoch": 0.07335907335907337, "grad_norm": 4.5468010902404785, "learning_rate": 1.6170212765957446e-05, "loss": 1.8764, "step": 76 }, { "epoch": 0.07432432432432433, "grad_norm": 4.610559940338135, "learning_rate": 1.6382978723404255e-05, "loss": 1.7964, "step": 77 }, { "epoch": 0.07528957528957529, "grad_norm": 4.943978786468506, "learning_rate": 1.6595744680851064e-05, "loss": 1.8379, "step": 78 }, { "epoch": 0.07625482625482626, "grad_norm": 4.75681209564209, "learning_rate": 1.6808510638297873e-05, "loss": 1.8266, "step": 79 }, { "epoch": 0.07722007722007722, "grad_norm": 5.278621196746826, "learning_rate": 1.7021276595744682e-05, "loss": 1.8639, "step": 80 }, { "epoch": 0.07818532818532818, "grad_norm": 4.590597629547119, "learning_rate": 1.723404255319149e-05, "loss": 1.8454, "step": 81 }, { "epoch": 0.07915057915057915, "grad_norm": 4.689551830291748, "learning_rate": 1.74468085106383e-05, "loss": 1.8563, "step": 82 }, { "epoch": 0.08011583011583012, "grad_norm": 5.145448207855225, "learning_rate": 1.765957446808511e-05, "loss": 1.91, "step": 83 }, { "epoch": 0.08108108108108109, "grad_norm": 5.2121968269348145, "learning_rate": 1.7872340425531915e-05, "loss": 1.8853, "step": 84 }, { "epoch": 0.08204633204633205, "grad_norm": 6.532495021820068, "learning_rate": 1.8085106382978724e-05, "loss": 1.923, "step": 85 }, { "epoch": 0.08301158301158301, "grad_norm": 6.2555251121521, "learning_rate": 1.8297872340425533e-05, "loss": 1.8334, "step": 86 }, { "epoch": 0.08397683397683398, "grad_norm": 5.264360427856445, "learning_rate": 1.8510638297872342e-05, "loss": 1.8977, "step": 87 }, { "epoch": 0.08494208494208494, "grad_norm": 6.3229756355285645, "learning_rate": 1.872340425531915e-05, "loss": 1.9133, "step": 88 }, { "epoch": 0.0859073359073359, "grad_norm": 4.155460834503174, "learning_rate": 1.893617021276596e-05, "loss": 1.8972, "step": 89 }, { "epoch": 0.08687258687258688, "grad_norm": 4.465296268463135, "learning_rate": 1.914893617021277e-05, "loss": 1.7892, "step": 90 }, { "epoch": 0.08783783783783784, "grad_norm": 5.7993388175964355, "learning_rate": 1.9361702127659575e-05, "loss": 1.8372, "step": 91 }, { "epoch": 0.0888030888030888, "grad_norm": 4.296961784362793, "learning_rate": 1.9574468085106384e-05, "loss": 1.8955, "step": 92 }, { "epoch": 0.08976833976833977, "grad_norm": 6.245906829833984, "learning_rate": 1.9787234042553193e-05, "loss": 1.8269, "step": 93 }, { "epoch": 0.09073359073359073, "grad_norm": 4.545825004577637, "learning_rate": 2e-05, "loss": 1.9008, "step": 94 }, { "epoch": 0.0916988416988417, "grad_norm": 7.551859378814697, "learning_rate": 1.999999456770665e-05, "loss": 1.8902, "step": 95 }, { "epoch": 0.09266409266409266, "grad_norm": 4.8623504638671875, "learning_rate": 1.9999978270832498e-05, "loss": 1.798, "step": 96 }, { "epoch": 0.09362934362934362, "grad_norm": 6.84691047668457, "learning_rate": 1.9999951109395245e-05, "loss": 1.8671, "step": 97 }, { "epoch": 0.0945945945945946, "grad_norm": 5.37424373626709, "learning_rate": 1.999991308342441e-05, "loss": 1.928, "step": 98 }, { "epoch": 0.09555984555984556, "grad_norm": 7.096020221710205, "learning_rate": 1.99998641929613e-05, "loss": 1.7758, "step": 99 }, { "epoch": 0.09652509652509653, "grad_norm": 4.507432460784912, "learning_rate": 1.9999804438059042e-05, "loss": 1.7964, "step": 100 }, { "epoch": 0.09749034749034749, "grad_norm": 7.82309103012085, "learning_rate": 1.9999733818782543e-05, "loss": 1.8715, "step": 101 }, { "epoch": 0.09845559845559845, "grad_norm": 4.694272518157959, "learning_rate": 1.999965233520854e-05, "loss": 1.814, "step": 102 }, { "epoch": 0.09942084942084942, "grad_norm": 6.026909828186035, "learning_rate": 1.9999559987425553e-05, "loss": 1.876, "step": 103 }, { "epoch": 0.10038610038610038, "grad_norm": 4.93021297454834, "learning_rate": 1.999945677553392e-05, "loss": 1.846, "step": 104 }, { "epoch": 0.10135135135135136, "grad_norm": 5.0024333000183105, "learning_rate": 1.9999342699645774e-05, "loss": 1.8466, "step": 105 }, { "epoch": 0.10231660231660232, "grad_norm": 5.216198444366455, "learning_rate": 1.999921775988505e-05, "loss": 1.9478, "step": 106 }, { "epoch": 0.10328185328185328, "grad_norm": 4.252792835235596, "learning_rate": 1.99990819563875e-05, "loss": 1.8523, "step": 107 }, { "epoch": 0.10424710424710425, "grad_norm": 5.397847652435303, "learning_rate": 1.999893528930066e-05, "loss": 1.8539, "step": 108 }, { "epoch": 0.10521235521235521, "grad_norm": 4.538277626037598, "learning_rate": 1.999877775878388e-05, "loss": 1.8018, "step": 109 }, { "epoch": 0.10617760617760617, "grad_norm": 4.547168731689453, "learning_rate": 1.9998609365008313e-05, "loss": 1.7968, "step": 110 }, { "epoch": 0.10714285714285714, "grad_norm": 4.570284366607666, "learning_rate": 1.999843010815691e-05, "loss": 1.737, "step": 111 }, { "epoch": 0.10810810810810811, "grad_norm": 5.048513412475586, "learning_rate": 1.9998239988424425e-05, "loss": 1.8962, "step": 112 }, { "epoch": 0.10907335907335908, "grad_norm": 4.87770938873291, "learning_rate": 1.9998039006017414e-05, "loss": 1.7879, "step": 113 }, { "epoch": 0.11003861003861004, "grad_norm": 4.417995929718018, "learning_rate": 1.999782716115424e-05, "loss": 1.8342, "step": 114 }, { "epoch": 0.111003861003861, "grad_norm": 5.057574272155762, "learning_rate": 1.9997604454065063e-05, "loss": 1.8057, "step": 115 }, { "epoch": 0.11196911196911197, "grad_norm": 4.49242639541626, "learning_rate": 1.9997370884991842e-05, "loss": 1.9251, "step": 116 }, { "epoch": 0.11293436293436293, "grad_norm": 8.180551528930664, "learning_rate": 1.9997126454188343e-05, "loss": 1.8771, "step": 117 }, { "epoch": 0.1138996138996139, "grad_norm": 4.551602840423584, "learning_rate": 1.999687116192013e-05, "loss": 1.8492, "step": 118 }, { "epoch": 0.11486486486486487, "grad_norm": 8.785762786865234, "learning_rate": 1.9996605008464564e-05, "loss": 1.8796, "step": 119 }, { "epoch": 0.11583011583011583, "grad_norm": 4.232022762298584, "learning_rate": 1.999632799411081e-05, "loss": 1.85, "step": 120 }, { "epoch": 0.1167953667953668, "grad_norm": 8.051064491271973, "learning_rate": 1.9996040119159842e-05, "loss": 1.7762, "step": 121 }, { "epoch": 0.11776061776061776, "grad_norm": 4.3684000968933105, "learning_rate": 1.9995741383924413e-05, "loss": 1.8294, "step": 122 }, { "epoch": 0.11872586872586872, "grad_norm": 7.8162841796875, "learning_rate": 1.999543178872909e-05, "loss": 1.8098, "step": 123 }, { "epoch": 0.11969111969111969, "grad_norm": 4.503332138061523, "learning_rate": 1.9995111333910238e-05, "loss": 1.84, "step": 124 }, { "epoch": 0.12065637065637065, "grad_norm": 8.560187339782715, "learning_rate": 1.999478001981602e-05, "loss": 1.8667, "step": 125 }, { "epoch": 0.12162162162162163, "grad_norm": 4.4390788078308105, "learning_rate": 1.9994437846806384e-05, "loss": 1.8341, "step": 126 }, { "epoch": 0.12258687258687259, "grad_norm": 5.524573802947998, "learning_rate": 1.99940848152531e-05, "loss": 1.8746, "step": 127 }, { "epoch": 0.12355212355212356, "grad_norm": 4.953189849853516, "learning_rate": 1.999372092553971e-05, "loss": 1.8883, "step": 128 }, { "epoch": 0.12451737451737452, "grad_norm": 4.29864501953125, "learning_rate": 1.9993346178061572e-05, "loss": 1.7821, "step": 129 }, { "epoch": 0.12548262548262548, "grad_norm": 5.282871246337891, "learning_rate": 1.9992960573225837e-05, "loss": 1.7862, "step": 130 }, { "epoch": 0.12644787644787644, "grad_norm": 4.665815353393555, "learning_rate": 1.9992564111451444e-05, "loss": 1.8474, "step": 131 }, { "epoch": 0.1274131274131274, "grad_norm": 4.723055839538574, "learning_rate": 1.999215679316913e-05, "loss": 1.838, "step": 132 }, { "epoch": 0.12837837837837837, "grad_norm": 4.902134418487549, "learning_rate": 1.9991738618821435e-05, "loss": 1.8074, "step": 133 }, { "epoch": 0.12934362934362933, "grad_norm": 4.841723918914795, "learning_rate": 1.9991309588862683e-05, "loss": 1.8605, "step": 134 }, { "epoch": 0.1303088803088803, "grad_norm": 4.6323723793029785, "learning_rate": 1.9990869703759e-05, "loss": 1.8063, "step": 135 }, { "epoch": 0.13127413127413126, "grad_norm": 4.535574436187744, "learning_rate": 1.9990418963988306e-05, "loss": 1.881, "step": 136 }, { "epoch": 0.13223938223938225, "grad_norm": 4.749152183532715, "learning_rate": 1.9989957370040305e-05, "loss": 1.7487, "step": 137 }, { "epoch": 0.13320463320463322, "grad_norm": 4.237002849578857, "learning_rate": 1.9989484922416503e-05, "loss": 1.8002, "step": 138 }, { "epoch": 0.13416988416988418, "grad_norm": 4.75019645690918, "learning_rate": 1.9989001621630196e-05, "loss": 1.8117, "step": 139 }, { "epoch": 0.13513513513513514, "grad_norm": 5.378240585327148, "learning_rate": 1.998850746820647e-05, "loss": 1.8826, "step": 140 }, { "epoch": 0.1361003861003861, "grad_norm": 5.388830661773682, "learning_rate": 1.9988002462682198e-05, "loss": 1.8002, "step": 141 }, { "epoch": 0.13706563706563707, "grad_norm": 5.59326171875, "learning_rate": 1.998748660560605e-05, "loss": 1.8897, "step": 142 }, { "epoch": 0.13803088803088803, "grad_norm": 5.555648326873779, "learning_rate": 1.9986959897538488e-05, "loss": 1.8162, "step": 143 }, { "epoch": 0.138996138996139, "grad_norm": 7.1515350341796875, "learning_rate": 1.9986422339051753e-05, "loss": 1.741, "step": 144 }, { "epoch": 0.13996138996138996, "grad_norm": 5.384510040283203, "learning_rate": 1.9985873930729883e-05, "loss": 1.8948, "step": 145 }, { "epoch": 0.14092664092664092, "grad_norm": 7.778863430023193, "learning_rate": 1.9985314673168696e-05, "loss": 1.8543, "step": 146 }, { "epoch": 0.14189189189189189, "grad_norm": 6.057487964630127, "learning_rate": 1.998474456697581e-05, "loss": 1.8448, "step": 147 }, { "epoch": 0.14285714285714285, "grad_norm": 5.474917888641357, "learning_rate": 1.9984163612770617e-05, "loss": 1.7643, "step": 148 }, { "epoch": 0.1438223938223938, "grad_norm": 7.61667013168335, "learning_rate": 1.9983571811184297e-05, "loss": 1.8911, "step": 149 }, { "epoch": 0.14478764478764478, "grad_norm": 4.5976643562316895, "learning_rate": 1.9982969162859828e-05, "loss": 1.815, "step": 150 }, { "epoch": 0.14575289575289574, "grad_norm": 6.352999687194824, "learning_rate": 1.9982355668451948e-05, "loss": 1.8259, "step": 151 }, { "epoch": 0.14671814671814673, "grad_norm": 5.403079509735107, "learning_rate": 1.9981731328627208e-05, "loss": 1.7196, "step": 152 }, { "epoch": 0.1476833976833977, "grad_norm": 4.54139518737793, "learning_rate": 1.9981096144063916e-05, "loss": 1.8072, "step": 153 }, { "epoch": 0.14864864864864866, "grad_norm": 4.9963507652282715, "learning_rate": 1.9980450115452183e-05, "loss": 1.8265, "step": 154 }, { "epoch": 0.14961389961389962, "grad_norm": 4.345424175262451, "learning_rate": 1.9979793243493887e-05, "loss": 1.8207, "step": 155 }, { "epoch": 0.15057915057915058, "grad_norm": 4.619804859161377, "learning_rate": 1.9979125528902688e-05, "loss": 1.7745, "step": 156 }, { "epoch": 0.15154440154440155, "grad_norm": 5.153815746307373, "learning_rate": 1.9978446972404043e-05, "loss": 1.8312, "step": 157 }, { "epoch": 0.1525096525096525, "grad_norm": 4.700374126434326, "learning_rate": 1.9977757574735164e-05, "loss": 1.8808, "step": 158 }, { "epoch": 0.15347490347490347, "grad_norm": 4.401470184326172, "learning_rate": 1.9977057336645057e-05, "loss": 1.8824, "step": 159 }, { "epoch": 0.15444015444015444, "grad_norm": 4.030679225921631, "learning_rate": 1.9976346258894502e-05, "loss": 1.8622, "step": 160 }, { "epoch": 0.1554054054054054, "grad_norm": 5.170019149780273, "learning_rate": 1.9975624342256057e-05, "loss": 1.8153, "step": 161 }, { "epoch": 0.15637065637065636, "grad_norm": 4.371926307678223, "learning_rate": 1.997489158751405e-05, "loss": 1.8585, "step": 162 }, { "epoch": 0.15733590733590733, "grad_norm": 5.159310340881348, "learning_rate": 1.9974147995464593e-05, "loss": 1.838, "step": 163 }, { "epoch": 0.1583011583011583, "grad_norm": 4.847809314727783, "learning_rate": 1.997339356691557e-05, "loss": 1.8734, "step": 164 }, { "epoch": 0.15926640926640925, "grad_norm": 4.6886305809021, "learning_rate": 1.997262830268663e-05, "loss": 1.7985, "step": 165 }, { "epoch": 0.16023166023166024, "grad_norm": 4.294079780578613, "learning_rate": 1.9971852203609204e-05, "loss": 1.8025, "step": 166 }, { "epoch": 0.1611969111969112, "grad_norm": 5.942923069000244, "learning_rate": 1.9971065270526494e-05, "loss": 1.8206, "step": 167 }, { "epoch": 0.16216216216216217, "grad_norm": 4.145232200622559, "learning_rate": 1.9970267504293466e-05, "loss": 1.8218, "step": 168 }, { "epoch": 0.16312741312741313, "grad_norm": 5.085411548614502, "learning_rate": 1.996945890577686e-05, "loss": 1.8176, "step": 169 }, { "epoch": 0.1640926640926641, "grad_norm": 5.151740074157715, "learning_rate": 1.9968639475855192e-05, "loss": 1.8552, "step": 170 }, { "epoch": 0.16505791505791506, "grad_norm": 4.554969310760498, "learning_rate": 1.996780921541873e-05, "loss": 1.8348, "step": 171 }, { "epoch": 0.16602316602316602, "grad_norm": 5.52618408203125, "learning_rate": 1.9966968125369522e-05, "loss": 1.8799, "step": 172 }, { "epoch": 0.166988416988417, "grad_norm": 4.414917469024658, "learning_rate": 1.996611620662138e-05, "loss": 1.7495, "step": 173 }, { "epoch": 0.16795366795366795, "grad_norm": 4.095779895782471, "learning_rate": 1.9965253460099872e-05, "loss": 1.7518, "step": 174 }, { "epoch": 0.16891891891891891, "grad_norm": 4.854794979095459, "learning_rate": 1.9964379886742344e-05, "loss": 1.7869, "step": 175 }, { "epoch": 0.16988416988416988, "grad_norm": 5.24516487121582, "learning_rate": 1.996349548749789e-05, "loss": 1.8371, "step": 176 }, { "epoch": 0.17084942084942084, "grad_norm": 4.558859825134277, "learning_rate": 1.9962600263327376e-05, "loss": 1.8612, "step": 177 }, { "epoch": 0.1718146718146718, "grad_norm": 4.1520538330078125, "learning_rate": 1.9961694215203424e-05, "loss": 1.859, "step": 178 }, { "epoch": 0.17277992277992277, "grad_norm": 5.53354024887085, "learning_rate": 1.9960777344110426e-05, "loss": 1.8372, "step": 179 }, { "epoch": 0.17374517374517376, "grad_norm": 4.267213821411133, "learning_rate": 1.9959849651044515e-05, "loss": 1.7665, "step": 180 }, { "epoch": 0.17471042471042472, "grad_norm": 4.841174602508545, "learning_rate": 1.9958911137013596e-05, "loss": 1.8199, "step": 181 }, { "epoch": 0.17567567567567569, "grad_norm": 5.410240173339844, "learning_rate": 1.9957961803037325e-05, "loss": 1.8309, "step": 182 }, { "epoch": 0.17664092664092665, "grad_norm": 5.265180587768555, "learning_rate": 1.9957001650147116e-05, "loss": 1.8111, "step": 183 }, { "epoch": 0.1776061776061776, "grad_norm": 4.3163251876831055, "learning_rate": 1.9956030679386132e-05, "loss": 1.8032, "step": 184 }, { "epoch": 0.17857142857142858, "grad_norm": 5.663985252380371, "learning_rate": 1.9955048891809293e-05, "loss": 1.7645, "step": 185 }, { "epoch": 0.17953667953667954, "grad_norm": 5.375406742095947, "learning_rate": 1.9954056288483273e-05, "loss": 1.8664, "step": 186 }, { "epoch": 0.1805019305019305, "grad_norm": 5.803011894226074, "learning_rate": 1.9953052870486492e-05, "loss": 1.69, "step": 187 }, { "epoch": 0.18146718146718147, "grad_norm": 5.244572162628174, "learning_rate": 1.995203863890912e-05, "loss": 1.8374, "step": 188 }, { "epoch": 0.18243243243243243, "grad_norm": 5.326415061950684, "learning_rate": 1.9951013594853087e-05, "loss": 1.8369, "step": 189 }, { "epoch": 0.1833976833976834, "grad_norm": 3.968106985092163, "learning_rate": 1.9949977739432052e-05, "loss": 1.7869, "step": 190 }, { "epoch": 0.18436293436293436, "grad_norm": 6.265425205230713, "learning_rate": 1.9948931073771433e-05, "loss": 1.8038, "step": 191 }, { "epoch": 0.18532818532818532, "grad_norm": 4.435483932495117, "learning_rate": 1.9947873599008387e-05, "loss": 1.7796, "step": 192 }, { "epoch": 0.18629343629343628, "grad_norm": 4.4179887771606445, "learning_rate": 1.9946805316291817e-05, "loss": 1.7582, "step": 193 }, { "epoch": 0.18725868725868725, "grad_norm": 5.019647121429443, "learning_rate": 1.994572622678237e-05, "loss": 1.7806, "step": 194 }, { "epoch": 0.18822393822393824, "grad_norm": 4.530157566070557, "learning_rate": 1.994463633165243e-05, "loss": 1.8291, "step": 195 }, { "epoch": 0.1891891891891892, "grad_norm": 4.435397624969482, "learning_rate": 1.9943535632086125e-05, "loss": 1.8541, "step": 196 }, { "epoch": 0.19015444015444016, "grad_norm": 4.224645137786865, "learning_rate": 1.9942424129279318e-05, "loss": 1.7747, "step": 197 }, { "epoch": 0.19111969111969113, "grad_norm": 4.534023284912109, "learning_rate": 1.9941301824439608e-05, "loss": 1.821, "step": 198 }, { "epoch": 0.1920849420849421, "grad_norm": 4.295060634613037, "learning_rate": 1.994016871878634e-05, "loss": 1.7816, "step": 199 }, { "epoch": 0.19305019305019305, "grad_norm": 4.451807498931885, "learning_rate": 1.993902481355058e-05, "loss": 1.8481, "step": 200 }, { "epoch": 0.19401544401544402, "grad_norm": 4.234887599945068, "learning_rate": 1.9937870109975138e-05, "loss": 1.8334, "step": 201 }, { "epoch": 0.19498069498069498, "grad_norm": 4.383005619049072, "learning_rate": 1.9936704609314553e-05, "loss": 1.85, "step": 202 }, { "epoch": 0.19594594594594594, "grad_norm": 4.373894691467285, "learning_rate": 1.9935528312835087e-05, "loss": 1.844, "step": 203 }, { "epoch": 0.1969111969111969, "grad_norm": 4.654743671417236, "learning_rate": 1.993434122181474e-05, "loss": 1.893, "step": 204 }, { "epoch": 0.19787644787644787, "grad_norm": 7.665431499481201, "learning_rate": 1.9933143337543238e-05, "loss": 1.7608, "step": 205 }, { "epoch": 0.19884169884169883, "grad_norm": 4.686855792999268, "learning_rate": 1.9931934661322035e-05, "loss": 1.8406, "step": 206 }, { "epoch": 0.1998069498069498, "grad_norm": 6.368899822235107, "learning_rate": 1.9930715194464304e-05, "loss": 1.7996, "step": 207 }, { "epoch": 0.20077220077220076, "grad_norm": 5.041531085968018, "learning_rate": 1.992948493829495e-05, "loss": 1.8371, "step": 208 }, { "epoch": 0.20173745173745175, "grad_norm": 3.9813947677612305, "learning_rate": 1.9928243894150586e-05, "loss": 1.8108, "step": 209 }, { "epoch": 0.20270270270270271, "grad_norm": 4.69705867767334, "learning_rate": 1.9926992063379568e-05, "loss": 1.788, "step": 210 }, { "epoch": 0.20366795366795368, "grad_norm": 5.991170883178711, "learning_rate": 1.992572944734195e-05, "loss": 1.8376, "step": 211 }, { "epoch": 0.20463320463320464, "grad_norm": 4.168196201324463, "learning_rate": 1.9924456047409517e-05, "loss": 1.8825, "step": 212 }, { "epoch": 0.2055984555984556, "grad_norm": 5.6147284507751465, "learning_rate": 1.992317186496576e-05, "loss": 1.8339, "step": 213 }, { "epoch": 0.20656370656370657, "grad_norm": 4.425781726837158, "learning_rate": 1.9921876901405894e-05, "loss": 1.7542, "step": 214 }, { "epoch": 0.20752895752895753, "grad_norm": 5.102436065673828, "learning_rate": 1.9920571158136837e-05, "loss": 1.8345, "step": 215 }, { "epoch": 0.2084942084942085, "grad_norm": 4.2920379638671875, "learning_rate": 1.9919254636577235e-05, "loss": 1.7899, "step": 216 }, { "epoch": 0.20945945945945946, "grad_norm": 4.998841285705566, "learning_rate": 1.9917927338157428e-05, "loss": 1.7901, "step": 217 }, { "epoch": 0.21042471042471042, "grad_norm": 4.545851707458496, "learning_rate": 1.9916589264319475e-05, "loss": 1.8788, "step": 218 }, { "epoch": 0.21138996138996138, "grad_norm": 4.769996643066406, "learning_rate": 1.9915240416517134e-05, "loss": 1.7339, "step": 219 }, { "epoch": 0.21235521235521235, "grad_norm": 5.561061859130859, "learning_rate": 1.991388079621587e-05, "loss": 1.8551, "step": 220 }, { "epoch": 0.2133204633204633, "grad_norm": 5.276782035827637, "learning_rate": 1.9912510404892865e-05, "loss": 1.7999, "step": 221 }, { "epoch": 0.21428571428571427, "grad_norm": 6.434167861938477, "learning_rate": 1.991112924403698e-05, "loss": 1.8323, "step": 222 }, { "epoch": 0.21525096525096524, "grad_norm": 5.014518737792969, "learning_rate": 1.9909737315148798e-05, "loss": 1.7433, "step": 223 }, { "epoch": 0.21621621621621623, "grad_norm": 6.130193710327148, "learning_rate": 1.9908334619740587e-05, "loss": 1.7799, "step": 224 }, { "epoch": 0.2171814671814672, "grad_norm": 3.872027635574341, "learning_rate": 1.9906921159336318e-05, "loss": 1.8152, "step": 225 }, { "epoch": 0.21814671814671815, "grad_norm": 6.342243671417236, "learning_rate": 1.990549693547166e-05, "loss": 1.8283, "step": 226 }, { "epoch": 0.21911196911196912, "grad_norm": 4.687882423400879, "learning_rate": 1.9904061949693972e-05, "loss": 1.8143, "step": 227 }, { "epoch": 0.22007722007722008, "grad_norm": 4.088676929473877, "learning_rate": 1.9902616203562307e-05, "loss": 1.7731, "step": 228 }, { "epoch": 0.22104247104247104, "grad_norm": 5.394243240356445, "learning_rate": 1.990115969864741e-05, "loss": 1.834, "step": 229 }, { "epoch": 0.222007722007722, "grad_norm": 4.424513816833496, "learning_rate": 1.989969243653171e-05, "loss": 1.8217, "step": 230 }, { "epoch": 0.22297297297297297, "grad_norm": 4.55377197265625, "learning_rate": 1.989821441880933e-05, "loss": 1.7464, "step": 231 }, { "epoch": 0.22393822393822393, "grad_norm": 4.709877014160156, "learning_rate": 1.9896725647086074e-05, "loss": 1.7462, "step": 232 }, { "epoch": 0.2249034749034749, "grad_norm": 4.184065818786621, "learning_rate": 1.9895226122979426e-05, "loss": 1.8002, "step": 233 }, { "epoch": 0.22586872586872586, "grad_norm": 4.604548931121826, "learning_rate": 1.9893715848118566e-05, "loss": 1.7847, "step": 234 }, { "epoch": 0.22683397683397682, "grad_norm": 5.295757293701172, "learning_rate": 1.9892194824144338e-05, "loss": 1.817, "step": 235 }, { "epoch": 0.2277992277992278, "grad_norm": 4.350464344024658, "learning_rate": 1.9890663052709276e-05, "loss": 1.7864, "step": 236 }, { "epoch": 0.22876447876447875, "grad_norm": 6.103113651275635, "learning_rate": 1.9889120535477584e-05, "loss": 1.8386, "step": 237 }, { "epoch": 0.22972972972972974, "grad_norm": 4.705803871154785, "learning_rate": 1.9887567274125143e-05, "loss": 1.7247, "step": 238 }, { "epoch": 0.2306949806949807, "grad_norm": 4.0369768142700195, "learning_rate": 1.988600327033951e-05, "loss": 1.811, "step": 239 }, { "epoch": 0.23166023166023167, "grad_norm": 4.492372512817383, "learning_rate": 1.9884428525819906e-05, "loss": 1.808, "step": 240 }, { "epoch": 0.23262548262548263, "grad_norm": 4.9545722007751465, "learning_rate": 1.988284304227723e-05, "loss": 1.7934, "step": 241 }, { "epoch": 0.2335907335907336, "grad_norm": 3.98966646194458, "learning_rate": 1.9881246821434047e-05, "loss": 1.7916, "step": 242 }, { "epoch": 0.23455598455598456, "grad_norm": 4.651267051696777, "learning_rate": 1.9879639865024573e-05, "loss": 1.7568, "step": 243 }, { "epoch": 0.23552123552123552, "grad_norm": 4.708108901977539, "learning_rate": 1.9878022174794714e-05, "loss": 1.711, "step": 244 }, { "epoch": 0.23648648648648649, "grad_norm": 4.429472923278809, "learning_rate": 1.9876393752502013e-05, "loss": 1.8032, "step": 245 }, { "epoch": 0.23745173745173745, "grad_norm": 4.605119228363037, "learning_rate": 1.987475459991569e-05, "loss": 1.8229, "step": 246 }, { "epoch": 0.2384169884169884, "grad_norm": 5.334287166595459, "learning_rate": 1.9873104718816613e-05, "loss": 1.7261, "step": 247 }, { "epoch": 0.23938223938223938, "grad_norm": 3.8895561695098877, "learning_rate": 1.987144411099731e-05, "loss": 1.7948, "step": 248 }, { "epoch": 0.24034749034749034, "grad_norm": 5.1954522132873535, "learning_rate": 1.9869772778261967e-05, "loss": 1.7504, "step": 249 }, { "epoch": 0.2413127413127413, "grad_norm": 4.352097988128662, "learning_rate": 1.986809072242641e-05, "loss": 1.7795, "step": 250 }, { "epoch": 0.24227799227799227, "grad_norm": 3.8593056201934814, "learning_rate": 1.9866397945318133e-05, "loss": 1.83, "step": 251 }, { "epoch": 0.24324324324324326, "grad_norm": 4.728880882263184, "learning_rate": 1.986469444877626e-05, "loss": 1.7474, "step": 252 }, { "epoch": 0.24420849420849422, "grad_norm": 5.145429611206055, "learning_rate": 1.9862980234651574e-05, "loss": 1.7845, "step": 253 }, { "epoch": 0.24517374517374518, "grad_norm": 3.842902183532715, "learning_rate": 1.9861255304806493e-05, "loss": 1.7645, "step": 254 }, { "epoch": 0.24613899613899615, "grad_norm": 4.171161651611328, "learning_rate": 1.9859519661115087e-05, "loss": 1.7954, "step": 255 }, { "epoch": 0.2471042471042471, "grad_norm": 4.398767471313477, "learning_rate": 1.985777330546306e-05, "loss": 1.8025, "step": 256 }, { "epoch": 0.24806949806949807, "grad_norm": 4.435847282409668, "learning_rate": 1.9856016239747754e-05, "loss": 1.7602, "step": 257 }, { "epoch": 0.24903474903474904, "grad_norm": 3.882884979248047, "learning_rate": 1.985424846587815e-05, "loss": 1.7644, "step": 258 }, { "epoch": 0.25, "grad_norm": 4.172622203826904, "learning_rate": 1.985246998577486e-05, "loss": 1.7587, "step": 259 }, { "epoch": 0.25096525096525096, "grad_norm": 4.639303207397461, "learning_rate": 1.985068080137013e-05, "loss": 1.7984, "step": 260 }, { "epoch": 0.2519305019305019, "grad_norm": 3.98333477973938, "learning_rate": 1.9848880914607834e-05, "loss": 1.7535, "step": 261 }, { "epoch": 0.2528957528957529, "grad_norm": 4.069552898406982, "learning_rate": 1.9847070327443476e-05, "loss": 1.6939, "step": 262 }, { "epoch": 0.25386100386100385, "grad_norm": 5.007661819458008, "learning_rate": 1.984524904184418e-05, "loss": 1.7565, "step": 263 }, { "epoch": 0.2548262548262548, "grad_norm": 4.196667194366455, "learning_rate": 1.9843417059788702e-05, "loss": 1.7631, "step": 264 }, { "epoch": 0.2557915057915058, "grad_norm": 5.928943157196045, "learning_rate": 1.9841574383267415e-05, "loss": 1.756, "step": 265 }, { "epoch": 0.25675675675675674, "grad_norm": 3.984234571456909, "learning_rate": 1.9839721014282307e-05, "loss": 1.7994, "step": 266 }, { "epoch": 0.2577220077220077, "grad_norm": 3.9803266525268555, "learning_rate": 1.983785695484699e-05, "loss": 1.7556, "step": 267 }, { "epoch": 0.25868725868725867, "grad_norm": 4.372330665588379, "learning_rate": 1.9835982206986685e-05, "loss": 1.6999, "step": 268 }, { "epoch": 0.25965250965250963, "grad_norm": 3.8054566383361816, "learning_rate": 1.9834096772738233e-05, "loss": 1.72, "step": 269 }, { "epoch": 0.2606177606177606, "grad_norm": 4.114311218261719, "learning_rate": 1.9832200654150077e-05, "loss": 1.8302, "step": 270 }, { "epoch": 0.26158301158301156, "grad_norm": 4.022603988647461, "learning_rate": 1.983029385328227e-05, "loss": 1.7972, "step": 271 }, { "epoch": 0.2625482625482625, "grad_norm": 4.409243583679199, "learning_rate": 1.982837637220647e-05, "loss": 1.7628, "step": 272 }, { "epoch": 0.2635135135135135, "grad_norm": 4.133562088012695, "learning_rate": 1.982644821300595e-05, "loss": 1.7827, "step": 273 }, { "epoch": 0.2644787644787645, "grad_norm": 5.861300945281982, "learning_rate": 1.9824509377775567e-05, "loss": 1.7821, "step": 274 }, { "epoch": 0.26544401544401547, "grad_norm": 4.026004791259766, "learning_rate": 1.982255986862179e-05, "loss": 1.753, "step": 275 }, { "epoch": 0.26640926640926643, "grad_norm": 6.394083023071289, "learning_rate": 1.9820599687662677e-05, "loss": 1.8002, "step": 276 }, { "epoch": 0.2673745173745174, "grad_norm": 3.9767532348632812, "learning_rate": 1.981862883702789e-05, "loss": 1.7948, "step": 277 }, { "epoch": 0.26833976833976836, "grad_norm": 7.201776504516602, "learning_rate": 1.9816647318858666e-05, "loss": 1.777, "step": 278 }, { "epoch": 0.2693050193050193, "grad_norm": 3.9289729595184326, "learning_rate": 1.981465513530785e-05, "loss": 1.7838, "step": 279 }, { "epoch": 0.2702702702702703, "grad_norm": 6.732141494750977, "learning_rate": 1.981265228853987e-05, "loss": 1.7321, "step": 280 }, { "epoch": 0.27123552123552125, "grad_norm": 5.770628929138184, "learning_rate": 1.981063878073073e-05, "loss": 1.7905, "step": 281 }, { "epoch": 0.2722007722007722, "grad_norm": 4.88922119140625, "learning_rate": 1.980861461406802e-05, "loss": 1.7441, "step": 282 }, { "epoch": 0.2731660231660232, "grad_norm": 5.953951358795166, "learning_rate": 1.9806579790750922e-05, "loss": 1.7527, "step": 283 }, { "epoch": 0.27413127413127414, "grad_norm": 4.962509632110596, "learning_rate": 1.9804534312990183e-05, "loss": 1.7342, "step": 284 }, { "epoch": 0.2750965250965251, "grad_norm": 6.674172401428223, "learning_rate": 1.980247818300813e-05, "loss": 1.7753, "step": 285 }, { "epoch": 0.27606177606177607, "grad_norm": 6.852287769317627, "learning_rate": 1.9800411403038664e-05, "loss": 1.7277, "step": 286 }, { "epoch": 0.27702702702702703, "grad_norm": 5.453775882720947, "learning_rate": 1.9798333975327258e-05, "loss": 1.8125, "step": 287 }, { "epoch": 0.277992277992278, "grad_norm": 7.326210021972656, "learning_rate": 1.9796245902130948e-05, "loss": 1.7298, "step": 288 }, { "epoch": 0.27895752895752896, "grad_norm": 4.287469863891602, "learning_rate": 1.979414718571834e-05, "loss": 1.7778, "step": 289 }, { "epoch": 0.2799227799227799, "grad_norm": 6.503158092498779, "learning_rate": 1.9792037828369607e-05, "loss": 1.7942, "step": 290 }, { "epoch": 0.2808880308880309, "grad_norm": 6.972282886505127, "learning_rate": 1.978991783237647e-05, "loss": 1.7601, "step": 291 }, { "epoch": 0.28185328185328185, "grad_norm": 6.6554460525512695, "learning_rate": 1.9787787200042224e-05, "loss": 1.7281, "step": 292 }, { "epoch": 0.2828185328185328, "grad_norm": 4.683424472808838, "learning_rate": 1.9785645933681713e-05, "loss": 1.7916, "step": 293 }, { "epoch": 0.28378378378378377, "grad_norm": 3.806309700012207, "learning_rate": 1.978349403562133e-05, "loss": 1.8447, "step": 294 }, { "epoch": 0.28474903474903474, "grad_norm": 5.332488059997559, "learning_rate": 1.978133150819902e-05, "loss": 1.7481, "step": 295 }, { "epoch": 0.2857142857142857, "grad_norm": 4.274888515472412, "learning_rate": 1.9779158353764293e-05, "loss": 1.7075, "step": 296 }, { "epoch": 0.28667953667953666, "grad_norm": 4.067286491394043, "learning_rate": 1.977697457467818e-05, "loss": 1.7307, "step": 297 }, { "epoch": 0.2876447876447876, "grad_norm": 4.854985237121582, "learning_rate": 1.9774780173313273e-05, "loss": 1.7905, "step": 298 }, { "epoch": 0.2886100386100386, "grad_norm": 4.431771278381348, "learning_rate": 1.9772575152053696e-05, "loss": 1.7558, "step": 299 }, { "epoch": 0.28957528957528955, "grad_norm": 7.292600154876709, "learning_rate": 1.977035951329511e-05, "loss": 1.7599, "step": 300 }, { "epoch": 0.2905405405405405, "grad_norm": 4.510945796966553, "learning_rate": 1.976813325944472e-05, "loss": 1.8491, "step": 301 }, { "epoch": 0.2915057915057915, "grad_norm": 5.671480178833008, "learning_rate": 1.9765896392921253e-05, "loss": 1.7796, "step": 302 }, { "epoch": 0.2924710424710425, "grad_norm": 5.7752227783203125, "learning_rate": 1.9763648916154982e-05, "loss": 1.8094, "step": 303 }, { "epoch": 0.29343629343629346, "grad_norm": 4.039534568786621, "learning_rate": 1.9761390831587687e-05, "loss": 1.7191, "step": 304 }, { "epoch": 0.2944015444015444, "grad_norm": 7.9651312828063965, "learning_rate": 1.975912214167269e-05, "loss": 1.7937, "step": 305 }, { "epoch": 0.2953667953667954, "grad_norm": 4.313594341278076, "learning_rate": 1.975684284887483e-05, "loss": 1.7014, "step": 306 }, { "epoch": 0.29633204633204635, "grad_norm": 4.699955463409424, "learning_rate": 1.975455295567046e-05, "loss": 1.738, "step": 307 }, { "epoch": 0.2972972972972973, "grad_norm": 5.305117130279541, "learning_rate": 1.9752252464547454e-05, "loss": 1.8219, "step": 308 }, { "epoch": 0.2982625482625483, "grad_norm": 5.326988220214844, "learning_rate": 1.97499413780052e-05, "loss": 1.8626, "step": 309 }, { "epoch": 0.29922779922779924, "grad_norm": 4.539814472198486, "learning_rate": 1.9747619698554606e-05, "loss": 1.7771, "step": 310 }, { "epoch": 0.3001930501930502, "grad_norm": 4.68597412109375, "learning_rate": 1.9745287428718072e-05, "loss": 1.8303, "step": 311 }, { "epoch": 0.30115830115830117, "grad_norm": 6.816437721252441, "learning_rate": 1.9742944571029515e-05, "loss": 1.8103, "step": 312 }, { "epoch": 0.30212355212355213, "grad_norm": 3.888559341430664, "learning_rate": 1.9740591128034357e-05, "loss": 1.7313, "step": 313 }, { "epoch": 0.3030888030888031, "grad_norm": 5.622889041900635, "learning_rate": 1.973822710228951e-05, "loss": 1.7976, "step": 314 }, { "epoch": 0.30405405405405406, "grad_norm": 5.591082572937012, "learning_rate": 1.9735852496363395e-05, "loss": 1.7817, "step": 315 }, { "epoch": 0.305019305019305, "grad_norm": 5.452985763549805, "learning_rate": 1.973346731283592e-05, "loss": 1.7994, "step": 316 }, { "epoch": 0.305984555984556, "grad_norm": 5.106136322021484, "learning_rate": 1.973107155429849e-05, "loss": 1.7853, "step": 317 }, { "epoch": 0.30694980694980695, "grad_norm": 4.438343524932861, "learning_rate": 1.9728665223354003e-05, "loss": 1.7399, "step": 318 }, { "epoch": 0.3079150579150579, "grad_norm": 4.4550251960754395, "learning_rate": 1.972624832261683e-05, "loss": 1.7319, "step": 319 }, { "epoch": 0.3088803088803089, "grad_norm": 5.767579555511475, "learning_rate": 1.9723820854712836e-05, "loss": 1.6821, "step": 320 }, { "epoch": 0.30984555984555984, "grad_norm": 4.572739601135254, "learning_rate": 1.972138282227937e-05, "loss": 1.8463, "step": 321 }, { "epoch": 0.3108108108108108, "grad_norm": 6.5078020095825195, "learning_rate": 1.971893422796524e-05, "loss": 1.7614, "step": 322 }, { "epoch": 0.31177606177606176, "grad_norm": 3.8346452713012695, "learning_rate": 1.9716475074430762e-05, "loss": 1.7916, "step": 323 }, { "epoch": 0.3127413127413127, "grad_norm": 4.194652557373047, "learning_rate": 1.9714005364347687e-05, "loss": 1.6699, "step": 324 }, { "epoch": 0.3137065637065637, "grad_norm": 5.606922149658203, "learning_rate": 1.971152510039926e-05, "loss": 1.7607, "step": 325 }, { "epoch": 0.31467181467181465, "grad_norm": 3.6330485343933105, "learning_rate": 1.9709034285280185e-05, "loss": 1.7754, "step": 326 }, { "epoch": 0.3156370656370656, "grad_norm": 5.310985565185547, "learning_rate": 1.9706532921696634e-05, "loss": 1.6881, "step": 327 }, { "epoch": 0.3166023166023166, "grad_norm": 4.781001567840576, "learning_rate": 1.9704021012366228e-05, "loss": 1.7821, "step": 328 }, { "epoch": 0.31756756756756754, "grad_norm": 6.116867542266846, "learning_rate": 1.9701498560018055e-05, "loss": 1.8737, "step": 329 }, { "epoch": 0.3185328185328185, "grad_norm": 5.587736129760742, "learning_rate": 1.9698965567392656e-05, "loss": 1.7625, "step": 330 }, { "epoch": 0.3194980694980695, "grad_norm": 3.6453404426574707, "learning_rate": 1.9696422037242023e-05, "loss": 1.7638, "step": 331 }, { "epoch": 0.3204633204633205, "grad_norm": 5.7377214431762695, "learning_rate": 1.96938679723296e-05, "loss": 1.7712, "step": 332 }, { "epoch": 0.32142857142857145, "grad_norm": 5.419057369232178, "learning_rate": 1.9691303375430265e-05, "loss": 1.7522, "step": 333 }, { "epoch": 0.3223938223938224, "grad_norm": 3.678049087524414, "learning_rate": 1.968872824933035e-05, "loss": 1.7662, "step": 334 }, { "epoch": 0.3233590733590734, "grad_norm": 5.382504940032959, "learning_rate": 1.9686142596827623e-05, "loss": 1.784, "step": 335 }, { "epoch": 0.32432432432432434, "grad_norm": 4.854099750518799, "learning_rate": 1.9683546420731292e-05, "loss": 1.7126, "step": 336 }, { "epoch": 0.3252895752895753, "grad_norm": 4.074610233306885, "learning_rate": 1.9680939723861993e-05, "loss": 1.7813, "step": 337 }, { "epoch": 0.32625482625482627, "grad_norm": 4.782020568847656, "learning_rate": 1.967832250905179e-05, "loss": 1.7093, "step": 338 }, { "epoch": 0.32722007722007723, "grad_norm": 4.857923984527588, "learning_rate": 1.9675694779144184e-05, "loss": 1.7251, "step": 339 }, { "epoch": 0.3281853281853282, "grad_norm": 4.356438636779785, "learning_rate": 1.9673056536994094e-05, "loss": 1.7315, "step": 340 }, { "epoch": 0.32915057915057916, "grad_norm": 4.294323444366455, "learning_rate": 1.967040778546786e-05, "loss": 1.7349, "step": 341 }, { "epoch": 0.3301158301158301, "grad_norm": 4.033375263214111, "learning_rate": 1.9667748527443248e-05, "loss": 1.6029, "step": 342 }, { "epoch": 0.3310810810810811, "grad_norm": 3.696950912475586, "learning_rate": 1.966507876580942e-05, "loss": 1.7534, "step": 343 }, { "epoch": 0.33204633204633205, "grad_norm": 3.718315839767456, "learning_rate": 1.9662398503466963e-05, "loss": 1.7581, "step": 344 }, { "epoch": 0.333011583011583, "grad_norm": 3.952453851699829, "learning_rate": 1.9659707743327878e-05, "loss": 1.7004, "step": 345 }, { "epoch": 0.333976833976834, "grad_norm": 4.828619956970215, "learning_rate": 1.9657006488315563e-05, "loss": 1.7381, "step": 346 }, { "epoch": 0.33494208494208494, "grad_norm": 4.196254730224609, "learning_rate": 1.965429474136482e-05, "loss": 1.7943, "step": 347 }, { "epoch": 0.3359073359073359, "grad_norm": 4.142396926879883, "learning_rate": 1.965157250542185e-05, "loss": 1.7326, "step": 348 }, { "epoch": 0.33687258687258687, "grad_norm": 4.161625385284424, "learning_rate": 1.9648839783444243e-05, "loss": 1.7341, "step": 349 }, { "epoch": 0.33783783783783783, "grad_norm": 4.203663349151611, "learning_rate": 1.9646096578401e-05, "loss": 1.7165, "step": 350 }, { "epoch": 0.3388030888030888, "grad_norm": 5.153465747833252, "learning_rate": 1.9643342893272493e-05, "loss": 1.803, "step": 351 }, { "epoch": 0.33976833976833976, "grad_norm": 3.938094139099121, "learning_rate": 1.9640578731050487e-05, "loss": 1.7499, "step": 352 }, { "epoch": 0.3407335907335907, "grad_norm": 5.379260063171387, "learning_rate": 1.9637804094738133e-05, "loss": 1.7332, "step": 353 }, { "epoch": 0.3416988416988417, "grad_norm": 3.7647342681884766, "learning_rate": 1.9635018987349955e-05, "loss": 1.7723, "step": 354 }, { "epoch": 0.34266409266409265, "grad_norm": 4.630080223083496, "learning_rate": 1.963222341191186e-05, "loss": 1.7612, "step": 355 }, { "epoch": 0.3436293436293436, "grad_norm": 4.819670677185059, "learning_rate": 1.9629417371461124e-05, "loss": 1.7614, "step": 356 }, { "epoch": 0.34459459459459457, "grad_norm": 5.193247318267822, "learning_rate": 1.962660086904639e-05, "loss": 1.7174, "step": 357 }, { "epoch": 0.34555984555984554, "grad_norm": 4.070240020751953, "learning_rate": 1.9623773907727682e-05, "loss": 1.7737, "step": 358 }, { "epoch": 0.3465250965250965, "grad_norm": 3.897179126739502, "learning_rate": 1.962093649057637e-05, "loss": 1.738, "step": 359 }, { "epoch": 0.3474903474903475, "grad_norm": 3.6528360843658447, "learning_rate": 1.9618088620675188e-05, "loss": 1.6711, "step": 360 }, { "epoch": 0.3484555984555985, "grad_norm": 3.6807613372802734, "learning_rate": 1.9615230301118234e-05, "loss": 1.7994, "step": 361 }, { "epoch": 0.34942084942084944, "grad_norm": 4.079460620880127, "learning_rate": 1.961236153501095e-05, "loss": 1.7327, "step": 362 }, { "epoch": 0.3503861003861004, "grad_norm": 4.715167999267578, "learning_rate": 1.9609482325470137e-05, "loss": 1.7292, "step": 363 }, { "epoch": 0.35135135135135137, "grad_norm": 3.709078073501587, "learning_rate": 1.960659267562393e-05, "loss": 1.7232, "step": 364 }, { "epoch": 0.35231660231660233, "grad_norm": 3.9571263790130615, "learning_rate": 1.9603692588611816e-05, "loss": 1.733, "step": 365 }, { "epoch": 0.3532818532818533, "grad_norm": 4.878260612487793, "learning_rate": 1.9600782067584625e-05, "loss": 1.7165, "step": 366 }, { "epoch": 0.35424710424710426, "grad_norm": 3.7100722789764404, "learning_rate": 1.9597861115704512e-05, "loss": 1.7196, "step": 367 }, { "epoch": 0.3552123552123552, "grad_norm": 5.340725898742676, "learning_rate": 1.9594929736144978e-05, "loss": 1.7328, "step": 368 }, { "epoch": 0.3561776061776062, "grad_norm": 3.7168092727661133, "learning_rate": 1.9591987932090836e-05, "loss": 1.7269, "step": 369 }, { "epoch": 0.35714285714285715, "grad_norm": 4.261868476867676, "learning_rate": 1.9589035706738237e-05, "loss": 1.7852, "step": 370 }, { "epoch": 0.3581081081081081, "grad_norm": 4.013570308685303, "learning_rate": 1.958607306329466e-05, "loss": 1.7234, "step": 371 }, { "epoch": 0.3590733590733591, "grad_norm": 3.7619221210479736, "learning_rate": 1.9583100004978886e-05, "loss": 1.6764, "step": 372 }, { "epoch": 0.36003861003861004, "grad_norm": 4.159510612487793, "learning_rate": 1.9580116535021022e-05, "loss": 1.7341, "step": 373 }, { "epoch": 0.361003861003861, "grad_norm": 3.8162317276000977, "learning_rate": 1.9577122656662485e-05, "loss": 1.7593, "step": 374 }, { "epoch": 0.36196911196911197, "grad_norm": 4.361527919769287, "learning_rate": 1.9574118373156e-05, "loss": 1.7437, "step": 375 }, { "epoch": 0.36293436293436293, "grad_norm": 4.29148006439209, "learning_rate": 1.95711036877656e-05, "loss": 1.8134, "step": 376 }, { "epoch": 0.3638996138996139, "grad_norm": 3.7081241607666016, "learning_rate": 1.9568078603766613e-05, "loss": 1.7224, "step": 377 }, { "epoch": 0.36486486486486486, "grad_norm": 4.411427974700928, "learning_rate": 1.9565043124445665e-05, "loss": 1.753, "step": 378 }, { "epoch": 0.3658301158301158, "grad_norm": 4.272647380828857, "learning_rate": 1.9561997253100684e-05, "loss": 1.7252, "step": 379 }, { "epoch": 0.3667953667953668, "grad_norm": 3.9347710609436035, "learning_rate": 1.9558940993040885e-05, "loss": 1.7363, "step": 380 }, { "epoch": 0.36776061776061775, "grad_norm": 3.8959333896636963, "learning_rate": 1.955587434758676e-05, "loss": 1.7543, "step": 381 }, { "epoch": 0.3687258687258687, "grad_norm": 3.861565113067627, "learning_rate": 1.9552797320070097e-05, "loss": 1.7333, "step": 382 }, { "epoch": 0.3696911196911197, "grad_norm": 4.173425197601318, "learning_rate": 1.954970991383396e-05, "loss": 1.749, "step": 383 }, { "epoch": 0.37065637065637064, "grad_norm": 5.294433116912842, "learning_rate": 1.9546612132232688e-05, "loss": 1.7576, "step": 384 }, { "epoch": 0.3716216216216216, "grad_norm": 4.082633972167969, "learning_rate": 1.9543503978631894e-05, "loss": 1.7569, "step": 385 }, { "epoch": 0.37258687258687256, "grad_norm": 4.692306995391846, "learning_rate": 1.9540385456408456e-05, "loss": 1.6946, "step": 386 }, { "epoch": 0.3735521235521235, "grad_norm": 4.214600563049316, "learning_rate": 1.953725656895052e-05, "loss": 1.7236, "step": 387 }, { "epoch": 0.3745173745173745, "grad_norm": 4.012565612792969, "learning_rate": 1.953411731965749e-05, "loss": 1.7263, "step": 388 }, { "epoch": 0.3754826254826255, "grad_norm": 4.862763404846191, "learning_rate": 1.9530967711940038e-05, "loss": 1.7547, "step": 389 }, { "epoch": 0.3764478764478765, "grad_norm": 4.253438472747803, "learning_rate": 1.952780774922008e-05, "loss": 1.6917, "step": 390 }, { "epoch": 0.37741312741312744, "grad_norm": 3.990596055984497, "learning_rate": 1.952463743493078e-05, "loss": 1.7482, "step": 391 }, { "epoch": 0.3783783783783784, "grad_norm": 4.393350601196289, "learning_rate": 1.9521456772516554e-05, "loss": 1.778, "step": 392 }, { "epoch": 0.37934362934362936, "grad_norm": 4.936257362365723, "learning_rate": 1.9518265765433066e-05, "loss": 1.7288, "step": 393 }, { "epoch": 0.3803088803088803, "grad_norm": 4.040771007537842, "learning_rate": 1.951506441714721e-05, "loss": 1.6901, "step": 394 }, { "epoch": 0.3812741312741313, "grad_norm": 4.402784824371338, "learning_rate": 1.951185273113712e-05, "loss": 1.7557, "step": 395 }, { "epoch": 0.38223938223938225, "grad_norm": 4.547336578369141, "learning_rate": 1.9508630710892157e-05, "loss": 1.7385, "step": 396 }, { "epoch": 0.3832046332046332, "grad_norm": 5.552551746368408, "learning_rate": 1.9505398359912914e-05, "loss": 1.7842, "step": 397 }, { "epoch": 0.3841698841698842, "grad_norm": 3.7976114749908447, "learning_rate": 1.950215568171121e-05, "loss": 1.6636, "step": 398 }, { "epoch": 0.38513513513513514, "grad_norm": 3.8602499961853027, "learning_rate": 1.9498902679810083e-05, "loss": 1.7281, "step": 399 }, { "epoch": 0.3861003861003861, "grad_norm": 3.753833055496216, "learning_rate": 1.9495639357743773e-05, "loss": 1.7545, "step": 400 }, { "epoch": 0.38706563706563707, "grad_norm": 3.642061471939087, "learning_rate": 1.9492365719057752e-05, "loss": 1.738, "step": 401 }, { "epoch": 0.38803088803088803, "grad_norm": 3.9402945041656494, "learning_rate": 1.9489081767308696e-05, "loss": 1.7693, "step": 402 }, { "epoch": 0.388996138996139, "grad_norm": 3.8130996227264404, "learning_rate": 1.9485787506064483e-05, "loss": 1.7868, "step": 403 }, { "epoch": 0.38996138996138996, "grad_norm": 3.8527071475982666, "learning_rate": 1.9482482938904183e-05, "loss": 1.7396, "step": 404 }, { "epoch": 0.3909266409266409, "grad_norm": 4.455222129821777, "learning_rate": 1.947916806941808e-05, "loss": 1.7663, "step": 405 }, { "epoch": 0.3918918918918919, "grad_norm": 3.8062307834625244, "learning_rate": 1.947584290120764e-05, "loss": 1.6825, "step": 406 }, { "epoch": 0.39285714285714285, "grad_norm": 3.9247851371765137, "learning_rate": 1.9472507437885523e-05, "loss": 1.7211, "step": 407 }, { "epoch": 0.3938223938223938, "grad_norm": 4.017999649047852, "learning_rate": 1.9469161683075568e-05, "loss": 1.6961, "step": 408 }, { "epoch": 0.3947876447876448, "grad_norm": 3.875453233718872, "learning_rate": 1.94658056404128e-05, "loss": 1.8132, "step": 409 }, { "epoch": 0.39575289575289574, "grad_norm": 4.510627269744873, "learning_rate": 1.9462439313543423e-05, "loss": 1.7492, "step": 410 }, { "epoch": 0.3967181467181467, "grad_norm": 4.282536506652832, "learning_rate": 1.9459062706124813e-05, "loss": 1.6867, "step": 411 }, { "epoch": 0.39768339768339767, "grad_norm": 4.016504764556885, "learning_rate": 1.945567582182551e-05, "loss": 1.6929, "step": 412 }, { "epoch": 0.39864864864864863, "grad_norm": 3.6763112545013428, "learning_rate": 1.9452278664325227e-05, "loss": 1.7274, "step": 413 }, { "epoch": 0.3996138996138996, "grad_norm": 4.243254661560059, "learning_rate": 1.9448871237314834e-05, "loss": 1.7558, "step": 414 }, { "epoch": 0.40057915057915056, "grad_norm": 3.863778591156006, "learning_rate": 1.9445453544496363e-05, "loss": 1.7452, "step": 415 }, { "epoch": 0.4015444015444015, "grad_norm": 3.8378090858459473, "learning_rate": 1.9442025589582988e-05, "loss": 1.8414, "step": 416 }, { "epoch": 0.4025096525096525, "grad_norm": 3.61639404296875, "learning_rate": 1.9438587376299047e-05, "loss": 1.6905, "step": 417 }, { "epoch": 0.4034749034749035, "grad_norm": 4.8080902099609375, "learning_rate": 1.9435138908380017e-05, "loss": 1.7115, "step": 418 }, { "epoch": 0.40444015444015446, "grad_norm": 4.0947418212890625, "learning_rate": 1.9431680189572514e-05, "loss": 1.7023, "step": 419 }, { "epoch": 0.40540540540540543, "grad_norm": 4.377869606018066, "learning_rate": 1.942821122363429e-05, "loss": 1.7552, "step": 420 }, { "epoch": 0.4063706563706564, "grad_norm": 4.720925331115723, "learning_rate": 1.942473201433424e-05, "loss": 1.7333, "step": 421 }, { "epoch": 0.40733590733590735, "grad_norm": 3.5857903957366943, "learning_rate": 1.9421242565452373e-05, "loss": 1.711, "step": 422 }, { "epoch": 0.4083011583011583, "grad_norm": 4.641952991485596, "learning_rate": 1.9417742880779835e-05, "loss": 1.7562, "step": 423 }, { "epoch": 0.4092664092664093, "grad_norm": 3.7906808853149414, "learning_rate": 1.9414232964118893e-05, "loss": 1.7231, "step": 424 }, { "epoch": 0.41023166023166024, "grad_norm": 3.551534414291382, "learning_rate": 1.941071281928292e-05, "loss": 1.6716, "step": 425 }, { "epoch": 0.4111969111969112, "grad_norm": 4.0085930824279785, "learning_rate": 1.940718245009641e-05, "loss": 1.6725, "step": 426 }, { "epoch": 0.41216216216216217, "grad_norm": 3.807429790496826, "learning_rate": 1.940364186039496e-05, "loss": 1.7575, "step": 427 }, { "epoch": 0.41312741312741313, "grad_norm": 3.671699047088623, "learning_rate": 1.9400091054025286e-05, "loss": 1.686, "step": 428 }, { "epoch": 0.4140926640926641, "grad_norm": 3.4760184288024902, "learning_rate": 1.939653003484518e-05, "loss": 1.7228, "step": 429 }, { "epoch": 0.41505791505791506, "grad_norm": 3.6061408519744873, "learning_rate": 1.9392958806723544e-05, "loss": 1.7131, "step": 430 }, { "epoch": 0.416023166023166, "grad_norm": 5.025588512420654, "learning_rate": 1.9389377373540373e-05, "loss": 1.6722, "step": 431 }, { "epoch": 0.416988416988417, "grad_norm": 5.2907280921936035, "learning_rate": 1.9385785739186746e-05, "loss": 1.7654, "step": 432 }, { "epoch": 0.41795366795366795, "grad_norm": 4.422195911407471, "learning_rate": 1.9382183907564824e-05, "loss": 1.6597, "step": 433 }, { "epoch": 0.4189189189189189, "grad_norm": 5.675929546356201, "learning_rate": 1.9378571882587848e-05, "loss": 1.7137, "step": 434 }, { "epoch": 0.4198841698841699, "grad_norm": 4.700954914093018, "learning_rate": 1.937494966818014e-05, "loss": 1.7297, "step": 435 }, { "epoch": 0.42084942084942084, "grad_norm": 3.9680557250976562, "learning_rate": 1.9371317268277075e-05, "loss": 1.6652, "step": 436 }, { "epoch": 0.4218146718146718, "grad_norm": 6.557024955749512, "learning_rate": 1.936767468682511e-05, "loss": 1.7385, "step": 437 }, { "epoch": 0.42277992277992277, "grad_norm": 3.592010498046875, "learning_rate": 1.9364021927781764e-05, "loss": 1.7594, "step": 438 }, { "epoch": 0.42374517374517373, "grad_norm": 5.884945392608643, "learning_rate": 1.9360358995115605e-05, "loss": 1.7047, "step": 439 }, { "epoch": 0.4247104247104247, "grad_norm": 4.135279655456543, "learning_rate": 1.9356685892806255e-05, "loss": 1.7335, "step": 440 }, { "epoch": 0.42567567567567566, "grad_norm": 4.141030311584473, "learning_rate": 1.9353002624844392e-05, "loss": 1.7431, "step": 441 }, { "epoch": 0.4266409266409266, "grad_norm": 5.444819450378418, "learning_rate": 1.9349309195231732e-05, "loss": 1.6846, "step": 442 }, { "epoch": 0.4276061776061776, "grad_norm": 3.9959070682525635, "learning_rate": 1.934560560798104e-05, "loss": 1.748, "step": 443 }, { "epoch": 0.42857142857142855, "grad_norm": 6.770045757293701, "learning_rate": 1.9341891867116102e-05, "loss": 1.7336, "step": 444 }, { "epoch": 0.4295366795366795, "grad_norm": 4.172303676605225, "learning_rate": 1.9338167976671747e-05, "loss": 1.7244, "step": 445 }, { "epoch": 0.4305019305019305, "grad_norm": 5.299330234527588, "learning_rate": 1.933443394069383e-05, "loss": 1.6752, "step": 446 }, { "epoch": 0.4314671814671815, "grad_norm": 5.519097805023193, "learning_rate": 1.9330689763239225e-05, "loss": 1.7469, "step": 447 }, { "epoch": 0.43243243243243246, "grad_norm": 4.377777099609375, "learning_rate": 1.9326935448375822e-05, "loss": 1.739, "step": 448 }, { "epoch": 0.4333976833976834, "grad_norm": 6.823882579803467, "learning_rate": 1.9323171000182538e-05, "loss": 1.7708, "step": 449 }, { "epoch": 0.4343629343629344, "grad_norm": 3.7352969646453857, "learning_rate": 1.9319396422749284e-05, "loss": 1.7035, "step": 450 }, { "epoch": 0.43532818532818535, "grad_norm": 5.113969802856445, "learning_rate": 1.9315611720176982e-05, "loss": 1.6949, "step": 451 }, { "epoch": 0.4362934362934363, "grad_norm": 5.587355613708496, "learning_rate": 1.931181689657756e-05, "loss": 1.7613, "step": 452 }, { "epoch": 0.4372586872586873, "grad_norm": 4.409136772155762, "learning_rate": 1.9308011956073933e-05, "loss": 1.7631, "step": 453 }, { "epoch": 0.43822393822393824, "grad_norm": 5.3545074462890625, "learning_rate": 1.9304196902800014e-05, "loss": 1.7496, "step": 454 }, { "epoch": 0.4391891891891892, "grad_norm": 3.6010336875915527, "learning_rate": 1.9300371740900698e-05, "loss": 1.7381, "step": 455 }, { "epoch": 0.44015444015444016, "grad_norm": 3.643707513809204, "learning_rate": 1.9296536474531864e-05, "loss": 1.7991, "step": 456 }, { "epoch": 0.4411196911196911, "grad_norm": 4.69381856918335, "learning_rate": 1.9292691107860374e-05, "loss": 1.6745, "step": 457 }, { "epoch": 0.4420849420849421, "grad_norm": 3.876347780227661, "learning_rate": 1.928883564506406e-05, "loss": 1.7045, "step": 458 }, { "epoch": 0.44305019305019305, "grad_norm": 3.8134143352508545, "learning_rate": 1.928497009033172e-05, "loss": 1.6769, "step": 459 }, { "epoch": 0.444015444015444, "grad_norm": 3.9049108028411865, "learning_rate": 1.9281094447863126e-05, "loss": 1.7107, "step": 460 }, { "epoch": 0.444980694980695, "grad_norm": 3.940433979034424, "learning_rate": 1.927720872186899e-05, "loss": 1.712, "step": 461 }, { "epoch": 0.44594594594594594, "grad_norm": 3.828815221786499, "learning_rate": 1.9273312916571008e-05, "loss": 1.765, "step": 462 }, { "epoch": 0.4469111969111969, "grad_norm": 3.92582106590271, "learning_rate": 1.9269407036201803e-05, "loss": 1.7118, "step": 463 }, { "epoch": 0.44787644787644787, "grad_norm": 4.062038421630859, "learning_rate": 1.9265491085004956e-05, "loss": 1.7196, "step": 464 }, { "epoch": 0.44884169884169883, "grad_norm": 4.037832736968994, "learning_rate": 1.926156506723498e-05, "loss": 1.7286, "step": 465 }, { "epoch": 0.4498069498069498, "grad_norm": 4.069302082061768, "learning_rate": 1.9257628987157343e-05, "loss": 1.692, "step": 466 }, { "epoch": 0.45077220077220076, "grad_norm": 4.033402442932129, "learning_rate": 1.9253682849048417e-05, "loss": 1.717, "step": 467 }, { "epoch": 0.4517374517374517, "grad_norm": 4.400399684906006, "learning_rate": 1.9249726657195534e-05, "loss": 1.7185, "step": 468 }, { "epoch": 0.4527027027027027, "grad_norm": 3.828836679458618, "learning_rate": 1.9245760415896923e-05, "loss": 1.6929, "step": 469 }, { "epoch": 0.45366795366795365, "grad_norm": 5.331608772277832, "learning_rate": 1.924178412946174e-05, "loss": 1.7378, "step": 470 }, { "epoch": 0.4546332046332046, "grad_norm": 4.237404823303223, "learning_rate": 1.9237797802210067e-05, "loss": 1.7607, "step": 471 }, { "epoch": 0.4555984555984556, "grad_norm": 4.649611949920654, "learning_rate": 1.9233801438472875e-05, "loss": 1.7569, "step": 472 }, { "epoch": 0.45656370656370654, "grad_norm": 6.196529865264893, "learning_rate": 1.922979504259205e-05, "loss": 1.7296, "step": 473 }, { "epoch": 0.4575289575289575, "grad_norm": 3.977447748184204, "learning_rate": 1.922577861892037e-05, "loss": 1.7504, "step": 474 }, { "epoch": 0.4584942084942085, "grad_norm": 9.30162525177002, "learning_rate": 1.9221752171821523e-05, "loss": 1.6945, "step": 475 }, { "epoch": 0.4594594594594595, "grad_norm": 3.6693522930145264, "learning_rate": 1.9217715705670072e-05, "loss": 1.6373, "step": 476 }, { "epoch": 0.46042471042471045, "grad_norm": 6.122697830200195, "learning_rate": 1.9213669224851473e-05, "loss": 1.6249, "step": 477 }, { "epoch": 0.4613899613899614, "grad_norm": 5.823074817657471, "learning_rate": 1.9209612733762056e-05, "loss": 1.6743, "step": 478 }, { "epoch": 0.4623552123552124, "grad_norm": 3.764169216156006, "learning_rate": 1.9205546236809037e-05, "loss": 1.7555, "step": 479 }, { "epoch": 0.46332046332046334, "grad_norm": 4.683211326599121, "learning_rate": 1.9201469738410493e-05, "loss": 1.7276, "step": 480 }, { "epoch": 0.4642857142857143, "grad_norm": 5.23283576965332, "learning_rate": 1.919738324299537e-05, "loss": 1.6407, "step": 481 }, { "epoch": 0.46525096525096526, "grad_norm": 4.238620281219482, "learning_rate": 1.9193286755003475e-05, "loss": 1.6999, "step": 482 }, { "epoch": 0.46621621621621623, "grad_norm": 4.695125579833984, "learning_rate": 1.918918027888548e-05, "loss": 1.7013, "step": 483 }, { "epoch": 0.4671814671814672, "grad_norm": 5.553123474121094, "learning_rate": 1.9185063819102896e-05, "loss": 1.7287, "step": 484 }, { "epoch": 0.46814671814671815, "grad_norm": 4.4962077140808105, "learning_rate": 1.918093738012809e-05, "loss": 1.7472, "step": 485 }, { "epoch": 0.4691119691119691, "grad_norm": 5.718868255615234, "learning_rate": 1.9176800966444263e-05, "loss": 1.717, "step": 486 }, { "epoch": 0.4700772200772201, "grad_norm": 4.19191837310791, "learning_rate": 1.917265458254546e-05, "loss": 1.6781, "step": 487 }, { "epoch": 0.47104247104247104, "grad_norm": 4.743502616882324, "learning_rate": 1.9168498232936557e-05, "loss": 1.7112, "step": 488 }, { "epoch": 0.472007722007722, "grad_norm": 3.346381902694702, "learning_rate": 1.9164331922133254e-05, "loss": 1.7331, "step": 489 }, { "epoch": 0.47297297297297297, "grad_norm": 3.966032028198242, "learning_rate": 1.9160155654662075e-05, "loss": 1.69, "step": 490 }, { "epoch": 0.47393822393822393, "grad_norm": 4.457366943359375, "learning_rate": 1.9155969435060363e-05, "loss": 1.6704, "step": 491 }, { "epoch": 0.4749034749034749, "grad_norm": 4.079778671264648, "learning_rate": 1.915177326787627e-05, "loss": 1.6572, "step": 492 }, { "epoch": 0.47586872586872586, "grad_norm": 5.250837802886963, "learning_rate": 1.914756715766877e-05, "loss": 1.7621, "step": 493 }, { "epoch": 0.4768339768339768, "grad_norm": 4.468987464904785, "learning_rate": 1.914335110900761e-05, "loss": 1.6141, "step": 494 }, { "epoch": 0.4777992277992278, "grad_norm": 5.575342178344727, "learning_rate": 1.9139125126473365e-05, "loss": 1.6715, "step": 495 }, { "epoch": 0.47876447876447875, "grad_norm": 6.624782562255859, "learning_rate": 1.9134889214657385e-05, "loss": 1.7033, "step": 496 }, { "epoch": 0.4797297297297297, "grad_norm": 4.400557518005371, "learning_rate": 1.9130643378161817e-05, "loss": 1.649, "step": 497 }, { "epoch": 0.4806949806949807, "grad_norm": 5.939478397369385, "learning_rate": 1.9126387621599585e-05, "loss": 1.7254, "step": 498 }, { "epoch": 0.48166023166023164, "grad_norm": 4.109172344207764, "learning_rate": 1.9122121949594392e-05, "loss": 1.6279, "step": 499 }, { "epoch": 0.4826254826254826, "grad_norm": 4.4112701416015625, "learning_rate": 1.9117846366780714e-05, "loss": 1.7164, "step": 500 }, { "epoch": 0.48359073359073357, "grad_norm": 6.475155353546143, "learning_rate": 1.9113560877803798e-05, "loss": 1.663, "step": 501 }, { "epoch": 0.48455598455598453, "grad_norm": 4.766849040985107, "learning_rate": 1.9109265487319645e-05, "loss": 1.6338, "step": 502 }, { "epoch": 0.4855212355212355, "grad_norm": 5.895227432250977, "learning_rate": 1.910496019999502e-05, "loss": 1.7569, "step": 503 }, { "epoch": 0.4864864864864865, "grad_norm": 4.210174083709717, "learning_rate": 1.9100645020507448e-05, "loss": 1.7424, "step": 504 }, { "epoch": 0.4874517374517375, "grad_norm": 4.535717487335205, "learning_rate": 1.9096319953545186e-05, "loss": 1.7557, "step": 505 }, { "epoch": 0.48841698841698844, "grad_norm": 4.628802299499512, "learning_rate": 1.9091985003807242e-05, "loss": 1.6228, "step": 506 }, { "epoch": 0.4893822393822394, "grad_norm": 4.0775017738342285, "learning_rate": 1.9087640176003356e-05, "loss": 1.6476, "step": 507 }, { "epoch": 0.49034749034749037, "grad_norm": 5.471363067626953, "learning_rate": 1.9083285474854012e-05, "loss": 1.6965, "step": 508 }, { "epoch": 0.49131274131274133, "grad_norm": 3.517397880554199, "learning_rate": 1.9078920905090408e-05, "loss": 1.6821, "step": 509 }, { "epoch": 0.4922779922779923, "grad_norm": 5.442646503448486, "learning_rate": 1.907454647145447e-05, "loss": 1.7628, "step": 510 }, { "epoch": 0.49324324324324326, "grad_norm": 3.913710355758667, "learning_rate": 1.9070162178698838e-05, "loss": 1.7004, "step": 511 }, { "epoch": 0.4942084942084942, "grad_norm": 3.85947322845459, "learning_rate": 1.9065768031586864e-05, "loss": 1.6542, "step": 512 }, { "epoch": 0.4951737451737452, "grad_norm": 4.821897983551025, "learning_rate": 1.906136403489261e-05, "loss": 1.6969, "step": 513 }, { "epoch": 0.49613899613899615, "grad_norm": 3.840449094772339, "learning_rate": 1.905695019340083e-05, "loss": 1.7016, "step": 514 }, { "epoch": 0.4971042471042471, "grad_norm": 5.120670795440674, "learning_rate": 1.9052526511906993e-05, "loss": 1.7121, "step": 515 }, { "epoch": 0.4980694980694981, "grad_norm": 4.671112060546875, "learning_rate": 1.9048092995217234e-05, "loss": 1.7133, "step": 516 }, { "epoch": 0.49903474903474904, "grad_norm": 4.036538600921631, "learning_rate": 1.904364964814839e-05, "loss": 1.6991, "step": 517 }, { "epoch": 0.5, "grad_norm": 4.0499701499938965, "learning_rate": 1.9039196475527974e-05, "loss": 1.7296, "step": 518 }, { "epoch": 0.500965250965251, "grad_norm": 4.240586757659912, "learning_rate": 1.9034733482194176e-05, "loss": 1.6821, "step": 519 }, { "epoch": 0.5019305019305019, "grad_norm": 3.5308685302734375, "learning_rate": 1.903026067299585e-05, "loss": 1.6224, "step": 520 }, { "epoch": 0.502895752895753, "grad_norm": 3.4554286003112793, "learning_rate": 1.9025778052792518e-05, "loss": 1.7154, "step": 521 }, { "epoch": 0.5038610038610039, "grad_norm": 3.995757818222046, "learning_rate": 1.902128562645437e-05, "loss": 1.7159, "step": 522 }, { "epoch": 0.5048262548262549, "grad_norm": 3.843982696533203, "learning_rate": 1.901678339886223e-05, "loss": 1.7729, "step": 523 }, { "epoch": 0.5057915057915058, "grad_norm": 3.784400463104248, "learning_rate": 1.901227137490759e-05, "loss": 1.6779, "step": 524 }, { "epoch": 0.5067567567567568, "grad_norm": 4.4152350425720215, "learning_rate": 1.9007749559492575e-05, "loss": 1.6983, "step": 525 }, { "epoch": 0.5077220077220077, "grad_norm": 3.8352880477905273, "learning_rate": 1.9003217957529948e-05, "loss": 1.6381, "step": 526 }, { "epoch": 0.5086872586872587, "grad_norm": 3.8873074054718018, "learning_rate": 1.899867657394311e-05, "loss": 1.7291, "step": 527 }, { "epoch": 0.5096525096525096, "grad_norm": 4.419820785522461, "learning_rate": 1.899412541366609e-05, "loss": 1.6595, "step": 528 }, { "epoch": 0.5106177606177607, "grad_norm": 3.6705615520477295, "learning_rate": 1.898956448164353e-05, "loss": 1.6919, "step": 529 }, { "epoch": 0.5115830115830116, "grad_norm": 3.7511744499206543, "learning_rate": 1.8984993782830695e-05, "loss": 1.6396, "step": 530 }, { "epoch": 0.5125482625482626, "grad_norm": 4.202239990234375, "learning_rate": 1.8980413322193462e-05, "loss": 1.6633, "step": 531 }, { "epoch": 0.5135135135135135, "grad_norm": 3.8342583179473877, "learning_rate": 1.8975823104708314e-05, "loss": 1.6892, "step": 532 }, { "epoch": 0.5144787644787645, "grad_norm": 3.90169095993042, "learning_rate": 1.8971223135362328e-05, "loss": 1.7225, "step": 533 }, { "epoch": 0.5154440154440154, "grad_norm": 4.5637688636779785, "learning_rate": 1.896661341915318e-05, "loss": 1.7099, "step": 534 }, { "epoch": 0.5164092664092664, "grad_norm": 3.929818868637085, "learning_rate": 1.8961993961089145e-05, "loss": 1.7169, "step": 535 }, { "epoch": 0.5173745173745173, "grad_norm": 4.224812984466553, "learning_rate": 1.895736476618906e-05, "loss": 1.7094, "step": 536 }, { "epoch": 0.5183397683397684, "grad_norm": 4.2393317222595215, "learning_rate": 1.8952725839482368e-05, "loss": 1.6806, "step": 537 }, { "epoch": 0.5193050193050193, "grad_norm": 4.0183305740356445, "learning_rate": 1.894807718600906e-05, "loss": 1.7971, "step": 538 }, { "epoch": 0.5202702702702703, "grad_norm": 4.289072513580322, "learning_rate": 1.8943418810819715e-05, "loss": 1.6686, "step": 539 }, { "epoch": 0.5212355212355212, "grad_norm": 4.444197654724121, "learning_rate": 1.8938750718975456e-05, "loss": 1.7366, "step": 540 }, { "epoch": 0.5222007722007722, "grad_norm": 3.5191142559051514, "learning_rate": 1.8934072915547978e-05, "loss": 1.7118, "step": 541 }, { "epoch": 0.5231660231660231, "grad_norm": 3.8551297187805176, "learning_rate": 1.892938540561952e-05, "loss": 1.6944, "step": 542 }, { "epoch": 0.5241312741312741, "grad_norm": 4.650771141052246, "learning_rate": 1.8924688194282862e-05, "loss": 1.6372, "step": 543 }, { "epoch": 0.525096525096525, "grad_norm": 4.20098352432251, "learning_rate": 1.8919981286641336e-05, "loss": 1.6815, "step": 544 }, { "epoch": 0.5260617760617761, "grad_norm": 3.853182315826416, "learning_rate": 1.891526468780881e-05, "loss": 1.6536, "step": 545 }, { "epoch": 0.527027027027027, "grad_norm": 3.648350954055786, "learning_rate": 1.8910538402909658e-05, "loss": 1.6387, "step": 546 }, { "epoch": 0.527992277992278, "grad_norm": 3.5757272243499756, "learning_rate": 1.89058024370788e-05, "loss": 1.6526, "step": 547 }, { "epoch": 0.528957528957529, "grad_norm": 3.996739149093628, "learning_rate": 1.890105679546167e-05, "loss": 1.7263, "step": 548 }, { "epoch": 0.5299227799227799, "grad_norm": 3.7277631759643555, "learning_rate": 1.8896301483214207e-05, "loss": 1.6864, "step": 549 }, { "epoch": 0.5308880308880309, "grad_norm": 3.406763792037964, "learning_rate": 1.8891536505502865e-05, "loss": 1.6984, "step": 550 }, { "epoch": 0.5318532818532818, "grad_norm": 4.051969051361084, "learning_rate": 1.8886761867504595e-05, "loss": 1.657, "step": 551 }, { "epoch": 0.5328185328185329, "grad_norm": 3.5501787662506104, "learning_rate": 1.8881977574406842e-05, "loss": 1.6886, "step": 552 }, { "epoch": 0.5337837837837838, "grad_norm": 3.881225109100342, "learning_rate": 1.887718363140754e-05, "loss": 1.7374, "step": 553 }, { "epoch": 0.5347490347490348, "grad_norm": 4.034554481506348, "learning_rate": 1.8872380043715117e-05, "loss": 1.6904, "step": 554 }, { "epoch": 0.5357142857142857, "grad_norm": 3.6535398960113525, "learning_rate": 1.8867566816548468e-05, "loss": 1.7026, "step": 555 }, { "epoch": 0.5366795366795367, "grad_norm": 3.7659504413604736, "learning_rate": 1.8862743955136966e-05, "loss": 1.6495, "step": 556 }, { "epoch": 0.5376447876447876, "grad_norm": 4.140193939208984, "learning_rate": 1.885791146472045e-05, "loss": 1.7337, "step": 557 }, { "epoch": 0.5386100386100386, "grad_norm": 3.7351629734039307, "learning_rate": 1.8853069350549223e-05, "loss": 1.7096, "step": 558 }, { "epoch": 0.5395752895752896, "grad_norm": 5.067814826965332, "learning_rate": 1.884821761788404e-05, "loss": 1.6834, "step": 559 }, { "epoch": 0.5405405405405406, "grad_norm": 3.812654733657837, "learning_rate": 1.8843356271996106e-05, "loss": 1.6414, "step": 560 }, { "epoch": 0.5415057915057915, "grad_norm": 4.1293158531188965, "learning_rate": 1.8838485318167078e-05, "loss": 1.7375, "step": 561 }, { "epoch": 0.5424710424710425, "grad_norm": 4.6960625648498535, "learning_rate": 1.8833604761689045e-05, "loss": 1.6657, "step": 562 }, { "epoch": 0.5434362934362934, "grad_norm": 3.738614320755005, "learning_rate": 1.8828714607864524e-05, "loss": 1.6944, "step": 563 }, { "epoch": 0.5444015444015444, "grad_norm": 4.760887145996094, "learning_rate": 1.882381486200647e-05, "loss": 1.6449, "step": 564 }, { "epoch": 0.5453667953667953, "grad_norm": 3.6925551891326904, "learning_rate": 1.881890552943826e-05, "loss": 1.6279, "step": 565 }, { "epoch": 0.5463320463320464, "grad_norm": 3.980975389480591, "learning_rate": 1.8813986615493664e-05, "loss": 1.7143, "step": 566 }, { "epoch": 0.5472972972972973, "grad_norm": 3.7689743041992188, "learning_rate": 1.8809058125516894e-05, "loss": 1.6662, "step": 567 }, { "epoch": 0.5482625482625483, "grad_norm": 3.663180351257324, "learning_rate": 1.8804120064862547e-05, "loss": 1.671, "step": 568 }, { "epoch": 0.5492277992277992, "grad_norm": 4.199854373931885, "learning_rate": 1.879917243889562e-05, "loss": 1.7025, "step": 569 }, { "epoch": 0.5501930501930502, "grad_norm": 3.8860297203063965, "learning_rate": 1.8794215252991504e-05, "loss": 1.7047, "step": 570 }, { "epoch": 0.5511583011583011, "grad_norm": 4.648730754852295, "learning_rate": 1.8789248512535978e-05, "loss": 1.6904, "step": 571 }, { "epoch": 0.5521235521235521, "grad_norm": 4.810901641845703, "learning_rate": 1.87842722229252e-05, "loss": 1.7054, "step": 572 }, { "epoch": 0.553088803088803, "grad_norm": 4.9697041511535645, "learning_rate": 1.87792863895657e-05, "loss": 1.6567, "step": 573 }, { "epoch": 0.5540540540540541, "grad_norm": 3.7920072078704834, "learning_rate": 1.8774291017874384e-05, "loss": 1.6878, "step": 574 }, { "epoch": 0.555019305019305, "grad_norm": 3.894486904144287, "learning_rate": 1.8769286113278515e-05, "loss": 1.6685, "step": 575 }, { "epoch": 0.555984555984556, "grad_norm": 3.9627492427825928, "learning_rate": 1.8764271681215714e-05, "loss": 1.6961, "step": 576 }, { "epoch": 0.556949806949807, "grad_norm": 4.0938591957092285, "learning_rate": 1.8759247727133956e-05, "loss": 1.6871, "step": 577 }, { "epoch": 0.5579150579150579, "grad_norm": 3.893789768218994, "learning_rate": 1.8754214256491564e-05, "loss": 1.6215, "step": 578 }, { "epoch": 0.5588803088803089, "grad_norm": 3.9757490158081055, "learning_rate": 1.8749171274757184e-05, "loss": 1.6701, "step": 579 }, { "epoch": 0.5598455598455598, "grad_norm": 4.111543655395508, "learning_rate": 1.874411878740982e-05, "loss": 1.7002, "step": 580 }, { "epoch": 0.5608108108108109, "grad_norm": 4.3574299812316895, "learning_rate": 1.873905679993878e-05, "loss": 1.656, "step": 581 }, { "epoch": 0.5617760617760618, "grad_norm": 4.900210857391357, "learning_rate": 1.873398531784371e-05, "loss": 1.6724, "step": 582 }, { "epoch": 0.5627413127413128, "grad_norm": 4.666040420532227, "learning_rate": 1.8728904346634563e-05, "loss": 1.6004, "step": 583 }, { "epoch": 0.5637065637065637, "grad_norm": 4.129469871520996, "learning_rate": 1.872381389183161e-05, "loss": 1.7019, "step": 584 }, { "epoch": 0.5646718146718147, "grad_norm": 4.4677934646606445, "learning_rate": 1.8718713958965413e-05, "loss": 1.7334, "step": 585 }, { "epoch": 0.5656370656370656, "grad_norm": 3.5727407932281494, "learning_rate": 1.8713604553576842e-05, "loss": 1.6449, "step": 586 }, { "epoch": 0.5666023166023166, "grad_norm": 3.8256876468658447, "learning_rate": 1.870848568121705e-05, "loss": 1.6638, "step": 587 }, { "epoch": 0.5675675675675675, "grad_norm": 3.4252405166625977, "learning_rate": 1.8703357347447487e-05, "loss": 1.6841, "step": 588 }, { "epoch": 0.5685328185328186, "grad_norm": 4.170039176940918, "learning_rate": 1.8698219557839875e-05, "loss": 1.5711, "step": 589 }, { "epoch": 0.5694980694980695, "grad_norm": 3.6318793296813965, "learning_rate": 1.8693072317976205e-05, "loss": 1.6643, "step": 590 }, { "epoch": 0.5704633204633205, "grad_norm": 3.745554208755493, "learning_rate": 1.868791563344874e-05, "loss": 1.6813, "step": 591 }, { "epoch": 0.5714285714285714, "grad_norm": 5.6081461906433105, "learning_rate": 1.868274950986001e-05, "loss": 1.6829, "step": 592 }, { "epoch": 0.5723938223938224, "grad_norm": 4.0746541023254395, "learning_rate": 1.8677573952822793e-05, "loss": 1.7016, "step": 593 }, { "epoch": 0.5733590733590733, "grad_norm": 5.50642728805542, "learning_rate": 1.867238896796012e-05, "loss": 1.6679, "step": 594 }, { "epoch": 0.5743243243243243, "grad_norm": 4.05189847946167, "learning_rate": 1.8667194560905255e-05, "loss": 1.6469, "step": 595 }, { "epoch": 0.5752895752895753, "grad_norm": 4.773981094360352, "learning_rate": 1.8661990737301714e-05, "loss": 1.6685, "step": 596 }, { "epoch": 0.5762548262548263, "grad_norm": 4.233147144317627, "learning_rate": 1.865677750280323e-05, "loss": 1.7635, "step": 597 }, { "epoch": 0.5772200772200772, "grad_norm": 5.221375465393066, "learning_rate": 1.8651554863073776e-05, "loss": 1.6962, "step": 598 }, { "epoch": 0.5781853281853282, "grad_norm": 3.8712682723999023, "learning_rate": 1.864632282378753e-05, "loss": 1.6369, "step": 599 }, { "epoch": 0.5791505791505791, "grad_norm": 4.542074203491211, "learning_rate": 1.864108139062888e-05, "loss": 1.6512, "step": 600 }, { "epoch": 0.5801158301158301, "grad_norm": 5.4050374031066895, "learning_rate": 1.8635830569292434e-05, "loss": 1.7831, "step": 601 }, { "epoch": 0.581081081081081, "grad_norm": 4.166981220245361, "learning_rate": 1.863057036548299e-05, "loss": 1.7131, "step": 602 }, { "epoch": 0.582046332046332, "grad_norm": 5.486953258514404, "learning_rate": 1.862530078491554e-05, "loss": 1.6969, "step": 603 }, { "epoch": 0.583011583011583, "grad_norm": 4.638899326324463, "learning_rate": 1.862002183331527e-05, "loss": 1.6511, "step": 604 }, { "epoch": 0.583976833976834, "grad_norm": 4.579814910888672, "learning_rate": 1.8614733516417543e-05, "loss": 1.6889, "step": 605 }, { "epoch": 0.584942084942085, "grad_norm": 5.79071044921875, "learning_rate": 1.860943583996789e-05, "loss": 1.654, "step": 606 }, { "epoch": 0.5859073359073359, "grad_norm": 3.7253856658935547, "learning_rate": 1.8604128809722024e-05, "loss": 1.6732, "step": 607 }, { "epoch": 0.5868725868725869, "grad_norm": 4.581823825836182, "learning_rate": 1.859881243144581e-05, "loss": 1.6502, "step": 608 }, { "epoch": 0.5878378378378378, "grad_norm": 4.885397911071777, "learning_rate": 1.8593486710915277e-05, "loss": 1.7279, "step": 609 }, { "epoch": 0.5888030888030888, "grad_norm": 3.970855236053467, "learning_rate": 1.8588151653916596e-05, "loss": 1.6674, "step": 610 }, { "epoch": 0.5897683397683398, "grad_norm": 4.714148044586182, "learning_rate": 1.858280726624609e-05, "loss": 1.6763, "step": 611 }, { "epoch": 0.5907335907335908, "grad_norm": 6.035184860229492, "learning_rate": 1.8577453553710213e-05, "loss": 1.7452, "step": 612 }, { "epoch": 0.5916988416988417, "grad_norm": 4.146158695220947, "learning_rate": 1.8572090522125553e-05, "loss": 1.6846, "step": 613 }, { "epoch": 0.5926640926640927, "grad_norm": 5.308503150939941, "learning_rate": 1.8566718177318825e-05, "loss": 1.6529, "step": 614 }, { "epoch": 0.5936293436293436, "grad_norm": 4.472123146057129, "learning_rate": 1.856133652512685e-05, "loss": 1.734, "step": 615 }, { "epoch": 0.5945945945945946, "grad_norm": 3.907163143157959, "learning_rate": 1.8555945571396584e-05, "loss": 1.6416, "step": 616 }, { "epoch": 0.5955598455598455, "grad_norm": 4.636082172393799, "learning_rate": 1.8550545321985065e-05, "loss": 1.6213, "step": 617 }, { "epoch": 0.5965250965250966, "grad_norm": 3.847663640975952, "learning_rate": 1.8545135782759446e-05, "loss": 1.6378, "step": 618 }, { "epoch": 0.5974903474903475, "grad_norm": 4.3884501457214355, "learning_rate": 1.8539716959596967e-05, "loss": 1.7255, "step": 619 }, { "epoch": 0.5984555984555985, "grad_norm": 3.7094476222991943, "learning_rate": 1.853428885838495e-05, "loss": 1.6741, "step": 620 }, { "epoch": 0.5994208494208494, "grad_norm": 3.8970134258270264, "learning_rate": 1.852885148502081e-05, "loss": 1.6916, "step": 621 }, { "epoch": 0.6003861003861004, "grad_norm": 3.7777106761932373, "learning_rate": 1.8523404845412028e-05, "loss": 1.6407, "step": 622 }, { "epoch": 0.6013513513513513, "grad_norm": 3.556906223297119, "learning_rate": 1.851794894547615e-05, "loss": 1.6999, "step": 623 }, { "epoch": 0.6023166023166023, "grad_norm": 4.2413105964660645, "learning_rate": 1.8512483791140787e-05, "loss": 1.6502, "step": 624 }, { "epoch": 0.6032818532818532, "grad_norm": 3.9186458587646484, "learning_rate": 1.85070093883436e-05, "loss": 1.5961, "step": 625 }, { "epoch": 0.6042471042471043, "grad_norm": 3.8897972106933594, "learning_rate": 1.8501525743032305e-05, "loss": 1.7424, "step": 626 }, { "epoch": 0.6052123552123552, "grad_norm": 4.06066370010376, "learning_rate": 1.8496032861164657e-05, "loss": 1.6369, "step": 627 }, { "epoch": 0.6061776061776062, "grad_norm": 3.7971489429473877, "learning_rate": 1.8490530748708442e-05, "loss": 1.7034, "step": 628 }, { "epoch": 0.6071428571428571, "grad_norm": 3.7843589782714844, "learning_rate": 1.8485019411641477e-05, "loss": 1.6269, "step": 629 }, { "epoch": 0.6081081081081081, "grad_norm": 3.9630510807037354, "learning_rate": 1.8479498855951607e-05, "loss": 1.7404, "step": 630 }, { "epoch": 0.609073359073359, "grad_norm": 3.554232120513916, "learning_rate": 1.8473969087636683e-05, "loss": 1.7156, "step": 631 }, { "epoch": 0.61003861003861, "grad_norm": 4.078527450561523, "learning_rate": 1.8468430112704574e-05, "loss": 1.6219, "step": 632 }, { "epoch": 0.611003861003861, "grad_norm": 3.7438254356384277, "learning_rate": 1.8462881937173144e-05, "loss": 1.6589, "step": 633 }, { "epoch": 0.611969111969112, "grad_norm": 3.675246238708496, "learning_rate": 1.8457324567070255e-05, "loss": 1.7171, "step": 634 }, { "epoch": 0.612934362934363, "grad_norm": 4.8114519119262695, "learning_rate": 1.845175800843376e-05, "loss": 1.6456, "step": 635 }, { "epoch": 0.6138996138996139, "grad_norm": 4.08369255065918, "learning_rate": 1.84461822673115e-05, "loss": 1.7574, "step": 636 }, { "epoch": 0.6148648648648649, "grad_norm": 4.7406206130981445, "learning_rate": 1.8440597349761286e-05, "loss": 1.6294, "step": 637 }, { "epoch": 0.6158301158301158, "grad_norm": 4.751537322998047, "learning_rate": 1.8435003261850895e-05, "loss": 1.641, "step": 638 }, { "epoch": 0.6167953667953668, "grad_norm": 4.042719841003418, "learning_rate": 1.8429400009658076e-05, "loss": 1.5904, "step": 639 }, { "epoch": 0.6177606177606177, "grad_norm": 3.7075490951538086, "learning_rate": 1.842378759927053e-05, "loss": 1.6742, "step": 640 }, { "epoch": 0.6187258687258688, "grad_norm": 3.8176772594451904, "learning_rate": 1.8418166036785912e-05, "loss": 1.6862, "step": 641 }, { "epoch": 0.6196911196911197, "grad_norm": 3.418301820755005, "learning_rate": 1.8412535328311813e-05, "loss": 1.6231, "step": 642 }, { "epoch": 0.6206563706563707, "grad_norm": 3.5319151878356934, "learning_rate": 1.8406895479965766e-05, "loss": 1.7387, "step": 643 }, { "epoch": 0.6216216216216216, "grad_norm": 4.254901885986328, "learning_rate": 1.8401246497875238e-05, "loss": 1.6421, "step": 644 }, { "epoch": 0.6225868725868726, "grad_norm": 4.021042823791504, "learning_rate": 1.8395588388177606e-05, "loss": 1.6828, "step": 645 }, { "epoch": 0.6235521235521235, "grad_norm": 3.8519716262817383, "learning_rate": 1.838992115702018e-05, "loss": 1.613, "step": 646 }, { "epoch": 0.6245173745173745, "grad_norm": 5.556555271148682, "learning_rate": 1.8384244810560166e-05, "loss": 1.6835, "step": 647 }, { "epoch": 0.6254826254826255, "grad_norm": 5.064825057983398, "learning_rate": 1.8378559354964686e-05, "loss": 1.6579, "step": 648 }, { "epoch": 0.6264478764478765, "grad_norm": 3.874995470046997, "learning_rate": 1.8372864796410748e-05, "loss": 1.6936, "step": 649 }, { "epoch": 0.6274131274131274, "grad_norm": 3.5412216186523438, "learning_rate": 1.8367161141085256e-05, "loss": 1.6235, "step": 650 }, { "epoch": 0.6283783783783784, "grad_norm": 3.8917672634124756, "learning_rate": 1.8361448395184995e-05, "loss": 1.7078, "step": 651 }, { "epoch": 0.6293436293436293, "grad_norm": 3.8688552379608154, "learning_rate": 1.835572656491663e-05, "loss": 1.6606, "step": 652 }, { "epoch": 0.6303088803088803, "grad_norm": 4.006204605102539, "learning_rate": 1.8349995656496692e-05, "loss": 1.66, "step": 653 }, { "epoch": 0.6312741312741312, "grad_norm": 4.492037773132324, "learning_rate": 1.8344255676151572e-05, "loss": 1.6344, "step": 654 }, { "epoch": 0.6322393822393823, "grad_norm": 5.894545555114746, "learning_rate": 1.8338506630117527e-05, "loss": 1.7011, "step": 655 }, { "epoch": 0.6332046332046332, "grad_norm": 3.9362146854400635, "learning_rate": 1.8332748524640657e-05, "loss": 1.5894, "step": 656 }, { "epoch": 0.6341698841698842, "grad_norm": 4.3061113357543945, "learning_rate": 1.8326981365976903e-05, "loss": 1.6885, "step": 657 }, { "epoch": 0.6351351351351351, "grad_norm": 5.677443504333496, "learning_rate": 1.832120516039205e-05, "loss": 1.6736, "step": 658 }, { "epoch": 0.6361003861003861, "grad_norm": 3.6564252376556396, "learning_rate": 1.8315419914161697e-05, "loss": 1.5815, "step": 659 }, { "epoch": 0.637065637065637, "grad_norm": 8.331316947937012, "learning_rate": 1.8309625633571283e-05, "loss": 1.6649, "step": 660 }, { "epoch": 0.638030888030888, "grad_norm": 4.027668476104736, "learning_rate": 1.830382232491605e-05, "loss": 1.645, "step": 661 }, { "epoch": 0.638996138996139, "grad_norm": 5.594932556152344, "learning_rate": 1.829800999450106e-05, "loss": 1.687, "step": 662 }, { "epoch": 0.63996138996139, "grad_norm": 4.846828937530518, "learning_rate": 1.829218864864116e-05, "loss": 1.6671, "step": 663 }, { "epoch": 0.640926640926641, "grad_norm": 3.602202892303467, "learning_rate": 1.8286358293661007e-05, "loss": 1.6521, "step": 664 }, { "epoch": 0.6418918918918919, "grad_norm": 5.479918956756592, "learning_rate": 1.828051893589504e-05, "loss": 1.6227, "step": 665 }, { "epoch": 0.6428571428571429, "grad_norm": 4.227159023284912, "learning_rate": 1.827467058168748e-05, "loss": 1.6611, "step": 666 }, { "epoch": 0.6438223938223938, "grad_norm": 4.943421840667725, "learning_rate": 1.8268813237392323e-05, "loss": 1.6425, "step": 667 }, { "epoch": 0.6447876447876448, "grad_norm": 3.53085994720459, "learning_rate": 1.826294690937333e-05, "loss": 1.6765, "step": 668 }, { "epoch": 0.6457528957528957, "grad_norm": 3.979865312576294, "learning_rate": 1.8257071604004025e-05, "loss": 1.7101, "step": 669 }, { "epoch": 0.6467181467181468, "grad_norm": 3.7377078533172607, "learning_rate": 1.8251187327667682e-05, "loss": 1.6569, "step": 670 }, { "epoch": 0.6476833976833977, "grad_norm": 3.7672817707061768, "learning_rate": 1.824529408675733e-05, "loss": 1.6771, "step": 671 }, { "epoch": 0.6486486486486487, "grad_norm": 3.766531467437744, "learning_rate": 1.8239391887675725e-05, "loss": 1.6386, "step": 672 }, { "epoch": 0.6496138996138996, "grad_norm": 4.239291667938232, "learning_rate": 1.8233480736835366e-05, "loss": 1.7249, "step": 673 }, { "epoch": 0.6505791505791506, "grad_norm": 6.503885269165039, "learning_rate": 1.822756064065847e-05, "loss": 1.6037, "step": 674 }, { "epoch": 0.6515444015444015, "grad_norm": 3.618699073791504, "learning_rate": 1.8221631605576983e-05, "loss": 1.6653, "step": 675 }, { "epoch": 0.6525096525096525, "grad_norm": 4.299498081207275, "learning_rate": 1.8215693638032555e-05, "loss": 1.6083, "step": 676 }, { "epoch": 0.6534749034749034, "grad_norm": 6.6480631828308105, "learning_rate": 1.8209746744476538e-05, "loss": 1.7064, "step": 677 }, { "epoch": 0.6544401544401545, "grad_norm": 3.544414520263672, "learning_rate": 1.820379093136999e-05, "loss": 1.6489, "step": 678 }, { "epoch": 0.6554054054054054, "grad_norm": 5.126187324523926, "learning_rate": 1.8197826205183657e-05, "loss": 1.6937, "step": 679 }, { "epoch": 0.6563706563706564, "grad_norm": 5.806528091430664, "learning_rate": 1.819185257239796e-05, "loss": 1.7678, "step": 680 }, { "epoch": 0.6573359073359073, "grad_norm": 4.069647312164307, "learning_rate": 1.8185870039503014e-05, "loss": 1.6307, "step": 681 }, { "epoch": 0.6583011583011583, "grad_norm": 4.525701522827148, "learning_rate": 1.8179878612998586e-05, "loss": 1.6497, "step": 682 }, { "epoch": 0.6592664092664092, "grad_norm": 5.206305027008057, "learning_rate": 1.8173878299394118e-05, "loss": 1.7508, "step": 683 }, { "epoch": 0.6602316602316602, "grad_norm": 3.8882358074188232, "learning_rate": 1.8167869105208703e-05, "loss": 1.6425, "step": 684 }, { "epoch": 0.6611969111969112, "grad_norm": 5.368058681488037, "learning_rate": 1.816185103697108e-05, "loss": 1.6902, "step": 685 }, { "epoch": 0.6621621621621622, "grad_norm": 3.8239119052886963, "learning_rate": 1.8155824101219626e-05, "loss": 1.5893, "step": 686 }, { "epoch": 0.6631274131274131, "grad_norm": 5.2074198722839355, "learning_rate": 1.814978830450237e-05, "loss": 1.7116, "step": 687 }, { "epoch": 0.6640926640926641, "grad_norm": 3.5977511405944824, "learning_rate": 1.8143743653376944e-05, "loss": 1.6007, "step": 688 }, { "epoch": 0.665057915057915, "grad_norm": 4.01193380355835, "learning_rate": 1.8137690154410617e-05, "loss": 1.649, "step": 689 }, { "epoch": 0.666023166023166, "grad_norm": 4.035015106201172, "learning_rate": 1.8131627814180264e-05, "loss": 1.6641, "step": 690 }, { "epoch": 0.666988416988417, "grad_norm": 3.439990282058716, "learning_rate": 1.8125556639272375e-05, "loss": 1.627, "step": 691 }, { "epoch": 0.667953667953668, "grad_norm": 3.9323132038116455, "learning_rate": 1.8119476636283018e-05, "loss": 1.688, "step": 692 }, { "epoch": 0.668918918918919, "grad_norm": 4.216135025024414, "learning_rate": 1.811338781181787e-05, "loss": 1.6948, "step": 693 }, { "epoch": 0.6698841698841699, "grad_norm": 3.865110158920288, "learning_rate": 1.8107290172492193e-05, "loss": 1.6122, "step": 694 }, { "epoch": 0.6708494208494209, "grad_norm": 4.306641578674316, "learning_rate": 1.8101183724930812e-05, "loss": 1.7233, "step": 695 }, { "epoch": 0.6718146718146718, "grad_norm": 3.8325014114379883, "learning_rate": 1.809506847576813e-05, "loss": 1.6368, "step": 696 }, { "epoch": 0.6727799227799228, "grad_norm": 5.356362819671631, "learning_rate": 1.8088944431648118e-05, "loss": 1.651, "step": 697 }, { "epoch": 0.6737451737451737, "grad_norm": 4.689187526702881, "learning_rate": 1.8082811599224293e-05, "loss": 1.6605, "step": 698 }, { "epoch": 0.6747104247104247, "grad_norm": 4.608354568481445, "learning_rate": 1.8076669985159726e-05, "loss": 1.6396, "step": 699 }, { "epoch": 0.6756756756756757, "grad_norm": 5.5505757331848145, "learning_rate": 1.8070519596127027e-05, "loss": 1.7028, "step": 700 }, { "epoch": 0.6766409266409267, "grad_norm": 3.8463668823242188, "learning_rate": 1.8064360438808335e-05, "loss": 1.6099, "step": 701 }, { "epoch": 0.6776061776061776, "grad_norm": 5.928972244262695, "learning_rate": 1.805819251989533e-05, "loss": 1.6517, "step": 702 }, { "epoch": 0.6785714285714286, "grad_norm": 3.801046848297119, "learning_rate": 1.8052015846089187e-05, "loss": 1.6492, "step": 703 }, { "epoch": 0.6795366795366795, "grad_norm": 5.3330302238464355, "learning_rate": 1.804583042410062e-05, "loss": 1.6213, "step": 704 }, { "epoch": 0.6805019305019305, "grad_norm": 4.596437931060791, "learning_rate": 1.8039636260649823e-05, "loss": 1.6368, "step": 705 }, { "epoch": 0.6814671814671814, "grad_norm": 4.642145156860352, "learning_rate": 1.8033433362466506e-05, "loss": 1.618, "step": 706 }, { "epoch": 0.6824324324324325, "grad_norm": 5.829033374786377, "learning_rate": 1.802722173628986e-05, "loss": 1.6833, "step": 707 }, { "epoch": 0.6833976833976834, "grad_norm": 4.1263017654418945, "learning_rate": 1.802100138886856e-05, "loss": 1.717, "step": 708 }, { "epoch": 0.6843629343629344, "grad_norm": 5.381926536560059, "learning_rate": 1.8014772326960758e-05, "loss": 1.6017, "step": 709 }, { "epoch": 0.6853281853281853, "grad_norm": 6.798498630523682, "learning_rate": 1.8008534557334064e-05, "loss": 1.6535, "step": 710 }, { "epoch": 0.6862934362934363, "grad_norm": 3.4682328701019287, "learning_rate": 1.800228808676557e-05, "loss": 1.5921, "step": 711 }, { "epoch": 0.6872586872586872, "grad_norm": 5.673734188079834, "learning_rate": 1.7996032922041797e-05, "loss": 1.6136, "step": 712 }, { "epoch": 0.6882239382239382, "grad_norm": 6.470987319946289, "learning_rate": 1.7989769069958728e-05, "loss": 1.7331, "step": 713 }, { "epoch": 0.6891891891891891, "grad_norm": 3.7526819705963135, "learning_rate": 1.798349653732178e-05, "loss": 1.6889, "step": 714 }, { "epoch": 0.6901544401544402, "grad_norm": 7.26838493347168, "learning_rate": 1.79772153309458e-05, "loss": 1.6973, "step": 715 }, { "epoch": 0.6911196911196911, "grad_norm": 3.74743390083313, "learning_rate": 1.7970925457655055e-05, "loss": 1.682, "step": 716 }, { "epoch": 0.6920849420849421, "grad_norm": 4.3754191398620605, "learning_rate": 1.7964626924283238e-05, "loss": 1.5958, "step": 717 }, { "epoch": 0.693050193050193, "grad_norm": 4.73179292678833, "learning_rate": 1.7958319737673444e-05, "loss": 1.6401, "step": 718 }, { "epoch": 0.694015444015444, "grad_norm": 4.429830074310303, "learning_rate": 1.795200390467817e-05, "loss": 1.6231, "step": 719 }, { "epoch": 0.694980694980695, "grad_norm": 3.9136927127838135, "learning_rate": 1.7945679432159302e-05, "loss": 1.5997, "step": 720 }, { "epoch": 0.6959459459459459, "grad_norm": 4.052205562591553, "learning_rate": 1.7939346326988127e-05, "loss": 1.6286, "step": 721 }, { "epoch": 0.696911196911197, "grad_norm": 4.2305450439453125, "learning_rate": 1.7933004596045294e-05, "loss": 1.6407, "step": 722 }, { "epoch": 0.6978764478764479, "grad_norm": 3.89178729057312, "learning_rate": 1.7926654246220843e-05, "loss": 1.7152, "step": 723 }, { "epoch": 0.6988416988416989, "grad_norm": 4.049131870269775, "learning_rate": 1.7920295284414153e-05, "loss": 1.7144, "step": 724 }, { "epoch": 0.6998069498069498, "grad_norm": 3.665480136871338, "learning_rate": 1.791392771753398e-05, "loss": 1.6139, "step": 725 }, { "epoch": 0.7007722007722008, "grad_norm": 4.4121928215026855, "learning_rate": 1.7907551552498423e-05, "loss": 1.6958, "step": 726 }, { "epoch": 0.7017374517374517, "grad_norm": 4.521336078643799, "learning_rate": 1.7901166796234922e-05, "loss": 1.7486, "step": 727 }, { "epoch": 0.7027027027027027, "grad_norm": 3.2848575115203857, "learning_rate": 1.789477345568025e-05, "loss": 1.6371, "step": 728 }, { "epoch": 0.7036679536679536, "grad_norm": 3.7929630279541016, "learning_rate": 1.78883715377805e-05, "loss": 1.5856, "step": 729 }, { "epoch": 0.7046332046332047, "grad_norm": 3.8128833770751953, "learning_rate": 1.78819610494911e-05, "loss": 1.6768, "step": 730 }, { "epoch": 0.7055984555984556, "grad_norm": 3.6518142223358154, "learning_rate": 1.7875541997776783e-05, "loss": 1.5678, "step": 731 }, { "epoch": 0.7065637065637066, "grad_norm": 3.605537176132202, "learning_rate": 1.7869114389611574e-05, "loss": 1.6645, "step": 732 }, { "epoch": 0.7075289575289575, "grad_norm": 3.7172515392303467, "learning_rate": 1.786267823197881e-05, "loss": 1.6761, "step": 733 }, { "epoch": 0.7084942084942085, "grad_norm": 3.7993557453155518, "learning_rate": 1.785623353187111e-05, "loss": 1.5896, "step": 734 }, { "epoch": 0.7094594594594594, "grad_norm": 3.8158602714538574, "learning_rate": 1.7849780296290368e-05, "loss": 1.5756, "step": 735 }, { "epoch": 0.7104247104247104, "grad_norm": 3.8875603675842285, "learning_rate": 1.7843318532247767e-05, "loss": 1.6682, "step": 736 }, { "epoch": 0.7113899613899614, "grad_norm": 4.242062568664551, "learning_rate": 1.783684824676374e-05, "loss": 1.6656, "step": 737 }, { "epoch": 0.7123552123552124, "grad_norm": 3.6864781379699707, "learning_rate": 1.783036944686799e-05, "loss": 1.6468, "step": 738 }, { "epoch": 0.7133204633204633, "grad_norm": 4.057651042938232, "learning_rate": 1.782388213959946e-05, "loss": 1.6584, "step": 739 }, { "epoch": 0.7142857142857143, "grad_norm": 4.905232906341553, "learning_rate": 1.7817386332006346e-05, "loss": 1.6772, "step": 740 }, { "epoch": 0.7152509652509652, "grad_norm": 3.8806796073913574, "learning_rate": 1.781088203114607e-05, "loss": 1.6806, "step": 741 }, { "epoch": 0.7162162162162162, "grad_norm": 5.752893447875977, "learning_rate": 1.7804369244085283e-05, "loss": 1.6034, "step": 742 }, { "epoch": 0.7171814671814671, "grad_norm": 4.513430118560791, "learning_rate": 1.7797847977899873e-05, "loss": 1.6523, "step": 743 }, { "epoch": 0.7181467181467182, "grad_norm": 4.028271198272705, "learning_rate": 1.7791318239674914e-05, "loss": 1.6314, "step": 744 }, { "epoch": 0.7191119691119691, "grad_norm": 6.31402063369751, "learning_rate": 1.7784780036504696e-05, "loss": 1.7083, "step": 745 }, { "epoch": 0.7200772200772201, "grad_norm": 3.543903350830078, "learning_rate": 1.7778233375492716e-05, "loss": 1.6528, "step": 746 }, { "epoch": 0.721042471042471, "grad_norm": 5.082342624664307, "learning_rate": 1.777167826375164e-05, "loss": 1.679, "step": 747 }, { "epoch": 0.722007722007722, "grad_norm": 4.935264587402344, "learning_rate": 1.7765114708403332e-05, "loss": 1.6521, "step": 748 }, { "epoch": 0.722972972972973, "grad_norm": 3.5462920665740967, "learning_rate": 1.7758542716578826e-05, "loss": 1.6503, "step": 749 }, { "epoch": 0.7239382239382239, "grad_norm": 6.806006908416748, "learning_rate": 1.7751962295418314e-05, "loss": 1.6455, "step": 750 }, { "epoch": 0.724903474903475, "grad_norm": 3.723386526107788, "learning_rate": 1.7745373452071154e-05, "loss": 1.7378, "step": 751 }, { "epoch": 0.7258687258687259, "grad_norm": 4.961573600769043, "learning_rate": 1.7738776193695853e-05, "loss": 1.6893, "step": 752 }, { "epoch": 0.7268339768339769, "grad_norm": 3.9472904205322266, "learning_rate": 1.7732170527460058e-05, "loss": 1.6363, "step": 753 }, { "epoch": 0.7277992277992278, "grad_norm": 4.425893783569336, "learning_rate": 1.7725556460540553e-05, "loss": 1.6562, "step": 754 }, { "epoch": 0.7287644787644788, "grad_norm": 5.396843433380127, "learning_rate": 1.7718934000123247e-05, "loss": 1.6127, "step": 755 }, { "epoch": 0.7297297297297297, "grad_norm": 4.616225719451904, "learning_rate": 1.7712303153403167e-05, "loss": 1.6582, "step": 756 }, { "epoch": 0.7306949806949807, "grad_norm": 4.454399585723877, "learning_rate": 1.7705663927584464e-05, "loss": 1.6997, "step": 757 }, { "epoch": 0.7316602316602316, "grad_norm": 4.703960418701172, "learning_rate": 1.769901632988037e-05, "loss": 1.6392, "step": 758 }, { "epoch": 0.7326254826254827, "grad_norm": 3.2728753089904785, "learning_rate": 1.7692360367513238e-05, "loss": 1.6287, "step": 759 }, { "epoch": 0.7335907335907336, "grad_norm": 5.88960075378418, "learning_rate": 1.7685696047714486e-05, "loss": 1.6408, "step": 760 }, { "epoch": 0.7345559845559846, "grad_norm": 3.615349054336548, "learning_rate": 1.7679023377724623e-05, "loss": 1.61, "step": 761 }, { "epoch": 0.7355212355212355, "grad_norm": 4.272396087646484, "learning_rate": 1.7672342364793233e-05, "loss": 1.7013, "step": 762 }, { "epoch": 0.7364864864864865, "grad_norm": 5.247366905212402, "learning_rate": 1.7665653016178957e-05, "loss": 1.5842, "step": 763 }, { "epoch": 0.7374517374517374, "grad_norm": 3.7255361080169678, "learning_rate": 1.7658955339149497e-05, "loss": 1.6795, "step": 764 }, { "epoch": 0.7384169884169884, "grad_norm": 5.505873203277588, "learning_rate": 1.7652249340981608e-05, "loss": 1.6291, "step": 765 }, { "epoch": 0.7393822393822393, "grad_norm": 3.6880943775177, "learning_rate": 1.764553502896107e-05, "loss": 1.6307, "step": 766 }, { "epoch": 0.7403474903474904, "grad_norm": 4.854920864105225, "learning_rate": 1.7638812410382714e-05, "loss": 1.6176, "step": 767 }, { "epoch": 0.7413127413127413, "grad_norm": 4.184886455535889, "learning_rate": 1.763208149255038e-05, "loss": 1.5853, "step": 768 }, { "epoch": 0.7422779922779923, "grad_norm": 3.6255054473876953, "learning_rate": 1.7625342282776935e-05, "loss": 1.5887, "step": 769 }, { "epoch": 0.7432432432432432, "grad_norm": 4.314204692840576, "learning_rate": 1.7618594788384252e-05, "loss": 1.672, "step": 770 }, { "epoch": 0.7442084942084942, "grad_norm": 3.8578593730926514, "learning_rate": 1.7611839016703208e-05, "loss": 1.6656, "step": 771 }, { "epoch": 0.7451737451737451, "grad_norm": 3.6460721492767334, "learning_rate": 1.7605074975073667e-05, "loss": 1.589, "step": 772 }, { "epoch": 0.7461389961389961, "grad_norm": 3.87589430809021, "learning_rate": 1.759830267084448e-05, "loss": 1.5959, "step": 773 }, { "epoch": 0.747104247104247, "grad_norm": 3.7357277870178223, "learning_rate": 1.7591522111373475e-05, "loss": 1.6224, "step": 774 }, { "epoch": 0.7480694980694981, "grad_norm": 3.794887065887451, "learning_rate": 1.758473330402745e-05, "loss": 1.6611, "step": 775 }, { "epoch": 0.749034749034749, "grad_norm": 3.853482961654663, "learning_rate": 1.757793625618217e-05, "loss": 1.676, "step": 776 }, { "epoch": 0.75, "grad_norm": 4.004652500152588, "learning_rate": 1.7571130975222335e-05, "loss": 1.5552, "step": 777 }, { "epoch": 0.750965250965251, "grad_norm": 4.246206283569336, "learning_rate": 1.7564317468541608e-05, "loss": 1.6533, "step": 778 }, { "epoch": 0.7519305019305019, "grad_norm": 4.195562839508057, "learning_rate": 1.7557495743542586e-05, "loss": 1.6361, "step": 779 }, { "epoch": 0.752895752895753, "grad_norm": 4.303326606750488, "learning_rate": 1.7550665807636783e-05, "loss": 1.6806, "step": 780 }, { "epoch": 0.7538610038610039, "grad_norm": 3.555691719055176, "learning_rate": 1.7543827668244648e-05, "loss": 1.67, "step": 781 }, { "epoch": 0.7548262548262549, "grad_norm": 3.820612907409668, "learning_rate": 1.7536981332795535e-05, "loss": 1.6068, "step": 782 }, { "epoch": 0.7557915057915058, "grad_norm": 3.809476375579834, "learning_rate": 1.7530126808727703e-05, "loss": 1.6098, "step": 783 }, { "epoch": 0.7567567567567568, "grad_norm": 4.754481792449951, "learning_rate": 1.7523264103488314e-05, "loss": 1.635, "step": 784 }, { "epoch": 0.7577220077220077, "grad_norm": 3.7754509449005127, "learning_rate": 1.7516393224533406e-05, "loss": 1.6649, "step": 785 }, { "epoch": 0.7586872586872587, "grad_norm": 5.550960540771484, "learning_rate": 1.7509514179327914e-05, "loss": 1.6158, "step": 786 }, { "epoch": 0.7596525096525096, "grad_norm": 3.930851459503174, "learning_rate": 1.750262697534563e-05, "loss": 1.6389, "step": 787 }, { "epoch": 0.7606177606177607, "grad_norm": 6.380732536315918, "learning_rate": 1.7495731620069216e-05, "loss": 1.694, "step": 788 }, { "epoch": 0.7615830115830116, "grad_norm": 3.5392532348632812, "learning_rate": 1.7488828120990194e-05, "loss": 1.5979, "step": 789 }, { "epoch": 0.7625482625482626, "grad_norm": 6.182634353637695, "learning_rate": 1.748191648560893e-05, "loss": 1.6096, "step": 790 }, { "epoch": 0.7635135135135135, "grad_norm": 4.290624618530273, "learning_rate": 1.7474996721434628e-05, "loss": 1.6527, "step": 791 }, { "epoch": 0.7644787644787645, "grad_norm": 4.417576313018799, "learning_rate": 1.7468068835985326e-05, "loss": 1.5916, "step": 792 }, { "epoch": 0.7654440154440154, "grad_norm": 4.773061275482178, "learning_rate": 1.746113283678789e-05, "loss": 1.5791, "step": 793 }, { "epoch": 0.7664092664092664, "grad_norm": 3.7683379650115967, "learning_rate": 1.7454188731377988e-05, "loss": 1.601, "step": 794 }, { "epoch": 0.7673745173745173, "grad_norm": 5.138625144958496, "learning_rate": 1.744723652730011e-05, "loss": 1.5937, "step": 795 }, { "epoch": 0.7683397683397684, "grad_norm": 3.958002805709839, "learning_rate": 1.7440276232107537e-05, "loss": 1.6804, "step": 796 }, { "epoch": 0.7693050193050193, "grad_norm": 3.6182544231414795, "learning_rate": 1.7433307853362344e-05, "loss": 1.6123, "step": 797 }, { "epoch": 0.7702702702702703, "grad_norm": 4.057281970977783, "learning_rate": 1.742633139863538e-05, "loss": 1.6334, "step": 798 }, { "epoch": 0.7712355212355212, "grad_norm": 3.439039707183838, "learning_rate": 1.7419346875506282e-05, "loss": 1.5937, "step": 799 }, { "epoch": 0.7722007722007722, "grad_norm": 3.634779453277588, "learning_rate": 1.7412354291563442e-05, "loss": 1.587, "step": 800 }, { "epoch": 0.7731660231660231, "grad_norm": 3.980450391769409, "learning_rate": 1.7405353654404017e-05, "loss": 1.6383, "step": 801 }, { "epoch": 0.7741312741312741, "grad_norm": 3.7642734050750732, "learning_rate": 1.7398344971633902e-05, "loss": 1.6375, "step": 802 }, { "epoch": 0.775096525096525, "grad_norm": 3.817166566848755, "learning_rate": 1.7391328250867748e-05, "loss": 1.6397, "step": 803 }, { "epoch": 0.7760617760617761, "grad_norm": 5.095615863800049, "learning_rate": 1.738430349972893e-05, "loss": 1.6001, "step": 804 }, { "epoch": 0.777027027027027, "grad_norm": 3.6880061626434326, "learning_rate": 1.737727072584955e-05, "loss": 1.6828, "step": 805 }, { "epoch": 0.777992277992278, "grad_norm": 3.768186569213867, "learning_rate": 1.7370229936870425e-05, "loss": 1.5754, "step": 806 }, { "epoch": 0.778957528957529, "grad_norm": 4.330945014953613, "learning_rate": 1.7363181140441086e-05, "loss": 1.513, "step": 807 }, { "epoch": 0.7799227799227799, "grad_norm": 4.447457790374756, "learning_rate": 1.7356124344219757e-05, "loss": 1.6115, "step": 808 }, { "epoch": 0.7808880308880309, "grad_norm": 4.9061279296875, "learning_rate": 1.7349059555873348e-05, "loss": 1.6968, "step": 809 }, { "epoch": 0.7818532818532818, "grad_norm": 5.54481840133667, "learning_rate": 1.7341986783077466e-05, "loss": 1.6252, "step": 810 }, { "epoch": 0.7828185328185329, "grad_norm": 3.801086664199829, "learning_rate": 1.7334906033516388e-05, "loss": 1.5357, "step": 811 }, { "epoch": 0.7837837837837838, "grad_norm": 3.63675594329834, "learning_rate": 1.7327817314883055e-05, "loss": 1.5803, "step": 812 }, { "epoch": 0.7847490347490348, "grad_norm": 4.226805686950684, "learning_rate": 1.732072063487906e-05, "loss": 1.6519, "step": 813 }, { "epoch": 0.7857142857142857, "grad_norm": 4.242192268371582, "learning_rate": 1.7313616001214666e-05, "loss": 1.629, "step": 814 }, { "epoch": 0.7866795366795367, "grad_norm": 4.957813739776611, "learning_rate": 1.730650342160875e-05, "loss": 1.6566, "step": 815 }, { "epoch": 0.7876447876447876, "grad_norm": 4.036060333251953, "learning_rate": 1.7299382903788844e-05, "loss": 1.604, "step": 816 }, { "epoch": 0.7886100386100386, "grad_norm": 4.0332794189453125, "learning_rate": 1.7292254455491095e-05, "loss": 1.639, "step": 817 }, { "epoch": 0.7895752895752896, "grad_norm": 5.2410807609558105, "learning_rate": 1.7285118084460265e-05, "loss": 1.6602, "step": 818 }, { "epoch": 0.7905405405405406, "grad_norm": 3.6184473037719727, "learning_rate": 1.727797379844972e-05, "loss": 1.6424, "step": 819 }, { "epoch": 0.7915057915057915, "grad_norm": 4.824613571166992, "learning_rate": 1.7270821605221448e-05, "loss": 1.655, "step": 820 }, { "epoch": 0.7924710424710425, "grad_norm": 5.858402252197266, "learning_rate": 1.7263661512546e-05, "loss": 1.6421, "step": 821 }, { "epoch": 0.7934362934362934, "grad_norm": 3.8016343116760254, "learning_rate": 1.7256493528202523e-05, "loss": 1.5408, "step": 822 }, { "epoch": 0.7944015444015444, "grad_norm": 5.814817428588867, "learning_rate": 1.7249317659978736e-05, "loss": 1.7092, "step": 823 }, { "epoch": 0.7953667953667953, "grad_norm": 3.719472885131836, "learning_rate": 1.7242133915670923e-05, "loss": 1.6248, "step": 824 }, { "epoch": 0.7963320463320464, "grad_norm": 4.978057861328125, "learning_rate": 1.7234942303083925e-05, "loss": 1.6871, "step": 825 }, { "epoch": 0.7972972972972973, "grad_norm": 4.768524646759033, "learning_rate": 1.7227742830031132e-05, "loss": 1.6106, "step": 826 }, { "epoch": 0.7982625482625483, "grad_norm": 3.8294737339019775, "learning_rate": 1.7220535504334477e-05, "loss": 1.6351, "step": 827 }, { "epoch": 0.7992277992277992, "grad_norm": 4.684476375579834, "learning_rate": 1.7213320333824417e-05, "loss": 1.5997, "step": 828 }, { "epoch": 0.8001930501930502, "grad_norm": 3.5390968322753906, "learning_rate": 1.7206097326339937e-05, "loss": 1.6052, "step": 829 }, { "epoch": 0.8011583011583011, "grad_norm": 3.9698216915130615, "learning_rate": 1.7198866489728537e-05, "loss": 1.5969, "step": 830 }, { "epoch": 0.8021235521235521, "grad_norm": 4.555296421051025, "learning_rate": 1.7191627831846226e-05, "loss": 1.6549, "step": 831 }, { "epoch": 0.803088803088803, "grad_norm": 4.702564239501953, "learning_rate": 1.71843813605575e-05, "loss": 1.5909, "step": 832 }, { "epoch": 0.8040540540540541, "grad_norm": 3.884777307510376, "learning_rate": 1.7177127083735352e-05, "loss": 1.6226, "step": 833 }, { "epoch": 0.805019305019305, "grad_norm": 3.7954657077789307, "learning_rate": 1.716986500926126e-05, "loss": 1.6062, "step": 834 }, { "epoch": 0.805984555984556, "grad_norm": 4.036071300506592, "learning_rate": 1.716259514502516e-05, "loss": 1.6093, "step": 835 }, { "epoch": 0.806949806949807, "grad_norm": 4.345009803771973, "learning_rate": 1.715531749892546e-05, "loss": 1.5864, "step": 836 }, { "epoch": 0.8079150579150579, "grad_norm": 3.6161437034606934, "learning_rate": 1.7148032078869027e-05, "loss": 1.6097, "step": 837 }, { "epoch": 0.8088803088803089, "grad_norm": 3.721926689147949, "learning_rate": 1.714073889277117e-05, "loss": 1.6151, "step": 838 }, { "epoch": 0.8098455598455598, "grad_norm": 4.898469924926758, "learning_rate": 1.713343794855562e-05, "loss": 1.6506, "step": 839 }, { "epoch": 0.8108108108108109, "grad_norm": 4.193187236785889, "learning_rate": 1.7126129254154568e-05, "loss": 1.5663, "step": 840 }, { "epoch": 0.8117760617760618, "grad_norm": 5.0784406661987305, "learning_rate": 1.7118812817508603e-05, "loss": 1.6037, "step": 841 }, { "epoch": 0.8127413127413128, "grad_norm": 3.959920644760132, "learning_rate": 1.7111488646566728e-05, "loss": 1.6617, "step": 842 }, { "epoch": 0.8137065637065637, "grad_norm": 4.190287113189697, "learning_rate": 1.7104156749286352e-05, "loss": 1.5909, "step": 843 }, { "epoch": 0.8146718146718147, "grad_norm": 3.578152894973755, "learning_rate": 1.7096817133633283e-05, "loss": 1.6875, "step": 844 }, { "epoch": 0.8156370656370656, "grad_norm": 3.6228625774383545, "learning_rate": 1.7089469807581705e-05, "loss": 1.6502, "step": 845 }, { "epoch": 0.8166023166023166, "grad_norm": 3.968024730682373, "learning_rate": 1.708211477911419e-05, "loss": 1.562, "step": 846 }, { "epoch": 0.8175675675675675, "grad_norm": 4.099898815155029, "learning_rate": 1.7074752056221663e-05, "loss": 1.6169, "step": 847 }, { "epoch": 0.8185328185328186, "grad_norm": 4.45789098739624, "learning_rate": 1.7067381646903426e-05, "loss": 1.6571, "step": 848 }, { "epoch": 0.8194980694980695, "grad_norm": 3.8516762256622314, "learning_rate": 1.706000355916712e-05, "loss": 1.5939, "step": 849 }, { "epoch": 0.8204633204633205, "grad_norm": 3.3639023303985596, "learning_rate": 1.705261780102874e-05, "loss": 1.6121, "step": 850 }, { "epoch": 0.8214285714285714, "grad_norm": 4.166355609893799, "learning_rate": 1.7045224380512597e-05, "loss": 1.534, "step": 851 }, { "epoch": 0.8223938223938224, "grad_norm": 3.9245617389678955, "learning_rate": 1.7037823305651345e-05, "loss": 1.6817, "step": 852 }, { "epoch": 0.8233590733590733, "grad_norm": 4.970295429229736, "learning_rate": 1.7030414584485938e-05, "loss": 1.6512, "step": 853 }, { "epoch": 0.8243243243243243, "grad_norm": 3.748103380203247, "learning_rate": 1.702299822506565e-05, "loss": 1.6636, "step": 854 }, { "epoch": 0.8252895752895753, "grad_norm": 4.198969841003418, "learning_rate": 1.701557423544805e-05, "loss": 1.6339, "step": 855 }, { "epoch": 0.8262548262548263, "grad_norm": 3.630988359451294, "learning_rate": 1.700814262369899e-05, "loss": 1.6347, "step": 856 }, { "epoch": 0.8272200772200772, "grad_norm": 3.7846217155456543, "learning_rate": 1.700070339789262e-05, "loss": 1.5582, "step": 857 }, { "epoch": 0.8281853281853282, "grad_norm": 4.339964866638184, "learning_rate": 1.699325656611134e-05, "loss": 1.631, "step": 858 }, { "epoch": 0.8291505791505791, "grad_norm": 3.7031326293945312, "learning_rate": 1.6985802136445832e-05, "loss": 1.5525, "step": 859 }, { "epoch": 0.8301158301158301, "grad_norm": 4.129475116729736, "learning_rate": 1.6978340116995022e-05, "loss": 1.6242, "step": 860 }, { "epoch": 0.831081081081081, "grad_norm": 3.5283150672912598, "learning_rate": 1.6970870515866086e-05, "loss": 1.5839, "step": 861 }, { "epoch": 0.832046332046332, "grad_norm": 3.6117632389068604, "learning_rate": 1.6963393341174438e-05, "loss": 1.6413, "step": 862 }, { "epoch": 0.833011583011583, "grad_norm": 3.7649621963500977, "learning_rate": 1.6955908601043724e-05, "loss": 1.5737, "step": 863 }, { "epoch": 0.833976833976834, "grad_norm": 3.8086862564086914, "learning_rate": 1.6948416303605796e-05, "loss": 1.6382, "step": 864 }, { "epoch": 0.834942084942085, "grad_norm": 4.5638508796691895, "learning_rate": 1.694091645700073e-05, "loss": 1.5265, "step": 865 }, { "epoch": 0.8359073359073359, "grad_norm": 3.99662184715271, "learning_rate": 1.6933409069376797e-05, "loss": 1.6353, "step": 866 }, { "epoch": 0.8368725868725869, "grad_norm": 4.2375407218933105, "learning_rate": 1.6925894148890473e-05, "loss": 1.6229, "step": 867 }, { "epoch": 0.8378378378378378, "grad_norm": 3.9651808738708496, "learning_rate": 1.691837170370639e-05, "loss": 1.6518, "step": 868 }, { "epoch": 0.8388030888030888, "grad_norm": 4.371737480163574, "learning_rate": 1.6910841741997395e-05, "loss": 1.6, "step": 869 }, { "epoch": 0.8397683397683398, "grad_norm": 3.907066583633423, "learning_rate": 1.6903304271944467e-05, "loss": 1.5402, "step": 870 }, { "epoch": 0.8407335907335908, "grad_norm": 4.892452239990234, "learning_rate": 1.6895759301736757e-05, "loss": 1.6335, "step": 871 }, { "epoch": 0.8416988416988417, "grad_norm": 4.123228073120117, "learning_rate": 1.688820683957156e-05, "loss": 1.6499, "step": 872 }, { "epoch": 0.8426640926640927, "grad_norm": 4.737846374511719, "learning_rate": 1.6880646893654327e-05, "loss": 1.6568, "step": 873 }, { "epoch": 0.8436293436293436, "grad_norm": 5.440667152404785, "learning_rate": 1.6873079472198613e-05, "loss": 1.6075, "step": 874 }, { "epoch": 0.8445945945945946, "grad_norm": 4.515224456787109, "learning_rate": 1.6865504583426117e-05, "loss": 1.586, "step": 875 }, { "epoch": 0.8455598455598455, "grad_norm": 6.032504081726074, "learning_rate": 1.6857922235566642e-05, "loss": 1.6334, "step": 876 }, { "epoch": 0.8465250965250966, "grad_norm": 4.806612968444824, "learning_rate": 1.6850332436858087e-05, "loss": 1.6048, "step": 877 }, { "epoch": 0.8474903474903475, "grad_norm": 5.170681476593018, "learning_rate": 1.6842735195546466e-05, "loss": 1.6263, "step": 878 }, { "epoch": 0.8484555984555985, "grad_norm": 8.467127799987793, "learning_rate": 1.683513051988586e-05, "loss": 1.6341, "step": 879 }, { "epoch": 0.8494208494208494, "grad_norm": 4.348141670227051, "learning_rate": 1.682751841813844e-05, "loss": 1.5846, "step": 880 }, { "epoch": 0.8503861003861004, "grad_norm": 5.740167140960693, "learning_rate": 1.681989889857444e-05, "loss": 1.598, "step": 881 }, { "epoch": 0.8513513513513513, "grad_norm": 4.303435802459717, "learning_rate": 1.6812271969472145e-05, "loss": 1.626, "step": 882 }, { "epoch": 0.8523166023166023, "grad_norm": 4.012930393218994, "learning_rate": 1.680463763911791e-05, "loss": 1.6634, "step": 883 }, { "epoch": 0.8532818532818532, "grad_norm": 5.010847568511963, "learning_rate": 1.679699591580611e-05, "loss": 1.6119, "step": 884 }, { "epoch": 0.8542471042471043, "grad_norm": 5.053982734680176, "learning_rate": 1.6789346807839165e-05, "loss": 1.5977, "step": 885 }, { "epoch": 0.8552123552123552, "grad_norm": 4.631558418273926, "learning_rate": 1.6781690323527512e-05, "loss": 1.6513, "step": 886 }, { "epoch": 0.8561776061776062, "grad_norm": 4.911116600036621, "learning_rate": 1.677402647118961e-05, "loss": 1.6718, "step": 887 }, { "epoch": 0.8571428571428571, "grad_norm": 4.912961006164551, "learning_rate": 1.676635525915191e-05, "loss": 1.6031, "step": 888 }, { "epoch": 0.8581081081081081, "grad_norm": 6.79826021194458, "learning_rate": 1.675867669574888e-05, "loss": 1.6854, "step": 889 }, { "epoch": 0.859073359073359, "grad_norm": 3.9823529720306396, "learning_rate": 1.6750990789322948e-05, "loss": 1.6156, "step": 890 }, { "epoch": 0.86003861003861, "grad_norm": 3.468554735183716, "learning_rate": 1.6743297548224543e-05, "loss": 1.5787, "step": 891 }, { "epoch": 0.861003861003861, "grad_norm": 5.073884010314941, "learning_rate": 1.673559698081205e-05, "loss": 1.5174, "step": 892 }, { "epoch": 0.861969111969112, "grad_norm": 4.256384372711182, "learning_rate": 1.6727889095451814e-05, "loss": 1.6093, "step": 893 }, { "epoch": 0.862934362934363, "grad_norm": 3.446455478668213, "learning_rate": 1.6720173900518137e-05, "loss": 1.5703, "step": 894 }, { "epoch": 0.8638996138996139, "grad_norm": 6.296940803527832, "learning_rate": 1.671245140439326e-05, "loss": 1.6112, "step": 895 }, { "epoch": 0.8648648648648649, "grad_norm": 3.773432493209839, "learning_rate": 1.670472161546736e-05, "loss": 1.6364, "step": 896 }, { "epoch": 0.8658301158301158, "grad_norm": 5.01470422744751, "learning_rate": 1.669698454213852e-05, "loss": 1.5684, "step": 897 }, { "epoch": 0.8667953667953668, "grad_norm": 4.782303810119629, "learning_rate": 1.6689240192812765e-05, "loss": 1.5945, "step": 898 }, { "epoch": 0.8677606177606177, "grad_norm": 4.499415397644043, "learning_rate": 1.6681488575904003e-05, "loss": 1.6817, "step": 899 }, { "epoch": 0.8687258687258688, "grad_norm": 4.399867057800293, "learning_rate": 1.6673729699834046e-05, "loss": 1.6501, "step": 900 }, { "epoch": 0.8696911196911197, "grad_norm": 3.5721189975738525, "learning_rate": 1.6665963573032593e-05, "loss": 1.5793, "step": 901 }, { "epoch": 0.8706563706563707, "grad_norm": 4.1614274978637695, "learning_rate": 1.6658190203937218e-05, "loss": 1.6342, "step": 902 }, { "epoch": 0.8716216216216216, "grad_norm": 4.05487585067749, "learning_rate": 1.6650409600993373e-05, "loss": 1.6461, "step": 903 }, { "epoch": 0.8725868725868726, "grad_norm": 5.089929103851318, "learning_rate": 1.6642621772654347e-05, "loss": 1.5484, "step": 904 }, { "epoch": 0.8735521235521235, "grad_norm": 3.659691572189331, "learning_rate": 1.663482672738131e-05, "loss": 1.6263, "step": 905 }, { "epoch": 0.8745173745173745, "grad_norm": 5.445408344268799, "learning_rate": 1.6627024473643244e-05, "loss": 1.6315, "step": 906 }, { "epoch": 0.8754826254826255, "grad_norm": 3.7327609062194824, "learning_rate": 1.6619215019916983e-05, "loss": 1.6925, "step": 907 }, { "epoch": 0.8764478764478765, "grad_norm": 4.179985046386719, "learning_rate": 1.6611398374687172e-05, "loss": 1.5642, "step": 908 }, { "epoch": 0.8774131274131274, "grad_norm": 3.8566882610321045, "learning_rate": 1.6603574546446274e-05, "loss": 1.5556, "step": 909 }, { "epoch": 0.8783783783783784, "grad_norm": 3.8949801921844482, "learning_rate": 1.659574354369455e-05, "loss": 1.6518, "step": 910 }, { "epoch": 0.8793436293436293, "grad_norm": 3.5222058296203613, "learning_rate": 1.6587905374940067e-05, "loss": 1.6216, "step": 911 }, { "epoch": 0.8803088803088803, "grad_norm": 3.6448230743408203, "learning_rate": 1.6580060048698673e-05, "loss": 1.6862, "step": 912 }, { "epoch": 0.8812741312741312, "grad_norm": 3.809236526489258, "learning_rate": 1.6572207573493987e-05, "loss": 1.549, "step": 913 }, { "epoch": 0.8822393822393823, "grad_norm": 4.041685581207275, "learning_rate": 1.6564347957857395e-05, "loss": 1.5944, "step": 914 }, { "epoch": 0.8832046332046332, "grad_norm": 3.648710012435913, "learning_rate": 1.6556481210328053e-05, "loss": 1.6113, "step": 915 }, { "epoch": 0.8841698841698842, "grad_norm": 3.8146796226501465, "learning_rate": 1.6548607339452853e-05, "loss": 1.602, "step": 916 }, { "epoch": 0.8851351351351351, "grad_norm": 4.380678653717041, "learning_rate": 1.6540726353786428e-05, "loss": 1.6612, "step": 917 }, { "epoch": 0.8861003861003861, "grad_norm": 4.221372127532959, "learning_rate": 1.6532838261891145e-05, "loss": 1.6558, "step": 918 }, { "epoch": 0.887065637065637, "grad_norm": 5.726127624511719, "learning_rate": 1.6524943072337094e-05, "loss": 1.5827, "step": 919 }, { "epoch": 0.888030888030888, "grad_norm": 4.030428409576416, "learning_rate": 1.6517040793702066e-05, "loss": 1.5973, "step": 920 }, { "epoch": 0.888996138996139, "grad_norm": 5.082960605621338, "learning_rate": 1.6509131434571564e-05, "loss": 1.6047, "step": 921 }, { "epoch": 0.88996138996139, "grad_norm": 4.497354984283447, "learning_rate": 1.6501215003538776e-05, "loss": 1.582, "step": 922 }, { "epoch": 0.890926640926641, "grad_norm": 4.039971828460693, "learning_rate": 1.6493291509204582e-05, "loss": 1.5584, "step": 923 }, { "epoch": 0.8918918918918919, "grad_norm": 4.771897792816162, "learning_rate": 1.6485360960177525e-05, "loss": 1.567, "step": 924 }, { "epoch": 0.8928571428571429, "grad_norm": 3.8797128200531006, "learning_rate": 1.6477423365073825e-05, "loss": 1.6056, "step": 925 }, { "epoch": 0.8938223938223938, "grad_norm": 5.256024360656738, "learning_rate": 1.646947873251735e-05, "loss": 1.61, "step": 926 }, { "epoch": 0.8947876447876448, "grad_norm": 4.378505706787109, "learning_rate": 1.6461527071139613e-05, "loss": 1.5766, "step": 927 }, { "epoch": 0.8957528957528957, "grad_norm": 3.816833257675171, "learning_rate": 1.645356838957976e-05, "loss": 1.6357, "step": 928 }, { "epoch": 0.8967181467181468, "grad_norm": 3.6121904850006104, "learning_rate": 1.644560269648458e-05, "loss": 1.5561, "step": 929 }, { "epoch": 0.8976833976833977, "grad_norm": 4.310316562652588, "learning_rate": 1.6437630000508466e-05, "loss": 1.5876, "step": 930 }, { "epoch": 0.8986486486486487, "grad_norm": 3.828895092010498, "learning_rate": 1.642965031031342e-05, "loss": 1.611, "step": 931 }, { "epoch": 0.8996138996138996, "grad_norm": 3.4154212474823, "learning_rate": 1.6421663634569045e-05, "loss": 1.6266, "step": 932 }, { "epoch": 0.9005791505791506, "grad_norm": 4.432816028594971, "learning_rate": 1.641366998195254e-05, "loss": 1.5617, "step": 933 }, { "epoch": 0.9015444015444015, "grad_norm": 4.561981201171875, "learning_rate": 1.6405669361148668e-05, "loss": 1.5748, "step": 934 }, { "epoch": 0.9025096525096525, "grad_norm": 3.6001009941101074, "learning_rate": 1.6397661780849785e-05, "loss": 1.5985, "step": 935 }, { "epoch": 0.9034749034749034, "grad_norm": 3.6477224826812744, "learning_rate": 1.6389647249755788e-05, "loss": 1.5923, "step": 936 }, { "epoch": 0.9044401544401545, "grad_norm": 4.120946407318115, "learning_rate": 1.6381625776574135e-05, "loss": 1.6292, "step": 937 }, { "epoch": 0.9054054054054054, "grad_norm": 3.8856008052825928, "learning_rate": 1.6373597370019825e-05, "loss": 1.5703, "step": 938 }, { "epoch": 0.9063706563706564, "grad_norm": 4.099815368652344, "learning_rate": 1.6365562038815394e-05, "loss": 1.5469, "step": 939 }, { "epoch": 0.9073359073359073, "grad_norm": 3.5519697666168213, "learning_rate": 1.6357519791690893e-05, "loss": 1.6077, "step": 940 }, { "epoch": 0.9083011583011583, "grad_norm": 4.991375923156738, "learning_rate": 1.634947063738389e-05, "loss": 1.6249, "step": 941 }, { "epoch": 0.9092664092664092, "grad_norm": 4.817924976348877, "learning_rate": 1.6341414584639464e-05, "loss": 1.6458, "step": 942 }, { "epoch": 0.9102316602316602, "grad_norm": 4.194146156311035, "learning_rate": 1.633335164221018e-05, "loss": 1.5919, "step": 943 }, { "epoch": 0.9111969111969112, "grad_norm": 6.744943618774414, "learning_rate": 1.632528181885609e-05, "loss": 1.5598, "step": 944 }, { "epoch": 0.9121621621621622, "grad_norm": 3.4473798274993896, "learning_rate": 1.6317205123344725e-05, "loss": 1.6495, "step": 945 }, { "epoch": 0.9131274131274131, "grad_norm": 6.090734004974365, "learning_rate": 1.6309121564451084e-05, "loss": 1.611, "step": 946 }, { "epoch": 0.9140926640926641, "grad_norm": 4.007740497589111, "learning_rate": 1.6301031150957616e-05, "loss": 1.5825, "step": 947 }, { "epoch": 0.915057915057915, "grad_norm": 3.920323133468628, "learning_rate": 1.6292933891654217e-05, "loss": 1.5947, "step": 948 }, { "epoch": 0.916023166023166, "grad_norm": 5.101447582244873, "learning_rate": 1.6284829795338235e-05, "loss": 1.5809, "step": 949 }, { "epoch": 0.916988416988417, "grad_norm": 4.230952262878418, "learning_rate": 1.627671887081443e-05, "loss": 1.5773, "step": 950 }, { "epoch": 0.917953667953668, "grad_norm": 4.483916759490967, "learning_rate": 1.6268601126894986e-05, "loss": 1.6321, "step": 951 }, { "epoch": 0.918918918918919, "grad_norm": 3.5368592739105225, "learning_rate": 1.6260476572399494e-05, "loss": 1.6531, "step": 952 }, { "epoch": 0.9198841698841699, "grad_norm": 4.395543098449707, "learning_rate": 1.625234521615495e-05, "loss": 1.5416, "step": 953 }, { "epoch": 0.9208494208494209, "grad_norm": 3.7641806602478027, "learning_rate": 1.6244207066995737e-05, "loss": 1.6085, "step": 954 }, { "epoch": 0.9218146718146718, "grad_norm": 4.261475086212158, "learning_rate": 1.6236062133763617e-05, "loss": 1.5874, "step": 955 }, { "epoch": 0.9227799227799228, "grad_norm": 5.187191009521484, "learning_rate": 1.6227910425307723e-05, "loss": 1.5936, "step": 956 }, { "epoch": 0.9237451737451737, "grad_norm": 3.669710159301758, "learning_rate": 1.6219751950484547e-05, "loss": 1.549, "step": 957 }, { "epoch": 0.9247104247104247, "grad_norm": 5.148052215576172, "learning_rate": 1.6211586718157935e-05, "loss": 1.5487, "step": 958 }, { "epoch": 0.9256756756756757, "grad_norm": 3.8471758365631104, "learning_rate": 1.6203414737199083e-05, "loss": 1.601, "step": 959 }, { "epoch": 0.9266409266409267, "grad_norm": 4.013639450073242, "learning_rate": 1.61952360164865e-05, "loss": 1.6314, "step": 960 }, { "epoch": 0.9276061776061776, "grad_norm": 3.7651102542877197, "learning_rate": 1.618705056490603e-05, "loss": 1.5831, "step": 961 }, { "epoch": 0.9285714285714286, "grad_norm": 4.085816860198975, "learning_rate": 1.6178858391350835e-05, "loss": 1.6087, "step": 962 }, { "epoch": 0.9295366795366795, "grad_norm": 3.8041088581085205, "learning_rate": 1.6170659504721365e-05, "loss": 1.6063, "step": 963 }, { "epoch": 0.9305019305019305, "grad_norm": 5.450389385223389, "learning_rate": 1.6162453913925373e-05, "loss": 1.5322, "step": 964 }, { "epoch": 0.9314671814671814, "grad_norm": 4.215041637420654, "learning_rate": 1.6154241627877894e-05, "loss": 1.6165, "step": 965 }, { "epoch": 0.9324324324324325, "grad_norm": 3.923663377761841, "learning_rate": 1.614602265550124e-05, "loss": 1.5717, "step": 966 }, { "epoch": 0.9333976833976834, "grad_norm": 3.8256661891937256, "learning_rate": 1.613779700572499e-05, "loss": 1.6302, "step": 967 }, { "epoch": 0.9343629343629344, "grad_norm": 4.588598251342773, "learning_rate": 1.612956468748596e-05, "loss": 1.6229, "step": 968 }, { "epoch": 0.9353281853281853, "grad_norm": 4.931879043579102, "learning_rate": 1.6121325709728228e-05, "loss": 1.5349, "step": 969 }, { "epoch": 0.9362934362934363, "grad_norm": 3.869826555252075, "learning_rate": 1.611308008140311e-05, "loss": 1.6005, "step": 970 }, { "epoch": 0.9372586872586872, "grad_norm": 4.035933971405029, "learning_rate": 1.610482781146913e-05, "loss": 1.6232, "step": 971 }, { "epoch": 0.9382239382239382, "grad_norm": 3.962907075881958, "learning_rate": 1.6096568908892048e-05, "loss": 1.6566, "step": 972 }, { "epoch": 0.9391891891891891, "grad_norm": 3.8459584712982178, "learning_rate": 1.608830338264481e-05, "loss": 1.5392, "step": 973 }, { "epoch": 0.9401544401544402, "grad_norm": 3.968818187713623, "learning_rate": 1.608003124170758e-05, "loss": 1.5504, "step": 974 }, { "epoch": 0.9411196911196911, "grad_norm": 3.915367364883423, "learning_rate": 1.6071752495067684e-05, "loss": 1.57, "step": 975 }, { "epoch": 0.9420849420849421, "grad_norm": 3.6851232051849365, "learning_rate": 1.6063467151719653e-05, "loss": 1.569, "step": 976 }, { "epoch": 0.943050193050193, "grad_norm": 4.471630096435547, "learning_rate": 1.6055175220665158e-05, "loss": 1.6378, "step": 977 }, { "epoch": 0.944015444015444, "grad_norm": 4.199156761169434, "learning_rate": 1.6046876710913043e-05, "loss": 1.5974, "step": 978 }, { "epoch": 0.944980694980695, "grad_norm": 5.314173221588135, "learning_rate": 1.60385716314793e-05, "loss": 1.5, "step": 979 }, { "epoch": 0.9459459459459459, "grad_norm": 4.639838695526123, "learning_rate": 1.603025999138705e-05, "loss": 1.601, "step": 980 }, { "epoch": 0.946911196911197, "grad_norm": 5.4433159828186035, "learning_rate": 1.602194179966655e-05, "loss": 1.5841, "step": 981 }, { "epoch": 0.9478764478764479, "grad_norm": 5.501407623291016, "learning_rate": 1.6013617065355167e-05, "loss": 1.6584, "step": 982 }, { "epoch": 0.9488416988416989, "grad_norm": 3.7589755058288574, "learning_rate": 1.6005285797497383e-05, "loss": 1.5948, "step": 983 }, { "epoch": 0.9498069498069498, "grad_norm": 3.967992067337036, "learning_rate": 1.5996948005144774e-05, "loss": 1.6061, "step": 984 }, { "epoch": 0.9507722007722008, "grad_norm": 5.262543678283691, "learning_rate": 1.598860369735601e-05, "loss": 1.5848, "step": 985 }, { "epoch": 0.9517374517374517, "grad_norm": 3.765303611755371, "learning_rate": 1.598025288319684e-05, "loss": 1.5626, "step": 986 }, { "epoch": 0.9527027027027027, "grad_norm": 4.409090995788574, "learning_rate": 1.5971895571740072e-05, "loss": 1.6094, "step": 987 }, { "epoch": 0.9536679536679536, "grad_norm": 5.229574203491211, "learning_rate": 1.596353177206558e-05, "loss": 1.6579, "step": 988 }, { "epoch": 0.9546332046332047, "grad_norm": 4.5636887550354, "learning_rate": 1.5955161493260288e-05, "loss": 1.6182, "step": 989 }, { "epoch": 0.9555984555984556, "grad_norm": 3.9794015884399414, "learning_rate": 1.5946784744418162e-05, "loss": 1.612, "step": 990 }, { "epoch": 0.9565637065637066, "grad_norm": 3.775864601135254, "learning_rate": 1.5938401534640185e-05, "loss": 1.6424, "step": 991 }, { "epoch": 0.9575289575289575, "grad_norm": 3.8312482833862305, "learning_rate": 1.5930011873034377e-05, "loss": 1.6274, "step": 992 }, { "epoch": 0.9584942084942085, "grad_norm": 3.593261480331421, "learning_rate": 1.592161576871575e-05, "loss": 1.6042, "step": 993 }, { "epoch": 0.9594594594594594, "grad_norm": 3.6945395469665527, "learning_rate": 1.591321323080633e-05, "loss": 1.5665, "step": 994 }, { "epoch": 0.9604247104247104, "grad_norm": 4.436593532562256, "learning_rate": 1.5904804268435127e-05, "loss": 1.6183, "step": 995 }, { "epoch": 0.9613899613899614, "grad_norm": 5.163026332855225, "learning_rate": 1.589638889073813e-05, "loss": 1.4779, "step": 996 }, { "epoch": 0.9623552123552124, "grad_norm": 3.6680943965911865, "learning_rate": 1.5887967106858296e-05, "loss": 1.5125, "step": 997 }, { "epoch": 0.9633204633204633, "grad_norm": 3.6280436515808105, "learning_rate": 1.587953892594555e-05, "loss": 1.5655, "step": 998 }, { "epoch": 0.9642857142857143, "grad_norm": 3.6882362365722656, "learning_rate": 1.587110435715676e-05, "loss": 1.6317, "step": 999 }, { "epoch": 0.9652509652509652, "grad_norm": 4.365917205810547, "learning_rate": 1.5862663409655733e-05, "loss": 1.5443, "step": 1000 }, { "epoch": 0.9662162162162162, "grad_norm": 3.6691746711730957, "learning_rate": 1.585421609261322e-05, "loss": 1.6013, "step": 1001 }, { "epoch": 0.9671814671814671, "grad_norm": 4.16331672668457, "learning_rate": 1.584576241520687e-05, "loss": 1.4935, "step": 1002 }, { "epoch": 0.9681467181467182, "grad_norm": 4.23085355758667, "learning_rate": 1.583730238662126e-05, "loss": 1.6189, "step": 1003 }, { "epoch": 0.9691119691119691, "grad_norm": 4.181711673736572, "learning_rate": 1.582883601604786e-05, "loss": 1.6259, "step": 1004 }, { "epoch": 0.9700772200772201, "grad_norm": 4.299831867218018, "learning_rate": 1.5820363312685032e-05, "loss": 1.5885, "step": 1005 }, { "epoch": 0.971042471042471, "grad_norm": 3.6799392700195312, "learning_rate": 1.581188428573802e-05, "loss": 1.608, "step": 1006 }, { "epoch": 0.972007722007722, "grad_norm": 4.73622989654541, "learning_rate": 1.5803398944418934e-05, "loss": 1.5512, "step": 1007 }, { "epoch": 0.972972972972973, "grad_norm": 3.8058416843414307, "learning_rate": 1.579490729794675e-05, "loss": 1.5421, "step": 1008 }, { "epoch": 0.9739382239382239, "grad_norm": 3.8185997009277344, "learning_rate": 1.578640935554728e-05, "loss": 1.7025, "step": 1009 }, { "epoch": 0.974903474903475, "grad_norm": 3.7489326000213623, "learning_rate": 1.5777905126453204e-05, "loss": 1.5785, "step": 1010 }, { "epoch": 0.9758687258687259, "grad_norm": 4.161064147949219, "learning_rate": 1.5769394619904002e-05, "loss": 1.5742, "step": 1011 }, { "epoch": 0.9768339768339769, "grad_norm": 4.197792053222656, "learning_rate": 1.5760877845145996e-05, "loss": 1.5541, "step": 1012 }, { "epoch": 0.9777992277992278, "grad_norm": 5.174560546875, "learning_rate": 1.57523548114323e-05, "loss": 1.5785, "step": 1013 }, { "epoch": 0.9787644787644788, "grad_norm": 6.082184314727783, "learning_rate": 1.5743825528022848e-05, "loss": 1.616, "step": 1014 }, { "epoch": 0.9797297297297297, "grad_norm": 4.4992899894714355, "learning_rate": 1.5735290004184354e-05, "loss": 1.5737, "step": 1015 }, { "epoch": 0.9806949806949807, "grad_norm": 8.02248764038086, "learning_rate": 1.57267482491903e-05, "loss": 1.6128, "step": 1016 }, { "epoch": 0.9816602316602316, "grad_norm": 3.4630367755889893, "learning_rate": 1.5718200272320965e-05, "loss": 1.5705, "step": 1017 }, { "epoch": 0.9826254826254827, "grad_norm": 5.955785751342773, "learning_rate": 1.570964608286336e-05, "loss": 1.5853, "step": 1018 }, { "epoch": 0.9835907335907336, "grad_norm": 5.194217681884766, "learning_rate": 1.5701085690111268e-05, "loss": 1.6157, "step": 1019 }, { "epoch": 0.9845559845559846, "grad_norm": 3.750546932220459, "learning_rate": 1.5692519103365194e-05, "loss": 1.5675, "step": 1020 }, { "epoch": 0.9855212355212355, "grad_norm": 5.692684173583984, "learning_rate": 1.5683946331932385e-05, "loss": 1.6369, "step": 1021 }, { "epoch": 0.9864864864864865, "grad_norm": 4.506756782531738, "learning_rate": 1.56753673851268e-05, "loss": 1.5216, "step": 1022 }, { "epoch": 0.9874517374517374, "grad_norm": 4.121953964233398, "learning_rate": 1.5666782272269113e-05, "loss": 1.6066, "step": 1023 }, { "epoch": 0.9884169884169884, "grad_norm": 4.941500186920166, "learning_rate": 1.5658191002686694e-05, "loss": 1.6255, "step": 1024 }, { "epoch": 0.9893822393822393, "grad_norm": 3.950207471847534, "learning_rate": 1.5649593585713604e-05, "loss": 1.5147, "step": 1025 }, { "epoch": 0.9903474903474904, "grad_norm": 4.301034927368164, "learning_rate": 1.5640990030690576e-05, "loss": 1.5777, "step": 1026 }, { "epoch": 0.9913127413127413, "grad_norm": 4.856922149658203, "learning_rate": 1.563238034696502e-05, "loss": 1.536, "step": 1027 }, { "epoch": 0.9922779922779923, "grad_norm": 3.9627559185028076, "learning_rate": 1.5623764543891e-05, "loss": 1.5217, "step": 1028 }, { "epoch": 0.9932432432432432, "grad_norm": 4.766280174255371, "learning_rate": 1.561514263082923e-05, "loss": 1.558, "step": 1029 }, { "epoch": 0.9942084942084942, "grad_norm": 4.64534330368042, "learning_rate": 1.560651461714707e-05, "loss": 1.6071, "step": 1030 }, { "epoch": 0.9951737451737451, "grad_norm": 3.7165379524230957, "learning_rate": 1.5597880512218486e-05, "loss": 1.5449, "step": 1031 }, { "epoch": 0.9961389961389961, "grad_norm": 4.039422512054443, "learning_rate": 1.5589240325424088e-05, "loss": 1.6113, "step": 1032 }, { "epoch": 0.997104247104247, "grad_norm": 4.411838531494141, "learning_rate": 1.5580594066151076e-05, "loss": 1.5712, "step": 1033 }, { "epoch": 0.9980694980694981, "grad_norm": 3.8884987831115723, "learning_rate": 1.5571941743793256e-05, "loss": 1.5246, "step": 1034 }, { "epoch": 0.999034749034749, "grad_norm": 3.771638870239258, "learning_rate": 1.556328336775102e-05, "loss": 1.5772, "step": 1035 }, { "epoch": 1.0, "grad_norm": 3.959677219390869, "learning_rate": 1.5554618947431326e-05, "loss": 1.5756, "step": 1036 }, { "epoch": 1.000965250965251, "grad_norm": 3.3843886852264404, "learning_rate": 1.5545948492247723e-05, "loss": 1.2667, "step": 1037 }, { "epoch": 1.001930501930502, "grad_norm": 3.923689603805542, "learning_rate": 1.5537272011620294e-05, "loss": 1.2317, "step": 1038 }, { "epoch": 1.002895752895753, "grad_norm": 3.9356627464294434, "learning_rate": 1.5528589514975675e-05, "loss": 1.1738, "step": 1039 }, { "epoch": 1.0038610038610039, "grad_norm": 3.580815553665161, "learning_rate": 1.5519901011747046e-05, "loss": 1.2487, "step": 1040 }, { "epoch": 1.0048262548262548, "grad_norm": 4.160267353057861, "learning_rate": 1.55112065113741e-05, "loss": 1.1911, "step": 1041 }, { "epoch": 1.005791505791506, "grad_norm": 4.202048301696777, "learning_rate": 1.5502506023303058e-05, "loss": 1.2276, "step": 1042 }, { "epoch": 1.0067567567567568, "grad_norm": 5.078566074371338, "learning_rate": 1.5493799556986634e-05, "loss": 1.262, "step": 1043 }, { "epoch": 1.0077220077220077, "grad_norm": 4.21602201461792, "learning_rate": 1.5485087121884056e-05, "loss": 1.19, "step": 1044 }, { "epoch": 1.0086872586872586, "grad_norm": 4.7379608154296875, "learning_rate": 1.547636872746101e-05, "loss": 1.1399, "step": 1045 }, { "epoch": 1.0096525096525097, "grad_norm": 4.72409200668335, "learning_rate": 1.546764438318968e-05, "loss": 1.1955, "step": 1046 }, { "epoch": 1.0106177606177607, "grad_norm": 3.7184345722198486, "learning_rate": 1.5458914098548702e-05, "loss": 1.1271, "step": 1047 }, { "epoch": 1.0115830115830116, "grad_norm": 5.027003765106201, "learning_rate": 1.5450177883023174e-05, "loss": 1.21, "step": 1048 }, { "epoch": 1.0125482625482625, "grad_norm": 5.185225009918213, "learning_rate": 1.5441435746104628e-05, "loss": 1.2206, "step": 1049 }, { "epoch": 1.0135135135135136, "grad_norm": 3.985189437866211, "learning_rate": 1.5432687697291037e-05, "loss": 1.1847, "step": 1050 }, { "epoch": 1.0144787644787645, "grad_norm": 6.368143081665039, "learning_rate": 1.5423933746086793e-05, "loss": 1.138, "step": 1051 }, { "epoch": 1.0154440154440154, "grad_norm": 3.65655517578125, "learning_rate": 1.5415173902002703e-05, "loss": 1.1519, "step": 1052 }, { "epoch": 1.0164092664092663, "grad_norm": 4.059977054595947, "learning_rate": 1.5406408174555978e-05, "loss": 1.2186, "step": 1053 }, { "epoch": 1.0173745173745175, "grad_norm": 5.151540279388428, "learning_rate": 1.5397636573270213e-05, "loss": 1.1636, "step": 1054 }, { "epoch": 1.0183397683397684, "grad_norm": 4.020742893218994, "learning_rate": 1.5388859107675396e-05, "loss": 1.1955, "step": 1055 }, { "epoch": 1.0193050193050193, "grad_norm": 4.228591442108154, "learning_rate": 1.5380075787307875e-05, "loss": 1.1862, "step": 1056 }, { "epoch": 1.0202702702702702, "grad_norm": 5.486896991729736, "learning_rate": 1.5371286621710374e-05, "loss": 1.1797, "step": 1057 }, { "epoch": 1.0212355212355213, "grad_norm": 4.261332035064697, "learning_rate": 1.536249162043194e-05, "loss": 1.2069, "step": 1058 }, { "epoch": 1.0222007722007722, "grad_norm": 5.16762113571167, "learning_rate": 1.5353690793028e-05, "loss": 1.2037, "step": 1059 }, { "epoch": 1.0231660231660231, "grad_norm": 5.293707370758057, "learning_rate": 1.534488414906028e-05, "loss": 1.1758, "step": 1060 }, { "epoch": 1.024131274131274, "grad_norm": 4.2502336502075195, "learning_rate": 1.533607169809683e-05, "loss": 1.1617, "step": 1061 }, { "epoch": 1.0250965250965252, "grad_norm": 4.9595441818237305, "learning_rate": 1.532725344971202e-05, "loss": 1.1538, "step": 1062 }, { "epoch": 1.026061776061776, "grad_norm": 4.1361541748046875, "learning_rate": 1.5318429413486507e-05, "loss": 1.0488, "step": 1063 }, { "epoch": 1.027027027027027, "grad_norm": 3.9857394695281982, "learning_rate": 1.530959959900725e-05, "loss": 1.103, "step": 1064 }, { "epoch": 1.0279922779922779, "grad_norm": 3.9534642696380615, "learning_rate": 1.5300764015867467e-05, "loss": 1.1534, "step": 1065 }, { "epoch": 1.028957528957529, "grad_norm": 5.401212692260742, "learning_rate": 1.5291922673666664e-05, "loss": 1.2082, "step": 1066 }, { "epoch": 1.02992277992278, "grad_norm": 3.9076457023620605, "learning_rate": 1.5283075582010586e-05, "loss": 1.1735, "step": 1067 }, { "epoch": 1.0308880308880308, "grad_norm": 4.63938045501709, "learning_rate": 1.527422275051124e-05, "loss": 1.1487, "step": 1068 }, { "epoch": 1.031853281853282, "grad_norm": 4.2600836753845215, "learning_rate": 1.5265364188786852e-05, "loss": 1.205, "step": 1069 }, { "epoch": 1.0328185328185329, "grad_norm": 4.081650733947754, "learning_rate": 1.5256499906461893e-05, "loss": 1.1628, "step": 1070 }, { "epoch": 1.0337837837837838, "grad_norm": 4.65988826751709, "learning_rate": 1.5247629913167034e-05, "loss": 1.1916, "step": 1071 }, { "epoch": 1.0347490347490347, "grad_norm": 4.270849227905273, "learning_rate": 1.5238754218539155e-05, "loss": 1.2091, "step": 1072 }, { "epoch": 1.0357142857142858, "grad_norm": 4.084417343139648, "learning_rate": 1.5229872832221336e-05, "loss": 1.1886, "step": 1073 }, { "epoch": 1.0366795366795367, "grad_norm": 4.581521034240723, "learning_rate": 1.522098576386283e-05, "loss": 1.1868, "step": 1074 }, { "epoch": 1.0376447876447876, "grad_norm": 5.107059955596924, "learning_rate": 1.5212093023119075e-05, "loss": 1.1687, "step": 1075 }, { "epoch": 1.0386100386100385, "grad_norm": 4.153777599334717, "learning_rate": 1.5203194619651664e-05, "loss": 1.0946, "step": 1076 }, { "epoch": 1.0395752895752897, "grad_norm": 4.025907516479492, "learning_rate": 1.5194290563128346e-05, "loss": 1.1327, "step": 1077 }, { "epoch": 1.0405405405405406, "grad_norm": 4.304230213165283, "learning_rate": 1.5185380863223007e-05, "loss": 1.2248, "step": 1078 }, { "epoch": 1.0415057915057915, "grad_norm": 4.770325183868408, "learning_rate": 1.5176465529615672e-05, "loss": 1.2188, "step": 1079 }, { "epoch": 1.0424710424710424, "grad_norm": 4.624781608581543, "learning_rate": 1.5167544571992478e-05, "loss": 1.2264, "step": 1080 }, { "epoch": 1.0434362934362935, "grad_norm": 4.707468032836914, "learning_rate": 1.5158618000045679e-05, "loss": 1.2513, "step": 1081 }, { "epoch": 1.0444015444015444, "grad_norm": 4.72183895111084, "learning_rate": 1.5149685823473633e-05, "loss": 1.2276, "step": 1082 }, { "epoch": 1.0453667953667953, "grad_norm": 4.330811023712158, "learning_rate": 1.5140748051980767e-05, "loss": 1.2213, "step": 1083 }, { "epoch": 1.0463320463320462, "grad_norm": 4.455382823944092, "learning_rate": 1.5131804695277612e-05, "loss": 1.1809, "step": 1084 }, { "epoch": 1.0472972972972974, "grad_norm": 3.680363178253174, "learning_rate": 1.5122855763080748e-05, "loss": 1.2191, "step": 1085 }, { "epoch": 1.0482625482625483, "grad_norm": 3.820376396179199, "learning_rate": 1.5113901265112829e-05, "loss": 1.1232, "step": 1086 }, { "epoch": 1.0492277992277992, "grad_norm": 4.394763469696045, "learning_rate": 1.5104941211102536e-05, "loss": 1.26, "step": 1087 }, { "epoch": 1.05019305019305, "grad_norm": 4.531584739685059, "learning_rate": 1.5095975610784603e-05, "loss": 1.2006, "step": 1088 }, { "epoch": 1.0511583011583012, "grad_norm": 3.9240989685058594, "learning_rate": 1.5087004473899783e-05, "loss": 1.2255, "step": 1089 }, { "epoch": 1.0521235521235521, "grad_norm": 4.53143835067749, "learning_rate": 1.5078027810194847e-05, "loss": 1.2436, "step": 1090 }, { "epoch": 1.053088803088803, "grad_norm": 4.420854091644287, "learning_rate": 1.5069045629422569e-05, "loss": 1.1336, "step": 1091 }, { "epoch": 1.054054054054054, "grad_norm": 4.524373531341553, "learning_rate": 1.5060057941341717e-05, "loss": 1.2893, "step": 1092 }, { "epoch": 1.055019305019305, "grad_norm": 4.0564188957214355, "learning_rate": 1.5051064755717041e-05, "loss": 1.0964, "step": 1093 }, { "epoch": 1.055984555984556, "grad_norm": 4.545648574829102, "learning_rate": 1.5042066082319266e-05, "loss": 1.2065, "step": 1094 }, { "epoch": 1.056949806949807, "grad_norm": 3.6719343662261963, "learning_rate": 1.5033061930925081e-05, "loss": 1.1443, "step": 1095 }, { "epoch": 1.057915057915058, "grad_norm": 5.457347393035889, "learning_rate": 1.5024052311317122e-05, "loss": 1.1244, "step": 1096 }, { "epoch": 1.058880308880309, "grad_norm": 4.415368556976318, "learning_rate": 1.5015037233283969e-05, "loss": 1.2113, "step": 1097 }, { "epoch": 1.0598455598455598, "grad_norm": 4.450148582458496, "learning_rate": 1.5006016706620132e-05, "loss": 1.2133, "step": 1098 }, { "epoch": 1.0608108108108107, "grad_norm": 4.177610397338867, "learning_rate": 1.499699074112604e-05, "loss": 1.2148, "step": 1099 }, { "epoch": 1.0617760617760619, "grad_norm": 4.037389755249023, "learning_rate": 1.4987959346608029e-05, "loss": 1.1619, "step": 1100 }, { "epoch": 1.0627413127413128, "grad_norm": 4.289162635803223, "learning_rate": 1.497892253287834e-05, "loss": 1.1953, "step": 1101 }, { "epoch": 1.0637065637065637, "grad_norm": 4.018803119659424, "learning_rate": 1.4969880309755098e-05, "loss": 1.1564, "step": 1102 }, { "epoch": 1.0646718146718146, "grad_norm": 4.385936737060547, "learning_rate": 1.4960832687062303e-05, "loss": 1.1924, "step": 1103 }, { "epoch": 1.0656370656370657, "grad_norm": 3.9476888179779053, "learning_rate": 1.495177967462982e-05, "loss": 1.1805, "step": 1104 }, { "epoch": 1.0666023166023166, "grad_norm": 3.9989492893218994, "learning_rate": 1.4942721282293378e-05, "loss": 1.1573, "step": 1105 }, { "epoch": 1.0675675675675675, "grad_norm": 3.8703742027282715, "learning_rate": 1.4933657519894542e-05, "loss": 1.15, "step": 1106 }, { "epoch": 1.0685328185328185, "grad_norm": 4.762238025665283, "learning_rate": 1.4924588397280717e-05, "loss": 1.2102, "step": 1107 }, { "epoch": 1.0694980694980696, "grad_norm": 4.119187355041504, "learning_rate": 1.4915513924305132e-05, "loss": 1.1715, "step": 1108 }, { "epoch": 1.0704633204633205, "grad_norm": 4.430844306945801, "learning_rate": 1.4906434110826822e-05, "loss": 1.1914, "step": 1109 }, { "epoch": 1.0714285714285714, "grad_norm": 4.662785053253174, "learning_rate": 1.4897348966710632e-05, "loss": 1.237, "step": 1110 }, { "epoch": 1.0723938223938223, "grad_norm": 4.0754547119140625, "learning_rate": 1.4888258501827197e-05, "loss": 1.1837, "step": 1111 }, { "epoch": 1.0733590733590734, "grad_norm": 4.096596717834473, "learning_rate": 1.4879162726052926e-05, "loss": 1.1665, "step": 1112 }, { "epoch": 1.0743243243243243, "grad_norm": 4.2532782554626465, "learning_rate": 1.4870061649270014e-05, "loss": 1.1448, "step": 1113 }, { "epoch": 1.0752895752895753, "grad_norm": 4.306217193603516, "learning_rate": 1.4860955281366392e-05, "loss": 1.1413, "step": 1114 }, { "epoch": 1.0762548262548262, "grad_norm": 4.049156665802002, "learning_rate": 1.4851843632235761e-05, "loss": 1.1803, "step": 1115 }, { "epoch": 1.0772200772200773, "grad_norm": 4.349161624908447, "learning_rate": 1.4842726711777546e-05, "loss": 1.202, "step": 1116 }, { "epoch": 1.0781853281853282, "grad_norm": 4.6802778244018555, "learning_rate": 1.483360452989691e-05, "loss": 1.2489, "step": 1117 }, { "epoch": 1.079150579150579, "grad_norm": 4.26700496673584, "learning_rate": 1.482447709650472e-05, "loss": 1.1868, "step": 1118 }, { "epoch": 1.08011583011583, "grad_norm": 4.153388500213623, "learning_rate": 1.4815344421517561e-05, "loss": 1.1798, "step": 1119 }, { "epoch": 1.0810810810810811, "grad_norm": 4.498805046081543, "learning_rate": 1.4806206514857703e-05, "loss": 1.2535, "step": 1120 }, { "epoch": 1.082046332046332, "grad_norm": 4.3766655921936035, "learning_rate": 1.4797063386453107e-05, "loss": 1.2405, "step": 1121 }, { "epoch": 1.083011583011583, "grad_norm": 4.181150913238525, "learning_rate": 1.4787915046237401e-05, "loss": 1.2012, "step": 1122 }, { "epoch": 1.083976833976834, "grad_norm": 4.513302326202393, "learning_rate": 1.4778761504149882e-05, "loss": 1.2228, "step": 1123 }, { "epoch": 1.084942084942085, "grad_norm": 3.881986618041992, "learning_rate": 1.4769602770135492e-05, "loss": 1.2317, "step": 1124 }, { "epoch": 1.085907335907336, "grad_norm": 4.63210916519165, "learning_rate": 1.4760438854144816e-05, "loss": 1.2549, "step": 1125 }, { "epoch": 1.0868725868725868, "grad_norm": 4.1304473876953125, "learning_rate": 1.4751269766134075e-05, "loss": 1.196, "step": 1126 }, { "epoch": 1.087837837837838, "grad_norm": 3.830836534500122, "learning_rate": 1.47420955160651e-05, "loss": 1.2011, "step": 1127 }, { "epoch": 1.0888030888030888, "grad_norm": 4.23577880859375, "learning_rate": 1.4732916113905336e-05, "loss": 1.1752, "step": 1128 }, { "epoch": 1.0897683397683398, "grad_norm": 4.022891521453857, "learning_rate": 1.4723731569627827e-05, "loss": 1.1643, "step": 1129 }, { "epoch": 1.0907335907335907, "grad_norm": 4.168032169342041, "learning_rate": 1.4714541893211195e-05, "loss": 1.1538, "step": 1130 }, { "epoch": 1.0916988416988418, "grad_norm": 4.2090044021606445, "learning_rate": 1.4705347094639643e-05, "loss": 1.2151, "step": 1131 }, { "epoch": 1.0926640926640927, "grad_norm": 4.411066055297852, "learning_rate": 1.4696147183902949e-05, "loss": 1.1558, "step": 1132 }, { "epoch": 1.0936293436293436, "grad_norm": 4.503549575805664, "learning_rate": 1.4686942170996426e-05, "loss": 1.1777, "step": 1133 }, { "epoch": 1.0945945945945945, "grad_norm": 4.157337188720703, "learning_rate": 1.4677732065920945e-05, "loss": 1.213, "step": 1134 }, { "epoch": 1.0955598455598456, "grad_norm": 4.277044773101807, "learning_rate": 1.4668516878682902e-05, "loss": 1.2097, "step": 1135 }, { "epoch": 1.0965250965250966, "grad_norm": 4.421609878540039, "learning_rate": 1.465929661929422e-05, "loss": 1.19, "step": 1136 }, { "epoch": 1.0974903474903475, "grad_norm": 4.0304436683654785, "learning_rate": 1.4650071297772326e-05, "loss": 1.2228, "step": 1137 }, { "epoch": 1.0984555984555984, "grad_norm": 4.247618198394775, "learning_rate": 1.4640840924140155e-05, "loss": 1.1725, "step": 1138 }, { "epoch": 1.0994208494208495, "grad_norm": 4.0964813232421875, "learning_rate": 1.4631605508426124e-05, "loss": 1.1491, "step": 1139 }, { "epoch": 1.1003861003861004, "grad_norm": 4.424109935760498, "learning_rate": 1.4622365060664132e-05, "loss": 1.1315, "step": 1140 }, { "epoch": 1.1013513513513513, "grad_norm": 3.909346342086792, "learning_rate": 1.4613119590893536e-05, "loss": 1.1786, "step": 1141 }, { "epoch": 1.1023166023166022, "grad_norm": 4.109957218170166, "learning_rate": 1.4603869109159172e-05, "loss": 1.2014, "step": 1142 }, { "epoch": 1.1032818532818534, "grad_norm": 4.767273902893066, "learning_rate": 1.4594613625511288e-05, "loss": 1.2175, "step": 1143 }, { "epoch": 1.1042471042471043, "grad_norm": 4.3955559730529785, "learning_rate": 1.45853531500056e-05, "loss": 1.1938, "step": 1144 }, { "epoch": 1.1052123552123552, "grad_norm": 4.092763900756836, "learning_rate": 1.4576087692703223e-05, "loss": 1.1766, "step": 1145 }, { "epoch": 1.106177606177606, "grad_norm": 6.0021538734436035, "learning_rate": 1.4566817263670698e-05, "loss": 1.1753, "step": 1146 }, { "epoch": 1.1071428571428572, "grad_norm": 4.584158420562744, "learning_rate": 1.455754187297996e-05, "loss": 1.2373, "step": 1147 }, { "epoch": 1.1081081081081081, "grad_norm": 4.043362140655518, "learning_rate": 1.4548261530708336e-05, "loss": 1.1766, "step": 1148 }, { "epoch": 1.109073359073359, "grad_norm": 4.277347564697266, "learning_rate": 1.4538976246938541e-05, "loss": 1.1801, "step": 1149 }, { "epoch": 1.1100386100386102, "grad_norm": 4.641721248626709, "learning_rate": 1.4529686031758642e-05, "loss": 1.2534, "step": 1150 }, { "epoch": 1.111003861003861, "grad_norm": 4.423067092895508, "learning_rate": 1.4520390895262089e-05, "loss": 1.1882, "step": 1151 }, { "epoch": 1.111969111969112, "grad_norm": 4.20358943939209, "learning_rate": 1.4511090847547643e-05, "loss": 1.1374, "step": 1152 }, { "epoch": 1.1129343629343629, "grad_norm": 4.543899059295654, "learning_rate": 1.4501785898719443e-05, "loss": 1.1742, "step": 1153 }, { "epoch": 1.1138996138996138, "grad_norm": 4.224864482879639, "learning_rate": 1.4492476058886915e-05, "loss": 1.1377, "step": 1154 }, { "epoch": 1.114864864864865, "grad_norm": 4.120309352874756, "learning_rate": 1.4483161338164827e-05, "loss": 1.1776, "step": 1155 }, { "epoch": 1.1158301158301158, "grad_norm": 4.584604263305664, "learning_rate": 1.4473841746673225e-05, "loss": 1.1792, "step": 1156 }, { "epoch": 1.1167953667953667, "grad_norm": 4.499202728271484, "learning_rate": 1.4464517294537472e-05, "loss": 1.1798, "step": 1157 }, { "epoch": 1.1177606177606179, "grad_norm": 4.687706470489502, "learning_rate": 1.4455187991888195e-05, "loss": 1.1704, "step": 1158 }, { "epoch": 1.1187258687258688, "grad_norm": 5.001594543457031, "learning_rate": 1.4445853848861295e-05, "loss": 1.2071, "step": 1159 }, { "epoch": 1.1196911196911197, "grad_norm": 4.183671474456787, "learning_rate": 1.4436514875597935e-05, "loss": 1.1657, "step": 1160 }, { "epoch": 1.1206563706563706, "grad_norm": 4.838707447052002, "learning_rate": 1.4427171082244523e-05, "loss": 1.151, "step": 1161 }, { "epoch": 1.1216216216216217, "grad_norm": 5.016688823699951, "learning_rate": 1.4417822478952701e-05, "loss": 1.1944, "step": 1162 }, { "epoch": 1.1225868725868726, "grad_norm": 4.297835826873779, "learning_rate": 1.4408469075879344e-05, "loss": 1.1759, "step": 1163 }, { "epoch": 1.1235521235521235, "grad_norm": 4.285523414611816, "learning_rate": 1.4399110883186538e-05, "loss": 1.2142, "step": 1164 }, { "epoch": 1.1245173745173744, "grad_norm": 4.139347553253174, "learning_rate": 1.4389747911041568e-05, "loss": 1.216, "step": 1165 }, { "epoch": 1.1254826254826256, "grad_norm": 4.459876537322998, "learning_rate": 1.438038016961692e-05, "loss": 1.1942, "step": 1166 }, { "epoch": 1.1264478764478765, "grad_norm": 4.0485124588012695, "learning_rate": 1.4371007669090257e-05, "loss": 1.2403, "step": 1167 }, { "epoch": 1.1274131274131274, "grad_norm": 4.578690528869629, "learning_rate": 1.4361630419644417e-05, "loss": 1.182, "step": 1168 }, { "epoch": 1.1283783783783783, "grad_norm": 4.086171627044678, "learning_rate": 1.4352248431467386e-05, "loss": 1.1811, "step": 1169 }, { "epoch": 1.1293436293436294, "grad_norm": 4.123401165008545, "learning_rate": 1.4342861714752315e-05, "loss": 1.223, "step": 1170 }, { "epoch": 1.1303088803088803, "grad_norm": 4.041556358337402, "learning_rate": 1.433347027969748e-05, "loss": 1.1837, "step": 1171 }, { "epoch": 1.1312741312741312, "grad_norm": 4.314547538757324, "learning_rate": 1.4324074136506283e-05, "loss": 1.226, "step": 1172 }, { "epoch": 1.1322393822393821, "grad_norm": 3.764941692352295, "learning_rate": 1.4314673295387251e-05, "loss": 1.1753, "step": 1173 }, { "epoch": 1.1332046332046333, "grad_norm": 4.116872787475586, "learning_rate": 1.4305267766554007e-05, "loss": 1.203, "step": 1174 }, { "epoch": 1.1341698841698842, "grad_norm": 3.960906505584717, "learning_rate": 1.4295857560225268e-05, "loss": 1.1122, "step": 1175 }, { "epoch": 1.135135135135135, "grad_norm": 4.113468170166016, "learning_rate": 1.4286442686624842e-05, "loss": 1.2211, "step": 1176 }, { "epoch": 1.1361003861003862, "grad_norm": 4.512895584106445, "learning_rate": 1.4277023155981587e-05, "loss": 1.1981, "step": 1177 }, { "epoch": 1.1370656370656371, "grad_norm": 4.439307689666748, "learning_rate": 1.4267598978529447e-05, "loss": 1.231, "step": 1178 }, { "epoch": 1.138030888030888, "grad_norm": 4.276900291442871, "learning_rate": 1.425817016450739e-05, "loss": 1.1545, "step": 1179 }, { "epoch": 1.138996138996139, "grad_norm": 4.67651891708374, "learning_rate": 1.4248736724159442e-05, "loss": 1.2395, "step": 1180 }, { "epoch": 1.1399613899613898, "grad_norm": 4.468369007110596, "learning_rate": 1.4239298667734638e-05, "loss": 1.1793, "step": 1181 }, { "epoch": 1.140926640926641, "grad_norm": 4.818966388702393, "learning_rate": 1.4229856005487044e-05, "loss": 1.3307, "step": 1182 }, { "epoch": 1.1418918918918919, "grad_norm": 4.063793182373047, "learning_rate": 1.4220408747675714e-05, "loss": 1.214, "step": 1183 }, { "epoch": 1.1428571428571428, "grad_norm": 4.494610786437988, "learning_rate": 1.421095690456471e-05, "loss": 1.1376, "step": 1184 }, { "epoch": 1.143822393822394, "grad_norm": 4.65255069732666, "learning_rate": 1.4201500486423067e-05, "loss": 1.2266, "step": 1185 }, { "epoch": 1.1447876447876448, "grad_norm": 4.087616443634033, "learning_rate": 1.4192039503524792e-05, "loss": 1.1821, "step": 1186 }, { "epoch": 1.1457528957528957, "grad_norm": 4.112333297729492, "learning_rate": 1.4182573966148851e-05, "loss": 1.2152, "step": 1187 }, { "epoch": 1.1467181467181466, "grad_norm": 3.8481669425964355, "learning_rate": 1.4173103884579163e-05, "loss": 1.1877, "step": 1188 }, { "epoch": 1.1476833976833978, "grad_norm": 4.059640407562256, "learning_rate": 1.4163629269104573e-05, "loss": 1.2537, "step": 1189 }, { "epoch": 1.1486486486486487, "grad_norm": 4.0942230224609375, "learning_rate": 1.4154150130018867e-05, "loss": 1.238, "step": 1190 }, { "epoch": 1.1496138996138996, "grad_norm": 4.230532169342041, "learning_rate": 1.414466647762073e-05, "loss": 1.1883, "step": 1191 }, { "epoch": 1.1505791505791505, "grad_norm": 4.073517322540283, "learning_rate": 1.4135178322213765e-05, "loss": 1.1941, "step": 1192 }, { "epoch": 1.1515444015444016, "grad_norm": 4.37415885925293, "learning_rate": 1.412568567410646e-05, "loss": 1.1625, "step": 1193 }, { "epoch": 1.1525096525096525, "grad_norm": 4.088471412658691, "learning_rate": 1.4116188543612182e-05, "loss": 1.1843, "step": 1194 }, { "epoch": 1.1534749034749034, "grad_norm": 4.376766204833984, "learning_rate": 1.4106686941049171e-05, "loss": 1.1423, "step": 1195 }, { "epoch": 1.1544401544401544, "grad_norm": 4.660998821258545, "learning_rate": 1.4097180876740525e-05, "loss": 1.271, "step": 1196 }, { "epoch": 1.1554054054054055, "grad_norm": 4.198565483093262, "learning_rate": 1.4087670361014194e-05, "loss": 1.2037, "step": 1197 }, { "epoch": 1.1563706563706564, "grad_norm": 4.214802265167236, "learning_rate": 1.4078155404202957e-05, "loss": 1.2044, "step": 1198 }, { "epoch": 1.1573359073359073, "grad_norm": 4.1454362869262695, "learning_rate": 1.4068636016644424e-05, "loss": 1.1671, "step": 1199 }, { "epoch": 1.1583011583011582, "grad_norm": 4.182562351226807, "learning_rate": 1.4059112208681012e-05, "loss": 1.1981, "step": 1200 }, { "epoch": 1.1592664092664093, "grad_norm": 4.076030731201172, "learning_rate": 1.4049583990659947e-05, "loss": 1.2083, "step": 1201 }, { "epoch": 1.1602316602316602, "grad_norm": 4.5406270027160645, "learning_rate": 1.4040051372933244e-05, "loss": 1.2541, "step": 1202 }, { "epoch": 1.1611969111969112, "grad_norm": 4.326791286468506, "learning_rate": 1.4030514365857696e-05, "loss": 1.1372, "step": 1203 }, { "epoch": 1.1621621621621623, "grad_norm": 4.059686183929443, "learning_rate": 1.402097297979487e-05, "loss": 1.1778, "step": 1204 }, { "epoch": 1.1631274131274132, "grad_norm": 3.9642210006713867, "learning_rate": 1.4011427225111091e-05, "loss": 1.1207, "step": 1205 }, { "epoch": 1.164092664092664, "grad_norm": 4.094119071960449, "learning_rate": 1.4001877112177418e-05, "loss": 1.1544, "step": 1206 }, { "epoch": 1.165057915057915, "grad_norm": 4.555078506469727, "learning_rate": 1.3992322651369663e-05, "loss": 1.1621, "step": 1207 }, { "epoch": 1.166023166023166, "grad_norm": 4.552582740783691, "learning_rate": 1.3982763853068345e-05, "loss": 1.175, "step": 1208 }, { "epoch": 1.166988416988417, "grad_norm": 4.766264915466309, "learning_rate": 1.3973200727658709e-05, "loss": 1.1716, "step": 1209 }, { "epoch": 1.167953667953668, "grad_norm": 4.566944599151611, "learning_rate": 1.3963633285530688e-05, "loss": 1.2204, "step": 1210 }, { "epoch": 1.1689189189189189, "grad_norm": 4.437086582183838, "learning_rate": 1.3954061537078927e-05, "loss": 1.1336, "step": 1211 }, { "epoch": 1.16988416988417, "grad_norm": 4.496171474456787, "learning_rate": 1.3944485492702717e-05, "loss": 1.2572, "step": 1212 }, { "epoch": 1.170849420849421, "grad_norm": 4.487252235412598, "learning_rate": 1.393490516280605e-05, "loss": 1.2293, "step": 1213 }, { "epoch": 1.1718146718146718, "grad_norm": 4.405369281768799, "learning_rate": 1.3925320557797549e-05, "loss": 1.2419, "step": 1214 }, { "epoch": 1.1727799227799227, "grad_norm": 4.267056941986084, "learning_rate": 1.3915731688090492e-05, "loss": 1.1946, "step": 1215 }, { "epoch": 1.1737451737451738, "grad_norm": 4.19378137588501, "learning_rate": 1.3906138564102794e-05, "loss": 1.1232, "step": 1216 }, { "epoch": 1.1747104247104247, "grad_norm": 4.049899101257324, "learning_rate": 1.3896541196256985e-05, "loss": 1.2202, "step": 1217 }, { "epoch": 1.1756756756756757, "grad_norm": 4.774388313293457, "learning_rate": 1.3886939594980209e-05, "loss": 1.2143, "step": 1218 }, { "epoch": 1.1766409266409266, "grad_norm": 3.9391682147979736, "learning_rate": 1.3877333770704208e-05, "loss": 1.259, "step": 1219 }, { "epoch": 1.1776061776061777, "grad_norm": 4.408590793609619, "learning_rate": 1.3867723733865314e-05, "loss": 1.2118, "step": 1220 }, { "epoch": 1.1785714285714286, "grad_norm": 4.427639007568359, "learning_rate": 1.3858109494904437e-05, "loss": 1.1951, "step": 1221 }, { "epoch": 1.1795366795366795, "grad_norm": 4.049868106842041, "learning_rate": 1.3848491064267046e-05, "loss": 1.187, "step": 1222 }, { "epoch": 1.1805019305019304, "grad_norm": 3.7526743412017822, "learning_rate": 1.3838868452403172e-05, "loss": 1.1953, "step": 1223 }, { "epoch": 1.1814671814671815, "grad_norm": 4.315376281738281, "learning_rate": 1.3829241669767381e-05, "loss": 1.1979, "step": 1224 }, { "epoch": 1.1824324324324325, "grad_norm": 4.181025505065918, "learning_rate": 1.381961072681878e-05, "loss": 1.1545, "step": 1225 }, { "epoch": 1.1833976833976834, "grad_norm": 4.2237548828125, "learning_rate": 1.3809975634020985e-05, "loss": 1.1436, "step": 1226 }, { "epoch": 1.1843629343629343, "grad_norm": 4.464808464050293, "learning_rate": 1.380033640184213e-05, "loss": 1.1657, "step": 1227 }, { "epoch": 1.1853281853281854, "grad_norm": 4.275068759918213, "learning_rate": 1.3790693040754839e-05, "loss": 1.2498, "step": 1228 }, { "epoch": 1.1862934362934363, "grad_norm": 4.397349834442139, "learning_rate": 1.3781045561236228e-05, "loss": 1.2087, "step": 1229 }, { "epoch": 1.1872586872586872, "grad_norm": 4.3661980628967285, "learning_rate": 1.3771393973767886e-05, "loss": 1.2416, "step": 1230 }, { "epoch": 1.1882239382239383, "grad_norm": 4.0517802238464355, "learning_rate": 1.376173828883586e-05, "loss": 1.1691, "step": 1231 }, { "epoch": 1.1891891891891893, "grad_norm": 4.048524856567383, "learning_rate": 1.3752078516930653e-05, "loss": 1.1568, "step": 1232 }, { "epoch": 1.1901544401544402, "grad_norm": 4.841455936431885, "learning_rate": 1.374241466854721e-05, "loss": 1.1866, "step": 1233 }, { "epoch": 1.191119691119691, "grad_norm": 4.119144439697266, "learning_rate": 1.3732746754184905e-05, "loss": 1.2027, "step": 1234 }, { "epoch": 1.192084942084942, "grad_norm": 4.689269542694092, "learning_rate": 1.3723074784347521e-05, "loss": 1.2116, "step": 1235 }, { "epoch": 1.193050193050193, "grad_norm": 5.033204078674316, "learning_rate": 1.371339876954326e-05, "loss": 1.2976, "step": 1236 }, { "epoch": 1.194015444015444, "grad_norm": 4.489188194274902, "learning_rate": 1.3703718720284707e-05, "loss": 1.0913, "step": 1237 }, { "epoch": 1.194980694980695, "grad_norm": 4.7308669090271, "learning_rate": 1.369403464708884e-05, "loss": 1.216, "step": 1238 }, { "epoch": 1.195945945945946, "grad_norm": 4.222895622253418, "learning_rate": 1.3684346560477e-05, "loss": 1.1872, "step": 1239 }, { "epoch": 1.196911196911197, "grad_norm": 3.997390031814575, "learning_rate": 1.3674654470974897e-05, "loss": 1.1965, "step": 1240 }, { "epoch": 1.1978764478764479, "grad_norm": 4.557868003845215, "learning_rate": 1.3664958389112581e-05, "loss": 1.2309, "step": 1241 }, { "epoch": 1.1988416988416988, "grad_norm": 4.424615383148193, "learning_rate": 1.3655258325424447e-05, "loss": 1.2487, "step": 1242 }, { "epoch": 1.1998069498069497, "grad_norm": 4.087277889251709, "learning_rate": 1.3645554290449215e-05, "loss": 1.2148, "step": 1243 }, { "epoch": 1.2007722007722008, "grad_norm": 3.749039649963379, "learning_rate": 1.3635846294729915e-05, "loss": 1.1904, "step": 1244 }, { "epoch": 1.2017374517374517, "grad_norm": 4.259382724761963, "learning_rate": 1.3626134348813885e-05, "loss": 1.2536, "step": 1245 }, { "epoch": 1.2027027027027026, "grad_norm": 4.148220062255859, "learning_rate": 1.3616418463252747e-05, "loss": 1.2717, "step": 1246 }, { "epoch": 1.2036679536679538, "grad_norm": 4.54502534866333, "learning_rate": 1.360669864860242e-05, "loss": 1.2153, "step": 1247 }, { "epoch": 1.2046332046332047, "grad_norm": 3.793149471282959, "learning_rate": 1.3596974915423071e-05, "loss": 1.1592, "step": 1248 }, { "epoch": 1.2055984555984556, "grad_norm": 4.2187652587890625, "learning_rate": 1.358724727427914e-05, "loss": 1.1378, "step": 1249 }, { "epoch": 1.2065637065637065, "grad_norm": 4.317118167877197, "learning_rate": 1.3577515735739302e-05, "loss": 1.2069, "step": 1250 }, { "epoch": 1.2075289575289576, "grad_norm": 4.604206562042236, "learning_rate": 1.3567780310376476e-05, "loss": 1.1417, "step": 1251 }, { "epoch": 1.2084942084942085, "grad_norm": 4.524065017700195, "learning_rate": 1.3558041008767799e-05, "loss": 1.1692, "step": 1252 }, { "epoch": 1.2094594594594594, "grad_norm": 4.630565166473389, "learning_rate": 1.354829784149462e-05, "loss": 1.217, "step": 1253 }, { "epoch": 1.2104247104247103, "grad_norm": 4.320188045501709, "learning_rate": 1.3538550819142484e-05, "loss": 1.2048, "step": 1254 }, { "epoch": 1.2113899613899615, "grad_norm": 4.694369792938232, "learning_rate": 1.3528799952301128e-05, "loss": 1.2499, "step": 1255 }, { "epoch": 1.2123552123552124, "grad_norm": 4.20485782623291, "learning_rate": 1.351904525156447e-05, "loss": 1.2518, "step": 1256 }, { "epoch": 1.2133204633204633, "grad_norm": 4.132609844207764, "learning_rate": 1.3509286727530588e-05, "loss": 1.1469, "step": 1257 }, { "epoch": 1.2142857142857142, "grad_norm": 4.2905592918396, "learning_rate": 1.3499524390801716e-05, "loss": 1.2003, "step": 1258 }, { "epoch": 1.2152509652509653, "grad_norm": 4.843905448913574, "learning_rate": 1.3489758251984222e-05, "loss": 1.2131, "step": 1259 }, { "epoch": 1.2162162162162162, "grad_norm": 4.21188497543335, "learning_rate": 1.3479988321688619e-05, "loss": 1.2113, "step": 1260 }, { "epoch": 1.2171814671814671, "grad_norm": 4.492689609527588, "learning_rate": 1.3470214610529536e-05, "loss": 1.2315, "step": 1261 }, { "epoch": 1.218146718146718, "grad_norm": 4.724948883056641, "learning_rate": 1.3460437129125694e-05, "loss": 1.2238, "step": 1262 }, { "epoch": 1.2191119691119692, "grad_norm": 4.601674556732178, "learning_rate": 1.3450655888099935e-05, "loss": 1.1761, "step": 1263 }, { "epoch": 1.22007722007722, "grad_norm": 5.123815059661865, "learning_rate": 1.3440870898079166e-05, "loss": 1.2684, "step": 1264 }, { "epoch": 1.221042471042471, "grad_norm": 4.135692119598389, "learning_rate": 1.3431082169694376e-05, "loss": 1.1713, "step": 1265 }, { "epoch": 1.2220077220077221, "grad_norm": 4.55994176864624, "learning_rate": 1.3421289713580616e-05, "loss": 1.2257, "step": 1266 }, { "epoch": 1.222972972972973, "grad_norm": 4.352341175079346, "learning_rate": 1.341149354037698e-05, "loss": 1.223, "step": 1267 }, { "epoch": 1.223938223938224, "grad_norm": 4.42564058303833, "learning_rate": 1.340169366072661e-05, "loss": 1.1589, "step": 1268 }, { "epoch": 1.2249034749034748, "grad_norm": 4.8221611976623535, "learning_rate": 1.3391890085276669e-05, "loss": 1.2191, "step": 1269 }, { "epoch": 1.2258687258687258, "grad_norm": 4.877390384674072, "learning_rate": 1.3382082824678336e-05, "loss": 1.2135, "step": 1270 }, { "epoch": 1.2268339768339769, "grad_norm": 4.110681533813477, "learning_rate": 1.337227188958679e-05, "loss": 1.172, "step": 1271 }, { "epoch": 1.2277992277992278, "grad_norm": 5.047957420349121, "learning_rate": 1.3362457290661215e-05, "loss": 1.2651, "step": 1272 }, { "epoch": 1.2287644787644787, "grad_norm": 4.183019161224365, "learning_rate": 1.335263903856476e-05, "loss": 1.2434, "step": 1273 }, { "epoch": 1.2297297297297298, "grad_norm": 5.047949314117432, "learning_rate": 1.3342817143964557e-05, "loss": 1.2027, "step": 1274 }, { "epoch": 1.2306949806949807, "grad_norm": 4.732578277587891, "learning_rate": 1.333299161753168e-05, "loss": 1.2546, "step": 1275 }, { "epoch": 1.2316602316602316, "grad_norm": 4.286832809448242, "learning_rate": 1.3323162469941164e-05, "loss": 1.2295, "step": 1276 }, { "epoch": 1.2326254826254825, "grad_norm": 4.886010646820068, "learning_rate": 1.3313329711871969e-05, "loss": 1.1589, "step": 1277 }, { "epoch": 1.2335907335907337, "grad_norm": 4.348087310791016, "learning_rate": 1.3303493354006985e-05, "loss": 1.2417, "step": 1278 }, { "epoch": 1.2345559845559846, "grad_norm": 4.194649696350098, "learning_rate": 1.3293653407032997e-05, "loss": 1.197, "step": 1279 }, { "epoch": 1.2355212355212355, "grad_norm": 4.147604942321777, "learning_rate": 1.328380988164071e-05, "loss": 1.2041, "step": 1280 }, { "epoch": 1.2364864864864864, "grad_norm": 4.729609489440918, "learning_rate": 1.3273962788524705e-05, "loss": 1.2053, "step": 1281 }, { "epoch": 1.2374517374517375, "grad_norm": 5.589478969573975, "learning_rate": 1.3264112138383445e-05, "loss": 1.262, "step": 1282 }, { "epoch": 1.2384169884169884, "grad_norm": 3.983586072921753, "learning_rate": 1.3254257941919249e-05, "loss": 1.1723, "step": 1283 }, { "epoch": 1.2393822393822393, "grad_norm": 4.140334606170654, "learning_rate": 1.3244400209838297e-05, "loss": 1.2247, "step": 1284 }, { "epoch": 1.2403474903474903, "grad_norm": 4.859688758850098, "learning_rate": 1.3234538952850606e-05, "loss": 1.1928, "step": 1285 }, { "epoch": 1.2413127413127414, "grad_norm": 4.2499189376831055, "learning_rate": 1.3224674181670025e-05, "loss": 1.158, "step": 1286 }, { "epoch": 1.2422779922779923, "grad_norm": 3.908384323120117, "learning_rate": 1.3214805907014223e-05, "loss": 1.2375, "step": 1287 }, { "epoch": 1.2432432432432432, "grad_norm": 4.4078168869018555, "learning_rate": 1.3204934139604668e-05, "loss": 1.1594, "step": 1288 }, { "epoch": 1.244208494208494, "grad_norm": 4.318722724914551, "learning_rate": 1.3195058890166628e-05, "loss": 1.2081, "step": 1289 }, { "epoch": 1.2451737451737452, "grad_norm": 4.253969192504883, "learning_rate": 1.3185180169429155e-05, "loss": 1.1388, "step": 1290 }, { "epoch": 1.2461389961389961, "grad_norm": 4.060301780700684, "learning_rate": 1.3175297988125071e-05, "loss": 1.2017, "step": 1291 }, { "epoch": 1.247104247104247, "grad_norm": 4.686599254608154, "learning_rate": 1.3165412356990954e-05, "loss": 1.1764, "step": 1292 }, { "epoch": 1.2480694980694982, "grad_norm": 4.2792067527771, "learning_rate": 1.315552328676714e-05, "loss": 1.1698, "step": 1293 }, { "epoch": 1.249034749034749, "grad_norm": 4.322442531585693, "learning_rate": 1.314563078819769e-05, "loss": 1.1632, "step": 1294 }, { "epoch": 1.25, "grad_norm": 4.505197525024414, "learning_rate": 1.3135734872030397e-05, "loss": 1.2016, "step": 1295 }, { "epoch": 1.250965250965251, "grad_norm": 4.2458906173706055, "learning_rate": 1.3125835549016763e-05, "loss": 1.1946, "step": 1296 }, { "epoch": 1.2519305019305018, "grad_norm": 4.345571517944336, "learning_rate": 1.3115932829911997e-05, "loss": 1.2077, "step": 1297 }, { "epoch": 1.252895752895753, "grad_norm": 4.5000433921813965, "learning_rate": 1.3106026725474987e-05, "loss": 1.1879, "step": 1298 }, { "epoch": 1.2538610038610039, "grad_norm": 4.399386405944824, "learning_rate": 1.3096117246468317e-05, "loss": 1.2398, "step": 1299 }, { "epoch": 1.2548262548262548, "grad_norm": 4.267080307006836, "learning_rate": 1.3086204403658216e-05, "loss": 1.2535, "step": 1300 }, { "epoch": 1.255791505791506, "grad_norm": 4.634487628936768, "learning_rate": 1.3076288207814585e-05, "loss": 1.3211, "step": 1301 }, { "epoch": 1.2567567567567568, "grad_norm": 3.9481966495513916, "learning_rate": 1.3066368669710953e-05, "loss": 1.1489, "step": 1302 }, { "epoch": 1.2577220077220077, "grad_norm": 4.762670993804932, "learning_rate": 1.3056445800124497e-05, "loss": 1.217, "step": 1303 }, { "epoch": 1.2586872586872586, "grad_norm": 4.1158766746521, "learning_rate": 1.3046519609836002e-05, "loss": 1.1996, "step": 1304 }, { "epoch": 1.2596525096525095, "grad_norm": 4.694605827331543, "learning_rate": 1.303659010962986e-05, "loss": 1.1895, "step": 1305 }, { "epoch": 1.2606177606177607, "grad_norm": 5.370503902435303, "learning_rate": 1.3026657310294067e-05, "loss": 1.2432, "step": 1306 }, { "epoch": 1.2615830115830116, "grad_norm": 4.1840362548828125, "learning_rate": 1.3016721222620197e-05, "loss": 1.2045, "step": 1307 }, { "epoch": 1.2625482625482625, "grad_norm": 5.977203845977783, "learning_rate": 1.3006781857403394e-05, "loss": 1.2218, "step": 1308 }, { "epoch": 1.2635135135135136, "grad_norm": 4.654899597167969, "learning_rate": 1.2996839225442377e-05, "loss": 1.3054, "step": 1309 }, { "epoch": 1.2644787644787645, "grad_norm": 4.7028727531433105, "learning_rate": 1.2986893337539397e-05, "loss": 1.2712, "step": 1310 }, { "epoch": 1.2654440154440154, "grad_norm": 5.325465679168701, "learning_rate": 1.2976944204500255e-05, "loss": 1.1635, "step": 1311 }, { "epoch": 1.2664092664092665, "grad_norm": 4.675345420837402, "learning_rate": 1.2966991837134271e-05, "loss": 1.267, "step": 1312 }, { "epoch": 1.2673745173745175, "grad_norm": 4.558609962463379, "learning_rate": 1.295703624625428e-05, "loss": 1.2272, "step": 1313 }, { "epoch": 1.2683397683397684, "grad_norm": 4.645543575286865, "learning_rate": 1.294707744267662e-05, "loss": 1.2366, "step": 1314 }, { "epoch": 1.2693050193050193, "grad_norm": 4.8329620361328125, "learning_rate": 1.2937115437221119e-05, "loss": 1.193, "step": 1315 }, { "epoch": 1.2702702702702702, "grad_norm": 4.300382137298584, "learning_rate": 1.2927150240711089e-05, "loss": 1.1907, "step": 1316 }, { "epoch": 1.2712355212355213, "grad_norm": 4.584596633911133, "learning_rate": 1.2917181863973298e-05, "loss": 1.2358, "step": 1317 }, { "epoch": 1.2722007722007722, "grad_norm": 4.23142671585083, "learning_rate": 1.290721031783798e-05, "loss": 1.2462, "step": 1318 }, { "epoch": 1.2731660231660231, "grad_norm": 4.2517828941345215, "learning_rate": 1.2897235613138804e-05, "loss": 1.2021, "step": 1319 }, { "epoch": 1.2741312741312742, "grad_norm": 4.605195999145508, "learning_rate": 1.2887257760712875e-05, "loss": 1.246, "step": 1320 }, { "epoch": 1.2750965250965252, "grad_norm": 4.4199347496032715, "learning_rate": 1.287727677140072e-05, "loss": 1.1999, "step": 1321 }, { "epoch": 1.276061776061776, "grad_norm": 3.882155656814575, "learning_rate": 1.2867292656046268e-05, "loss": 1.1948, "step": 1322 }, { "epoch": 1.277027027027027, "grad_norm": 4.638187885284424, "learning_rate": 1.285730542549685e-05, "loss": 1.2795, "step": 1323 }, { "epoch": 1.2779922779922779, "grad_norm": 4.155126571655273, "learning_rate": 1.284731509060318e-05, "loss": 1.1946, "step": 1324 }, { "epoch": 1.278957528957529, "grad_norm": 4.521946907043457, "learning_rate": 1.2837321662219342e-05, "loss": 1.2178, "step": 1325 }, { "epoch": 1.27992277992278, "grad_norm": 4.1087212562561035, "learning_rate": 1.2827325151202783e-05, "loss": 1.2762, "step": 1326 }, { "epoch": 1.2808880308880308, "grad_norm": 3.956718683242798, "learning_rate": 1.2817325568414299e-05, "loss": 1.2021, "step": 1327 }, { "epoch": 1.281853281853282, "grad_norm": 4.165088176727295, "learning_rate": 1.2807322924718024e-05, "loss": 1.1841, "step": 1328 }, { "epoch": 1.2828185328185329, "grad_norm": 4.392886161804199, "learning_rate": 1.2797317230981416e-05, "loss": 1.1817, "step": 1329 }, { "epoch": 1.2837837837837838, "grad_norm": 4.641284465789795, "learning_rate": 1.2787308498075253e-05, "loss": 1.1638, "step": 1330 }, { "epoch": 1.2847490347490347, "grad_norm": 4.319555759429932, "learning_rate": 1.27772967368736e-05, "loss": 1.268, "step": 1331 }, { "epoch": 1.2857142857142856, "grad_norm": 4.382452964782715, "learning_rate": 1.276728195825383e-05, "loss": 1.1895, "step": 1332 }, { "epoch": 1.2866795366795367, "grad_norm": 4.566013813018799, "learning_rate": 1.2757264173096582e-05, "loss": 1.213, "step": 1333 }, { "epoch": 1.2876447876447876, "grad_norm": 4.029094219207764, "learning_rate": 1.2747243392285769e-05, "loss": 1.2524, "step": 1334 }, { "epoch": 1.2886100386100385, "grad_norm": 4.885570049285889, "learning_rate": 1.2737219626708549e-05, "loss": 1.2848, "step": 1335 }, { "epoch": 1.2895752895752897, "grad_norm": 4.629604816436768, "learning_rate": 1.2727192887255332e-05, "loss": 1.1951, "step": 1336 }, { "epoch": 1.2905405405405406, "grad_norm": 4.667120933532715, "learning_rate": 1.2717163184819761e-05, "loss": 1.1794, "step": 1337 }, { "epoch": 1.2915057915057915, "grad_norm": 4.037268161773682, "learning_rate": 1.2707130530298688e-05, "loss": 1.2396, "step": 1338 }, { "epoch": 1.2924710424710426, "grad_norm": 4.660993576049805, "learning_rate": 1.2697094934592177e-05, "loss": 1.2385, "step": 1339 }, { "epoch": 1.2934362934362935, "grad_norm": 4.749906539916992, "learning_rate": 1.2687056408603493e-05, "loss": 1.1782, "step": 1340 }, { "epoch": 1.2944015444015444, "grad_norm": 4.700593948364258, "learning_rate": 1.2677014963239075e-05, "loss": 1.2046, "step": 1341 }, { "epoch": 1.2953667953667953, "grad_norm": 4.356866359710693, "learning_rate": 1.2666970609408535e-05, "loss": 1.2227, "step": 1342 }, { "epoch": 1.2963320463320462, "grad_norm": 5.177303791046143, "learning_rate": 1.2656923358024659e-05, "loss": 1.2335, "step": 1343 }, { "epoch": 1.2972972972972974, "grad_norm": 5.240119934082031, "learning_rate": 1.264687322000336e-05, "loss": 1.1615, "step": 1344 }, { "epoch": 1.2982625482625483, "grad_norm": 4.722900867462158, "learning_rate": 1.2636820206263704e-05, "loss": 1.211, "step": 1345 }, { "epoch": 1.2992277992277992, "grad_norm": 5.814818859100342, "learning_rate": 1.2626764327727871e-05, "loss": 1.2365, "step": 1346 }, { "epoch": 1.3001930501930503, "grad_norm": 4.204340934753418, "learning_rate": 1.261670559532116e-05, "loss": 1.1819, "step": 1347 }, { "epoch": 1.3011583011583012, "grad_norm": 4.929001331329346, "learning_rate": 1.2606644019971967e-05, "loss": 1.2352, "step": 1348 }, { "epoch": 1.3021235521235521, "grad_norm": 4.845025539398193, "learning_rate": 1.259657961261178e-05, "loss": 1.2067, "step": 1349 }, { "epoch": 1.303088803088803, "grad_norm": 4.197154998779297, "learning_rate": 1.2586512384175156e-05, "loss": 1.2472, "step": 1350 }, { "epoch": 1.304054054054054, "grad_norm": 4.3102922439575195, "learning_rate": 1.2576442345599729e-05, "loss": 1.2445, "step": 1351 }, { "epoch": 1.305019305019305, "grad_norm": 4.416947364807129, "learning_rate": 1.2566369507826175e-05, "loss": 1.2305, "step": 1352 }, { "epoch": 1.305984555984556, "grad_norm": 4.655396461486816, "learning_rate": 1.2556293881798218e-05, "loss": 1.2124, "step": 1353 }, { "epoch": 1.306949806949807, "grad_norm": 4.648707866668701, "learning_rate": 1.2546215478462611e-05, "loss": 1.1722, "step": 1354 }, { "epoch": 1.307915057915058, "grad_norm": 4.69895076751709, "learning_rate": 1.2536134308769118e-05, "loss": 1.1696, "step": 1355 }, { "epoch": 1.308880308880309, "grad_norm": 4.79047966003418, "learning_rate": 1.2526050383670516e-05, "loss": 1.1772, "step": 1356 }, { "epoch": 1.3098455598455598, "grad_norm": 4.19055700302124, "learning_rate": 1.2515963714122577e-05, "loss": 1.1856, "step": 1357 }, { "epoch": 1.3108108108108107, "grad_norm": 4.9402241706848145, "learning_rate": 1.2505874311084041e-05, "loss": 1.2117, "step": 1358 }, { "epoch": 1.3117760617760617, "grad_norm": 5.0677924156188965, "learning_rate": 1.2495782185516638e-05, "loss": 1.2339, "step": 1359 }, { "epoch": 1.3127413127413128, "grad_norm": 4.164126396179199, "learning_rate": 1.2485687348385038e-05, "loss": 1.1588, "step": 1360 }, { "epoch": 1.3137065637065637, "grad_norm": 5.482615947723389, "learning_rate": 1.2475589810656868e-05, "loss": 1.1945, "step": 1361 }, { "epoch": 1.3146718146718146, "grad_norm": 5.1199750900268555, "learning_rate": 1.246548958330268e-05, "loss": 1.1531, "step": 1362 }, { "epoch": 1.3156370656370657, "grad_norm": 5.075857162475586, "learning_rate": 1.2455386677295962e-05, "loss": 1.1786, "step": 1363 }, { "epoch": 1.3166023166023166, "grad_norm": 6.397689342498779, "learning_rate": 1.2445281103613096e-05, "loss": 1.2092, "step": 1364 }, { "epoch": 1.3175675675675675, "grad_norm": 4.174447059631348, "learning_rate": 1.2435172873233371e-05, "loss": 1.1784, "step": 1365 }, { "epoch": 1.3185328185328185, "grad_norm": 5.897645950317383, "learning_rate": 1.2425061997138972e-05, "loss": 1.2342, "step": 1366 }, { "epoch": 1.3194980694980696, "grad_norm": 5.2558794021606445, "learning_rate": 1.2414948486314932e-05, "loss": 1.263, "step": 1367 }, { "epoch": 1.3204633204633205, "grad_norm": 4.3572845458984375, "learning_rate": 1.2404832351749179e-05, "loss": 1.2517, "step": 1368 }, { "epoch": 1.3214285714285714, "grad_norm": 5.881748676300049, "learning_rate": 1.239471360443246e-05, "loss": 1.2184, "step": 1369 }, { "epoch": 1.3223938223938223, "grad_norm": 4.852202415466309, "learning_rate": 1.2384592255358385e-05, "loss": 1.176, "step": 1370 }, { "epoch": 1.3233590733590734, "grad_norm": 4.252193450927734, "learning_rate": 1.2374468315523375e-05, "loss": 1.2621, "step": 1371 }, { "epoch": 1.3243243243243243, "grad_norm": 6.008481025695801, "learning_rate": 1.2364341795926684e-05, "loss": 1.1949, "step": 1372 }, { "epoch": 1.3252895752895753, "grad_norm": 5.349253177642822, "learning_rate": 1.2354212707570344e-05, "loss": 1.217, "step": 1373 }, { "epoch": 1.3262548262548264, "grad_norm": 5.8846435546875, "learning_rate": 1.2344081061459195e-05, "loss": 1.2277, "step": 1374 }, { "epoch": 1.3272200772200773, "grad_norm": 6.014350891113281, "learning_rate": 1.2333946868600854e-05, "loss": 1.1879, "step": 1375 }, { "epoch": 1.3281853281853282, "grad_norm": 3.9860541820526123, "learning_rate": 1.2323810140005701e-05, "loss": 1.1859, "step": 1376 }, { "epoch": 1.329150579150579, "grad_norm": 6.001570701599121, "learning_rate": 1.2313670886686874e-05, "loss": 1.1648, "step": 1377 }, { "epoch": 1.33011583011583, "grad_norm": 5.547324180603027, "learning_rate": 1.230352911966025e-05, "loss": 1.2331, "step": 1378 }, { "epoch": 1.3310810810810811, "grad_norm": 4.244880199432373, "learning_rate": 1.2293384849944445e-05, "loss": 1.1715, "step": 1379 }, { "epoch": 1.332046332046332, "grad_norm": 5.324029445648193, "learning_rate": 1.2283238088560783e-05, "loss": 1.2098, "step": 1380 }, { "epoch": 1.333011583011583, "grad_norm": 5.186968803405762, "learning_rate": 1.2273088846533303e-05, "loss": 1.1639, "step": 1381 }, { "epoch": 1.333976833976834, "grad_norm": 4.450983047485352, "learning_rate": 1.226293713488874e-05, "loss": 1.253, "step": 1382 }, { "epoch": 1.334942084942085, "grad_norm": 4.1214680671691895, "learning_rate": 1.2252782964656502e-05, "loss": 1.1587, "step": 1383 }, { "epoch": 1.335907335907336, "grad_norm": 4.0626349449157715, "learning_rate": 1.2242626346868682e-05, "loss": 1.1852, "step": 1384 }, { "epoch": 1.3368725868725868, "grad_norm": 4.6084723472595215, "learning_rate": 1.2232467292560025e-05, "loss": 1.2039, "step": 1385 }, { "epoch": 1.3378378378378377, "grad_norm": 4.472030162811279, "learning_rate": 1.2222305812767919e-05, "loss": 1.2213, "step": 1386 }, { "epoch": 1.3388030888030888, "grad_norm": 4.617482662200928, "learning_rate": 1.2212141918532394e-05, "loss": 1.2015, "step": 1387 }, { "epoch": 1.3397683397683398, "grad_norm": 4.271813869476318, "learning_rate": 1.2201975620896102e-05, "loss": 1.1907, "step": 1388 }, { "epoch": 1.3407335907335907, "grad_norm": 4.236230373382568, "learning_rate": 1.2191806930904304e-05, "loss": 1.2442, "step": 1389 }, { "epoch": 1.3416988416988418, "grad_norm": 4.1355204582214355, "learning_rate": 1.2181635859604864e-05, "loss": 1.1606, "step": 1390 }, { "epoch": 1.3426640926640927, "grad_norm": 4.975465297698975, "learning_rate": 1.2171462418048228e-05, "loss": 1.1792, "step": 1391 }, { "epoch": 1.3436293436293436, "grad_norm": 4.425232410430908, "learning_rate": 1.216128661728742e-05, "loss": 1.2268, "step": 1392 }, { "epoch": 1.3445945945945945, "grad_norm": 4.223307132720947, "learning_rate": 1.2151108468378028e-05, "loss": 1.2575, "step": 1393 }, { "epoch": 1.3455598455598454, "grad_norm": 4.804199695587158, "learning_rate": 1.2140927982378189e-05, "loss": 1.2522, "step": 1394 }, { "epoch": 1.3465250965250966, "grad_norm": 4.646818161010742, "learning_rate": 1.2130745170348584e-05, "loss": 1.2433, "step": 1395 }, { "epoch": 1.3474903474903475, "grad_norm": 3.9267654418945312, "learning_rate": 1.212056004335241e-05, "loss": 1.1664, "step": 1396 }, { "epoch": 1.3484555984555984, "grad_norm": 4.747241497039795, "learning_rate": 1.2110372612455394e-05, "loss": 1.2441, "step": 1397 }, { "epoch": 1.3494208494208495, "grad_norm": 4.446130275726318, "learning_rate": 1.2100182888725751e-05, "loss": 1.1992, "step": 1398 }, { "epoch": 1.3503861003861004, "grad_norm": 4.628722190856934, "learning_rate": 1.2089990883234206e-05, "loss": 1.207, "step": 1399 }, { "epoch": 1.3513513513513513, "grad_norm": 5.560851573944092, "learning_rate": 1.2079796607053939e-05, "loss": 1.2201, "step": 1400 }, { "epoch": 1.3523166023166024, "grad_norm": 3.9680185317993164, "learning_rate": 1.2069600071260621e-05, "loss": 1.1057, "step": 1401 }, { "epoch": 1.3532818532818534, "grad_norm": 5.056230068206787, "learning_rate": 1.2059401286932354e-05, "loss": 1.1756, "step": 1402 }, { "epoch": 1.3542471042471043, "grad_norm": 5.653619766235352, "learning_rate": 1.204920026514971e-05, "loss": 1.2339, "step": 1403 }, { "epoch": 1.3552123552123552, "grad_norm": 4.381045818328857, "learning_rate": 1.2038997016995668e-05, "loss": 1.2332, "step": 1404 }, { "epoch": 1.356177606177606, "grad_norm": 5.91451358795166, "learning_rate": 1.202879155355564e-05, "loss": 1.2509, "step": 1405 }, { "epoch": 1.3571428571428572, "grad_norm": 5.2406535148620605, "learning_rate": 1.2018583885917434e-05, "loss": 1.1803, "step": 1406 }, { "epoch": 1.3581081081081081, "grad_norm": 3.9339377880096436, "learning_rate": 1.2008374025171265e-05, "loss": 1.142, "step": 1407 }, { "epoch": 1.359073359073359, "grad_norm": 5.352104187011719, "learning_rate": 1.1998161982409724e-05, "loss": 1.2281, "step": 1408 }, { "epoch": 1.3600386100386102, "grad_norm": 5.339706897735596, "learning_rate": 1.1987947768727773e-05, "loss": 1.1884, "step": 1409 }, { "epoch": 1.361003861003861, "grad_norm": 4.254153728485107, "learning_rate": 1.1977731395222731e-05, "loss": 1.2148, "step": 1410 }, { "epoch": 1.361969111969112, "grad_norm": 4.582601547241211, "learning_rate": 1.1967512872994266e-05, "loss": 1.198, "step": 1411 }, { "epoch": 1.3629343629343629, "grad_norm": 4.7897844314575195, "learning_rate": 1.1957292213144386e-05, "loss": 1.1731, "step": 1412 }, { "epoch": 1.3638996138996138, "grad_norm": 4.098704814910889, "learning_rate": 1.1947069426777404e-05, "loss": 1.1872, "step": 1413 }, { "epoch": 1.364864864864865, "grad_norm": 5.669726371765137, "learning_rate": 1.1936844524999966e-05, "loss": 1.1926, "step": 1414 }, { "epoch": 1.3658301158301158, "grad_norm": 4.694494247436523, "learning_rate": 1.1926617518921e-05, "loss": 1.189, "step": 1415 }, { "epoch": 1.3667953667953667, "grad_norm": 4.791158199310303, "learning_rate": 1.1916388419651725e-05, "loss": 1.179, "step": 1416 }, { "epoch": 1.3677606177606179, "grad_norm": 5.525454521179199, "learning_rate": 1.1906157238305635e-05, "loss": 1.2432, "step": 1417 }, { "epoch": 1.3687258687258688, "grad_norm": 4.639889717102051, "learning_rate": 1.189592398599849e-05, "loss": 1.189, "step": 1418 }, { "epoch": 1.3696911196911197, "grad_norm": 5.080594539642334, "learning_rate": 1.1885688673848289e-05, "loss": 1.1942, "step": 1419 }, { "epoch": 1.3706563706563706, "grad_norm": 4.258522033691406, "learning_rate": 1.1875451312975278e-05, "loss": 1.2235, "step": 1420 }, { "epoch": 1.3716216216216215, "grad_norm": 4.478529453277588, "learning_rate": 1.1865211914501926e-05, "loss": 1.1379, "step": 1421 }, { "epoch": 1.3725868725868726, "grad_norm": 4.141099452972412, "learning_rate": 1.1854970489552924e-05, "loss": 1.1214, "step": 1422 }, { "epoch": 1.3735521235521235, "grad_norm": 4.646026611328125, "learning_rate": 1.1844727049255143e-05, "loss": 1.229, "step": 1423 }, { "epoch": 1.3745173745173744, "grad_norm": 4.416215419769287, "learning_rate": 1.1834481604737671e-05, "loss": 1.1703, "step": 1424 }, { "epoch": 1.3754826254826256, "grad_norm": 4.077454090118408, "learning_rate": 1.1824234167131748e-05, "loss": 1.1772, "step": 1425 }, { "epoch": 1.3764478764478765, "grad_norm": 4.6979660987854, "learning_rate": 1.1813984747570802e-05, "loss": 1.2484, "step": 1426 }, { "epoch": 1.3774131274131274, "grad_norm": 4.771971225738525, "learning_rate": 1.1803733357190396e-05, "loss": 1.1678, "step": 1427 }, { "epoch": 1.3783783783783785, "grad_norm": 4.0661420822143555, "learning_rate": 1.1793480007128248e-05, "loss": 1.2631, "step": 1428 }, { "epoch": 1.3793436293436294, "grad_norm": 4.452084541320801, "learning_rate": 1.1783224708524192e-05, "loss": 1.2458, "step": 1429 }, { "epoch": 1.3803088803088803, "grad_norm": 4.443688869476318, "learning_rate": 1.1772967472520196e-05, "loss": 1.2118, "step": 1430 }, { "epoch": 1.3812741312741312, "grad_norm": 4.165465354919434, "learning_rate": 1.1762708310260312e-05, "loss": 1.1932, "step": 1431 }, { "epoch": 1.3822393822393821, "grad_norm": 4.817536354064941, "learning_rate": 1.1752447232890702e-05, "loss": 1.1686, "step": 1432 }, { "epoch": 1.3832046332046333, "grad_norm": 4.203273296356201, "learning_rate": 1.1742184251559602e-05, "loss": 1.2121, "step": 1433 }, { "epoch": 1.3841698841698842, "grad_norm": 4.254453182220459, "learning_rate": 1.1731919377417318e-05, "loss": 1.2411, "step": 1434 }, { "epoch": 1.385135135135135, "grad_norm": 3.992889404296875, "learning_rate": 1.1721652621616208e-05, "loss": 1.1499, "step": 1435 }, { "epoch": 1.3861003861003862, "grad_norm": 4.445054054260254, "learning_rate": 1.171138399531068e-05, "loss": 1.2199, "step": 1436 }, { "epoch": 1.3870656370656371, "grad_norm": 4.198911666870117, "learning_rate": 1.1701113509657174e-05, "loss": 1.1976, "step": 1437 }, { "epoch": 1.388030888030888, "grad_norm": 4.214139938354492, "learning_rate": 1.1690841175814145e-05, "loss": 1.2287, "step": 1438 }, { "epoch": 1.388996138996139, "grad_norm": 4.2203369140625, "learning_rate": 1.1680567004942062e-05, "loss": 1.2446, "step": 1439 }, { "epoch": 1.3899613899613898, "grad_norm": 4.274831295013428, "learning_rate": 1.1670291008203383e-05, "loss": 1.2148, "step": 1440 }, { "epoch": 1.390926640926641, "grad_norm": 4.5262675285339355, "learning_rate": 1.1660013196762556e-05, "loss": 1.1929, "step": 1441 }, { "epoch": 1.3918918918918919, "grad_norm": 4.146603584289551, "learning_rate": 1.1649733581786001e-05, "loss": 1.2134, "step": 1442 }, { "epoch": 1.3928571428571428, "grad_norm": 4.600409984588623, "learning_rate": 1.163945217444209e-05, "loss": 1.1481, "step": 1443 }, { "epoch": 1.393822393822394, "grad_norm": 5.549932956695557, "learning_rate": 1.1629168985901149e-05, "loss": 1.1883, "step": 1444 }, { "epoch": 1.3947876447876448, "grad_norm": 4.512418746948242, "learning_rate": 1.1618884027335438e-05, "loss": 1.2328, "step": 1445 }, { "epoch": 1.3957528957528957, "grad_norm": 4.620266914367676, "learning_rate": 1.1608597309919141e-05, "loss": 1.2499, "step": 1446 }, { "epoch": 1.3967181467181466, "grad_norm": 4.4497528076171875, "learning_rate": 1.1598308844828348e-05, "loss": 1.2454, "step": 1447 }, { "epoch": 1.3976833976833976, "grad_norm": 4.136813163757324, "learning_rate": 1.1588018643241053e-05, "loss": 1.2569, "step": 1448 }, { "epoch": 1.3986486486486487, "grad_norm": 5.220927715301514, "learning_rate": 1.1577726716337134e-05, "loss": 1.2982, "step": 1449 }, { "epoch": 1.3996138996138996, "grad_norm": 4.751706123352051, "learning_rate": 1.1567433075298341e-05, "loss": 1.1566, "step": 1450 }, { "epoch": 1.4005791505791505, "grad_norm": 5.164363384246826, "learning_rate": 1.1557137731308302e-05, "loss": 1.1564, "step": 1451 }, { "epoch": 1.4015444015444016, "grad_norm": 4.481420040130615, "learning_rate": 1.1546840695552465e-05, "loss": 1.2552, "step": 1452 }, { "epoch": 1.4025096525096525, "grad_norm": 4.251123428344727, "learning_rate": 1.153654197921815e-05, "loss": 1.166, "step": 1453 }, { "epoch": 1.4034749034749034, "grad_norm": 5.288816928863525, "learning_rate": 1.1526241593494474e-05, "loss": 1.1846, "step": 1454 }, { "epoch": 1.4044401544401546, "grad_norm": 4.537790775299072, "learning_rate": 1.1515939549572389e-05, "loss": 1.2219, "step": 1455 }, { "epoch": 1.4054054054054055, "grad_norm": 4.063901901245117, "learning_rate": 1.1505635858644635e-05, "loss": 1.2029, "step": 1456 }, { "epoch": 1.4063706563706564, "grad_norm": 5.178050994873047, "learning_rate": 1.1495330531905749e-05, "loss": 1.1267, "step": 1457 }, { "epoch": 1.4073359073359073, "grad_norm": 4.601864337921143, "learning_rate": 1.1485023580552039e-05, "loss": 1.2514, "step": 1458 }, { "epoch": 1.4083011583011582, "grad_norm": 3.957143545150757, "learning_rate": 1.1474715015781584e-05, "loss": 1.194, "step": 1459 }, { "epoch": 1.4092664092664093, "grad_norm": 4.590023994445801, "learning_rate": 1.1464404848794218e-05, "loss": 1.2605, "step": 1460 }, { "epoch": 1.4102316602316602, "grad_norm": 4.938705921173096, "learning_rate": 1.1454093090791503e-05, "loss": 1.2199, "step": 1461 }, { "epoch": 1.4111969111969112, "grad_norm": 5.405674457550049, "learning_rate": 1.1443779752976745e-05, "loss": 1.1818, "step": 1462 }, { "epoch": 1.4121621621621623, "grad_norm": 6.0514817237854, "learning_rate": 1.1433464846554953e-05, "loss": 1.1749, "step": 1463 }, { "epoch": 1.4131274131274132, "grad_norm": 6.1509904861450195, "learning_rate": 1.1423148382732854e-05, "loss": 1.2601, "step": 1464 }, { "epoch": 1.414092664092664, "grad_norm": 6.008955001831055, "learning_rate": 1.1412830372718851e-05, "loss": 1.2095, "step": 1465 }, { "epoch": 1.415057915057915, "grad_norm": 7.071529388427734, "learning_rate": 1.1402510827723043e-05, "loss": 1.1539, "step": 1466 }, { "epoch": 1.416023166023166, "grad_norm": 4.7585954666137695, "learning_rate": 1.1392189758957183e-05, "loss": 1.1916, "step": 1467 }, { "epoch": 1.416988416988417, "grad_norm": 5.787749290466309, "learning_rate": 1.1381867177634693e-05, "loss": 1.1798, "step": 1468 }, { "epoch": 1.417953667953668, "grad_norm": 4.924343109130859, "learning_rate": 1.1371543094970624e-05, "loss": 1.2515, "step": 1469 }, { "epoch": 1.4189189189189189, "grad_norm": 4.688542366027832, "learning_rate": 1.1361217522181667e-05, "loss": 1.1731, "step": 1470 }, { "epoch": 1.41988416988417, "grad_norm": 5.041671276092529, "learning_rate": 1.135089047048613e-05, "loss": 1.2632, "step": 1471 }, { "epoch": 1.420849420849421, "grad_norm": 4.252134799957275, "learning_rate": 1.134056195110393e-05, "loss": 1.1982, "step": 1472 }, { "epoch": 1.4218146718146718, "grad_norm": 4.020081520080566, "learning_rate": 1.1330231975256576e-05, "loss": 1.2558, "step": 1473 }, { "epoch": 1.4227799227799227, "grad_norm": 5.2743024826049805, "learning_rate": 1.1319900554167158e-05, "loss": 1.228, "step": 1474 }, { "epoch": 1.4237451737451736, "grad_norm": 4.727425575256348, "learning_rate": 1.130956769906034e-05, "loss": 1.1643, "step": 1475 }, { "epoch": 1.4247104247104247, "grad_norm": 4.66657829284668, "learning_rate": 1.1299233421162339e-05, "loss": 1.2035, "step": 1476 }, { "epoch": 1.4256756756756757, "grad_norm": 4.683762550354004, "learning_rate": 1.1288897731700922e-05, "loss": 1.2453, "step": 1477 }, { "epoch": 1.4266409266409266, "grad_norm": 4.349846839904785, "learning_rate": 1.1278560641905392e-05, "loss": 1.2206, "step": 1478 }, { "epoch": 1.4276061776061777, "grad_norm": 4.372389793395996, "learning_rate": 1.1268222163006566e-05, "loss": 1.141, "step": 1479 }, { "epoch": 1.4285714285714286, "grad_norm": 4.824862957000732, "learning_rate": 1.1257882306236776e-05, "loss": 1.2172, "step": 1480 }, { "epoch": 1.4295366795366795, "grad_norm": 4.0703887939453125, "learning_rate": 1.1247541082829847e-05, "loss": 1.1931, "step": 1481 }, { "epoch": 1.4305019305019304, "grad_norm": 5.336752414703369, "learning_rate": 1.1237198504021091e-05, "loss": 1.2588, "step": 1482 }, { "epoch": 1.4314671814671815, "grad_norm": 4.761801719665527, "learning_rate": 1.1226854581047295e-05, "loss": 1.3055, "step": 1483 }, { "epoch": 1.4324324324324325, "grad_norm": 4.233315467834473, "learning_rate": 1.12165093251467e-05, "loss": 1.1884, "step": 1484 }, { "epoch": 1.4333976833976834, "grad_norm": 5.077378749847412, "learning_rate": 1.1206162747559001e-05, "loss": 1.1707, "step": 1485 }, { "epoch": 1.4343629343629343, "grad_norm": 4.529723644256592, "learning_rate": 1.1195814859525332e-05, "loss": 1.2067, "step": 1486 }, { "epoch": 1.4353281853281854, "grad_norm": 4.297308444976807, "learning_rate": 1.1185465672288237e-05, "loss": 1.1901, "step": 1487 }, { "epoch": 1.4362934362934363, "grad_norm": 4.214022636413574, "learning_rate": 1.1175115197091682e-05, "loss": 1.1859, "step": 1488 }, { "epoch": 1.4372586872586872, "grad_norm": 5.2057881355285645, "learning_rate": 1.1164763445181034e-05, "loss": 1.1926, "step": 1489 }, { "epoch": 1.4382239382239383, "grad_norm": 4.289044380187988, "learning_rate": 1.1154410427803042e-05, "loss": 1.1837, "step": 1490 }, { "epoch": 1.4391891891891893, "grad_norm": 5.051332473754883, "learning_rate": 1.1144056156205834e-05, "loss": 1.2201, "step": 1491 }, { "epoch": 1.4401544401544402, "grad_norm": 4.821202754974365, "learning_rate": 1.1133700641638892e-05, "loss": 1.1524, "step": 1492 }, { "epoch": 1.441119691119691, "grad_norm": 4.25753116607666, "learning_rate": 1.1123343895353063e-05, "loss": 1.199, "step": 1493 }, { "epoch": 1.442084942084942, "grad_norm": 4.807949542999268, "learning_rate": 1.1112985928600513e-05, "loss": 1.1852, "step": 1494 }, { "epoch": 1.443050193050193, "grad_norm": 4.423308849334717, "learning_rate": 1.1102626752634758e-05, "loss": 1.2182, "step": 1495 }, { "epoch": 1.444015444015444, "grad_norm": 4.054159641265869, "learning_rate": 1.1092266378710602e-05, "loss": 1.2056, "step": 1496 }, { "epoch": 1.444980694980695, "grad_norm": 4.658977508544922, "learning_rate": 1.1081904818084169e-05, "loss": 1.26, "step": 1497 }, { "epoch": 1.445945945945946, "grad_norm": 4.359392166137695, "learning_rate": 1.1071542082012865e-05, "loss": 1.2023, "step": 1498 }, { "epoch": 1.446911196911197, "grad_norm": 4.363769054412842, "learning_rate": 1.106117818175538e-05, "loss": 1.2128, "step": 1499 }, { "epoch": 1.4478764478764479, "grad_norm": 4.644874095916748, "learning_rate": 1.1050813128571654e-05, "loss": 1.2804, "step": 1500 }, { "epoch": 1.4488416988416988, "grad_norm": 3.9422566890716553, "learning_rate": 1.1040446933722894e-05, "loss": 1.1319, "step": 1501 }, { "epoch": 1.4498069498069497, "grad_norm": 4.872453689575195, "learning_rate": 1.1030079608471544e-05, "loss": 1.2062, "step": 1502 }, { "epoch": 1.4507722007722008, "grad_norm": 4.18463659286499, "learning_rate": 1.1019711164081272e-05, "loss": 1.223, "step": 1503 }, { "epoch": 1.4517374517374517, "grad_norm": 4.284641265869141, "learning_rate": 1.1009341611816963e-05, "loss": 1.1816, "step": 1504 }, { "epoch": 1.4527027027027026, "grad_norm": 4.588273525238037, "learning_rate": 1.099897096294471e-05, "loss": 1.2189, "step": 1505 }, { "epoch": 1.4536679536679538, "grad_norm": 4.396002292633057, "learning_rate": 1.0988599228731794e-05, "loss": 1.2246, "step": 1506 }, { "epoch": 1.4546332046332047, "grad_norm": 4.360711574554443, "learning_rate": 1.0978226420446671e-05, "loss": 1.1902, "step": 1507 }, { "epoch": 1.4555984555984556, "grad_norm": 4.676750183105469, "learning_rate": 1.0967852549358975e-05, "loss": 1.237, "step": 1508 }, { "epoch": 1.4565637065637065, "grad_norm": 5.116564750671387, "learning_rate": 1.0957477626739483e-05, "loss": 1.1937, "step": 1509 }, { "epoch": 1.4575289575289574, "grad_norm": 4.32762336730957, "learning_rate": 1.0947101663860123e-05, "loss": 1.1836, "step": 1510 }, { "epoch": 1.4584942084942085, "grad_norm": 4.377365589141846, "learning_rate": 1.0936724671993948e-05, "loss": 1.1516, "step": 1511 }, { "epoch": 1.4594594594594594, "grad_norm": 4.5395002365112305, "learning_rate": 1.092634666241513e-05, "loss": 1.1631, "step": 1512 }, { "epoch": 1.4604247104247103, "grad_norm": 4.971916198730469, "learning_rate": 1.091596764639895e-05, "loss": 1.2129, "step": 1513 }, { "epoch": 1.4613899613899615, "grad_norm": 4.18382453918457, "learning_rate": 1.0905587635221782e-05, "loss": 1.1798, "step": 1514 }, { "epoch": 1.4623552123552124, "grad_norm": 4.545231342315674, "learning_rate": 1.089520664016107e-05, "loss": 1.1676, "step": 1515 }, { "epoch": 1.4633204633204633, "grad_norm": 4.12399435043335, "learning_rate": 1.0884824672495346e-05, "loss": 1.194, "step": 1516 }, { "epoch": 1.4642857142857144, "grad_norm": 5.3278489112854, "learning_rate": 1.0874441743504182e-05, "loss": 1.2275, "step": 1517 }, { "epoch": 1.4652509652509653, "grad_norm": 4.356189727783203, "learning_rate": 1.086405786446821e-05, "loss": 1.1639, "step": 1518 }, { "epoch": 1.4662162162162162, "grad_norm": 4.2782793045043945, "learning_rate": 1.0853673046669071e-05, "loss": 1.2363, "step": 1519 }, { "epoch": 1.4671814671814671, "grad_norm": 4.346798896789551, "learning_rate": 1.0843287301389455e-05, "loss": 1.2136, "step": 1520 }, { "epoch": 1.468146718146718, "grad_norm": 4.449827194213867, "learning_rate": 1.0832900639913035e-05, "loss": 1.2552, "step": 1521 }, { "epoch": 1.4691119691119692, "grad_norm": 4.275735855102539, "learning_rate": 1.0822513073524498e-05, "loss": 1.1877, "step": 1522 }, { "epoch": 1.47007722007722, "grad_norm": 5.334815502166748, "learning_rate": 1.0812124613509496e-05, "loss": 1.2357, "step": 1523 }, { "epoch": 1.471042471042471, "grad_norm": 5.0692853927612305, "learning_rate": 1.080173527115467e-05, "loss": 1.2511, "step": 1524 }, { "epoch": 1.4720077220077221, "grad_norm": 5.489389896392822, "learning_rate": 1.0791345057747604e-05, "loss": 1.2368, "step": 1525 }, { "epoch": 1.472972972972973, "grad_norm": 4.0504469871521, "learning_rate": 1.0780953984576841e-05, "loss": 1.146, "step": 1526 }, { "epoch": 1.473938223938224, "grad_norm": 4.052753448486328, "learning_rate": 1.0770562062931849e-05, "loss": 1.1287, "step": 1527 }, { "epoch": 1.4749034749034748, "grad_norm": 5.282724380493164, "learning_rate": 1.0760169304103023e-05, "loss": 1.2229, "step": 1528 }, { "epoch": 1.4758687258687258, "grad_norm": 4.906439304351807, "learning_rate": 1.0749775719381665e-05, "loss": 1.2455, "step": 1529 }, { "epoch": 1.4768339768339769, "grad_norm": 5.4240031242370605, "learning_rate": 1.0739381320059978e-05, "loss": 1.1586, "step": 1530 }, { "epoch": 1.4777992277992278, "grad_norm": 4.4138875007629395, "learning_rate": 1.0728986117431044e-05, "loss": 1.2136, "step": 1531 }, { "epoch": 1.4787644787644787, "grad_norm": 4.2003350257873535, "learning_rate": 1.0718590122788822e-05, "loss": 1.2395, "step": 1532 }, { "epoch": 1.4797297297297298, "grad_norm": 4.415201187133789, "learning_rate": 1.0708193347428128e-05, "loss": 1.2035, "step": 1533 }, { "epoch": 1.4806949806949807, "grad_norm": 3.981343984603882, "learning_rate": 1.0697795802644635e-05, "loss": 1.1869, "step": 1534 }, { "epoch": 1.4816602316602316, "grad_norm": 4.455201148986816, "learning_rate": 1.0687397499734842e-05, "loss": 1.1693, "step": 1535 }, { "epoch": 1.4826254826254825, "grad_norm": 4.045246124267578, "learning_rate": 1.0676998449996075e-05, "loss": 1.2343, "step": 1536 }, { "epoch": 1.4835907335907335, "grad_norm": 4.094812393188477, "learning_rate": 1.0666598664726471e-05, "loss": 1.2074, "step": 1537 }, { "epoch": 1.4845559845559846, "grad_norm": 5.180531978607178, "learning_rate": 1.0656198155224968e-05, "loss": 1.2136, "step": 1538 }, { "epoch": 1.4855212355212355, "grad_norm": 4.258422374725342, "learning_rate": 1.064579693279129e-05, "loss": 1.1904, "step": 1539 }, { "epoch": 1.4864864864864864, "grad_norm": 4.394831657409668, "learning_rate": 1.0635395008725936e-05, "loss": 1.2433, "step": 1540 }, { "epoch": 1.4874517374517375, "grad_norm": 4.647140979766846, "learning_rate": 1.0624992394330165e-05, "loss": 1.1996, "step": 1541 }, { "epoch": 1.4884169884169884, "grad_norm": 4.8291521072387695, "learning_rate": 1.061458910090599e-05, "loss": 1.2498, "step": 1542 }, { "epoch": 1.4893822393822393, "grad_norm": 4.660562038421631, "learning_rate": 1.0604185139756155e-05, "loss": 1.2261, "step": 1543 }, { "epoch": 1.4903474903474905, "grad_norm": 4.097156524658203, "learning_rate": 1.0593780522184139e-05, "loss": 1.2191, "step": 1544 }, { "epoch": 1.4913127413127414, "grad_norm": 4.38627815246582, "learning_rate": 1.0583375259494125e-05, "loss": 1.2229, "step": 1545 }, { "epoch": 1.4922779922779923, "grad_norm": 4.089046001434326, "learning_rate": 1.0572969362991e-05, "loss": 1.2135, "step": 1546 }, { "epoch": 1.4932432432432432, "grad_norm": 4.13787317276001, "learning_rate": 1.0562562843980346e-05, "loss": 1.1843, "step": 1547 }, { "epoch": 1.494208494208494, "grad_norm": 4.3745832443237305, "learning_rate": 1.0552155713768409e-05, "loss": 1.2013, "step": 1548 }, { "epoch": 1.4951737451737452, "grad_norm": 4.971855640411377, "learning_rate": 1.0541747983662114e-05, "loss": 1.162, "step": 1549 }, { "epoch": 1.4961389961389961, "grad_norm": 4.968060493469238, "learning_rate": 1.0531339664969021e-05, "loss": 1.2469, "step": 1550 }, { "epoch": 1.497104247104247, "grad_norm": 4.692889213562012, "learning_rate": 1.0520930768997347e-05, "loss": 1.23, "step": 1551 }, { "epoch": 1.4980694980694982, "grad_norm": 4.44988489151001, "learning_rate": 1.0510521307055914e-05, "loss": 1.1305, "step": 1552 }, { "epoch": 1.499034749034749, "grad_norm": 4.482110023498535, "learning_rate": 1.0500111290454189e-05, "loss": 1.1734, "step": 1553 }, { "epoch": 1.5, "grad_norm": 4.157884120941162, "learning_rate": 1.0489700730502212e-05, "loss": 1.198, "step": 1554 }, { "epoch": 1.5009652509652511, "grad_norm": 4.152207851409912, "learning_rate": 1.047928963851063e-05, "loss": 1.205, "step": 1555 }, { "epoch": 1.5019305019305018, "grad_norm": 4.259824275970459, "learning_rate": 1.046887802579067e-05, "loss": 1.2947, "step": 1556 }, { "epoch": 1.502895752895753, "grad_norm": 4.260583877563477, "learning_rate": 1.0458465903654107e-05, "loss": 1.1686, "step": 1557 }, { "epoch": 1.5038610038610039, "grad_norm": 4.16087007522583, "learning_rate": 1.0448053283413293e-05, "loss": 1.1988, "step": 1558 }, { "epoch": 1.5048262548262548, "grad_norm": 3.9509360790252686, "learning_rate": 1.0437640176381099e-05, "loss": 1.1873, "step": 1559 }, { "epoch": 1.505791505791506, "grad_norm": 4.118860244750977, "learning_rate": 1.0427226593870947e-05, "loss": 1.1767, "step": 1560 }, { "epoch": 1.5067567567567568, "grad_norm": 4.576681137084961, "learning_rate": 1.0416812547196753e-05, "loss": 1.1545, "step": 1561 }, { "epoch": 1.5077220077220077, "grad_norm": 3.929964542388916, "learning_rate": 1.0406398047672956e-05, "loss": 1.209, "step": 1562 }, { "epoch": 1.5086872586872588, "grad_norm": 4.916624069213867, "learning_rate": 1.0395983106614474e-05, "loss": 1.1555, "step": 1563 }, { "epoch": 1.5096525096525095, "grad_norm": 4.800134181976318, "learning_rate": 1.0385567735336712e-05, "loss": 1.2684, "step": 1564 }, { "epoch": 1.5106177606177607, "grad_norm": 5.12754487991333, "learning_rate": 1.0375151945155544e-05, "loss": 1.1961, "step": 1565 }, { "epoch": 1.5115830115830116, "grad_norm": 4.375315189361572, "learning_rate": 1.0364735747387292e-05, "loss": 1.2706, "step": 1566 }, { "epoch": 1.5125482625482625, "grad_norm": 3.993865489959717, "learning_rate": 1.0354319153348724e-05, "loss": 1.2037, "step": 1567 }, { "epoch": 1.5135135135135136, "grad_norm": 4.187656402587891, "learning_rate": 1.034390217435704e-05, "loss": 1.2099, "step": 1568 }, { "epoch": 1.5144787644787645, "grad_norm": 4.1158447265625, "learning_rate": 1.0333484821729857e-05, "loss": 1.1932, "step": 1569 }, { "epoch": 1.5154440154440154, "grad_norm": 4.137932300567627, "learning_rate": 1.0323067106785199e-05, "loss": 1.2271, "step": 1570 }, { "epoch": 1.5164092664092665, "grad_norm": 3.928439140319824, "learning_rate": 1.031264904084148e-05, "loss": 1.1639, "step": 1571 }, { "epoch": 1.5173745173745172, "grad_norm": 4.347318172454834, "learning_rate": 1.03022306352175e-05, "loss": 1.2631, "step": 1572 }, { "epoch": 1.5183397683397684, "grad_norm": 4.589171886444092, "learning_rate": 1.0291811901232428e-05, "loss": 1.2417, "step": 1573 }, { "epoch": 1.5193050193050193, "grad_norm": 4.574836254119873, "learning_rate": 1.0281392850205783e-05, "loss": 1.1509, "step": 1574 }, { "epoch": 1.5202702702702702, "grad_norm": 4.554994106292725, "learning_rate": 1.0270973493457437e-05, "loss": 1.2337, "step": 1575 }, { "epoch": 1.5212355212355213, "grad_norm": 4.198336124420166, "learning_rate": 1.026055384230759e-05, "loss": 1.2079, "step": 1576 }, { "epoch": 1.5222007722007722, "grad_norm": 4.500290870666504, "learning_rate": 1.025013390807676e-05, "loss": 1.1258, "step": 1577 }, { "epoch": 1.5231660231660231, "grad_norm": 4.623536586761475, "learning_rate": 1.0239713702085781e-05, "loss": 1.2186, "step": 1578 }, { "epoch": 1.5241312741312742, "grad_norm": 4.773224830627441, "learning_rate": 1.0229293235655768e-05, "loss": 1.1922, "step": 1579 }, { "epoch": 1.525096525096525, "grad_norm": 3.701310396194458, "learning_rate": 1.0218872520108135e-05, "loss": 1.2281, "step": 1580 }, { "epoch": 1.526061776061776, "grad_norm": 4.196020603179932, "learning_rate": 1.020845156676455e-05, "loss": 1.1747, "step": 1581 }, { "epoch": 1.527027027027027, "grad_norm": 4.553520202636719, "learning_rate": 1.0198030386946956e-05, "loss": 1.3491, "step": 1582 }, { "epoch": 1.5279922779922779, "grad_norm": 4.003087997436523, "learning_rate": 1.0187608991977531e-05, "loss": 1.2327, "step": 1583 }, { "epoch": 1.528957528957529, "grad_norm": 5.1173882484436035, "learning_rate": 1.0177187393178687e-05, "loss": 1.178, "step": 1584 }, { "epoch": 1.52992277992278, "grad_norm": 4.88187313079834, "learning_rate": 1.0166765601873067e-05, "loss": 1.1399, "step": 1585 }, { "epoch": 1.5308880308880308, "grad_norm": 4.126231670379639, "learning_rate": 1.0156343629383509e-05, "loss": 1.2101, "step": 1586 }, { "epoch": 1.531853281853282, "grad_norm": 4.237364292144775, "learning_rate": 1.0145921487033065e-05, "loss": 1.2188, "step": 1587 }, { "epoch": 1.5328185328185329, "grad_norm": 3.97042179107666, "learning_rate": 1.0135499186144952e-05, "loss": 1.1823, "step": 1588 }, { "epoch": 1.5337837837837838, "grad_norm": 5.60618257522583, "learning_rate": 1.0125076738042577e-05, "loss": 1.2448, "step": 1589 }, { "epoch": 1.534749034749035, "grad_norm": 4.790480136871338, "learning_rate": 1.011465415404949e-05, "loss": 1.1589, "step": 1590 }, { "epoch": 1.5357142857142856, "grad_norm": 4.756139278411865, "learning_rate": 1.010423144548941e-05, "loss": 1.1934, "step": 1591 }, { "epoch": 1.5366795366795367, "grad_norm": 5.73478889465332, "learning_rate": 1.0093808623686165e-05, "loss": 1.1871, "step": 1592 }, { "epoch": 1.5376447876447876, "grad_norm": 3.9583542346954346, "learning_rate": 1.0083385699963733e-05, "loss": 1.2149, "step": 1593 }, { "epoch": 1.5386100386100385, "grad_norm": 4.312107086181641, "learning_rate": 1.0072962685646182e-05, "loss": 1.1849, "step": 1594 }, { "epoch": 1.5395752895752897, "grad_norm": 5.653817653656006, "learning_rate": 1.0062539592057689e-05, "loss": 1.2231, "step": 1595 }, { "epoch": 1.5405405405405406, "grad_norm": 4.556292533874512, "learning_rate": 1.0052116430522514e-05, "loss": 1.2432, "step": 1596 }, { "epoch": 1.5415057915057915, "grad_norm": 5.225686550140381, "learning_rate": 1.0041693212364992e-05, "loss": 1.1649, "step": 1597 }, { "epoch": 1.5424710424710426, "grad_norm": 5.1191606521606445, "learning_rate": 1.0031269948909517e-05, "loss": 1.1819, "step": 1598 }, { "epoch": 1.5434362934362933, "grad_norm": 4.36271333694458, "learning_rate": 1.0020846651480534e-05, "loss": 1.2678, "step": 1599 }, { "epoch": 1.5444015444015444, "grad_norm": 4.28533411026001, "learning_rate": 1.0010423331402528e-05, "loss": 1.1961, "step": 1600 }, { "epoch": 1.5453667953667953, "grad_norm": 6.018980503082275, "learning_rate": 1e-05, "loss": 1.187, "step": 1601 }, { "epoch": 1.5463320463320462, "grad_norm": 4.648509502410889, "learning_rate": 9.989576668597476e-06, "loss": 1.1882, "step": 1602 }, { "epoch": 1.5472972972972974, "grad_norm": 6.939119815826416, "learning_rate": 9.979153348519466e-06, "loss": 1.2215, "step": 1603 }, { "epoch": 1.5482625482625483, "grad_norm": 6.063055992126465, "learning_rate": 9.968730051090487e-06, "loss": 1.2178, "step": 1604 }, { "epoch": 1.5492277992277992, "grad_norm": 5.186483860015869, "learning_rate": 9.958306787635011e-06, "loss": 1.3228, "step": 1605 }, { "epoch": 1.5501930501930503, "grad_norm": 5.201595783233643, "learning_rate": 9.94788356947749e-06, "loss": 1.1799, "step": 1606 }, { "epoch": 1.551158301158301, "grad_norm": 4.202732563018799, "learning_rate": 9.937460407942312e-06, "loss": 1.1754, "step": 1607 }, { "epoch": 1.5521235521235521, "grad_norm": 4.17235803604126, "learning_rate": 9.927037314353821e-06, "loss": 1.1695, "step": 1608 }, { "epoch": 1.553088803088803, "grad_norm": 4.822647571563721, "learning_rate": 9.91661430003627e-06, "loss": 1.2279, "step": 1609 }, { "epoch": 1.554054054054054, "grad_norm": 4.211466312408447, "learning_rate": 9.906191376313837e-06, "loss": 1.1714, "step": 1610 }, { "epoch": 1.555019305019305, "grad_norm": 4.621142387390137, "learning_rate": 9.895768554510593e-06, "loss": 1.1973, "step": 1611 }, { "epoch": 1.555984555984556, "grad_norm": 5.649501323699951, "learning_rate": 9.88534584595051e-06, "loss": 1.1692, "step": 1612 }, { "epoch": 1.556949806949807, "grad_norm": 4.738320350646973, "learning_rate": 9.874923261957428e-06, "loss": 1.1858, "step": 1613 }, { "epoch": 1.557915057915058, "grad_norm": 4.468520641326904, "learning_rate": 9.864500813855051e-06, "loss": 1.2527, "step": 1614 }, { "epoch": 1.558880308880309, "grad_norm": 5.1193037033081055, "learning_rate": 9.854078512966937e-06, "loss": 1.1502, "step": 1615 }, { "epoch": 1.5598455598455598, "grad_norm": 5.412008285522461, "learning_rate": 9.84365637061649e-06, "loss": 1.3047, "step": 1616 }, { "epoch": 1.560810810810811, "grad_norm": 5.719449996948242, "learning_rate": 9.833234398126938e-06, "loss": 1.2183, "step": 1617 }, { "epoch": 1.5617760617760617, "grad_norm": 4.692526817321777, "learning_rate": 9.822812606821317e-06, "loss": 1.1853, "step": 1618 }, { "epoch": 1.5627413127413128, "grad_norm": 4.056661605834961, "learning_rate": 9.812391008022472e-06, "loss": 1.1964, "step": 1619 }, { "epoch": 1.5637065637065637, "grad_norm": 5.884062767028809, "learning_rate": 9.801969613053047e-06, "loss": 1.1945, "step": 1620 }, { "epoch": 1.5646718146718146, "grad_norm": 4.1736626625061035, "learning_rate": 9.791548433235452e-06, "loss": 1.2004, "step": 1621 }, { "epoch": 1.5656370656370657, "grad_norm": 4.439320087432861, "learning_rate": 9.781127479891872e-06, "loss": 1.1776, "step": 1622 }, { "epoch": 1.5666023166023166, "grad_norm": 4.904225826263428, "learning_rate": 9.770706764344235e-06, "loss": 1.1812, "step": 1623 }, { "epoch": 1.5675675675675675, "grad_norm": 4.182514667510986, "learning_rate": 9.760286297914224e-06, "loss": 1.2016, "step": 1624 }, { "epoch": 1.5685328185328187, "grad_norm": 4.227596282958984, "learning_rate": 9.749866091923241e-06, "loss": 1.2117, "step": 1625 }, { "epoch": 1.5694980694980694, "grad_norm": 4.459351062774658, "learning_rate": 9.739446157692416e-06, "loss": 1.1541, "step": 1626 }, { "epoch": 1.5704633204633205, "grad_norm": 4.430898666381836, "learning_rate": 9.729026506542566e-06, "loss": 1.1422, "step": 1627 }, { "epoch": 1.5714285714285714, "grad_norm": 4.316422939300537, "learning_rate": 9.71860714979422e-06, "loss": 1.1945, "step": 1628 }, { "epoch": 1.5723938223938223, "grad_norm": 4.788571357727051, "learning_rate": 9.708188098767577e-06, "loss": 1.1744, "step": 1629 }, { "epoch": 1.5733590733590734, "grad_norm": 4.579430103302002, "learning_rate": 9.697769364782502e-06, "loss": 1.2298, "step": 1630 }, { "epoch": 1.5743243243243243, "grad_norm": 4.8498687744140625, "learning_rate": 9.687350959158525e-06, "loss": 1.2028, "step": 1631 }, { "epoch": 1.5752895752895753, "grad_norm": 4.816117763519287, "learning_rate": 9.676932893214806e-06, "loss": 1.1988, "step": 1632 }, { "epoch": 1.5762548262548264, "grad_norm": 4.2359514236450195, "learning_rate": 9.666515178270146e-06, "loss": 1.1664, "step": 1633 }, { "epoch": 1.577220077220077, "grad_norm": 4.571994781494141, "learning_rate": 9.65609782564296e-06, "loss": 1.1995, "step": 1634 }, { "epoch": 1.5781853281853282, "grad_norm": 4.250718116760254, "learning_rate": 9.645680846651281e-06, "loss": 1.1907, "step": 1635 }, { "epoch": 1.579150579150579, "grad_norm": 4.135900497436523, "learning_rate": 9.635264252612712e-06, "loss": 1.1788, "step": 1636 }, { "epoch": 1.58011583011583, "grad_norm": 4.255411148071289, "learning_rate": 9.62484805484446e-06, "loss": 1.2615, "step": 1637 }, { "epoch": 1.5810810810810811, "grad_norm": 4.202883243560791, "learning_rate": 9.614432264663288e-06, "loss": 1.2089, "step": 1638 }, { "epoch": 1.582046332046332, "grad_norm": 4.191743850708008, "learning_rate": 9.604016893385528e-06, "loss": 1.2093, "step": 1639 }, { "epoch": 1.583011583011583, "grad_norm": 5.0460686683654785, "learning_rate": 9.593601952327049e-06, "loss": 1.168, "step": 1640 }, { "epoch": 1.583976833976834, "grad_norm": 4.46348237991333, "learning_rate": 9.58318745280325e-06, "loss": 1.188, "step": 1641 }, { "epoch": 1.584942084942085, "grad_norm": 5.175261497497559, "learning_rate": 9.572773406129056e-06, "loss": 1.185, "step": 1642 }, { "epoch": 1.585907335907336, "grad_norm": 4.250487804412842, "learning_rate": 9.562359823618901e-06, "loss": 1.167, "step": 1643 }, { "epoch": 1.586872586872587, "grad_norm": 4.666817665100098, "learning_rate": 9.551946716586712e-06, "loss": 1.1994, "step": 1644 }, { "epoch": 1.5878378378378377, "grad_norm": 6.0936713218688965, "learning_rate": 9.541534096345896e-06, "loss": 1.2028, "step": 1645 }, { "epoch": 1.5888030888030888, "grad_norm": 4.434663772583008, "learning_rate": 9.531121974209334e-06, "loss": 1.2667, "step": 1646 }, { "epoch": 1.5897683397683398, "grad_norm": 5.738936901092529, "learning_rate": 9.52071036148937e-06, "loss": 1.192, "step": 1647 }, { "epoch": 1.5907335907335907, "grad_norm": 4.914812088012695, "learning_rate": 9.510299269497791e-06, "loss": 1.2331, "step": 1648 }, { "epoch": 1.5916988416988418, "grad_norm": 4.0906901359558105, "learning_rate": 9.499888709545816e-06, "loss": 1.2696, "step": 1649 }, { "epoch": 1.5926640926640927, "grad_norm": 5.1826324462890625, "learning_rate": 9.489478692944087e-06, "loss": 1.1439, "step": 1650 }, { "epoch": 1.5936293436293436, "grad_norm": 4.445993423461914, "learning_rate": 9.479069231002658e-06, "loss": 1.1957, "step": 1651 }, { "epoch": 1.5945945945945947, "grad_norm": 4.667313098907471, "learning_rate": 9.46866033503098e-06, "loss": 1.2407, "step": 1652 }, { "epoch": 1.5955598455598454, "grad_norm": 4.472392559051514, "learning_rate": 9.458252016337891e-06, "loss": 1.2802, "step": 1653 }, { "epoch": 1.5965250965250966, "grad_norm": 4.763835906982422, "learning_rate": 9.447844286231594e-06, "loss": 1.2293, "step": 1654 }, { "epoch": 1.5974903474903475, "grad_norm": 3.701159715652466, "learning_rate": 9.437437156019657e-06, "loss": 1.1311, "step": 1655 }, { "epoch": 1.5984555984555984, "grad_norm": 5.122166633605957, "learning_rate": 9.427030637009002e-06, "loss": 1.1632, "step": 1656 }, { "epoch": 1.5994208494208495, "grad_norm": 4.4899678230285645, "learning_rate": 9.416624740505882e-06, "loss": 1.2298, "step": 1657 }, { "epoch": 1.6003861003861004, "grad_norm": 4.307829856872559, "learning_rate": 9.406219477815868e-06, "loss": 1.2088, "step": 1658 }, { "epoch": 1.6013513513513513, "grad_norm": 4.44758939743042, "learning_rate": 9.395814860243848e-06, "loss": 1.204, "step": 1659 }, { "epoch": 1.6023166023166024, "grad_norm": 4.317739963531494, "learning_rate": 9.385410899094014e-06, "loss": 1.2388, "step": 1660 }, { "epoch": 1.6032818532818531, "grad_norm": 4.105050563812256, "learning_rate": 9.375007605669835e-06, "loss": 1.2187, "step": 1661 }, { "epoch": 1.6042471042471043, "grad_norm": 4.454011917114258, "learning_rate": 9.364604991274069e-06, "loss": 1.2077, "step": 1662 }, { "epoch": 1.6052123552123552, "grad_norm": 4.6989426612854, "learning_rate": 9.354203067208711e-06, "loss": 1.1345, "step": 1663 }, { "epoch": 1.606177606177606, "grad_norm": 4.412756443023682, "learning_rate": 9.343801844775035e-06, "loss": 1.198, "step": 1664 }, { "epoch": 1.6071428571428572, "grad_norm": 4.618637561798096, "learning_rate": 9.33340133527353e-06, "loss": 1.2647, "step": 1665 }, { "epoch": 1.6081081081081081, "grad_norm": 4.564658164978027, "learning_rate": 9.32300155000393e-06, "loss": 1.224, "step": 1666 }, { "epoch": 1.609073359073359, "grad_norm": 4.380929470062256, "learning_rate": 9.312602500265162e-06, "loss": 1.1752, "step": 1667 }, { "epoch": 1.6100386100386102, "grad_norm": 4.754830360412598, "learning_rate": 9.302204197355368e-06, "loss": 1.2209, "step": 1668 }, { "epoch": 1.6110038610038608, "grad_norm": 3.8590304851531982, "learning_rate": 9.291806652571873e-06, "loss": 1.171, "step": 1669 }, { "epoch": 1.611969111969112, "grad_norm": 4.317570209503174, "learning_rate": 9.28140987721118e-06, "loss": 1.136, "step": 1670 }, { "epoch": 1.612934362934363, "grad_norm": 4.725559234619141, "learning_rate": 9.27101388256896e-06, "loss": 1.1823, "step": 1671 }, { "epoch": 1.6138996138996138, "grad_norm": 4.458336353302002, "learning_rate": 9.260618679940025e-06, "loss": 1.2175, "step": 1672 }, { "epoch": 1.614864864864865, "grad_norm": 4.144418239593506, "learning_rate": 9.250224280618337e-06, "loss": 1.2245, "step": 1673 }, { "epoch": 1.6158301158301158, "grad_norm": 4.354918956756592, "learning_rate": 9.239830695896978e-06, "loss": 1.1639, "step": 1674 }, { "epoch": 1.6167953667953667, "grad_norm": 4.767379283905029, "learning_rate": 9.229437937068155e-06, "loss": 1.2238, "step": 1675 }, { "epoch": 1.6177606177606179, "grad_norm": 4.249136924743652, "learning_rate": 9.219046015423162e-06, "loss": 1.209, "step": 1676 }, { "epoch": 1.6187258687258688, "grad_norm": 4.206948280334473, "learning_rate": 9.2086549422524e-06, "loss": 1.1854, "step": 1677 }, { "epoch": 1.6196911196911197, "grad_norm": 4.812625408172607, "learning_rate": 9.198264728845332e-06, "loss": 1.174, "step": 1678 }, { "epoch": 1.6206563706563708, "grad_norm": 4.244357109069824, "learning_rate": 9.187875386490504e-06, "loss": 1.1699, "step": 1679 }, { "epoch": 1.6216216216216215, "grad_norm": 5.580759048461914, "learning_rate": 9.177486926475507e-06, "loss": 1.1849, "step": 1680 }, { "epoch": 1.6225868725868726, "grad_norm": 4.3491692543029785, "learning_rate": 9.167099360086967e-06, "loss": 1.1835, "step": 1681 }, { "epoch": 1.6235521235521235, "grad_norm": 5.0731940269470215, "learning_rate": 9.156712698610546e-06, "loss": 1.1614, "step": 1682 }, { "epoch": 1.6245173745173744, "grad_norm": 4.256218910217285, "learning_rate": 9.146326953330929e-06, "loss": 1.2258, "step": 1683 }, { "epoch": 1.6254826254826256, "grad_norm": 4.389143943786621, "learning_rate": 9.135942135531797e-06, "loss": 1.2423, "step": 1684 }, { "epoch": 1.6264478764478765, "grad_norm": 5.3131561279296875, "learning_rate": 9.12555825649582e-06, "loss": 1.208, "step": 1685 }, { "epoch": 1.6274131274131274, "grad_norm": 4.0794548988342285, "learning_rate": 9.115175327504656e-06, "loss": 1.2044, "step": 1686 }, { "epoch": 1.6283783783783785, "grad_norm": 5.013195037841797, "learning_rate": 9.104793359838931e-06, "loss": 1.2365, "step": 1687 }, { "epoch": 1.6293436293436292, "grad_norm": 3.932877779006958, "learning_rate": 9.094412364778222e-06, "loss": 1.2031, "step": 1688 }, { "epoch": 1.6303088803088803, "grad_norm": 4.152234077453613, "learning_rate": 9.084032353601053e-06, "loss": 1.1921, "step": 1689 }, { "epoch": 1.6312741312741312, "grad_norm": 4.233832836151123, "learning_rate": 9.073653337584872e-06, "loss": 1.2151, "step": 1690 }, { "epoch": 1.6322393822393821, "grad_norm": 4.629663467407227, "learning_rate": 9.063275328006055e-06, "loss": 1.1999, "step": 1691 }, { "epoch": 1.6332046332046333, "grad_norm": 4.575491905212402, "learning_rate": 9.05289833613988e-06, "loss": 1.183, "step": 1692 }, { "epoch": 1.6341698841698842, "grad_norm": 4.473489284515381, "learning_rate": 9.042522373260522e-06, "loss": 1.2094, "step": 1693 }, { "epoch": 1.635135135135135, "grad_norm": 4.254086494445801, "learning_rate": 9.03214745064103e-06, "loss": 1.2092, "step": 1694 }, { "epoch": 1.6361003861003862, "grad_norm": 4.169577598571777, "learning_rate": 9.02177357955333e-06, "loss": 1.1868, "step": 1695 }, { "epoch": 1.637065637065637, "grad_norm": 4.918846607208252, "learning_rate": 9.01140077126821e-06, "loss": 1.214, "step": 1696 }, { "epoch": 1.638030888030888, "grad_norm": 4.24448823928833, "learning_rate": 9.00102903705529e-06, "loss": 1.1783, "step": 1697 }, { "epoch": 1.6389961389961392, "grad_norm": 4.124368667602539, "learning_rate": 8.990658388183038e-06, "loss": 1.148, "step": 1698 }, { "epoch": 1.6399613899613898, "grad_norm": 4.327722072601318, "learning_rate": 8.98028883591873e-06, "loss": 1.2095, "step": 1699 }, { "epoch": 1.640926640926641, "grad_norm": 4.191107749938965, "learning_rate": 8.969920391528459e-06, "loss": 1.1779, "step": 1700 }, { "epoch": 1.6418918918918919, "grad_norm": 4.929723739624023, "learning_rate": 8.959553066277104e-06, "loss": 1.1678, "step": 1701 }, { "epoch": 1.6428571428571428, "grad_norm": 4.273260593414307, "learning_rate": 8.949186871428347e-06, "loss": 1.1998, "step": 1702 }, { "epoch": 1.643822393822394, "grad_norm": 5.487876892089844, "learning_rate": 8.938821818244626e-06, "loss": 1.2464, "step": 1703 }, { "epoch": 1.6447876447876448, "grad_norm": 3.923842191696167, "learning_rate": 8.928457917987137e-06, "loss": 1.2016, "step": 1704 }, { "epoch": 1.6457528957528957, "grad_norm": 4.900607585906982, "learning_rate": 8.918095181915833e-06, "loss": 1.2306, "step": 1705 }, { "epoch": 1.6467181467181469, "grad_norm": 5.042302131652832, "learning_rate": 8.9077336212894e-06, "loss": 1.1821, "step": 1706 }, { "epoch": 1.6476833976833976, "grad_norm": 4.501278400421143, "learning_rate": 8.897373247365249e-06, "loss": 1.2297, "step": 1707 }, { "epoch": 1.6486486486486487, "grad_norm": 4.577908039093018, "learning_rate": 8.887014071399488e-06, "loss": 1.231, "step": 1708 }, { "epoch": 1.6496138996138996, "grad_norm": 5.391479015350342, "learning_rate": 8.876656104646942e-06, "loss": 1.2098, "step": 1709 }, { "epoch": 1.6505791505791505, "grad_norm": 4.688028812408447, "learning_rate": 8.86629935836111e-06, "loss": 1.1722, "step": 1710 }, { "epoch": 1.6515444015444016, "grad_norm": 5.170119762420654, "learning_rate": 8.855943843794171e-06, "loss": 1.2354, "step": 1711 }, { "epoch": 1.6525096525096525, "grad_norm": 4.34480619430542, "learning_rate": 8.84558957219696e-06, "loss": 1.1562, "step": 1712 }, { "epoch": 1.6534749034749034, "grad_norm": 4.477398872375488, "learning_rate": 8.835236554818967e-06, "loss": 1.2139, "step": 1713 }, { "epoch": 1.6544401544401546, "grad_norm": 4.357746124267578, "learning_rate": 8.82488480290832e-06, "loss": 1.1659, "step": 1714 }, { "epoch": 1.6554054054054053, "grad_norm": 4.401113033294678, "learning_rate": 8.814534327711768e-06, "loss": 1.2111, "step": 1715 }, { "epoch": 1.6563706563706564, "grad_norm": 4.574923992156982, "learning_rate": 8.804185140474675e-06, "loss": 1.1184, "step": 1716 }, { "epoch": 1.6573359073359073, "grad_norm": 4.733505725860596, "learning_rate": 8.793837252441e-06, "loss": 1.182, "step": 1717 }, { "epoch": 1.6583011583011582, "grad_norm": 4.71107816696167, "learning_rate": 8.783490674853302e-06, "loss": 1.2215, "step": 1718 }, { "epoch": 1.6592664092664093, "grad_norm": 5.102229595184326, "learning_rate": 8.773145418952708e-06, "loss": 1.1866, "step": 1719 }, { "epoch": 1.6602316602316602, "grad_norm": 4.5466132164001465, "learning_rate": 8.762801495978914e-06, "loss": 1.2404, "step": 1720 }, { "epoch": 1.6611969111969112, "grad_norm": 4.194196701049805, "learning_rate": 8.752458917170157e-06, "loss": 1.1256, "step": 1721 }, { "epoch": 1.6621621621621623, "grad_norm": 5.514886379241943, "learning_rate": 8.742117693763229e-06, "loss": 1.2062, "step": 1722 }, { "epoch": 1.663127413127413, "grad_norm": 4.242153167724609, "learning_rate": 8.731777836993437e-06, "loss": 1.1739, "step": 1723 }, { "epoch": 1.664092664092664, "grad_norm": 4.451455116271973, "learning_rate": 8.721439358094613e-06, "loss": 1.2389, "step": 1724 }, { "epoch": 1.665057915057915, "grad_norm": 5.862081527709961, "learning_rate": 8.71110226829908e-06, "loss": 1.2433, "step": 1725 }, { "epoch": 1.666023166023166, "grad_norm": 4.439873695373535, "learning_rate": 8.700766578837664e-06, "loss": 1.1859, "step": 1726 }, { "epoch": 1.666988416988417, "grad_norm": 4.343622207641602, "learning_rate": 8.690432300939664e-06, "loss": 1.1274, "step": 1727 }, { "epoch": 1.667953667953668, "grad_norm": 4.739264011383057, "learning_rate": 8.680099445832843e-06, "loss": 1.149, "step": 1728 }, { "epoch": 1.6689189189189189, "grad_norm": 4.081526756286621, "learning_rate": 8.669768024743429e-06, "loss": 1.1458, "step": 1729 }, { "epoch": 1.66988416988417, "grad_norm": 4.270155906677246, "learning_rate": 8.659438048896072e-06, "loss": 1.2105, "step": 1730 }, { "epoch": 1.670849420849421, "grad_norm": 4.286427021026611, "learning_rate": 8.649109529513873e-06, "loss": 1.1492, "step": 1731 }, { "epoch": 1.6718146718146718, "grad_norm": 4.289936065673828, "learning_rate": 8.638782477818335e-06, "loss": 1.1017, "step": 1732 }, { "epoch": 1.672779922779923, "grad_norm": 4.637660980224609, "learning_rate": 8.628456905029383e-06, "loss": 1.2084, "step": 1733 }, { "epoch": 1.6737451737451736, "grad_norm": 4.21026086807251, "learning_rate": 8.618132822365312e-06, "loss": 1.2537, "step": 1734 }, { "epoch": 1.6747104247104247, "grad_norm": 3.7908935546875, "learning_rate": 8.607810241042819e-06, "loss": 1.1979, "step": 1735 }, { "epoch": 1.6756756756756757, "grad_norm": 4.276378154754639, "learning_rate": 8.597489172276959e-06, "loss": 1.1713, "step": 1736 }, { "epoch": 1.6766409266409266, "grad_norm": 4.769236087799072, "learning_rate": 8.587169627281149e-06, "loss": 1.204, "step": 1737 }, { "epoch": 1.6776061776061777, "grad_norm": 4.216571807861328, "learning_rate": 8.576851617267151e-06, "loss": 1.2026, "step": 1738 }, { "epoch": 1.6785714285714286, "grad_norm": 4.589860916137695, "learning_rate": 8.566535153445049e-06, "loss": 1.2139, "step": 1739 }, { "epoch": 1.6795366795366795, "grad_norm": 4.3696722984313965, "learning_rate": 8.556220247023257e-06, "loss": 1.1702, "step": 1740 }, { "epoch": 1.6805019305019306, "grad_norm": 4.10421085357666, "learning_rate": 8.545906909208498e-06, "loss": 1.1449, "step": 1741 }, { "epoch": 1.6814671814671813, "grad_norm": 4.313359260559082, "learning_rate": 8.535595151205787e-06, "loss": 1.2174, "step": 1742 }, { "epoch": 1.6824324324324325, "grad_norm": 4.988412857055664, "learning_rate": 8.525284984218418e-06, "loss": 1.2489, "step": 1743 }, { "epoch": 1.6833976833976834, "grad_norm": 5.241071701049805, "learning_rate": 8.514976419447963e-06, "loss": 1.2148, "step": 1744 }, { "epoch": 1.6843629343629343, "grad_norm": 5.049296855926514, "learning_rate": 8.504669468094256e-06, "loss": 1.1645, "step": 1745 }, { "epoch": 1.6853281853281854, "grad_norm": 5.042121887207031, "learning_rate": 8.494364141355368e-06, "loss": 1.1522, "step": 1746 }, { "epoch": 1.6862934362934363, "grad_norm": 4.717390060424805, "learning_rate": 8.484060450427616e-06, "loss": 1.1816, "step": 1747 }, { "epoch": 1.6872586872586872, "grad_norm": 4.598116874694824, "learning_rate": 8.47375840650553e-06, "loss": 1.2413, "step": 1748 }, { "epoch": 1.6882239382239383, "grad_norm": 4.408492088317871, "learning_rate": 8.463458020781855e-06, "loss": 1.1495, "step": 1749 }, { "epoch": 1.689189189189189, "grad_norm": 4.6580586433410645, "learning_rate": 8.453159304447535e-06, "loss": 1.2852, "step": 1750 }, { "epoch": 1.6901544401544402, "grad_norm": 4.525180816650391, "learning_rate": 8.442862268691703e-06, "loss": 1.1859, "step": 1751 }, { "epoch": 1.691119691119691, "grad_norm": 4.853588104248047, "learning_rate": 8.43256692470166e-06, "loss": 1.2238, "step": 1752 }, { "epoch": 1.692084942084942, "grad_norm": 5.207756519317627, "learning_rate": 8.422273283662869e-06, "loss": 1.1487, "step": 1753 }, { "epoch": 1.693050193050193, "grad_norm": 4.643857002258301, "learning_rate": 8.41198135675895e-06, "loss": 1.1702, "step": 1754 }, { "epoch": 1.694015444015444, "grad_norm": 4.395835876464844, "learning_rate": 8.401691155171654e-06, "loss": 1.2509, "step": 1755 }, { "epoch": 1.694980694980695, "grad_norm": 4.321347713470459, "learning_rate": 8.391402690080864e-06, "loss": 1.1961, "step": 1756 }, { "epoch": 1.695945945945946, "grad_norm": 4.597218990325928, "learning_rate": 8.381115972664563e-06, "loss": 1.2019, "step": 1757 }, { "epoch": 1.696911196911197, "grad_norm": 4.090878486633301, "learning_rate": 8.370831014098854e-06, "loss": 1.2323, "step": 1758 }, { "epoch": 1.6978764478764479, "grad_norm": 4.209794521331787, "learning_rate": 8.360547825557912e-06, "loss": 1.1784, "step": 1759 }, { "epoch": 1.698841698841699, "grad_norm": 4.162314414978027, "learning_rate": 8.350266418214006e-06, "loss": 1.1674, "step": 1760 }, { "epoch": 1.6998069498069497, "grad_norm": 3.905789852142334, "learning_rate": 8.339986803237446e-06, "loss": 1.1917, "step": 1761 }, { "epoch": 1.7007722007722008, "grad_norm": 4.145606994628906, "learning_rate": 8.32970899179662e-06, "loss": 1.2348, "step": 1762 }, { "epoch": 1.7017374517374517, "grad_norm": 3.782750368118286, "learning_rate": 8.31943299505794e-06, "loss": 1.2221, "step": 1763 }, { "epoch": 1.7027027027027026, "grad_norm": 4.163419723510742, "learning_rate": 8.309158824185855e-06, "loss": 1.1523, "step": 1764 }, { "epoch": 1.7036679536679538, "grad_norm": 4.2257771492004395, "learning_rate": 8.298886490342827e-06, "loss": 1.1298, "step": 1765 }, { "epoch": 1.7046332046332047, "grad_norm": 4.363615989685059, "learning_rate": 8.288616004689321e-06, "loss": 1.1604, "step": 1766 }, { "epoch": 1.7055984555984556, "grad_norm": 4.496586799621582, "learning_rate": 8.278347378383794e-06, "loss": 1.2354, "step": 1767 }, { "epoch": 1.7065637065637067, "grad_norm": 4.663826942443848, "learning_rate": 8.268080622582682e-06, "loss": 1.221, "step": 1768 }, { "epoch": 1.7075289575289574, "grad_norm": 4.14992094039917, "learning_rate": 8.2578157484404e-06, "loss": 1.2565, "step": 1769 }, { "epoch": 1.7084942084942085, "grad_norm": 3.9128541946411133, "learning_rate": 8.247552767109301e-06, "loss": 1.1902, "step": 1770 }, { "epoch": 1.7094594594594594, "grad_norm": 4.58083438873291, "learning_rate": 8.237291689739692e-06, "loss": 1.2451, "step": 1771 }, { "epoch": 1.7104247104247103, "grad_norm": 4.241626262664795, "learning_rate": 8.227032527479807e-06, "loss": 1.1586, "step": 1772 }, { "epoch": 1.7113899613899615, "grad_norm": 5.163703441619873, "learning_rate": 8.216775291475808e-06, "loss": 1.1513, "step": 1773 }, { "epoch": 1.7123552123552124, "grad_norm": 3.928961992263794, "learning_rate": 8.206519992871757e-06, "loss": 1.2216, "step": 1774 }, { "epoch": 1.7133204633204633, "grad_norm": 4.547893047332764, "learning_rate": 8.196266642809607e-06, "loss": 1.2212, "step": 1775 }, { "epoch": 1.7142857142857144, "grad_norm": 4.903558254241943, "learning_rate": 8.1860152524292e-06, "loss": 1.2437, "step": 1776 }, { "epoch": 1.715250965250965, "grad_norm": 4.899398326873779, "learning_rate": 8.175765832868252e-06, "loss": 1.2081, "step": 1777 }, { "epoch": 1.7162162162162162, "grad_norm": 4.3335442543029785, "learning_rate": 8.165518395262335e-06, "loss": 1.1931, "step": 1778 }, { "epoch": 1.7171814671814671, "grad_norm": 4.198944091796875, "learning_rate": 8.155272950744859e-06, "loss": 1.1742, "step": 1779 }, { "epoch": 1.718146718146718, "grad_norm": 4.206162452697754, "learning_rate": 8.14502951044708e-06, "loss": 1.1192, "step": 1780 }, { "epoch": 1.7191119691119692, "grad_norm": 4.380505561828613, "learning_rate": 8.134788085498073e-06, "loss": 1.1848, "step": 1781 }, { "epoch": 1.72007722007722, "grad_norm": 4.550119400024414, "learning_rate": 8.124548687024723e-06, "loss": 1.2425, "step": 1782 }, { "epoch": 1.721042471042471, "grad_norm": 4.7923479080200195, "learning_rate": 8.114311326151716e-06, "loss": 1.2471, "step": 1783 }, { "epoch": 1.7220077220077221, "grad_norm": 6.3003058433532715, "learning_rate": 8.104076014001514e-06, "loss": 1.1869, "step": 1784 }, { "epoch": 1.722972972972973, "grad_norm": 4.867679119110107, "learning_rate": 8.093842761694366e-06, "loss": 1.2735, "step": 1785 }, { "epoch": 1.723938223938224, "grad_norm": 5.90767240524292, "learning_rate": 8.083611580348278e-06, "loss": 1.1858, "step": 1786 }, { "epoch": 1.724903474903475, "grad_norm": 4.679843425750732, "learning_rate": 8.073382481079005e-06, "loss": 1.1917, "step": 1787 }, { "epoch": 1.7258687258687258, "grad_norm": 4.496260643005371, "learning_rate": 8.063155475000037e-06, "loss": 1.1404, "step": 1788 }, { "epoch": 1.7268339768339769, "grad_norm": 4.13546085357666, "learning_rate": 8.052930573222597e-06, "loss": 1.1757, "step": 1789 }, { "epoch": 1.7277992277992278, "grad_norm": 4.124424934387207, "learning_rate": 8.042707786855618e-06, "loss": 1.1352, "step": 1790 }, { "epoch": 1.7287644787644787, "grad_norm": 4.252573013305664, "learning_rate": 8.032487127005732e-06, "loss": 1.2342, "step": 1791 }, { "epoch": 1.7297297297297298, "grad_norm": 4.479787349700928, "learning_rate": 8.02226860477727e-06, "loss": 1.2001, "step": 1792 }, { "epoch": 1.7306949806949807, "grad_norm": 4.56220006942749, "learning_rate": 8.01205223127223e-06, "loss": 1.1628, "step": 1793 }, { "epoch": 1.7316602316602316, "grad_norm": 4.39263916015625, "learning_rate": 8.001838017590278e-06, "loss": 1.1549, "step": 1794 }, { "epoch": 1.7326254826254828, "grad_norm": 4.32893705368042, "learning_rate": 7.991625974828735e-06, "loss": 1.1817, "step": 1795 }, { "epoch": 1.7335907335907335, "grad_norm": 5.986409664154053, "learning_rate": 7.981416114082568e-06, "loss": 1.1799, "step": 1796 }, { "epoch": 1.7345559845559846, "grad_norm": 5.146429061889648, "learning_rate": 7.971208446444366e-06, "loss": 1.1745, "step": 1797 }, { "epoch": 1.7355212355212355, "grad_norm": 4.632083892822266, "learning_rate": 7.961002983004335e-06, "loss": 1.2565, "step": 1798 }, { "epoch": 1.7364864864864864, "grad_norm": 5.8181962966918945, "learning_rate": 7.950799734850292e-06, "loss": 1.1844, "step": 1799 }, { "epoch": 1.7374517374517375, "grad_norm": 5.046428680419922, "learning_rate": 7.940598713067647e-06, "loss": 1.1407, "step": 1800 }, { "epoch": 1.7384169884169884, "grad_norm": 3.968395709991455, "learning_rate": 7.930399928739385e-06, "loss": 1.1586, "step": 1801 }, { "epoch": 1.7393822393822393, "grad_norm": 5.228263854980469, "learning_rate": 7.920203392946064e-06, "loss": 1.2388, "step": 1802 }, { "epoch": 1.7403474903474905, "grad_norm": 4.870275020599365, "learning_rate": 7.910009116765796e-06, "loss": 1.1788, "step": 1803 }, { "epoch": 1.7413127413127412, "grad_norm": 5.449280261993408, "learning_rate": 7.899817111274249e-06, "loss": 1.2277, "step": 1804 }, { "epoch": 1.7422779922779923, "grad_norm": 4.336848258972168, "learning_rate": 7.889627387544611e-06, "loss": 1.2175, "step": 1805 }, { "epoch": 1.7432432432432432, "grad_norm": 4.2892374992370605, "learning_rate": 7.879439956647594e-06, "loss": 1.1233, "step": 1806 }, { "epoch": 1.744208494208494, "grad_norm": 4.4046244621276855, "learning_rate": 7.86925482965142e-06, "loss": 1.2245, "step": 1807 }, { "epoch": 1.7451737451737452, "grad_norm": 4.251980304718018, "learning_rate": 7.859072017621811e-06, "loss": 1.1628, "step": 1808 }, { "epoch": 1.7461389961389961, "grad_norm": 4.342248916625977, "learning_rate": 7.848891531621977e-06, "loss": 1.2555, "step": 1809 }, { "epoch": 1.747104247104247, "grad_norm": 4.569221019744873, "learning_rate": 7.838713382712583e-06, "loss": 1.1232, "step": 1810 }, { "epoch": 1.7480694980694982, "grad_norm": 4.1841044425964355, "learning_rate": 7.828537581951773e-06, "loss": 1.2064, "step": 1811 }, { "epoch": 1.7490347490347489, "grad_norm": 4.37945556640625, "learning_rate": 7.818364140395138e-06, "loss": 1.1761, "step": 1812 }, { "epoch": 1.75, "grad_norm": 4.35389518737793, "learning_rate": 7.808193069095698e-06, "loss": 1.1817, "step": 1813 }, { "epoch": 1.7509652509652511, "grad_norm": 4.284439563751221, "learning_rate": 7.798024379103903e-06, "loss": 1.1517, "step": 1814 }, { "epoch": 1.7519305019305018, "grad_norm": 4.608835220336914, "learning_rate": 7.787858081467609e-06, "loss": 1.2335, "step": 1815 }, { "epoch": 1.752895752895753, "grad_norm": 4.245322227478027, "learning_rate": 7.777694187232085e-06, "loss": 1.1761, "step": 1816 }, { "epoch": 1.7538610038610039, "grad_norm": 4.203044891357422, "learning_rate": 7.76753270743998e-06, "loss": 1.1422, "step": 1817 }, { "epoch": 1.7548262548262548, "grad_norm": 4.512558460235596, "learning_rate": 7.757373653131323e-06, "loss": 1.1788, "step": 1818 }, { "epoch": 1.755791505791506, "grad_norm": 4.879763603210449, "learning_rate": 7.7472170353435e-06, "loss": 1.179, "step": 1819 }, { "epoch": 1.7567567567567568, "grad_norm": 4.212410926818848, "learning_rate": 7.737062865111265e-06, "loss": 1.2456, "step": 1820 }, { "epoch": 1.7577220077220077, "grad_norm": 4.991441249847412, "learning_rate": 7.726911153466699e-06, "loss": 1.1862, "step": 1821 }, { "epoch": 1.7586872586872588, "grad_norm": 3.9241816997528076, "learning_rate": 7.716761911439217e-06, "loss": 1.2542, "step": 1822 }, { "epoch": 1.7596525096525095, "grad_norm": 3.840327024459839, "learning_rate": 7.706615150055557e-06, "loss": 1.1657, "step": 1823 }, { "epoch": 1.7606177606177607, "grad_norm": 4.376457214355469, "learning_rate": 7.696470880339751e-06, "loss": 1.2348, "step": 1824 }, { "epoch": 1.7615830115830116, "grad_norm": 4.3103508949279785, "learning_rate": 7.686329113313129e-06, "loss": 1.1787, "step": 1825 }, { "epoch": 1.7625482625482625, "grad_norm": 4.218089580535889, "learning_rate": 7.676189859994299e-06, "loss": 1.1962, "step": 1826 }, { "epoch": 1.7635135135135136, "grad_norm": 4.236190319061279, "learning_rate": 7.666053131399151e-06, "loss": 1.1673, "step": 1827 }, { "epoch": 1.7644787644787645, "grad_norm": 4.285384178161621, "learning_rate": 7.655918938540808e-06, "loss": 1.1574, "step": 1828 }, { "epoch": 1.7654440154440154, "grad_norm": 4.414580345153809, "learning_rate": 7.64578729242966e-06, "loss": 1.1686, "step": 1829 }, { "epoch": 1.7664092664092665, "grad_norm": 4.828358173370361, "learning_rate": 7.63565820407332e-06, "loss": 1.2469, "step": 1830 }, { "epoch": 1.7673745173745172, "grad_norm": 4.314386367797852, "learning_rate": 7.625531684476624e-06, "loss": 1.1844, "step": 1831 }, { "epoch": 1.7683397683397684, "grad_norm": 4.553681373596191, "learning_rate": 7.615407744641618e-06, "loss": 1.1458, "step": 1832 }, { "epoch": 1.7693050193050193, "grad_norm": 4.936324119567871, "learning_rate": 7.605286395567544e-06, "loss": 1.2037, "step": 1833 }, { "epoch": 1.7702702702702702, "grad_norm": 4.830184459686279, "learning_rate": 7.595167648250825e-06, "loss": 1.2002, "step": 1834 }, { "epoch": 1.7712355212355213, "grad_norm": 4.841872692108154, "learning_rate": 7.585051513685067e-06, "loss": 1.2497, "step": 1835 }, { "epoch": 1.7722007722007722, "grad_norm": 4.141880512237549, "learning_rate": 7.574938002861033e-06, "loss": 1.2083, "step": 1836 }, { "epoch": 1.7731660231660231, "grad_norm": 4.67802095413208, "learning_rate": 7.5648271267666295e-06, "loss": 1.2196, "step": 1837 }, { "epoch": 1.7741312741312742, "grad_norm": 4.651124954223633, "learning_rate": 7.554718896386907e-06, "loss": 1.1569, "step": 1838 }, { "epoch": 1.775096525096525, "grad_norm": 5.021517276763916, "learning_rate": 7.5446133227040415e-06, "loss": 1.2042, "step": 1839 }, { "epoch": 1.776061776061776, "grad_norm": 5.392575263977051, "learning_rate": 7.534510416697322e-06, "loss": 1.2277, "step": 1840 }, { "epoch": 1.777027027027027, "grad_norm": 4.053293704986572, "learning_rate": 7.524410189343139e-06, "loss": 1.1792, "step": 1841 }, { "epoch": 1.7779922779922779, "grad_norm": 4.763055801391602, "learning_rate": 7.514312651614964e-06, "loss": 1.209, "step": 1842 }, { "epoch": 1.778957528957529, "grad_norm": 4.666579723358154, "learning_rate": 7.504217814483364e-06, "loss": 1.211, "step": 1843 }, { "epoch": 1.77992277992278, "grad_norm": 4.091919898986816, "learning_rate": 7.494125688915959e-06, "loss": 1.2149, "step": 1844 }, { "epoch": 1.7808880308880308, "grad_norm": 5.1892595291137695, "learning_rate": 7.484036285877429e-06, "loss": 1.1575, "step": 1845 }, { "epoch": 1.781853281853282, "grad_norm": 4.62443733215332, "learning_rate": 7.4739496163294875e-06, "loss": 1.1507, "step": 1846 }, { "epoch": 1.7828185328185329, "grad_norm": 4.525411605834961, "learning_rate": 7.463865691230885e-06, "loss": 1.1297, "step": 1847 }, { "epoch": 1.7837837837837838, "grad_norm": 4.4371724128723145, "learning_rate": 7.453784521537393e-06, "loss": 1.217, "step": 1848 }, { "epoch": 1.784749034749035, "grad_norm": 4.973574638366699, "learning_rate": 7.443706118201782e-06, "loss": 1.207, "step": 1849 }, { "epoch": 1.7857142857142856, "grad_norm": 5.308924674987793, "learning_rate": 7.433630492173831e-06, "loss": 1.1676, "step": 1850 }, { "epoch": 1.7866795366795367, "grad_norm": 4.378430366516113, "learning_rate": 7.423557654400275e-06, "loss": 1.1859, "step": 1851 }, { "epoch": 1.7876447876447876, "grad_norm": 5.219233512878418, "learning_rate": 7.413487615824847e-06, "loss": 1.1863, "step": 1852 }, { "epoch": 1.7886100386100385, "grad_norm": 4.620556831359863, "learning_rate": 7.403420387388223e-06, "loss": 1.1951, "step": 1853 }, { "epoch": 1.7895752895752897, "grad_norm": 4.507449626922607, "learning_rate": 7.393355980028039e-06, "loss": 1.1938, "step": 1854 }, { "epoch": 1.7905405405405406, "grad_norm": 4.811728000640869, "learning_rate": 7.383294404678842e-06, "loss": 1.1331, "step": 1855 }, { "epoch": 1.7915057915057915, "grad_norm": 4.3953375816345215, "learning_rate": 7.373235672272132e-06, "loss": 1.135, "step": 1856 }, { "epoch": 1.7924710424710426, "grad_norm": 5.127755165100098, "learning_rate": 7.363179793736297e-06, "loss": 1.2286, "step": 1857 }, { "epoch": 1.7934362934362933, "grad_norm": 4.2956132888793945, "learning_rate": 7.35312677999664e-06, "loss": 1.1705, "step": 1858 }, { "epoch": 1.7944015444015444, "grad_norm": 4.386326313018799, "learning_rate": 7.343076641975345e-06, "loss": 1.2242, "step": 1859 }, { "epoch": 1.7953667953667953, "grad_norm": 4.163666725158691, "learning_rate": 7.333029390591466e-06, "loss": 1.1547, "step": 1860 }, { "epoch": 1.7963320463320462, "grad_norm": 4.53470516204834, "learning_rate": 7.322985036760928e-06, "loss": 1.1772, "step": 1861 }, { "epoch": 1.7972972972972974, "grad_norm": 4.408353805541992, "learning_rate": 7.312943591396508e-06, "loss": 1.154, "step": 1862 }, { "epoch": 1.7982625482625483, "grad_norm": 4.333704948425293, "learning_rate": 7.3029050654078235e-06, "loss": 1.2424, "step": 1863 }, { "epoch": 1.7992277992277992, "grad_norm": 4.769482135772705, "learning_rate": 7.292869469701315e-06, "loss": 1.2207, "step": 1864 }, { "epoch": 1.8001930501930503, "grad_norm": 4.205842971801758, "learning_rate": 7.282836815180241e-06, "loss": 1.1742, "step": 1865 }, { "epoch": 1.801158301158301, "grad_norm": 4.177207946777344, "learning_rate": 7.272807112744666e-06, "loss": 1.1845, "step": 1866 }, { "epoch": 1.8021235521235521, "grad_norm": 4.250517845153809, "learning_rate": 7.262780373291452e-06, "loss": 1.2649, "step": 1867 }, { "epoch": 1.803088803088803, "grad_norm": 4.70888614654541, "learning_rate": 7.252756607714238e-06, "loss": 1.1456, "step": 1868 }, { "epoch": 1.804054054054054, "grad_norm": 4.450895309448242, "learning_rate": 7.242735826903422e-06, "loss": 1.1862, "step": 1869 }, { "epoch": 1.805019305019305, "grad_norm": 4.835483074188232, "learning_rate": 7.2327180417461716e-06, "loss": 1.1659, "step": 1870 }, { "epoch": 1.805984555984556, "grad_norm": 4.576693534851074, "learning_rate": 7.222703263126402e-06, "loss": 1.1301, "step": 1871 }, { "epoch": 1.806949806949807, "grad_norm": 4.504087924957275, "learning_rate": 7.212691501924753e-06, "loss": 1.118, "step": 1872 }, { "epoch": 1.807915057915058, "grad_norm": 4.55885124206543, "learning_rate": 7.2026827690185874e-06, "loss": 1.2146, "step": 1873 }, { "epoch": 1.808880308880309, "grad_norm": 5.224696636199951, "learning_rate": 7.192677075281978e-06, "loss": 1.1357, "step": 1874 }, { "epoch": 1.8098455598455598, "grad_norm": 4.089498519897461, "learning_rate": 7.182674431585703e-06, "loss": 1.2009, "step": 1875 }, { "epoch": 1.810810810810811, "grad_norm": 5.025662899017334, "learning_rate": 7.172674848797218e-06, "loss": 1.2124, "step": 1876 }, { "epoch": 1.8117760617760617, "grad_norm": 4.546243667602539, "learning_rate": 7.162678337780664e-06, "loss": 1.2037, "step": 1877 }, { "epoch": 1.8127413127413128, "grad_norm": 4.327416896820068, "learning_rate": 7.1526849093968234e-06, "loss": 1.2691, "step": 1878 }, { "epoch": 1.8137065637065637, "grad_norm": 4.666914939880371, "learning_rate": 7.142694574503151e-06, "loss": 1.2732, "step": 1879 }, { "epoch": 1.8146718146718146, "grad_norm": 4.3911333084106445, "learning_rate": 7.132707343953732e-06, "loss": 1.2423, "step": 1880 }, { "epoch": 1.8156370656370657, "grad_norm": 4.461322784423828, "learning_rate": 7.1227232285992855e-06, "loss": 1.1411, "step": 1881 }, { "epoch": 1.8166023166023166, "grad_norm": 4.10672664642334, "learning_rate": 7.112742239287127e-06, "loss": 1.2356, "step": 1882 }, { "epoch": 1.8175675675675675, "grad_norm": 4.363541126251221, "learning_rate": 7.1027643868612e-06, "loss": 1.1945, "step": 1883 }, { "epoch": 1.8185328185328187, "grad_norm": 4.521150588989258, "learning_rate": 7.092789682162024e-06, "loss": 1.24, "step": 1884 }, { "epoch": 1.8194980694980694, "grad_norm": 4.348193168640137, "learning_rate": 7.082818136026706e-06, "loss": 1.1318, "step": 1885 }, { "epoch": 1.8204633204633205, "grad_norm": 4.012368202209473, "learning_rate": 7.072849759288914e-06, "loss": 1.2535, "step": 1886 }, { "epoch": 1.8214285714285714, "grad_norm": 4.3002471923828125, "learning_rate": 7.062884562778883e-06, "loss": 1.1929, "step": 1887 }, { "epoch": 1.8223938223938223, "grad_norm": 4.529809474945068, "learning_rate": 7.052922557323383e-06, "loss": 1.1595, "step": 1888 }, { "epoch": 1.8233590733590734, "grad_norm": 4.302705764770508, "learning_rate": 7.042963753745722e-06, "loss": 1.2402, "step": 1889 }, { "epoch": 1.8243243243243243, "grad_norm": 4.311285495758057, "learning_rate": 7.033008162865732e-06, "loss": 1.1861, "step": 1890 }, { "epoch": 1.8252895752895753, "grad_norm": 4.44135046005249, "learning_rate": 7.023055795499747e-06, "loss": 1.2708, "step": 1891 }, { "epoch": 1.8262548262548264, "grad_norm": 4.508293151855469, "learning_rate": 7.013106662460605e-06, "loss": 1.1778, "step": 1892 }, { "epoch": 1.827220077220077, "grad_norm": 4.0937395095825195, "learning_rate": 7.003160774557625e-06, "loss": 1.184, "step": 1893 }, { "epoch": 1.8281853281853282, "grad_norm": 4.576756954193115, "learning_rate": 6.993218142596609e-06, "loss": 1.171, "step": 1894 }, { "epoch": 1.829150579150579, "grad_norm": 4.2449164390563965, "learning_rate": 6.98327877737981e-06, "loss": 1.1839, "step": 1895 }, { "epoch": 1.83011583011583, "grad_norm": 4.2238688468933105, "learning_rate": 6.9733426897059375e-06, "loss": 1.2014, "step": 1896 }, { "epoch": 1.8310810810810811, "grad_norm": 3.9330780506134033, "learning_rate": 6.963409890370141e-06, "loss": 1.1652, "step": 1897 }, { "epoch": 1.832046332046332, "grad_norm": 4.291113376617432, "learning_rate": 6.953480390164001e-06, "loss": 1.1836, "step": 1898 }, { "epoch": 1.833011583011583, "grad_norm": 3.825181484222412, "learning_rate": 6.943554199875506e-06, "loss": 1.1542, "step": 1899 }, { "epoch": 1.833976833976834, "grad_norm": 4.255434513092041, "learning_rate": 6.93363133028905e-06, "loss": 1.2163, "step": 1900 }, { "epoch": 1.834942084942085, "grad_norm": 4.05705451965332, "learning_rate": 6.923711792185419e-06, "loss": 1.2347, "step": 1901 }, { "epoch": 1.835907335907336, "grad_norm": 4.2765631675720215, "learning_rate": 6.913795596341786e-06, "loss": 1.2278, "step": 1902 }, { "epoch": 1.836872586872587, "grad_norm": 3.9429051876068115, "learning_rate": 6.903882753531687e-06, "loss": 1.2926, "step": 1903 }, { "epoch": 1.8378378378378377, "grad_norm": 4.436086654663086, "learning_rate": 6.8939732745250145e-06, "loss": 1.2279, "step": 1904 }, { "epoch": 1.8388030888030888, "grad_norm": 4.028366565704346, "learning_rate": 6.884067170088007e-06, "loss": 1.1804, "step": 1905 }, { "epoch": 1.8397683397683398, "grad_norm": 4.2818427085876465, "learning_rate": 6.874164450983238e-06, "loss": 1.1952, "step": 1906 }, { "epoch": 1.8407335907335907, "grad_norm": 4.631438732147217, "learning_rate": 6.864265127969606e-06, "loss": 1.2501, "step": 1907 }, { "epoch": 1.8416988416988418, "grad_norm": 4.176527976989746, "learning_rate": 6.854369211802314e-06, "loss": 1.2307, "step": 1908 }, { "epoch": 1.8426640926640927, "grad_norm": 4.664109230041504, "learning_rate": 6.844476713232863e-06, "loss": 1.1214, "step": 1909 }, { "epoch": 1.8436293436293436, "grad_norm": 5.156432151794434, "learning_rate": 6.834587643009047e-06, "loss": 1.2163, "step": 1910 }, { "epoch": 1.8445945945945947, "grad_norm": 3.8033342361450195, "learning_rate": 6.824702011874933e-06, "loss": 1.14, "step": 1911 }, { "epoch": 1.8455598455598454, "grad_norm": 5.293600082397461, "learning_rate": 6.81481983057085e-06, "loss": 1.1455, "step": 1912 }, { "epoch": 1.8465250965250966, "grad_norm": 5.740131855010986, "learning_rate": 6.804941109833375e-06, "loss": 1.2523, "step": 1913 }, { "epoch": 1.8474903474903475, "grad_norm": 4.465753555297852, "learning_rate": 6.7950658603953355e-06, "loss": 1.1601, "step": 1914 }, { "epoch": 1.8484555984555984, "grad_norm": 6.7862324714660645, "learning_rate": 6.78519409298578e-06, "loss": 1.1743, "step": 1915 }, { "epoch": 1.8494208494208495, "grad_norm": 5.846525192260742, "learning_rate": 6.775325818329974e-06, "loss": 1.2319, "step": 1916 }, { "epoch": 1.8503861003861004, "grad_norm": 5.151252269744873, "learning_rate": 6.765461047149397e-06, "loss": 1.2479, "step": 1917 }, { "epoch": 1.8513513513513513, "grad_norm": 6.499340057373047, "learning_rate": 6.755599790161707e-06, "loss": 1.2277, "step": 1918 }, { "epoch": 1.8523166023166024, "grad_norm": 4.451925754547119, "learning_rate": 6.745742058080753e-06, "loss": 1.1884, "step": 1919 }, { "epoch": 1.8532818532818531, "grad_norm": 6.6447319984436035, "learning_rate": 6.735887861616555e-06, "loss": 1.1883, "step": 1920 }, { "epoch": 1.8542471042471043, "grad_norm": 4.626460075378418, "learning_rate": 6.726037211475296e-06, "loss": 1.1998, "step": 1921 }, { "epoch": 1.8552123552123552, "grad_norm": 5.135448932647705, "learning_rate": 6.716190118359293e-06, "loss": 1.1687, "step": 1922 }, { "epoch": 1.856177606177606, "grad_norm": 5.576619625091553, "learning_rate": 6.706346592967006e-06, "loss": 1.1574, "step": 1923 }, { "epoch": 1.8571428571428572, "grad_norm": 5.447540760040283, "learning_rate": 6.6965066459930195e-06, "loss": 1.1473, "step": 1924 }, { "epoch": 1.8581081081081081, "grad_norm": 7.530019283294678, "learning_rate": 6.686670288128032e-06, "loss": 1.1573, "step": 1925 }, { "epoch": 1.859073359073359, "grad_norm": 6.108087539672852, "learning_rate": 6.67683753005884e-06, "loss": 1.245, "step": 1926 }, { "epoch": 1.8600386100386102, "grad_norm": 4.211674690246582, "learning_rate": 6.667008382468323e-06, "loss": 1.1678, "step": 1927 }, { "epoch": 1.8610038610038608, "grad_norm": 6.624056339263916, "learning_rate": 6.657182856035447e-06, "loss": 1.2613, "step": 1928 }, { "epoch": 1.861969111969112, "grad_norm": 4.441121578216553, "learning_rate": 6.6473609614352405e-06, "loss": 1.1993, "step": 1929 }, { "epoch": 1.862934362934363, "grad_norm": 4.745296478271484, "learning_rate": 6.637542709338789e-06, "loss": 1.2113, "step": 1930 }, { "epoch": 1.8638996138996138, "grad_norm": 4.846886157989502, "learning_rate": 6.627728110413214e-06, "loss": 1.159, "step": 1931 }, { "epoch": 1.864864864864865, "grad_norm": 4.237720012664795, "learning_rate": 6.617917175321668e-06, "loss": 1.1422, "step": 1932 }, { "epoch": 1.8658301158301158, "grad_norm": 5.267391681671143, "learning_rate": 6.6081099147233344e-06, "loss": 1.2238, "step": 1933 }, { "epoch": 1.8667953667953667, "grad_norm": 4.453833103179932, "learning_rate": 6.598306339273393e-06, "loss": 1.1406, "step": 1934 }, { "epoch": 1.8677606177606179, "grad_norm": 5.526133060455322, "learning_rate": 6.588506459623023e-06, "loss": 1.2287, "step": 1935 }, { "epoch": 1.8687258687258688, "grad_norm": 4.279199600219727, "learning_rate": 6.578710286419388e-06, "loss": 1.2199, "step": 1936 }, { "epoch": 1.8696911196911197, "grad_norm": 4.354450225830078, "learning_rate": 6.568917830305625e-06, "loss": 1.1645, "step": 1937 }, { "epoch": 1.8706563706563708, "grad_norm": 5.210305690765381, "learning_rate": 6.559129101920836e-06, "loss": 1.1394, "step": 1938 }, { "epoch": 1.8716216216216215, "grad_norm": 4.386662483215332, "learning_rate": 6.54934411190007e-06, "loss": 1.175, "step": 1939 }, { "epoch": 1.8725868725868726, "grad_norm": 4.03883695602417, "learning_rate": 6.539562870874309e-06, "loss": 1.1861, "step": 1940 }, { "epoch": 1.8735521235521235, "grad_norm": 5.33974552154541, "learning_rate": 6.529785389470469e-06, "loss": 1.1526, "step": 1941 }, { "epoch": 1.8745173745173744, "grad_norm": 4.374886512756348, "learning_rate": 6.520011678311382e-06, "loss": 1.2011, "step": 1942 }, { "epoch": 1.8754826254826256, "grad_norm": 4.7918829917907715, "learning_rate": 6.510241748015779e-06, "loss": 1.2436, "step": 1943 }, { "epoch": 1.8764478764478765, "grad_norm": 4.391792297363281, "learning_rate": 6.500475609198291e-06, "loss": 1.1879, "step": 1944 }, { "epoch": 1.8774131274131274, "grad_norm": 4.195566177368164, "learning_rate": 6.490713272469414e-06, "loss": 1.1857, "step": 1945 }, { "epoch": 1.8783783783783785, "grad_norm": 4.680538654327393, "learning_rate": 6.480954748435531e-06, "loss": 1.2311, "step": 1946 }, { "epoch": 1.8793436293436292, "grad_norm": 4.65995979309082, "learning_rate": 6.4712000476988715e-06, "loss": 1.152, "step": 1947 }, { "epoch": 1.8803088803088803, "grad_norm": 4.1404829025268555, "learning_rate": 6.4614491808575215e-06, "loss": 1.215, "step": 1948 }, { "epoch": 1.8812741312741312, "grad_norm": 4.770052433013916, "learning_rate": 6.451702158505384e-06, "loss": 1.1552, "step": 1949 }, { "epoch": 1.8822393822393821, "grad_norm": 4.840427875518799, "learning_rate": 6.4419589912322025e-06, "loss": 1.2019, "step": 1950 }, { "epoch": 1.8832046332046333, "grad_norm": 4.4432148933410645, "learning_rate": 6.4322196896235236e-06, "loss": 1.1779, "step": 1951 }, { "epoch": 1.8841698841698842, "grad_norm": 4.31553316116333, "learning_rate": 6.422484264260698e-06, "loss": 1.1712, "step": 1952 }, { "epoch": 1.885135135135135, "grad_norm": 4.212576389312744, "learning_rate": 6.412752725720864e-06, "loss": 1.1775, "step": 1953 }, { "epoch": 1.8861003861003862, "grad_norm": 4.3694353103637695, "learning_rate": 6.403025084576932e-06, "loss": 1.2055, "step": 1954 }, { "epoch": 1.887065637065637, "grad_norm": 5.076420783996582, "learning_rate": 6.393301351397583e-06, "loss": 1.2423, "step": 1955 }, { "epoch": 1.888030888030888, "grad_norm": 4.373606204986572, "learning_rate": 6.383581536747252e-06, "loss": 1.2749, "step": 1956 }, { "epoch": 1.8889961389961392, "grad_norm": 5.308967113494873, "learning_rate": 6.3738656511861195e-06, "loss": 1.1588, "step": 1957 }, { "epoch": 1.8899613899613898, "grad_norm": 4.475518703460693, "learning_rate": 6.364153705270088e-06, "loss": 1.1719, "step": 1958 }, { "epoch": 1.890926640926641, "grad_norm": 4.570695877075195, "learning_rate": 6.3544457095507885e-06, "loss": 1.1194, "step": 1959 }, { "epoch": 1.8918918918918919, "grad_norm": 4.763030529022217, "learning_rate": 6.344741674575553e-06, "loss": 1.1407, "step": 1960 }, { "epoch": 1.8928571428571428, "grad_norm": 4.580066680908203, "learning_rate": 6.335041610887422e-06, "loss": 1.188, "step": 1961 }, { "epoch": 1.893822393822394, "grad_norm": 4.779040336608887, "learning_rate": 6.325345529025107e-06, "loss": 1.1896, "step": 1962 }, { "epoch": 1.8947876447876448, "grad_norm": 4.4460129737854, "learning_rate": 6.315653439523003e-06, "loss": 1.1016, "step": 1963 }, { "epoch": 1.8957528957528957, "grad_norm": 4.613245487213135, "learning_rate": 6.305965352911162e-06, "loss": 1.1686, "step": 1964 }, { "epoch": 1.8967181467181469, "grad_norm": 5.302827835083008, "learning_rate": 6.296281279715294e-06, "loss": 1.1598, "step": 1965 }, { "epoch": 1.8976833976833976, "grad_norm": 4.315168380737305, "learning_rate": 6.2866012304567436e-06, "loss": 1.1639, "step": 1966 }, { "epoch": 1.8986486486486487, "grad_norm": 5.385082721710205, "learning_rate": 6.276925215652482e-06, "loss": 1.2059, "step": 1967 }, { "epoch": 1.8996138996138996, "grad_norm": 3.810722589492798, "learning_rate": 6.267253245815098e-06, "loss": 1.214, "step": 1968 }, { "epoch": 1.9005791505791505, "grad_norm": 4.369156360626221, "learning_rate": 6.257585331452791e-06, "loss": 1.2251, "step": 1969 }, { "epoch": 1.9015444015444016, "grad_norm": 4.052249908447266, "learning_rate": 6.247921483069352e-06, "loss": 1.2096, "step": 1970 }, { "epoch": 1.9025096525096525, "grad_norm": 4.842185974121094, "learning_rate": 6.238261711164146e-06, "loss": 1.1637, "step": 1971 }, { "epoch": 1.9034749034749034, "grad_norm": 4.0050787925720215, "learning_rate": 6.228606026232118e-06, "loss": 1.2152, "step": 1972 }, { "epoch": 1.9044401544401546, "grad_norm": 4.679233074188232, "learning_rate": 6.218954438763773e-06, "loss": 1.1746, "step": 1973 }, { "epoch": 1.9054054054054053, "grad_norm": 4.082679748535156, "learning_rate": 6.209306959245161e-06, "loss": 1.1686, "step": 1974 }, { "epoch": 1.9063706563706564, "grad_norm": 4.452788352966309, "learning_rate": 6.1996635981578755e-06, "loss": 1.1397, "step": 1975 }, { "epoch": 1.9073359073359073, "grad_norm": 4.730610370635986, "learning_rate": 6.190024365979017e-06, "loss": 1.2131, "step": 1976 }, { "epoch": 1.9083011583011582, "grad_norm": 5.205382347106934, "learning_rate": 6.180389273181224e-06, "loss": 1.1978, "step": 1977 }, { "epoch": 1.9092664092664093, "grad_norm": 4.563416481018066, "learning_rate": 6.170758330232621e-06, "loss": 1.2212, "step": 1978 }, { "epoch": 1.9102316602316602, "grad_norm": 4.994668483734131, "learning_rate": 6.161131547596833e-06, "loss": 1.1968, "step": 1979 }, { "epoch": 1.9111969111969112, "grad_norm": 4.848903179168701, "learning_rate": 6.1515089357329565e-06, "loss": 1.2414, "step": 1980 }, { "epoch": 1.9121621621621623, "grad_norm": 4.824883937835693, "learning_rate": 6.141890505095566e-06, "loss": 1.1529, "step": 1981 }, { "epoch": 1.913127413127413, "grad_norm": 4.200015068054199, "learning_rate": 6.1322762661346865e-06, "loss": 1.2305, "step": 1982 }, { "epoch": 1.914092664092664, "grad_norm": 4.15867280960083, "learning_rate": 6.122666229295792e-06, "loss": 1.2336, "step": 1983 }, { "epoch": 1.915057915057915, "grad_norm": 4.84639835357666, "learning_rate": 6.113060405019794e-06, "loss": 1.1236, "step": 1984 }, { "epoch": 1.916023166023166, "grad_norm": 4.204685688018799, "learning_rate": 6.103458803743018e-06, "loss": 1.1784, "step": 1985 }, { "epoch": 1.916988416988417, "grad_norm": 4.868194580078125, "learning_rate": 6.093861435897208e-06, "loss": 1.1808, "step": 1986 }, { "epoch": 1.917953667953668, "grad_norm": 4.526357173919678, "learning_rate": 6.0842683119095095e-06, "loss": 1.1911, "step": 1987 }, { "epoch": 1.9189189189189189, "grad_norm": 5.651543617248535, "learning_rate": 6.0746794422024555e-06, "loss": 1.2539, "step": 1988 }, { "epoch": 1.91988416988417, "grad_norm": 5.241189002990723, "learning_rate": 6.065094837193954e-06, "loss": 1.1517, "step": 1989 }, { "epoch": 1.920849420849421, "grad_norm": 4.834219455718994, "learning_rate": 6.055514507297284e-06, "loss": 1.1307, "step": 1990 }, { "epoch": 1.9218146718146718, "grad_norm": 5.599984645843506, "learning_rate": 6.045938462921077e-06, "loss": 1.2073, "step": 1991 }, { "epoch": 1.922779922779923, "grad_norm": 4.227385520935059, "learning_rate": 6.03636671446931e-06, "loss": 1.2211, "step": 1992 }, { "epoch": 1.9237451737451736, "grad_norm": 6.122756004333496, "learning_rate": 6.026799272341296e-06, "loss": 1.1712, "step": 1993 }, { "epoch": 1.9247104247104247, "grad_norm": 4.500845909118652, "learning_rate": 6.0172361469316585e-06, "loss": 1.1913, "step": 1994 }, { "epoch": 1.9256756756756757, "grad_norm": 5.436099052429199, "learning_rate": 6.00767734863034e-06, "loss": 1.0837, "step": 1995 }, { "epoch": 1.9266409266409266, "grad_norm": 5.637316703796387, "learning_rate": 5.998122887822582e-06, "loss": 1.188, "step": 1996 }, { "epoch": 1.9276061776061777, "grad_norm": 4.430753231048584, "learning_rate": 5.988572774888913e-06, "loss": 1.1335, "step": 1997 }, { "epoch": 1.9285714285714286, "grad_norm": 6.754420757293701, "learning_rate": 5.979027020205131e-06, "loss": 1.2139, "step": 1998 }, { "epoch": 1.9295366795366795, "grad_norm": 4.46309232711792, "learning_rate": 5.9694856341423045e-06, "loss": 1.1497, "step": 1999 }, { "epoch": 1.9305019305019306, "grad_norm": 5.311478137969971, "learning_rate": 5.9599486270667595e-06, "loss": 1.2128, "step": 2000 }, { "epoch": 1.9314671814671813, "grad_norm": 5.918153762817383, "learning_rate": 5.950416009340056e-06, "loss": 1.233, "step": 2001 }, { "epoch": 1.9324324324324325, "grad_norm": 5.034909248352051, "learning_rate": 5.940887791318993e-06, "loss": 1.2033, "step": 2002 }, { "epoch": 1.9333976833976834, "grad_norm": 5.40713357925415, "learning_rate": 5.93136398335558e-06, "loss": 1.266, "step": 2003 }, { "epoch": 1.9343629343629343, "grad_norm": 6.276338577270508, "learning_rate": 5.9218445957970435e-06, "loss": 1.2289, "step": 2004 }, { "epoch": 1.9353281853281854, "grad_norm": 4.238287448883057, "learning_rate": 5.912329638985808e-06, "loss": 1.1408, "step": 2005 }, { "epoch": 1.9362934362934363, "grad_norm": 7.066354274749756, "learning_rate": 5.902819123259478e-06, "loss": 1.1703, "step": 2006 }, { "epoch": 1.9372586872586872, "grad_norm": 5.24070405960083, "learning_rate": 5.8933130589508314e-06, "loss": 1.2035, "step": 2007 }, { "epoch": 1.9382239382239383, "grad_norm": 4.5445404052734375, "learning_rate": 5.883811456387821e-06, "loss": 1.2079, "step": 2008 }, { "epoch": 1.939189189189189, "grad_norm": 7.246978759765625, "learning_rate": 5.8743143258935435e-06, "loss": 1.2059, "step": 2009 }, { "epoch": 1.9401544401544402, "grad_norm": 4.268136978149414, "learning_rate": 5.8648216777862345e-06, "loss": 1.1996, "step": 2010 }, { "epoch": 1.941119691119691, "grad_norm": 5.60009241104126, "learning_rate": 5.855333522379272e-06, "loss": 1.2054, "step": 2011 }, { "epoch": 1.942084942084942, "grad_norm": 5.696875095367432, "learning_rate": 5.845849869981137e-06, "loss": 1.2104, "step": 2012 }, { "epoch": 1.943050193050193, "grad_norm": 4.394375324249268, "learning_rate": 5.836370730895427e-06, "loss": 1.1857, "step": 2013 }, { "epoch": 1.944015444015444, "grad_norm": 5.050985336303711, "learning_rate": 5.8268961154208395e-06, "loss": 1.2082, "step": 2014 }, { "epoch": 1.944980694980695, "grad_norm": 4.772952556610107, "learning_rate": 5.817426033851152e-06, "loss": 1.211, "step": 2015 }, { "epoch": 1.945945945945946, "grad_norm": 4.337421894073486, "learning_rate": 5.8079604964752126e-06, "loss": 1.1849, "step": 2016 }, { "epoch": 1.946911196911197, "grad_norm": 5.666919231414795, "learning_rate": 5.798499513576934e-06, "loss": 1.2063, "step": 2017 }, { "epoch": 1.9478764478764479, "grad_norm": 4.089428424835205, "learning_rate": 5.7890430954352895e-06, "loss": 1.2211, "step": 2018 }, { "epoch": 1.948841698841699, "grad_norm": 5.097423553466797, "learning_rate": 5.779591252324286e-06, "loss": 1.1857, "step": 2019 }, { "epoch": 1.9498069498069497, "grad_norm": 4.608874797821045, "learning_rate": 5.7701439945129625e-06, "loss": 1.1794, "step": 2020 }, { "epoch": 1.9507722007722008, "grad_norm": 4.127477169036865, "learning_rate": 5.7607013322653635e-06, "loss": 1.2072, "step": 2021 }, { "epoch": 1.9517374517374517, "grad_norm": 5.385368824005127, "learning_rate": 5.751263275840562e-06, "loss": 1.1942, "step": 2022 }, { "epoch": 1.9527027027027026, "grad_norm": 4.392673492431641, "learning_rate": 5.741829835492611e-06, "loss": 1.1737, "step": 2023 }, { "epoch": 1.9536679536679538, "grad_norm": 4.49373197555542, "learning_rate": 5.73240102147056e-06, "loss": 1.1957, "step": 2024 }, { "epoch": 1.9546332046332047, "grad_norm": 5.873342037200928, "learning_rate": 5.722976844018414e-06, "loss": 1.1902, "step": 2025 }, { "epoch": 1.9555984555984556, "grad_norm": 4.4943766593933105, "learning_rate": 5.713557313375162e-06, "loss": 1.1804, "step": 2026 }, { "epoch": 1.9565637065637067, "grad_norm": 4.722905158996582, "learning_rate": 5.704142439774732e-06, "loss": 1.2106, "step": 2027 }, { "epoch": 1.9575289575289574, "grad_norm": 5.092207908630371, "learning_rate": 5.694732233445996e-06, "loss": 1.179, "step": 2028 }, { "epoch": 1.9584942084942085, "grad_norm": 4.395949363708496, "learning_rate": 5.685326704612753e-06, "loss": 1.2379, "step": 2029 }, { "epoch": 1.9594594594594594, "grad_norm": 4.497315406799316, "learning_rate": 5.675925863493721e-06, "loss": 1.154, "step": 2030 }, { "epoch": 1.9604247104247103, "grad_norm": 5.878604412078857, "learning_rate": 5.666529720302525e-06, "loss": 1.1275, "step": 2031 }, { "epoch": 1.9613899613899615, "grad_norm": 4.746928691864014, "learning_rate": 5.657138285247687e-06, "loss": 1.2247, "step": 2032 }, { "epoch": 1.9623552123552124, "grad_norm": 4.125810623168945, "learning_rate": 5.647751568532616e-06, "loss": 1.2614, "step": 2033 }, { "epoch": 1.9633204633204633, "grad_norm": 5.916082859039307, "learning_rate": 5.638369580355587e-06, "loss": 1.2408, "step": 2034 }, { "epoch": 1.9642857142857144, "grad_norm": 5.335667610168457, "learning_rate": 5.628992330909744e-06, "loss": 1.2014, "step": 2035 }, { "epoch": 1.965250965250965, "grad_norm": 5.12222957611084, "learning_rate": 5.619619830383083e-06, "loss": 1.237, "step": 2036 }, { "epoch": 1.9662162162162162, "grad_norm": 4.592029094696045, "learning_rate": 5.610252088958434e-06, "loss": 1.2112, "step": 2037 }, { "epoch": 1.9671814671814671, "grad_norm": 4.4699225425720215, "learning_rate": 5.600889116813466e-06, "loss": 1.2412, "step": 2038 }, { "epoch": 1.968146718146718, "grad_norm": 4.312219142913818, "learning_rate": 5.591530924120659e-06, "loss": 1.1981, "step": 2039 }, { "epoch": 1.9691119691119692, "grad_norm": 5.5404815673828125, "learning_rate": 5.5821775210473015e-06, "loss": 1.1716, "step": 2040 }, { "epoch": 1.97007722007722, "grad_norm": 4.853655815124512, "learning_rate": 5.5728289177554805e-06, "loss": 1.0895, "step": 2041 }, { "epoch": 1.971042471042471, "grad_norm": 4.7461419105529785, "learning_rate": 5.563485124402068e-06, "loss": 1.2178, "step": 2042 }, { "epoch": 1.9720077220077221, "grad_norm": 5.6155595779418945, "learning_rate": 5.554146151138706e-06, "loss": 1.2066, "step": 2043 }, { "epoch": 1.972972972972973, "grad_norm": 5.168851375579834, "learning_rate": 5.5448120081118085e-06, "loss": 1.1773, "step": 2044 }, { "epoch": 1.973938223938224, "grad_norm": 5.844036102294922, "learning_rate": 5.53548270546253e-06, "loss": 1.1823, "step": 2045 }, { "epoch": 1.974903474903475, "grad_norm": 6.1245341300964355, "learning_rate": 5.526158253326777e-06, "loss": 1.2009, "step": 2046 }, { "epoch": 1.9758687258687258, "grad_norm": 4.416387557983398, "learning_rate": 5.5168386618351785e-06, "loss": 1.2369, "step": 2047 }, { "epoch": 1.9768339768339769, "grad_norm": 6.00404167175293, "learning_rate": 5.507523941113086e-06, "loss": 1.1976, "step": 2048 }, { "epoch": 1.9777992277992278, "grad_norm": 7.08099365234375, "learning_rate": 5.498214101280561e-06, "loss": 1.1504, "step": 2049 }, { "epoch": 1.9787644787644787, "grad_norm": 4.339968204498291, "learning_rate": 5.488909152452357e-06, "loss": 1.1816, "step": 2050 }, { "epoch": 1.9797297297297298, "grad_norm": 6.9030537605285645, "learning_rate": 5.479609104737917e-06, "loss": 1.2805, "step": 2051 }, { "epoch": 1.9806949806949807, "grad_norm": 6.546342372894287, "learning_rate": 5.4703139682413585e-06, "loss": 1.1604, "step": 2052 }, { "epoch": 1.9816602316602316, "grad_norm": 4.142830848693848, "learning_rate": 5.461023753061463e-06, "loss": 1.1628, "step": 2053 }, { "epoch": 1.9826254826254828, "grad_norm": 6.560509204864502, "learning_rate": 5.451738469291666e-06, "loss": 1.1804, "step": 2054 }, { "epoch": 1.9835907335907335, "grad_norm": 6.8526811599731445, "learning_rate": 5.442458127020045e-06, "loss": 1.1882, "step": 2055 }, { "epoch": 1.9845559845559846, "grad_norm": 4.52562952041626, "learning_rate": 5.433182736329305e-06, "loss": 1.1942, "step": 2056 }, { "epoch": 1.9855212355212355, "grad_norm": 6.92567777633667, "learning_rate": 5.423912307296779e-06, "loss": 1.1957, "step": 2057 }, { "epoch": 1.9864864864864864, "grad_norm": 5.3068366050720215, "learning_rate": 5.414646849994403e-06, "loss": 1.1577, "step": 2058 }, { "epoch": 1.9874517374517375, "grad_norm": 4.130660057067871, "learning_rate": 5.405386374488709e-06, "loss": 1.1749, "step": 2059 }, { "epoch": 1.9884169884169884, "grad_norm": 4.9182209968566895, "learning_rate": 5.396130890840834e-06, "loss": 1.2134, "step": 2060 }, { "epoch": 1.9893822393822393, "grad_norm": 4.714017391204834, "learning_rate": 5.386880409106465e-06, "loss": 1.1217, "step": 2061 }, { "epoch": 1.9903474903474905, "grad_norm": 4.295403480529785, "learning_rate": 5.377634939335874e-06, "loss": 1.1718, "step": 2062 }, { "epoch": 1.9913127413127412, "grad_norm": 4.523777484893799, "learning_rate": 5.368394491573876e-06, "loss": 1.2635, "step": 2063 }, { "epoch": 1.9922779922779923, "grad_norm": 5.195492744445801, "learning_rate": 5.359159075859848e-06, "loss": 1.1536, "step": 2064 }, { "epoch": 1.9932432432432432, "grad_norm": 3.8529839515686035, "learning_rate": 5.349928702227677e-06, "loss": 1.1507, "step": 2065 }, { "epoch": 1.994208494208494, "grad_norm": 4.830549716949463, "learning_rate": 5.3407033807057855e-06, "loss": 1.1928, "step": 2066 }, { "epoch": 1.9951737451737452, "grad_norm": 4.1817402839660645, "learning_rate": 5.331483121317098e-06, "loss": 1.1438, "step": 2067 }, { "epoch": 1.9961389961389961, "grad_norm": 4.0853495597839355, "learning_rate": 5.322267934079056e-06, "loss": 1.2292, "step": 2068 }, { "epoch": 1.997104247104247, "grad_norm": 4.8644866943359375, "learning_rate": 5.313057829003578e-06, "loss": 1.2231, "step": 2069 }, { "epoch": 1.9980694980694982, "grad_norm": 4.481793403625488, "learning_rate": 5.303852816097057e-06, "loss": 1.1979, "step": 2070 }, { "epoch": 1.9990347490347489, "grad_norm": 4.819639205932617, "learning_rate": 5.294652905360356e-06, "loss": 1.2558, "step": 2071 }, { "epoch": 2.0, "grad_norm": 4.601913928985596, "learning_rate": 5.285458106788808e-06, "loss": 1.0367, "step": 2072 }, { "epoch": 2.000965250965251, "grad_norm": 3.6043078899383545, "learning_rate": 5.27626843037218e-06, "loss": 0.7693, "step": 2073 }, { "epoch": 2.001930501930502, "grad_norm": 3.638916015625, "learning_rate": 5.267083886094668e-06, "loss": 0.7657, "step": 2074 }, { "epoch": 2.002895752895753, "grad_norm": 4.656620025634766, "learning_rate": 5.2579044839349e-06, "loss": 0.7083, "step": 2075 }, { "epoch": 2.003861003861004, "grad_norm": 3.8113861083984375, "learning_rate": 5.248730233865926e-06, "loss": 0.7342, "step": 2076 }, { "epoch": 2.0048262548262548, "grad_norm": 4.38107442855835, "learning_rate": 5.239561145855183e-06, "loss": 0.6649, "step": 2077 }, { "epoch": 2.005791505791506, "grad_norm": 5.1227498054504395, "learning_rate": 5.230397229864515e-06, "loss": 0.694, "step": 2078 }, { "epoch": 2.0067567567567566, "grad_norm": 5.051416873931885, "learning_rate": 5.221238495850121e-06, "loss": 0.7027, "step": 2079 }, { "epoch": 2.0077220077220077, "grad_norm": 5.646663665771484, "learning_rate": 5.212084953762599e-06, "loss": 0.7197, "step": 2080 }, { "epoch": 2.008687258687259, "grad_norm": 5.704290390014648, "learning_rate": 5.202936613546894e-06, "loss": 0.7074, "step": 2081 }, { "epoch": 2.0096525096525095, "grad_norm": 5.16745662689209, "learning_rate": 5.193793485142301e-06, "loss": 0.6318, "step": 2082 }, { "epoch": 2.0106177606177607, "grad_norm": 5.382175445556641, "learning_rate": 5.1846555784824404e-06, "loss": 0.6801, "step": 2083 }, { "epoch": 2.011583011583012, "grad_norm": 5.676861763000488, "learning_rate": 5.17552290349528e-06, "loss": 0.7505, "step": 2084 }, { "epoch": 2.0125482625482625, "grad_norm": 4.259150505065918, "learning_rate": 5.166395470103092e-06, "loss": 0.7028, "step": 2085 }, { "epoch": 2.0135135135135136, "grad_norm": 4.849636077880859, "learning_rate": 5.157273288222455e-06, "loss": 0.7002, "step": 2086 }, { "epoch": 2.0144787644787643, "grad_norm": 5.219135761260986, "learning_rate": 5.1481563677642455e-06, "loss": 0.7052, "step": 2087 }, { "epoch": 2.0154440154440154, "grad_norm": 4.577015399932861, "learning_rate": 5.139044718633609e-06, "loss": 0.6767, "step": 2088 }, { "epoch": 2.0164092664092665, "grad_norm": 4.252452850341797, "learning_rate": 5.129938350729989e-06, "loss": 0.6946, "step": 2089 }, { "epoch": 2.0173745173745172, "grad_norm": 4.529285907745361, "learning_rate": 5.120837273947073e-06, "loss": 0.6331, "step": 2090 }, { "epoch": 2.0183397683397684, "grad_norm": 4.3095927238464355, "learning_rate": 5.111741498172808e-06, "loss": 0.6511, "step": 2091 }, { "epoch": 2.0193050193050195, "grad_norm": 4.548341274261475, "learning_rate": 5.102651033289369e-06, "loss": 0.7208, "step": 2092 }, { "epoch": 2.02027027027027, "grad_norm": 4.232816696166992, "learning_rate": 5.09356588917318e-06, "loss": 0.687, "step": 2093 }, { "epoch": 2.0212355212355213, "grad_norm": 4.521832466125488, "learning_rate": 5.084486075694871e-06, "loss": 0.7281, "step": 2094 }, { "epoch": 2.022200772200772, "grad_norm": 4.4892754554748535, "learning_rate": 5.075411602719284e-06, "loss": 0.6967, "step": 2095 }, { "epoch": 2.023166023166023, "grad_norm": 4.672778129577637, "learning_rate": 5.06634248010546e-06, "loss": 0.6521, "step": 2096 }, { "epoch": 2.0241312741312742, "grad_norm": 4.826206684112549, "learning_rate": 5.057278717706624e-06, "loss": 0.6482, "step": 2097 }, { "epoch": 2.025096525096525, "grad_norm": 4.64094877243042, "learning_rate": 5.0482203253701815e-06, "loss": 0.6018, "step": 2098 }, { "epoch": 2.026061776061776, "grad_norm": 4.916898250579834, "learning_rate": 5.0391673129377e-06, "loss": 0.6544, "step": 2099 }, { "epoch": 2.027027027027027, "grad_norm": 4.634807109832764, "learning_rate": 5.030119690244903e-06, "loss": 0.6411, "step": 2100 }, { "epoch": 2.027992277992278, "grad_norm": 4.873452186584473, "learning_rate": 5.02107746712166e-06, "loss": 0.6592, "step": 2101 }, { "epoch": 2.028957528957529, "grad_norm": 4.586275577545166, "learning_rate": 5.012040653391972e-06, "loss": 0.6816, "step": 2102 }, { "epoch": 2.02992277992278, "grad_norm": 4.615923881530762, "learning_rate": 5.003009258873963e-06, "loss": 0.6463, "step": 2103 }, { "epoch": 2.030888030888031, "grad_norm": 4.522491931915283, "learning_rate": 4.993983293379871e-06, "loss": 0.6184, "step": 2104 }, { "epoch": 2.031853281853282, "grad_norm": 4.397302627563477, "learning_rate": 4.984962766716034e-06, "loss": 0.6795, "step": 2105 }, { "epoch": 2.0328185328185326, "grad_norm": 4.632791042327881, "learning_rate": 4.975947688682881e-06, "loss": 0.6398, "step": 2106 }, { "epoch": 2.0337837837837838, "grad_norm": 4.260961532592773, "learning_rate": 4.9669380690749215e-06, "loss": 0.6514, "step": 2107 }, { "epoch": 2.034749034749035, "grad_norm": 5.073614597320557, "learning_rate": 4.957933917680735e-06, "loss": 0.6558, "step": 2108 }, { "epoch": 2.0357142857142856, "grad_norm": 4.8688483238220215, "learning_rate": 4.948935244282961e-06, "loss": 0.6295, "step": 2109 }, { "epoch": 2.0366795366795367, "grad_norm": 4.682999610900879, "learning_rate": 4.939942058658286e-06, "loss": 0.6618, "step": 2110 }, { "epoch": 2.037644787644788, "grad_norm": 4.4976983070373535, "learning_rate": 4.9309543705774335e-06, "loss": 0.6809, "step": 2111 }, { "epoch": 2.0386100386100385, "grad_norm": 4.920668601989746, "learning_rate": 4.921972189805154e-06, "loss": 0.7048, "step": 2112 }, { "epoch": 2.0395752895752897, "grad_norm": 4.667609214782715, "learning_rate": 4.912995526100216e-06, "loss": 0.6312, "step": 2113 }, { "epoch": 2.0405405405405403, "grad_norm": 4.698695182800293, "learning_rate": 4.904024389215401e-06, "loss": 0.6086, "step": 2114 }, { "epoch": 2.0415057915057915, "grad_norm": 4.660233974456787, "learning_rate": 4.895058788897469e-06, "loss": 0.6518, "step": 2115 }, { "epoch": 2.0424710424710426, "grad_norm": 5.051577091217041, "learning_rate": 4.886098734887177e-06, "loss": 0.6578, "step": 2116 }, { "epoch": 2.0434362934362933, "grad_norm": 4.889238357543945, "learning_rate": 4.87714423691925e-06, "loss": 0.6245, "step": 2117 }, { "epoch": 2.0444015444015444, "grad_norm": 4.660045623779297, "learning_rate": 4.868195304722391e-06, "loss": 0.695, "step": 2118 }, { "epoch": 2.0453667953667956, "grad_norm": 4.155313968658447, "learning_rate": 4.859251948019236e-06, "loss": 0.6629, "step": 2119 }, { "epoch": 2.0463320463320462, "grad_norm": 4.660876750946045, "learning_rate": 4.850314176526373e-06, "loss": 0.6384, "step": 2120 }, { "epoch": 2.0472972972972974, "grad_norm": 4.83228874206543, "learning_rate": 4.8413819999543185e-06, "loss": 0.6539, "step": 2121 }, { "epoch": 2.048262548262548, "grad_norm": 4.625710964202881, "learning_rate": 4.8324554280075255e-06, "loss": 0.7382, "step": 2122 }, { "epoch": 2.049227799227799, "grad_norm": 4.783905506134033, "learning_rate": 4.823534470384333e-06, "loss": 0.7325, "step": 2123 }, { "epoch": 2.0501930501930503, "grad_norm": 4.6731038093566895, "learning_rate": 4.814619136776998e-06, "loss": 0.686, "step": 2124 }, { "epoch": 2.051158301158301, "grad_norm": 5.126330852508545, "learning_rate": 4.805709436871659e-06, "loss": 0.65, "step": 2125 }, { "epoch": 2.052123552123552, "grad_norm": 5.164779186248779, "learning_rate": 4.796805380348336e-06, "loss": 0.6678, "step": 2126 }, { "epoch": 2.0530888030888033, "grad_norm": 4.569568157196045, "learning_rate": 4.78790697688093e-06, "loss": 0.637, "step": 2127 }, { "epoch": 2.054054054054054, "grad_norm": 5.218832015991211, "learning_rate": 4.7790142361371735e-06, "loss": 0.6875, "step": 2128 }, { "epoch": 2.055019305019305, "grad_norm": 5.137186050415039, "learning_rate": 4.77012716777867e-06, "loss": 0.6989, "step": 2129 }, { "epoch": 2.0559845559845558, "grad_norm": 4.738580226898193, "learning_rate": 4.761245781460846e-06, "loss": 0.6949, "step": 2130 }, { "epoch": 2.056949806949807, "grad_norm": 5.230045795440674, "learning_rate": 4.752370086832971e-06, "loss": 0.6575, "step": 2131 }, { "epoch": 2.057915057915058, "grad_norm": 4.877504348754883, "learning_rate": 4.743500093538111e-06, "loss": 0.7114, "step": 2132 }, { "epoch": 2.0588803088803087, "grad_norm": 4.573655605316162, "learning_rate": 4.73463581121315e-06, "loss": 0.6913, "step": 2133 }, { "epoch": 2.05984555984556, "grad_norm": 4.723101615905762, "learning_rate": 4.725777249488761e-06, "loss": 0.674, "step": 2134 }, { "epoch": 2.060810810810811, "grad_norm": 4.4217023849487305, "learning_rate": 4.716924417989413e-06, "loss": 0.6296, "step": 2135 }, { "epoch": 2.0617760617760617, "grad_norm": 4.694399356842041, "learning_rate": 4.7080773263333405e-06, "loss": 0.6748, "step": 2136 }, { "epoch": 2.062741312741313, "grad_norm": 4.912079334259033, "learning_rate": 4.699235984132535e-06, "loss": 0.6427, "step": 2137 }, { "epoch": 2.063706563706564, "grad_norm": 4.5845723152160645, "learning_rate": 4.690400400992752e-06, "loss": 0.6512, "step": 2138 }, { "epoch": 2.0646718146718146, "grad_norm": 4.9580512046813965, "learning_rate": 4.681570586513492e-06, "loss": 0.6706, "step": 2139 }, { "epoch": 2.0656370656370657, "grad_norm": 4.75042200088501, "learning_rate": 4.672746550287985e-06, "loss": 0.6449, "step": 2140 }, { "epoch": 2.0666023166023164, "grad_norm": 5.1743245124816895, "learning_rate": 4.663928301903175e-06, "loss": 0.6604, "step": 2141 }, { "epoch": 2.0675675675675675, "grad_norm": 4.8098955154418945, "learning_rate": 4.655115850939722e-06, "loss": 0.6772, "step": 2142 }, { "epoch": 2.0685328185328187, "grad_norm": 5.1585259437561035, "learning_rate": 4.6463092069719996e-06, "loss": 0.6969, "step": 2143 }, { "epoch": 2.0694980694980694, "grad_norm": 4.868752956390381, "learning_rate": 4.637508379568057e-06, "loss": 0.6561, "step": 2144 }, { "epoch": 2.0704633204633205, "grad_norm": 4.869829177856445, "learning_rate": 4.628713378289634e-06, "loss": 0.6787, "step": 2145 }, { "epoch": 2.0714285714285716, "grad_norm": 4.9792633056640625, "learning_rate": 4.619924212692125e-06, "loss": 0.6789, "step": 2146 }, { "epoch": 2.0723938223938223, "grad_norm": 4.65444278717041, "learning_rate": 4.611140892324606e-06, "loss": 0.6409, "step": 2147 }, { "epoch": 2.0733590733590734, "grad_norm": 4.609977722167969, "learning_rate": 4.602363426729788e-06, "loss": 0.6254, "step": 2148 }, { "epoch": 2.074324324324324, "grad_norm": 4.691058158874512, "learning_rate": 4.593591825444028e-06, "loss": 0.6442, "step": 2149 }, { "epoch": 2.0752895752895753, "grad_norm": 4.959919452667236, "learning_rate": 4.584826097997297e-06, "loss": 0.6973, "step": 2150 }, { "epoch": 2.0762548262548264, "grad_norm": 4.790925979614258, "learning_rate": 4.576066253913209e-06, "loss": 0.6681, "step": 2151 }, { "epoch": 2.077220077220077, "grad_norm": 5.478535175323486, "learning_rate": 4.567312302708965e-06, "loss": 0.6703, "step": 2152 }, { "epoch": 2.078185328185328, "grad_norm": 4.707211494445801, "learning_rate": 4.558564253895374e-06, "loss": 0.6848, "step": 2153 }, { "epoch": 2.0791505791505793, "grad_norm": 5.1434006690979, "learning_rate": 4.549822116976827e-06, "loss": 0.6834, "step": 2154 }, { "epoch": 2.08011583011583, "grad_norm": 5.3403730392456055, "learning_rate": 4.5410859014512975e-06, "loss": 0.6803, "step": 2155 }, { "epoch": 2.081081081081081, "grad_norm": 4.581385135650635, "learning_rate": 4.532355616810321e-06, "loss": 0.6413, "step": 2156 }, { "epoch": 2.082046332046332, "grad_norm": 4.807558536529541, "learning_rate": 4.523631272538992e-06, "loss": 0.6617, "step": 2157 }, { "epoch": 2.083011583011583, "grad_norm": 4.22251033782959, "learning_rate": 4.514912878115948e-06, "loss": 0.6829, "step": 2158 }, { "epoch": 2.083976833976834, "grad_norm": 4.495012283325195, "learning_rate": 4.5062004430133655e-06, "loss": 0.7071, "step": 2159 }, { "epoch": 2.0849420849420848, "grad_norm": 4.16301965713501, "learning_rate": 4.497493976696945e-06, "loss": 0.6848, "step": 2160 }, { "epoch": 2.085907335907336, "grad_norm": 5.111375331878662, "learning_rate": 4.4887934886259025e-06, "loss": 0.6817, "step": 2161 }, { "epoch": 2.086872586872587, "grad_norm": 4.352403163909912, "learning_rate": 4.480098988252958e-06, "loss": 0.7081, "step": 2162 }, { "epoch": 2.0878378378378377, "grad_norm": 4.8227152824401855, "learning_rate": 4.471410485024327e-06, "loss": 0.6611, "step": 2163 }, { "epoch": 2.088803088803089, "grad_norm": 4.887004375457764, "learning_rate": 4.4627279883797096e-06, "loss": 0.7235, "step": 2164 }, { "epoch": 2.08976833976834, "grad_norm": 4.713346481323242, "learning_rate": 4.454051507752279e-06, "loss": 0.6352, "step": 2165 }, { "epoch": 2.0907335907335907, "grad_norm": 4.931870460510254, "learning_rate": 4.445381052568674e-06, "loss": 0.6878, "step": 2166 }, { "epoch": 2.091698841698842, "grad_norm": 4.754703044891357, "learning_rate": 4.436716632248985e-06, "loss": 0.6186, "step": 2167 }, { "epoch": 2.0926640926640925, "grad_norm": 5.296566486358643, "learning_rate": 4.4280582562067465e-06, "loss": 0.6907, "step": 2168 }, { "epoch": 2.0936293436293436, "grad_norm": 4.916853427886963, "learning_rate": 4.419405933848927e-06, "loss": 0.6636, "step": 2169 }, { "epoch": 2.0945945945945947, "grad_norm": 4.9489288330078125, "learning_rate": 4.410759674575915e-06, "loss": 0.7377, "step": 2170 }, { "epoch": 2.0955598455598454, "grad_norm": 5.000365257263184, "learning_rate": 4.4021194877815165e-06, "loss": 0.6841, "step": 2171 }, { "epoch": 2.0965250965250966, "grad_norm": 4.6656575202941895, "learning_rate": 4.393485382852936e-06, "loss": 0.7516, "step": 2172 }, { "epoch": 2.0974903474903477, "grad_norm": 4.6403021812438965, "learning_rate": 4.384857369170772e-06, "loss": 0.6799, "step": 2173 }, { "epoch": 2.0984555984555984, "grad_norm": 4.789555549621582, "learning_rate": 4.376235456109003e-06, "loss": 0.6836, "step": 2174 }, { "epoch": 2.0994208494208495, "grad_norm": 4.911081314086914, "learning_rate": 4.3676196530349845e-06, "loss": 0.6228, "step": 2175 }, { "epoch": 2.1003861003861, "grad_norm": 4.8095879554748535, "learning_rate": 4.359009969309429e-06, "loss": 0.6683, "step": 2176 }, { "epoch": 2.1013513513513513, "grad_norm": 5.17523717880249, "learning_rate": 4.350406414286401e-06, "loss": 0.7043, "step": 2177 }, { "epoch": 2.1023166023166024, "grad_norm": 4.903587818145752, "learning_rate": 4.341808997313308e-06, "loss": 0.6637, "step": 2178 }, { "epoch": 2.103281853281853, "grad_norm": 4.904983043670654, "learning_rate": 4.3332177277308885e-06, "loss": 0.7013, "step": 2179 }, { "epoch": 2.1042471042471043, "grad_norm": 5.028541564941406, "learning_rate": 4.324632614873199e-06, "loss": 0.6973, "step": 2180 }, { "epoch": 2.1052123552123554, "grad_norm": 4.6789093017578125, "learning_rate": 4.316053668067619e-06, "loss": 0.6755, "step": 2181 }, { "epoch": 2.106177606177606, "grad_norm": 5.516636848449707, "learning_rate": 4.3074808966348105e-06, "loss": 0.6592, "step": 2182 }, { "epoch": 2.107142857142857, "grad_norm": 4.884287357330322, "learning_rate": 4.298914309888736e-06, "loss": 0.6969, "step": 2183 }, { "epoch": 2.108108108108108, "grad_norm": 4.955660343170166, "learning_rate": 4.290353917136639e-06, "loss": 0.6357, "step": 2184 }, { "epoch": 2.109073359073359, "grad_norm": 4.877090930938721, "learning_rate": 4.281799727679039e-06, "loss": 0.6607, "step": 2185 }, { "epoch": 2.11003861003861, "grad_norm": 4.54473876953125, "learning_rate": 4.2732517508097016e-06, "loss": 0.6441, "step": 2186 }, { "epoch": 2.111003861003861, "grad_norm": 4.789745330810547, "learning_rate": 4.264709995815651e-06, "loss": 0.6652, "step": 2187 }, { "epoch": 2.111969111969112, "grad_norm": 4.660126686096191, "learning_rate": 4.25617447197715e-06, "loss": 0.6766, "step": 2188 }, { "epoch": 2.112934362934363, "grad_norm": 5.456777095794678, "learning_rate": 4.247645188567698e-06, "loss": 0.674, "step": 2189 }, { "epoch": 2.113899613899614, "grad_norm": 4.886148452758789, "learning_rate": 4.2391221548540094e-06, "loss": 0.6653, "step": 2190 }, { "epoch": 2.114864864864865, "grad_norm": 4.867854595184326, "learning_rate": 4.230605380096001e-06, "loss": 0.7063, "step": 2191 }, { "epoch": 2.115830115830116, "grad_norm": 5.060470104217529, "learning_rate": 4.222094873546797e-06, "loss": 0.6619, "step": 2192 }, { "epoch": 2.1167953667953667, "grad_norm": 4.72755241394043, "learning_rate": 4.213590644452717e-06, "loss": 0.7273, "step": 2193 }, { "epoch": 2.117760617760618, "grad_norm": 5.060878276824951, "learning_rate": 4.205092702053257e-06, "loss": 0.685, "step": 2194 }, { "epoch": 2.1187258687258685, "grad_norm": 4.600476264953613, "learning_rate": 4.1966010555810696e-06, "loss": 0.7129, "step": 2195 }, { "epoch": 2.1196911196911197, "grad_norm": 4.683096408843994, "learning_rate": 4.18811571426198e-06, "loss": 0.7142, "step": 2196 }, { "epoch": 2.120656370656371, "grad_norm": 4.833113193511963, "learning_rate": 4.179636687314967e-06, "loss": 0.7439, "step": 2197 }, { "epoch": 2.1216216216216215, "grad_norm": 4.970816135406494, "learning_rate": 4.171163983952144e-06, "loss": 0.6971, "step": 2198 }, { "epoch": 2.1225868725868726, "grad_norm": 4.352797985076904, "learning_rate": 4.1626976133787445e-06, "loss": 0.7058, "step": 2199 }, { "epoch": 2.1235521235521237, "grad_norm": 4.643060684204102, "learning_rate": 4.154237584793132e-06, "loss": 0.6841, "step": 2200 }, { "epoch": 2.1245173745173744, "grad_norm": 4.573018550872803, "learning_rate": 4.145783907386782e-06, "loss": 0.7119, "step": 2201 }, { "epoch": 2.1254826254826256, "grad_norm": 4.699937343597412, "learning_rate": 4.1373365903442665e-06, "loss": 0.667, "step": 2202 }, { "epoch": 2.1264478764478763, "grad_norm": 4.921900749206543, "learning_rate": 4.1288956428432455e-06, "loss": 0.6998, "step": 2203 }, { "epoch": 2.1274131274131274, "grad_norm": 4.901946544647217, "learning_rate": 4.120461074054455e-06, "loss": 0.6726, "step": 2204 }, { "epoch": 2.1283783783783785, "grad_norm": 5.032048225402832, "learning_rate": 4.112032893141706e-06, "loss": 0.7078, "step": 2205 }, { "epoch": 2.129343629343629, "grad_norm": 4.5773162841796875, "learning_rate": 4.1036111092618725e-06, "loss": 0.6699, "step": 2206 }, { "epoch": 2.1303088803088803, "grad_norm": 5.37986421585083, "learning_rate": 4.095195731564878e-06, "loss": 0.7211, "step": 2207 }, { "epoch": 2.1312741312741315, "grad_norm": 4.763434410095215, "learning_rate": 4.086786769193673e-06, "loss": 0.6721, "step": 2208 }, { "epoch": 2.132239382239382, "grad_norm": 4.474204063415527, "learning_rate": 4.0783842312842506e-06, "loss": 0.7242, "step": 2209 }, { "epoch": 2.1332046332046333, "grad_norm": 4.982334136962891, "learning_rate": 4.069988126965625e-06, "loss": 0.6732, "step": 2210 }, { "epoch": 2.134169884169884, "grad_norm": 5.002752304077148, "learning_rate": 4.061598465359815e-06, "loss": 0.662, "step": 2211 }, { "epoch": 2.135135135135135, "grad_norm": 5.2900238037109375, "learning_rate": 4.053215255581843e-06, "loss": 0.6901, "step": 2212 }, { "epoch": 2.136100386100386, "grad_norm": 4.898741245269775, "learning_rate": 4.044838506739711e-06, "loss": 0.7002, "step": 2213 }, { "epoch": 2.137065637065637, "grad_norm": 4.9963884353637695, "learning_rate": 4.03646822793442e-06, "loss": 0.6949, "step": 2214 }, { "epoch": 2.138030888030888, "grad_norm": 4.586988925933838, "learning_rate": 4.028104428259929e-06, "loss": 0.6962, "step": 2215 }, { "epoch": 2.138996138996139, "grad_norm": 4.49263334274292, "learning_rate": 4.019747116803164e-06, "loss": 0.6961, "step": 2216 }, { "epoch": 2.13996138996139, "grad_norm": 4.701002597808838, "learning_rate": 4.011396302643989e-06, "loss": 0.7185, "step": 2217 }, { "epoch": 2.140926640926641, "grad_norm": 4.800919532775879, "learning_rate": 4.003051994855226e-06, "loss": 0.6886, "step": 2218 }, { "epoch": 2.141891891891892, "grad_norm": 4.730530738830566, "learning_rate": 3.99471420250262e-06, "loss": 0.6515, "step": 2219 }, { "epoch": 2.142857142857143, "grad_norm": 5.242630958557129, "learning_rate": 3.986382934644836e-06, "loss": 0.694, "step": 2220 }, { "epoch": 2.143822393822394, "grad_norm": 4.6426005363464355, "learning_rate": 3.978058200333453e-06, "loss": 0.68, "step": 2221 }, { "epoch": 2.1447876447876446, "grad_norm": 4.800093650817871, "learning_rate": 3.969740008612951e-06, "loss": 0.6869, "step": 2222 }, { "epoch": 2.1457528957528957, "grad_norm": 4.5579681396484375, "learning_rate": 3.961428368520702e-06, "loss": 0.6762, "step": 2223 }, { "epoch": 2.146718146718147, "grad_norm": 5.217144012451172, "learning_rate": 3.953123289086958e-06, "loss": 0.6761, "step": 2224 }, { "epoch": 2.1476833976833976, "grad_norm": 4.891463756561279, "learning_rate": 3.944824779334845e-06, "loss": 0.6881, "step": 2225 }, { "epoch": 2.1486486486486487, "grad_norm": 4.874589920043945, "learning_rate": 3.936532848280351e-06, "loss": 0.7041, "step": 2226 }, { "epoch": 2.1496138996139, "grad_norm": 4.724242210388184, "learning_rate": 3.928247504932317e-06, "loss": 0.7238, "step": 2227 }, { "epoch": 2.1505791505791505, "grad_norm": 4.630641460418701, "learning_rate": 3.919968758292425e-06, "loss": 0.6537, "step": 2228 }, { "epoch": 2.1515444015444016, "grad_norm": 4.487358093261719, "learning_rate": 3.9116966173551915e-06, "loss": 0.671, "step": 2229 }, { "epoch": 2.1525096525096523, "grad_norm": 4.850062370300293, "learning_rate": 3.903431091107957e-06, "loss": 0.7052, "step": 2230 }, { "epoch": 2.1534749034749034, "grad_norm": 4.9462504386901855, "learning_rate": 3.895172188530871e-06, "loss": 0.7052, "step": 2231 }, { "epoch": 2.1544401544401546, "grad_norm": 4.509797096252441, "learning_rate": 3.8869199185968944e-06, "loss": 0.7541, "step": 2232 }, { "epoch": 2.1554054054054053, "grad_norm": 4.4127302169799805, "learning_rate": 3.878674290271773e-06, "loss": 0.6825, "step": 2233 }, { "epoch": 2.1563706563706564, "grad_norm": 5.578569412231445, "learning_rate": 3.8704353125140436e-06, "loss": 0.6822, "step": 2234 }, { "epoch": 2.1573359073359075, "grad_norm": 4.80642032623291, "learning_rate": 3.862202994275015e-06, "loss": 0.6905, "step": 2235 }, { "epoch": 2.158301158301158, "grad_norm": 4.7371134757995605, "learning_rate": 3.853977344498761e-06, "loss": 0.6829, "step": 2236 }, { "epoch": 2.1592664092664093, "grad_norm": 5.159573078155518, "learning_rate": 3.845758372122107e-06, "loss": 0.7233, "step": 2237 }, { "epoch": 2.16023166023166, "grad_norm": 4.449493885040283, "learning_rate": 3.83754608607463e-06, "loss": 0.6616, "step": 2238 }, { "epoch": 2.161196911196911, "grad_norm": 4.797781467437744, "learning_rate": 3.82934049527864e-06, "loss": 0.6864, "step": 2239 }, { "epoch": 2.1621621621621623, "grad_norm": 4.917870998382568, "learning_rate": 3.821141608649169e-06, "loss": 0.7103, "step": 2240 }, { "epoch": 2.163127413127413, "grad_norm": 4.734244346618652, "learning_rate": 3.812949435093971e-06, "loss": 0.6438, "step": 2241 }, { "epoch": 2.164092664092664, "grad_norm": 5.052276611328125, "learning_rate": 3.804763983513503e-06, "loss": 0.7217, "step": 2242 }, { "epoch": 2.1650579150579152, "grad_norm": 4.778468608856201, "learning_rate": 3.7965852628009204e-06, "loss": 0.6653, "step": 2243 }, { "epoch": 2.166023166023166, "grad_norm": 5.127320289611816, "learning_rate": 3.788413281842065e-06, "loss": 0.7194, "step": 2244 }, { "epoch": 2.166988416988417, "grad_norm": 4.7970805168151855, "learning_rate": 3.780248049515456e-06, "loss": 0.6861, "step": 2245 }, { "epoch": 2.167953667953668, "grad_norm": 5.094338893890381, "learning_rate": 3.7720895746922812e-06, "loss": 0.7325, "step": 2246 }, { "epoch": 2.168918918918919, "grad_norm": 4.787800312042236, "learning_rate": 3.7639378662363823e-06, "loss": 0.682, "step": 2247 }, { "epoch": 2.16988416988417, "grad_norm": 5.029555320739746, "learning_rate": 3.7557929330042654e-06, "loss": 0.7349, "step": 2248 }, { "epoch": 2.1708494208494207, "grad_norm": 4.596803188323975, "learning_rate": 3.747654783845052e-06, "loss": 0.702, "step": 2249 }, { "epoch": 2.171814671814672, "grad_norm": 4.6137566566467285, "learning_rate": 3.739523427600509e-06, "loss": 0.6399, "step": 2250 }, { "epoch": 2.172779922779923, "grad_norm": 4.858090877532959, "learning_rate": 3.7313988731050156e-06, "loss": 0.7122, "step": 2251 }, { "epoch": 2.1737451737451736, "grad_norm": 4.909048080444336, "learning_rate": 3.723281129185574e-06, "loss": 0.7443, "step": 2252 }, { "epoch": 2.1747104247104247, "grad_norm": 4.852362632751465, "learning_rate": 3.7151702046617677e-06, "loss": 0.6985, "step": 2253 }, { "epoch": 2.175675675675676, "grad_norm": 4.483128547668457, "learning_rate": 3.7070661083457838e-06, "loss": 0.6975, "step": 2254 }, { "epoch": 2.1766409266409266, "grad_norm": 4.893508434295654, "learning_rate": 3.698968849042386e-06, "loss": 0.7307, "step": 2255 }, { "epoch": 2.1776061776061777, "grad_norm": 5.04038667678833, "learning_rate": 3.690878435548917e-06, "loss": 0.7117, "step": 2256 }, { "epoch": 2.1785714285714284, "grad_norm": 4.4502949714660645, "learning_rate": 3.682794876655278e-06, "loss": 0.6626, "step": 2257 }, { "epoch": 2.1795366795366795, "grad_norm": 4.678581237792969, "learning_rate": 3.6747181811439146e-06, "loss": 0.6716, "step": 2258 }, { "epoch": 2.1805019305019306, "grad_norm": 4.655972480773926, "learning_rate": 3.6666483577898215e-06, "loss": 0.6374, "step": 2259 }, { "epoch": 2.1814671814671813, "grad_norm": 5.271326541900635, "learning_rate": 3.658585415360537e-06, "loss": 0.6766, "step": 2260 }, { "epoch": 2.1824324324324325, "grad_norm": 4.562073707580566, "learning_rate": 3.650529362616113e-06, "loss": 0.7137, "step": 2261 }, { "epoch": 2.1833976833976836, "grad_norm": 5.000593185424805, "learning_rate": 3.6424802083091124e-06, "loss": 0.6968, "step": 2262 }, { "epoch": 2.1843629343629343, "grad_norm": 4.962578773498535, "learning_rate": 3.634437961184608e-06, "loss": 0.6495, "step": 2263 }, { "epoch": 2.1853281853281854, "grad_norm": 4.9937262535095215, "learning_rate": 3.6264026299801758e-06, "loss": 0.7027, "step": 2264 }, { "epoch": 2.186293436293436, "grad_norm": 4.739039897918701, "learning_rate": 3.618374223425868e-06, "loss": 0.7133, "step": 2265 }, { "epoch": 2.187258687258687, "grad_norm": 5.114686965942383, "learning_rate": 3.6103527502442183e-06, "loss": 0.7089, "step": 2266 }, { "epoch": 2.1882239382239383, "grad_norm": 4.9151835441589355, "learning_rate": 3.6023382191502175e-06, "loss": 0.663, "step": 2267 }, { "epoch": 2.189189189189189, "grad_norm": 5.096463203430176, "learning_rate": 3.5943306388513333e-06, "loss": 0.7173, "step": 2268 }, { "epoch": 2.19015444015444, "grad_norm": 4.799716472625732, "learning_rate": 3.586330018047465e-06, "loss": 0.6574, "step": 2269 }, { "epoch": 2.1911196911196913, "grad_norm": 4.545619010925293, "learning_rate": 3.57833636543096e-06, "loss": 0.751, "step": 2270 }, { "epoch": 2.192084942084942, "grad_norm": 4.805863380432129, "learning_rate": 3.570349689686583e-06, "loss": 0.6806, "step": 2271 }, { "epoch": 2.193050193050193, "grad_norm": 4.8336286544799805, "learning_rate": 3.5623699994915363e-06, "loss": 0.6843, "step": 2272 }, { "epoch": 2.1940154440154442, "grad_norm": 4.659176826477051, "learning_rate": 3.554397303515421e-06, "loss": 0.6987, "step": 2273 }, { "epoch": 2.194980694980695, "grad_norm": 4.7669830322265625, "learning_rate": 3.5464316104202403e-06, "loss": 0.7378, "step": 2274 }, { "epoch": 2.195945945945946, "grad_norm": 5.27888298034668, "learning_rate": 3.5384729288603912e-06, "loss": 0.6762, "step": 2275 }, { "epoch": 2.1969111969111967, "grad_norm": 4.9906535148620605, "learning_rate": 3.5305212674826526e-06, "loss": 0.7317, "step": 2276 }, { "epoch": 2.197876447876448, "grad_norm": 4.950737476348877, "learning_rate": 3.522576634926176e-06, "loss": 0.7051, "step": 2277 }, { "epoch": 2.198841698841699, "grad_norm": 5.027462005615234, "learning_rate": 3.514639039822475e-06, "loss": 0.6887, "step": 2278 }, { "epoch": 2.1998069498069497, "grad_norm": 5.140257358551025, "learning_rate": 3.506708490795424e-06, "loss": 0.6656, "step": 2279 }, { "epoch": 2.200772200772201, "grad_norm": 5.102508544921875, "learning_rate": 3.498784996461225e-06, "loss": 0.6501, "step": 2280 }, { "epoch": 2.201737451737452, "grad_norm": 5.576716899871826, "learning_rate": 3.490868565428439e-06, "loss": 0.7047, "step": 2281 }, { "epoch": 2.2027027027027026, "grad_norm": 5.020506381988525, "learning_rate": 3.4829592062979358e-06, "loss": 0.7033, "step": 2282 }, { "epoch": 2.2036679536679538, "grad_norm": 5.018290996551514, "learning_rate": 3.475056927662912e-06, "loss": 0.7177, "step": 2283 }, { "epoch": 2.2046332046332044, "grad_norm": 5.415291786193848, "learning_rate": 3.467161738108855e-06, "loss": 0.7378, "step": 2284 }, { "epoch": 2.2055984555984556, "grad_norm": 4.696223735809326, "learning_rate": 3.4592736462135734e-06, "loss": 0.6622, "step": 2285 }, { "epoch": 2.2065637065637067, "grad_norm": 5.0370635986328125, "learning_rate": 3.4513926605471504e-06, "loss": 0.7077, "step": 2286 }, { "epoch": 2.2075289575289574, "grad_norm": 5.2918782234191895, "learning_rate": 3.443518789671949e-06, "loss": 0.7347, "step": 2287 }, { "epoch": 2.2084942084942085, "grad_norm": 4.940263271331787, "learning_rate": 3.4356520421426063e-06, "loss": 0.7493, "step": 2288 }, { "epoch": 2.2094594594594597, "grad_norm": 4.90836763381958, "learning_rate": 3.427792426506017e-06, "loss": 0.6653, "step": 2289 }, { "epoch": 2.2104247104247103, "grad_norm": 4.963898658752441, "learning_rate": 3.4199399513013297e-06, "loss": 0.709, "step": 2290 }, { "epoch": 2.2113899613899615, "grad_norm": 4.532299041748047, "learning_rate": 3.412094625059934e-06, "loss": 0.6859, "step": 2291 }, { "epoch": 2.212355212355212, "grad_norm": 5.141048908233643, "learning_rate": 3.4042564563054525e-06, "loss": 0.68, "step": 2292 }, { "epoch": 2.2133204633204633, "grad_norm": 5.234963417053223, "learning_rate": 3.3964254535537322e-06, "loss": 0.7707, "step": 2293 }, { "epoch": 2.2142857142857144, "grad_norm": 5.121557235717773, "learning_rate": 3.388601625312833e-06, "loss": 0.7329, "step": 2294 }, { "epoch": 2.215250965250965, "grad_norm": 4.665938377380371, "learning_rate": 3.3807849800830207e-06, "loss": 0.7215, "step": 2295 }, { "epoch": 2.2162162162162162, "grad_norm": 5.325049877166748, "learning_rate": 3.3729755263567577e-06, "loss": 0.7204, "step": 2296 }, { "epoch": 2.2171814671814674, "grad_norm": 4.914840221405029, "learning_rate": 3.3651732726186924e-06, "loss": 0.7348, "step": 2297 }, { "epoch": 2.218146718146718, "grad_norm": 4.860520362854004, "learning_rate": 3.3573782273456523e-06, "loss": 0.7138, "step": 2298 }, { "epoch": 2.219111969111969, "grad_norm": 5.827704429626465, "learning_rate": 3.3495903990066315e-06, "loss": 0.6851, "step": 2299 }, { "epoch": 2.2200772200772203, "grad_norm": 4.654504776000977, "learning_rate": 3.341809796062783e-06, "loss": 0.6777, "step": 2300 }, { "epoch": 2.221042471042471, "grad_norm": 5.1855268478393555, "learning_rate": 3.33403642696741e-06, "loss": 0.734, "step": 2301 }, { "epoch": 2.222007722007722, "grad_norm": 4.997435092926025, "learning_rate": 3.3262703001659567e-06, "loss": 0.7135, "step": 2302 }, { "epoch": 2.222972972972973, "grad_norm": 5.055532932281494, "learning_rate": 3.318511424096e-06, "loss": 0.7069, "step": 2303 }, { "epoch": 2.223938223938224, "grad_norm": 5.179004192352295, "learning_rate": 3.310759807187238e-06, "loss": 0.7401, "step": 2304 }, { "epoch": 2.224903474903475, "grad_norm": 6.071235656738281, "learning_rate": 3.3030154578614783e-06, "loss": 0.7607, "step": 2305 }, { "epoch": 2.2258687258687258, "grad_norm": 4.733060836791992, "learning_rate": 3.2952783845326465e-06, "loss": 0.6579, "step": 2306 }, { "epoch": 2.226833976833977, "grad_norm": 4.739913463592529, "learning_rate": 3.287548595606742e-06, "loss": 0.722, "step": 2307 }, { "epoch": 2.2277992277992276, "grad_norm": 5.429046154022217, "learning_rate": 3.2798260994818663e-06, "loss": 0.6672, "step": 2308 }, { "epoch": 2.2287644787644787, "grad_norm": 4.519412994384766, "learning_rate": 3.2721109045481867e-06, "loss": 0.6304, "step": 2309 }, { "epoch": 2.22972972972973, "grad_norm": 4.585423946380615, "learning_rate": 3.2644030191879563e-06, "loss": 0.6756, "step": 2310 }, { "epoch": 2.2306949806949805, "grad_norm": 4.876979351043701, "learning_rate": 3.256702451775461e-06, "loss": 0.7007, "step": 2311 }, { "epoch": 2.2316602316602316, "grad_norm": 5.468658447265625, "learning_rate": 3.2490092106770545e-06, "loss": 0.7515, "step": 2312 }, { "epoch": 2.2326254826254828, "grad_norm": 4.656357288360596, "learning_rate": 3.2413233042511203e-06, "loss": 0.7121, "step": 2313 }, { "epoch": 2.2335907335907335, "grad_norm": 4.834227085113525, "learning_rate": 3.2336447408480874e-06, "loss": 0.6866, "step": 2314 }, { "epoch": 2.2345559845559846, "grad_norm": 4.927269458770752, "learning_rate": 3.2259735288103956e-06, "loss": 0.7328, "step": 2315 }, { "epoch": 2.2355212355212357, "grad_norm": 4.478086471557617, "learning_rate": 3.218309676472492e-06, "loss": 0.651, "step": 2316 }, { "epoch": 2.2364864864864864, "grad_norm": 5.098201751708984, "learning_rate": 3.210653192160842e-06, "loss": 0.7066, "step": 2317 }, { "epoch": 2.2374517374517375, "grad_norm": 5.49979829788208, "learning_rate": 3.2030040841938927e-06, "loss": 0.7486, "step": 2318 }, { "epoch": 2.238416988416988, "grad_norm": 5.119729995727539, "learning_rate": 3.195362360882096e-06, "loss": 0.6676, "step": 2319 }, { "epoch": 2.2393822393822393, "grad_norm": 5.401087760925293, "learning_rate": 3.1877280305278568e-06, "loss": 0.7287, "step": 2320 }, { "epoch": 2.2403474903474905, "grad_norm": 4.851002216339111, "learning_rate": 3.1801011014255645e-06, "loss": 0.7232, "step": 2321 }, { "epoch": 2.241312741312741, "grad_norm": 5.39402437210083, "learning_rate": 3.1724815818615584e-06, "loss": 0.7458, "step": 2322 }, { "epoch": 2.2422779922779923, "grad_norm": 5.446462631225586, "learning_rate": 3.164869480114138e-06, "loss": 0.7149, "step": 2323 }, { "epoch": 2.2432432432432434, "grad_norm": 4.565616130828857, "learning_rate": 3.157264804453537e-06, "loss": 0.6811, "step": 2324 }, { "epoch": 2.244208494208494, "grad_norm": 5.289637088775635, "learning_rate": 3.1496675631419147e-06, "loss": 0.7427, "step": 2325 }, { "epoch": 2.2451737451737452, "grad_norm": 5.112846851348877, "learning_rate": 3.1420777644333613e-06, "loss": 0.7339, "step": 2326 }, { "epoch": 2.2461389961389964, "grad_norm": 5.118411540985107, "learning_rate": 3.134495416573884e-06, "loss": 0.6778, "step": 2327 }, { "epoch": 2.247104247104247, "grad_norm": 4.570213794708252, "learning_rate": 3.1269205278013903e-06, "loss": 0.6523, "step": 2328 }, { "epoch": 2.248069498069498, "grad_norm": 4.825456142425537, "learning_rate": 3.1193531063456785e-06, "loss": 0.6886, "step": 2329 }, { "epoch": 2.249034749034749, "grad_norm": 4.914466381072998, "learning_rate": 3.111793160428439e-06, "loss": 0.7074, "step": 2330 }, { "epoch": 2.25, "grad_norm": 5.233219146728516, "learning_rate": 3.1042406982632467e-06, "loss": 0.7019, "step": 2331 }, { "epoch": 2.250965250965251, "grad_norm": 5.194098472595215, "learning_rate": 3.0966957280555367e-06, "loss": 0.7208, "step": 2332 }, { "epoch": 2.251930501930502, "grad_norm": 4.804540157318115, "learning_rate": 3.08915825800261e-06, "loss": 0.7185, "step": 2333 }, { "epoch": 2.252895752895753, "grad_norm": 5.275886535644531, "learning_rate": 3.0816282962936083e-06, "loss": 0.7112, "step": 2334 }, { "epoch": 2.2538610038610036, "grad_norm": 5.222113132476807, "learning_rate": 3.074105851109531e-06, "loss": 0.6777, "step": 2335 }, { "epoch": 2.2548262548262548, "grad_norm": 4.733977317810059, "learning_rate": 3.0665909306232023e-06, "loss": 0.6635, "step": 2336 }, { "epoch": 2.255791505791506, "grad_norm": 4.772053241729736, "learning_rate": 3.0590835429992747e-06, "loss": 0.7097, "step": 2337 }, { "epoch": 2.2567567567567566, "grad_norm": 5.085040092468262, "learning_rate": 3.0515836963942056e-06, "loss": 0.715, "step": 2338 }, { "epoch": 2.2577220077220077, "grad_norm": 5.072872161865234, "learning_rate": 3.044091398956279e-06, "loss": 0.7268, "step": 2339 }, { "epoch": 2.258687258687259, "grad_norm": 4.757853031158447, "learning_rate": 3.0366066588255616e-06, "loss": 0.7046, "step": 2340 }, { "epoch": 2.2596525096525095, "grad_norm": 5.22769021987915, "learning_rate": 3.029129484133915e-06, "loss": 0.734, "step": 2341 }, { "epoch": 2.2606177606177607, "grad_norm": 4.824207305908203, "learning_rate": 3.02165988300498e-06, "loss": 0.7313, "step": 2342 }, { "epoch": 2.261583011583012, "grad_norm": 4.599949359893799, "learning_rate": 3.014197863554171e-06, "loss": 0.7436, "step": 2343 }, { "epoch": 2.2625482625482625, "grad_norm": 4.781702041625977, "learning_rate": 3.0067434338886623e-06, "loss": 0.7186, "step": 2344 }, { "epoch": 2.2635135135135136, "grad_norm": 4.790346145629883, "learning_rate": 2.9992966021073834e-06, "loss": 0.7361, "step": 2345 }, { "epoch": 2.2644787644787643, "grad_norm": 4.375913143157959, "learning_rate": 2.9918573763010106e-06, "loss": 0.7012, "step": 2346 }, { "epoch": 2.2654440154440154, "grad_norm": 4.515613079071045, "learning_rate": 2.984425764551954e-06, "loss": 0.6792, "step": 2347 }, { "epoch": 2.2664092664092665, "grad_norm": 5.460712909698486, "learning_rate": 2.9770017749343526e-06, "loss": 0.7358, "step": 2348 }, { "epoch": 2.2673745173745172, "grad_norm": 4.83195686340332, "learning_rate": 2.9695854155140648e-06, "loss": 0.6767, "step": 2349 }, { "epoch": 2.2683397683397684, "grad_norm": 4.955194473266602, "learning_rate": 2.9621766943486597e-06, "loss": 0.7982, "step": 2350 }, { "epoch": 2.2693050193050195, "grad_norm": 4.737084865570068, "learning_rate": 2.9547756194874046e-06, "loss": 0.7212, "step": 2351 }, { "epoch": 2.27027027027027, "grad_norm": 4.95520544052124, "learning_rate": 2.9473821989712626e-06, "loss": 0.7618, "step": 2352 }, { "epoch": 2.2712355212355213, "grad_norm": 5.034097194671631, "learning_rate": 2.9399964408328795e-06, "loss": 0.7017, "step": 2353 }, { "epoch": 2.2722007722007724, "grad_norm": 5.054235458374023, "learning_rate": 2.932618353096576e-06, "loss": 0.7333, "step": 2354 }, { "epoch": 2.273166023166023, "grad_norm": 4.866247653961182, "learning_rate": 2.925247943778339e-06, "loss": 0.7135, "step": 2355 }, { "epoch": 2.2741312741312742, "grad_norm": 4.6588640213012695, "learning_rate": 2.9178852208858146e-06, "loss": 0.7327, "step": 2356 }, { "epoch": 2.275096525096525, "grad_norm": 5.072292327880859, "learning_rate": 2.9105301924182972e-06, "loss": 0.7306, "step": 2357 }, { "epoch": 2.276061776061776, "grad_norm": 5.094712257385254, "learning_rate": 2.9031828663667194e-06, "loss": 0.7504, "step": 2358 }, { "epoch": 2.277027027027027, "grad_norm": 4.930572032928467, "learning_rate": 2.8958432507136492e-06, "loss": 0.7214, "step": 2359 }, { "epoch": 2.277992277992278, "grad_norm": 5.065821647644043, "learning_rate": 2.8885113534332742e-06, "loss": 0.6841, "step": 2360 }, { "epoch": 2.278957528957529, "grad_norm": 4.5738043785095215, "learning_rate": 2.8811871824913993e-06, "loss": 0.7253, "step": 2361 }, { "epoch": 2.2799227799227797, "grad_norm": 5.412510395050049, "learning_rate": 2.873870745845433e-06, "loss": 0.7534, "step": 2362 }, { "epoch": 2.280888030888031, "grad_norm": 5.102321624755859, "learning_rate": 2.8665620514443805e-06, "loss": 0.7576, "step": 2363 }, { "epoch": 2.281853281853282, "grad_norm": 4.808786869049072, "learning_rate": 2.8592611072288356e-06, "loss": 0.7234, "step": 2364 }, { "epoch": 2.2828185328185326, "grad_norm": 5.819774150848389, "learning_rate": 2.8519679211309746e-06, "loss": 0.7136, "step": 2365 }, { "epoch": 2.2837837837837838, "grad_norm": 5.263688564300537, "learning_rate": 2.8446825010745427e-06, "loss": 0.7136, "step": 2366 }, { "epoch": 2.284749034749035, "grad_norm": 5.062206745147705, "learning_rate": 2.8374048549748457e-06, "loss": 0.7149, "step": 2367 }, { "epoch": 2.2857142857142856, "grad_norm": 5.310696601867676, "learning_rate": 2.8301349907387464e-06, "loss": 0.7304, "step": 2368 }, { "epoch": 2.2866795366795367, "grad_norm": 5.125802516937256, "learning_rate": 2.8228729162646518e-06, "loss": 0.7395, "step": 2369 }, { "epoch": 2.287644787644788, "grad_norm": 4.984714508056641, "learning_rate": 2.8156186394425054e-06, "loss": 0.7306, "step": 2370 }, { "epoch": 2.2886100386100385, "grad_norm": 5.050137042999268, "learning_rate": 2.80837216815378e-06, "loss": 0.6995, "step": 2371 }, { "epoch": 2.2895752895752897, "grad_norm": 4.933598518371582, "learning_rate": 2.801133510271463e-06, "loss": 0.7144, "step": 2372 }, { "epoch": 2.2905405405405403, "grad_norm": 4.596554279327393, "learning_rate": 2.7939026736600673e-06, "loss": 0.7064, "step": 2373 }, { "epoch": 2.2915057915057915, "grad_norm": 5.07573127746582, "learning_rate": 2.7866796661755866e-06, "loss": 0.7067, "step": 2374 }, { "epoch": 2.2924710424710426, "grad_norm": 5.1375908851623535, "learning_rate": 2.7794644956655272e-06, "loss": 0.7172, "step": 2375 }, { "epoch": 2.2934362934362933, "grad_norm": 4.794432163238525, "learning_rate": 2.772257169968866e-06, "loss": 0.7097, "step": 2376 }, { "epoch": 2.2944015444015444, "grad_norm": 4.823713779449463, "learning_rate": 2.765057696916077e-06, "loss": 0.7465, "step": 2377 }, { "epoch": 2.2953667953667956, "grad_norm": 4.790879249572754, "learning_rate": 2.7578660843290796e-06, "loss": 0.7635, "step": 2378 }, { "epoch": 2.2963320463320462, "grad_norm": 5.0206298828125, "learning_rate": 2.7506823400212657e-06, "loss": 0.7172, "step": 2379 }, { "epoch": 2.2972972972972974, "grad_norm": 4.932159423828125, "learning_rate": 2.743506471797476e-06, "loss": 0.7068, "step": 2380 }, { "epoch": 2.2982625482625485, "grad_norm": 4.736297607421875, "learning_rate": 2.736338487453999e-06, "loss": 0.6866, "step": 2381 }, { "epoch": 2.299227799227799, "grad_norm": 4.674778461456299, "learning_rate": 2.7291783947785544e-06, "loss": 0.7199, "step": 2382 }, { "epoch": 2.3001930501930503, "grad_norm": 4.690591335296631, "learning_rate": 2.7220262015502796e-06, "loss": 0.69, "step": 2383 }, { "epoch": 2.301158301158301, "grad_norm": 4.7273268699646, "learning_rate": 2.714881915539739e-06, "loss": 0.7404, "step": 2384 }, { "epoch": 2.302123552123552, "grad_norm": 4.441293239593506, "learning_rate": 2.7077455445089075e-06, "loss": 0.7021, "step": 2385 }, { "epoch": 2.3030888030888033, "grad_norm": 5.465465545654297, "learning_rate": 2.70061709621116e-06, "loss": 0.7992, "step": 2386 }, { "epoch": 2.304054054054054, "grad_norm": 4.92411470413208, "learning_rate": 2.6934965783912524e-06, "loss": 0.7286, "step": 2387 }, { "epoch": 2.305019305019305, "grad_norm": 4.971828460693359, "learning_rate": 2.686383998785336e-06, "loss": 0.7351, "step": 2388 }, { "epoch": 2.3059845559845558, "grad_norm": 4.93446159362793, "learning_rate": 2.679279365120937e-06, "loss": 0.6786, "step": 2389 }, { "epoch": 2.306949806949807, "grad_norm": 4.765533924102783, "learning_rate": 2.6721826851169462e-06, "loss": 0.6873, "step": 2390 }, { "epoch": 2.307915057915058, "grad_norm": 5.01896858215332, "learning_rate": 2.665093966483615e-06, "loss": 0.7841, "step": 2391 }, { "epoch": 2.3088803088803087, "grad_norm": 4.646848678588867, "learning_rate": 2.6580132169225335e-06, "loss": 0.6952, "step": 2392 }, { "epoch": 2.30984555984556, "grad_norm": 5.088491439819336, "learning_rate": 2.650940444126654e-06, "loss": 0.7163, "step": 2393 }, { "epoch": 2.310810810810811, "grad_norm": 4.857966899871826, "learning_rate": 2.643875655780248e-06, "loss": 0.728, "step": 2394 }, { "epoch": 2.3117760617760617, "grad_norm": 5.659574031829834, "learning_rate": 2.6368188595589183e-06, "loss": 0.7304, "step": 2395 }, { "epoch": 2.312741312741313, "grad_norm": 4.86353874206543, "learning_rate": 2.6297700631295776e-06, "loss": 0.7056, "step": 2396 }, { "epoch": 2.313706563706564, "grad_norm": 4.285360813140869, "learning_rate": 2.6227292741504517e-06, "loss": 0.6679, "step": 2397 }, { "epoch": 2.3146718146718146, "grad_norm": 4.985572338104248, "learning_rate": 2.615696500271072e-06, "loss": 0.6983, "step": 2398 }, { "epoch": 2.3156370656370657, "grad_norm": 5.162213325500488, "learning_rate": 2.6086717491322545e-06, "loss": 0.714, "step": 2399 }, { "epoch": 2.3166023166023164, "grad_norm": 4.907494068145752, "learning_rate": 2.6016550283661037e-06, "loss": 0.7243, "step": 2400 }, { "epoch": 2.3175675675675675, "grad_norm": 4.992336750030518, "learning_rate": 2.5946463455959868e-06, "loss": 0.7138, "step": 2401 }, { "epoch": 2.3185328185328187, "grad_norm": 4.930575370788574, "learning_rate": 2.5876457084365593e-06, "loss": 0.7534, "step": 2402 }, { "epoch": 2.3194980694980694, "grad_norm": 4.708988189697266, "learning_rate": 2.5806531244937193e-06, "loss": 0.7042, "step": 2403 }, { "epoch": 2.3204633204633205, "grad_norm": 4.472352504730225, "learning_rate": 2.5736686013646226e-06, "loss": 0.7182, "step": 2404 }, { "epoch": 2.3214285714285716, "grad_norm": 4.982179164886475, "learning_rate": 2.5666921466376583e-06, "loss": 0.6801, "step": 2405 }, { "epoch": 2.3223938223938223, "grad_norm": 5.021693706512451, "learning_rate": 2.5597237678924637e-06, "loss": 0.7223, "step": 2406 }, { "epoch": 2.3233590733590734, "grad_norm": 4.943729400634766, "learning_rate": 2.5527634726998906e-06, "loss": 0.7029, "step": 2407 }, { "epoch": 2.3243243243243246, "grad_norm": 5.22763204574585, "learning_rate": 2.5458112686220136e-06, "loss": 0.7379, "step": 2408 }, { "epoch": 2.3252895752895753, "grad_norm": 5.044573783874512, "learning_rate": 2.5388671632121144e-06, "loss": 0.7035, "step": 2409 }, { "epoch": 2.3262548262548264, "grad_norm": 4.836273670196533, "learning_rate": 2.531931164014676e-06, "loss": 0.7247, "step": 2410 }, { "epoch": 2.327220077220077, "grad_norm": 5.079002857208252, "learning_rate": 2.5250032785653744e-06, "loss": 0.6896, "step": 2411 }, { "epoch": 2.328185328185328, "grad_norm": 5.008070945739746, "learning_rate": 2.518083514391073e-06, "loss": 0.7174, "step": 2412 }, { "epoch": 2.3291505791505793, "grad_norm": 4.878530502319336, "learning_rate": 2.511171879009808e-06, "loss": 0.7195, "step": 2413 }, { "epoch": 2.33011583011583, "grad_norm": 4.816442489624023, "learning_rate": 2.5042683799307864e-06, "loss": 0.6529, "step": 2414 }, { "epoch": 2.331081081081081, "grad_norm": 4.850268363952637, "learning_rate": 2.4973730246543736e-06, "loss": 0.7392, "step": 2415 }, { "epoch": 2.332046332046332, "grad_norm": 5.145356178283691, "learning_rate": 2.4904858206720892e-06, "loss": 0.7219, "step": 2416 }, { "epoch": 2.333011583011583, "grad_norm": 4.7043375968933105, "learning_rate": 2.483606775466595e-06, "loss": 0.6624, "step": 2417 }, { "epoch": 2.333976833976834, "grad_norm": 4.561450004577637, "learning_rate": 2.4767358965116893e-06, "loss": 0.6903, "step": 2418 }, { "epoch": 2.3349420849420848, "grad_norm": 4.79498291015625, "learning_rate": 2.4698731912722983e-06, "loss": 0.749, "step": 2419 }, { "epoch": 2.335907335907336, "grad_norm": 4.445528984069824, "learning_rate": 2.4630186672044687e-06, "loss": 0.7012, "step": 2420 }, { "epoch": 2.336872586872587, "grad_norm": 4.964884281158447, "learning_rate": 2.4561723317553556e-06, "loss": 0.7092, "step": 2421 }, { "epoch": 2.3378378378378377, "grad_norm": 4.606667995452881, "learning_rate": 2.4493341923632195e-06, "loss": 0.7299, "step": 2422 }, { "epoch": 2.338803088803089, "grad_norm": 5.229047775268555, "learning_rate": 2.4425042564574186e-06, "loss": 0.7875, "step": 2423 }, { "epoch": 2.33976833976834, "grad_norm": 5.045839786529541, "learning_rate": 2.435682531458393e-06, "loss": 0.7264, "step": 2424 }, { "epoch": 2.3407335907335907, "grad_norm": 5.008421421051025, "learning_rate": 2.428869024777668e-06, "loss": 0.7375, "step": 2425 }, { "epoch": 2.341698841698842, "grad_norm": 5.249884605407715, "learning_rate": 2.422063743817832e-06, "loss": 0.7812, "step": 2426 }, { "epoch": 2.3426640926640925, "grad_norm": 4.918494701385498, "learning_rate": 2.4152666959725513e-06, "loss": 0.7217, "step": 2427 }, { "epoch": 2.3436293436293436, "grad_norm": 4.714098930358887, "learning_rate": 2.408477888626527e-06, "loss": 0.7027, "step": 2428 }, { "epoch": 2.3445945945945947, "grad_norm": 5.372392177581787, "learning_rate": 2.4016973291555235e-06, "loss": 0.7622, "step": 2429 }, { "epoch": 2.3455598455598454, "grad_norm": 4.705463409423828, "learning_rate": 2.3949250249263333e-06, "loss": 0.6922, "step": 2430 }, { "epoch": 2.3465250965250966, "grad_norm": 4.9439520835876465, "learning_rate": 2.3881609832967934e-06, "loss": 0.7497, "step": 2431 }, { "epoch": 2.3474903474903477, "grad_norm": 4.718496799468994, "learning_rate": 2.3814052116157495e-06, "loss": 0.7282, "step": 2432 }, { "epoch": 2.3484555984555984, "grad_norm": 4.703372955322266, "learning_rate": 2.374657717223068e-06, "loss": 0.7032, "step": 2433 }, { "epoch": 2.3494208494208495, "grad_norm": 4.857315540313721, "learning_rate": 2.3679185074496247e-06, "loss": 0.7654, "step": 2434 }, { "epoch": 2.3503861003861006, "grad_norm": 5.146064758300781, "learning_rate": 2.361187589617292e-06, "loss": 0.7491, "step": 2435 }, { "epoch": 2.3513513513513513, "grad_norm": 4.706638813018799, "learning_rate": 2.354464971038933e-06, "loss": 0.7369, "step": 2436 }, { "epoch": 2.3523166023166024, "grad_norm": 4.841254234313965, "learning_rate": 2.347750659018397e-06, "loss": 0.7545, "step": 2437 }, { "epoch": 2.353281853281853, "grad_norm": 4.904670715332031, "learning_rate": 2.3410446608505044e-06, "loss": 0.7085, "step": 2438 }, { "epoch": 2.3542471042471043, "grad_norm": 4.515750885009766, "learning_rate": 2.3343469838210444e-06, "loss": 0.6409, "step": 2439 }, { "epoch": 2.3552123552123554, "grad_norm": 4.757826805114746, "learning_rate": 2.3276576352067725e-06, "loss": 0.7561, "step": 2440 }, { "epoch": 2.356177606177606, "grad_norm": 4.912268161773682, "learning_rate": 2.3209766222753814e-06, "loss": 0.7219, "step": 2441 }, { "epoch": 2.357142857142857, "grad_norm": 4.534762382507324, "learning_rate": 2.3143039522855204e-06, "loss": 0.7202, "step": 2442 }, { "epoch": 2.358108108108108, "grad_norm": 4.946676254272461, "learning_rate": 2.3076396324867643e-06, "loss": 0.8139, "step": 2443 }, { "epoch": 2.359073359073359, "grad_norm": 4.679795742034912, "learning_rate": 2.300983670119632e-06, "loss": 0.6618, "step": 2444 }, { "epoch": 2.36003861003861, "grad_norm": 4.652212619781494, "learning_rate": 2.2943360724155406e-06, "loss": 0.7275, "step": 2445 }, { "epoch": 2.361003861003861, "grad_norm": 4.834445953369141, "learning_rate": 2.2876968465968354e-06, "loss": 0.7166, "step": 2446 }, { "epoch": 2.361969111969112, "grad_norm": 4.824753761291504, "learning_rate": 2.281065999876756e-06, "loss": 0.7317, "step": 2447 }, { "epoch": 2.362934362934363, "grad_norm": 5.067233085632324, "learning_rate": 2.27444353945945e-06, "loss": 0.7301, "step": 2448 }, { "epoch": 2.363899613899614, "grad_norm": 5.02011775970459, "learning_rate": 2.2678294725399465e-06, "loss": 0.7139, "step": 2449 }, { "epoch": 2.364864864864865, "grad_norm": 4.794853687286377, "learning_rate": 2.2612238063041515e-06, "loss": 0.7047, "step": 2450 }, { "epoch": 2.365830115830116, "grad_norm": 5.3170599937438965, "learning_rate": 2.2546265479288464e-06, "loss": 0.7272, "step": 2451 }, { "epoch": 2.3667953667953667, "grad_norm": 4.975372791290283, "learning_rate": 2.2480377045816858e-06, "loss": 0.7372, "step": 2452 }, { "epoch": 2.367760617760618, "grad_norm": 5.004469394683838, "learning_rate": 2.241457283421178e-06, "loss": 0.7281, "step": 2453 }, { "epoch": 2.3687258687258685, "grad_norm": 5.1172380447387695, "learning_rate": 2.23488529159667e-06, "loss": 0.7877, "step": 2454 }, { "epoch": 2.3696911196911197, "grad_norm": 4.929131984710693, "learning_rate": 2.22832173624836e-06, "loss": 0.7335, "step": 2455 }, { "epoch": 2.370656370656371, "grad_norm": 4.840761184692383, "learning_rate": 2.2217666245072864e-06, "loss": 0.6732, "step": 2456 }, { "epoch": 2.3716216216216215, "grad_norm": 5.0260329246521, "learning_rate": 2.2152199634953043e-06, "loss": 0.7467, "step": 2457 }, { "epoch": 2.3725868725868726, "grad_norm": 4.6075439453125, "learning_rate": 2.208681760325092e-06, "loss": 0.6816, "step": 2458 }, { "epoch": 2.3735521235521233, "grad_norm": 5.479339599609375, "learning_rate": 2.2021520221001304e-06, "loss": 0.7389, "step": 2459 }, { "epoch": 2.3745173745173744, "grad_norm": 4.929660797119141, "learning_rate": 2.195630755914716e-06, "loss": 0.7207, "step": 2460 }, { "epoch": 2.3754826254826256, "grad_norm": 5.115887641906738, "learning_rate": 2.189117968853933e-06, "loss": 0.7422, "step": 2461 }, { "epoch": 2.3764478764478767, "grad_norm": 4.739163398742676, "learning_rate": 2.1826136679936605e-06, "loss": 0.7021, "step": 2462 }, { "epoch": 2.3774131274131274, "grad_norm": 4.605644226074219, "learning_rate": 2.1761178604005418e-06, "loss": 0.6962, "step": 2463 }, { "epoch": 2.3783783783783785, "grad_norm": 4.777822971343994, "learning_rate": 2.1696305531320115e-06, "loss": 0.6872, "step": 2464 }, { "epoch": 2.379343629343629, "grad_norm": 5.216665744781494, "learning_rate": 2.16315175323626e-06, "loss": 0.733, "step": 2465 }, { "epoch": 2.3803088803088803, "grad_norm": 4.957310676574707, "learning_rate": 2.156681467752234e-06, "loss": 0.7233, "step": 2466 }, { "epoch": 2.3812741312741315, "grad_norm": 4.596826553344727, "learning_rate": 2.1502197037096317e-06, "loss": 0.7304, "step": 2467 }, { "epoch": 2.382239382239382, "grad_norm": 4.776923656463623, "learning_rate": 2.1437664681288926e-06, "loss": 0.7066, "step": 2468 }, { "epoch": 2.3832046332046333, "grad_norm": 4.835448741912842, "learning_rate": 2.1373217680211924e-06, "loss": 0.7082, "step": 2469 }, { "epoch": 2.384169884169884, "grad_norm": 4.912600040435791, "learning_rate": 2.130885610388428e-06, "loss": 0.7089, "step": 2470 }, { "epoch": 2.385135135135135, "grad_norm": 5.509889125823975, "learning_rate": 2.12445800222322e-06, "loss": 0.7146, "step": 2471 }, { "epoch": 2.386100386100386, "grad_norm": 4.585874080657959, "learning_rate": 2.1180389505089007e-06, "loss": 0.7554, "step": 2472 }, { "epoch": 2.387065637065637, "grad_norm": 4.63887357711792, "learning_rate": 2.1116284622195018e-06, "loss": 0.7053, "step": 2473 }, { "epoch": 2.388030888030888, "grad_norm": 5.154671669006348, "learning_rate": 2.105226544319756e-06, "loss": 0.7142, "step": 2474 }, { "epoch": 2.388996138996139, "grad_norm": 4.6797566413879395, "learning_rate": 2.0988332037650816e-06, "loss": 0.685, "step": 2475 }, { "epoch": 2.38996138996139, "grad_norm": 4.851925373077393, "learning_rate": 2.0924484475015783e-06, "loss": 0.7357, "step": 2476 }, { "epoch": 2.390926640926641, "grad_norm": 5.000600337982178, "learning_rate": 2.0860722824660217e-06, "loss": 0.7511, "step": 2477 }, { "epoch": 2.391891891891892, "grad_norm": 4.8918352127075195, "learning_rate": 2.0797047155858495e-06, "loss": 0.7263, "step": 2478 }, { "epoch": 2.392857142857143, "grad_norm": 4.8164591789245605, "learning_rate": 2.0733457537791614e-06, "loss": 0.7303, "step": 2479 }, { "epoch": 2.393822393822394, "grad_norm": 5.150565147399902, "learning_rate": 2.066995403954706e-06, "loss": 0.7473, "step": 2480 }, { "epoch": 2.3947876447876446, "grad_norm": 4.9704742431640625, "learning_rate": 2.0606536730118767e-06, "loss": 0.7989, "step": 2481 }, { "epoch": 2.3957528957528957, "grad_norm": 4.806553363800049, "learning_rate": 2.0543205678407e-06, "loss": 0.7698, "step": 2482 }, { "epoch": 2.396718146718147, "grad_norm": 4.680107593536377, "learning_rate": 2.047996095321835e-06, "loss": 0.7222, "step": 2483 }, { "epoch": 2.3976833976833976, "grad_norm": 4.774284362792969, "learning_rate": 2.0416802623265596e-06, "loss": 0.7484, "step": 2484 }, { "epoch": 2.3986486486486487, "grad_norm": 5.096017360687256, "learning_rate": 2.0353730757167634e-06, "loss": 0.7417, "step": 2485 }, { "epoch": 2.3996138996138994, "grad_norm": 4.331223011016846, "learning_rate": 2.0290745423449465e-06, "loss": 0.6919, "step": 2486 }, { "epoch": 2.4005791505791505, "grad_norm": 4.920547008514404, "learning_rate": 2.0227846690542043e-06, "loss": 0.7301, "step": 2487 }, { "epoch": 2.4015444015444016, "grad_norm": 5.212554454803467, "learning_rate": 2.016503462678222e-06, "loss": 0.7452, "step": 2488 }, { "epoch": 2.4025096525096523, "grad_norm": 5.136991500854492, "learning_rate": 2.010230930041274e-06, "loss": 0.7219, "step": 2489 }, { "epoch": 2.4034749034749034, "grad_norm": 4.706001281738281, "learning_rate": 2.0039670779582066e-06, "loss": 0.6659, "step": 2490 }, { "epoch": 2.4044401544401546, "grad_norm": 5.0232133865356445, "learning_rate": 1.997711913234435e-06, "loss": 0.7514, "step": 2491 }, { "epoch": 2.4054054054054053, "grad_norm": 5.077817440032959, "learning_rate": 1.9914654426659374e-06, "loss": 0.7229, "step": 2492 }, { "epoch": 2.4063706563706564, "grad_norm": 5.236948490142822, "learning_rate": 1.985227673039245e-06, "loss": 0.7808, "step": 2493 }, { "epoch": 2.4073359073359075, "grad_norm": 4.87125825881958, "learning_rate": 1.978998611131443e-06, "loss": 0.7886, "step": 2494 }, { "epoch": 2.408301158301158, "grad_norm": 5.276411533355713, "learning_rate": 1.9727782637101435e-06, "loss": 0.7445, "step": 2495 }, { "epoch": 2.4092664092664093, "grad_norm": 5.497075080871582, "learning_rate": 1.966566637533497e-06, "loss": 0.7403, "step": 2496 }, { "epoch": 2.41023166023166, "grad_norm": 5.233363151550293, "learning_rate": 1.960363739350177e-06, "loss": 0.714, "step": 2497 }, { "epoch": 2.411196911196911, "grad_norm": 5.141330718994141, "learning_rate": 1.954169575899386e-06, "loss": 0.7169, "step": 2498 }, { "epoch": 2.4121621621621623, "grad_norm": 4.863707542419434, "learning_rate": 1.9479841539108156e-06, "loss": 0.7577, "step": 2499 }, { "epoch": 2.413127413127413, "grad_norm": 5.165623188018799, "learning_rate": 1.941807480104676e-06, "loss": 0.7102, "step": 2500 }, { "epoch": 2.414092664092664, "grad_norm": 4.734787464141846, "learning_rate": 1.935639561191663e-06, "loss": 0.7166, "step": 2501 }, { "epoch": 2.4150579150579152, "grad_norm": 5.163553714752197, "learning_rate": 1.9294804038729743e-06, "loss": 0.7574, "step": 2502 }, { "epoch": 2.416023166023166, "grad_norm": 5.120594024658203, "learning_rate": 1.9233300148402767e-06, "loss": 0.6955, "step": 2503 }, { "epoch": 2.416988416988417, "grad_norm": 5.021568298339844, "learning_rate": 1.9171884007757093e-06, "loss": 0.7281, "step": 2504 }, { "epoch": 2.417953667953668, "grad_norm": 4.861301422119141, "learning_rate": 1.911055568351883e-06, "loss": 0.7986, "step": 2505 }, { "epoch": 2.418918918918919, "grad_norm": 5.154460430145264, "learning_rate": 1.90493152423187e-06, "loss": 0.6855, "step": 2506 }, { "epoch": 2.41988416988417, "grad_norm": 4.927275657653809, "learning_rate": 1.898816275069193e-06, "loss": 0.7509, "step": 2507 }, { "epoch": 2.4208494208494207, "grad_norm": 5.266952991485596, "learning_rate": 1.8927098275078116e-06, "loss": 0.7192, "step": 2508 }, { "epoch": 2.421814671814672, "grad_norm": 4.80417537689209, "learning_rate": 1.886612188182132e-06, "loss": 0.7568, "step": 2509 }, { "epoch": 2.422779922779923, "grad_norm": 5.437583923339844, "learning_rate": 1.8805233637169829e-06, "loss": 0.6822, "step": 2510 }, { "epoch": 2.4237451737451736, "grad_norm": 4.913822650909424, "learning_rate": 1.8744433607276269e-06, "loss": 0.7758, "step": 2511 }, { "epoch": 2.4247104247104247, "grad_norm": 5.271089553833008, "learning_rate": 1.8683721858197369e-06, "loss": 0.7822, "step": 2512 }, { "epoch": 2.4256756756756754, "grad_norm": 4.6216139793396, "learning_rate": 1.8623098455893862e-06, "loss": 0.7506, "step": 2513 }, { "epoch": 2.4266409266409266, "grad_norm": 4.506705284118652, "learning_rate": 1.8562563466230577e-06, "loss": 0.7118, "step": 2514 }, { "epoch": 2.4276061776061777, "grad_norm": 4.784862041473389, "learning_rate": 1.8502116954976334e-06, "loss": 0.7221, "step": 2515 }, { "epoch": 2.4285714285714284, "grad_norm": 5.663732528686523, "learning_rate": 1.844175898780377e-06, "loss": 0.7711, "step": 2516 }, { "epoch": 2.4295366795366795, "grad_norm": 4.902246952056885, "learning_rate": 1.8381489630289273e-06, "loss": 0.73, "step": 2517 }, { "epoch": 2.4305019305019306, "grad_norm": 4.988832473754883, "learning_rate": 1.8321308947912997e-06, "loss": 0.6607, "step": 2518 }, { "epoch": 2.4314671814671813, "grad_norm": 4.855826377868652, "learning_rate": 1.8261217006058828e-06, "loss": 0.7027, "step": 2519 }, { "epoch": 2.4324324324324325, "grad_norm": 4.841279029846191, "learning_rate": 1.8201213870014168e-06, "loss": 0.7623, "step": 2520 }, { "epoch": 2.4333976833976836, "grad_norm": 5.101052761077881, "learning_rate": 1.8141299604969908e-06, "loss": 0.7258, "step": 2521 }, { "epoch": 2.4343629343629343, "grad_norm": 5.110799312591553, "learning_rate": 1.8081474276020406e-06, "loss": 0.6804, "step": 2522 }, { "epoch": 2.4353281853281854, "grad_norm": 5.3934125900268555, "learning_rate": 1.8021737948163465e-06, "loss": 0.7129, "step": 2523 }, { "epoch": 2.436293436293436, "grad_norm": 5.148820877075195, "learning_rate": 1.7962090686300116e-06, "loss": 0.7749, "step": 2524 }, { "epoch": 2.437258687258687, "grad_norm": 4.828898906707764, "learning_rate": 1.7902532555234653e-06, "loss": 0.7241, "step": 2525 }, { "epoch": 2.4382239382239383, "grad_norm": 4.930122375488281, "learning_rate": 1.7843063619674473e-06, "loss": 0.7818, "step": 2526 }, { "epoch": 2.439189189189189, "grad_norm": 4.935842514038086, "learning_rate": 1.778368394423018e-06, "loss": 0.7692, "step": 2527 }, { "epoch": 2.44015444015444, "grad_norm": 5.2509074211120605, "learning_rate": 1.7724393593415302e-06, "loss": 0.7382, "step": 2528 }, { "epoch": 2.4411196911196913, "grad_norm": 4.885476112365723, "learning_rate": 1.76651926316464e-06, "loss": 0.7435, "step": 2529 }, { "epoch": 2.442084942084942, "grad_norm": 5.482947826385498, "learning_rate": 1.760608112324278e-06, "loss": 0.7631, "step": 2530 }, { "epoch": 2.443050193050193, "grad_norm": 5.086392879486084, "learning_rate": 1.7547059132426735e-06, "loss": 0.7685, "step": 2531 }, { "epoch": 2.4440154440154442, "grad_norm": 5.346570014953613, "learning_rate": 1.7488126723323185e-06, "loss": 0.7837, "step": 2532 }, { "epoch": 2.444980694980695, "grad_norm": 6.160499095916748, "learning_rate": 1.7429283959959776e-06, "loss": 0.7581, "step": 2533 }, { "epoch": 2.445945945945946, "grad_norm": 5.070865154266357, "learning_rate": 1.7370530906266724e-06, "loss": 0.7536, "step": 2534 }, { "epoch": 2.4469111969111967, "grad_norm": 5.467529296875, "learning_rate": 1.7311867626076795e-06, "loss": 0.6851, "step": 2535 }, { "epoch": 2.447876447876448, "grad_norm": 5.569533824920654, "learning_rate": 1.7253294183125223e-06, "loss": 0.7724, "step": 2536 }, { "epoch": 2.448841698841699, "grad_norm": 4.970886707305908, "learning_rate": 1.719481064104962e-06, "loss": 0.7247, "step": 2537 }, { "epoch": 2.4498069498069497, "grad_norm": 5.251448631286621, "learning_rate": 1.7136417063389964e-06, "loss": 0.7191, "step": 2538 }, { "epoch": 2.450772200772201, "grad_norm": 5.369867324829102, "learning_rate": 1.7078113513588434e-06, "loss": 0.7434, "step": 2539 }, { "epoch": 2.4517374517374515, "grad_norm": 4.895583152770996, "learning_rate": 1.7019900054989446e-06, "loss": 0.7787, "step": 2540 }, { "epoch": 2.4527027027027026, "grad_norm": 5.491269588470459, "learning_rate": 1.6961776750839509e-06, "loss": 0.7176, "step": 2541 }, { "epoch": 2.4536679536679538, "grad_norm": 4.857673645019531, "learning_rate": 1.6903743664287197e-06, "loss": 0.7052, "step": 2542 }, { "epoch": 2.4546332046332044, "grad_norm": 5.4430623054504395, "learning_rate": 1.6845800858383055e-06, "loss": 0.7157, "step": 2543 }, { "epoch": 2.4555984555984556, "grad_norm": 5.276530742645264, "learning_rate": 1.6787948396079544e-06, "loss": 0.7353, "step": 2544 }, { "epoch": 2.4565637065637067, "grad_norm": 5.065103054046631, "learning_rate": 1.6730186340230981e-06, "loss": 0.6974, "step": 2545 }, { "epoch": 2.4575289575289574, "grad_norm": 5.144730567932129, "learning_rate": 1.667251475359344e-06, "loss": 0.7688, "step": 2546 }, { "epoch": 2.4584942084942085, "grad_norm": 5.180382251739502, "learning_rate": 1.661493369882473e-06, "loss": 0.7028, "step": 2547 }, { "epoch": 2.4594594594594597, "grad_norm": 5.192044734954834, "learning_rate": 1.6557443238484283e-06, "loss": 0.7423, "step": 2548 }, { "epoch": 2.4604247104247103, "grad_norm": 5.120516300201416, "learning_rate": 1.6500043435033109e-06, "loss": 0.6733, "step": 2549 }, { "epoch": 2.4613899613899615, "grad_norm": 4.712008953094482, "learning_rate": 1.6442734350833721e-06, "loss": 0.7562, "step": 2550 }, { "epoch": 2.462355212355212, "grad_norm": 5.138809680938721, "learning_rate": 1.6385516048150063e-06, "loss": 0.7383, "step": 2551 }, { "epoch": 2.4633204633204633, "grad_norm": 5.04536247253418, "learning_rate": 1.632838858914747e-06, "loss": 0.7934, "step": 2552 }, { "epoch": 2.4642857142857144, "grad_norm": 5.438734531402588, "learning_rate": 1.6271352035892552e-06, "loss": 0.7686, "step": 2553 }, { "epoch": 2.465250965250965, "grad_norm": 4.890718460083008, "learning_rate": 1.621440645035318e-06, "loss": 0.7321, "step": 2554 }, { "epoch": 2.4662162162162162, "grad_norm": 5.157116413116455, "learning_rate": 1.6157551894398371e-06, "loss": 0.7909, "step": 2555 }, { "epoch": 2.4671814671814674, "grad_norm": 4.768998146057129, "learning_rate": 1.6100788429798242e-06, "loss": 0.7467, "step": 2556 }, { "epoch": 2.468146718146718, "grad_norm": 5.029572486877441, "learning_rate": 1.6044116118223973e-06, "loss": 0.7762, "step": 2557 }, { "epoch": 2.469111969111969, "grad_norm": 5.010144233703613, "learning_rate": 1.5987535021247668e-06, "loss": 0.77, "step": 2558 }, { "epoch": 2.4700772200772203, "grad_norm": 4.539247989654541, "learning_rate": 1.5931045200342355e-06, "loss": 0.7277, "step": 2559 }, { "epoch": 2.471042471042471, "grad_norm": 4.728090286254883, "learning_rate": 1.587464671688187e-06, "loss": 0.7132, "step": 2560 }, { "epoch": 2.472007722007722, "grad_norm": 4.758804798126221, "learning_rate": 1.581833963214091e-06, "loss": 0.7277, "step": 2561 }, { "epoch": 2.472972972972973, "grad_norm": 5.113906383514404, "learning_rate": 1.5762124007294721e-06, "loss": 0.7661, "step": 2562 }, { "epoch": 2.473938223938224, "grad_norm": 5.134305953979492, "learning_rate": 1.5705999903419267e-06, "loss": 0.7827, "step": 2563 }, { "epoch": 2.474903474903475, "grad_norm": 5.465189456939697, "learning_rate": 1.564996738149106e-06, "loss": 0.7497, "step": 2564 }, { "epoch": 2.4758687258687258, "grad_norm": 5.321998119354248, "learning_rate": 1.5594026502387184e-06, "loss": 0.759, "step": 2565 }, { "epoch": 2.476833976833977, "grad_norm": 5.111255168914795, "learning_rate": 1.553817732688503e-06, "loss": 0.7343, "step": 2566 }, { "epoch": 2.4777992277992276, "grad_norm": 5.24220085144043, "learning_rate": 1.5482419915662418e-06, "loss": 0.7317, "step": 2567 }, { "epoch": 2.4787644787644787, "grad_norm": 4.93293571472168, "learning_rate": 1.5426754329297477e-06, "loss": 0.7427, "step": 2568 }, { "epoch": 2.47972972972973, "grad_norm": 4.531349182128906, "learning_rate": 1.537118062826859e-06, "loss": 0.7641, "step": 2569 }, { "epoch": 2.4806949806949805, "grad_norm": 4.671472549438477, "learning_rate": 1.5315698872954299e-06, "loss": 0.7524, "step": 2570 }, { "epoch": 2.4816602316602316, "grad_norm": 5.036346435546875, "learning_rate": 1.5260309123633189e-06, "loss": 0.7977, "step": 2571 }, { "epoch": 2.4826254826254828, "grad_norm": 5.153173446655273, "learning_rate": 1.520501144048393e-06, "loss": 0.7819, "step": 2572 }, { "epoch": 2.4835907335907335, "grad_norm": 4.5762858390808105, "learning_rate": 1.5149805883585221e-06, "loss": 0.6937, "step": 2573 }, { "epoch": 2.4845559845559846, "grad_norm": 4.898032188415527, "learning_rate": 1.5094692512915621e-06, "loss": 0.7118, "step": 2574 }, { "epoch": 2.4855212355212357, "grad_norm": 4.975152969360352, "learning_rate": 1.5039671388353472e-06, "loss": 0.7288, "step": 2575 }, { "epoch": 2.4864864864864864, "grad_norm": 4.596181869506836, "learning_rate": 1.4984742569676957e-06, "loss": 0.7118, "step": 2576 }, { "epoch": 2.4874517374517375, "grad_norm": 5.141880512237549, "learning_rate": 1.4929906116564008e-06, "loss": 0.793, "step": 2577 }, { "epoch": 2.488416988416988, "grad_norm": 4.9118194580078125, "learning_rate": 1.4875162088592155e-06, "loss": 0.7374, "step": 2578 }, { "epoch": 2.4893822393822393, "grad_norm": 5.6777849197387695, "learning_rate": 1.4820510545238542e-06, "loss": 0.7466, "step": 2579 }, { "epoch": 2.4903474903474905, "grad_norm": 5.995486259460449, "learning_rate": 1.4765951545879732e-06, "loss": 0.7335, "step": 2580 }, { "epoch": 2.491312741312741, "grad_norm": 4.940185546875, "learning_rate": 1.4711485149791894e-06, "loss": 0.7946, "step": 2581 }, { "epoch": 2.4922779922779923, "grad_norm": 4.679786682128906, "learning_rate": 1.4657111416150504e-06, "loss": 0.7744, "step": 2582 }, { "epoch": 2.4932432432432434, "grad_norm": 5.902797222137451, "learning_rate": 1.4602830404030387e-06, "loss": 0.7534, "step": 2583 }, { "epoch": 2.494208494208494, "grad_norm": 5.293987274169922, "learning_rate": 1.4548642172405569e-06, "loss": 0.7632, "step": 2584 }, { "epoch": 2.4951737451737452, "grad_norm": 5.381636142730713, "learning_rate": 1.4494546780149366e-06, "loss": 0.7585, "step": 2585 }, { "epoch": 2.4961389961389964, "grad_norm": 5.651233673095703, "learning_rate": 1.4440544286034186e-06, "loss": 0.7293, "step": 2586 }, { "epoch": 2.497104247104247, "grad_norm": 5.446207046508789, "learning_rate": 1.4386634748731499e-06, "loss": 0.7867, "step": 2587 }, { "epoch": 2.498069498069498, "grad_norm": 5.752668857574463, "learning_rate": 1.4332818226811807e-06, "loss": 0.692, "step": 2588 }, { "epoch": 2.499034749034749, "grad_norm": 5.912139892578125, "learning_rate": 1.4279094778744473e-06, "loss": 0.8358, "step": 2589 }, { "epoch": 2.5, "grad_norm": 5.182679176330566, "learning_rate": 1.4225464462897876e-06, "loss": 0.7419, "step": 2590 }, { "epoch": 2.500965250965251, "grad_norm": 5.262983798980713, "learning_rate": 1.4171927337539103e-06, "loss": 0.76, "step": 2591 }, { "epoch": 2.501930501930502, "grad_norm": 4.978514671325684, "learning_rate": 1.4118483460834065e-06, "loss": 0.7344, "step": 2592 }, { "epoch": 2.502895752895753, "grad_norm": 5.735084533691406, "learning_rate": 1.4065132890847245e-06, "loss": 0.7396, "step": 2593 }, { "epoch": 2.5038610038610036, "grad_norm": 5.626955986022949, "learning_rate": 1.4011875685541898e-06, "loss": 0.7295, "step": 2594 }, { "epoch": 2.5048262548262548, "grad_norm": 4.752853870391846, "learning_rate": 1.3958711902779775e-06, "loss": 0.6756, "step": 2595 }, { "epoch": 2.505791505791506, "grad_norm": 5.7477569580078125, "learning_rate": 1.390564160032113e-06, "loss": 0.7603, "step": 2596 }, { "epoch": 2.506756756756757, "grad_norm": 4.698903560638428, "learning_rate": 1.3852664835824592e-06, "loss": 0.7626, "step": 2597 }, { "epoch": 2.5077220077220077, "grad_norm": 5.006353378295898, "learning_rate": 1.3799781666847301e-06, "loss": 0.7236, "step": 2598 }, { "epoch": 2.508687258687259, "grad_norm": 6.56751823425293, "learning_rate": 1.3746992150844595e-06, "loss": 0.7565, "step": 2599 }, { "epoch": 2.5096525096525095, "grad_norm": 4.768529415130615, "learning_rate": 1.369429634517012e-06, "loss": 0.7437, "step": 2600 }, { "epoch": 2.5106177606177607, "grad_norm": 5.13110876083374, "learning_rate": 1.3641694307075682e-06, "loss": 0.7305, "step": 2601 }, { "epoch": 2.511583011583012, "grad_norm": 4.623649597167969, "learning_rate": 1.3589186093711227e-06, "loss": 0.7282, "step": 2602 }, { "epoch": 2.5125482625482625, "grad_norm": 4.7761759757995605, "learning_rate": 1.3536771762124745e-06, "loss": 0.7194, "step": 2603 }, { "epoch": 2.5135135135135136, "grad_norm": 5.18038272857666, "learning_rate": 1.3484451369262251e-06, "loss": 0.7949, "step": 2604 }, { "epoch": 2.5144787644787643, "grad_norm": 4.617712020874023, "learning_rate": 1.34322249719677e-06, "loss": 0.7398, "step": 2605 }, { "epoch": 2.5154440154440154, "grad_norm": 5.1880784034729, "learning_rate": 1.3380092626982887e-06, "loss": 0.6584, "step": 2606 }, { "epoch": 2.5164092664092665, "grad_norm": 5.24415922164917, "learning_rate": 1.3328054390947475e-06, "loss": 0.735, "step": 2607 }, { "epoch": 2.5173745173745172, "grad_norm": 4.836585521697998, "learning_rate": 1.3276110320398849e-06, "loss": 0.7223, "step": 2608 }, { "epoch": 2.5183397683397684, "grad_norm": 5.203280925750732, "learning_rate": 1.3224260471772088e-06, "loss": 0.7055, "step": 2609 }, { "epoch": 2.519305019305019, "grad_norm": 5.076012134552002, "learning_rate": 1.317250490139992e-06, "loss": 0.745, "step": 2610 }, { "epoch": 2.52027027027027, "grad_norm": 5.236003398895264, "learning_rate": 1.3120843665512617e-06, "loss": 0.7623, "step": 2611 }, { "epoch": 2.5212355212355213, "grad_norm": 5.240046977996826, "learning_rate": 1.3069276820237997e-06, "loss": 0.7261, "step": 2612 }, { "epoch": 2.5222007722007724, "grad_norm": 5.299560546875, "learning_rate": 1.3017804421601298e-06, "loss": 0.8015, "step": 2613 }, { "epoch": 2.523166023166023, "grad_norm": 5.6360015869140625, "learning_rate": 1.2966426525525144e-06, "loss": 0.7365, "step": 2614 }, { "epoch": 2.5241312741312742, "grad_norm": 5.0282158851623535, "learning_rate": 1.2915143187829516e-06, "loss": 0.728, "step": 2615 }, { "epoch": 2.525096525096525, "grad_norm": 4.797794342041016, "learning_rate": 1.286395446423162e-06, "loss": 0.7516, "step": 2616 }, { "epoch": 2.526061776061776, "grad_norm": 4.75595235824585, "learning_rate": 1.2812860410345894e-06, "loss": 0.7432, "step": 2617 }, { "epoch": 2.527027027027027, "grad_norm": 4.534605503082275, "learning_rate": 1.2761861081683914e-06, "loss": 0.8098, "step": 2618 }, { "epoch": 2.527992277992278, "grad_norm": 4.790217399597168, "learning_rate": 1.2710956533654384e-06, "loss": 0.7833, "step": 2619 }, { "epoch": 2.528957528957529, "grad_norm": 4.9109721183776855, "learning_rate": 1.2660146821562936e-06, "loss": 0.7315, "step": 2620 }, { "epoch": 2.5299227799227797, "grad_norm": 4.9708781242370605, "learning_rate": 1.2609432000612232e-06, "loss": 0.7869, "step": 2621 }, { "epoch": 2.530888030888031, "grad_norm": 5.2403082847595215, "learning_rate": 1.255881212590183e-06, "loss": 0.742, "step": 2622 }, { "epoch": 2.531853281853282, "grad_norm": 5.221992015838623, "learning_rate": 1.2508287252428175e-06, "loss": 0.7807, "step": 2623 }, { "epoch": 2.532818532818533, "grad_norm": 4.320801734924316, "learning_rate": 1.245785743508441e-06, "loss": 0.6937, "step": 2624 }, { "epoch": 2.5337837837837838, "grad_norm": 5.410978317260742, "learning_rate": 1.2407522728660447e-06, "loss": 0.7456, "step": 2625 }, { "epoch": 2.534749034749035, "grad_norm": 5.150344371795654, "learning_rate": 1.2357283187842872e-06, "loss": 0.7478, "step": 2626 }, { "epoch": 2.5357142857142856, "grad_norm": 5.0697174072265625, "learning_rate": 1.2307138867214863e-06, "loss": 0.7477, "step": 2627 }, { "epoch": 2.5366795366795367, "grad_norm": 5.012232780456543, "learning_rate": 1.2257089821256196e-06, "loss": 0.764, "step": 2628 }, { "epoch": 2.537644787644788, "grad_norm": 4.838174819946289, "learning_rate": 1.220713610434303e-06, "loss": 0.7387, "step": 2629 }, { "epoch": 2.5386100386100385, "grad_norm": 5.2238311767578125, "learning_rate": 1.2157277770748044e-06, "loss": 0.7306, "step": 2630 }, { "epoch": 2.5395752895752897, "grad_norm": 4.850396633148193, "learning_rate": 1.2107514874640235e-06, "loss": 0.7291, "step": 2631 }, { "epoch": 2.5405405405405403, "grad_norm": 4.7762322425842285, "learning_rate": 1.2057847470084993e-06, "loss": 0.6686, "step": 2632 }, { "epoch": 2.5415057915057915, "grad_norm": 5.246326446533203, "learning_rate": 1.200827561104383e-06, "loss": 0.7399, "step": 2633 }, { "epoch": 2.5424710424710426, "grad_norm": 5.028595924377441, "learning_rate": 1.195879935137455e-06, "loss": 0.7027, "step": 2634 }, { "epoch": 2.5434362934362933, "grad_norm": 4.831367492675781, "learning_rate": 1.1909418744831048e-06, "loss": 0.7693, "step": 2635 }, { "epoch": 2.5444015444015444, "grad_norm": 5.539283275604248, "learning_rate": 1.186013384506335e-06, "loss": 0.7085, "step": 2636 }, { "epoch": 2.545366795366795, "grad_norm": 5.044233798980713, "learning_rate": 1.1810944705617466e-06, "loss": 0.7303, "step": 2637 }, { "epoch": 2.5463320463320462, "grad_norm": 4.916104316711426, "learning_rate": 1.1761851379935307e-06, "loss": 0.7268, "step": 2638 }, { "epoch": 2.5472972972972974, "grad_norm": 5.304819583892822, "learning_rate": 1.171285392135476e-06, "loss": 0.8003, "step": 2639 }, { "epoch": 2.5482625482625485, "grad_norm": 4.751123428344727, "learning_rate": 1.1663952383109566e-06, "loss": 0.7035, "step": 2640 }, { "epoch": 2.549227799227799, "grad_norm": 4.719592094421387, "learning_rate": 1.1615146818329237e-06, "loss": 0.738, "step": 2641 }, { "epoch": 2.5501930501930503, "grad_norm": 4.5217180252075195, "learning_rate": 1.1566437280038956e-06, "loss": 0.7048, "step": 2642 }, { "epoch": 2.551158301158301, "grad_norm": 4.832311630249023, "learning_rate": 1.1517823821159624e-06, "loss": 0.7116, "step": 2643 }, { "epoch": 2.552123552123552, "grad_norm": 5.188864231109619, "learning_rate": 1.1469306494507792e-06, "loss": 0.6957, "step": 2644 }, { "epoch": 2.5530888030888033, "grad_norm": 4.863763809204102, "learning_rate": 1.1420885352795519e-06, "loss": 0.7794, "step": 2645 }, { "epoch": 2.554054054054054, "grad_norm": 4.8022003173828125, "learning_rate": 1.1372560448630377e-06, "loss": 0.7648, "step": 2646 }, { "epoch": 2.555019305019305, "grad_norm": 4.957478046417236, "learning_rate": 1.1324331834515335e-06, "loss": 0.7656, "step": 2647 }, { "epoch": 2.5559845559845558, "grad_norm": 4.873451232910156, "learning_rate": 1.1276199562848845e-06, "loss": 0.752, "step": 2648 }, { "epoch": 2.556949806949807, "grad_norm": 5.428621768951416, "learning_rate": 1.1228163685924599e-06, "loss": 0.7545, "step": 2649 }, { "epoch": 2.557915057915058, "grad_norm": 4.873611927032471, "learning_rate": 1.1180224255931626e-06, "loss": 0.7147, "step": 2650 }, { "epoch": 2.558880308880309, "grad_norm": 4.914524555206299, "learning_rate": 1.1132381324954067e-06, "loss": 0.7488, "step": 2651 }, { "epoch": 2.55984555984556, "grad_norm": 4.954052448272705, "learning_rate": 1.108463494497135e-06, "loss": 0.79, "step": 2652 }, { "epoch": 2.560810810810811, "grad_norm": 4.874828815460205, "learning_rate": 1.1036985167857938e-06, "loss": 0.7259, "step": 2653 }, { "epoch": 2.5617760617760617, "grad_norm": 5.06655740737915, "learning_rate": 1.098943204538332e-06, "loss": 0.7827, "step": 2654 }, { "epoch": 2.562741312741313, "grad_norm": 4.803916931152344, "learning_rate": 1.094197562921202e-06, "loss": 0.7496, "step": 2655 }, { "epoch": 2.563706563706564, "grad_norm": 5.09578800201416, "learning_rate": 1.0894615970903454e-06, "loss": 0.7658, "step": 2656 }, { "epoch": 2.5646718146718146, "grad_norm": 4.7430219650268555, "learning_rate": 1.0847353121911952e-06, "loss": 0.7573, "step": 2657 }, { "epoch": 2.5656370656370657, "grad_norm": 5.272439479827881, "learning_rate": 1.0800187133586626e-06, "loss": 0.7421, "step": 2658 }, { "epoch": 2.5666023166023164, "grad_norm": 4.683495998382568, "learning_rate": 1.0753118057171396e-06, "loss": 0.7881, "step": 2659 }, { "epoch": 2.5675675675675675, "grad_norm": 5.072673797607422, "learning_rate": 1.0706145943804846e-06, "loss": 0.7413, "step": 2660 }, { "epoch": 2.5685328185328187, "grad_norm": 4.405806541442871, "learning_rate": 1.0659270844520253e-06, "loss": 0.7472, "step": 2661 }, { "epoch": 2.5694980694980694, "grad_norm": 4.982200622558594, "learning_rate": 1.0612492810245466e-06, "loss": 0.7096, "step": 2662 }, { "epoch": 2.5704633204633205, "grad_norm": 4.694039821624756, "learning_rate": 1.056581189180288e-06, "loss": 0.7413, "step": 2663 }, { "epoch": 2.571428571428571, "grad_norm": 4.818814754486084, "learning_rate": 1.05192281399094e-06, "loss": 0.7524, "step": 2664 }, { "epoch": 2.5723938223938223, "grad_norm": 4.986637592315674, "learning_rate": 1.0472741605176339e-06, "loss": 0.788, "step": 2665 }, { "epoch": 2.5733590733590734, "grad_norm": 4.786144733428955, "learning_rate": 1.0426352338109392e-06, "loss": 0.7132, "step": 2666 }, { "epoch": 2.5743243243243246, "grad_norm": 5.189314365386963, "learning_rate": 1.038006038910858e-06, "loss": 0.7619, "step": 2667 }, { "epoch": 2.5752895752895753, "grad_norm": 4.994247913360596, "learning_rate": 1.0333865808468203e-06, "loss": 0.7311, "step": 2668 }, { "epoch": 2.5762548262548264, "grad_norm": 5.909048557281494, "learning_rate": 1.0287768646376762e-06, "loss": 0.7497, "step": 2669 }, { "epoch": 2.577220077220077, "grad_norm": 4.981646537780762, "learning_rate": 1.0241768952916898e-06, "loss": 0.778, "step": 2670 }, { "epoch": 2.578185328185328, "grad_norm": 5.011918067932129, "learning_rate": 1.01958667780654e-06, "loss": 0.7496, "step": 2671 }, { "epoch": 2.5791505791505793, "grad_norm": 4.762938976287842, "learning_rate": 1.0150062171693075e-06, "loss": 0.7419, "step": 2672 }, { "epoch": 2.58011583011583, "grad_norm": 5.066151142120361, "learning_rate": 1.010435518356474e-06, "loss": 0.736, "step": 2673 }, { "epoch": 2.581081081081081, "grad_norm": 5.22730827331543, "learning_rate": 1.0058745863339136e-06, "loss": 0.7731, "step": 2674 }, { "epoch": 2.582046332046332, "grad_norm": 5.501034736633301, "learning_rate": 1.0013234260568915e-06, "loss": 0.7585, "step": 2675 }, { "epoch": 2.583011583011583, "grad_norm": 5.197367191314697, "learning_rate": 9.967820424700548e-07, "loss": 0.7524, "step": 2676 }, { "epoch": 2.583976833976834, "grad_norm": 5.409716606140137, "learning_rate": 9.922504405074295e-07, "loss": 0.7613, "step": 2677 }, { "epoch": 2.584942084942085, "grad_norm": 5.282143592834473, "learning_rate": 9.877286250924145e-07, "loss": 0.7369, "step": 2678 }, { "epoch": 2.585907335907336, "grad_norm": 5.7618842124938965, "learning_rate": 9.83216601137773e-07, "loss": 0.7047, "step": 2679 }, { "epoch": 2.586872586872587, "grad_norm": 4.768263339996338, "learning_rate": 9.787143735456351e-07, "loss": 0.8062, "step": 2680 }, { "epoch": 2.5878378378378377, "grad_norm": 4.665050029754639, "learning_rate": 9.742219472074833e-07, "loss": 0.7731, "step": 2681 }, { "epoch": 2.588803088803089, "grad_norm": 4.817323207855225, "learning_rate": 9.697393270041533e-07, "loss": 0.7083, "step": 2682 }, { "epoch": 2.58976833976834, "grad_norm": 4.751098155975342, "learning_rate": 9.652665178058285e-07, "loss": 0.8147, "step": 2683 }, { "epoch": 2.5907335907335907, "grad_norm": 4.651120662689209, "learning_rate": 9.608035244720293e-07, "loss": 0.7698, "step": 2684 }, { "epoch": 2.591698841698842, "grad_norm": 5.18508768081665, "learning_rate": 9.563503518516114e-07, "loss": 0.7625, "step": 2685 }, { "epoch": 2.5926640926640925, "grad_norm": 4.685789108276367, "learning_rate": 9.519070047827705e-07, "loss": 0.7353, "step": 2686 }, { "epoch": 2.5936293436293436, "grad_norm": 4.945529937744141, "learning_rate": 9.474734880930114e-07, "loss": 0.7487, "step": 2687 }, { "epoch": 2.5945945945945947, "grad_norm": 4.744202136993408, "learning_rate": 9.430498065991712e-07, "loss": 0.7357, "step": 2688 }, { "epoch": 2.5955598455598454, "grad_norm": 5.035425186157227, "learning_rate": 9.386359651073929e-07, "loss": 0.7739, "step": 2689 }, { "epoch": 2.5965250965250966, "grad_norm": 4.681142330169678, "learning_rate": 9.342319684131396e-07, "loss": 0.7957, "step": 2690 }, { "epoch": 2.5974903474903472, "grad_norm": 5.089695930480957, "learning_rate": 9.298378213011661e-07, "loss": 0.717, "step": 2691 }, { "epoch": 2.5984555984555984, "grad_norm": 4.988779544830322, "learning_rate": 9.254535285455335e-07, "loss": 0.6986, "step": 2692 }, { "epoch": 2.5994208494208495, "grad_norm": 4.7568254470825195, "learning_rate": 9.210790949095916e-07, "loss": 0.7577, "step": 2693 }, { "epoch": 2.6003861003861006, "grad_norm": 4.567595481872559, "learning_rate": 9.167145251459875e-07, "loss": 0.7213, "step": 2694 }, { "epoch": 2.6013513513513513, "grad_norm": 4.910531044006348, "learning_rate": 9.123598239966447e-07, "loss": 0.7957, "step": 2695 }, { "epoch": 2.6023166023166024, "grad_norm": 5.033100605010986, "learning_rate": 9.08014996192762e-07, "loss": 0.7237, "step": 2696 }, { "epoch": 2.603281853281853, "grad_norm": 4.96120023727417, "learning_rate": 9.036800464548157e-07, "loss": 0.6612, "step": 2697 }, { "epoch": 2.6042471042471043, "grad_norm": 4.651401519775391, "learning_rate": 8.993549794925538e-07, "loss": 0.7619, "step": 2698 }, { "epoch": 2.6052123552123554, "grad_norm": 5.121455192565918, "learning_rate": 8.95039800004982e-07, "loss": 0.7651, "step": 2699 }, { "epoch": 2.606177606177606, "grad_norm": 4.698480606079102, "learning_rate": 8.907345126803602e-07, "loss": 0.7207, "step": 2700 }, { "epoch": 2.607142857142857, "grad_norm": 4.910830020904541, "learning_rate": 8.864391221962065e-07, "loss": 0.7565, "step": 2701 }, { "epoch": 2.608108108108108, "grad_norm": 4.684881687164307, "learning_rate": 8.821536332192882e-07, "loss": 0.7329, "step": 2702 }, { "epoch": 2.609073359073359, "grad_norm": 4.938525676727295, "learning_rate": 8.778780504056106e-07, "loss": 0.7575, "step": 2703 }, { "epoch": 2.61003861003861, "grad_norm": 5.000374794006348, "learning_rate": 8.73612378400418e-07, "loss": 0.7125, "step": 2704 }, { "epoch": 2.611003861003861, "grad_norm": 5.096011638641357, "learning_rate": 8.693566218381854e-07, "loss": 0.7542, "step": 2705 }, { "epoch": 2.611969111969112, "grad_norm": 4.76540470123291, "learning_rate": 8.651107853426144e-07, "loss": 0.7275, "step": 2706 }, { "epoch": 2.612934362934363, "grad_norm": 5.00642204284668, "learning_rate": 8.608748735266359e-07, "loss": 0.7828, "step": 2707 }, { "epoch": 2.613899613899614, "grad_norm": 5.640402793884277, "learning_rate": 8.56648890992392e-07, "loss": 0.7276, "step": 2708 }, { "epoch": 2.614864864864865, "grad_norm": 4.735208034515381, "learning_rate": 8.524328423312356e-07, "loss": 0.7951, "step": 2709 }, { "epoch": 2.615830115830116, "grad_norm": 5.167140960693359, "learning_rate": 8.482267321237281e-07, "loss": 0.8058, "step": 2710 }, { "epoch": 2.6167953667953667, "grad_norm": 4.77217960357666, "learning_rate": 8.440305649396385e-07, "loss": 0.7804, "step": 2711 }, { "epoch": 2.617760617760618, "grad_norm": 4.561986446380615, "learning_rate": 8.398443453379268e-07, "loss": 0.7066, "step": 2712 }, { "epoch": 2.6187258687258685, "grad_norm": 4.737310886383057, "learning_rate": 8.356680778667503e-07, "loss": 0.7665, "step": 2713 }, { "epoch": 2.6196911196911197, "grad_norm": 4.607104778289795, "learning_rate": 8.31501767063445e-07, "loss": 0.7518, "step": 2714 }, { "epoch": 2.620656370656371, "grad_norm": 4.817365646362305, "learning_rate": 8.273454174545415e-07, "loss": 0.7109, "step": 2715 }, { "epoch": 2.6216216216216215, "grad_norm": 4.927873611450195, "learning_rate": 8.231990335557383e-07, "loss": 0.7881, "step": 2716 }, { "epoch": 2.6225868725868726, "grad_norm": 4.71023416519165, "learning_rate": 8.190626198719143e-07, "loss": 0.7334, "step": 2717 }, { "epoch": 2.6235521235521233, "grad_norm": 4.634425163269043, "learning_rate": 8.149361808971045e-07, "loss": 0.7472, "step": 2718 }, { "epoch": 2.6245173745173744, "grad_norm": 4.826690196990967, "learning_rate": 8.108197211145219e-07, "loss": 0.7295, "step": 2719 }, { "epoch": 2.6254826254826256, "grad_norm": 4.9166483879089355, "learning_rate": 8.067132449965254e-07, "loss": 0.7745, "step": 2720 }, { "epoch": 2.6264478764478767, "grad_norm": 4.875312805175781, "learning_rate": 8.02616757004634e-07, "loss": 0.7593, "step": 2721 }, { "epoch": 2.6274131274131274, "grad_norm": 5.300418853759766, "learning_rate": 7.985302615895119e-07, "loss": 0.7508, "step": 2722 }, { "epoch": 2.6283783783783785, "grad_norm": 5.059196949005127, "learning_rate": 7.944537631909666e-07, "loss": 0.7335, "step": 2723 }, { "epoch": 2.629343629343629, "grad_norm": 5.13384485244751, "learning_rate": 7.903872662379453e-07, "loss": 0.7502, "step": 2724 }, { "epoch": 2.6303088803088803, "grad_norm": 7.005805492401123, "learning_rate": 7.863307751485294e-07, "loss": 0.7461, "step": 2725 }, { "epoch": 2.6312741312741315, "grad_norm": 4.686102867126465, "learning_rate": 7.822842943299292e-07, "loss": 0.6944, "step": 2726 }, { "epoch": 2.632239382239382, "grad_norm": 4.866771697998047, "learning_rate": 7.782478281784789e-07, "loss": 0.7448, "step": 2727 }, { "epoch": 2.6332046332046333, "grad_norm": 5.382166385650635, "learning_rate": 7.742213810796306e-07, "loss": 0.7336, "step": 2728 }, { "epoch": 2.634169884169884, "grad_norm": 4.498585224151611, "learning_rate": 7.70204957407954e-07, "loss": 0.7431, "step": 2729 }, { "epoch": 2.635135135135135, "grad_norm": 4.91465950012207, "learning_rate": 7.661985615271283e-07, "loss": 0.7126, "step": 2730 }, { "epoch": 2.636100386100386, "grad_norm": 4.96571683883667, "learning_rate": 7.62202197789934e-07, "loss": 0.7496, "step": 2731 }, { "epoch": 2.637065637065637, "grad_norm": 4.990151405334473, "learning_rate": 7.582158705382581e-07, "loss": 0.7053, "step": 2732 }, { "epoch": 2.638030888030888, "grad_norm": 5.347865581512451, "learning_rate": 7.542395841030792e-07, "loss": 0.7827, "step": 2733 }, { "epoch": 2.638996138996139, "grad_norm": 5.1041646003723145, "learning_rate": 7.502733428044684e-07, "loss": 0.7695, "step": 2734 }, { "epoch": 2.63996138996139, "grad_norm": 4.740970611572266, "learning_rate": 7.463171509515832e-07, "loss": 0.7504, "step": 2735 }, { "epoch": 2.640926640926641, "grad_norm": 5.272965908050537, "learning_rate": 7.423710128426608e-07, "loss": 0.7869, "step": 2736 }, { "epoch": 2.641891891891892, "grad_norm": 5.517708778381348, "learning_rate": 7.384349327650198e-07, "loss": 0.7672, "step": 2737 }, { "epoch": 2.642857142857143, "grad_norm": 4.674668312072754, "learning_rate": 7.345089149950468e-07, "loss": 0.677, "step": 2738 }, { "epoch": 2.643822393822394, "grad_norm": 5.323797225952148, "learning_rate": 7.305929637981968e-07, "loss": 0.7509, "step": 2739 }, { "epoch": 2.6447876447876446, "grad_norm": 4.841979026794434, "learning_rate": 7.26687083428993e-07, "loss": 0.7427, "step": 2740 }, { "epoch": 2.6457528957528957, "grad_norm": 4.6486639976501465, "learning_rate": 7.227912781310099e-07, "loss": 0.7341, "step": 2741 }, { "epoch": 2.646718146718147, "grad_norm": 5.009174823760986, "learning_rate": 7.189055521368793e-07, "loss": 0.7128, "step": 2742 }, { "epoch": 2.6476833976833976, "grad_norm": 4.864299297332764, "learning_rate": 7.150299096682812e-07, "loss": 0.6911, "step": 2743 }, { "epoch": 2.6486486486486487, "grad_norm": 4.875516414642334, "learning_rate": 7.111643549359426e-07, "loss": 0.7303, "step": 2744 }, { "epoch": 2.6496138996138994, "grad_norm": 4.866146564483643, "learning_rate": 7.073088921396287e-07, "loss": 0.7547, "step": 2745 }, { "epoch": 2.6505791505791505, "grad_norm": 5.089334011077881, "learning_rate": 7.034635254681388e-07, "loss": 0.7418, "step": 2746 }, { "epoch": 2.6515444015444016, "grad_norm": 4.971090793609619, "learning_rate": 6.996282590993064e-07, "loss": 0.7802, "step": 2747 }, { "epoch": 2.6525096525096528, "grad_norm": 4.696529865264893, "learning_rate": 6.958030971999885e-07, "loss": 0.7193, "step": 2748 }, { "epoch": 2.6534749034749034, "grad_norm": 4.897533416748047, "learning_rate": 6.919880439260685e-07, "loss": 0.7503, "step": 2749 }, { "epoch": 2.6544401544401546, "grad_norm": 4.853023052215576, "learning_rate": 6.881831034224417e-07, "loss": 0.7827, "step": 2750 }, { "epoch": 2.6554054054054053, "grad_norm": 5.431102275848389, "learning_rate": 6.843882798230184e-07, "loss": 0.7609, "step": 2751 }, { "epoch": 2.6563706563706564, "grad_norm": 5.013240337371826, "learning_rate": 6.806035772507169e-07, "loss": 0.7242, "step": 2752 }, { "epoch": 2.6573359073359075, "grad_norm": 6.189661026000977, "learning_rate": 6.768289998174649e-07, "loss": 0.7656, "step": 2753 }, { "epoch": 2.658301158301158, "grad_norm": 5.869439125061035, "learning_rate": 6.730645516241785e-07, "loss": 0.7486, "step": 2754 }, { "epoch": 2.6592664092664093, "grad_norm": 5.316051006317139, "learning_rate": 6.693102367607795e-07, "loss": 0.736, "step": 2755 }, { "epoch": 2.66023166023166, "grad_norm": 5.744836330413818, "learning_rate": 6.655660593061719e-07, "loss": 0.7973, "step": 2756 }, { "epoch": 2.661196911196911, "grad_norm": 6.7306809425354, "learning_rate": 6.618320233282549e-07, "loss": 0.7991, "step": 2757 }, { "epoch": 2.6621621621621623, "grad_norm": 5.0219268798828125, "learning_rate": 6.581081328839012e-07, "loss": 0.7495, "step": 2758 }, { "epoch": 2.663127413127413, "grad_norm": 6.37299108505249, "learning_rate": 6.543943920189633e-07, "loss": 0.8041, "step": 2759 }, { "epoch": 2.664092664092664, "grad_norm": 5.728517055511475, "learning_rate": 6.50690804768267e-07, "loss": 0.794, "step": 2760 }, { "epoch": 2.665057915057915, "grad_norm": 5.198686122894287, "learning_rate": 6.46997375155608e-07, "loss": 0.7419, "step": 2761 }, { "epoch": 2.666023166023166, "grad_norm": 6.464615821838379, "learning_rate": 6.433141071937476e-07, "loss": 0.7804, "step": 2762 }, { "epoch": 2.666988416988417, "grad_norm": 5.360623836517334, "learning_rate": 6.396410048843992e-07, "loss": 0.7121, "step": 2763 }, { "epoch": 2.667953667953668, "grad_norm": 5.197407245635986, "learning_rate": 6.359780722182374e-07, "loss": 0.7632, "step": 2764 }, { "epoch": 2.668918918918919, "grad_norm": 5.302798748016357, "learning_rate": 6.323253131748908e-07, "loss": 0.7396, "step": 2765 }, { "epoch": 2.66988416988417, "grad_norm": 4.82261323928833, "learning_rate": 6.286827317229293e-07, "loss": 0.7525, "step": 2766 }, { "epoch": 2.6708494208494207, "grad_norm": 5.102636814117432, "learning_rate": 6.250503318198664e-07, "loss": 0.7265, "step": 2767 }, { "epoch": 2.671814671814672, "grad_norm": 5.377608299255371, "learning_rate": 6.214281174121518e-07, "loss": 0.7501, "step": 2768 }, { "epoch": 2.672779922779923, "grad_norm": 5.08730411529541, "learning_rate": 6.178160924351772e-07, "loss": 0.7851, "step": 2769 }, { "epoch": 2.6737451737451736, "grad_norm": 5.115046501159668, "learning_rate": 6.142142608132551e-07, "loss": 0.7517, "step": 2770 }, { "epoch": 2.6747104247104247, "grad_norm": 4.9304022789001465, "learning_rate": 6.106226264596293e-07, "loss": 0.796, "step": 2771 }, { "epoch": 2.6756756756756754, "grad_norm": 5.703690052032471, "learning_rate": 6.070411932764586e-07, "loss": 0.8081, "step": 2772 }, { "epoch": 2.6766409266409266, "grad_norm": 4.732430934906006, "learning_rate": 6.034699651548237e-07, "loss": 0.7497, "step": 2773 }, { "epoch": 2.6776061776061777, "grad_norm": 5.347548484802246, "learning_rate": 5.999089459747176e-07, "loss": 0.7851, "step": 2774 }, { "epoch": 2.678571428571429, "grad_norm": 5.401774883270264, "learning_rate": 5.9635813960504e-07, "loss": 0.7221, "step": 2775 }, { "epoch": 2.6795366795366795, "grad_norm": 4.879080295562744, "learning_rate": 5.928175499035926e-07, "loss": 0.7224, "step": 2776 }, { "epoch": 2.6805019305019306, "grad_norm": 4.759188175201416, "learning_rate": 5.89287180717083e-07, "loss": 0.7287, "step": 2777 }, { "epoch": 2.6814671814671813, "grad_norm": 5.15285587310791, "learning_rate": 5.857670358811096e-07, "loss": 0.7374, "step": 2778 }, { "epoch": 2.6824324324324325, "grad_norm": 5.157477378845215, "learning_rate": 5.822571192201654e-07, "loss": 0.7771, "step": 2779 }, { "epoch": 2.6833976833976836, "grad_norm": 5.011092662811279, "learning_rate": 5.787574345476299e-07, "loss": 0.733, "step": 2780 }, { "epoch": 2.6843629343629343, "grad_norm": 5.20009708404541, "learning_rate": 5.752679856657639e-07, "loss": 0.7647, "step": 2781 }, { "epoch": 2.6853281853281854, "grad_norm": 5.075290679931641, "learning_rate": 5.717887763657115e-07, "loss": 0.7759, "step": 2782 }, { "epoch": 2.686293436293436, "grad_norm": 4.610193252563477, "learning_rate": 5.68319810427489e-07, "loss": 0.7193, "step": 2783 }, { "epoch": 2.687258687258687, "grad_norm": 5.327815055847168, "learning_rate": 5.64861091619987e-07, "loss": 0.7594, "step": 2784 }, { "epoch": 2.6882239382239383, "grad_norm": 5.171527862548828, "learning_rate": 5.614126237009543e-07, "loss": 0.7818, "step": 2785 }, { "epoch": 2.689189189189189, "grad_norm": 5.051294326782227, "learning_rate": 5.579744104170137e-07, "loss": 0.753, "step": 2786 }, { "epoch": 2.69015444015444, "grad_norm": 5.165798664093018, "learning_rate": 5.545464555036418e-07, "loss": 0.7747, "step": 2787 }, { "epoch": 2.691119691119691, "grad_norm": 5.248556137084961, "learning_rate": 5.511287626851681e-07, "loss": 0.7073, "step": 2788 }, { "epoch": 2.692084942084942, "grad_norm": 4.877704620361328, "learning_rate": 5.477213356747746e-07, "loss": 0.7038, "step": 2789 }, { "epoch": 2.693050193050193, "grad_norm": 5.552155017852783, "learning_rate": 5.443241781744923e-07, "loss": 0.7444, "step": 2790 }, { "epoch": 2.6940154440154442, "grad_norm": 5.087784767150879, "learning_rate": 5.4093729387519e-07, "loss": 0.7423, "step": 2791 }, { "epoch": 2.694980694980695, "grad_norm": 4.976923942565918, "learning_rate": 5.375606864565786e-07, "loss": 0.7266, "step": 2792 }, { "epoch": 2.695945945945946, "grad_norm": 5.3527302742004395, "learning_rate": 5.341943595872012e-07, "loss": 0.7541, "step": 2793 }, { "epoch": 2.6969111969111967, "grad_norm": 4.737863063812256, "learning_rate": 5.308383169244346e-07, "loss": 0.7284, "step": 2794 }, { "epoch": 2.697876447876448, "grad_norm": 4.960092067718506, "learning_rate": 5.274925621144799e-07, "loss": 0.7621, "step": 2795 }, { "epoch": 2.698841698841699, "grad_norm": 5.724166393280029, "learning_rate": 5.241570987923616e-07, "loss": 0.7272, "step": 2796 }, { "epoch": 2.6998069498069497, "grad_norm": 4.756486892700195, "learning_rate": 5.208319305819209e-07, "loss": 0.7545, "step": 2797 }, { "epoch": 2.700772200772201, "grad_norm": 4.879261016845703, "learning_rate": 5.17517061095818e-07, "loss": 0.77, "step": 2798 }, { "epoch": 2.7017374517374515, "grad_norm": 5.662070274353027, "learning_rate": 5.142124939355209e-07, "loss": 0.7919, "step": 2799 }, { "epoch": 2.7027027027027026, "grad_norm": 4.820952892303467, "learning_rate": 5.109182326913053e-07, "loss": 0.7558, "step": 2800 }, { "epoch": 2.7036679536679538, "grad_norm": 4.854535102844238, "learning_rate": 5.076342809422485e-07, "loss": 0.7414, "step": 2801 }, { "epoch": 2.704633204633205, "grad_norm": 5.061697483062744, "learning_rate": 5.043606422562308e-07, "loss": 0.7599, "step": 2802 }, { "epoch": 2.7055984555984556, "grad_norm": 4.653299331665039, "learning_rate": 5.010973201899228e-07, "loss": 0.7372, "step": 2803 }, { "epoch": 2.7065637065637067, "grad_norm": 4.892110824584961, "learning_rate": 4.978443182887904e-07, "loss": 0.745, "step": 2804 }, { "epoch": 2.7075289575289574, "grad_norm": 4.729368686676025, "learning_rate": 4.946016400870857e-07, "loss": 0.6932, "step": 2805 }, { "epoch": 2.7084942084942085, "grad_norm": 5.155167579650879, "learning_rate": 4.913692891078437e-07, "loss": 0.7692, "step": 2806 }, { "epoch": 2.7094594594594597, "grad_norm": 4.721940040588379, "learning_rate": 4.88147268862883e-07, "loss": 0.7415, "step": 2807 }, { "epoch": 2.7104247104247103, "grad_norm": 5.228612422943115, "learning_rate": 4.849355828527924e-07, "loss": 0.7987, "step": 2808 }, { "epoch": 2.7113899613899615, "grad_norm": 5.587022304534912, "learning_rate": 4.817342345669363e-07, "loss": 0.7307, "step": 2809 }, { "epoch": 2.712355212355212, "grad_norm": 5.123222351074219, "learning_rate": 4.785432274834468e-07, "loss": 0.7939, "step": 2810 }, { "epoch": 2.7133204633204633, "grad_norm": 4.989898204803467, "learning_rate": 4.7536256506922507e-07, "loss": 0.7548, "step": 2811 }, { "epoch": 2.7142857142857144, "grad_norm": 5.328431606292725, "learning_rate": 4.721922507799248e-07, "loss": 0.7577, "step": 2812 }, { "epoch": 2.715250965250965, "grad_norm": 4.798675060272217, "learning_rate": 4.6903228805996406e-07, "loss": 0.757, "step": 2813 }, { "epoch": 2.7162162162162162, "grad_norm": 4.8693952560424805, "learning_rate": 4.658826803425087e-07, "loss": 0.7315, "step": 2814 }, { "epoch": 2.717181467181467, "grad_norm": 5.654747486114502, "learning_rate": 4.6274343104948163e-07, "loss": 0.7477, "step": 2815 }, { "epoch": 2.718146718146718, "grad_norm": 4.620950222015381, "learning_rate": 4.596145435915467e-07, "loss": 0.7357, "step": 2816 }, { "epoch": 2.719111969111969, "grad_norm": 4.7026567459106445, "learning_rate": 4.56496021368108e-07, "loss": 0.7102, "step": 2817 }, { "epoch": 2.7200772200772203, "grad_norm": 5.4267730712890625, "learning_rate": 4.533878677673131e-07, "loss": 0.7338, "step": 2818 }, { "epoch": 2.721042471042471, "grad_norm": 4.934582233428955, "learning_rate": 4.502900861660398e-07, "loss": 0.7352, "step": 2819 }, { "epoch": 2.722007722007722, "grad_norm": 5.04288911819458, "learning_rate": 4.472026799299045e-07, "loss": 0.7491, "step": 2820 }, { "epoch": 2.722972972972973, "grad_norm": 5.7374467849731445, "learning_rate": 4.441256524132431e-07, "loss": 0.7443, "step": 2821 }, { "epoch": 2.723938223938224, "grad_norm": 5.140861988067627, "learning_rate": 4.410590069591192e-07, "loss": 0.7479, "step": 2822 }, { "epoch": 2.724903474903475, "grad_norm": 5.024938583374023, "learning_rate": 4.380027468993153e-07, "loss": 0.7909, "step": 2823 }, { "epoch": 2.7258687258687258, "grad_norm": 6.299558639526367, "learning_rate": 4.3495687555433544e-07, "loss": 0.7606, "step": 2824 }, { "epoch": 2.726833976833977, "grad_norm": 5.034975051879883, "learning_rate": 4.3192139623339125e-07, "loss": 0.7507, "step": 2825 }, { "epoch": 2.7277992277992276, "grad_norm": 4.692763328552246, "learning_rate": 4.2889631223440255e-07, "loss": 0.7735, "step": 2826 }, { "epoch": 2.7287644787644787, "grad_norm": 5.218514919281006, "learning_rate": 4.2588162684400027e-07, "loss": 0.7663, "step": 2827 }, { "epoch": 2.72972972972973, "grad_norm": 5.270998954772949, "learning_rate": 4.228773433375166e-07, "loss": 0.795, "step": 2828 }, { "epoch": 2.730694980694981, "grad_norm": 4.963054656982422, "learning_rate": 4.198834649789818e-07, "loss": 0.7967, "step": 2829 }, { "epoch": 2.7316602316602316, "grad_norm": 4.898987293243408, "learning_rate": 4.1689999502111835e-07, "loss": 0.7473, "step": 2830 }, { "epoch": 2.7326254826254828, "grad_norm": 6.245449066162109, "learning_rate": 4.1392693670534223e-07, "loss": 0.7795, "step": 2831 }, { "epoch": 2.7335907335907335, "grad_norm": 5.4050421714782715, "learning_rate": 4.10964293261763e-07, "loss": 0.7471, "step": 2832 }, { "epoch": 2.7345559845559846, "grad_norm": 5.049526691436768, "learning_rate": 4.0801206790916815e-07, "loss": 0.7621, "step": 2833 }, { "epoch": 2.7355212355212357, "grad_norm": 4.990163326263428, "learning_rate": 4.0507026385502747e-07, "loss": 0.7178, "step": 2834 }, { "epoch": 2.7364864864864864, "grad_norm": 5.667870998382568, "learning_rate": 4.0213888429548877e-07, "loss": 0.8263, "step": 2835 }, { "epoch": 2.7374517374517375, "grad_norm": 5.045688152313232, "learning_rate": 3.9921793241537666e-07, "loss": 0.7756, "step": 2836 }, { "epoch": 2.738416988416988, "grad_norm": 5.432678699493408, "learning_rate": 3.9630741138818484e-07, "loss": 0.756, "step": 2837 }, { "epoch": 2.7393822393822393, "grad_norm": 5.021127700805664, "learning_rate": 3.9340732437607496e-07, "loss": 0.7377, "step": 2838 }, { "epoch": 2.7403474903474905, "grad_norm": 4.582510948181152, "learning_rate": 3.9051767452986777e-07, "loss": 0.7567, "step": 2839 }, { "epoch": 2.741312741312741, "grad_norm": 6.305225849151611, "learning_rate": 3.8763846498905187e-07, "loss": 0.7504, "step": 2840 }, { "epoch": 2.7422779922779923, "grad_norm": 5.327244758605957, "learning_rate": 3.8476969888176843e-07, "loss": 0.7837, "step": 2841 }, { "epoch": 2.743243243243243, "grad_norm": 5.5716233253479, "learning_rate": 3.8191137932481414e-07, "loss": 0.7646, "step": 2842 }, { "epoch": 2.744208494208494, "grad_norm": 6.495911598205566, "learning_rate": 3.790635094236317e-07, "loss": 0.7561, "step": 2843 }, { "epoch": 2.7451737451737452, "grad_norm": 4.862516403198242, "learning_rate": 3.762260922723182e-07, "loss": 0.7373, "step": 2844 }, { "epoch": 2.7461389961389964, "grad_norm": 5.6016998291015625, "learning_rate": 3.733991309536078e-07, "loss": 0.7469, "step": 2845 }, { "epoch": 2.747104247104247, "grad_norm": 5.131885051727295, "learning_rate": 3.7058262853887805e-07, "loss": 0.7512, "step": 2846 }, { "epoch": 2.748069498069498, "grad_norm": 4.686842441558838, "learning_rate": 3.6777658808814234e-07, "loss": 0.7179, "step": 2847 }, { "epoch": 2.749034749034749, "grad_norm": 5.378399848937988, "learning_rate": 3.649810126500475e-07, "loss": 0.8191, "step": 2848 }, { "epoch": 2.75, "grad_norm": 5.172610282897949, "learning_rate": 3.6219590526187063e-07, "loss": 0.7824, "step": 2849 }, { "epoch": 2.750965250965251, "grad_norm": 4.6634321212768555, "learning_rate": 3.594212689495147e-07, "loss": 0.7388, "step": 2850 }, { "epoch": 2.751930501930502, "grad_norm": 5.370039463043213, "learning_rate": 3.566571067275093e-07, "loss": 0.6907, "step": 2851 }, { "epoch": 2.752895752895753, "grad_norm": 4.233461380004883, "learning_rate": 3.5390342159900225e-07, "loss": 0.7345, "step": 2852 }, { "epoch": 2.7538610038610036, "grad_norm": 5.249642372131348, "learning_rate": 3.511602165557571e-07, "loss": 0.7928, "step": 2853 }, { "epoch": 2.7548262548262548, "grad_norm": 4.909470081329346, "learning_rate": 3.4842749457815406e-07, "loss": 0.7511, "step": 2854 }, { "epoch": 2.755791505791506, "grad_norm": 4.895900726318359, "learning_rate": 3.457052586351817e-07, "loss": 0.7404, "step": 2855 }, { "epoch": 2.756756756756757, "grad_norm": 5.3179216384887695, "learning_rate": 3.429935116844374e-07, "loss": 0.806, "step": 2856 }, { "epoch": 2.7577220077220077, "grad_norm": 4.825327396392822, "learning_rate": 3.402922566721223e-07, "loss": 0.7589, "step": 2857 }, { "epoch": 2.758687258687259, "grad_norm": 5.171874523162842, "learning_rate": 3.3760149653303877e-07, "loss": 0.7502, "step": 2858 }, { "epoch": 2.7596525096525095, "grad_norm": 4.987893104553223, "learning_rate": 3.3492123419058497e-07, "loss": 0.7794, "step": 2859 }, { "epoch": 2.7606177606177607, "grad_norm": 5.187206745147705, "learning_rate": 3.3225147255675716e-07, "loss": 0.8164, "step": 2860 }, { "epoch": 2.761583011583012, "grad_norm": 4.622416973114014, "learning_rate": 3.295922145321384e-07, "loss": 0.728, "step": 2861 }, { "epoch": 2.7625482625482625, "grad_norm": 5.248575687408447, "learning_rate": 3.269434630059054e-07, "loss": 0.7108, "step": 2862 }, { "epoch": 2.7635135135135136, "grad_norm": 5.089781761169434, "learning_rate": 3.243052208558151e-07, "loss": 0.7604, "step": 2863 }, { "epoch": 2.7644787644787643, "grad_norm": 5.344825744628906, "learning_rate": 3.216774909482101e-07, "loss": 0.8224, "step": 2864 }, { "epoch": 2.7654440154440154, "grad_norm": 4.867672920227051, "learning_rate": 3.1906027613800903e-07, "loss": 0.7378, "step": 2865 }, { "epoch": 2.7664092664092665, "grad_norm": 5.2442216873168945, "learning_rate": 3.1645357926870957e-07, "loss": 0.8109, "step": 2866 }, { "epoch": 2.7673745173745172, "grad_norm": 4.97896146774292, "learning_rate": 3.1385740317237733e-07, "loss": 0.7306, "step": 2867 }, { "epoch": 2.7683397683397684, "grad_norm": 4.8300089836120605, "learning_rate": 3.1127175066965276e-07, "loss": 0.7886, "step": 2868 }, { "epoch": 2.769305019305019, "grad_norm": 4.99078369140625, "learning_rate": 3.086966245697387e-07, "loss": 0.7539, "step": 2869 }, { "epoch": 2.77027027027027, "grad_norm": 5.40231466293335, "learning_rate": 3.06132027670405e-07, "loss": 0.763, "step": 2870 }, { "epoch": 2.7712355212355213, "grad_norm": 5.105173587799072, "learning_rate": 3.035779627579782e-07, "loss": 0.7722, "step": 2871 }, { "epoch": 2.7722007722007724, "grad_norm": 5.617315292358398, "learning_rate": 3.0103443260734557e-07, "loss": 0.8002, "step": 2872 }, { "epoch": 2.773166023166023, "grad_norm": 5.030704021453857, "learning_rate": 2.9850143998194635e-07, "loss": 0.7475, "step": 2873 }, { "epoch": 2.7741312741312742, "grad_norm": 5.375137805938721, "learning_rate": 2.9597898763377595e-07, "loss": 0.7464, "step": 2874 }, { "epoch": 2.775096525096525, "grad_norm": 4.688440799713135, "learning_rate": 2.9346707830336997e-07, "loss": 0.7595, "step": 2875 }, { "epoch": 2.776061776061776, "grad_norm": 4.799991130828857, "learning_rate": 2.9096571471981636e-07, "loss": 0.7795, "step": 2876 }, { "epoch": 2.777027027027027, "grad_norm": 5.409911632537842, "learning_rate": 2.8847489960074136e-07, "loss": 0.7786, "step": 2877 }, { "epoch": 2.777992277992278, "grad_norm": 5.098364353179932, "learning_rate": 2.8599463565231687e-07, "loss": 0.8011, "step": 2878 }, { "epoch": 2.778957528957529, "grad_norm": 5.285067081451416, "learning_rate": 2.8352492556924295e-07, "loss": 0.7708, "step": 2879 }, { "epoch": 2.7799227799227797, "grad_norm": 4.846736431121826, "learning_rate": 2.810657720347587e-07, "loss": 0.7992, "step": 2880 }, { "epoch": 2.780888030888031, "grad_norm": 4.905336380004883, "learning_rate": 2.786171777206337e-07, "loss": 0.7135, "step": 2881 }, { "epoch": 2.781853281853282, "grad_norm": 4.571996212005615, "learning_rate": 2.7617914528716317e-07, "loss": 0.79, "step": 2882 }, { "epoch": 2.782818532818533, "grad_norm": 5.4335527420043945, "learning_rate": 2.737516773831717e-07, "loss": 0.7503, "step": 2883 }, { "epoch": 2.7837837837837838, "grad_norm": 4.6285223960876465, "learning_rate": 2.713347766459984e-07, "loss": 0.7467, "step": 2884 }, { "epoch": 2.784749034749035, "grad_norm": 4.802557945251465, "learning_rate": 2.689284457015073e-07, "loss": 0.7849, "step": 2885 }, { "epoch": 2.7857142857142856, "grad_norm": 5.64417028427124, "learning_rate": 2.6653268716407944e-07, "loss": 0.7614, "step": 2886 }, { "epoch": 2.7866795366795367, "grad_norm": 5.090156555175781, "learning_rate": 2.6414750363660814e-07, "loss": 0.7759, "step": 2887 }, { "epoch": 2.787644787644788, "grad_norm": 5.5157880783081055, "learning_rate": 2.6177289771049274e-07, "loss": 0.7346, "step": 2888 }, { "epoch": 2.7886100386100385, "grad_norm": 4.951792240142822, "learning_rate": 2.5940887196564625e-07, "loss": 0.7408, "step": 2889 }, { "epoch": 2.7895752895752897, "grad_norm": 4.753869533538818, "learning_rate": 2.570554289704863e-07, "loss": 0.7802, "step": 2890 }, { "epoch": 2.7905405405405403, "grad_norm": 5.767086982727051, "learning_rate": 2.5471257128193094e-07, "loss": 0.7615, "step": 2891 }, { "epoch": 2.7915057915057915, "grad_norm": 5.374652862548828, "learning_rate": 2.5238030144539737e-07, "loss": 0.6966, "step": 2892 }, { "epoch": 2.7924710424710426, "grad_norm": 5.251793384552002, "learning_rate": 2.500586219947998e-07, "loss": 0.7848, "step": 2893 }, { "epoch": 2.7934362934362933, "grad_norm": 5.737508296966553, "learning_rate": 2.4774753545254936e-07, "loss": 0.7652, "step": 2894 }, { "epoch": 2.7944015444015444, "grad_norm": 5.297785758972168, "learning_rate": 2.4544704432954313e-07, "loss": 0.7817, "step": 2895 }, { "epoch": 2.795366795366795, "grad_norm": 5.0009636878967285, "learning_rate": 2.431571511251729e-07, "loss": 0.8392, "step": 2896 }, { "epoch": 2.7963320463320462, "grad_norm": 4.5383100509643555, "learning_rate": 2.408778583273108e-07, "loss": 0.7715, "step": 2897 }, { "epoch": 2.7972972972972974, "grad_norm": 4.934019565582275, "learning_rate": 2.3860916841231265e-07, "loss": 0.7585, "step": 2898 }, { "epoch": 2.7982625482625485, "grad_norm": 4.724292755126953, "learning_rate": 2.3635108384502003e-07, "loss": 0.7789, "step": 2899 }, { "epoch": 2.799227799227799, "grad_norm": 5.758883476257324, "learning_rate": 2.3410360707874724e-07, "loss": 0.7811, "step": 2900 }, { "epoch": 2.8001930501930503, "grad_norm": 4.7594475746154785, "learning_rate": 2.3186674055528436e-07, "loss": 0.7249, "step": 2901 }, { "epoch": 2.801158301158301, "grad_norm": 4.8631815910339355, "learning_rate": 2.2964048670489293e-07, "loss": 0.7789, "step": 2902 }, { "epoch": 2.802123552123552, "grad_norm": 5.352580547332764, "learning_rate": 2.2742484794630813e-07, "loss": 0.7284, "step": 2903 }, { "epoch": 2.8030888030888033, "grad_norm": 5.275538921356201, "learning_rate": 2.2521982668673003e-07, "loss": 0.6938, "step": 2904 }, { "epoch": 2.804054054054054, "grad_norm": 5.052992820739746, "learning_rate": 2.2302542532182226e-07, "loss": 0.7923, "step": 2905 }, { "epoch": 2.805019305019305, "grad_norm": 5.165040969848633, "learning_rate": 2.208416462357088e-07, "loss": 0.7193, "step": 2906 }, { "epoch": 2.8059845559845558, "grad_norm": 4.947180271148682, "learning_rate": 2.186684918009796e-07, "loss": 0.7712, "step": 2907 }, { "epoch": 2.806949806949807, "grad_norm": 5.16188907623291, "learning_rate": 2.165059643786749e-07, "loss": 0.7505, "step": 2908 }, { "epoch": 2.807915057915058, "grad_norm": 4.7499680519104, "learning_rate": 2.1435406631829192e-07, "loss": 0.7404, "step": 2909 }, { "epoch": 2.808880308880309, "grad_norm": 5.207340717315674, "learning_rate": 2.1221279995777833e-07, "loss": 0.7567, "step": 2910 }, { "epoch": 2.80984555984556, "grad_norm": 4.701102256774902, "learning_rate": 2.1008216762353207e-07, "loss": 0.7376, "step": 2911 }, { "epoch": 2.810810810810811, "grad_norm": 4.868410110473633, "learning_rate": 2.0796217163039588e-07, "loss": 0.7915, "step": 2912 }, { "epoch": 2.8117760617760617, "grad_norm": 5.2054667472839355, "learning_rate": 2.0585281428166071e-07, "loss": 0.8091, "step": 2913 }, { "epoch": 2.812741312741313, "grad_norm": 4.966453552246094, "learning_rate": 2.037540978690533e-07, "loss": 0.7822, "step": 2914 }, { "epoch": 2.813706563706564, "grad_norm": 4.5922532081604, "learning_rate": 2.016660246727431e-07, "loss": 0.7637, "step": 2915 }, { "epoch": 2.8146718146718146, "grad_norm": 4.752979755401611, "learning_rate": 1.995885969613365e-07, "loss": 0.7835, "step": 2916 }, { "epoch": 2.8156370656370657, "grad_norm": 4.5410027503967285, "learning_rate": 1.9752181699187133e-07, "loss": 0.7396, "step": 2917 }, { "epoch": 2.8166023166023164, "grad_norm": 5.296461582183838, "learning_rate": 1.9546568700981927e-07, "loss": 0.7387, "step": 2918 }, { "epoch": 2.8175675675675675, "grad_norm": 5.087576866149902, "learning_rate": 1.9342020924908e-07, "loss": 0.7756, "step": 2919 }, { "epoch": 2.8185328185328187, "grad_norm": 5.443266868591309, "learning_rate": 1.9138538593198141e-07, "loss": 0.7282, "step": 2920 }, { "epoch": 2.8194980694980694, "grad_norm": 5.033198356628418, "learning_rate": 1.8936121926927508e-07, "loss": 0.7677, "step": 2921 }, { "epoch": 2.8204633204633205, "grad_norm": 4.999009132385254, "learning_rate": 1.8734771146013297e-07, "loss": 0.7393, "step": 2922 }, { "epoch": 2.821428571428571, "grad_norm": 5.431258201599121, "learning_rate": 1.8534486469214962e-07, "loss": 0.7643, "step": 2923 }, { "epoch": 2.8223938223938223, "grad_norm": 5.312672138214111, "learning_rate": 1.8335268114133443e-07, "loss": 0.7246, "step": 2924 }, { "epoch": 2.8233590733590734, "grad_norm": 4.68110466003418, "learning_rate": 1.8137116297211266e-07, "loss": 0.7418, "step": 2925 }, { "epoch": 2.8243243243243246, "grad_norm": 5.30441951751709, "learning_rate": 1.7940031233732224e-07, "loss": 0.7541, "step": 2926 }, { "epoch": 2.8252895752895753, "grad_norm": 5.100651264190674, "learning_rate": 1.7744013137821038e-07, "loss": 0.7718, "step": 2927 }, { "epoch": 2.8262548262548264, "grad_norm": 4.791323184967041, "learning_rate": 1.7549062222443347e-07, "loss": 0.79, "step": 2928 }, { "epoch": 2.827220077220077, "grad_norm": 4.730557441711426, "learning_rate": 1.7355178699405172e-07, "loss": 0.7372, "step": 2929 }, { "epoch": 2.828185328185328, "grad_norm": 4.413017749786377, "learning_rate": 1.7162362779353014e-07, "loss": 0.7552, "step": 2930 }, { "epoch": 2.8291505791505793, "grad_norm": 4.423736572265625, "learning_rate": 1.69706146717733e-07, "loss": 0.768, "step": 2931 }, { "epoch": 2.83011583011583, "grad_norm": 4.955407619476318, "learning_rate": 1.677993458499272e-07, "loss": 0.7212, "step": 2932 }, { "epoch": 2.831081081081081, "grad_norm": 5.887596130371094, "learning_rate": 1.659032272617689e-07, "loss": 0.7751, "step": 2933 }, { "epoch": 2.832046332046332, "grad_norm": 5.196951866149902, "learning_rate": 1.6401779301331467e-07, "loss": 0.7516, "step": 2934 }, { "epoch": 2.833011583011583, "grad_norm": 5.311980247497559, "learning_rate": 1.621430451530115e-07, "loss": 0.749, "step": 2935 }, { "epoch": 2.833976833976834, "grad_norm": 5.174939155578613, "learning_rate": 1.602789857176945e-07, "loss": 0.7625, "step": 2936 }, { "epoch": 2.834942084942085, "grad_norm": 4.715925216674805, "learning_rate": 1.5842561673258705e-07, "loss": 0.8064, "step": 2937 }, { "epoch": 2.835907335907336, "grad_norm": 4.72916841506958, "learning_rate": 1.565829402112995e-07, "loss": 0.7617, "step": 2938 }, { "epoch": 2.836872586872587, "grad_norm": 5.083279132843018, "learning_rate": 1.5475095815582152e-07, "loss": 0.8057, "step": 2939 }, { "epoch": 2.8378378378378377, "grad_norm": 4.998275279998779, "learning_rate": 1.5292967255652657e-07, "loss": 0.7013, "step": 2940 }, { "epoch": 2.838803088803089, "grad_norm": 4.9365553855896, "learning_rate": 1.5111908539216845e-07, "loss": 0.7415, "step": 2941 }, { "epoch": 2.83976833976834, "grad_norm": 4.951539039611816, "learning_rate": 1.4931919862987143e-07, "loss": 0.761, "step": 2942 }, { "epoch": 2.8407335907335907, "grad_norm": 5.639662265777588, "learning_rate": 1.4753001422514125e-07, "loss": 0.7355, "step": 2943 }, { "epoch": 2.841698841698842, "grad_norm": 5.438138961791992, "learning_rate": 1.4575153412185072e-07, "loss": 0.7831, "step": 2944 }, { "epoch": 2.8426640926640925, "grad_norm": 5.091670513153076, "learning_rate": 1.4398376025224758e-07, "loss": 0.7536, "step": 2945 }, { "epoch": 2.8436293436293436, "grad_norm": 5.158112525939941, "learning_rate": 1.4222669453694215e-07, "loss": 0.7313, "step": 2946 }, { "epoch": 2.8445945945945947, "grad_norm": 5.176087379455566, "learning_rate": 1.4048033888491518e-07, "loss": 0.7136, "step": 2947 }, { "epoch": 2.8455598455598454, "grad_norm": 4.935799598693848, "learning_rate": 1.3874469519350896e-07, "loss": 0.8053, "step": 2948 }, { "epoch": 2.8465250965250966, "grad_norm": 4.985904216766357, "learning_rate": 1.3701976534842953e-07, "loss": 0.7894, "step": 2949 }, { "epoch": 2.8474903474903472, "grad_norm": 4.679719924926758, "learning_rate": 1.3530555122374335e-07, "loss": 0.7374, "step": 2950 }, { "epoch": 2.8484555984555984, "grad_norm": 5.193964004516602, "learning_rate": 1.336020546818706e-07, "loss": 0.704, "step": 2951 }, { "epoch": 2.8494208494208495, "grad_norm": 4.97507905960083, "learning_rate": 1.3190927757358974e-07, "loss": 0.7801, "step": 2952 }, { "epoch": 2.8503861003861006, "grad_norm": 4.854886531829834, "learning_rate": 1.3022722173803516e-07, "loss": 0.7937, "step": 2953 }, { "epoch": 2.8513513513513513, "grad_norm": 5.049645900726318, "learning_rate": 1.2855588900269057e-07, "loss": 0.8131, "step": 2954 }, { "epoch": 2.8523166023166024, "grad_norm": 4.761787414550781, "learning_rate": 1.26895281183389e-07, "loss": 0.7051, "step": 2955 }, { "epoch": 2.853281853281853, "grad_norm": 4.782556533813477, "learning_rate": 1.2524540008431173e-07, "loss": 0.7114, "step": 2956 }, { "epoch": 2.8542471042471043, "grad_norm": 5.348335266113281, "learning_rate": 1.2360624749798822e-07, "loss": 0.7948, "step": 2957 }, { "epoch": 2.8552123552123554, "grad_norm": 5.1443657875061035, "learning_rate": 1.2197782520528834e-07, "loss": 0.7879, "step": 2958 }, { "epoch": 2.856177606177606, "grad_norm": 5.073531627655029, "learning_rate": 1.2036013497542798e-07, "loss": 0.7857, "step": 2959 }, { "epoch": 2.857142857142857, "grad_norm": 4.599514961242676, "learning_rate": 1.187531785659568e-07, "loss": 0.7515, "step": 2960 }, { "epoch": 2.858108108108108, "grad_norm": 4.990761756896973, "learning_rate": 1.1715695772277047e-07, "loss": 0.7379, "step": 2961 }, { "epoch": 2.859073359073359, "grad_norm": 4.7035346031188965, "learning_rate": 1.1557147418009396e-07, "loss": 0.7611, "step": 2962 }, { "epoch": 2.86003861003861, "grad_norm": 5.617299556732178, "learning_rate": 1.1399672966049269e-07, "loss": 0.7622, "step": 2963 }, { "epoch": 2.861003861003861, "grad_norm": 4.790339946746826, "learning_rate": 1.1243272587485809e-07, "loss": 0.7545, "step": 2964 }, { "epoch": 2.861969111969112, "grad_norm": 5.162209510803223, "learning_rate": 1.1087946452241871e-07, "loss": 0.7503, "step": 2965 }, { "epoch": 2.862934362934363, "grad_norm": 4.988022327423096, "learning_rate": 1.0933694729072686e-07, "loss": 0.7562, "step": 2966 }, { "epoch": 2.863899613899614, "grad_norm": 4.976837635040283, "learning_rate": 1.0780517585566419e-07, "loss": 0.7398, "step": 2967 }, { "epoch": 2.864864864864865, "grad_norm": 4.8899922370910645, "learning_rate": 1.0628415188143726e-07, "loss": 0.7984, "step": 2968 }, { "epoch": 2.865830115830116, "grad_norm": 4.969272613525391, "learning_rate": 1.0477387702057529e-07, "loss": 0.7793, "step": 2969 }, { "epoch": 2.8667953667953667, "grad_norm": 5.026739597320557, "learning_rate": 1.0327435291393018e-07, "loss": 0.76, "step": 2970 }, { "epoch": 2.867760617760618, "grad_norm": 4.789086818695068, "learning_rate": 1.0178558119067316e-07, "loss": 0.7797, "step": 2971 }, { "epoch": 2.8687258687258685, "grad_norm": 4.7164225578308105, "learning_rate": 1.0030756346829151e-07, "loss": 0.7619, "step": 2972 }, { "epoch": 2.8696911196911197, "grad_norm": 4.880717754364014, "learning_rate": 9.884030135259182e-08, "loss": 0.736, "step": 2973 }, { "epoch": 2.870656370656371, "grad_norm": 4.9621171951293945, "learning_rate": 9.738379643769446e-08, "loss": 0.7874, "step": 2974 }, { "epoch": 2.8716216216216215, "grad_norm": 4.993303298950195, "learning_rate": 9.593805030602921e-08, "loss": 0.7836, "step": 2975 }, { "epoch": 2.8725868725868726, "grad_norm": 5.17083740234375, "learning_rate": 9.45030645283418e-08, "loss": 0.7683, "step": 2976 }, { "epoch": 2.8735521235521233, "grad_norm": 5.310078144073486, "learning_rate": 9.3078840663684e-08, "loss": 0.7087, "step": 2977 }, { "epoch": 2.8745173745173744, "grad_norm": 4.97788667678833, "learning_rate": 9.166538025941695e-08, "loss": 0.791, "step": 2978 }, { "epoch": 2.8754826254826256, "grad_norm": 4.916007041931152, "learning_rate": 9.02626848512067e-08, "loss": 0.7509, "step": 2979 }, { "epoch": 2.8764478764478767, "grad_norm": 4.999144554138184, "learning_rate": 8.887075596302197e-08, "loss": 0.7848, "step": 2980 }, { "epoch": 2.8774131274131274, "grad_norm": 5.314653396606445, "learning_rate": 8.748959510713861e-08, "loss": 0.7556, "step": 2981 }, { "epoch": 2.8783783783783785, "grad_norm": 5.080662727355957, "learning_rate": 8.611920378412963e-08, "loss": 0.793, "step": 2982 }, { "epoch": 2.879343629343629, "grad_norm": 5.135159015655518, "learning_rate": 8.475958348286962e-08, "loss": 0.7788, "step": 2983 }, { "epoch": 2.8803088803088803, "grad_norm": 4.982119083404541, "learning_rate": 8.341073568052804e-08, "loss": 0.7507, "step": 2984 }, { "epoch": 2.8812741312741315, "grad_norm": 4.8198018074035645, "learning_rate": 8.207266184257268e-08, "loss": 0.7585, "step": 2985 }, { "epoch": 2.882239382239382, "grad_norm": 4.909982681274414, "learning_rate": 8.074536342276618e-08, "loss": 0.7613, "step": 2986 }, { "epoch": 2.8832046332046333, "grad_norm": 4.988618850708008, "learning_rate": 7.94288418631639e-08, "loss": 0.8453, "step": 2987 }, { "epoch": 2.884169884169884, "grad_norm": 4.8787689208984375, "learning_rate": 7.81230985941106e-08, "loss": 0.7528, "step": 2988 }, { "epoch": 2.885135135135135, "grad_norm": 4.880171775817871, "learning_rate": 7.68281350342437e-08, "loss": 0.7982, "step": 2989 }, { "epoch": 2.886100386100386, "grad_norm": 4.924539566040039, "learning_rate": 7.554395259048664e-08, "loss": 0.7348, "step": 2990 }, { "epoch": 2.887065637065637, "grad_norm": 4.793140888214111, "learning_rate": 7.427055265805006e-08, "loss": 0.7723, "step": 2991 }, { "epoch": 2.888030888030888, "grad_norm": 4.872824668884277, "learning_rate": 7.300793662043282e-08, "loss": 0.746, "step": 2992 }, { "epoch": 2.888996138996139, "grad_norm": 4.631551742553711, "learning_rate": 7.175610584941317e-08, "loss": 0.7264, "step": 2993 }, { "epoch": 2.88996138996139, "grad_norm": 4.891885757446289, "learning_rate": 7.051506170505318e-08, "loss": 0.8092, "step": 2994 }, { "epoch": 2.890926640926641, "grad_norm": 5.292384624481201, "learning_rate": 6.92848055356976e-08, "loss": 0.7981, "step": 2995 }, { "epoch": 2.891891891891892, "grad_norm": 4.880120754241943, "learning_rate": 6.806533867796728e-08, "loss": 0.788, "step": 2996 }, { "epoch": 2.892857142857143, "grad_norm": 4.910423278808594, "learning_rate": 6.685666245676348e-08, "loss": 0.7714, "step": 2997 }, { "epoch": 2.893822393822394, "grad_norm": 4.717833042144775, "learning_rate": 6.565877818526245e-08, "loss": 0.7899, "step": 2998 }, { "epoch": 2.8947876447876446, "grad_norm": 5.327867031097412, "learning_rate": 6.447168716491648e-08, "loss": 0.8035, "step": 2999 }, { "epoch": 2.8957528957528957, "grad_norm": 4.775325298309326, "learning_rate": 6.329539068545054e-08, "loss": 0.728, "step": 3000 }, { "epoch": 2.896718146718147, "grad_norm": 5.513738632202148, "learning_rate": 6.212989002486236e-08, "loss": 0.7387, "step": 3001 }, { "epoch": 2.8976833976833976, "grad_norm": 4.775712490081787, "learning_rate": 6.097518644942013e-08, "loss": 0.7119, "step": 3002 }, { "epoch": 2.8986486486486487, "grad_norm": 4.993870258331299, "learning_rate": 5.983128121366255e-08, "loss": 0.7261, "step": 3003 }, { "epoch": 2.8996138996138994, "grad_norm": 4.865166187286377, "learning_rate": 5.869817556039325e-08, "loss": 0.7597, "step": 3004 }, { "epoch": 2.9005791505791505, "grad_norm": 5.616902828216553, "learning_rate": 5.757587072068638e-08, "loss": 0.8111, "step": 3005 }, { "epoch": 2.9015444015444016, "grad_norm": 5.1008429527282715, "learning_rate": 5.646436791387766e-08, "loss": 0.734, "step": 3006 }, { "epoch": 2.9025096525096528, "grad_norm": 4.946638107299805, "learning_rate": 5.536366834757223e-08, "loss": 0.7565, "step": 3007 }, { "epoch": 2.9034749034749034, "grad_norm": 4.6390061378479, "learning_rate": 5.427377321763239e-08, "loss": 0.7207, "step": 3008 }, { "epoch": 2.9044401544401546, "grad_norm": 4.830086708068848, "learning_rate": 5.319468370818537e-08, "loss": 0.7857, "step": 3009 }, { "epoch": 2.9054054054054053, "grad_norm": 4.865915298461914, "learning_rate": 5.212640099161559e-08, "loss": 0.8014, "step": 3010 }, { "epoch": 2.9063706563706564, "grad_norm": 4.121964454650879, "learning_rate": 5.106892622857018e-08, "loss": 0.7236, "step": 3011 }, { "epoch": 2.9073359073359075, "grad_norm": 5.154491901397705, "learning_rate": 5.002226056795123e-08, "loss": 0.7661, "step": 3012 }, { "epoch": 2.908301158301158, "grad_norm": 5.089937686920166, "learning_rate": 4.898640514691577e-08, "loss": 0.7765, "step": 3013 }, { "epoch": 2.9092664092664093, "grad_norm": 4.757026195526123, "learning_rate": 4.796136109088023e-08, "loss": 0.7503, "step": 3014 }, { "epoch": 2.91023166023166, "grad_norm": 4.677746295928955, "learning_rate": 4.694712951351044e-08, "loss": 0.7914, "step": 3015 }, { "epoch": 2.911196911196911, "grad_norm": 4.440855026245117, "learning_rate": 4.594371151673049e-08, "loss": 0.7548, "step": 3016 }, { "epoch": 2.9121621621621623, "grad_norm": 5.054029941558838, "learning_rate": 4.4951108190709466e-08, "loss": 0.7766, "step": 3017 }, { "epoch": 2.913127413127413, "grad_norm": 4.783627986907959, "learning_rate": 4.396932061387138e-08, "loss": 0.7003, "step": 3018 }, { "epoch": 2.914092664092664, "grad_norm": 4.913122653961182, "learning_rate": 4.299834985288631e-08, "loss": 0.7566, "step": 3019 }, { "epoch": 2.915057915057915, "grad_norm": 5.131570339202881, "learning_rate": 4.203819696267486e-08, "loss": 0.7938, "step": 3020 }, { "epoch": 2.916023166023166, "grad_norm": 4.725548267364502, "learning_rate": 4.10888629864048e-08, "loss": 0.7956, "step": 3021 }, { "epoch": 2.916988416988417, "grad_norm": 4.535856246948242, "learning_rate": 4.015034895548664e-08, "loss": 0.7068, "step": 3022 }, { "epoch": 2.917953667953668, "grad_norm": 4.978786945343018, "learning_rate": 3.922265588957586e-08, "loss": 0.7321, "step": 3023 }, { "epoch": 2.918918918918919, "grad_norm": 4.783323764801025, "learning_rate": 3.8305784796576203e-08, "loss": 0.7729, "step": 3024 }, { "epoch": 2.91988416988417, "grad_norm": 4.754459381103516, "learning_rate": 3.739973667262642e-08, "loss": 0.7401, "step": 3025 }, { "epoch": 2.9208494208494207, "grad_norm": 4.781511306762695, "learning_rate": 3.6504512502113506e-08, "loss": 0.7532, "step": 3026 }, { "epoch": 2.921814671814672, "grad_norm": 4.7534565925598145, "learning_rate": 3.562011325765946e-08, "loss": 0.7489, "step": 3027 }, { "epoch": 2.922779922779923, "grad_norm": 4.938658714294434, "learning_rate": 3.4746539900129e-08, "loss": 0.7472, "step": 3028 }, { "epoch": 2.9237451737451736, "grad_norm": 4.837621688842773, "learning_rate": 3.38837933786218e-08, "loss": 0.7241, "step": 3029 }, { "epoch": 2.9247104247104247, "grad_norm": 4.8044939041137695, "learning_rate": 3.3031874630478076e-08, "loss": 0.7251, "step": 3030 }, { "epoch": 2.9256756756756754, "grad_norm": 4.589970588684082, "learning_rate": 3.2190784581270786e-08, "loss": 0.7609, "step": 3031 }, { "epoch": 2.9266409266409266, "grad_norm": 5.225532531738281, "learning_rate": 3.1360524144810055e-08, "loss": 0.8266, "step": 3032 }, { "epoch": 2.9276061776061777, "grad_norm": 4.62082576751709, "learning_rate": 3.054109422313989e-08, "loss": 0.7959, "step": 3033 }, { "epoch": 2.928571428571429, "grad_norm": 4.918736457824707, "learning_rate": 2.9732495706535913e-08, "loss": 0.7869, "step": 3034 }, { "epoch": 2.9295366795366795, "grad_norm": 4.885461330413818, "learning_rate": 2.8934729473508726e-08, "loss": 0.7288, "step": 3035 }, { "epoch": 2.9305019305019306, "grad_norm": 4.932406425476074, "learning_rate": 2.814779639079723e-08, "loss": 0.7362, "step": 3036 }, { "epoch": 2.9314671814671813, "grad_norm": 5.010663032531738, "learning_rate": 2.737169731337197e-08, "loss": 0.7421, "step": 3037 }, { "epoch": 2.9324324324324325, "grad_norm": 5.282849311828613, "learning_rate": 2.6606433084432892e-08, "loss": 0.7304, "step": 3038 }, { "epoch": 2.9333976833976836, "grad_norm": 4.926002025604248, "learning_rate": 2.585200453540715e-08, "loss": 0.7978, "step": 3039 }, { "epoch": 2.9343629343629343, "grad_norm": 4.964679718017578, "learning_rate": 2.5108412485951306e-08, "loss": 0.7656, "step": 3040 }, { "epoch": 2.9353281853281854, "grad_norm": 4.717856407165527, "learning_rate": 2.437565774394579e-08, "loss": 0.8208, "step": 3041 }, { "epoch": 2.936293436293436, "grad_norm": 4.755115032196045, "learning_rate": 2.3653741105499338e-08, "loss": 0.7739, "step": 3042 }, { "epoch": 2.937258687258687, "grad_norm": 4.922940731048584, "learning_rate": 2.294266335494566e-08, "loss": 0.7458, "step": 3043 }, { "epoch": 2.9382239382239383, "grad_norm": 4.712069988250732, "learning_rate": 2.2242425264838997e-08, "loss": 0.7617, "step": 3044 }, { "epoch": 2.939189189189189, "grad_norm": 4.8516106605529785, "learning_rate": 2.1553027595959676e-08, "loss": 0.8564, "step": 3045 }, { "epoch": 2.94015444015444, "grad_norm": 5.042926788330078, "learning_rate": 2.0874471097311886e-08, "loss": 0.7718, "step": 3046 }, { "epoch": 2.941119691119691, "grad_norm": 4.81359338760376, "learning_rate": 2.020675650611703e-08, "loss": 0.6968, "step": 3047 }, { "epoch": 2.942084942084942, "grad_norm": 5.007816314697266, "learning_rate": 1.9549884547819253e-08, "loss": 0.7924, "step": 3048 }, { "epoch": 2.943050193050193, "grad_norm": 4.917684078216553, "learning_rate": 1.890385593608435e-08, "loss": 0.7483, "step": 3049 }, { "epoch": 2.9440154440154442, "grad_norm": 4.8487114906311035, "learning_rate": 1.82686713727942e-08, "loss": 0.7795, "step": 3050 }, { "epoch": 2.944980694980695, "grad_norm": 4.602555274963379, "learning_rate": 1.7644331548052342e-08, "loss": 0.7561, "step": 3051 }, { "epoch": 2.945945945945946, "grad_norm": 5.175610065460205, "learning_rate": 1.703083714017617e-08, "loss": 0.783, "step": 3052 }, { "epoch": 2.9469111969111967, "grad_norm": 5.09755277633667, "learning_rate": 1.6428188815703627e-08, "loss": 0.7955, "step": 3053 }, { "epoch": 2.947876447876448, "grad_norm": 4.747951030731201, "learning_rate": 1.583638722938652e-08, "loss": 0.7569, "step": 3054 }, { "epoch": 2.948841698841699, "grad_norm": 4.651163578033447, "learning_rate": 1.5255433024192746e-08, "loss": 0.756, "step": 3055 }, { "epoch": 2.9498069498069497, "grad_norm": 4.646449565887451, "learning_rate": 1.4685326831305192e-08, "loss": 0.7379, "step": 3056 }, { "epoch": 2.950772200772201, "grad_norm": 4.513665199279785, "learning_rate": 1.4126069270120613e-08, "loss": 0.7685, "step": 3057 }, { "epoch": 2.9517374517374515, "grad_norm": 4.776824474334717, "learning_rate": 1.3577660948249637e-08, "loss": 0.7659, "step": 3058 }, { "epoch": 2.9527027027027026, "grad_norm": 4.89701509475708, "learning_rate": 1.3040102461514548e-08, "loss": 0.7627, "step": 3059 }, { "epoch": 2.9536679536679538, "grad_norm": 4.525649070739746, "learning_rate": 1.2513394393950384e-08, "loss": 0.7146, "step": 3060 }, { "epoch": 2.954633204633205, "grad_norm": 4.8755316734313965, "learning_rate": 1.1997537317804953e-08, "loss": 0.7616, "step": 3061 }, { "epoch": 2.9555984555984556, "grad_norm": 4.713533878326416, "learning_rate": 1.1492531793534378e-08, "loss": 0.6995, "step": 3062 }, { "epoch": 2.9565637065637067, "grad_norm": 4.9669342041015625, "learning_rate": 1.0998378369806439e-08, "loss": 0.7321, "step": 3063 }, { "epoch": 2.9575289575289574, "grad_norm": 4.8890180587768555, "learning_rate": 1.0515077583498346e-08, "loss": 0.7737, "step": 3064 }, { "epoch": 2.9584942084942085, "grad_norm": 5.029232501983643, "learning_rate": 1.0042629959696737e-08, "loss": 0.7617, "step": 3065 }, { "epoch": 2.9594594594594597, "grad_norm": 4.720171928405762, "learning_rate": 9.581036011696577e-09, "loss": 0.7643, "step": 3066 }, { "epoch": 2.9604247104247103, "grad_norm": 4.735867977142334, "learning_rate": 9.130296241001146e-09, "loss": 0.7611, "step": 3067 }, { "epoch": 2.9613899613899615, "grad_norm": 4.768474578857422, "learning_rate": 8.690411137318722e-09, "loss": 0.7638, "step": 3068 }, { "epoch": 2.962355212355212, "grad_norm": 5.289783954620361, "learning_rate": 8.261381178568117e-09, "loss": 0.7812, "step": 3069 }, { "epoch": 2.9633204633204633, "grad_norm": 4.914121150970459, "learning_rate": 7.84320683087203e-09, "loss": 0.8026, "step": 3070 }, { "epoch": 2.9642857142857144, "grad_norm": 5.514883995056152, "learning_rate": 7.435888548559256e-09, "loss": 0.7871, "step": 3071 }, { "epoch": 2.965250965250965, "grad_norm": 5.027594089508057, "learning_rate": 7.0394267741646925e-09, "loss": 0.8206, "step": 3072 }, { "epoch": 2.9662162162162162, "grad_norm": 5.023658752441406, "learning_rate": 6.653821938427119e-09, "loss": 0.7488, "step": 3073 }, { "epoch": 2.967181467181467, "grad_norm": 5.08175802230835, "learning_rate": 6.279074460291412e-09, "loss": 0.7689, "step": 3074 }, { "epoch": 2.968146718146718, "grad_norm": 5.168396472930908, "learning_rate": 5.9151847469041125e-09, "loss": 0.7105, "step": 3075 }, { "epoch": 2.969111969111969, "grad_norm": 4.691485404968262, "learning_rate": 5.562153193616749e-09, "loss": 0.732, "step": 3076 }, { "epoch": 2.9700772200772203, "grad_norm": 4.773744106292725, "learning_rate": 5.2199801839836195e-09, "loss": 0.7813, "step": 3077 }, { "epoch": 2.971042471042471, "grad_norm": 4.5500617027282715, "learning_rate": 4.8886660897617955e-09, "loss": 0.8096, "step": 3078 }, { "epoch": 2.972007722007722, "grad_norm": 4.949380874633789, "learning_rate": 4.568211270910005e-09, "loss": 0.7535, "step": 3079 }, { "epoch": 2.972972972972973, "grad_norm": 4.675527572631836, "learning_rate": 4.25861607558864e-09, "loss": 0.7593, "step": 3080 }, { "epoch": 2.973938223938224, "grad_norm": 5.3298773765563965, "learning_rate": 3.9598808401597466e-09, "loss": 0.808, "step": 3081 }, { "epoch": 2.974903474903475, "grad_norm": 4.755771160125732, "learning_rate": 3.6720058891892564e-09, "loss": 0.7844, "step": 3082 }, { "epoch": 2.9758687258687258, "grad_norm": 4.750619411468506, "learning_rate": 3.3949915354380968e-09, "loss": 0.754, "step": 3083 }, { "epoch": 2.976833976833977, "grad_norm": 4.936907768249512, "learning_rate": 3.1288380798744078e-09, "loss": 0.7575, "step": 3084 }, { "epoch": 2.9777992277992276, "grad_norm": 4.750420093536377, "learning_rate": 2.8735458116602164e-09, "loss": 0.7027, "step": 3085 }, { "epoch": 2.9787644787644787, "grad_norm": 5.266950607299805, "learning_rate": 2.629115008160321e-09, "loss": 0.7765, "step": 3086 }, { "epoch": 2.97972972972973, "grad_norm": 5.024481773376465, "learning_rate": 2.395545934940069e-09, "loss": 0.7972, "step": 3087 }, { "epoch": 2.980694980694981, "grad_norm": 5.134521007537842, "learning_rate": 2.1728388457620263e-09, "loss": 0.76, "step": 3088 }, { "epoch": 2.9816602316602316, "grad_norm": 5.2135138511657715, "learning_rate": 1.9609939825881996e-09, "loss": 0.731, "step": 3089 }, { "epoch": 2.9826254826254828, "grad_norm": 4.94840669631958, "learning_rate": 1.7600115755789237e-09, "loss": 0.757, "step": 3090 }, { "epoch": 2.9835907335907335, "grad_norm": 5.121521472930908, "learning_rate": 1.5698918430928634e-09, "loss": 0.7751, "step": 3091 }, { "epoch": 2.9845559845559846, "grad_norm": 5.261178016662598, "learning_rate": 1.3906349916881223e-09, "loss": 0.7607, "step": 3092 }, { "epoch": 2.9855212355212357, "grad_norm": 4.680968284606934, "learning_rate": 1.2222412161200237e-09, "loss": 0.7714, "step": 3093 }, { "epoch": 2.9864864864864864, "grad_norm": 5.011970520019531, "learning_rate": 1.0647106993411094e-09, "loss": 0.7493, "step": 3094 }, { "epoch": 2.9874517374517375, "grad_norm": 4.68505334854126, "learning_rate": 9.180436125011405e-10, "loss": 0.7664, "step": 3095 }, { "epoch": 2.988416988416988, "grad_norm": 5.225539684295654, "learning_rate": 7.822401149482073e-10, "loss": 0.7783, "step": 3096 }, { "epoch": 2.9893822393822393, "grad_norm": 4.609559535980225, "learning_rate": 6.573003542276191e-10, "loss": 0.7553, "step": 3097 }, { "epoch": 2.9903474903474905, "grad_norm": 4.871151924133301, "learning_rate": 5.432244660819041e-10, "loss": 0.7406, "step": 3098 }, { "epoch": 2.991312741312741, "grad_norm": 4.843747615814209, "learning_rate": 4.4001257444747924e-10, "loss": 0.823, "step": 3099 }, { "epoch": 2.9922779922779923, "grad_norm": 4.668362140655518, "learning_rate": 3.4766479146242095e-10, "loss": 0.7192, "step": 3100 }, { "epoch": 2.993243243243243, "grad_norm": 5.2617316246032715, "learning_rate": 2.66181217457584e-10, "loss": 0.7802, "step": 3101 }, { "epoch": 2.994208494208494, "grad_norm": 5.22498083114624, "learning_rate": 1.955619409610421e-10, "loss": 0.7905, "step": 3102 }, { "epoch": 2.9951737451737452, "grad_norm": 5.287078857421875, "learning_rate": 1.358070386980881e-10, "loss": 0.7469, "step": 3103 }, { "epoch": 2.9961389961389964, "grad_norm": 4.943191051483154, "learning_rate": 8.691657559012356e-11, "loss": 0.8264, "step": 3104 }, { "epoch": 2.997104247104247, "grad_norm": 5.016753673553467, "learning_rate": 4.88906047546589e-11, "loss": 0.7621, "step": 3105 }, { "epoch": 2.998069498069498, "grad_norm": 4.927329063415527, "learning_rate": 2.1729167505313288e-11, "loss": 0.702, "step": 3106 }, { "epoch": 2.999034749034749, "grad_norm": 4.855465412139893, "learning_rate": 5.432293351814721e-12, "loss": 0.7606, "step": 3107 }, { "epoch": 3.0, "grad_norm": 4.297211170196533, "learning_rate": 0.0, "loss": 0.6057, "step": 3108 }, { "epoch": 3.0, "step": 3108, "total_flos": 8.955621288433418e+17, "train_loss": 0.6617098891558642, "train_runtime": 8230.5312, "train_samples_per_second": 36.226, "train_steps_per_second": 0.378 } ], "logging_steps": 1.0, "max_steps": 3108, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 1000, "total_flos": 8.955621288433418e+17, "train_batch_size": 24, "trial_name": null, "trial_params": null }