{ "best_metric": null, "best_model_checkpoint": null, "epoch": 15.0, "eval_steps": 500, "global_step": 1620, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.009259259259259259, "grad_norm": 17.116676330566406, "learning_rate": 2e-05, "loss": 2.3189, "step": 1 }, { "epoch": 0.018518518518518517, "grad_norm": 19.404932022094727, "learning_rate": 4e-05, "loss": 2.3402, "step": 2 }, { "epoch": 0.027777777777777776, "grad_norm": 15.749666213989258, "learning_rate": 6e-05, "loss": 1.8406, "step": 3 }, { "epoch": 0.037037037037037035, "grad_norm": 14.914640426635742, "learning_rate": 8e-05, "loss": 2.0528, "step": 4 }, { "epoch": 0.046296296296296294, "grad_norm": 11.358346939086914, "learning_rate": 0.0001, "loss": 1.1907, "step": 5 }, { "epoch": 0.05555555555555555, "grad_norm": 13.200481414794922, "learning_rate": 0.00012, "loss": 1.1732, "step": 6 }, { "epoch": 0.06481481481481481, "grad_norm": 11.002130508422852, "learning_rate": 0.00014, "loss": 0.6805, "step": 7 }, { "epoch": 0.07407407407407407, "grad_norm": 13.887001991271973, "learning_rate": 0.00016, "loss": 0.6608, "step": 8 }, { "epoch": 0.08333333333333333, "grad_norm": 8.605554580688477, "learning_rate": 0.00018, "loss": 0.6947, "step": 9 }, { "epoch": 0.09259259259259259, "grad_norm": 10.935001373291016, "learning_rate": 0.0002, "loss": 0.6098, "step": 10 }, { "epoch": 0.10185185185185185, "grad_norm": 10.06436538696289, "learning_rate": 0.00019987577639751553, "loss": 0.9718, "step": 11 }, { "epoch": 0.1111111111111111, "grad_norm": 11.505407333374023, "learning_rate": 0.00019975155279503107, "loss": 0.7152, "step": 12 }, { "epoch": 0.12037037037037036, "grad_norm": 5.3426899909973145, "learning_rate": 0.00019962732919254659, "loss": 0.2549, "step": 13 }, { "epoch": 0.12962962962962962, "grad_norm": Infinity, "learning_rate": 0.00019962732919254659, "loss": 1.1325, "step": 14 }, { "epoch": 0.1388888888888889, "grad_norm": 10.246034622192383, "learning_rate": 0.0001995031055900621, "loss": 1.418, "step": 15 }, { "epoch": 0.14814814814814814, "grad_norm": 10.24806022644043, "learning_rate": 0.00019937888198757767, "loss": 0.6118, "step": 16 }, { "epoch": 0.1574074074074074, "grad_norm": 13.599936485290527, "learning_rate": 0.0001992546583850932, "loss": 0.8439, "step": 17 }, { "epoch": 0.16666666666666666, "grad_norm": 8.4192476272583, "learning_rate": 0.0001991304347826087, "loss": 0.7078, "step": 18 }, { "epoch": 0.17592592592592593, "grad_norm": 7.700663089752197, "learning_rate": 0.00019900621118012425, "loss": 0.6345, "step": 19 }, { "epoch": 0.18518518518518517, "grad_norm": 9.339193344116211, "learning_rate": 0.00019888198757763977, "loss": 0.4559, "step": 20 }, { "epoch": 0.19444444444444445, "grad_norm": 16.562761306762695, "learning_rate": 0.00019875776397515528, "loss": 1.4581, "step": 21 }, { "epoch": 0.2037037037037037, "grad_norm": 13.53317928314209, "learning_rate": 0.00019863354037267082, "loss": 1.3443, "step": 22 }, { "epoch": 0.21296296296296297, "grad_norm": 9.449267387390137, "learning_rate": 0.00019850931677018634, "loss": 1.0892, "step": 23 }, { "epoch": 0.2222222222222222, "grad_norm": 19.80406951904297, "learning_rate": 0.00019838509316770186, "loss": 1.1021, "step": 24 }, { "epoch": 0.23148148148148148, "grad_norm": 9.569581985473633, "learning_rate": 0.0001982608695652174, "loss": 0.8147, "step": 25 }, { "epoch": 0.24074074074074073, "grad_norm": 6.435489654541016, "learning_rate": 0.00019813664596273294, "loss": 0.4822, "step": 26 }, { "epoch": 0.25, "grad_norm": 9.351337432861328, "learning_rate": 0.00019801242236024846, "loss": 1.7875, "step": 27 }, { "epoch": 0.25925925925925924, "grad_norm": 6.918874263763428, "learning_rate": 0.00019788819875776398, "loss": 1.0019, "step": 28 }, { "epoch": 0.26851851851851855, "grad_norm": 7.385853290557861, "learning_rate": 0.00019776397515527952, "loss": 0.9338, "step": 29 }, { "epoch": 0.2777777777777778, "grad_norm": 8.40998363494873, "learning_rate": 0.00019763975155279504, "loss": 0.4633, "step": 30 }, { "epoch": 0.28703703703703703, "grad_norm": 12.839616775512695, "learning_rate": 0.00019751552795031055, "loss": 0.8168, "step": 31 }, { "epoch": 0.2962962962962963, "grad_norm": 9.741667747497559, "learning_rate": 0.0001973913043478261, "loss": 0.7139, "step": 32 }, { "epoch": 0.3055555555555556, "grad_norm": 10.23273754119873, "learning_rate": 0.0001972670807453416, "loss": 0.5477, "step": 33 }, { "epoch": 0.3148148148148148, "grad_norm": 10.880776405334473, "learning_rate": 0.00019714285714285716, "loss": 1.6291, "step": 34 }, { "epoch": 0.32407407407407407, "grad_norm": 11.059282302856445, "learning_rate": 0.0001970186335403727, "loss": 1.1802, "step": 35 }, { "epoch": 0.3333333333333333, "grad_norm": 8.465683937072754, "learning_rate": 0.00019689440993788822, "loss": 1.5225, "step": 36 }, { "epoch": 0.3425925925925926, "grad_norm": 8.37371826171875, "learning_rate": 0.00019677018633540373, "loss": 0.5846, "step": 37 }, { "epoch": 0.35185185185185186, "grad_norm": 5.269484043121338, "learning_rate": 0.00019664596273291928, "loss": 0.3792, "step": 38 }, { "epoch": 0.3611111111111111, "grad_norm": 5.26035213470459, "learning_rate": 0.0001965217391304348, "loss": 0.4575, "step": 39 }, { "epoch": 0.37037037037037035, "grad_norm": 7.327594757080078, "learning_rate": 0.0001963975155279503, "loss": 0.6059, "step": 40 }, { "epoch": 0.37962962962962965, "grad_norm": 10.461058616638184, "learning_rate": 0.00019627329192546585, "loss": 0.8153, "step": 41 }, { "epoch": 0.3888888888888889, "grad_norm": 10.147456169128418, "learning_rate": 0.00019614906832298137, "loss": 0.4933, "step": 42 }, { "epoch": 0.39814814814814814, "grad_norm": 6.203906536102295, "learning_rate": 0.0001960248447204969, "loss": 0.4631, "step": 43 }, { "epoch": 0.4074074074074074, "grad_norm": 10.029688835144043, "learning_rate": 0.00019590062111801243, "loss": 1.1486, "step": 44 }, { "epoch": 0.4166666666666667, "grad_norm": 8.837681770324707, "learning_rate": 0.00019577639751552797, "loss": 1.4024, "step": 45 }, { "epoch": 0.42592592592592593, "grad_norm": 8.74865436553955, "learning_rate": 0.0001956521739130435, "loss": 1.5168, "step": 46 }, { "epoch": 0.4351851851851852, "grad_norm": 5.9198479652404785, "learning_rate": 0.000195527950310559, "loss": 0.4223, "step": 47 }, { "epoch": 0.4444444444444444, "grad_norm": 3.298370122909546, "learning_rate": 0.00019540372670807455, "loss": 0.1694, "step": 48 }, { "epoch": 0.4537037037037037, "grad_norm": 8.61251163482666, "learning_rate": 0.00019527950310559006, "loss": 1.2626, "step": 49 }, { "epoch": 0.46296296296296297, "grad_norm": 9.859038352966309, "learning_rate": 0.00019515527950310558, "loss": 1.4776, "step": 50 }, { "epoch": 0.4722222222222222, "grad_norm": 10.356538772583008, "learning_rate": 0.00019503105590062112, "loss": 1.2718, "step": 51 }, { "epoch": 0.48148148148148145, "grad_norm": 5.658873081207275, "learning_rate": 0.00019490683229813667, "loss": 0.1858, "step": 52 }, { "epoch": 0.49074074074074076, "grad_norm": 10.131982803344727, "learning_rate": 0.00019478260869565218, "loss": 0.6037, "step": 53 }, { "epoch": 0.5, "grad_norm": 11.389909744262695, "learning_rate": 0.00019465838509316773, "loss": 1.4944, "step": 54 }, { "epoch": 0.5092592592592593, "grad_norm": 12.465958595275879, "learning_rate": 0.00019453416149068324, "loss": 2.2114, "step": 55 }, { "epoch": 0.5185185185185185, "grad_norm": 4.682929039001465, "learning_rate": 0.00019440993788819876, "loss": 0.1766, "step": 56 }, { "epoch": 0.5277777777777778, "grad_norm": 6.179277420043945, "learning_rate": 0.0001942857142857143, "loss": 0.4002, "step": 57 }, { "epoch": 0.5370370370370371, "grad_norm": 6.188724994659424, "learning_rate": 0.00019416149068322982, "loss": 0.4758, "step": 58 }, { "epoch": 0.5462962962962963, "grad_norm": 8.731796264648438, "learning_rate": 0.00019403726708074534, "loss": 1.4991, "step": 59 }, { "epoch": 0.5555555555555556, "grad_norm": 8.394841194152832, "learning_rate": 0.00019391304347826088, "loss": 0.506, "step": 60 }, { "epoch": 0.5648148148148148, "grad_norm": 13.798285484313965, "learning_rate": 0.00019378881987577642, "loss": 2.0693, "step": 61 }, { "epoch": 0.5740740740740741, "grad_norm": 6.169430255889893, "learning_rate": 0.00019366459627329194, "loss": 0.4175, "step": 62 }, { "epoch": 0.5833333333333334, "grad_norm": 7.522382736206055, "learning_rate": 0.00019354037267080746, "loss": 0.6897, "step": 63 }, { "epoch": 0.5925925925925926, "grad_norm": 5.0495219230651855, "learning_rate": 0.000193416149068323, "loss": 0.2197, "step": 64 }, { "epoch": 0.6018518518518519, "grad_norm": 2.5212173461914062, "learning_rate": 0.00019329192546583852, "loss": 0.1039, "step": 65 }, { "epoch": 0.6111111111111112, "grad_norm": 9.070035934448242, "learning_rate": 0.00019316770186335403, "loss": 1.0819, "step": 66 }, { "epoch": 0.6203703703703703, "grad_norm": Infinity, "learning_rate": 0.00019316770186335403, "loss": 0.7222, "step": 67 }, { "epoch": 0.6296296296296297, "grad_norm": 8.20578384399414, "learning_rate": 0.00019304347826086958, "loss": 1.1301, "step": 68 }, { "epoch": 0.6388888888888888, "grad_norm": 8.681023597717285, "learning_rate": 0.0001929192546583851, "loss": 1.3619, "step": 69 }, { "epoch": 0.6481481481481481, "grad_norm": NaN, "learning_rate": 0.0001929192546583851, "loss": 1.2046, "step": 70 }, { "epoch": 0.6574074074074074, "grad_norm": NaN, "learning_rate": 0.0001929192546583851, "loss": 0.6509, "step": 71 }, { "epoch": 0.6666666666666666, "grad_norm": Infinity, "learning_rate": 0.0001929192546583851, "loss": 1.5734, "step": 72 }, { "epoch": 0.6759259259259259, "grad_norm": Infinity, "learning_rate": 0.0001929192546583851, "loss": 1.1999, "step": 73 }, { "epoch": 0.6851851851851852, "grad_norm": 120.05274963378906, "learning_rate": 0.0001927950310559006, "loss": 1.0121, "step": 74 }, { "epoch": 0.6944444444444444, "grad_norm": 6.013508319854736, "learning_rate": 0.00019267080745341618, "loss": 0.245, "step": 75 }, { "epoch": 0.7037037037037037, "grad_norm": 10.709644317626953, "learning_rate": 0.0001925465838509317, "loss": 1.7426, "step": 76 }, { "epoch": 0.7129629629629629, "grad_norm": 8.936174392700195, "learning_rate": 0.0001924223602484472, "loss": 2.0067, "step": 77 }, { "epoch": 0.7222222222222222, "grad_norm": 10.918343544006348, "learning_rate": 0.00019229813664596275, "loss": 1.7653, "step": 78 }, { "epoch": 0.7314814814814815, "grad_norm": 7.559531211853027, "learning_rate": 0.00019217391304347827, "loss": 0.7266, "step": 79 }, { "epoch": 0.7407407407407407, "grad_norm": 5.118418216705322, "learning_rate": 0.0001920496894409938, "loss": 0.2286, "step": 80 }, { "epoch": 0.75, "grad_norm": 8.13199234008789, "learning_rate": 0.00019192546583850933, "loss": 1.4382, "step": 81 }, { "epoch": 0.7592592592592593, "grad_norm": 5.355680465698242, "learning_rate": 0.00019180124223602485, "loss": 0.2843, "step": 82 }, { "epoch": 0.7685185185185185, "grad_norm": 10.049469947814941, "learning_rate": 0.00019167701863354036, "loss": 0.9342, "step": 83 }, { "epoch": 0.7777777777777778, "grad_norm": 5.9033918380737305, "learning_rate": 0.0001915527950310559, "loss": 0.3322, "step": 84 }, { "epoch": 0.7870370370370371, "grad_norm": 11.78792953491211, "learning_rate": 0.00019142857142857145, "loss": 1.5358, "step": 85 }, { "epoch": 0.7962962962962963, "grad_norm": 9.4061861038208, "learning_rate": 0.00019130434782608697, "loss": 1.7081, "step": 86 }, { "epoch": 0.8055555555555556, "grad_norm": 9.609070777893066, "learning_rate": 0.0001911801242236025, "loss": 1.2752, "step": 87 }, { "epoch": 0.8148148148148148, "grad_norm": 7.137119770050049, "learning_rate": 0.00019105590062111803, "loss": 0.6856, "step": 88 }, { "epoch": 0.8240740740740741, "grad_norm": 8.402849197387695, "learning_rate": 0.00019093167701863354, "loss": 0.8608, "step": 89 }, { "epoch": 0.8333333333333334, "grad_norm": 7.3328680992126465, "learning_rate": 0.0001908074534161491, "loss": 0.9065, "step": 90 }, { "epoch": 0.8425925925925926, "grad_norm": 7.892049312591553, "learning_rate": 0.0001906832298136646, "loss": 0.7751, "step": 91 }, { "epoch": 0.8518518518518519, "grad_norm": 7.179913520812988, "learning_rate": 0.00019055900621118012, "loss": 0.854, "step": 92 }, { "epoch": 0.8611111111111112, "grad_norm": 8.245528221130371, "learning_rate": 0.00019043478260869566, "loss": 1.0387, "step": 93 }, { "epoch": 0.8703703703703703, "grad_norm": 8.712199211120605, "learning_rate": 0.0001903105590062112, "loss": 1.2923, "step": 94 }, { "epoch": 0.8796296296296297, "grad_norm": 4.641773223876953, "learning_rate": 0.00019018633540372672, "loss": 0.2925, "step": 95 }, { "epoch": 0.8888888888888888, "grad_norm": 7.5324177742004395, "learning_rate": 0.00019006211180124224, "loss": 0.7769, "step": 96 }, { "epoch": 0.8981481481481481, "grad_norm": 9.859864234924316, "learning_rate": 0.00018993788819875778, "loss": 1.7021, "step": 97 }, { "epoch": 0.9074074074074074, "grad_norm": 9.314699172973633, "learning_rate": 0.0001898136645962733, "loss": 1.0719, "step": 98 }, { "epoch": 0.9166666666666666, "grad_norm": 11.670292854309082, "learning_rate": 0.00018968944099378881, "loss": 1.6333, "step": 99 }, { "epoch": 0.9259259259259259, "grad_norm": 12.171489715576172, "learning_rate": 0.00018956521739130436, "loss": 0.8229, "step": 100 }, { "epoch": 0.9351851851851852, "grad_norm": 8.804001808166504, "learning_rate": 0.00018944099378881987, "loss": 1.3885, "step": 101 }, { "epoch": 0.9444444444444444, "grad_norm": 8.818242073059082, "learning_rate": 0.00018931677018633542, "loss": 1.3888, "step": 102 }, { "epoch": 0.9537037037037037, "grad_norm": 7.723133563995361, "learning_rate": 0.00018919254658385096, "loss": 0.8324, "step": 103 }, { "epoch": 0.9629629629629629, "grad_norm": 8.167946815490723, "learning_rate": 0.00018906832298136648, "loss": 1.1976, "step": 104 }, { "epoch": 0.9722222222222222, "grad_norm": 8.978958129882812, "learning_rate": 0.000188944099378882, "loss": 0.7289, "step": 105 }, { "epoch": 0.9814814814814815, "grad_norm": 8.901488304138184, "learning_rate": 0.00018881987577639754, "loss": 1.2306, "step": 106 }, { "epoch": 0.9907407407407407, "grad_norm": 5.3030877113342285, "learning_rate": 0.00018869565217391305, "loss": 0.2643, "step": 107 }, { "epoch": 1.0, "grad_norm": 10.44385814666748, "learning_rate": 0.00018857142857142857, "loss": 1.4146, "step": 108 }, { "epoch": 1.0092592592592593, "grad_norm": 2.322072744369507, "learning_rate": 0.00018844720496894411, "loss": 0.0682, "step": 109 }, { "epoch": 1.0185185185185186, "grad_norm": 9.836686134338379, "learning_rate": 0.00018832298136645963, "loss": 0.8827, "step": 110 }, { "epoch": 1.0277777777777777, "grad_norm": 5.424779415130615, "learning_rate": 0.00018819875776397517, "loss": 0.385, "step": 111 }, { "epoch": 1.037037037037037, "grad_norm": 8.20822525024414, "learning_rate": 0.0001880745341614907, "loss": 0.6324, "step": 112 }, { "epoch": 1.0462962962962963, "grad_norm": 6.136359691619873, "learning_rate": 0.00018795031055900623, "loss": 0.3507, "step": 113 }, { "epoch": 1.0555555555555556, "grad_norm": 12.00852108001709, "learning_rate": 0.00018782608695652175, "loss": 0.7757, "step": 114 }, { "epoch": 1.0648148148148149, "grad_norm": 10.327400207519531, "learning_rate": 0.00018770186335403727, "loss": 0.6816, "step": 115 }, { "epoch": 1.074074074074074, "grad_norm": 8.669729232788086, "learning_rate": 0.0001875776397515528, "loss": 0.5291, "step": 116 }, { "epoch": 1.0833333333333333, "grad_norm": 7.052165985107422, "learning_rate": 0.00018745341614906833, "loss": 0.5359, "step": 117 }, { "epoch": 1.0925925925925926, "grad_norm": 11.054783821105957, "learning_rate": 0.00018732919254658384, "loss": 0.8106, "step": 118 }, { "epoch": 1.1018518518518519, "grad_norm": 10.095086097717285, "learning_rate": 0.00018720496894409939, "loss": 0.903, "step": 119 }, { "epoch": 1.1111111111111112, "grad_norm": 23.136199951171875, "learning_rate": 0.0001870807453416149, "loss": 1.0593, "step": 120 }, { "epoch": 1.1203703703703705, "grad_norm": 8.750686645507812, "learning_rate": 0.00018695652173913045, "loss": 0.2787, "step": 121 }, { "epoch": 1.1296296296296295, "grad_norm": 7.291873931884766, "learning_rate": 0.000186832298136646, "loss": 0.5422, "step": 122 }, { "epoch": 1.1388888888888888, "grad_norm": 9.136726379394531, "learning_rate": 0.0001867080745341615, "loss": 0.28, "step": 123 }, { "epoch": 1.1481481481481481, "grad_norm": 7.925637245178223, "learning_rate": 0.00018658385093167702, "loss": 0.6995, "step": 124 }, { "epoch": 1.1574074074074074, "grad_norm": 4.114215850830078, "learning_rate": 0.00018645962732919257, "loss": 0.1883, "step": 125 }, { "epoch": 1.1666666666666667, "grad_norm": 8.64969539642334, "learning_rate": 0.00018633540372670808, "loss": 0.7698, "step": 126 }, { "epoch": 1.175925925925926, "grad_norm": 10.038046836853027, "learning_rate": 0.0001862111801242236, "loss": 1.2307, "step": 127 }, { "epoch": 1.1851851851851851, "grad_norm": 10.920683860778809, "learning_rate": 0.00018608695652173914, "loss": 1.0984, "step": 128 }, { "epoch": 1.1944444444444444, "grad_norm": 5.161189556121826, "learning_rate": 0.00018596273291925466, "loss": 0.1952, "step": 129 }, { "epoch": 1.2037037037037037, "grad_norm": 10.901082038879395, "learning_rate": 0.0001858385093167702, "loss": 0.5871, "step": 130 }, { "epoch": 1.212962962962963, "grad_norm": 5.366225719451904, "learning_rate": 0.00018571428571428572, "loss": 0.1007, "step": 131 }, { "epoch": 1.2222222222222223, "grad_norm": 7.6988959312438965, "learning_rate": 0.00018559006211180126, "loss": 0.6182, "step": 132 }, { "epoch": 1.2314814814814814, "grad_norm": 5.929910659790039, "learning_rate": 0.00018546583850931678, "loss": 0.3714, "step": 133 }, { "epoch": 1.2407407407407407, "grad_norm": 4.514689922332764, "learning_rate": 0.0001853416149068323, "loss": 0.1787, "step": 134 }, { "epoch": 1.25, "grad_norm": 12.674291610717773, "learning_rate": 0.00018521739130434784, "loss": 1.6911, "step": 135 }, { "epoch": 1.2592592592592593, "grad_norm": 9.21278190612793, "learning_rate": 0.00018509316770186335, "loss": 0.7563, "step": 136 }, { "epoch": 1.2685185185185186, "grad_norm": 8.567093849182129, "learning_rate": 0.00018496894409937887, "loss": 0.4734, "step": 137 }, { "epoch": 1.2777777777777777, "grad_norm": 8.084855079650879, "learning_rate": 0.0001848447204968944, "loss": 1.0525, "step": 138 }, { "epoch": 1.287037037037037, "grad_norm": 5.928450107574463, "learning_rate": 0.00018472049689440996, "loss": 0.5145, "step": 139 }, { "epoch": 1.2962962962962963, "grad_norm": 7.130023956298828, "learning_rate": 0.00018459627329192547, "loss": 0.6849, "step": 140 }, { "epoch": 1.3055555555555556, "grad_norm": 10.378190994262695, "learning_rate": 0.00018447204968944102, "loss": 0.986, "step": 141 }, { "epoch": 1.3148148148148149, "grad_norm": 8.725743293762207, "learning_rate": 0.00018434782608695653, "loss": 0.8948, "step": 142 }, { "epoch": 1.324074074074074, "grad_norm": 9.408001899719238, "learning_rate": 0.00018422360248447205, "loss": 1.6738, "step": 143 }, { "epoch": 1.3333333333333333, "grad_norm": 8.074913024902344, "learning_rate": 0.0001840993788819876, "loss": 0.6298, "step": 144 }, { "epoch": 1.3425925925925926, "grad_norm": 7.581278324127197, "learning_rate": 0.0001839751552795031, "loss": 0.3591, "step": 145 }, { "epoch": 1.3518518518518519, "grad_norm": 6.366710186004639, "learning_rate": 0.00018385093167701863, "loss": 0.6695, "step": 146 }, { "epoch": 1.3611111111111112, "grad_norm": 7.74171781539917, "learning_rate": 0.00018372670807453417, "loss": 0.4759, "step": 147 }, { "epoch": 1.3703703703703702, "grad_norm": 7.979270935058594, "learning_rate": 0.0001836024844720497, "loss": 0.4445, "step": 148 }, { "epoch": 1.3796296296296298, "grad_norm": 5.539051532745361, "learning_rate": 0.00018347826086956523, "loss": 0.2873, "step": 149 }, { "epoch": 1.3888888888888888, "grad_norm": 8.086750030517578, "learning_rate": 0.00018335403726708074, "loss": 0.1622, "step": 150 }, { "epoch": 1.3981481481481481, "grad_norm": 4.4308671951293945, "learning_rate": 0.0001832298136645963, "loss": 0.21, "step": 151 }, { "epoch": 1.4074074074074074, "grad_norm": 8.82310676574707, "learning_rate": 0.0001831055900621118, "loss": 0.8503, "step": 152 }, { "epoch": 1.4166666666666667, "grad_norm": 6.758204460144043, "learning_rate": 0.00018298136645962735, "loss": 0.3063, "step": 153 }, { "epoch": 1.425925925925926, "grad_norm": 5.448921203613281, "learning_rate": 0.00018285714285714286, "loss": 0.2057, "step": 154 }, { "epoch": 1.4351851851851851, "grad_norm": 5.913497447967529, "learning_rate": 0.00018273291925465838, "loss": 0.3591, "step": 155 }, { "epoch": 1.4444444444444444, "grad_norm": 8.914529800415039, "learning_rate": 0.00018260869565217392, "loss": 0.8959, "step": 156 }, { "epoch": 1.4537037037037037, "grad_norm": 10.08364486694336, "learning_rate": 0.00018248447204968947, "loss": 0.768, "step": 157 }, { "epoch": 1.462962962962963, "grad_norm": 7.685210704803467, "learning_rate": 0.00018236024844720498, "loss": 0.8345, "step": 158 }, { "epoch": 1.4722222222222223, "grad_norm": 4.343817710876465, "learning_rate": 0.0001822360248447205, "loss": 0.2126, "step": 159 }, { "epoch": 1.4814814814814814, "grad_norm": 3.930588483810425, "learning_rate": 0.00018211180124223604, "loss": 0.1506, "step": 160 }, { "epoch": 1.4907407407407407, "grad_norm": 6.4112958908081055, "learning_rate": 0.00018198757763975156, "loss": 0.2868, "step": 161 }, { "epoch": 1.5, "grad_norm": 9.307744026184082, "learning_rate": 0.00018186335403726708, "loss": 1.069, "step": 162 }, { "epoch": 1.5092592592592593, "grad_norm": 11.457921028137207, "learning_rate": 0.00018173913043478262, "loss": 0.6819, "step": 163 }, { "epoch": 1.5185185185185186, "grad_norm": 4.507647514343262, "learning_rate": 0.00018161490683229814, "loss": 0.292, "step": 164 }, { "epoch": 1.5277777777777777, "grad_norm": 5.411582946777344, "learning_rate": 0.00018149068322981365, "loss": 0.2228, "step": 165 }, { "epoch": 1.5370370370370372, "grad_norm": 9.850244522094727, "learning_rate": 0.00018136645962732922, "loss": 0.9859, "step": 166 }, { "epoch": 1.5462962962962963, "grad_norm": 8.918268203735352, "learning_rate": 0.00018124223602484474, "loss": 0.7327, "step": 167 }, { "epoch": 1.5555555555555556, "grad_norm": 13.585978507995605, "learning_rate": 0.00018111801242236026, "loss": 1.3928, "step": 168 }, { "epoch": 1.5648148148148149, "grad_norm": 8.466153144836426, "learning_rate": 0.0001809937888198758, "loss": 0.8465, "step": 169 }, { "epoch": 1.574074074074074, "grad_norm": 7.285000324249268, "learning_rate": 0.00018086956521739132, "loss": 0.5777, "step": 170 }, { "epoch": 1.5833333333333335, "grad_norm": 3.1247687339782715, "learning_rate": 0.00018074534161490683, "loss": 0.1439, "step": 171 }, { "epoch": 1.5925925925925926, "grad_norm": 7.981088161468506, "learning_rate": 0.00018062111801242238, "loss": 0.5934, "step": 172 }, { "epoch": 1.6018518518518519, "grad_norm": 5.762646198272705, "learning_rate": 0.0001804968944099379, "loss": 0.1859, "step": 173 }, { "epoch": 1.6111111111111112, "grad_norm": 3.394270181655884, "learning_rate": 0.0001803726708074534, "loss": 0.093, "step": 174 }, { "epoch": 1.6203703703703702, "grad_norm": 8.87936019897461, "learning_rate": 0.00018024844720496895, "loss": 1.2075, "step": 175 }, { "epoch": 1.6296296296296298, "grad_norm": 6.698455810546875, "learning_rate": 0.0001801242236024845, "loss": 0.2976, "step": 176 }, { "epoch": 1.6388888888888888, "grad_norm": 6.693928241729736, "learning_rate": 0.00018, "loss": 0.478, "step": 177 }, { "epoch": 1.6481481481481481, "grad_norm": 9.64937686920166, "learning_rate": 0.00017987577639751553, "loss": 1.1285, "step": 178 }, { "epoch": 1.6574074074074074, "grad_norm": 7.241469860076904, "learning_rate": 0.00017975155279503107, "loss": 0.5239, "step": 179 }, { "epoch": 1.6666666666666665, "grad_norm": 7.1578545570373535, "learning_rate": 0.0001796273291925466, "loss": 0.7933, "step": 180 }, { "epoch": 1.675925925925926, "grad_norm": 6.9873046875, "learning_rate": 0.0001795031055900621, "loss": 0.4719, "step": 181 }, { "epoch": 1.6851851851851851, "grad_norm": 6.540338516235352, "learning_rate": 0.00017937888198757765, "loss": 0.5329, "step": 182 }, { "epoch": 1.6944444444444444, "grad_norm": 9.244357109069824, "learning_rate": 0.00017925465838509316, "loss": 1.0934, "step": 183 }, { "epoch": 1.7037037037037037, "grad_norm": 8.517329216003418, "learning_rate": 0.0001791304347826087, "loss": 1.0569, "step": 184 }, { "epoch": 1.7129629629629628, "grad_norm": 7.156026840209961, "learning_rate": 0.00017900621118012425, "loss": 0.4594, "step": 185 }, { "epoch": 1.7222222222222223, "grad_norm": 9.52519702911377, "learning_rate": 0.00017888198757763977, "loss": 0.9583, "step": 186 }, { "epoch": 1.7314814814814814, "grad_norm": 7.694867134094238, "learning_rate": 0.00017875776397515528, "loss": 1.013, "step": 187 }, { "epoch": 1.7407407407407407, "grad_norm": 8.308947563171387, "learning_rate": 0.00017863354037267083, "loss": 0.8631, "step": 188 }, { "epoch": 1.75, "grad_norm": 4.8953022956848145, "learning_rate": 0.00017850931677018634, "loss": 0.1827, "step": 189 }, { "epoch": 1.7592592592592593, "grad_norm": 5.308168411254883, "learning_rate": 0.00017838509316770186, "loss": 0.3858, "step": 190 }, { "epoch": 1.7685185185185186, "grad_norm": 4.636107444763184, "learning_rate": 0.0001782608695652174, "loss": 0.2085, "step": 191 }, { "epoch": 1.7777777777777777, "grad_norm": 11.884739875793457, "learning_rate": 0.00017813664596273292, "loss": 0.6365, "step": 192 }, { "epoch": 1.7870370370370372, "grad_norm": 6.900138854980469, "learning_rate": 0.00017801242236024846, "loss": 0.3309, "step": 193 }, { "epoch": 1.7962962962962963, "grad_norm": 5.4100494384765625, "learning_rate": 0.00017788819875776398, "loss": 0.315, "step": 194 }, { "epoch": 1.8055555555555556, "grad_norm": 5.625341415405273, "learning_rate": 0.00017776397515527952, "loss": 0.2598, "step": 195 }, { "epoch": 1.8148148148148149, "grad_norm": 8.942736625671387, "learning_rate": 0.00017763975155279504, "loss": 1.0631, "step": 196 }, { "epoch": 1.824074074074074, "grad_norm": 7.598841190338135, "learning_rate": 0.00017751552795031056, "loss": 0.9039, "step": 197 }, { "epoch": 1.8333333333333335, "grad_norm": 8.496445655822754, "learning_rate": 0.0001773913043478261, "loss": 1.4777, "step": 198 }, { "epoch": 1.8425925925925926, "grad_norm": 7.111565113067627, "learning_rate": 0.00017726708074534161, "loss": 0.7539, "step": 199 }, { "epoch": 1.8518518518518519, "grad_norm": 8.147820472717285, "learning_rate": 0.00017714285714285713, "loss": 0.8148, "step": 200 }, { "epoch": 1.8611111111111112, "grad_norm": 9.675362586975098, "learning_rate": 0.00017701863354037267, "loss": 0.8903, "step": 201 }, { "epoch": 1.8703703703703702, "grad_norm": 3.753589630126953, "learning_rate": 0.00017689440993788822, "loss": 0.1457, "step": 202 }, { "epoch": 1.8796296296296298, "grad_norm": 4.512214183807373, "learning_rate": 0.00017677018633540373, "loss": 0.1948, "step": 203 }, { "epoch": 1.8888888888888888, "grad_norm": 4.206667423248291, "learning_rate": 0.00017664596273291928, "loss": 0.2643, "step": 204 }, { "epoch": 1.8981481481481481, "grad_norm": 8.376415252685547, "learning_rate": 0.0001765217391304348, "loss": 1.0019, "step": 205 }, { "epoch": 1.9074074074074074, "grad_norm": 9.681005477905273, "learning_rate": 0.0001763975155279503, "loss": 1.0884, "step": 206 }, { "epoch": 1.9166666666666665, "grad_norm": 4.311055660247803, "learning_rate": 0.00017627329192546585, "loss": 0.1727, "step": 207 }, { "epoch": 1.925925925925926, "grad_norm": 10.78451156616211, "learning_rate": 0.00017614906832298137, "loss": 0.3261, "step": 208 }, { "epoch": 1.9351851851851851, "grad_norm": 3.912935972213745, "learning_rate": 0.0001760248447204969, "loss": 0.1821, "step": 209 }, { "epoch": 1.9444444444444444, "grad_norm": 7.964453220367432, "learning_rate": 0.00017590062111801243, "loss": 0.6643, "step": 210 }, { "epoch": 1.9537037037037037, "grad_norm": 5.46901273727417, "learning_rate": 0.00017577639751552797, "loss": 0.3486, "step": 211 }, { "epoch": 1.9629629629629628, "grad_norm": 7.396164894104004, "learning_rate": 0.0001756521739130435, "loss": 0.7032, "step": 212 }, { "epoch": 1.9722222222222223, "grad_norm": 3.847592830657959, "learning_rate": 0.000175527950310559, "loss": 0.0872, "step": 213 }, { "epoch": 1.9814814814814814, "grad_norm": 7.062877178192139, "learning_rate": 0.00017540372670807455, "loss": 0.7447, "step": 214 }, { "epoch": 1.9907407407407407, "grad_norm": 5.658631801605225, "learning_rate": 0.00017527950310559007, "loss": 0.3057, "step": 215 }, { "epoch": 2.0, "grad_norm": 10.587858200073242, "learning_rate": 0.00017515527950310558, "loss": 0.8977, "step": 216 }, { "epoch": 2.009259259259259, "grad_norm": 9.634937286376953, "learning_rate": 0.00017503105590062113, "loss": 0.8675, "step": 217 }, { "epoch": 2.0185185185185186, "grad_norm": 3.696709156036377, "learning_rate": 0.00017490683229813664, "loss": 0.1366, "step": 218 }, { "epoch": 2.0277777777777777, "grad_norm": 5.684165000915527, "learning_rate": 0.00017478260869565219, "loss": 0.2566, "step": 219 }, { "epoch": 2.037037037037037, "grad_norm": 7.722327709197998, "learning_rate": 0.00017465838509316773, "loss": 0.285, "step": 220 }, { "epoch": 2.0462962962962963, "grad_norm": 4.990947246551514, "learning_rate": 0.00017453416149068325, "loss": 0.3135, "step": 221 }, { "epoch": 2.0555555555555554, "grad_norm": 3.8842992782592773, "learning_rate": 0.00017440993788819876, "loss": 0.0931, "step": 222 }, { "epoch": 2.064814814814815, "grad_norm": 3.498350143432617, "learning_rate": 0.0001742857142857143, "loss": 0.2054, "step": 223 }, { "epoch": 2.074074074074074, "grad_norm": 7.796590805053711, "learning_rate": 0.00017416149068322982, "loss": 0.7749, "step": 224 }, { "epoch": 2.0833333333333335, "grad_norm": 5.396276950836182, "learning_rate": 0.00017403726708074534, "loss": 0.3065, "step": 225 }, { "epoch": 2.0925925925925926, "grad_norm": 5.663061618804932, "learning_rate": 0.00017391304347826088, "loss": 0.2401, "step": 226 }, { "epoch": 2.1018518518518516, "grad_norm": 2.867725133895874, "learning_rate": 0.0001737888198757764, "loss": 0.0901, "step": 227 }, { "epoch": 2.111111111111111, "grad_norm": 4.327350616455078, "learning_rate": 0.00017366459627329191, "loss": 0.1796, "step": 228 }, { "epoch": 2.1203703703703702, "grad_norm": 8.092239379882812, "learning_rate": 0.00017354037267080748, "loss": 0.4362, "step": 229 }, { "epoch": 2.1296296296296298, "grad_norm": 8.841894149780273, "learning_rate": 0.000173416149068323, "loss": 0.4565, "step": 230 }, { "epoch": 2.138888888888889, "grad_norm": 3.5413870811462402, "learning_rate": 0.00017329192546583852, "loss": 0.1747, "step": 231 }, { "epoch": 2.148148148148148, "grad_norm": 11.68136978149414, "learning_rate": 0.00017316770186335406, "loss": 0.2018, "step": 232 }, { "epoch": 2.1574074074074074, "grad_norm": 10.268793106079102, "learning_rate": 0.00017304347826086958, "loss": 0.765, "step": 233 }, { "epoch": 2.1666666666666665, "grad_norm": 9.073164939880371, "learning_rate": 0.0001729192546583851, "loss": 0.4725, "step": 234 }, { "epoch": 2.175925925925926, "grad_norm": 3.81775164604187, "learning_rate": 0.00017279503105590064, "loss": 0.1851, "step": 235 }, { "epoch": 2.185185185185185, "grad_norm": 7.331940174102783, "learning_rate": 0.00017267080745341615, "loss": 0.4361, "step": 236 }, { "epoch": 2.1944444444444446, "grad_norm": 5.943619251251221, "learning_rate": 0.00017254658385093167, "loss": 0.3958, "step": 237 }, { "epoch": 2.2037037037037037, "grad_norm": 9.69388484954834, "learning_rate": 0.0001724223602484472, "loss": 0.2486, "step": 238 }, { "epoch": 2.212962962962963, "grad_norm": 9.289849281311035, "learning_rate": 0.00017229813664596276, "loss": 0.4469, "step": 239 }, { "epoch": 2.2222222222222223, "grad_norm": 6.857956886291504, "learning_rate": 0.00017217391304347827, "loss": 0.3926, "step": 240 }, { "epoch": 2.2314814814814814, "grad_norm": 10.366476058959961, "learning_rate": 0.0001720496894409938, "loss": 1.3493, "step": 241 }, { "epoch": 2.240740740740741, "grad_norm": 5.455699443817139, "learning_rate": 0.00017192546583850933, "loss": 0.2597, "step": 242 }, { "epoch": 2.25, "grad_norm": 7.438530921936035, "learning_rate": 0.00017180124223602485, "loss": 0.5654, "step": 243 }, { "epoch": 2.259259259259259, "grad_norm": 7.767147541046143, "learning_rate": 0.00017167701863354037, "loss": 0.5442, "step": 244 }, { "epoch": 2.2685185185185186, "grad_norm": 9.872796058654785, "learning_rate": 0.0001715527950310559, "loss": 0.7294, "step": 245 }, { "epoch": 2.2777777777777777, "grad_norm": 10.673378944396973, "learning_rate": 0.00017142857142857143, "loss": 1.1546, "step": 246 }, { "epoch": 2.287037037037037, "grad_norm": 21.295475006103516, "learning_rate": 0.00017130434782608697, "loss": 0.2201, "step": 247 }, { "epoch": 2.2962962962962963, "grad_norm": 8.360919952392578, "learning_rate": 0.0001711801242236025, "loss": 0.4692, "step": 248 }, { "epoch": 2.3055555555555554, "grad_norm": 7.903579235076904, "learning_rate": 0.00017105590062111803, "loss": 0.2641, "step": 249 }, { "epoch": 2.314814814814815, "grad_norm": 6.9065399169921875, "learning_rate": 0.00017093167701863354, "loss": 0.3394, "step": 250 }, { "epoch": 2.324074074074074, "grad_norm": 4.162428379058838, "learning_rate": 0.0001708074534161491, "loss": 0.1332, "step": 251 }, { "epoch": 2.3333333333333335, "grad_norm": 9.210295677185059, "learning_rate": 0.0001706832298136646, "loss": 0.6607, "step": 252 }, { "epoch": 2.3425925925925926, "grad_norm": 10.019932746887207, "learning_rate": 0.00017055900621118012, "loss": 0.5042, "step": 253 }, { "epoch": 2.351851851851852, "grad_norm": 4.602824687957764, "learning_rate": 0.00017043478260869566, "loss": 0.1234, "step": 254 }, { "epoch": 2.361111111111111, "grad_norm": 10.391927719116211, "learning_rate": 0.00017031055900621118, "loss": 0.6339, "step": 255 }, { "epoch": 2.3703703703703702, "grad_norm": 5.13262414932251, "learning_rate": 0.00017018633540372672, "loss": 0.2344, "step": 256 }, { "epoch": 2.3796296296296298, "grad_norm": 8.059638023376465, "learning_rate": 0.00017006211180124224, "loss": 0.5158, "step": 257 }, { "epoch": 2.388888888888889, "grad_norm": 34.22102355957031, "learning_rate": 0.00016993788819875778, "loss": 0.5373, "step": 258 }, { "epoch": 2.398148148148148, "grad_norm": 13.332140922546387, "learning_rate": 0.0001698136645962733, "loss": 0.7577, "step": 259 }, { "epoch": 2.4074074074074074, "grad_norm": 4.515472412109375, "learning_rate": 0.00016968944099378882, "loss": 0.1429, "step": 260 }, { "epoch": 2.4166666666666665, "grad_norm": 6.849452972412109, "learning_rate": 0.00016956521739130436, "loss": 0.1432, "step": 261 }, { "epoch": 2.425925925925926, "grad_norm": 5.058333873748779, "learning_rate": 0.00016944099378881988, "loss": 0.2357, "step": 262 }, { "epoch": 2.435185185185185, "grad_norm": 9.91899299621582, "learning_rate": 0.0001693167701863354, "loss": 0.3807, "step": 263 }, { "epoch": 2.4444444444444446, "grad_norm": 10.223733901977539, "learning_rate": 0.00016919254658385094, "loss": 0.9076, "step": 264 }, { "epoch": 2.4537037037037037, "grad_norm": 6.5338335037231445, "learning_rate": 0.00016906832298136648, "loss": 0.1497, "step": 265 }, { "epoch": 2.462962962962963, "grad_norm": 5.8402299880981445, "learning_rate": 0.000168944099378882, "loss": 0.5041, "step": 266 }, { "epoch": 2.4722222222222223, "grad_norm": 3.3992233276367188, "learning_rate": 0.00016881987577639754, "loss": 0.1547, "step": 267 }, { "epoch": 2.4814814814814814, "grad_norm": 7.095219612121582, "learning_rate": 0.00016869565217391306, "loss": 0.5186, "step": 268 }, { "epoch": 2.490740740740741, "grad_norm": 6.44674825668335, "learning_rate": 0.00016857142857142857, "loss": 0.1233, "step": 269 }, { "epoch": 2.5, "grad_norm": 4.471312522888184, "learning_rate": 0.00016844720496894412, "loss": 0.24, "step": 270 }, { "epoch": 2.5092592592592595, "grad_norm": 5.365081310272217, "learning_rate": 0.00016832298136645963, "loss": 0.3773, "step": 271 }, { "epoch": 2.5185185185185186, "grad_norm": 9.75694465637207, "learning_rate": 0.00016819875776397515, "loss": 0.5362, "step": 272 }, { "epoch": 2.5277777777777777, "grad_norm": 8.214263916015625, "learning_rate": 0.0001680745341614907, "loss": 0.4716, "step": 273 }, { "epoch": 2.537037037037037, "grad_norm": 10.777454376220703, "learning_rate": 0.00016795031055900624, "loss": 0.6536, "step": 274 }, { "epoch": 2.5462962962962963, "grad_norm": 8.224162101745605, "learning_rate": 0.00016782608695652175, "loss": 0.5819, "step": 275 }, { "epoch": 2.5555555555555554, "grad_norm": 10.33161449432373, "learning_rate": 0.00016770186335403727, "loss": 0.6333, "step": 276 }, { "epoch": 2.564814814814815, "grad_norm": 6.827465057373047, "learning_rate": 0.0001675776397515528, "loss": 0.2474, "step": 277 }, { "epoch": 2.574074074074074, "grad_norm": 7.896028995513916, "learning_rate": 0.00016745341614906833, "loss": 0.5632, "step": 278 }, { "epoch": 2.5833333333333335, "grad_norm": 4.075222492218018, "learning_rate": 0.00016732919254658384, "loss": 0.1224, "step": 279 }, { "epoch": 2.5925925925925926, "grad_norm": 4.907321929931641, "learning_rate": 0.0001672049689440994, "loss": 0.1726, "step": 280 }, { "epoch": 2.601851851851852, "grad_norm": 8.992182731628418, "learning_rate": 0.0001670807453416149, "loss": 0.6303, "step": 281 }, { "epoch": 2.611111111111111, "grad_norm": 7.870790958404541, "learning_rate": 0.00016695652173913042, "loss": 0.4438, "step": 282 }, { "epoch": 2.6203703703703702, "grad_norm": 5.292234420776367, "learning_rate": 0.000166832298136646, "loss": 0.1207, "step": 283 }, { "epoch": 2.6296296296296298, "grad_norm": 1.3497140407562256, "learning_rate": 0.0001667080745341615, "loss": 0.0298, "step": 284 }, { "epoch": 2.638888888888889, "grad_norm": 7.471672058105469, "learning_rate": 0.00016658385093167702, "loss": 0.5523, "step": 285 }, { "epoch": 2.648148148148148, "grad_norm": 1.7936245203018188, "learning_rate": 0.00016645962732919257, "loss": 0.1268, "step": 286 }, { "epoch": 2.6574074074074074, "grad_norm": 8.869959831237793, "learning_rate": 0.00016633540372670808, "loss": 0.7427, "step": 287 }, { "epoch": 2.6666666666666665, "grad_norm": 7.612359523773193, "learning_rate": 0.0001662111801242236, "loss": 0.5962, "step": 288 }, { "epoch": 2.675925925925926, "grad_norm": 3.3016366958618164, "learning_rate": 0.00016608695652173914, "loss": 0.124, "step": 289 }, { "epoch": 2.685185185185185, "grad_norm": 4.570164680480957, "learning_rate": 0.00016596273291925466, "loss": 0.2078, "step": 290 }, { "epoch": 2.6944444444444446, "grad_norm": 4.242143630981445, "learning_rate": 0.00016583850931677018, "loss": 0.1053, "step": 291 }, { "epoch": 2.7037037037037037, "grad_norm": 3.7191407680511475, "learning_rate": 0.00016571428571428575, "loss": 0.1582, "step": 292 }, { "epoch": 2.712962962962963, "grad_norm": 5.166412830352783, "learning_rate": 0.00016559006211180126, "loss": 0.2469, "step": 293 }, { "epoch": 2.7222222222222223, "grad_norm": 4.837871551513672, "learning_rate": 0.00016546583850931678, "loss": 0.2485, "step": 294 }, { "epoch": 2.7314814814814814, "grad_norm": 8.435837745666504, "learning_rate": 0.00016534161490683232, "loss": 0.5547, "step": 295 }, { "epoch": 2.7407407407407405, "grad_norm": 13.024179458618164, "learning_rate": 0.00016521739130434784, "loss": 0.7232, "step": 296 }, { "epoch": 2.75, "grad_norm": 9.188488960266113, "learning_rate": 0.00016509316770186336, "loss": 0.9133, "step": 297 }, { "epoch": 2.7592592592592595, "grad_norm": 6.474546432495117, "learning_rate": 0.0001649689440993789, "loss": 0.5251, "step": 298 }, { "epoch": 2.7685185185185186, "grad_norm": 8.005181312561035, "learning_rate": 0.00016484472049689441, "loss": 0.6408, "step": 299 }, { "epoch": 2.7777777777777777, "grad_norm": 7.383902549743652, "learning_rate": 0.00016472049689440993, "loss": 0.5634, "step": 300 }, { "epoch": 2.787037037037037, "grad_norm": 5.368563652038574, "learning_rate": 0.00016459627329192547, "loss": 0.264, "step": 301 }, { "epoch": 2.7962962962962963, "grad_norm": 3.9778330326080322, "learning_rate": 0.00016447204968944102, "loss": 0.1811, "step": 302 }, { "epoch": 2.8055555555555554, "grad_norm": 6.301558971405029, "learning_rate": 0.00016434782608695653, "loss": 0.3352, "step": 303 }, { "epoch": 2.814814814814815, "grad_norm": 11.769688606262207, "learning_rate": 0.00016422360248447205, "loss": 0.5181, "step": 304 }, { "epoch": 2.824074074074074, "grad_norm": 11.051255226135254, "learning_rate": 0.0001640993788819876, "loss": 0.9957, "step": 305 }, { "epoch": 2.8333333333333335, "grad_norm": 14.160080909729004, "learning_rate": 0.0001639751552795031, "loss": 0.3998, "step": 306 }, { "epoch": 2.8425925925925926, "grad_norm": 5.162957668304443, "learning_rate": 0.00016385093167701863, "loss": 0.2078, "step": 307 }, { "epoch": 2.851851851851852, "grad_norm": 4.017885208129883, "learning_rate": 0.00016372670807453417, "loss": 0.1804, "step": 308 }, { "epoch": 2.861111111111111, "grad_norm": 8.482934951782227, "learning_rate": 0.0001636024844720497, "loss": 0.6282, "step": 309 }, { "epoch": 2.8703703703703702, "grad_norm": 5.554912090301514, "learning_rate": 0.00016347826086956523, "loss": 0.2312, "step": 310 }, { "epoch": 2.8796296296296298, "grad_norm": 9.502184867858887, "learning_rate": 0.00016335403726708077, "loss": 0.6199, "step": 311 }, { "epoch": 2.888888888888889, "grad_norm": 8.019617080688477, "learning_rate": 0.0001632298136645963, "loss": 0.6601, "step": 312 }, { "epoch": 2.898148148148148, "grad_norm": 5.383131504058838, "learning_rate": 0.0001631055900621118, "loss": 0.3231, "step": 313 }, { "epoch": 2.9074074074074074, "grad_norm": 5.896549224853516, "learning_rate": 0.00016298136645962735, "loss": 0.3919, "step": 314 }, { "epoch": 2.9166666666666665, "grad_norm": 7.678750038146973, "learning_rate": 0.00016285714285714287, "loss": 0.4805, "step": 315 }, { "epoch": 2.925925925925926, "grad_norm": 9.610578536987305, "learning_rate": 0.00016273291925465838, "loss": 0.9259, "step": 316 }, { "epoch": 2.935185185185185, "grad_norm": 9.18185806274414, "learning_rate": 0.00016260869565217393, "loss": 1.0033, "step": 317 }, { "epoch": 2.9444444444444446, "grad_norm": 3.519683599472046, "learning_rate": 0.00016248447204968944, "loss": 0.1465, "step": 318 }, { "epoch": 2.9537037037037037, "grad_norm": 5.032393455505371, "learning_rate": 0.00016236024844720496, "loss": 0.2461, "step": 319 }, { "epoch": 2.962962962962963, "grad_norm": 7.032131671905518, "learning_rate": 0.0001622360248447205, "loss": 0.6732, "step": 320 }, { "epoch": 2.9722222222222223, "grad_norm": 3.6258678436279297, "learning_rate": 0.00016211180124223605, "loss": 0.1176, "step": 321 }, { "epoch": 2.9814814814814814, "grad_norm": 1.865622878074646, "learning_rate": 0.00016198757763975156, "loss": 0.0598, "step": 322 }, { "epoch": 2.9907407407407405, "grad_norm": 6.845046043395996, "learning_rate": 0.00016186335403726708, "loss": 0.4092, "step": 323 }, { "epoch": 3.0, "grad_norm": 5.940382480621338, "learning_rate": 0.00016173913043478262, "loss": 0.1281, "step": 324 }, { "epoch": 3.009259259259259, "grad_norm": 3.2817156314849854, "learning_rate": 0.00016161490683229814, "loss": 0.076, "step": 325 }, { "epoch": 3.0185185185185186, "grad_norm": 4.320040225982666, "learning_rate": 0.00016149068322981365, "loss": 0.164, "step": 326 }, { "epoch": 3.0277777777777777, "grad_norm": 6.783189296722412, "learning_rate": 0.0001613664596273292, "loss": 0.312, "step": 327 }, { "epoch": 3.037037037037037, "grad_norm": 9.855510711669922, "learning_rate": 0.00016124223602484471, "loss": 0.7179, "step": 328 }, { "epoch": 3.0462962962962963, "grad_norm": 11.480785369873047, "learning_rate": 0.00016111801242236026, "loss": 0.7069, "step": 329 }, { "epoch": 3.0555555555555554, "grad_norm": 6.975889205932617, "learning_rate": 0.0001609937888198758, "loss": 0.3169, "step": 330 }, { "epoch": 3.064814814814815, "grad_norm": 3.5186703205108643, "learning_rate": 0.00016086956521739132, "loss": 0.1183, "step": 331 }, { "epoch": 3.074074074074074, "grad_norm": 9.784034729003906, "learning_rate": 0.00016074534161490683, "loss": 0.601, "step": 332 }, { "epoch": 3.0833333333333335, "grad_norm": 2.755011796951294, "learning_rate": 0.00016062111801242238, "loss": 0.1034, "step": 333 }, { "epoch": 3.0925925925925926, "grad_norm": 3.478494167327881, "learning_rate": 0.0001604968944099379, "loss": 0.1482, "step": 334 }, { "epoch": 3.1018518518518516, "grad_norm": 5.404292583465576, "learning_rate": 0.0001603726708074534, "loss": 0.2123, "step": 335 }, { "epoch": 3.111111111111111, "grad_norm": 7.036523342132568, "learning_rate": 0.00016024844720496895, "loss": 0.4788, "step": 336 }, { "epoch": 3.1203703703703702, "grad_norm": 4.069665908813477, "learning_rate": 0.00016012422360248447, "loss": 0.2323, "step": 337 }, { "epoch": 3.1296296296296298, "grad_norm": 2.5665032863616943, "learning_rate": 0.00016, "loss": 0.0713, "step": 338 }, { "epoch": 3.138888888888889, "grad_norm": 8.244379997253418, "learning_rate": 0.00015987577639751553, "loss": 0.4594, "step": 339 }, { "epoch": 3.148148148148148, "grad_norm": 3.4311187267303467, "learning_rate": 0.00015975155279503107, "loss": 0.1345, "step": 340 }, { "epoch": 3.1574074074074074, "grad_norm": 6.329025745391846, "learning_rate": 0.0001596273291925466, "loss": 0.2205, "step": 341 }, { "epoch": 3.1666666666666665, "grad_norm": 8.342190742492676, "learning_rate": 0.0001595031055900621, "loss": 0.397, "step": 342 }, { "epoch": 3.175925925925926, "grad_norm": 3.3707950115203857, "learning_rate": 0.00015937888198757765, "loss": 0.1535, "step": 343 }, { "epoch": 3.185185185185185, "grad_norm": 2.9287197589874268, "learning_rate": 0.00015925465838509317, "loss": 0.114, "step": 344 }, { "epoch": 3.1944444444444446, "grad_norm": 1.9324220418930054, "learning_rate": 0.00015913043478260868, "loss": 0.0537, "step": 345 }, { "epoch": 3.2037037037037037, "grad_norm": 8.233223915100098, "learning_rate": 0.00015900621118012423, "loss": 0.2895, "step": 346 }, { "epoch": 3.212962962962963, "grad_norm": 8.415253639221191, "learning_rate": 0.00015888198757763977, "loss": 0.2266, "step": 347 }, { "epoch": 3.2222222222222223, "grad_norm": 7.05043888092041, "learning_rate": 0.00015875776397515528, "loss": 0.247, "step": 348 }, { "epoch": 3.2314814814814814, "grad_norm": 7.294909477233887, "learning_rate": 0.00015863354037267083, "loss": 0.3165, "step": 349 }, { "epoch": 3.240740740740741, "grad_norm": 4.877599716186523, "learning_rate": 0.00015850931677018634, "loss": 0.1593, "step": 350 }, { "epoch": 3.25, "grad_norm": 5.8372039794921875, "learning_rate": 0.00015838509316770186, "loss": 0.2842, "step": 351 }, { "epoch": 3.259259259259259, "grad_norm": 5.851783275604248, "learning_rate": 0.0001582608695652174, "loss": 0.2469, "step": 352 }, { "epoch": 3.2685185185185186, "grad_norm": 2.9726572036743164, "learning_rate": 0.00015813664596273292, "loss": 0.1536, "step": 353 }, { "epoch": 3.2777777777777777, "grad_norm": 8.282976150512695, "learning_rate": 0.00015801242236024844, "loss": 0.3309, "step": 354 }, { "epoch": 3.287037037037037, "grad_norm": 9.234270095825195, "learning_rate": 0.00015788819875776398, "loss": 0.4955, "step": 355 }, { "epoch": 3.2962962962962963, "grad_norm": 9.033778190612793, "learning_rate": 0.00015776397515527952, "loss": 0.3431, "step": 356 }, { "epoch": 3.3055555555555554, "grad_norm": 4.390086650848389, "learning_rate": 0.00015763975155279504, "loss": 0.2213, "step": 357 }, { "epoch": 3.314814814814815, "grad_norm": 7.078484058380127, "learning_rate": 0.00015751552795031058, "loss": 0.4093, "step": 358 }, { "epoch": 3.324074074074074, "grad_norm": 1.9521043300628662, "learning_rate": 0.0001573913043478261, "loss": 0.0501, "step": 359 }, { "epoch": 3.3333333333333335, "grad_norm": 6.5932159423828125, "learning_rate": 0.00015726708074534162, "loss": 0.2902, "step": 360 }, { "epoch": 3.3425925925925926, "grad_norm": 1.199304223060608, "learning_rate": 0.00015714285714285716, "loss": 0.0348, "step": 361 }, { "epoch": 3.351851851851852, "grad_norm": 7.945189476013184, "learning_rate": 0.00015701863354037268, "loss": 0.4226, "step": 362 }, { "epoch": 3.361111111111111, "grad_norm": 6.304150104522705, "learning_rate": 0.0001568944099378882, "loss": 0.3419, "step": 363 }, { "epoch": 3.3703703703703702, "grad_norm": 0.575678825378418, "learning_rate": 0.00015677018633540374, "loss": 0.0143, "step": 364 }, { "epoch": 3.3796296296296298, "grad_norm": 55.86056137084961, "learning_rate": 0.00015664596273291928, "loss": 1.0001, "step": 365 }, { "epoch": 3.388888888888889, "grad_norm": 5.551873683929443, "learning_rate": 0.0001565217391304348, "loss": 0.316, "step": 366 }, { "epoch": 3.398148148148148, "grad_norm": 4.200618743896484, "learning_rate": 0.0001563975155279503, "loss": 0.1263, "step": 367 }, { "epoch": 3.4074074074074074, "grad_norm": 9.562318801879883, "learning_rate": 0.00015627329192546586, "loss": 0.8528, "step": 368 }, { "epoch": 3.4166666666666665, "grad_norm": 3.7029190063476562, "learning_rate": 0.00015614906832298137, "loss": 0.1608, "step": 369 }, { "epoch": 3.425925925925926, "grad_norm": 5.9052252769470215, "learning_rate": 0.0001560248447204969, "loss": 0.2919, "step": 370 }, { "epoch": 3.435185185185185, "grad_norm": 5.278835773468018, "learning_rate": 0.00015590062111801243, "loss": 0.1226, "step": 371 }, { "epoch": 3.4444444444444446, "grad_norm": 6.618929862976074, "learning_rate": 0.00015577639751552795, "loss": 0.3602, "step": 372 }, { "epoch": 3.4537037037037037, "grad_norm": 3.45408296585083, "learning_rate": 0.00015565217391304346, "loss": 0.0969, "step": 373 }, { "epoch": 3.462962962962963, "grad_norm": 4.552202224731445, "learning_rate": 0.00015552795031055904, "loss": 0.1777, "step": 374 }, { "epoch": 3.4722222222222223, "grad_norm": 2.348261833190918, "learning_rate": 0.00015540372670807455, "loss": 0.0569, "step": 375 }, { "epoch": 3.4814814814814814, "grad_norm": 1.7154549360275269, "learning_rate": 0.00015527950310559007, "loss": 0.0647, "step": 376 }, { "epoch": 3.490740740740741, "grad_norm": 7.577757358551025, "learning_rate": 0.0001551552795031056, "loss": 0.4125, "step": 377 }, { "epoch": 3.5, "grad_norm": 6.152462005615234, "learning_rate": 0.00015503105590062113, "loss": 0.4456, "step": 378 }, { "epoch": 3.5092592592592595, "grad_norm": 6.738099575042725, "learning_rate": 0.00015490683229813664, "loss": 0.4373, "step": 379 }, { "epoch": 3.5185185185185186, "grad_norm": 6.085219860076904, "learning_rate": 0.0001547826086956522, "loss": 0.2929, "step": 380 }, { "epoch": 3.5277777777777777, "grad_norm": 3.194652557373047, "learning_rate": 0.0001546583850931677, "loss": 0.2306, "step": 381 }, { "epoch": 3.537037037037037, "grad_norm": 7.23938512802124, "learning_rate": 0.00015453416149068322, "loss": 0.5282, "step": 382 }, { "epoch": 3.5462962962962963, "grad_norm": 5.027129173278809, "learning_rate": 0.00015440993788819876, "loss": 0.2126, "step": 383 }, { "epoch": 3.5555555555555554, "grad_norm": 11.30113410949707, "learning_rate": 0.0001542857142857143, "loss": 0.4177, "step": 384 }, { "epoch": 3.564814814814815, "grad_norm": 9.71517276763916, "learning_rate": 0.00015416149068322982, "loss": 0.4661, "step": 385 }, { "epoch": 3.574074074074074, "grad_norm": 5.263823509216309, "learning_rate": 0.00015403726708074534, "loss": 0.0915, "step": 386 }, { "epoch": 3.5833333333333335, "grad_norm": 8.799884796142578, "learning_rate": 0.00015391304347826088, "loss": 0.2908, "step": 387 }, { "epoch": 3.5925925925925926, "grad_norm": 5.173797130584717, "learning_rate": 0.0001537888198757764, "loss": 0.2236, "step": 388 }, { "epoch": 3.601851851851852, "grad_norm": 4.034602165222168, "learning_rate": 0.00015366459627329192, "loss": 0.2318, "step": 389 }, { "epoch": 3.611111111111111, "grad_norm": 6.178292751312256, "learning_rate": 0.00015354037267080746, "loss": 0.4033, "step": 390 }, { "epoch": 3.6203703703703702, "grad_norm": 6.233795166015625, "learning_rate": 0.00015341614906832298, "loss": 0.3785, "step": 391 }, { "epoch": 3.6296296296296298, "grad_norm": 9.079377174377441, "learning_rate": 0.00015329192546583852, "loss": 0.2154, "step": 392 }, { "epoch": 3.638888888888889, "grad_norm": 7.057285785675049, "learning_rate": 0.00015316770186335406, "loss": 0.4248, "step": 393 }, { "epoch": 3.648148148148148, "grad_norm": 1.6093757152557373, "learning_rate": 0.00015304347826086958, "loss": 0.0697, "step": 394 }, { "epoch": 3.6574074074074074, "grad_norm": 5.713193416595459, "learning_rate": 0.0001529192546583851, "loss": 0.304, "step": 395 }, { "epoch": 3.6666666666666665, "grad_norm": 4.769229888916016, "learning_rate": 0.00015279503105590064, "loss": 0.1802, "step": 396 }, { "epoch": 3.675925925925926, "grad_norm": 11.877578735351562, "learning_rate": 0.00015267080745341616, "loss": 0.952, "step": 397 }, { "epoch": 3.685185185185185, "grad_norm": 3.5577473640441895, "learning_rate": 0.00015254658385093167, "loss": 0.1702, "step": 398 }, { "epoch": 3.6944444444444446, "grad_norm": 4.835704803466797, "learning_rate": 0.00015242236024844721, "loss": 0.1728, "step": 399 }, { "epoch": 3.7037037037037037, "grad_norm": 8.712724685668945, "learning_rate": 0.00015229813664596273, "loss": 0.5967, "step": 400 }, { "epoch": 3.712962962962963, "grad_norm": 3.5652997493743896, "learning_rate": 0.00015217391304347827, "loss": 0.1029, "step": 401 }, { "epoch": 3.7222222222222223, "grad_norm": 7.978386402130127, "learning_rate": 0.0001520496894409938, "loss": 0.456, "step": 402 }, { "epoch": 3.7314814814814814, "grad_norm": 8.434885025024414, "learning_rate": 0.00015192546583850933, "loss": 0.4298, "step": 403 }, { "epoch": 3.7407407407407405, "grad_norm": 5.78842306137085, "learning_rate": 0.00015180124223602485, "loss": 0.3563, "step": 404 }, { "epoch": 3.75, "grad_norm": 5.0681047439575195, "learning_rate": 0.00015167701863354037, "loss": 0.2031, "step": 405 }, { "epoch": 3.7592592592592595, "grad_norm": 4.695619583129883, "learning_rate": 0.0001515527950310559, "loss": 0.1284, "step": 406 }, { "epoch": 3.7685185185185186, "grad_norm": 4.437154769897461, "learning_rate": 0.00015142857142857143, "loss": 0.2112, "step": 407 }, { "epoch": 3.7777777777777777, "grad_norm": 3.3746249675750732, "learning_rate": 0.00015130434782608694, "loss": 0.1581, "step": 408 }, { "epoch": 3.787037037037037, "grad_norm": 7.2292585372924805, "learning_rate": 0.0001511801242236025, "loss": 0.4854, "step": 409 }, { "epoch": 3.7962962962962963, "grad_norm": 7.788264751434326, "learning_rate": 0.00015105590062111803, "loss": 0.4422, "step": 410 }, { "epoch": 3.8055555555555554, "grad_norm": 7.971909999847412, "learning_rate": 0.00015093167701863355, "loss": 0.5586, "step": 411 }, { "epoch": 3.814814814814815, "grad_norm": 7.098501682281494, "learning_rate": 0.0001508074534161491, "loss": 0.2834, "step": 412 }, { "epoch": 3.824074074074074, "grad_norm": 3.429027795791626, "learning_rate": 0.0001506832298136646, "loss": 0.1501, "step": 413 }, { "epoch": 3.8333333333333335, "grad_norm": 8.535176277160645, "learning_rate": 0.00015055900621118012, "loss": 0.3973, "step": 414 }, { "epoch": 3.8425925925925926, "grad_norm": 7.709643363952637, "learning_rate": 0.00015043478260869567, "loss": 0.6272, "step": 415 }, { "epoch": 3.851851851851852, "grad_norm": 5.064123630523682, "learning_rate": 0.00015031055900621118, "loss": 0.2087, "step": 416 }, { "epoch": 3.861111111111111, "grad_norm": 2.5206053256988525, "learning_rate": 0.0001501863354037267, "loss": 0.095, "step": 417 }, { "epoch": 3.8703703703703702, "grad_norm": 8.01754093170166, "learning_rate": 0.00015006211180124224, "loss": 0.2509, "step": 418 }, { "epoch": 3.8796296296296298, "grad_norm": 3.0863702297210693, "learning_rate": 0.00014993788819875779, "loss": 0.1075, "step": 419 }, { "epoch": 3.888888888888889, "grad_norm": 4.70889139175415, "learning_rate": 0.0001498136645962733, "loss": 0.1924, "step": 420 }, { "epoch": 3.898148148148148, "grad_norm": 7.025390148162842, "learning_rate": 0.00014968944099378885, "loss": 0.359, "step": 421 }, { "epoch": 3.9074074074074074, "grad_norm": 5.652724266052246, "learning_rate": 0.00014956521739130436, "loss": 0.307, "step": 422 }, { "epoch": 3.9166666666666665, "grad_norm": 6.922481536865234, "learning_rate": 0.00014944099378881988, "loss": 0.2555, "step": 423 }, { "epoch": 3.925925925925926, "grad_norm": 3.9260189533233643, "learning_rate": 0.00014931677018633542, "loss": 0.2264, "step": 424 }, { "epoch": 3.935185185185185, "grad_norm": 7.77918004989624, "learning_rate": 0.00014919254658385094, "loss": 0.5299, "step": 425 }, { "epoch": 3.9444444444444446, "grad_norm": 8.371830940246582, "learning_rate": 0.00014906832298136645, "loss": 0.5051, "step": 426 }, { "epoch": 3.9537037037037037, "grad_norm": 4.843681812286377, "learning_rate": 0.000148944099378882, "loss": 0.23, "step": 427 }, { "epoch": 3.962962962962963, "grad_norm": 5.792315483093262, "learning_rate": 0.00014881987577639754, "loss": 0.2242, "step": 428 }, { "epoch": 3.9722222222222223, "grad_norm": 3.9335224628448486, "learning_rate": 0.00014869565217391306, "loss": 0.1533, "step": 429 }, { "epoch": 3.9814814814814814, "grad_norm": 12.831199645996094, "learning_rate": 0.00014857142857142857, "loss": 0.7754, "step": 430 }, { "epoch": 3.9907407407407405, "grad_norm": 3.6635866165161133, "learning_rate": 0.00014844720496894412, "loss": 0.1512, "step": 431 }, { "epoch": 4.0, "grad_norm": 7.196122646331787, "learning_rate": 0.00014832298136645963, "loss": 0.3633, "step": 432 }, { "epoch": 4.0092592592592595, "grad_norm": 3.1862425804138184, "learning_rate": 0.00014819875776397515, "loss": 0.1247, "step": 433 }, { "epoch": 4.018518518518518, "grad_norm": 8.481317520141602, "learning_rate": 0.0001480745341614907, "loss": 0.4654, "step": 434 }, { "epoch": 4.027777777777778, "grad_norm": 5.601878643035889, "learning_rate": 0.0001479503105590062, "loss": 0.3271, "step": 435 }, { "epoch": 4.037037037037037, "grad_norm": 2.543224573135376, "learning_rate": 0.00014782608695652173, "loss": 0.0809, "step": 436 }, { "epoch": 4.046296296296297, "grad_norm": 4.2349395751953125, "learning_rate": 0.0001477018633540373, "loss": 0.1481, "step": 437 }, { "epoch": 4.055555555555555, "grad_norm": 1.989195704460144, "learning_rate": 0.0001475776397515528, "loss": 0.0472, "step": 438 }, { "epoch": 4.064814814814815, "grad_norm": 3.653602123260498, "learning_rate": 0.00014745341614906833, "loss": 0.1803, "step": 439 }, { "epoch": 4.074074074074074, "grad_norm": 5.280818939208984, "learning_rate": 0.00014732919254658387, "loss": 0.1864, "step": 440 }, { "epoch": 4.083333333333333, "grad_norm": 2.3477160930633545, "learning_rate": 0.0001472049689440994, "loss": 0.0735, "step": 441 }, { "epoch": 4.092592592592593, "grad_norm": 7.4617414474487305, "learning_rate": 0.0001470807453416149, "loss": 0.281, "step": 442 }, { "epoch": 4.101851851851852, "grad_norm": 4.8935546875, "learning_rate": 0.00014695652173913045, "loss": 0.257, "step": 443 }, { "epoch": 4.111111111111111, "grad_norm": 7.706578731536865, "learning_rate": 0.00014683229813664597, "loss": 0.3459, "step": 444 }, { "epoch": 4.12037037037037, "grad_norm": 2.809396266937256, "learning_rate": 0.00014670807453416148, "loss": 0.0632, "step": 445 }, { "epoch": 4.12962962962963, "grad_norm": 6.64403772354126, "learning_rate": 0.00014658385093167703, "loss": 0.3485, "step": 446 }, { "epoch": 4.138888888888889, "grad_norm": 5.006530284881592, "learning_rate": 0.00014645962732919257, "loss": 0.2197, "step": 447 }, { "epoch": 4.148148148148148, "grad_norm": 2.446791410446167, "learning_rate": 0.00014633540372670808, "loss": 0.0841, "step": 448 }, { "epoch": 4.157407407407407, "grad_norm": 2.2801015377044678, "learning_rate": 0.0001462111801242236, "loss": 0.0983, "step": 449 }, { "epoch": 4.166666666666667, "grad_norm": 3.368048667907715, "learning_rate": 0.00014608695652173914, "loss": 0.1286, "step": 450 }, { "epoch": 4.175925925925926, "grad_norm": 6.958519458770752, "learning_rate": 0.00014596273291925466, "loss": 0.2685, "step": 451 }, { "epoch": 4.185185185185185, "grad_norm": 2.6293435096740723, "learning_rate": 0.00014583850931677018, "loss": 0.118, "step": 452 }, { "epoch": 4.194444444444445, "grad_norm": 2.426443338394165, "learning_rate": 0.00014571428571428572, "loss": 0.0909, "step": 453 }, { "epoch": 4.203703703703703, "grad_norm": 3.9959075450897217, "learning_rate": 0.00014559006211180124, "loss": 0.1808, "step": 454 }, { "epoch": 4.212962962962963, "grad_norm": 4.146143436431885, "learning_rate": 0.00014546583850931678, "loss": 0.2056, "step": 455 }, { "epoch": 4.222222222222222, "grad_norm": 6.517556190490723, "learning_rate": 0.00014534161490683232, "loss": 0.2656, "step": 456 }, { "epoch": 4.231481481481482, "grad_norm": 4.0375237464904785, "learning_rate": 0.00014521739130434784, "loss": 0.2048, "step": 457 }, { "epoch": 4.2407407407407405, "grad_norm": 3.206602096557617, "learning_rate": 0.00014509316770186336, "loss": 0.1343, "step": 458 }, { "epoch": 4.25, "grad_norm": 4.9624786376953125, "learning_rate": 0.0001449689440993789, "loss": 0.2606, "step": 459 }, { "epoch": 4.2592592592592595, "grad_norm": 4.6561455726623535, "learning_rate": 0.00014484472049689442, "loss": 0.1632, "step": 460 }, { "epoch": 4.268518518518518, "grad_norm": 3.209395408630371, "learning_rate": 0.00014472049689440993, "loss": 0.1272, "step": 461 }, { "epoch": 4.277777777777778, "grad_norm": 5.021265983581543, "learning_rate": 0.00014459627329192548, "loss": 0.2286, "step": 462 }, { "epoch": 4.287037037037037, "grad_norm": 7.752100467681885, "learning_rate": 0.000144472049689441, "loss": 0.4169, "step": 463 }, { "epoch": 4.296296296296296, "grad_norm": 7.6476850509643555, "learning_rate": 0.00014434782608695654, "loss": 0.276, "step": 464 }, { "epoch": 4.305555555555555, "grad_norm": 4.68563175201416, "learning_rate": 0.00014422360248447205, "loss": 0.113, "step": 465 }, { "epoch": 4.314814814814815, "grad_norm": 2.981081485748291, "learning_rate": 0.0001440993788819876, "loss": 0.1054, "step": 466 }, { "epoch": 4.324074074074074, "grad_norm": 4.140884876251221, "learning_rate": 0.0001439751552795031, "loss": 0.1545, "step": 467 }, { "epoch": 4.333333333333333, "grad_norm": 4.200994968414307, "learning_rate": 0.00014385093167701863, "loss": 0.1325, "step": 468 }, { "epoch": 4.342592592592593, "grad_norm": 4.808220386505127, "learning_rate": 0.00014372670807453417, "loss": 0.2498, "step": 469 }, { "epoch": 4.351851851851852, "grad_norm": 3.549621820449829, "learning_rate": 0.0001436024844720497, "loss": 0.1205, "step": 470 }, { "epoch": 4.361111111111111, "grad_norm": 4.576113224029541, "learning_rate": 0.0001434782608695652, "loss": 0.1623, "step": 471 }, { "epoch": 4.37037037037037, "grad_norm": 9.293719291687012, "learning_rate": 0.00014335403726708075, "loss": 0.5254, "step": 472 }, { "epoch": 4.37962962962963, "grad_norm": 4.397647380828857, "learning_rate": 0.0001432298136645963, "loss": 0.1521, "step": 473 }, { "epoch": 4.388888888888889, "grad_norm": 3.319638252258301, "learning_rate": 0.0001431055900621118, "loss": 0.1145, "step": 474 }, { "epoch": 4.398148148148148, "grad_norm": 6.22605562210083, "learning_rate": 0.00014298136645962735, "loss": 0.2767, "step": 475 }, { "epoch": 4.407407407407407, "grad_norm": 8.628198623657227, "learning_rate": 0.00014285714285714287, "loss": 0.502, "step": 476 }, { "epoch": 4.416666666666667, "grad_norm": 5.40584659576416, "learning_rate": 0.00014273291925465838, "loss": 0.2086, "step": 477 }, { "epoch": 4.425925925925926, "grad_norm": 7.446430683135986, "learning_rate": 0.00014260869565217393, "loss": 0.3835, "step": 478 }, { "epoch": 4.435185185185185, "grad_norm": 3.44232439994812, "learning_rate": 0.00014248447204968944, "loss": 0.1593, "step": 479 }, { "epoch": 4.444444444444445, "grad_norm": 5.173421382904053, "learning_rate": 0.00014236024844720496, "loss": 0.1365, "step": 480 }, { "epoch": 4.453703703703704, "grad_norm": 4.879393100738525, "learning_rate": 0.0001422360248447205, "loss": 0.1906, "step": 481 }, { "epoch": 4.462962962962963, "grad_norm": 4.784718036651611, "learning_rate": 0.00014211180124223605, "loss": 0.2266, "step": 482 }, { "epoch": 4.472222222222222, "grad_norm": 2.3085122108459473, "learning_rate": 0.00014198757763975156, "loss": 0.0864, "step": 483 }, { "epoch": 4.481481481481482, "grad_norm": 6.118433475494385, "learning_rate": 0.00014186335403726708, "loss": 0.2416, "step": 484 }, { "epoch": 4.4907407407407405, "grad_norm": 4.493291854858398, "learning_rate": 0.00014173913043478262, "loss": 0.174, "step": 485 }, { "epoch": 4.5, "grad_norm": 3.1890506744384766, "learning_rate": 0.00014161490683229814, "loss": 0.1136, "step": 486 }, { "epoch": 4.5092592592592595, "grad_norm": 4.983994007110596, "learning_rate": 0.00014149068322981368, "loss": 0.2204, "step": 487 }, { "epoch": 4.518518518518518, "grad_norm": 4.625283718109131, "learning_rate": 0.0001413664596273292, "loss": 0.2092, "step": 488 }, { "epoch": 4.527777777777778, "grad_norm": 5.007232666015625, "learning_rate": 0.00014124223602484472, "loss": 0.1789, "step": 489 }, { "epoch": 4.537037037037037, "grad_norm": 3.370143413543701, "learning_rate": 0.00014111801242236026, "loss": 0.1262, "step": 490 }, { "epoch": 4.546296296296296, "grad_norm": 5.218165874481201, "learning_rate": 0.0001409937888198758, "loss": 0.1389, "step": 491 }, { "epoch": 4.555555555555555, "grad_norm": 7.572610378265381, "learning_rate": 0.00014086956521739132, "loss": 0.4608, "step": 492 }, { "epoch": 4.564814814814815, "grad_norm": 8.756231307983398, "learning_rate": 0.00014074534161490684, "loss": 0.4648, "step": 493 }, { "epoch": 4.574074074074074, "grad_norm": 3.8681464195251465, "learning_rate": 0.00014062111801242238, "loss": 0.1241, "step": 494 }, { "epoch": 4.583333333333333, "grad_norm": 2.857240915298462, "learning_rate": 0.0001404968944099379, "loss": 0.1154, "step": 495 }, { "epoch": 4.592592592592593, "grad_norm": 3.749955654144287, "learning_rate": 0.0001403726708074534, "loss": 0.1234, "step": 496 }, { "epoch": 4.601851851851852, "grad_norm": 7.389172077178955, "learning_rate": 0.00014024844720496896, "loss": 0.341, "step": 497 }, { "epoch": 4.611111111111111, "grad_norm": 2.2326736450195312, "learning_rate": 0.00014012422360248447, "loss": 0.0911, "step": 498 }, { "epoch": 4.62037037037037, "grad_norm": 5.606997489929199, "learning_rate": 0.00014, "loss": 0.2355, "step": 499 }, { "epoch": 4.62962962962963, "grad_norm": 4.593019008636475, "learning_rate": 0.00013987577639751556, "loss": 0.2544, "step": 500 }, { "epoch": 4.638888888888889, "grad_norm": 4.68584680557251, "learning_rate": 0.00013975155279503107, "loss": 0.1838, "step": 501 }, { "epoch": 4.648148148148148, "grad_norm": 5.606670379638672, "learning_rate": 0.0001396273291925466, "loss": 0.2747, "step": 502 }, { "epoch": 4.657407407407407, "grad_norm": 6.980198860168457, "learning_rate": 0.00013950310559006213, "loss": 0.3134, "step": 503 }, { "epoch": 4.666666666666667, "grad_norm": 5.77571439743042, "learning_rate": 0.00013937888198757765, "loss": 0.3051, "step": 504 }, { "epoch": 4.675925925925926, "grad_norm": 8.558889389038086, "learning_rate": 0.00013925465838509317, "loss": 0.5136, "step": 505 }, { "epoch": 4.685185185185185, "grad_norm": 6.231207847595215, "learning_rate": 0.0001391304347826087, "loss": 0.2484, "step": 506 }, { "epoch": 4.694444444444445, "grad_norm": 4.745160102844238, "learning_rate": 0.00013900621118012423, "loss": 0.2174, "step": 507 }, { "epoch": 4.703703703703704, "grad_norm": 6.144798278808594, "learning_rate": 0.00013888198757763974, "loss": 0.2192, "step": 508 }, { "epoch": 4.712962962962963, "grad_norm": 6.802698612213135, "learning_rate": 0.0001387577639751553, "loss": 0.2729, "step": 509 }, { "epoch": 4.722222222222222, "grad_norm": 5.089367389678955, "learning_rate": 0.00013863354037267083, "loss": 0.1986, "step": 510 }, { "epoch": 4.731481481481482, "grad_norm": 2.46826434135437, "learning_rate": 0.00013850931677018635, "loss": 0.0684, "step": 511 }, { "epoch": 4.7407407407407405, "grad_norm": 3.246137857437134, "learning_rate": 0.00013838509316770186, "loss": 0.106, "step": 512 }, { "epoch": 4.75, "grad_norm": 4.2868266105651855, "learning_rate": 0.0001382608695652174, "loss": 0.1894, "step": 513 }, { "epoch": 4.7592592592592595, "grad_norm": 6.733697891235352, "learning_rate": 0.00013813664596273292, "loss": 0.2563, "step": 514 }, { "epoch": 4.768518518518518, "grad_norm": 5.327560901641846, "learning_rate": 0.00013801242236024844, "loss": 0.2412, "step": 515 }, { "epoch": 4.777777777777778, "grad_norm": 2.9705471992492676, "learning_rate": 0.00013788819875776398, "loss": 0.0845, "step": 516 }, { "epoch": 4.787037037037037, "grad_norm": 2.2280571460723877, "learning_rate": 0.0001377639751552795, "loss": 0.0572, "step": 517 }, { "epoch": 4.796296296296296, "grad_norm": 13.944962501525879, "learning_rate": 0.00013763975155279504, "loss": 0.4604, "step": 518 }, { "epoch": 4.805555555555555, "grad_norm": 5.684059143066406, "learning_rate": 0.00013751552795031059, "loss": 0.184, "step": 519 }, { "epoch": 4.814814814814815, "grad_norm": 3.9679698944091797, "learning_rate": 0.0001373913043478261, "loss": 0.1452, "step": 520 }, { "epoch": 4.824074074074074, "grad_norm": 1.851914405822754, "learning_rate": 0.00013726708074534162, "loss": 0.065, "step": 521 }, { "epoch": 4.833333333333333, "grad_norm": 6.1766133308410645, "learning_rate": 0.00013714285714285716, "loss": 0.2719, "step": 522 }, { "epoch": 4.842592592592593, "grad_norm": 6.558730602264404, "learning_rate": 0.00013701863354037268, "loss": 0.3054, "step": 523 }, { "epoch": 4.851851851851852, "grad_norm": 2.0593979358673096, "learning_rate": 0.0001368944099378882, "loss": 0.0425, "step": 524 }, { "epoch": 4.861111111111111, "grad_norm": 4.957576274871826, "learning_rate": 0.00013677018633540374, "loss": 0.2348, "step": 525 }, { "epoch": 4.87037037037037, "grad_norm": 5.931738376617432, "learning_rate": 0.00013664596273291925, "loss": 0.2931, "step": 526 }, { "epoch": 4.87962962962963, "grad_norm": 8.989755630493164, "learning_rate": 0.00013652173913043477, "loss": 0.513, "step": 527 }, { "epoch": 4.888888888888889, "grad_norm": 6.920965671539307, "learning_rate": 0.00013639751552795031, "loss": 0.3348, "step": 528 }, { "epoch": 4.898148148148148, "grad_norm": 2.210482358932495, "learning_rate": 0.00013627329192546586, "loss": 0.1061, "step": 529 }, { "epoch": 4.907407407407407, "grad_norm": 5.41426420211792, "learning_rate": 0.00013614906832298137, "loss": 0.1837, "step": 530 }, { "epoch": 4.916666666666667, "grad_norm": 8.46068000793457, "learning_rate": 0.0001360248447204969, "loss": 0.1896, "step": 531 }, { "epoch": 4.925925925925926, "grad_norm": 5.2427144050598145, "learning_rate": 0.00013590062111801243, "loss": 0.2031, "step": 532 }, { "epoch": 4.935185185185185, "grad_norm": 5.652583599090576, "learning_rate": 0.00013577639751552795, "loss": 0.209, "step": 533 }, { "epoch": 4.944444444444445, "grad_norm": 3.267430543899536, "learning_rate": 0.00013565217391304347, "loss": 0.1083, "step": 534 }, { "epoch": 4.953703703703704, "grad_norm": 7.333652973175049, "learning_rate": 0.000135527950310559, "loss": 0.191, "step": 535 }, { "epoch": 4.962962962962963, "grad_norm": 2.8522582054138184, "learning_rate": 0.00013540372670807453, "loss": 0.0938, "step": 536 }, { "epoch": 4.972222222222222, "grad_norm": 4.617889881134033, "learning_rate": 0.00013527950310559007, "loss": 0.1061, "step": 537 }, { "epoch": 4.981481481481482, "grad_norm": 7.658900737762451, "learning_rate": 0.0001351552795031056, "loss": 0.4003, "step": 538 }, { "epoch": 4.9907407407407405, "grad_norm": 7.5713653564453125, "learning_rate": 0.00013503105590062113, "loss": 0.3886, "step": 539 }, { "epoch": 5.0, "grad_norm": 9.257078170776367, "learning_rate": 0.00013490683229813665, "loss": 0.6839, "step": 540 }, { "epoch": 5.0092592592592595, "grad_norm": 6.371257305145264, "learning_rate": 0.0001347826086956522, "loss": 0.314, "step": 541 }, { "epoch": 5.018518518518518, "grad_norm": 2.907931327819824, "learning_rate": 0.0001346583850931677, "loss": 0.0629, "step": 542 }, { "epoch": 5.027777777777778, "grad_norm": 4.068233966827393, "learning_rate": 0.00013453416149068322, "loss": 0.1179, "step": 543 }, { "epoch": 5.037037037037037, "grad_norm": 2.4700989723205566, "learning_rate": 0.00013440993788819877, "loss": 0.0794, "step": 544 }, { "epoch": 5.046296296296297, "grad_norm": 3.4381418228149414, "learning_rate": 0.00013428571428571428, "loss": 0.1431, "step": 545 }, { "epoch": 5.055555555555555, "grad_norm": 0.883247971534729, "learning_rate": 0.00013416149068322983, "loss": 0.0141, "step": 546 }, { "epoch": 5.064814814814815, "grad_norm": 4.4683518409729, "learning_rate": 0.00013403726708074534, "loss": 0.0748, "step": 547 }, { "epoch": 5.074074074074074, "grad_norm": 5.847491264343262, "learning_rate": 0.00013391304347826088, "loss": 0.2232, "step": 548 }, { "epoch": 5.083333333333333, "grad_norm": 1.6964125633239746, "learning_rate": 0.0001337888198757764, "loss": 0.0555, "step": 549 }, { "epoch": 5.092592592592593, "grad_norm": 2.773496150970459, "learning_rate": 0.00013366459627329192, "loss": 0.0791, "step": 550 }, { "epoch": 5.101851851851852, "grad_norm": 4.492348670959473, "learning_rate": 0.00013354037267080746, "loss": 0.1072, "step": 551 }, { "epoch": 5.111111111111111, "grad_norm": 2.172301769256592, "learning_rate": 0.00013341614906832298, "loss": 0.0527, "step": 552 }, { "epoch": 5.12037037037037, "grad_norm": 2.18371319770813, "learning_rate": 0.00013329192546583852, "loss": 0.0565, "step": 553 }, { "epoch": 5.12962962962963, "grad_norm": 5.9107747077941895, "learning_rate": 0.00013316770186335404, "loss": 0.1583, "step": 554 }, { "epoch": 5.138888888888889, "grad_norm": 4.741661548614502, "learning_rate": 0.00013304347826086958, "loss": 0.1513, "step": 555 }, { "epoch": 5.148148148148148, "grad_norm": 4.259805202484131, "learning_rate": 0.0001329192546583851, "loss": 0.1908, "step": 556 }, { "epoch": 5.157407407407407, "grad_norm": 6.191773414611816, "learning_rate": 0.00013279503105590064, "loss": 0.2065, "step": 557 }, { "epoch": 5.166666666666667, "grad_norm": 2.9694433212280273, "learning_rate": 0.00013267080745341616, "loss": 0.0597, "step": 558 }, { "epoch": 5.175925925925926, "grad_norm": 3.9127731323242188, "learning_rate": 0.00013254658385093167, "loss": 0.1278, "step": 559 }, { "epoch": 5.185185185185185, "grad_norm": 4.475368499755859, "learning_rate": 0.00013242236024844722, "loss": 0.1604, "step": 560 }, { "epoch": 5.194444444444445, "grad_norm": 8.596575736999512, "learning_rate": 0.00013229813664596273, "loss": 0.2838, "step": 561 }, { "epoch": 5.203703703703703, "grad_norm": 3.6132686138153076, "learning_rate": 0.00013217391304347825, "loss": 0.0587, "step": 562 }, { "epoch": 5.212962962962963, "grad_norm": 9.223053932189941, "learning_rate": 0.0001320496894409938, "loss": 0.2524, "step": 563 }, { "epoch": 5.222222222222222, "grad_norm": 1.8150829076766968, "learning_rate": 0.00013192546583850934, "loss": 0.0615, "step": 564 }, { "epoch": 5.231481481481482, "grad_norm": 4.499701976776123, "learning_rate": 0.00013180124223602485, "loss": 0.1889, "step": 565 }, { "epoch": 5.2407407407407405, "grad_norm": 3.4126698970794678, "learning_rate": 0.0001316770186335404, "loss": 0.1661, "step": 566 }, { "epoch": 5.25, "grad_norm": 1.3917155265808105, "learning_rate": 0.0001315527950310559, "loss": 0.0479, "step": 567 }, { "epoch": 5.2592592592592595, "grad_norm": 4.621674060821533, "learning_rate": 0.00013142857142857143, "loss": 0.1365, "step": 568 }, { "epoch": 5.268518518518518, "grad_norm": 2.671595335006714, "learning_rate": 0.00013130434782608697, "loss": 0.0965, "step": 569 }, { "epoch": 5.277777777777778, "grad_norm": 2.2948710918426514, "learning_rate": 0.0001311801242236025, "loss": 0.057, "step": 570 }, { "epoch": 5.287037037037037, "grad_norm": 3.181098699569702, "learning_rate": 0.000131055900621118, "loss": 0.0844, "step": 571 }, { "epoch": 5.296296296296296, "grad_norm": 1.5297895669937134, "learning_rate": 0.00013093167701863355, "loss": 0.09, "step": 572 }, { "epoch": 5.305555555555555, "grad_norm": 5.39132022857666, "learning_rate": 0.0001308074534161491, "loss": 0.1515, "step": 573 }, { "epoch": 5.314814814814815, "grad_norm": 3.3082971572875977, "learning_rate": 0.0001306832298136646, "loss": 0.1473, "step": 574 }, { "epoch": 5.324074074074074, "grad_norm": 4.612159252166748, "learning_rate": 0.00013055900621118012, "loss": 0.2653, "step": 575 }, { "epoch": 5.333333333333333, "grad_norm": 8.018484115600586, "learning_rate": 0.00013043478260869567, "loss": 0.4067, "step": 576 }, { "epoch": 5.342592592592593, "grad_norm": 2.2573232650756836, "learning_rate": 0.00013031055900621118, "loss": 0.0973, "step": 577 }, { "epoch": 5.351851851851852, "grad_norm": 3.4523372650146484, "learning_rate": 0.0001301863354037267, "loss": 0.1417, "step": 578 }, { "epoch": 5.361111111111111, "grad_norm": 3.2705371379852295, "learning_rate": 0.00013006211180124224, "loss": 0.1203, "step": 579 }, { "epoch": 5.37037037037037, "grad_norm": 1.8731480836868286, "learning_rate": 0.00012993788819875776, "loss": 0.0674, "step": 580 }, { "epoch": 5.37962962962963, "grad_norm": 4.4362101554870605, "learning_rate": 0.00012981366459627328, "loss": 0.1952, "step": 581 }, { "epoch": 5.388888888888889, "grad_norm": 4.875374794006348, "learning_rate": 0.00012968944099378885, "loss": 0.1, "step": 582 }, { "epoch": 5.398148148148148, "grad_norm": 5.201648235321045, "learning_rate": 0.00012956521739130436, "loss": 0.277, "step": 583 }, { "epoch": 5.407407407407407, "grad_norm": 2.5301337242126465, "learning_rate": 0.00012944099378881988, "loss": 0.0705, "step": 584 }, { "epoch": 5.416666666666667, "grad_norm": 3.144968271255493, "learning_rate": 0.00012931677018633542, "loss": 0.0667, "step": 585 }, { "epoch": 5.425925925925926, "grad_norm": 2.509068489074707, "learning_rate": 0.00012919254658385094, "loss": 0.081, "step": 586 }, { "epoch": 5.435185185185185, "grad_norm": 7.523172855377197, "learning_rate": 0.00012906832298136646, "loss": 0.2569, "step": 587 }, { "epoch": 5.444444444444445, "grad_norm": 3.6031181812286377, "learning_rate": 0.000128944099378882, "loss": 0.1324, "step": 588 }, { "epoch": 5.453703703703704, "grad_norm": 3.194385528564453, "learning_rate": 0.00012881987577639752, "loss": 0.113, "step": 589 }, { "epoch": 5.462962962962963, "grad_norm": 1.4715389013290405, "learning_rate": 0.00012869565217391303, "loss": 0.073, "step": 590 }, { "epoch": 5.472222222222222, "grad_norm": 2.3477694988250732, "learning_rate": 0.00012857142857142858, "loss": 0.0708, "step": 591 }, { "epoch": 5.481481481481482, "grad_norm": 1.5223402976989746, "learning_rate": 0.00012844720496894412, "loss": 0.0464, "step": 592 }, { "epoch": 5.4907407407407405, "grad_norm": 6.1966352462768555, "learning_rate": 0.00012832298136645964, "loss": 0.2479, "step": 593 }, { "epoch": 5.5, "grad_norm": 10.558868408203125, "learning_rate": 0.00012819875776397515, "loss": 0.5108, "step": 594 }, { "epoch": 5.5092592592592595, "grad_norm": 5.110361576080322, "learning_rate": 0.0001280745341614907, "loss": 0.2486, "step": 595 }, { "epoch": 5.518518518518518, "grad_norm": 1.4477581977844238, "learning_rate": 0.0001279503105590062, "loss": 0.0472, "step": 596 }, { "epoch": 5.527777777777778, "grad_norm": 4.302581787109375, "learning_rate": 0.00012782608695652173, "loss": 0.2134, "step": 597 }, { "epoch": 5.537037037037037, "grad_norm": 4.25509786605835, "learning_rate": 0.00012770186335403727, "loss": 0.1519, "step": 598 }, { "epoch": 5.546296296296296, "grad_norm": 3.6477010250091553, "learning_rate": 0.0001275776397515528, "loss": 0.1756, "step": 599 }, { "epoch": 5.555555555555555, "grad_norm": 3.412436008453369, "learning_rate": 0.00012745341614906833, "loss": 0.1473, "step": 600 }, { "epoch": 5.564814814814815, "grad_norm": 2.7615256309509277, "learning_rate": 0.00012732919254658387, "loss": 0.0985, "step": 601 }, { "epoch": 5.574074074074074, "grad_norm": 4.826498985290527, "learning_rate": 0.0001272049689440994, "loss": 0.1564, "step": 602 }, { "epoch": 5.583333333333333, "grad_norm": 4.768690586090088, "learning_rate": 0.0001270807453416149, "loss": 0.2187, "step": 603 }, { "epoch": 5.592592592592593, "grad_norm": 4.710096836090088, "learning_rate": 0.00012695652173913045, "loss": 0.0982, "step": 604 }, { "epoch": 5.601851851851852, "grad_norm": 3.721019744873047, "learning_rate": 0.00012683229813664597, "loss": 0.1331, "step": 605 }, { "epoch": 5.611111111111111, "grad_norm": 10.823250770568848, "learning_rate": 0.00012670807453416148, "loss": 0.2592, "step": 606 }, { "epoch": 5.62037037037037, "grad_norm": 2.030266761779785, "learning_rate": 0.00012658385093167703, "loss": 0.0532, "step": 607 }, { "epoch": 5.62962962962963, "grad_norm": 4.477989673614502, "learning_rate": 0.00012645962732919254, "loss": 0.2144, "step": 608 }, { "epoch": 5.638888888888889, "grad_norm": 4.200387954711914, "learning_rate": 0.0001263354037267081, "loss": 0.1709, "step": 609 }, { "epoch": 5.648148148148148, "grad_norm": 6.254189968109131, "learning_rate": 0.0001262111801242236, "loss": 0.3155, "step": 610 }, { "epoch": 5.657407407407407, "grad_norm": 8.545554161071777, "learning_rate": 0.00012608695652173915, "loss": 0.2297, "step": 611 }, { "epoch": 5.666666666666667, "grad_norm": 6.469607830047607, "learning_rate": 0.00012596273291925466, "loss": 0.2023, "step": 612 }, { "epoch": 5.675925925925926, "grad_norm": 4.819572448730469, "learning_rate": 0.00012583850931677018, "loss": 0.1118, "step": 613 }, { "epoch": 5.685185185185185, "grad_norm": 5.746372699737549, "learning_rate": 0.00012571428571428572, "loss": 0.2105, "step": 614 }, { "epoch": 5.694444444444445, "grad_norm": 2.2594149112701416, "learning_rate": 0.00012559006211180124, "loss": 0.0692, "step": 615 }, { "epoch": 5.703703703703704, "grad_norm": 3.391378402709961, "learning_rate": 0.00012546583850931676, "loss": 0.1631, "step": 616 }, { "epoch": 5.712962962962963, "grad_norm": 2.416281223297119, "learning_rate": 0.0001253416149068323, "loss": 0.1129, "step": 617 }, { "epoch": 5.722222222222222, "grad_norm": 6.688156604766846, "learning_rate": 0.00012521739130434784, "loss": 0.1804, "step": 618 }, { "epoch": 5.731481481481482, "grad_norm": 5.007920265197754, "learning_rate": 0.00012509316770186336, "loss": 0.3592, "step": 619 }, { "epoch": 5.7407407407407405, "grad_norm": 4.108309268951416, "learning_rate": 0.0001249689440993789, "loss": 0.1732, "step": 620 }, { "epoch": 5.75, "grad_norm": 3.446714401245117, "learning_rate": 0.00012484472049689442, "loss": 0.1708, "step": 621 }, { "epoch": 5.7592592592592595, "grad_norm": 4.069822311401367, "learning_rate": 0.00012472049689440993, "loss": 0.198, "step": 622 }, { "epoch": 5.768518518518518, "grad_norm": 4.49385929107666, "learning_rate": 0.00012459627329192548, "loss": 0.2133, "step": 623 }, { "epoch": 5.777777777777778, "grad_norm": 2.8163678646087646, "learning_rate": 0.000124472049689441, "loss": 0.1077, "step": 624 }, { "epoch": 5.787037037037037, "grad_norm": 2.6389107704162598, "learning_rate": 0.0001243478260869565, "loss": 0.087, "step": 625 }, { "epoch": 5.796296296296296, "grad_norm": 4.10544490814209, "learning_rate": 0.00012422360248447205, "loss": 0.0834, "step": 626 }, { "epoch": 5.805555555555555, "grad_norm": 1.4941080808639526, "learning_rate": 0.0001240993788819876, "loss": 0.0485, "step": 627 }, { "epoch": 5.814814814814815, "grad_norm": 7.434000015258789, "learning_rate": 0.00012397515527950311, "loss": 0.3163, "step": 628 }, { "epoch": 5.824074074074074, "grad_norm": 5.449849605560303, "learning_rate": 0.00012385093167701866, "loss": 0.1665, "step": 629 }, { "epoch": 5.833333333333333, "grad_norm": 4.494096279144287, "learning_rate": 0.00012372670807453417, "loss": 0.1618, "step": 630 }, { "epoch": 5.842592592592593, "grad_norm": 3.1159920692443848, "learning_rate": 0.0001236024844720497, "loss": 0.0993, "step": 631 }, { "epoch": 5.851851851851852, "grad_norm": 7.402317523956299, "learning_rate": 0.00012347826086956523, "loss": 0.2546, "step": 632 }, { "epoch": 5.861111111111111, "grad_norm": 5.94581937789917, "learning_rate": 0.00012335403726708075, "loss": 0.2239, "step": 633 }, { "epoch": 5.87037037037037, "grad_norm": 1.6560418605804443, "learning_rate": 0.00012322981366459627, "loss": 0.0531, "step": 634 }, { "epoch": 5.87962962962963, "grad_norm": 2.7259700298309326, "learning_rate": 0.0001231055900621118, "loss": 0.1182, "step": 635 }, { "epoch": 5.888888888888889, "grad_norm": 6.228329181671143, "learning_rate": 0.00012298136645962735, "loss": 0.3486, "step": 636 }, { "epoch": 5.898148148148148, "grad_norm": 1.6788215637207031, "learning_rate": 0.00012285714285714287, "loss": 0.0312, "step": 637 }, { "epoch": 5.907407407407407, "grad_norm": 3.3095898628234863, "learning_rate": 0.00012273291925465839, "loss": 0.148, "step": 638 }, { "epoch": 5.916666666666667, "grad_norm": 2.946584701538086, "learning_rate": 0.00012260869565217393, "loss": 0.0905, "step": 639 }, { "epoch": 5.925925925925926, "grad_norm": 7.601954936981201, "learning_rate": 0.00012248447204968945, "loss": 0.3361, "step": 640 }, { "epoch": 5.935185185185185, "grad_norm": 3.047450542449951, "learning_rate": 0.00012236024844720496, "loss": 0.08, "step": 641 }, { "epoch": 5.944444444444445, "grad_norm": 2.289433717727661, "learning_rate": 0.0001222360248447205, "loss": 0.1263, "step": 642 }, { "epoch": 5.953703703703704, "grad_norm": 3.0301482677459717, "learning_rate": 0.00012211180124223602, "loss": 0.1434, "step": 643 }, { "epoch": 5.962962962962963, "grad_norm": 1.7519960403442383, "learning_rate": 0.00012198757763975155, "loss": 0.0757, "step": 644 }, { "epoch": 5.972222222222222, "grad_norm": 1.5886623859405518, "learning_rate": 0.0001218633540372671, "loss": 0.0571, "step": 645 }, { "epoch": 5.981481481481482, "grad_norm": 5.018250465393066, "learning_rate": 0.00012173913043478263, "loss": 0.3047, "step": 646 }, { "epoch": 5.9907407407407405, "grad_norm": 2.7874503135681152, "learning_rate": 0.00012161490683229814, "loss": 0.1017, "step": 647 }, { "epoch": 6.0, "grad_norm": 4.445015907287598, "learning_rate": 0.00012149068322981367, "loss": 0.1223, "step": 648 }, { "epoch": 6.0092592592592595, "grad_norm": 5.86106538772583, "learning_rate": 0.0001213664596273292, "loss": 0.1565, "step": 649 }, { "epoch": 6.018518518518518, "grad_norm": 3.1399528980255127, "learning_rate": 0.00012124223602484472, "loss": 0.1554, "step": 650 }, { "epoch": 6.027777777777778, "grad_norm": 3.606825828552246, "learning_rate": 0.00012111801242236025, "loss": 0.2018, "step": 651 }, { "epoch": 6.037037037037037, "grad_norm": 2.6407530307769775, "learning_rate": 0.00012099378881987578, "loss": 0.125, "step": 652 }, { "epoch": 6.046296296296297, "grad_norm": 2.7450759410858154, "learning_rate": 0.00012086956521739131, "loss": 0.0707, "step": 653 }, { "epoch": 6.055555555555555, "grad_norm": 4.761390209197998, "learning_rate": 0.00012074534161490685, "loss": 0.1126, "step": 654 }, { "epoch": 6.064814814814815, "grad_norm": 2.6973469257354736, "learning_rate": 0.00012062111801242237, "loss": 0.0593, "step": 655 }, { "epoch": 6.074074074074074, "grad_norm": 3.7341206073760986, "learning_rate": 0.0001204968944099379, "loss": 0.132, "step": 656 }, { "epoch": 6.083333333333333, "grad_norm": 8.52206039428711, "learning_rate": 0.00012037267080745343, "loss": 0.2624, "step": 657 }, { "epoch": 6.092592592592593, "grad_norm": 2.5560076236724854, "learning_rate": 0.00012024844720496896, "loss": 0.0968, "step": 658 }, { "epoch": 6.101851851851852, "grad_norm": 1.632588505744934, "learning_rate": 0.00012012422360248447, "loss": 0.0623, "step": 659 }, { "epoch": 6.111111111111111, "grad_norm": 3.6222176551818848, "learning_rate": 0.00012, "loss": 0.123, "step": 660 }, { "epoch": 6.12037037037037, "grad_norm": 3.070725202560425, "learning_rate": 0.00011987577639751553, "loss": 0.1106, "step": 661 }, { "epoch": 6.12962962962963, "grad_norm": 3.651214599609375, "learning_rate": 0.00011975155279503105, "loss": 0.1414, "step": 662 }, { "epoch": 6.138888888888889, "grad_norm": 3.094923734664917, "learning_rate": 0.0001196273291925466, "loss": 0.1233, "step": 663 }, { "epoch": 6.148148148148148, "grad_norm": 6.006683349609375, "learning_rate": 0.00011950310559006212, "loss": 0.1618, "step": 664 }, { "epoch": 6.157407407407407, "grad_norm": 1.4269198179244995, "learning_rate": 0.00011937888198757765, "loss": 0.0483, "step": 665 }, { "epoch": 6.166666666666667, "grad_norm": 3.9230058193206787, "learning_rate": 0.00011925465838509318, "loss": 0.0853, "step": 666 }, { "epoch": 6.175925925925926, "grad_norm": 5.348933696746826, "learning_rate": 0.0001191304347826087, "loss": 0.1022, "step": 667 }, { "epoch": 6.185185185185185, "grad_norm": 5.111001014709473, "learning_rate": 0.00011900621118012423, "loss": 0.1807, "step": 668 }, { "epoch": 6.194444444444445, "grad_norm": 8.565694808959961, "learning_rate": 0.00011888198757763976, "loss": 0.0918, "step": 669 }, { "epoch": 6.203703703703703, "grad_norm": 1.8871145248413086, "learning_rate": 0.00011875776397515527, "loss": 0.0277, "step": 670 }, { "epoch": 6.212962962962963, "grad_norm": 2.728764772415161, "learning_rate": 0.0001186335403726708, "loss": 0.0934, "step": 671 }, { "epoch": 6.222222222222222, "grad_norm": 4.937579154968262, "learning_rate": 0.00011850931677018635, "loss": 0.1346, "step": 672 }, { "epoch": 6.231481481481482, "grad_norm": 6.841187953948975, "learning_rate": 0.00011838509316770188, "loss": 0.1542, "step": 673 }, { "epoch": 6.2407407407407405, "grad_norm": 5.780396461486816, "learning_rate": 0.00011826086956521741, "loss": 0.2692, "step": 674 }, { "epoch": 6.25, "grad_norm": 4.7826080322265625, "learning_rate": 0.00011813664596273292, "loss": 0.1671, "step": 675 }, { "epoch": 6.2592592592592595, "grad_norm": 4.660366535186768, "learning_rate": 0.00011801242236024845, "loss": 0.1156, "step": 676 }, { "epoch": 6.268518518518518, "grad_norm": 2.414586067199707, "learning_rate": 0.00011788819875776398, "loss": 0.0844, "step": 677 }, { "epoch": 6.277777777777778, "grad_norm": 2.0737192630767822, "learning_rate": 0.0001177639751552795, "loss": 0.0828, "step": 678 }, { "epoch": 6.287037037037037, "grad_norm": 3.5586555004119873, "learning_rate": 0.00011763975155279503, "loss": 0.1531, "step": 679 }, { "epoch": 6.296296296296296, "grad_norm": 5.451366901397705, "learning_rate": 0.00011751552795031056, "loss": 0.1401, "step": 680 }, { "epoch": 6.305555555555555, "grad_norm": 2.353264570236206, "learning_rate": 0.0001173913043478261, "loss": 0.0969, "step": 681 }, { "epoch": 6.314814814814815, "grad_norm": 3.3806886672973633, "learning_rate": 0.00011726708074534163, "loss": 0.0933, "step": 682 }, { "epoch": 6.324074074074074, "grad_norm": 3.2616827487945557, "learning_rate": 0.00011714285714285715, "loss": 0.1034, "step": 683 }, { "epoch": 6.333333333333333, "grad_norm": 2.7368857860565186, "learning_rate": 0.00011701863354037268, "loss": 0.0752, "step": 684 }, { "epoch": 6.342592592592593, "grad_norm": 6.251594543457031, "learning_rate": 0.00011689440993788821, "loss": 0.1462, "step": 685 }, { "epoch": 6.351851851851852, "grad_norm": 4.142698764801025, "learning_rate": 0.00011677018633540373, "loss": 0.1326, "step": 686 }, { "epoch": 6.361111111111111, "grad_norm": 2.420140504837036, "learning_rate": 0.00011664596273291926, "loss": 0.0767, "step": 687 }, { "epoch": 6.37037037037037, "grad_norm": 3.2684707641601562, "learning_rate": 0.00011652173913043479, "loss": 0.0698, "step": 688 }, { "epoch": 6.37962962962963, "grad_norm": 2.8635385036468506, "learning_rate": 0.0001163975155279503, "loss": 0.1454, "step": 689 }, { "epoch": 6.388888888888889, "grad_norm": 1.000563144683838, "learning_rate": 0.00011627329192546586, "loss": 0.037, "step": 690 }, { "epoch": 6.398148148148148, "grad_norm": 5.726893424987793, "learning_rate": 0.00011614906832298138, "loss": 0.2738, "step": 691 }, { "epoch": 6.407407407407407, "grad_norm": 1.0472891330718994, "learning_rate": 0.0001160248447204969, "loss": 0.0284, "step": 692 }, { "epoch": 6.416666666666667, "grad_norm": 2.0266168117523193, "learning_rate": 0.00011590062111801244, "loss": 0.0703, "step": 693 }, { "epoch": 6.425925925925926, "grad_norm": 4.494514465332031, "learning_rate": 0.00011577639751552795, "loss": 0.1956, "step": 694 }, { "epoch": 6.435185185185185, "grad_norm": 5.9851837158203125, "learning_rate": 0.00011565217391304348, "loss": 0.1779, "step": 695 }, { "epoch": 6.444444444444445, "grad_norm": 3.127121686935425, "learning_rate": 0.00011552795031055901, "loss": 0.0926, "step": 696 }, { "epoch": 6.453703703703704, "grad_norm": 9.09554672241211, "learning_rate": 0.00011540372670807453, "loss": 0.236, "step": 697 }, { "epoch": 6.462962962962963, "grad_norm": 6.604413032531738, "learning_rate": 0.00011527950310559006, "loss": 0.2466, "step": 698 }, { "epoch": 6.472222222222222, "grad_norm": 1.7158292531967163, "learning_rate": 0.0001151552795031056, "loss": 0.0599, "step": 699 }, { "epoch": 6.481481481481482, "grad_norm": 2.3201727867126465, "learning_rate": 0.00011503105590062113, "loss": 0.1221, "step": 700 }, { "epoch": 6.4907407407407405, "grad_norm": 3.0037081241607666, "learning_rate": 0.00011490683229813666, "loss": 0.0752, "step": 701 }, { "epoch": 6.5, "grad_norm": 2.66445255279541, "learning_rate": 0.00011478260869565218, "loss": 0.0936, "step": 702 }, { "epoch": 6.5092592592592595, "grad_norm": 2.3784914016723633, "learning_rate": 0.00011465838509316771, "loss": 0.1362, "step": 703 }, { "epoch": 6.518518518518518, "grad_norm": 1.467207431793213, "learning_rate": 0.00011453416149068324, "loss": 0.033, "step": 704 }, { "epoch": 6.527777777777778, "grad_norm": 4.558538436889648, "learning_rate": 0.00011440993788819875, "loss": 0.2072, "step": 705 }, { "epoch": 6.537037037037037, "grad_norm": 2.80195951461792, "learning_rate": 0.00011428571428571428, "loss": 0.0687, "step": 706 }, { "epoch": 6.546296296296296, "grad_norm": 8.565617561340332, "learning_rate": 0.00011416149068322981, "loss": 0.2849, "step": 707 }, { "epoch": 6.555555555555555, "grad_norm": 1.682445764541626, "learning_rate": 0.00011403726708074536, "loss": 0.0688, "step": 708 }, { "epoch": 6.564814814814815, "grad_norm": 2.2818193435668945, "learning_rate": 0.00011391304347826089, "loss": 0.0762, "step": 709 }, { "epoch": 6.574074074074074, "grad_norm": 1.554701328277588, "learning_rate": 0.0001137888198757764, "loss": 0.0588, "step": 710 }, { "epoch": 6.583333333333333, "grad_norm": 0.8565337061882019, "learning_rate": 0.00011366459627329193, "loss": 0.0296, "step": 711 }, { "epoch": 6.592592592592593, "grad_norm": 3.192460060119629, "learning_rate": 0.00011354037267080746, "loss": 0.0978, "step": 712 }, { "epoch": 6.601851851851852, "grad_norm": 1.5544946193695068, "learning_rate": 0.00011341614906832298, "loss": 0.0615, "step": 713 }, { "epoch": 6.611111111111111, "grad_norm": 5.477461814880371, "learning_rate": 0.00011329192546583851, "loss": 0.2045, "step": 714 }, { "epoch": 6.62037037037037, "grad_norm": 4.519773960113525, "learning_rate": 0.00011316770186335404, "loss": 0.2205, "step": 715 }, { "epoch": 6.62962962962963, "grad_norm": 5.499269962310791, "learning_rate": 0.00011304347826086956, "loss": 0.1925, "step": 716 }, { "epoch": 6.638888888888889, "grad_norm": 3.747269868850708, "learning_rate": 0.00011291925465838511, "loss": 0.1676, "step": 717 }, { "epoch": 6.648148148148148, "grad_norm": 2.2231295108795166, "learning_rate": 0.00011279503105590063, "loss": 0.1009, "step": 718 }, { "epoch": 6.657407407407407, "grad_norm": 2.7668187618255615, "learning_rate": 0.00011267080745341616, "loss": 0.075, "step": 719 }, { "epoch": 6.666666666666667, "grad_norm": 3.6396594047546387, "learning_rate": 0.00011254658385093169, "loss": 0.1634, "step": 720 }, { "epoch": 6.675925925925926, "grad_norm": 2.8981094360351562, "learning_rate": 0.00011242236024844722, "loss": 0.167, "step": 721 }, { "epoch": 6.685185185185185, "grad_norm": 1.7787268161773682, "learning_rate": 0.00011229813664596273, "loss": 0.0541, "step": 722 }, { "epoch": 6.694444444444445, "grad_norm": 5.248884677886963, "learning_rate": 0.00011217391304347826, "loss": 0.1612, "step": 723 }, { "epoch": 6.703703703703704, "grad_norm": 6.716445446014404, "learning_rate": 0.0001120496894409938, "loss": 0.232, "step": 724 }, { "epoch": 6.712962962962963, "grad_norm": 4.387823581695557, "learning_rate": 0.00011192546583850931, "loss": 0.1634, "step": 725 }, { "epoch": 6.722222222222222, "grad_norm": 2.278259515762329, "learning_rate": 0.00011180124223602484, "loss": 0.0771, "step": 726 }, { "epoch": 6.731481481481482, "grad_norm": 3.915947198867798, "learning_rate": 0.00011167701863354038, "loss": 0.1595, "step": 727 }, { "epoch": 6.7407407407407405, "grad_norm": 3.473524332046509, "learning_rate": 0.00011155279503105591, "loss": 0.1075, "step": 728 }, { "epoch": 6.75, "grad_norm": 1.3765463829040527, "learning_rate": 0.00011142857142857144, "loss": 0.0532, "step": 729 }, { "epoch": 6.7592592592592595, "grad_norm": 3.4707448482513428, "learning_rate": 0.00011130434782608696, "loss": 0.1275, "step": 730 }, { "epoch": 6.768518518518518, "grad_norm": 4.364494800567627, "learning_rate": 0.00011118012422360249, "loss": 0.1482, "step": 731 }, { "epoch": 6.777777777777778, "grad_norm": 3.434342384338379, "learning_rate": 0.00011105590062111802, "loss": 0.091, "step": 732 }, { "epoch": 6.787037037037037, "grad_norm": 2.428468942642212, "learning_rate": 0.00011093167701863354, "loss": 0.1011, "step": 733 }, { "epoch": 6.796296296296296, "grad_norm": 4.1566057205200195, "learning_rate": 0.00011080745341614907, "loss": 0.1952, "step": 734 }, { "epoch": 6.805555555555555, "grad_norm": 3.701270818710327, "learning_rate": 0.0001106832298136646, "loss": 0.1344, "step": 735 }, { "epoch": 6.814814814814815, "grad_norm": 3.952190399169922, "learning_rate": 0.00011055900621118014, "loss": 0.1137, "step": 736 }, { "epoch": 6.824074074074074, "grad_norm": 1.9077421426773071, "learning_rate": 0.00011043478260869567, "loss": 0.0495, "step": 737 }, { "epoch": 6.833333333333333, "grad_norm": 7.015697956085205, "learning_rate": 0.00011031055900621119, "loss": 0.3821, "step": 738 }, { "epoch": 6.842592592592593, "grad_norm": 2.9424688816070557, "learning_rate": 0.00011018633540372672, "loss": 0.0788, "step": 739 }, { "epoch": 6.851851851851852, "grad_norm": 1.9170938730239868, "learning_rate": 0.00011006211180124225, "loss": 0.0735, "step": 740 }, { "epoch": 6.861111111111111, "grad_norm": 3.8930490016937256, "learning_rate": 0.00010993788819875776, "loss": 0.0723, "step": 741 }, { "epoch": 6.87037037037037, "grad_norm": 5.078344345092773, "learning_rate": 0.00010981366459627329, "loss": 0.1568, "step": 742 }, { "epoch": 6.87962962962963, "grad_norm": 1.1494572162628174, "learning_rate": 0.00010968944099378882, "loss": 0.0333, "step": 743 }, { "epoch": 6.888888888888889, "grad_norm": 4.472750663757324, "learning_rate": 0.00010956521739130434, "loss": 0.1611, "step": 744 }, { "epoch": 6.898148148148148, "grad_norm": 2.977266311645508, "learning_rate": 0.0001094409937888199, "loss": 0.1003, "step": 745 }, { "epoch": 6.907407407407407, "grad_norm": 2.7986366748809814, "learning_rate": 0.00010931677018633541, "loss": 0.104, "step": 746 }, { "epoch": 6.916666666666667, "grad_norm": 3.25215482711792, "learning_rate": 0.00010919254658385094, "loss": 0.1337, "step": 747 }, { "epoch": 6.925925925925926, "grad_norm": 4.609085559844971, "learning_rate": 0.00010906832298136647, "loss": 0.1346, "step": 748 }, { "epoch": 6.935185185185185, "grad_norm": 2.8187437057495117, "learning_rate": 0.00010894409937888199, "loss": 0.0817, "step": 749 }, { "epoch": 6.944444444444445, "grad_norm": 3.499835252761841, "learning_rate": 0.00010881987577639752, "loss": 0.1592, "step": 750 }, { "epoch": 6.953703703703704, "grad_norm": 5.405245304107666, "learning_rate": 0.00010869565217391305, "loss": 0.1918, "step": 751 }, { "epoch": 6.962962962962963, "grad_norm": 3.507462978363037, "learning_rate": 0.00010857142857142856, "loss": 0.1177, "step": 752 }, { "epoch": 6.972222222222222, "grad_norm": 4.5707316398620605, "learning_rate": 0.0001084472049689441, "loss": 0.2059, "step": 753 }, { "epoch": 6.981481481481482, "grad_norm": 4.220851898193359, "learning_rate": 0.00010832298136645964, "loss": 0.1032, "step": 754 }, { "epoch": 6.9907407407407405, "grad_norm": 1.384399652481079, "learning_rate": 0.00010819875776397517, "loss": 0.0396, "step": 755 }, { "epoch": 7.0, "grad_norm": 4.726045608520508, "learning_rate": 0.0001080745341614907, "loss": 0.1665, "step": 756 }, { "epoch": 7.0092592592592595, "grad_norm": 1.7653844356536865, "learning_rate": 0.00010795031055900621, "loss": 0.0523, "step": 757 }, { "epoch": 7.018518518518518, "grad_norm": 3.113438606262207, "learning_rate": 0.00010782608695652174, "loss": 0.1037, "step": 758 }, { "epoch": 7.027777777777778, "grad_norm": 0.7334030866622925, "learning_rate": 0.00010770186335403727, "loss": 0.018, "step": 759 }, { "epoch": 7.037037037037037, "grad_norm": 4.1807403564453125, "learning_rate": 0.00010757763975155279, "loss": 0.1268, "step": 760 }, { "epoch": 7.046296296296297, "grad_norm": 0.9362726807594299, "learning_rate": 0.00010745341614906832, "loss": 0.0297, "step": 761 }, { "epoch": 7.055555555555555, "grad_norm": 4.32288932800293, "learning_rate": 0.00010732919254658385, "loss": 0.1442, "step": 762 }, { "epoch": 7.064814814814815, "grad_norm": 5.067570209503174, "learning_rate": 0.00010720496894409939, "loss": 0.1025, "step": 763 }, { "epoch": 7.074074074074074, "grad_norm": 4.261683464050293, "learning_rate": 0.00010708074534161492, "loss": 0.0737, "step": 764 }, { "epoch": 7.083333333333333, "grad_norm": 1.640762448310852, "learning_rate": 0.00010695652173913044, "loss": 0.0539, "step": 765 }, { "epoch": 7.092592592592593, "grad_norm": 4.321948051452637, "learning_rate": 0.00010683229813664597, "loss": 0.1349, "step": 766 }, { "epoch": 7.101851851851852, "grad_norm": 3.3006460666656494, "learning_rate": 0.0001067080745341615, "loss": 0.1521, "step": 767 }, { "epoch": 7.111111111111111, "grad_norm": 4.936793804168701, "learning_rate": 0.00010658385093167702, "loss": 0.2134, "step": 768 }, { "epoch": 7.12037037037037, "grad_norm": 4.625700950622559, "learning_rate": 0.00010645962732919255, "loss": 0.1357, "step": 769 }, { "epoch": 7.12962962962963, "grad_norm": 3.6059927940368652, "learning_rate": 0.00010633540372670807, "loss": 0.1155, "step": 770 }, { "epoch": 7.138888888888889, "grad_norm": 7.513297080993652, "learning_rate": 0.00010621118012422359, "loss": 0.2876, "step": 771 }, { "epoch": 7.148148148148148, "grad_norm": 1.9764149188995361, "learning_rate": 0.00010608695652173915, "loss": 0.0494, "step": 772 }, { "epoch": 7.157407407407407, "grad_norm": 6.305065631866455, "learning_rate": 0.00010596273291925466, "loss": 0.1551, "step": 773 }, { "epoch": 7.166666666666667, "grad_norm": 7.061641216278076, "learning_rate": 0.0001058385093167702, "loss": 0.4477, "step": 774 }, { "epoch": 7.175925925925926, "grad_norm": 2.5913658142089844, "learning_rate": 0.00010571428571428572, "loss": 0.0891, "step": 775 }, { "epoch": 7.185185185185185, "grad_norm": 4.326108932495117, "learning_rate": 0.00010559006211180124, "loss": 0.1668, "step": 776 }, { "epoch": 7.194444444444445, "grad_norm": 1.2704784870147705, "learning_rate": 0.00010546583850931677, "loss": 0.0479, "step": 777 }, { "epoch": 7.203703703703703, "grad_norm": 4.410538673400879, "learning_rate": 0.0001053416149068323, "loss": 0.091, "step": 778 }, { "epoch": 7.212962962962963, "grad_norm": 4.080562114715576, "learning_rate": 0.00010521739130434782, "loss": 0.0864, "step": 779 }, { "epoch": 7.222222222222222, "grad_norm": 2.016847848892212, "learning_rate": 0.00010509316770186335, "loss": 0.0642, "step": 780 }, { "epoch": 7.231481481481482, "grad_norm": 1.3171155452728271, "learning_rate": 0.00010496894409937889, "loss": 0.0341, "step": 781 }, { "epoch": 7.2407407407407405, "grad_norm": 3.395470380783081, "learning_rate": 0.00010484472049689442, "loss": 0.0992, "step": 782 }, { "epoch": 7.25, "grad_norm": 3.640408515930176, "learning_rate": 0.00010472049689440995, "loss": 0.1221, "step": 783 }, { "epoch": 7.2592592592592595, "grad_norm": 2.7023701667785645, "learning_rate": 0.00010459627329192547, "loss": 0.0616, "step": 784 }, { "epoch": 7.268518518518518, "grad_norm": 1.2496731281280518, "learning_rate": 0.000104472049689441, "loss": 0.0395, "step": 785 }, { "epoch": 7.277777777777778, "grad_norm": 4.615080833435059, "learning_rate": 0.00010434782608695653, "loss": 0.1183, "step": 786 }, { "epoch": 7.287037037037037, "grad_norm": 2.8553638458251953, "learning_rate": 0.00010422360248447206, "loss": 0.065, "step": 787 }, { "epoch": 7.296296296296296, "grad_norm": 5.177341461181641, "learning_rate": 0.00010409937888198757, "loss": 0.1311, "step": 788 }, { "epoch": 7.305555555555555, "grad_norm": 4.220547199249268, "learning_rate": 0.0001039751552795031, "loss": 0.0963, "step": 789 }, { "epoch": 7.314814814814815, "grad_norm": 4.426061630249023, "learning_rate": 0.00010385093167701865, "loss": 0.0784, "step": 790 }, { "epoch": 7.324074074074074, "grad_norm": 3.388397693634033, "learning_rate": 0.00010372670807453418, "loss": 0.0874, "step": 791 }, { "epoch": 7.333333333333333, "grad_norm": 1.8489770889282227, "learning_rate": 0.0001036024844720497, "loss": 0.0512, "step": 792 }, { "epoch": 7.342592592592593, "grad_norm": 3.316032886505127, "learning_rate": 0.00010347826086956522, "loss": 0.0538, "step": 793 }, { "epoch": 7.351851851851852, "grad_norm": 2.2799925804138184, "learning_rate": 0.00010335403726708075, "loss": 0.0607, "step": 794 }, { "epoch": 7.361111111111111, "grad_norm": 6.567706108093262, "learning_rate": 0.00010322981366459628, "loss": 0.2587, "step": 795 }, { "epoch": 7.37037037037037, "grad_norm": 1.3475052118301392, "learning_rate": 0.0001031055900621118, "loss": 0.036, "step": 796 }, { "epoch": 7.37962962962963, "grad_norm": 4.109946250915527, "learning_rate": 0.00010298136645962733, "loss": 0.1113, "step": 797 }, { "epoch": 7.388888888888889, "grad_norm": 2.5367050170898438, "learning_rate": 0.00010285714285714286, "loss": 0.0818, "step": 798 }, { "epoch": 7.398148148148148, "grad_norm": 3.9039812088012695, "learning_rate": 0.0001027329192546584, "loss": 0.1218, "step": 799 }, { "epoch": 7.407407407407407, "grad_norm": 4.130105972290039, "learning_rate": 0.00010260869565217393, "loss": 0.111, "step": 800 }, { "epoch": 7.416666666666667, "grad_norm": 3.941784381866455, "learning_rate": 0.00010248447204968945, "loss": 0.1325, "step": 801 }, { "epoch": 7.425925925925926, "grad_norm": 1.690091848373413, "learning_rate": 0.00010236024844720498, "loss": 0.0302, "step": 802 }, { "epoch": 7.435185185185185, "grad_norm": 3.6138830184936523, "learning_rate": 0.00010223602484472051, "loss": 0.0976, "step": 803 }, { "epoch": 7.444444444444445, "grad_norm": 6.15052604675293, "learning_rate": 0.00010211180124223602, "loss": 0.2043, "step": 804 }, { "epoch": 7.453703703703704, "grad_norm": 0.9923351407051086, "learning_rate": 0.00010198757763975155, "loss": 0.0339, "step": 805 }, { "epoch": 7.462962962962963, "grad_norm": 1.4790726900100708, "learning_rate": 0.00010186335403726708, "loss": 0.0547, "step": 806 }, { "epoch": 7.472222222222222, "grad_norm": 2.1284537315368652, "learning_rate": 0.0001017391304347826, "loss": 0.0825, "step": 807 }, { "epoch": 7.481481481481482, "grad_norm": 6.377471923828125, "learning_rate": 0.00010161490683229816, "loss": 0.244, "step": 808 }, { "epoch": 7.4907407407407405, "grad_norm": 1.6442444324493408, "learning_rate": 0.00010149068322981367, "loss": 0.0599, "step": 809 }, { "epoch": 7.5, "grad_norm": 10.18280029296875, "learning_rate": 0.0001013664596273292, "loss": 0.1209, "step": 810 }, { "epoch": 7.5092592592592595, "grad_norm": 1.3371292352676392, "learning_rate": 0.00010124223602484473, "loss": 0.0328, "step": 811 }, { "epoch": 7.518518518518518, "grad_norm": 1.0884917974472046, "learning_rate": 0.00010111801242236025, "loss": 0.0481, "step": 812 }, { "epoch": 7.527777777777778, "grad_norm": 1.8597067594528198, "learning_rate": 0.00010099378881987578, "loss": 0.0612, "step": 813 }, { "epoch": 7.537037037037037, "grad_norm": 1.331111192703247, "learning_rate": 0.00010086956521739131, "loss": 0.0415, "step": 814 }, { "epoch": 7.546296296296296, "grad_norm": 5.1898322105407715, "learning_rate": 0.00010074534161490683, "loss": 0.2566, "step": 815 }, { "epoch": 7.555555555555555, "grad_norm": 3.672396183013916, "learning_rate": 0.00010062111801242236, "loss": 0.0728, "step": 816 }, { "epoch": 7.564814814814815, "grad_norm": 3.882545232772827, "learning_rate": 0.0001004968944099379, "loss": 0.1063, "step": 817 }, { "epoch": 7.574074074074074, "grad_norm": 4.563115119934082, "learning_rate": 0.00010037267080745343, "loss": 0.1335, "step": 818 }, { "epoch": 7.583333333333333, "grad_norm": 2.913468599319458, "learning_rate": 0.00010024844720496896, "loss": 0.1307, "step": 819 }, { "epoch": 7.592592592592593, "grad_norm": 5.023247241973877, "learning_rate": 0.00010012422360248447, "loss": 0.1065, "step": 820 }, { "epoch": 7.601851851851852, "grad_norm": 3.0828170776367188, "learning_rate": 0.0001, "loss": 0.0745, "step": 821 }, { "epoch": 7.611111111111111, "grad_norm": 1.2979748249053955, "learning_rate": 9.987577639751553e-05, "loss": 0.0521, "step": 822 }, { "epoch": 7.62037037037037, "grad_norm": 3.1594204902648926, "learning_rate": 9.975155279503105e-05, "loss": 0.1559, "step": 823 }, { "epoch": 7.62962962962963, "grad_norm": 5.367359638214111, "learning_rate": 9.96273291925466e-05, "loss": 0.2003, "step": 824 }, { "epoch": 7.638888888888889, "grad_norm": 4.4314284324646, "learning_rate": 9.950310559006212e-05, "loss": 0.1119, "step": 825 }, { "epoch": 7.648148148148148, "grad_norm": 2.753791570663452, "learning_rate": 9.937888198757764e-05, "loss": 0.05, "step": 826 }, { "epoch": 7.657407407407407, "grad_norm": 2.6093661785125732, "learning_rate": 9.925465838509317e-05, "loss": 0.0753, "step": 827 }, { "epoch": 7.666666666666667, "grad_norm": 3.5111002922058105, "learning_rate": 9.91304347826087e-05, "loss": 0.0774, "step": 828 }, { "epoch": 7.675925925925926, "grad_norm": 4.576236724853516, "learning_rate": 9.900621118012423e-05, "loss": 0.1454, "step": 829 }, { "epoch": 7.685185185185185, "grad_norm": 1.4670056104660034, "learning_rate": 9.888198757763976e-05, "loss": 0.059, "step": 830 }, { "epoch": 7.694444444444445, "grad_norm": 2.6183865070343018, "learning_rate": 9.875776397515528e-05, "loss": 0.0755, "step": 831 }, { "epoch": 7.703703703703704, "grad_norm": 2.1917359828948975, "learning_rate": 9.86335403726708e-05, "loss": 0.054, "step": 832 }, { "epoch": 7.712962962962963, "grad_norm": 1.6838301420211792, "learning_rate": 9.850931677018635e-05, "loss": 0.0606, "step": 833 }, { "epoch": 7.722222222222222, "grad_norm": 3.6826465129852295, "learning_rate": 9.838509316770187e-05, "loss": 0.1208, "step": 834 }, { "epoch": 7.731481481481482, "grad_norm": 1.396178126335144, "learning_rate": 9.82608695652174e-05, "loss": 0.0475, "step": 835 }, { "epoch": 7.7407407407407405, "grad_norm": 0.8179628849029541, "learning_rate": 9.813664596273293e-05, "loss": 0.0337, "step": 836 }, { "epoch": 7.75, "grad_norm": 1.4039196968078613, "learning_rate": 9.801242236024846e-05, "loss": 0.045, "step": 837 }, { "epoch": 7.7592592592592595, "grad_norm": 1.8538316488265991, "learning_rate": 9.788819875776399e-05, "loss": 0.0504, "step": 838 }, { "epoch": 7.768518518518518, "grad_norm": 3.978010654449463, "learning_rate": 9.77639751552795e-05, "loss": 0.1365, "step": 839 }, { "epoch": 7.777777777777778, "grad_norm": 6.751378059387207, "learning_rate": 9.763975155279503e-05, "loss": 0.2244, "step": 840 }, { "epoch": 7.787037037037037, "grad_norm": 1.3808748722076416, "learning_rate": 9.751552795031056e-05, "loss": 0.0637, "step": 841 }, { "epoch": 7.796296296296296, "grad_norm": 3.026418685913086, "learning_rate": 9.739130434782609e-05, "loss": 0.0923, "step": 842 }, { "epoch": 7.805555555555555, "grad_norm": 3.5097484588623047, "learning_rate": 9.726708074534162e-05, "loss": 0.1476, "step": 843 }, { "epoch": 7.814814814814815, "grad_norm": 1.7709219455718994, "learning_rate": 9.714285714285715e-05, "loss": 0.0732, "step": 844 }, { "epoch": 7.824074074074074, "grad_norm": 1.2974692583084106, "learning_rate": 9.701863354037267e-05, "loss": 0.0752, "step": 845 }, { "epoch": 7.833333333333333, "grad_norm": 0.6633884310722351, "learning_rate": 9.689440993788821e-05, "loss": 0.021, "step": 846 }, { "epoch": 7.842592592592593, "grad_norm": 2.9434750080108643, "learning_rate": 9.677018633540373e-05, "loss": 0.0862, "step": 847 }, { "epoch": 7.851851851851852, "grad_norm": 2.329890727996826, "learning_rate": 9.664596273291926e-05, "loss": 0.0689, "step": 848 }, { "epoch": 7.861111111111111, "grad_norm": 4.351027965545654, "learning_rate": 9.652173913043479e-05, "loss": 0.0473, "step": 849 }, { "epoch": 7.87037037037037, "grad_norm": 2.6241369247436523, "learning_rate": 9.63975155279503e-05, "loss": 0.0887, "step": 850 }, { "epoch": 7.87962962962963, "grad_norm": 1.7621710300445557, "learning_rate": 9.627329192546585e-05, "loss": 0.0686, "step": 851 }, { "epoch": 7.888888888888889, "grad_norm": 0.9145649075508118, "learning_rate": 9.614906832298138e-05, "loss": 0.0286, "step": 852 }, { "epoch": 7.898148148148148, "grad_norm": 5.4611101150512695, "learning_rate": 9.60248447204969e-05, "loss": 0.1921, "step": 853 }, { "epoch": 7.907407407407407, "grad_norm": 8.37093448638916, "learning_rate": 9.590062111801242e-05, "loss": 0.183, "step": 854 }, { "epoch": 7.916666666666667, "grad_norm": 1.575817346572876, "learning_rate": 9.577639751552795e-05, "loss": 0.0464, "step": 855 }, { "epoch": 7.925925925925926, "grad_norm": 5.839998722076416, "learning_rate": 9.565217391304348e-05, "loss": 0.1948, "step": 856 }, { "epoch": 7.935185185185185, "grad_norm": 3.813732624053955, "learning_rate": 9.552795031055901e-05, "loss": 0.1637, "step": 857 }, { "epoch": 7.944444444444445, "grad_norm": 6.688395023345947, "learning_rate": 9.540372670807454e-05, "loss": 0.1742, "step": 858 }, { "epoch": 7.953703703703704, "grad_norm": 1.701634168624878, "learning_rate": 9.527950310559006e-05, "loss": 0.0734, "step": 859 }, { "epoch": 7.962962962962963, "grad_norm": 6.329746246337891, "learning_rate": 9.51552795031056e-05, "loss": 0.4215, "step": 860 }, { "epoch": 7.972222222222222, "grad_norm": 1.4201380014419556, "learning_rate": 9.503105590062112e-05, "loss": 0.0552, "step": 861 }, { "epoch": 7.981481481481482, "grad_norm": 3.697819709777832, "learning_rate": 9.490683229813665e-05, "loss": 0.1865, "step": 862 }, { "epoch": 7.9907407407407405, "grad_norm": 2.9667913913726807, "learning_rate": 9.478260869565218e-05, "loss": 0.1707, "step": 863 }, { "epoch": 8.0, "grad_norm": 4.044425010681152, "learning_rate": 9.465838509316771e-05, "loss": 0.147, "step": 864 }, { "epoch": 8.00925925925926, "grad_norm": 3.552003860473633, "learning_rate": 9.453416149068324e-05, "loss": 0.2017, "step": 865 }, { "epoch": 8.018518518518519, "grad_norm": 1.0738625526428223, "learning_rate": 9.440993788819877e-05, "loss": 0.0394, "step": 866 }, { "epoch": 8.027777777777779, "grad_norm": 2.0488643646240234, "learning_rate": 9.428571428571429e-05, "loss": 0.0338, "step": 867 }, { "epoch": 8.037037037037036, "grad_norm": 0.6434540152549744, "learning_rate": 9.416149068322982e-05, "loss": 0.0203, "step": 868 }, { "epoch": 8.046296296296296, "grad_norm": 2.2168703079223633, "learning_rate": 9.403726708074534e-05, "loss": 0.0658, "step": 869 }, { "epoch": 8.055555555555555, "grad_norm": 4.93209171295166, "learning_rate": 9.391304347826087e-05, "loss": 0.0787, "step": 870 }, { "epoch": 8.064814814814815, "grad_norm": 1.0295863151550293, "learning_rate": 9.37888198757764e-05, "loss": 0.0422, "step": 871 }, { "epoch": 8.074074074074074, "grad_norm": 4.721346855163574, "learning_rate": 9.366459627329192e-05, "loss": 0.1582, "step": 872 }, { "epoch": 8.083333333333334, "grad_norm": 1.6571961641311646, "learning_rate": 9.354037267080745e-05, "loss": 0.0674, "step": 873 }, { "epoch": 8.092592592592593, "grad_norm": 1.5981183052062988, "learning_rate": 9.3416149068323e-05, "loss": 0.0582, "step": 874 }, { "epoch": 8.101851851851851, "grad_norm": 1.2692313194274902, "learning_rate": 9.329192546583851e-05, "loss": 0.0342, "step": 875 }, { "epoch": 8.11111111111111, "grad_norm": 1.5510468482971191, "learning_rate": 9.316770186335404e-05, "loss": 0.0303, "step": 876 }, { "epoch": 8.12037037037037, "grad_norm": 3.052006959915161, "learning_rate": 9.304347826086957e-05, "loss": 0.0613, "step": 877 }, { "epoch": 8.12962962962963, "grad_norm": 0.9419259428977966, "learning_rate": 9.29192546583851e-05, "loss": 0.0336, "step": 878 }, { "epoch": 8.13888888888889, "grad_norm": 1.4912816286087036, "learning_rate": 9.279503105590063e-05, "loss": 0.0579, "step": 879 }, { "epoch": 8.148148148148149, "grad_norm": 1.7151224613189697, "learning_rate": 9.267080745341615e-05, "loss": 0.0427, "step": 880 }, { "epoch": 8.157407407407407, "grad_norm": 2.1876168251037598, "learning_rate": 9.254658385093168e-05, "loss": 0.1089, "step": 881 }, { "epoch": 8.166666666666666, "grad_norm": 2.9372048377990723, "learning_rate": 9.24223602484472e-05, "loss": 0.0714, "step": 882 }, { "epoch": 8.175925925925926, "grad_norm": 1.4971363544464111, "learning_rate": 9.229813664596274e-05, "loss": 0.065, "step": 883 }, { "epoch": 8.185185185185185, "grad_norm": 0.9148528575897217, "learning_rate": 9.217391304347827e-05, "loss": 0.0339, "step": 884 }, { "epoch": 8.194444444444445, "grad_norm": 5.527726650238037, "learning_rate": 9.20496894409938e-05, "loss": 0.2355, "step": 885 }, { "epoch": 8.203703703703704, "grad_norm": 1.7691618204116821, "learning_rate": 9.192546583850931e-05, "loss": 0.0613, "step": 886 }, { "epoch": 8.212962962962964, "grad_norm": 5.282628536224365, "learning_rate": 9.180124223602486e-05, "loss": 0.2057, "step": 887 }, { "epoch": 8.222222222222221, "grad_norm": 4.678130626678467, "learning_rate": 9.167701863354037e-05, "loss": 0.0744, "step": 888 }, { "epoch": 8.231481481481481, "grad_norm": 1.7943284511566162, "learning_rate": 9.15527950310559e-05, "loss": 0.0748, "step": 889 }, { "epoch": 8.24074074074074, "grad_norm": 0.601938784122467, "learning_rate": 9.142857142857143e-05, "loss": 0.0185, "step": 890 }, { "epoch": 8.25, "grad_norm": 3.1283457279205322, "learning_rate": 9.130434782608696e-05, "loss": 0.0937, "step": 891 }, { "epoch": 8.25925925925926, "grad_norm": 3.6321136951446533, "learning_rate": 9.118012422360249e-05, "loss": 0.1203, "step": 892 }, { "epoch": 8.268518518518519, "grad_norm": 4.4610185623168945, "learning_rate": 9.105590062111802e-05, "loss": 0.2488, "step": 893 }, { "epoch": 8.277777777777779, "grad_norm": 1.2480530738830566, "learning_rate": 9.093167701863354e-05, "loss": 0.0595, "step": 894 }, { "epoch": 8.287037037037036, "grad_norm": 3.5497171878814697, "learning_rate": 9.080745341614907e-05, "loss": 0.1169, "step": 895 }, { "epoch": 8.296296296296296, "grad_norm": 1.2633136510849, "learning_rate": 9.068322981366461e-05, "loss": 0.0435, "step": 896 }, { "epoch": 8.305555555555555, "grad_norm": 0.9569624066352844, "learning_rate": 9.055900621118013e-05, "loss": 0.0457, "step": 897 }, { "epoch": 8.314814814814815, "grad_norm": 1.081705927848816, "learning_rate": 9.043478260869566e-05, "loss": 0.0537, "step": 898 }, { "epoch": 8.324074074074074, "grad_norm": 1.474846601486206, "learning_rate": 9.031055900621119e-05, "loss": 0.0465, "step": 899 }, { "epoch": 8.333333333333334, "grad_norm": 1.3447438478469849, "learning_rate": 9.01863354037267e-05, "loss": 0.0534, "step": 900 }, { "epoch": 8.342592592592593, "grad_norm": 4.491641998291016, "learning_rate": 9.006211180124225e-05, "loss": 0.1298, "step": 901 }, { "epoch": 8.351851851851851, "grad_norm": 0.8619861602783203, "learning_rate": 8.993788819875776e-05, "loss": 0.0237, "step": 902 }, { "epoch": 8.36111111111111, "grad_norm": 1.2432044744491577, "learning_rate": 8.98136645962733e-05, "loss": 0.0499, "step": 903 }, { "epoch": 8.37037037037037, "grad_norm": 2.284329891204834, "learning_rate": 8.968944099378882e-05, "loss": 0.0876, "step": 904 }, { "epoch": 8.37962962962963, "grad_norm": 4.381556034088135, "learning_rate": 8.956521739130435e-05, "loss": 0.1647, "step": 905 }, { "epoch": 8.38888888888889, "grad_norm": 1.341193675994873, "learning_rate": 8.944099378881988e-05, "loss": 0.0628, "step": 906 }, { "epoch": 8.398148148148149, "grad_norm": 1.5119465589523315, "learning_rate": 8.931677018633541e-05, "loss": 0.0429, "step": 907 }, { "epoch": 8.407407407407407, "grad_norm": 1.5674864053726196, "learning_rate": 8.919254658385093e-05, "loss": 0.066, "step": 908 }, { "epoch": 8.416666666666666, "grad_norm": 2.23675537109375, "learning_rate": 8.906832298136646e-05, "loss": 0.0567, "step": 909 }, { "epoch": 8.425925925925926, "grad_norm": 1.609828233718872, "learning_rate": 8.894409937888199e-05, "loss": 0.053, "step": 910 }, { "epoch": 8.435185185185185, "grad_norm": 2.384761095046997, "learning_rate": 8.881987577639752e-05, "loss": 0.0779, "step": 911 }, { "epoch": 8.444444444444445, "grad_norm": 1.0519819259643555, "learning_rate": 8.869565217391305e-05, "loss": 0.0365, "step": 912 }, { "epoch": 8.453703703703704, "grad_norm": 2.049403429031372, "learning_rate": 8.857142857142857e-05, "loss": 0.0595, "step": 913 }, { "epoch": 8.462962962962964, "grad_norm": 3.0430166721343994, "learning_rate": 8.844720496894411e-05, "loss": 0.1162, "step": 914 }, { "epoch": 8.472222222222221, "grad_norm": 2.900303602218628, "learning_rate": 8.832298136645964e-05, "loss": 0.0528, "step": 915 }, { "epoch": 8.481481481481481, "grad_norm": 3.444995164871216, "learning_rate": 8.819875776397516e-05, "loss": 0.0581, "step": 916 }, { "epoch": 8.49074074074074, "grad_norm": 3.1520862579345703, "learning_rate": 8.807453416149069e-05, "loss": 0.0796, "step": 917 }, { "epoch": 8.5, "grad_norm": 2.5778634548187256, "learning_rate": 8.795031055900622e-05, "loss": 0.0574, "step": 918 }, { "epoch": 8.50925925925926, "grad_norm": 1.2131502628326416, "learning_rate": 8.782608695652174e-05, "loss": 0.0589, "step": 919 }, { "epoch": 8.518518518518519, "grad_norm": 1.663877248764038, "learning_rate": 8.770186335403727e-05, "loss": 0.0636, "step": 920 }, { "epoch": 8.527777777777779, "grad_norm": 1.3171827793121338, "learning_rate": 8.757763975155279e-05, "loss": 0.0412, "step": 921 }, { "epoch": 8.537037037037036, "grad_norm": 3.0393033027648926, "learning_rate": 8.745341614906832e-05, "loss": 0.0926, "step": 922 }, { "epoch": 8.546296296296296, "grad_norm": 3.5446856021881104, "learning_rate": 8.732919254658386e-05, "loss": 0.0902, "step": 923 }, { "epoch": 8.555555555555555, "grad_norm": 4.914732933044434, "learning_rate": 8.720496894409938e-05, "loss": 0.1173, "step": 924 }, { "epoch": 8.564814814814815, "grad_norm": 6.190938949584961, "learning_rate": 8.708074534161491e-05, "loss": 0.1385, "step": 925 }, { "epoch": 8.574074074074074, "grad_norm": 1.2483842372894287, "learning_rate": 8.695652173913044e-05, "loss": 0.0517, "step": 926 }, { "epoch": 8.583333333333334, "grad_norm": 2.203432083129883, "learning_rate": 8.683229813664596e-05, "loss": 0.1012, "step": 927 }, { "epoch": 8.592592592592592, "grad_norm": 1.5815249681472778, "learning_rate": 8.67080745341615e-05, "loss": 0.0935, "step": 928 }, { "epoch": 8.601851851851851, "grad_norm": 6.097355842590332, "learning_rate": 8.658385093167703e-05, "loss": 0.1332, "step": 929 }, { "epoch": 8.61111111111111, "grad_norm": 1.781144618988037, "learning_rate": 8.645962732919255e-05, "loss": 0.0739, "step": 930 }, { "epoch": 8.62037037037037, "grad_norm": 10.972087860107422, "learning_rate": 8.633540372670808e-05, "loss": 0.2456, "step": 931 }, { "epoch": 8.62962962962963, "grad_norm": 2.080036163330078, "learning_rate": 8.62111801242236e-05, "loss": 0.0516, "step": 932 }, { "epoch": 8.63888888888889, "grad_norm": 3.2868905067443848, "learning_rate": 8.608695652173914e-05, "loss": 0.0682, "step": 933 }, { "epoch": 8.648148148148149, "grad_norm": 1.0348496437072754, "learning_rate": 8.596273291925467e-05, "loss": 0.0435, "step": 934 }, { "epoch": 8.657407407407408, "grad_norm": 2.882920026779175, "learning_rate": 8.583850931677018e-05, "loss": 0.0909, "step": 935 }, { "epoch": 8.666666666666666, "grad_norm": 2.1770124435424805, "learning_rate": 8.571428571428571e-05, "loss": 0.0953, "step": 936 }, { "epoch": 8.675925925925926, "grad_norm": 1.0724560022354126, "learning_rate": 8.559006211180126e-05, "loss": 0.0421, "step": 937 }, { "epoch": 8.685185185185185, "grad_norm": 1.1822600364685059, "learning_rate": 8.546583850931677e-05, "loss": 0.0382, "step": 938 }, { "epoch": 8.694444444444445, "grad_norm": 1.566285490989685, "learning_rate": 8.53416149068323e-05, "loss": 0.0548, "step": 939 }, { "epoch": 8.703703703703704, "grad_norm": 3.2085978984832764, "learning_rate": 8.521739130434783e-05, "loss": 0.0455, "step": 940 }, { "epoch": 8.712962962962964, "grad_norm": 1.2845832109451294, "learning_rate": 8.509316770186336e-05, "loss": 0.0439, "step": 941 }, { "epoch": 8.722222222222221, "grad_norm": 1.9509384632110596, "learning_rate": 8.496894409937889e-05, "loss": 0.0869, "step": 942 }, { "epoch": 8.731481481481481, "grad_norm": 1.4507602453231812, "learning_rate": 8.484472049689441e-05, "loss": 0.0512, "step": 943 }, { "epoch": 8.74074074074074, "grad_norm": 5.29793119430542, "learning_rate": 8.472049689440994e-05, "loss": 0.1188, "step": 944 }, { "epoch": 8.75, "grad_norm": 1.0561885833740234, "learning_rate": 8.459627329192547e-05, "loss": 0.0446, "step": 945 }, { "epoch": 8.75925925925926, "grad_norm": 0.9933180212974548, "learning_rate": 8.4472049689441e-05, "loss": 0.0251, "step": 946 }, { "epoch": 8.768518518518519, "grad_norm": 8.12339973449707, "learning_rate": 8.434782608695653e-05, "loss": 0.0529, "step": 947 }, { "epoch": 8.777777777777779, "grad_norm": 1.750126600265503, "learning_rate": 8.422360248447206e-05, "loss": 0.0537, "step": 948 }, { "epoch": 8.787037037037036, "grad_norm": 2.373239278793335, "learning_rate": 8.409937888198757e-05, "loss": 0.1339, "step": 949 }, { "epoch": 8.796296296296296, "grad_norm": 1.7778215408325195, "learning_rate": 8.397515527950312e-05, "loss": 0.0794, "step": 950 }, { "epoch": 8.805555555555555, "grad_norm": 4.6339430809021, "learning_rate": 8.385093167701863e-05, "loss": 0.0935, "step": 951 }, { "epoch": 8.814814814814815, "grad_norm": 2.691041946411133, "learning_rate": 8.372670807453416e-05, "loss": 0.0981, "step": 952 }, { "epoch": 8.824074074074074, "grad_norm": 1.270849347114563, "learning_rate": 8.36024844720497e-05, "loss": 0.0512, "step": 953 }, { "epoch": 8.833333333333334, "grad_norm": 0.8391767144203186, "learning_rate": 8.347826086956521e-05, "loss": 0.0233, "step": 954 }, { "epoch": 8.842592592592592, "grad_norm": 2.3432822227478027, "learning_rate": 8.335403726708075e-05, "loss": 0.0834, "step": 955 }, { "epoch": 8.851851851851851, "grad_norm": 1.2851802110671997, "learning_rate": 8.322981366459628e-05, "loss": 0.0468, "step": 956 }, { "epoch": 8.86111111111111, "grad_norm": 1.9697790145874023, "learning_rate": 8.31055900621118e-05, "loss": 0.0759, "step": 957 }, { "epoch": 8.87037037037037, "grad_norm": 2.5893197059631348, "learning_rate": 8.298136645962733e-05, "loss": 0.1465, "step": 958 }, { "epoch": 8.87962962962963, "grad_norm": 2.174973249435425, "learning_rate": 8.285714285714287e-05, "loss": 0.0834, "step": 959 }, { "epoch": 8.88888888888889, "grad_norm": 0.8256950974464417, "learning_rate": 8.273291925465839e-05, "loss": 0.0349, "step": 960 }, { "epoch": 8.898148148148149, "grad_norm": 1.0629161596298218, "learning_rate": 8.260869565217392e-05, "loss": 0.0447, "step": 961 }, { "epoch": 8.907407407407408, "grad_norm": 1.6304731369018555, "learning_rate": 8.248447204968945e-05, "loss": 0.0446, "step": 962 }, { "epoch": 8.916666666666666, "grad_norm": 4.195140361785889, "learning_rate": 8.236024844720497e-05, "loss": 0.1002, "step": 963 }, { "epoch": 8.925925925925926, "grad_norm": 0.7746004462242126, "learning_rate": 8.223602484472051e-05, "loss": 0.0266, "step": 964 }, { "epoch": 8.935185185185185, "grad_norm": 0.6275144815444946, "learning_rate": 8.211180124223603e-05, "loss": 0.0238, "step": 965 }, { "epoch": 8.944444444444445, "grad_norm": 2.2046384811401367, "learning_rate": 8.198757763975156e-05, "loss": 0.0676, "step": 966 }, { "epoch": 8.953703703703704, "grad_norm": 4.583156585693359, "learning_rate": 8.186335403726709e-05, "loss": 0.0885, "step": 967 }, { "epoch": 8.962962962962964, "grad_norm": 3.3135993480682373, "learning_rate": 8.173913043478262e-05, "loss": 0.0971, "step": 968 }, { "epoch": 8.972222222222221, "grad_norm": 1.8723350763320923, "learning_rate": 8.161490683229814e-05, "loss": 0.0651, "step": 969 }, { "epoch": 8.981481481481481, "grad_norm": 1.9317792654037476, "learning_rate": 8.149068322981367e-05, "loss": 0.0656, "step": 970 }, { "epoch": 8.99074074074074, "grad_norm": 1.3685884475708008, "learning_rate": 8.136645962732919e-05, "loss": 0.0618, "step": 971 }, { "epoch": 9.0, "grad_norm": 1.9626964330673218, "learning_rate": 8.124223602484472e-05, "loss": 0.0594, "step": 972 }, { "epoch": 9.00925925925926, "grad_norm": 2.1750247478485107, "learning_rate": 8.111801242236025e-05, "loss": 0.0503, "step": 973 }, { "epoch": 9.018518518518519, "grad_norm": 1.803802251815796, "learning_rate": 8.099378881987578e-05, "loss": 0.043, "step": 974 }, { "epoch": 9.027777777777779, "grad_norm": 3.9391958713531494, "learning_rate": 8.086956521739131e-05, "loss": 0.0847, "step": 975 }, { "epoch": 9.037037037037036, "grad_norm": 1.5313109159469604, "learning_rate": 8.074534161490683e-05, "loss": 0.0599, "step": 976 }, { "epoch": 9.046296296296296, "grad_norm": 3.568204164505005, "learning_rate": 8.062111801242236e-05, "loss": 0.1605, "step": 977 }, { "epoch": 9.055555555555555, "grad_norm": 0.7717824578285217, "learning_rate": 8.04968944099379e-05, "loss": 0.0243, "step": 978 }, { "epoch": 9.064814814814815, "grad_norm": 1.0113966464996338, "learning_rate": 8.037267080745342e-05, "loss": 0.0401, "step": 979 }, { "epoch": 9.074074074074074, "grad_norm": 1.8266679048538208, "learning_rate": 8.024844720496895e-05, "loss": 0.0533, "step": 980 }, { "epoch": 9.083333333333334, "grad_norm": 2.636436700820923, "learning_rate": 8.012422360248448e-05, "loss": 0.0923, "step": 981 }, { "epoch": 9.092592592592593, "grad_norm": 2.6462342739105225, "learning_rate": 8e-05, "loss": 0.0497, "step": 982 }, { "epoch": 9.101851851851851, "grad_norm": 1.8009244203567505, "learning_rate": 7.987577639751554e-05, "loss": 0.0718, "step": 983 }, { "epoch": 9.11111111111111, "grad_norm": 1.0373586416244507, "learning_rate": 7.975155279503105e-05, "loss": 0.0269, "step": 984 }, { "epoch": 9.12037037037037, "grad_norm": 1.1197093725204468, "learning_rate": 7.962732919254658e-05, "loss": 0.0375, "step": 985 }, { "epoch": 9.12962962962963, "grad_norm": 0.7301254272460938, "learning_rate": 7.950310559006211e-05, "loss": 0.0335, "step": 986 }, { "epoch": 9.13888888888889, "grad_norm": 1.0725760459899902, "learning_rate": 7.937888198757764e-05, "loss": 0.0392, "step": 987 }, { "epoch": 9.148148148148149, "grad_norm": 0.7969951033592224, "learning_rate": 7.925465838509317e-05, "loss": 0.0259, "step": 988 }, { "epoch": 9.157407407407407, "grad_norm": 1.0825790166854858, "learning_rate": 7.91304347826087e-05, "loss": 0.0346, "step": 989 }, { "epoch": 9.166666666666666, "grad_norm": 2.6907694339752197, "learning_rate": 7.900621118012422e-05, "loss": 0.0462, "step": 990 }, { "epoch": 9.175925925925926, "grad_norm": 1.146231770515442, "learning_rate": 7.888198757763976e-05, "loss": 0.0422, "step": 991 }, { "epoch": 9.185185185185185, "grad_norm": 0.8455992341041565, "learning_rate": 7.875776397515529e-05, "loss": 0.0356, "step": 992 }, { "epoch": 9.194444444444445, "grad_norm": 1.0626360177993774, "learning_rate": 7.863354037267081e-05, "loss": 0.0415, "step": 993 }, { "epoch": 9.203703703703704, "grad_norm": 1.5166810750961304, "learning_rate": 7.850931677018634e-05, "loss": 0.0477, "step": 994 }, { "epoch": 9.212962962962964, "grad_norm": 3.8809115886688232, "learning_rate": 7.838509316770187e-05, "loss": 0.0976, "step": 995 }, { "epoch": 9.222222222222221, "grad_norm": 2.389176368713379, "learning_rate": 7.82608695652174e-05, "loss": 0.0649, "step": 996 }, { "epoch": 9.231481481481481, "grad_norm": 0.7285131812095642, "learning_rate": 7.813664596273293e-05, "loss": 0.0192, "step": 997 }, { "epoch": 9.24074074074074, "grad_norm": 2.0538992881774902, "learning_rate": 7.801242236024844e-05, "loss": 0.0743, "step": 998 }, { "epoch": 9.25, "grad_norm": 3.144094944000244, "learning_rate": 7.788819875776397e-05, "loss": 0.0677, "step": 999 }, { "epoch": 9.25925925925926, "grad_norm": 1.6257270574569702, "learning_rate": 7.776397515527952e-05, "loss": 0.0739, "step": 1000 }, { "epoch": 9.268518518518519, "grad_norm": 3.6802332401275635, "learning_rate": 7.763975155279503e-05, "loss": 0.1243, "step": 1001 }, { "epoch": 9.277777777777779, "grad_norm": 0.8263121247291565, "learning_rate": 7.751552795031056e-05, "loss": 0.0376, "step": 1002 }, { "epoch": 9.287037037037036, "grad_norm": 2.52786922454834, "learning_rate": 7.73913043478261e-05, "loss": 0.0423, "step": 1003 }, { "epoch": 9.296296296296296, "grad_norm": 1.0051758289337158, "learning_rate": 7.726708074534161e-05, "loss": 0.0281, "step": 1004 }, { "epoch": 9.305555555555555, "grad_norm": 3.8244469165802, "learning_rate": 7.714285714285715e-05, "loss": 0.1488, "step": 1005 }, { "epoch": 9.314814814814815, "grad_norm": 5.202383518218994, "learning_rate": 7.701863354037267e-05, "loss": 0.1224, "step": 1006 }, { "epoch": 9.324074074074074, "grad_norm": 0.9186612963676453, "learning_rate": 7.68944099378882e-05, "loss": 0.0289, "step": 1007 }, { "epoch": 9.333333333333334, "grad_norm": 1.926774263381958, "learning_rate": 7.677018633540373e-05, "loss": 0.0516, "step": 1008 }, { "epoch": 9.342592592592593, "grad_norm": 4.624532699584961, "learning_rate": 7.664596273291926e-05, "loss": 0.2248, "step": 1009 }, { "epoch": 9.351851851851851, "grad_norm": 1.0460689067840576, "learning_rate": 7.652173913043479e-05, "loss": 0.0344, "step": 1010 }, { "epoch": 9.36111111111111, "grad_norm": 1.524003267288208, "learning_rate": 7.639751552795032e-05, "loss": 0.0372, "step": 1011 }, { "epoch": 9.37037037037037, "grad_norm": 2.419135570526123, "learning_rate": 7.627329192546584e-05, "loss": 0.0545, "step": 1012 }, { "epoch": 9.37962962962963, "grad_norm": 2.654794454574585, "learning_rate": 7.614906832298137e-05, "loss": 0.0649, "step": 1013 }, { "epoch": 9.38888888888889, "grad_norm": 2.0655901432037354, "learning_rate": 7.60248447204969e-05, "loss": 0.0731, "step": 1014 }, { "epoch": 9.398148148148149, "grad_norm": 2.0986382961273193, "learning_rate": 7.590062111801243e-05, "loss": 0.073, "step": 1015 }, { "epoch": 9.407407407407407, "grad_norm": 3.522204637527466, "learning_rate": 7.577639751552796e-05, "loss": 0.0648, "step": 1016 }, { "epoch": 9.416666666666666, "grad_norm": 1.7459393739700317, "learning_rate": 7.565217391304347e-05, "loss": 0.0258, "step": 1017 }, { "epoch": 9.425925925925926, "grad_norm": 2.7577157020568848, "learning_rate": 7.552795031055902e-05, "loss": 0.0928, "step": 1018 }, { "epoch": 9.435185185185185, "grad_norm": 2.878016710281372, "learning_rate": 7.540372670807454e-05, "loss": 0.0556, "step": 1019 }, { "epoch": 9.444444444444445, "grad_norm": 2.508056640625, "learning_rate": 7.527950310559006e-05, "loss": 0.1017, "step": 1020 }, { "epoch": 9.453703703703704, "grad_norm": 2.9259493350982666, "learning_rate": 7.515527950310559e-05, "loss": 0.1015, "step": 1021 }, { "epoch": 9.462962962962964, "grad_norm": 1.2046016454696655, "learning_rate": 7.503105590062112e-05, "loss": 0.052, "step": 1022 }, { "epoch": 9.472222222222221, "grad_norm": 0.83958500623703, "learning_rate": 7.490683229813665e-05, "loss": 0.0264, "step": 1023 }, { "epoch": 9.481481481481481, "grad_norm": 2.824143171310425, "learning_rate": 7.478260869565218e-05, "loss": 0.0801, "step": 1024 }, { "epoch": 9.49074074074074, "grad_norm": 3.1509366035461426, "learning_rate": 7.465838509316771e-05, "loss": 0.0832, "step": 1025 }, { "epoch": 9.5, "grad_norm": 2.522878885269165, "learning_rate": 7.453416149068323e-05, "loss": 0.0588, "step": 1026 }, { "epoch": 9.50925925925926, "grad_norm": 0.79912930727005, "learning_rate": 7.440993788819877e-05, "loss": 0.0423, "step": 1027 }, { "epoch": 9.518518518518519, "grad_norm": 2.4580273628234863, "learning_rate": 7.428571428571429e-05, "loss": 0.0771, "step": 1028 }, { "epoch": 9.527777777777779, "grad_norm": 2.6916251182556152, "learning_rate": 7.416149068322982e-05, "loss": 0.0663, "step": 1029 }, { "epoch": 9.537037037037036, "grad_norm": 0.7641211748123169, "learning_rate": 7.403726708074535e-05, "loss": 0.0276, "step": 1030 }, { "epoch": 9.546296296296296, "grad_norm": 2.337360382080078, "learning_rate": 7.391304347826086e-05, "loss": 0.0539, "step": 1031 }, { "epoch": 9.555555555555555, "grad_norm": 1.247176170349121, "learning_rate": 7.37888198757764e-05, "loss": 0.0633, "step": 1032 }, { "epoch": 9.564814814814815, "grad_norm": 0.9180092811584473, "learning_rate": 7.366459627329194e-05, "loss": 0.0353, "step": 1033 }, { "epoch": 9.574074074074074, "grad_norm": 3.825732707977295, "learning_rate": 7.354037267080745e-05, "loss": 0.0991, "step": 1034 }, { "epoch": 9.583333333333334, "grad_norm": 4.522989273071289, "learning_rate": 7.341614906832298e-05, "loss": 0.3405, "step": 1035 }, { "epoch": 9.592592592592592, "grad_norm": 4.249613285064697, "learning_rate": 7.329192546583851e-05, "loss": 0.1238, "step": 1036 }, { "epoch": 9.601851851851851, "grad_norm": 1.2015916109085083, "learning_rate": 7.316770186335404e-05, "loss": 0.0475, "step": 1037 }, { "epoch": 9.61111111111111, "grad_norm": 2.1656363010406494, "learning_rate": 7.304347826086957e-05, "loss": 0.0668, "step": 1038 }, { "epoch": 9.62037037037037, "grad_norm": 1.3153462409973145, "learning_rate": 7.291925465838509e-05, "loss": 0.0747, "step": 1039 }, { "epoch": 9.62962962962963, "grad_norm": 0.8693362474441528, "learning_rate": 7.279503105590062e-05, "loss": 0.0263, "step": 1040 }, { "epoch": 9.63888888888889, "grad_norm": 0.9345487952232361, "learning_rate": 7.267080745341616e-05, "loss": 0.031, "step": 1041 }, { "epoch": 9.648148148148149, "grad_norm": 3.389455795288086, "learning_rate": 7.254658385093168e-05, "loss": 0.1584, "step": 1042 }, { "epoch": 9.657407407407408, "grad_norm": 2.5855634212493896, "learning_rate": 7.242236024844721e-05, "loss": 0.0557, "step": 1043 }, { "epoch": 9.666666666666666, "grad_norm": 1.2336293458938599, "learning_rate": 7.229813664596274e-05, "loss": 0.0438, "step": 1044 }, { "epoch": 9.675925925925926, "grad_norm": 4.778486728668213, "learning_rate": 7.217391304347827e-05, "loss": 0.2042, "step": 1045 }, { "epoch": 9.685185185185185, "grad_norm": 1.5163942575454712, "learning_rate": 7.20496894409938e-05, "loss": 0.0564, "step": 1046 }, { "epoch": 9.694444444444445, "grad_norm": 0.900606632232666, "learning_rate": 7.192546583850931e-05, "loss": 0.0381, "step": 1047 }, { "epoch": 9.703703703703704, "grad_norm": 1.1757590770721436, "learning_rate": 7.180124223602484e-05, "loss": 0.0412, "step": 1048 }, { "epoch": 9.712962962962964, "grad_norm": 3.0201354026794434, "learning_rate": 7.167701863354037e-05, "loss": 0.0674, "step": 1049 }, { "epoch": 9.722222222222221, "grad_norm": 1.2021557092666626, "learning_rate": 7.15527950310559e-05, "loss": 0.0485, "step": 1050 }, { "epoch": 9.731481481481481, "grad_norm": 0.8072912096977234, "learning_rate": 7.142857142857143e-05, "loss": 0.0317, "step": 1051 }, { "epoch": 9.74074074074074, "grad_norm": 3.1552536487579346, "learning_rate": 7.130434782608696e-05, "loss": 0.0777, "step": 1052 }, { "epoch": 9.75, "grad_norm": 1.6829980611801147, "learning_rate": 7.118012422360248e-05, "loss": 0.1083, "step": 1053 }, { "epoch": 9.75925925925926, "grad_norm": 1.2367135286331177, "learning_rate": 7.105590062111802e-05, "loss": 0.061, "step": 1054 }, { "epoch": 9.768518518518519, "grad_norm": 2.295994997024536, "learning_rate": 7.093167701863354e-05, "loss": 0.1145, "step": 1055 }, { "epoch": 9.777777777777779, "grad_norm": 1.0769095420837402, "learning_rate": 7.080745341614907e-05, "loss": 0.0448, "step": 1056 }, { "epoch": 9.787037037037036, "grad_norm": 0.814005434513092, "learning_rate": 7.06832298136646e-05, "loss": 0.0279, "step": 1057 }, { "epoch": 9.796296296296296, "grad_norm": 2.416013240814209, "learning_rate": 7.055900621118013e-05, "loss": 0.1144, "step": 1058 }, { "epoch": 9.805555555555555, "grad_norm": 1.4185346364974976, "learning_rate": 7.043478260869566e-05, "loss": 0.0632, "step": 1059 }, { "epoch": 9.814814814814815, "grad_norm": 1.7376627922058105, "learning_rate": 7.031055900621119e-05, "loss": 0.0452, "step": 1060 }, { "epoch": 9.824074074074074, "grad_norm": 0.911581814289093, "learning_rate": 7.01863354037267e-05, "loss": 0.0404, "step": 1061 }, { "epoch": 9.833333333333334, "grad_norm": 0.774612545967102, "learning_rate": 7.006211180124224e-05, "loss": 0.0418, "step": 1062 }, { "epoch": 9.842592592592592, "grad_norm": 4.290308475494385, "learning_rate": 6.993788819875778e-05, "loss": 0.1378, "step": 1063 }, { "epoch": 9.851851851851851, "grad_norm": 2.3391199111938477, "learning_rate": 6.98136645962733e-05, "loss": 0.075, "step": 1064 }, { "epoch": 9.86111111111111, "grad_norm": 2.517441511154175, "learning_rate": 6.968944099378883e-05, "loss": 0.0692, "step": 1065 }, { "epoch": 9.87037037037037, "grad_norm": 2.425095558166504, "learning_rate": 6.956521739130436e-05, "loss": 0.1089, "step": 1066 }, { "epoch": 9.87962962962963, "grad_norm": 1.5073820352554321, "learning_rate": 6.944099378881987e-05, "loss": 0.045, "step": 1067 }, { "epoch": 9.88888888888889, "grad_norm": 1.071384072303772, "learning_rate": 6.931677018633542e-05, "loss": 0.0336, "step": 1068 }, { "epoch": 9.898148148148149, "grad_norm": 1.1007689237594604, "learning_rate": 6.919254658385093e-05, "loss": 0.0302, "step": 1069 }, { "epoch": 9.907407407407408, "grad_norm": 4.436883926391602, "learning_rate": 6.906832298136646e-05, "loss": 0.0853, "step": 1070 }, { "epoch": 9.916666666666666, "grad_norm": 0.7881323099136353, "learning_rate": 6.894409937888199e-05, "loss": 0.0275, "step": 1071 }, { "epoch": 9.925925925925926, "grad_norm": 4.63049840927124, "learning_rate": 6.881987577639752e-05, "loss": 0.1019, "step": 1072 }, { "epoch": 9.935185185185185, "grad_norm": 1.3890912532806396, "learning_rate": 6.869565217391305e-05, "loss": 0.0554, "step": 1073 }, { "epoch": 9.944444444444445, "grad_norm": 1.3758798837661743, "learning_rate": 6.857142857142858e-05, "loss": 0.0474, "step": 1074 }, { "epoch": 9.953703703703704, "grad_norm": 3.138035774230957, "learning_rate": 6.84472049689441e-05, "loss": 0.1148, "step": 1075 }, { "epoch": 9.962962962962964, "grad_norm": 3.5089962482452393, "learning_rate": 6.832298136645963e-05, "loss": 0.1058, "step": 1076 }, { "epoch": 9.972222222222221, "grad_norm": 2.708378314971924, "learning_rate": 6.819875776397516e-05, "loss": 0.1467, "step": 1077 }, { "epoch": 9.981481481481481, "grad_norm": 1.7580418586730957, "learning_rate": 6.807453416149069e-05, "loss": 0.1067, "step": 1078 }, { "epoch": 9.99074074074074, "grad_norm": 1.5734944343566895, "learning_rate": 6.795031055900622e-05, "loss": 0.0604, "step": 1079 }, { "epoch": 10.0, "grad_norm": 1.7437562942504883, "learning_rate": 6.782608695652173e-05, "loss": 0.078, "step": 1080 }, { "epoch": 10.00925925925926, "grad_norm": 1.2374075651168823, "learning_rate": 6.770186335403726e-05, "loss": 0.0517, "step": 1081 }, { "epoch": 10.018518518518519, "grad_norm": 3.730271816253662, "learning_rate": 6.75776397515528e-05, "loss": 0.0712, "step": 1082 }, { "epoch": 10.027777777777779, "grad_norm": 1.618425965309143, "learning_rate": 6.745341614906832e-05, "loss": 0.0471, "step": 1083 }, { "epoch": 10.037037037037036, "grad_norm": 2.101398468017578, "learning_rate": 6.732919254658385e-05, "loss": 0.0438, "step": 1084 }, { "epoch": 10.046296296296296, "grad_norm": 2.635305166244507, "learning_rate": 6.720496894409938e-05, "loss": 0.1445, "step": 1085 }, { "epoch": 10.055555555555555, "grad_norm": 1.2977538108825684, "learning_rate": 6.708074534161491e-05, "loss": 0.0367, "step": 1086 }, { "epoch": 10.064814814814815, "grad_norm": 2.383920192718506, "learning_rate": 6.695652173913044e-05, "loss": 0.066, "step": 1087 }, { "epoch": 10.074074074074074, "grad_norm": 2.3182485103607178, "learning_rate": 6.683229813664596e-05, "loss": 0.0399, "step": 1088 }, { "epoch": 10.083333333333334, "grad_norm": 2.031402826309204, "learning_rate": 6.670807453416149e-05, "loss": 0.037, "step": 1089 }, { "epoch": 10.092592592592593, "grad_norm": 4.036484718322754, "learning_rate": 6.658385093167702e-05, "loss": 0.0377, "step": 1090 }, { "epoch": 10.101851851851851, "grad_norm": 2.554116725921631, "learning_rate": 6.645962732919255e-05, "loss": 0.1237, "step": 1091 }, { "epoch": 10.11111111111111, "grad_norm": 2.1571269035339355, "learning_rate": 6.633540372670808e-05, "loss": 0.0415, "step": 1092 }, { "epoch": 10.12037037037037, "grad_norm": 0.994016706943512, "learning_rate": 6.621118012422361e-05, "loss": 0.0474, "step": 1093 }, { "epoch": 10.12962962962963, "grad_norm": 0.7662463784217834, "learning_rate": 6.608695652173912e-05, "loss": 0.0263, "step": 1094 }, { "epoch": 10.13888888888889, "grad_norm": 1.1817699670791626, "learning_rate": 6.596273291925467e-05, "loss": 0.0419, "step": 1095 }, { "epoch": 10.148148148148149, "grad_norm": 1.2163158655166626, "learning_rate": 6.58385093167702e-05, "loss": 0.0463, "step": 1096 }, { "epoch": 10.157407407407407, "grad_norm": 0.646700382232666, "learning_rate": 6.571428571428571e-05, "loss": 0.0178, "step": 1097 }, { "epoch": 10.166666666666666, "grad_norm": 0.7059783935546875, "learning_rate": 6.559006211180124e-05, "loss": 0.0289, "step": 1098 }, { "epoch": 10.175925925925926, "grad_norm": 4.205524921417236, "learning_rate": 6.546583850931677e-05, "loss": 0.1329, "step": 1099 }, { "epoch": 10.185185185185185, "grad_norm": 1.418074131011963, "learning_rate": 6.53416149068323e-05, "loss": 0.0412, "step": 1100 }, { "epoch": 10.194444444444445, "grad_norm": 0.9323771595954895, "learning_rate": 6.521739130434783e-05, "loss": 0.043, "step": 1101 }, { "epoch": 10.203703703703704, "grad_norm": 1.6822084188461304, "learning_rate": 6.509316770186335e-05, "loss": 0.0402, "step": 1102 }, { "epoch": 10.212962962962964, "grad_norm": 1.1018824577331543, "learning_rate": 6.496894409937888e-05, "loss": 0.0424, "step": 1103 }, { "epoch": 10.222222222222221, "grad_norm": 0.921039342880249, "learning_rate": 6.484472049689442e-05, "loss": 0.0251, "step": 1104 }, { "epoch": 10.231481481481481, "grad_norm": 2.9697976112365723, "learning_rate": 6.472049689440994e-05, "loss": 0.0627, "step": 1105 }, { "epoch": 10.24074074074074, "grad_norm": 1.5444436073303223, "learning_rate": 6.459627329192547e-05, "loss": 0.0361, "step": 1106 }, { "epoch": 10.25, "grad_norm": 1.545248031616211, "learning_rate": 6.4472049689441e-05, "loss": 0.0571, "step": 1107 }, { "epoch": 10.25925925925926, "grad_norm": 2.4655678272247314, "learning_rate": 6.434782608695652e-05, "loss": 0.0929, "step": 1108 }, { "epoch": 10.268518518518519, "grad_norm": 1.0610233545303345, "learning_rate": 6.422360248447206e-05, "loss": 0.0431, "step": 1109 }, { "epoch": 10.277777777777779, "grad_norm": 1.7690080404281616, "learning_rate": 6.409937888198758e-05, "loss": 0.0542, "step": 1110 }, { "epoch": 10.287037037037036, "grad_norm": 1.6236170530319214, "learning_rate": 6.39751552795031e-05, "loss": 0.0465, "step": 1111 }, { "epoch": 10.296296296296296, "grad_norm": 2.925307035446167, "learning_rate": 6.385093167701864e-05, "loss": 0.0608, "step": 1112 }, { "epoch": 10.305555555555555, "grad_norm": 1.1506783962249756, "learning_rate": 6.372670807453417e-05, "loss": 0.0494, "step": 1113 }, { "epoch": 10.314814814814815, "grad_norm": 1.6719893217086792, "learning_rate": 6.36024844720497e-05, "loss": 0.0527, "step": 1114 }, { "epoch": 10.324074074074074, "grad_norm": 1.290839433670044, "learning_rate": 6.347826086956523e-05, "loss": 0.0655, "step": 1115 }, { "epoch": 10.333333333333334, "grad_norm": 2.2520599365234375, "learning_rate": 6.335403726708074e-05, "loss": 0.0639, "step": 1116 }, { "epoch": 10.342592592592593, "grad_norm": 7.430666446685791, "learning_rate": 6.322981366459627e-05, "loss": 0.0843, "step": 1117 }, { "epoch": 10.351851851851851, "grad_norm": 1.3490774631500244, "learning_rate": 6.31055900621118e-05, "loss": 0.0582, "step": 1118 }, { "epoch": 10.36111111111111, "grad_norm": 2.280447244644165, "learning_rate": 6.298136645962733e-05, "loss": 0.0675, "step": 1119 }, { "epoch": 10.37037037037037, "grad_norm": 0.784392237663269, "learning_rate": 6.285714285714286e-05, "loss": 0.0369, "step": 1120 }, { "epoch": 10.37962962962963, "grad_norm": 2.3288660049438477, "learning_rate": 6.273291925465838e-05, "loss": 0.0548, "step": 1121 }, { "epoch": 10.38888888888889, "grad_norm": 0.6705870628356934, "learning_rate": 6.260869565217392e-05, "loss": 0.0257, "step": 1122 }, { "epoch": 10.398148148148149, "grad_norm": 1.070380449295044, "learning_rate": 6.248447204968945e-05, "loss": 0.0457, "step": 1123 }, { "epoch": 10.407407407407407, "grad_norm": 1.0458325147628784, "learning_rate": 6.236024844720497e-05, "loss": 0.0443, "step": 1124 }, { "epoch": 10.416666666666666, "grad_norm": 2.423663377761841, "learning_rate": 6.22360248447205e-05, "loss": 0.0768, "step": 1125 }, { "epoch": 10.425925925925926, "grad_norm": 0.707668662071228, "learning_rate": 6.211180124223603e-05, "loss": 0.0245, "step": 1126 }, { "epoch": 10.435185185185185, "grad_norm": 0.9784311056137085, "learning_rate": 6.198757763975156e-05, "loss": 0.0356, "step": 1127 }, { "epoch": 10.444444444444445, "grad_norm": 1.0185497999191284, "learning_rate": 6.186335403726709e-05, "loss": 0.0408, "step": 1128 }, { "epoch": 10.453703703703704, "grad_norm": 2.9687418937683105, "learning_rate": 6.173913043478262e-05, "loss": 0.0602, "step": 1129 }, { "epoch": 10.462962962962964, "grad_norm": 0.9770888090133667, "learning_rate": 6.161490683229813e-05, "loss": 0.0404, "step": 1130 }, { "epoch": 10.472222222222221, "grad_norm": 1.217678427696228, "learning_rate": 6.149068322981368e-05, "loss": 0.0381, "step": 1131 }, { "epoch": 10.481481481481481, "grad_norm": 0.7180342674255371, "learning_rate": 6.136645962732919e-05, "loss": 0.0266, "step": 1132 }, { "epoch": 10.49074074074074, "grad_norm": 3.5172603130340576, "learning_rate": 6.124223602484472e-05, "loss": 0.0743, "step": 1133 }, { "epoch": 10.5, "grad_norm": 1.9250236749649048, "learning_rate": 6.111801242236025e-05, "loss": 0.0557, "step": 1134 }, { "epoch": 10.50925925925926, "grad_norm": 3.3495516777038574, "learning_rate": 6.0993788819875776e-05, "loss": 0.0715, "step": 1135 }, { "epoch": 10.518518518518519, "grad_norm": 1.3475662469863892, "learning_rate": 6.086956521739131e-05, "loss": 0.0692, "step": 1136 }, { "epoch": 10.527777777777779, "grad_norm": 0.7849929928779602, "learning_rate": 6.0745341614906836e-05, "loss": 0.0263, "step": 1137 }, { "epoch": 10.537037037037036, "grad_norm": 1.3163514137268066, "learning_rate": 6.062111801242236e-05, "loss": 0.0406, "step": 1138 }, { "epoch": 10.546296296296296, "grad_norm": 0.7919923663139343, "learning_rate": 6.049689440993789e-05, "loss": 0.0361, "step": 1139 }, { "epoch": 10.555555555555555, "grad_norm": 1.334409475326538, "learning_rate": 6.0372670807453425e-05, "loss": 0.0446, "step": 1140 }, { "epoch": 10.564814814814815, "grad_norm": 3.7500407695770264, "learning_rate": 6.024844720496895e-05, "loss": 0.0969, "step": 1141 }, { "epoch": 10.574074074074074, "grad_norm": 1.9532806873321533, "learning_rate": 6.012422360248448e-05, "loss": 0.0426, "step": 1142 }, { "epoch": 10.583333333333334, "grad_norm": 1.429587483406067, "learning_rate": 6e-05, "loss": 0.0647, "step": 1143 }, { "epoch": 10.592592592592592, "grad_norm": 0.8298918604850769, "learning_rate": 5.9875776397515525e-05, "loss": 0.0336, "step": 1144 }, { "epoch": 10.601851851851851, "grad_norm": 2.578657388687134, "learning_rate": 5.975155279503106e-05, "loss": 0.1264, "step": 1145 }, { "epoch": 10.61111111111111, "grad_norm": 0.9398471117019653, "learning_rate": 5.962732919254659e-05, "loss": 0.0354, "step": 1146 }, { "epoch": 10.62037037037037, "grad_norm": 2.5847291946411133, "learning_rate": 5.9503105590062114e-05, "loss": 0.0763, "step": 1147 }, { "epoch": 10.62962962962963, "grad_norm": 1.649775505065918, "learning_rate": 5.937888198757764e-05, "loss": 0.0329, "step": 1148 }, { "epoch": 10.63888888888889, "grad_norm": 0.9102606177330017, "learning_rate": 5.9254658385093174e-05, "loss": 0.0396, "step": 1149 }, { "epoch": 10.648148148148149, "grad_norm": 0.9106652736663818, "learning_rate": 5.9130434782608704e-05, "loss": 0.0326, "step": 1150 }, { "epoch": 10.657407407407408, "grad_norm": 1.1053392887115479, "learning_rate": 5.900621118012423e-05, "loss": 0.0484, "step": 1151 }, { "epoch": 10.666666666666666, "grad_norm": 0.9101239442825317, "learning_rate": 5.888198757763975e-05, "loss": 0.0368, "step": 1152 }, { "epoch": 10.675925925925926, "grad_norm": 0.8673473000526428, "learning_rate": 5.875776397515528e-05, "loss": 0.0308, "step": 1153 }, { "epoch": 10.685185185185185, "grad_norm": 0.9925929307937622, "learning_rate": 5.863354037267082e-05, "loss": 0.0415, "step": 1154 }, { "epoch": 10.694444444444445, "grad_norm": 0.7710978388786316, "learning_rate": 5.850931677018634e-05, "loss": 0.0279, "step": 1155 }, { "epoch": 10.703703703703704, "grad_norm": 1.1936169862747192, "learning_rate": 5.838509316770186e-05, "loss": 0.0393, "step": 1156 }, { "epoch": 10.712962962962964, "grad_norm": 1.6933677196502686, "learning_rate": 5.826086956521739e-05, "loss": 0.0745, "step": 1157 }, { "epoch": 10.722222222222221, "grad_norm": 1.3728935718536377, "learning_rate": 5.813664596273293e-05, "loss": 0.0541, "step": 1158 }, { "epoch": 10.731481481481481, "grad_norm": 0.8559166193008423, "learning_rate": 5.801242236024845e-05, "loss": 0.0298, "step": 1159 }, { "epoch": 10.74074074074074, "grad_norm": 0.8479191660881042, "learning_rate": 5.7888198757763976e-05, "loss": 0.0291, "step": 1160 }, { "epoch": 10.75, "grad_norm": 3.2553343772888184, "learning_rate": 5.7763975155279506e-05, "loss": 0.0713, "step": 1161 }, { "epoch": 10.75925925925926, "grad_norm": 0.7682176828384399, "learning_rate": 5.763975155279503e-05, "loss": 0.0244, "step": 1162 }, { "epoch": 10.768518518518519, "grad_norm": 0.8638216257095337, "learning_rate": 5.7515527950310566e-05, "loss": 0.0238, "step": 1163 }, { "epoch": 10.777777777777779, "grad_norm": 1.0196231603622437, "learning_rate": 5.739130434782609e-05, "loss": 0.0309, "step": 1164 }, { "epoch": 10.787037037037036, "grad_norm": 0.9703147411346436, "learning_rate": 5.726708074534162e-05, "loss": 0.0308, "step": 1165 }, { "epoch": 10.796296296296296, "grad_norm": 0.8336690068244934, "learning_rate": 5.714285714285714e-05, "loss": 0.0272, "step": 1166 }, { "epoch": 10.805555555555555, "grad_norm": 0.5598458051681519, "learning_rate": 5.701863354037268e-05, "loss": 0.0182, "step": 1167 }, { "epoch": 10.814814814814815, "grad_norm": 0.8961607813835144, "learning_rate": 5.68944099378882e-05, "loss": 0.0313, "step": 1168 }, { "epoch": 10.824074074074074, "grad_norm": 4.9854207038879395, "learning_rate": 5.677018633540373e-05, "loss": 0.1544, "step": 1169 }, { "epoch": 10.833333333333334, "grad_norm": 1.967529058456421, "learning_rate": 5.6645962732919255e-05, "loss": 0.0656, "step": 1170 }, { "epoch": 10.842592592592592, "grad_norm": 1.4599827527999878, "learning_rate": 5.652173913043478e-05, "loss": 0.0564, "step": 1171 }, { "epoch": 10.851851851851851, "grad_norm": 0.9949111342430115, "learning_rate": 5.6397515527950314e-05, "loss": 0.0314, "step": 1172 }, { "epoch": 10.86111111111111, "grad_norm": 1.4174683094024658, "learning_rate": 5.6273291925465844e-05, "loss": 0.0552, "step": 1173 }, { "epoch": 10.87037037037037, "grad_norm": 1.6052273511886597, "learning_rate": 5.614906832298137e-05, "loss": 0.0557, "step": 1174 }, { "epoch": 10.87962962962963, "grad_norm": 1.5254520177841187, "learning_rate": 5.60248447204969e-05, "loss": 0.0595, "step": 1175 }, { "epoch": 10.88888888888889, "grad_norm": 1.0831390619277954, "learning_rate": 5.590062111801242e-05, "loss": 0.0364, "step": 1176 }, { "epoch": 10.898148148148149, "grad_norm": 2.996645212173462, "learning_rate": 5.577639751552796e-05, "loss": 0.1694, "step": 1177 }, { "epoch": 10.907407407407408, "grad_norm": 1.1233099699020386, "learning_rate": 5.565217391304348e-05, "loss": 0.0422, "step": 1178 }, { "epoch": 10.916666666666666, "grad_norm": 1.1308964490890503, "learning_rate": 5.552795031055901e-05, "loss": 0.0414, "step": 1179 }, { "epoch": 10.925925925925926, "grad_norm": 0.7682093381881714, "learning_rate": 5.540372670807453e-05, "loss": 0.0219, "step": 1180 }, { "epoch": 10.935185185185185, "grad_norm": 1.0228471755981445, "learning_rate": 5.527950310559007e-05, "loss": 0.0451, "step": 1181 }, { "epoch": 10.944444444444445, "grad_norm": 0.8916832208633423, "learning_rate": 5.515527950310559e-05, "loss": 0.0315, "step": 1182 }, { "epoch": 10.953703703703704, "grad_norm": 2.371509313583374, "learning_rate": 5.503105590062112e-05, "loss": 0.0667, "step": 1183 }, { "epoch": 10.962962962962964, "grad_norm": 1.1845130920410156, "learning_rate": 5.4906832298136646e-05, "loss": 0.0391, "step": 1184 }, { "epoch": 10.972222222222221, "grad_norm": 1.607428789138794, "learning_rate": 5.478260869565217e-05, "loss": 0.0506, "step": 1185 }, { "epoch": 10.981481481481481, "grad_norm": 3.609421968460083, "learning_rate": 5.4658385093167706e-05, "loss": 0.1342, "step": 1186 }, { "epoch": 10.99074074074074, "grad_norm": 2.8664135932922363, "learning_rate": 5.4534161490683236e-05, "loss": 0.1277, "step": 1187 }, { "epoch": 11.0, "grad_norm": 3.0398714542388916, "learning_rate": 5.440993788819876e-05, "loss": 0.102, "step": 1188 }, { "epoch": 11.00925925925926, "grad_norm": 0.8366669416427612, "learning_rate": 5.428571428571428e-05, "loss": 0.0293, "step": 1189 }, { "epoch": 11.018518518518519, "grad_norm": 0.8883122205734253, "learning_rate": 5.416149068322982e-05, "loss": 0.0401, "step": 1190 }, { "epoch": 11.027777777777779, "grad_norm": 0.9267494678497314, "learning_rate": 5.403726708074535e-05, "loss": 0.0403, "step": 1191 }, { "epoch": 11.037037037037036, "grad_norm": 0.8485496044158936, "learning_rate": 5.391304347826087e-05, "loss": 0.0272, "step": 1192 }, { "epoch": 11.046296296296296, "grad_norm": 0.8455182909965515, "learning_rate": 5.3788819875776395e-05, "loss": 0.0277, "step": 1193 }, { "epoch": 11.055555555555555, "grad_norm": 0.47915375232696533, "learning_rate": 5.3664596273291925e-05, "loss": 0.0132, "step": 1194 }, { "epoch": 11.064814814814815, "grad_norm": 1.502824068069458, "learning_rate": 5.354037267080746e-05, "loss": 0.0531, "step": 1195 }, { "epoch": 11.074074074074074, "grad_norm": 0.8204081654548645, "learning_rate": 5.3416149068322984e-05, "loss": 0.0349, "step": 1196 }, { "epoch": 11.083333333333334, "grad_norm": 0.6424718499183655, "learning_rate": 5.329192546583851e-05, "loss": 0.0163, "step": 1197 }, { "epoch": 11.092592592592593, "grad_norm": 1.192146897315979, "learning_rate": 5.316770186335404e-05, "loss": 0.0575, "step": 1198 }, { "epoch": 11.101851851851851, "grad_norm": 0.9034618735313416, "learning_rate": 5.3043478260869574e-05, "loss": 0.0339, "step": 1199 }, { "epoch": 11.11111111111111, "grad_norm": 0.9357187151908875, "learning_rate": 5.29192546583851e-05, "loss": 0.0301, "step": 1200 }, { "epoch": 11.12037037037037, "grad_norm": 0.802025318145752, "learning_rate": 5.279503105590062e-05, "loss": 0.033, "step": 1201 }, { "epoch": 11.12962962962963, "grad_norm": 0.8969548940658569, "learning_rate": 5.267080745341615e-05, "loss": 0.038, "step": 1202 }, { "epoch": 11.13888888888889, "grad_norm": 0.8712778091430664, "learning_rate": 5.254658385093167e-05, "loss": 0.0417, "step": 1203 }, { "epoch": 11.148148148148149, "grad_norm": 1.1687285900115967, "learning_rate": 5.242236024844721e-05, "loss": 0.0394, "step": 1204 }, { "epoch": 11.157407407407407, "grad_norm": 0.8223320841789246, "learning_rate": 5.229813664596273e-05, "loss": 0.0245, "step": 1205 }, { "epoch": 11.166666666666666, "grad_norm": 1.6229560375213623, "learning_rate": 5.217391304347826e-05, "loss": 0.0486, "step": 1206 }, { "epoch": 11.175925925925926, "grad_norm": 0.4597468376159668, "learning_rate": 5.2049689440993786e-05, "loss": 0.0116, "step": 1207 }, { "epoch": 11.185185185185185, "grad_norm": 0.9737257957458496, "learning_rate": 5.192546583850932e-05, "loss": 0.0296, "step": 1208 }, { "epoch": 11.194444444444445, "grad_norm": 1.6504062414169312, "learning_rate": 5.180124223602485e-05, "loss": 0.0711, "step": 1209 }, { "epoch": 11.203703703703704, "grad_norm": 1.0058763027191162, "learning_rate": 5.1677018633540376e-05, "loss": 0.0392, "step": 1210 }, { "epoch": 11.212962962962964, "grad_norm": 0.9670447707176208, "learning_rate": 5.15527950310559e-05, "loss": 0.0411, "step": 1211 }, { "epoch": 11.222222222222221, "grad_norm": 0.8737079501152039, "learning_rate": 5.142857142857143e-05, "loss": 0.0258, "step": 1212 }, { "epoch": 11.231481481481481, "grad_norm": 0.8489564061164856, "learning_rate": 5.1304347826086966e-05, "loss": 0.0355, "step": 1213 }, { "epoch": 11.24074074074074, "grad_norm": 0.8969002366065979, "learning_rate": 5.118012422360249e-05, "loss": 0.0415, "step": 1214 }, { "epoch": 11.25, "grad_norm": 0.8433552384376526, "learning_rate": 5.105590062111801e-05, "loss": 0.0359, "step": 1215 }, { "epoch": 11.25925925925926, "grad_norm": 1.3878159523010254, "learning_rate": 5.093167701863354e-05, "loss": 0.055, "step": 1216 }, { "epoch": 11.268518518518519, "grad_norm": 1.1686030626296997, "learning_rate": 5.080745341614908e-05, "loss": 0.0403, "step": 1217 }, { "epoch": 11.277777777777779, "grad_norm": 0.8048990964889526, "learning_rate": 5.06832298136646e-05, "loss": 0.0322, "step": 1218 }, { "epoch": 11.287037037037036, "grad_norm": 0.661250114440918, "learning_rate": 5.0559006211180125e-05, "loss": 0.0259, "step": 1219 }, { "epoch": 11.296296296296296, "grad_norm": 0.7250069975852966, "learning_rate": 5.0434782608695655e-05, "loss": 0.0239, "step": 1220 }, { "epoch": 11.305555555555555, "grad_norm": 1.1156517267227173, "learning_rate": 5.031055900621118e-05, "loss": 0.0347, "step": 1221 }, { "epoch": 11.314814814814815, "grad_norm": 0.941217303276062, "learning_rate": 5.0186335403726714e-05, "loss": 0.0318, "step": 1222 }, { "epoch": 11.324074074074074, "grad_norm": 4.407271385192871, "learning_rate": 5.006211180124224e-05, "loss": 0.0731, "step": 1223 }, { "epoch": 11.333333333333334, "grad_norm": 1.0587486028671265, "learning_rate": 4.993788819875777e-05, "loss": 0.042, "step": 1224 }, { "epoch": 11.342592592592593, "grad_norm": 1.2566124200820923, "learning_rate": 4.98136645962733e-05, "loss": 0.0387, "step": 1225 }, { "epoch": 11.351851851851851, "grad_norm": 0.9527769684791565, "learning_rate": 4.968944099378882e-05, "loss": 0.0415, "step": 1226 }, { "epoch": 11.36111111111111, "grad_norm": 1.3963499069213867, "learning_rate": 4.956521739130435e-05, "loss": 0.0543, "step": 1227 }, { "epoch": 11.37037037037037, "grad_norm": 0.9101176261901855, "learning_rate": 4.944099378881988e-05, "loss": 0.0387, "step": 1228 }, { "epoch": 11.37962962962963, "grad_norm": 1.1519862413406372, "learning_rate": 4.93167701863354e-05, "loss": 0.0381, "step": 1229 }, { "epoch": 11.38888888888889, "grad_norm": 1.6292718648910522, "learning_rate": 4.919254658385093e-05, "loss": 0.0682, "step": 1230 }, { "epoch": 11.398148148148149, "grad_norm": 0.717383623123169, "learning_rate": 4.906832298136646e-05, "loss": 0.0306, "step": 1231 }, { "epoch": 11.407407407407407, "grad_norm": 1.2665631771087646, "learning_rate": 4.894409937888199e-05, "loss": 0.064, "step": 1232 }, { "epoch": 11.416666666666666, "grad_norm": 0.914641261100769, "learning_rate": 4.8819875776397516e-05, "loss": 0.0319, "step": 1233 }, { "epoch": 11.425925925925926, "grad_norm": 3.159534454345703, "learning_rate": 4.8695652173913046e-05, "loss": 0.1141, "step": 1234 }, { "epoch": 11.435185185185185, "grad_norm": 0.8833385109901428, "learning_rate": 4.8571428571428576e-05, "loss": 0.0276, "step": 1235 }, { "epoch": 11.444444444444445, "grad_norm": 0.916418194770813, "learning_rate": 4.8447204968944106e-05, "loss": 0.0311, "step": 1236 }, { "epoch": 11.453703703703704, "grad_norm": 1.092687964439392, "learning_rate": 4.832298136645963e-05, "loss": 0.048, "step": 1237 }, { "epoch": 11.462962962962964, "grad_norm": 0.7379303574562073, "learning_rate": 4.819875776397515e-05, "loss": 0.0241, "step": 1238 }, { "epoch": 11.472222222222221, "grad_norm": 1.1241528987884521, "learning_rate": 4.807453416149069e-05, "loss": 0.0429, "step": 1239 }, { "epoch": 11.481481481481481, "grad_norm": 2.250471591949463, "learning_rate": 4.795031055900621e-05, "loss": 0.0408, "step": 1240 }, { "epoch": 11.49074074074074, "grad_norm": 1.1494890451431274, "learning_rate": 4.782608695652174e-05, "loss": 0.0501, "step": 1241 }, { "epoch": 11.5, "grad_norm": 0.8903128504753113, "learning_rate": 4.770186335403727e-05, "loss": 0.0281, "step": 1242 }, { "epoch": 11.50925925925926, "grad_norm": 0.6605562567710876, "learning_rate": 4.75776397515528e-05, "loss": 0.0235, "step": 1243 }, { "epoch": 11.518518518518519, "grad_norm": 1.227002739906311, "learning_rate": 4.7453416149068325e-05, "loss": 0.0506, "step": 1244 }, { "epoch": 11.527777777777779, "grad_norm": 1.0969829559326172, "learning_rate": 4.7329192546583855e-05, "loss": 0.0435, "step": 1245 }, { "epoch": 11.537037037037036, "grad_norm": 1.4205132722854614, "learning_rate": 4.7204968944099384e-05, "loss": 0.0762, "step": 1246 }, { "epoch": 11.546296296296296, "grad_norm": 1.1668907403945923, "learning_rate": 4.708074534161491e-05, "loss": 0.0478, "step": 1247 }, { "epoch": 11.555555555555555, "grad_norm": 0.9908825159072876, "learning_rate": 4.695652173913044e-05, "loss": 0.0437, "step": 1248 }, { "epoch": 11.564814814814815, "grad_norm": 1.2143094539642334, "learning_rate": 4.683229813664596e-05, "loss": 0.0452, "step": 1249 }, { "epoch": 11.574074074074074, "grad_norm": 0.7645125389099121, "learning_rate": 4.67080745341615e-05, "loss": 0.0297, "step": 1250 }, { "epoch": 11.583333333333334, "grad_norm": 0.8744638562202454, "learning_rate": 4.658385093167702e-05, "loss": 0.0342, "step": 1251 }, { "epoch": 11.592592592592592, "grad_norm": 0.9421960115432739, "learning_rate": 4.645962732919255e-05, "loss": 0.0422, "step": 1252 }, { "epoch": 11.601851851851851, "grad_norm": 0.9777348637580872, "learning_rate": 4.633540372670807e-05, "loss": 0.022, "step": 1253 }, { "epoch": 11.61111111111111, "grad_norm": 1.009817361831665, "learning_rate": 4.62111801242236e-05, "loss": 0.0458, "step": 1254 }, { "epoch": 11.62037037037037, "grad_norm": 1.0724588632583618, "learning_rate": 4.608695652173913e-05, "loss": 0.0441, "step": 1255 }, { "epoch": 11.62962962962963, "grad_norm": 0.9254492521286011, "learning_rate": 4.5962732919254656e-05, "loss": 0.037, "step": 1256 }, { "epoch": 11.63888888888889, "grad_norm": 1.018061637878418, "learning_rate": 4.5838509316770186e-05, "loss": 0.0378, "step": 1257 }, { "epoch": 11.648148148148149, "grad_norm": 0.8017347455024719, "learning_rate": 4.5714285714285716e-05, "loss": 0.0369, "step": 1258 }, { "epoch": 11.657407407407408, "grad_norm": 1.1639477014541626, "learning_rate": 4.5590062111801246e-05, "loss": 0.0428, "step": 1259 }, { "epoch": 11.666666666666666, "grad_norm": 1.008543610572815, "learning_rate": 4.546583850931677e-05, "loss": 0.0424, "step": 1260 }, { "epoch": 11.675925925925926, "grad_norm": 1.6025316715240479, "learning_rate": 4.5341614906832306e-05, "loss": 0.0614, "step": 1261 }, { "epoch": 11.685185185185185, "grad_norm": 0.9679161310195923, "learning_rate": 4.521739130434783e-05, "loss": 0.0328, "step": 1262 }, { "epoch": 11.694444444444445, "grad_norm": 1.2503728866577148, "learning_rate": 4.509316770186335e-05, "loss": 0.0471, "step": 1263 }, { "epoch": 11.703703703703704, "grad_norm": 1.0044113397598267, "learning_rate": 4.496894409937888e-05, "loss": 0.0361, "step": 1264 }, { "epoch": 11.712962962962964, "grad_norm": 0.9651626348495483, "learning_rate": 4.484472049689441e-05, "loss": 0.0413, "step": 1265 }, { "epoch": 11.722222222222221, "grad_norm": 2.5001490116119385, "learning_rate": 4.472049689440994e-05, "loss": 0.0573, "step": 1266 }, { "epoch": 11.731481481481481, "grad_norm": 0.8396499156951904, "learning_rate": 4.4596273291925465e-05, "loss": 0.033, "step": 1267 }, { "epoch": 11.74074074074074, "grad_norm": 0.7739114165306091, "learning_rate": 4.4472049689440995e-05, "loss": 0.0297, "step": 1268 }, { "epoch": 11.75, "grad_norm": 0.7845550775527954, "learning_rate": 4.4347826086956525e-05, "loss": 0.0384, "step": 1269 }, { "epoch": 11.75925925925926, "grad_norm": 0.9752318859100342, "learning_rate": 4.4223602484472055e-05, "loss": 0.0359, "step": 1270 }, { "epoch": 11.768518518518519, "grad_norm": 0.8686448335647583, "learning_rate": 4.409937888198758e-05, "loss": 0.0333, "step": 1271 }, { "epoch": 11.777777777777779, "grad_norm": 1.0658447742462158, "learning_rate": 4.397515527950311e-05, "loss": 0.0446, "step": 1272 }, { "epoch": 11.787037037037036, "grad_norm": 2.433940887451172, "learning_rate": 4.385093167701864e-05, "loss": 0.0614, "step": 1273 }, { "epoch": 11.796296296296296, "grad_norm": 1.705237865447998, "learning_rate": 4.372670807453416e-05, "loss": 0.0731, "step": 1274 }, { "epoch": 11.805555555555555, "grad_norm": 0.6588522791862488, "learning_rate": 4.360248447204969e-05, "loss": 0.0384, "step": 1275 }, { "epoch": 11.814814814814815, "grad_norm": 1.0283806324005127, "learning_rate": 4.347826086956522e-05, "loss": 0.0339, "step": 1276 }, { "epoch": 11.824074074074074, "grad_norm": 1.315185308456421, "learning_rate": 4.335403726708075e-05, "loss": 0.0578, "step": 1277 }, { "epoch": 11.833333333333334, "grad_norm": 0.930867612361908, "learning_rate": 4.322981366459627e-05, "loss": 0.0354, "step": 1278 }, { "epoch": 11.842592592592592, "grad_norm": 1.5080018043518066, "learning_rate": 4.31055900621118e-05, "loss": 0.0446, "step": 1279 }, { "epoch": 11.851851851851851, "grad_norm": 0.7906777262687683, "learning_rate": 4.298136645962733e-05, "loss": 0.0353, "step": 1280 }, { "epoch": 11.86111111111111, "grad_norm": 1.961120843887329, "learning_rate": 4.2857142857142856e-05, "loss": 0.0355, "step": 1281 }, { "epoch": 11.87037037037037, "grad_norm": 1.2218685150146484, "learning_rate": 4.2732919254658386e-05, "loss": 0.048, "step": 1282 }, { "epoch": 11.87962962962963, "grad_norm": 1.000490427017212, "learning_rate": 4.2608695652173916e-05, "loss": 0.0402, "step": 1283 }, { "epoch": 11.88888888888889, "grad_norm": 0.762229859828949, "learning_rate": 4.2484472049689446e-05, "loss": 0.0322, "step": 1284 }, { "epoch": 11.898148148148149, "grad_norm": 0.7762311100959778, "learning_rate": 4.236024844720497e-05, "loss": 0.0324, "step": 1285 }, { "epoch": 11.907407407407408, "grad_norm": 0.8216007947921753, "learning_rate": 4.22360248447205e-05, "loss": 0.0281, "step": 1286 }, { "epoch": 11.916666666666666, "grad_norm": 0.8371996879577637, "learning_rate": 4.211180124223603e-05, "loss": 0.0316, "step": 1287 }, { "epoch": 11.925925925925926, "grad_norm": 1.3675734996795654, "learning_rate": 4.198757763975156e-05, "loss": 0.063, "step": 1288 }, { "epoch": 11.935185185185185, "grad_norm": 0.5560950636863708, "learning_rate": 4.186335403726708e-05, "loss": 0.0242, "step": 1289 }, { "epoch": 11.944444444444445, "grad_norm": 0.775168240070343, "learning_rate": 4.1739130434782605e-05, "loss": 0.0339, "step": 1290 }, { "epoch": 11.953703703703704, "grad_norm": 1.0372427701950073, "learning_rate": 4.161490683229814e-05, "loss": 0.0398, "step": 1291 }, { "epoch": 11.962962962962964, "grad_norm": 2.2468655109405518, "learning_rate": 4.1490683229813665e-05, "loss": 0.0525, "step": 1292 }, { "epoch": 11.972222222222221, "grad_norm": 0.867691695690155, "learning_rate": 4.1366459627329195e-05, "loss": 0.0443, "step": 1293 }, { "epoch": 11.981481481481481, "grad_norm": 0.6092666387557983, "learning_rate": 4.1242236024844725e-05, "loss": 0.0258, "step": 1294 }, { "epoch": 11.99074074074074, "grad_norm": 0.9719471335411072, "learning_rate": 4.1118012422360255e-05, "loss": 0.0413, "step": 1295 }, { "epoch": 12.0, "grad_norm": 0.8400522470474243, "learning_rate": 4.099378881987578e-05, "loss": 0.0342, "step": 1296 }, { "epoch": 12.00925925925926, "grad_norm": 0.9376289248466492, "learning_rate": 4.086956521739131e-05, "loss": 0.0373, "step": 1297 }, { "epoch": 12.018518518518519, "grad_norm": 1.0187164545059204, "learning_rate": 4.074534161490684e-05, "loss": 0.0377, "step": 1298 }, { "epoch": 12.027777777777779, "grad_norm": 1.1379069089889526, "learning_rate": 4.062111801242236e-05, "loss": 0.0572, "step": 1299 }, { "epoch": 12.037037037037036, "grad_norm": 0.9860619902610779, "learning_rate": 4.049689440993789e-05, "loss": 0.0426, "step": 1300 }, { "epoch": 12.046296296296296, "grad_norm": 0.9077693223953247, "learning_rate": 4.0372670807453414e-05, "loss": 0.036, "step": 1301 }, { "epoch": 12.055555555555555, "grad_norm": 0.5804708003997803, "learning_rate": 4.024844720496895e-05, "loss": 0.0247, "step": 1302 }, { "epoch": 12.064814814814815, "grad_norm": 0.7517374753952026, "learning_rate": 4.012422360248447e-05, "loss": 0.0259, "step": 1303 }, { "epoch": 12.074074074074074, "grad_norm": 0.9251551032066345, "learning_rate": 4e-05, "loss": 0.0379, "step": 1304 }, { "epoch": 12.083333333333334, "grad_norm": 0.7882636189460754, "learning_rate": 3.9875776397515526e-05, "loss": 0.0287, "step": 1305 }, { "epoch": 12.092592592592593, "grad_norm": 0.6092426180839539, "learning_rate": 3.9751552795031056e-05, "loss": 0.0211, "step": 1306 }, { "epoch": 12.101851851851851, "grad_norm": 0.8440184593200684, "learning_rate": 3.9627329192546586e-05, "loss": 0.0282, "step": 1307 }, { "epoch": 12.11111111111111, "grad_norm": 0.9278969168663025, "learning_rate": 3.950310559006211e-05, "loss": 0.0304, "step": 1308 }, { "epoch": 12.12037037037037, "grad_norm": 0.6380085945129395, "learning_rate": 3.9378881987577646e-05, "loss": 0.0223, "step": 1309 }, { "epoch": 12.12962962962963, "grad_norm": 0.7007056474685669, "learning_rate": 3.925465838509317e-05, "loss": 0.0241, "step": 1310 }, { "epoch": 12.13888888888889, "grad_norm": 0.7670966982841492, "learning_rate": 3.91304347826087e-05, "loss": 0.0283, "step": 1311 }, { "epoch": 12.148148148148149, "grad_norm": 0.8422584533691406, "learning_rate": 3.900621118012422e-05, "loss": 0.0291, "step": 1312 }, { "epoch": 12.157407407407407, "grad_norm": 0.6846734881401062, "learning_rate": 3.888198757763976e-05, "loss": 0.0232, "step": 1313 }, { "epoch": 12.166666666666666, "grad_norm": 0.900596559047699, "learning_rate": 3.875776397515528e-05, "loss": 0.0295, "step": 1314 }, { "epoch": 12.175925925925926, "grad_norm": 0.6619224548339844, "learning_rate": 3.8633540372670805e-05, "loss": 0.0243, "step": 1315 }, { "epoch": 12.185185185185185, "grad_norm": 0.431782603263855, "learning_rate": 3.8509316770186335e-05, "loss": 0.0174, "step": 1316 }, { "epoch": 12.194444444444445, "grad_norm": 1.0050886869430542, "learning_rate": 3.8385093167701865e-05, "loss": 0.0391, "step": 1317 }, { "epoch": 12.203703703703704, "grad_norm": 0.8259010910987854, "learning_rate": 3.8260869565217395e-05, "loss": 0.0259, "step": 1318 }, { "epoch": 12.212962962962964, "grad_norm": 1.4165301322937012, "learning_rate": 3.813664596273292e-05, "loss": 0.0446, "step": 1319 }, { "epoch": 12.222222222222221, "grad_norm": 0.7267447113990784, "learning_rate": 3.801242236024845e-05, "loss": 0.022, "step": 1320 }, { "epoch": 12.231481481481481, "grad_norm": 0.8817240595817566, "learning_rate": 3.788819875776398e-05, "loss": 0.0349, "step": 1321 }, { "epoch": 12.24074074074074, "grad_norm": 0.815268337726593, "learning_rate": 3.776397515527951e-05, "loss": 0.027, "step": 1322 }, { "epoch": 12.25, "grad_norm": 0.6415846347808838, "learning_rate": 3.763975155279503e-05, "loss": 0.0212, "step": 1323 }, { "epoch": 12.25925925925926, "grad_norm": 0.7012866139411926, "learning_rate": 3.751552795031056e-05, "loss": 0.0219, "step": 1324 }, { "epoch": 12.268518518518519, "grad_norm": 0.9753007888793945, "learning_rate": 3.739130434782609e-05, "loss": 0.0339, "step": 1325 }, { "epoch": 12.277777777777779, "grad_norm": 0.8093515038490295, "learning_rate": 3.7267080745341614e-05, "loss": 0.0277, "step": 1326 }, { "epoch": 12.287037037037036, "grad_norm": 0.8891710638999939, "learning_rate": 3.7142857142857143e-05, "loss": 0.0413, "step": 1327 }, { "epoch": 12.296296296296296, "grad_norm": 0.5199636816978455, "learning_rate": 3.701863354037267e-05, "loss": 0.0169, "step": 1328 }, { "epoch": 12.305555555555555, "grad_norm": 0.8058778643608093, "learning_rate": 3.68944099378882e-05, "loss": 0.0324, "step": 1329 }, { "epoch": 12.314814814814815, "grad_norm": 0.7521347403526306, "learning_rate": 3.6770186335403726e-05, "loss": 0.0225, "step": 1330 }, { "epoch": 12.324074074074074, "grad_norm": 0.9405580163002014, "learning_rate": 3.6645962732919256e-05, "loss": 0.0321, "step": 1331 }, { "epoch": 12.333333333333334, "grad_norm": 0.6070406436920166, "learning_rate": 3.6521739130434786e-05, "loss": 0.0192, "step": 1332 }, { "epoch": 12.342592592592593, "grad_norm": 0.9515055418014526, "learning_rate": 3.639751552795031e-05, "loss": 0.0273, "step": 1333 }, { "epoch": 12.351851851851851, "grad_norm": 0.5170685052871704, "learning_rate": 3.627329192546584e-05, "loss": 0.0264, "step": 1334 }, { "epoch": 12.36111111111111, "grad_norm": 1.2589349746704102, "learning_rate": 3.614906832298137e-05, "loss": 0.0593, "step": 1335 }, { "epoch": 12.37037037037037, "grad_norm": 1.3736205101013184, "learning_rate": 3.60248447204969e-05, "loss": 0.038, "step": 1336 }, { "epoch": 12.37962962962963, "grad_norm": 0.7936609387397766, "learning_rate": 3.590062111801242e-05, "loss": 0.029, "step": 1337 }, { "epoch": 12.38888888888889, "grad_norm": 0.908733606338501, "learning_rate": 3.577639751552795e-05, "loss": 0.0343, "step": 1338 }, { "epoch": 12.398148148148149, "grad_norm": 0.8484895825386047, "learning_rate": 3.565217391304348e-05, "loss": 0.0275, "step": 1339 }, { "epoch": 12.407407407407407, "grad_norm": 1.0681447982788086, "learning_rate": 3.552795031055901e-05, "loss": 0.0328, "step": 1340 }, { "epoch": 12.416666666666666, "grad_norm": 1.075875997543335, "learning_rate": 3.5403726708074535e-05, "loss": 0.0418, "step": 1341 }, { "epoch": 12.425925925925926, "grad_norm": 0.6161890029907227, "learning_rate": 3.5279503105590065e-05, "loss": 0.0296, "step": 1342 }, { "epoch": 12.435185185185185, "grad_norm": 1.1916195154190063, "learning_rate": 3.5155279503105595e-05, "loss": 0.0486, "step": 1343 }, { "epoch": 12.444444444444445, "grad_norm": 0.6621356010437012, "learning_rate": 3.503105590062112e-05, "loss": 0.0221, "step": 1344 }, { "epoch": 12.453703703703704, "grad_norm": 1.6666837930679321, "learning_rate": 3.490683229813665e-05, "loss": 0.0459, "step": 1345 }, { "epoch": 12.462962962962964, "grad_norm": 0.7422270178794861, "learning_rate": 3.478260869565218e-05, "loss": 0.0249, "step": 1346 }, { "epoch": 12.472222222222221, "grad_norm": 0.7906531095504761, "learning_rate": 3.465838509316771e-05, "loss": 0.038, "step": 1347 }, { "epoch": 12.481481481481481, "grad_norm": 0.9974073767662048, "learning_rate": 3.453416149068323e-05, "loss": 0.0406, "step": 1348 }, { "epoch": 12.49074074074074, "grad_norm": 0.6397982835769653, "learning_rate": 3.440993788819876e-05, "loss": 0.0222, "step": 1349 }, { "epoch": 12.5, "grad_norm": 0.682849645614624, "learning_rate": 3.428571428571429e-05, "loss": 0.0298, "step": 1350 }, { "epoch": 12.50925925925926, "grad_norm": 1.4077941179275513, "learning_rate": 3.4161490683229814e-05, "loss": 0.0511, "step": 1351 }, { "epoch": 12.518518518518519, "grad_norm": 0.7027615308761597, "learning_rate": 3.4037267080745343e-05, "loss": 0.0203, "step": 1352 }, { "epoch": 12.527777777777779, "grad_norm": 1.0765162706375122, "learning_rate": 3.3913043478260867e-05, "loss": 0.0379, "step": 1353 }, { "epoch": 12.537037037037036, "grad_norm": 0.7544326782226562, "learning_rate": 3.37888198757764e-05, "loss": 0.0278, "step": 1354 }, { "epoch": 12.546296296296296, "grad_norm": 1.0442858934402466, "learning_rate": 3.3664596273291926e-05, "loss": 0.0441, "step": 1355 }, { "epoch": 12.555555555555555, "grad_norm": 1.0142107009887695, "learning_rate": 3.3540372670807456e-05, "loss": 0.0446, "step": 1356 }, { "epoch": 12.564814814814815, "grad_norm": 1.299988865852356, "learning_rate": 3.341614906832298e-05, "loss": 0.0447, "step": 1357 }, { "epoch": 12.574074074074074, "grad_norm": 0.6750516295433044, "learning_rate": 3.329192546583851e-05, "loss": 0.03, "step": 1358 }, { "epoch": 12.583333333333334, "grad_norm": 0.8981741070747375, "learning_rate": 3.316770186335404e-05, "loss": 0.0315, "step": 1359 }, { "epoch": 12.592592592592592, "grad_norm": 1.8234065771102905, "learning_rate": 3.304347826086956e-05, "loss": 0.0857, "step": 1360 }, { "epoch": 12.601851851851851, "grad_norm": 1.1053465604782104, "learning_rate": 3.29192546583851e-05, "loss": 0.0436, "step": 1361 }, { "epoch": 12.61111111111111, "grad_norm": 1.0005968809127808, "learning_rate": 3.279503105590062e-05, "loss": 0.042, "step": 1362 }, { "epoch": 12.62037037037037, "grad_norm": 0.952779233455658, "learning_rate": 3.267080745341615e-05, "loss": 0.0314, "step": 1363 }, { "epoch": 12.62962962962963, "grad_norm": 2.057094097137451, "learning_rate": 3.2546583850931675e-05, "loss": 0.0484, "step": 1364 }, { "epoch": 12.63888888888889, "grad_norm": 0.7208571434020996, "learning_rate": 3.242236024844721e-05, "loss": 0.0309, "step": 1365 }, { "epoch": 12.648148148148149, "grad_norm": 0.8814190626144409, "learning_rate": 3.2298136645962735e-05, "loss": 0.0373, "step": 1366 }, { "epoch": 12.657407407407408, "grad_norm": 0.7553854584693909, "learning_rate": 3.217391304347826e-05, "loss": 0.0259, "step": 1367 }, { "epoch": 12.666666666666666, "grad_norm": 0.8120577335357666, "learning_rate": 3.204968944099379e-05, "loss": 0.0332, "step": 1368 }, { "epoch": 12.675925925925926, "grad_norm": 0.8838486075401306, "learning_rate": 3.192546583850932e-05, "loss": 0.0324, "step": 1369 }, { "epoch": 12.685185185185185, "grad_norm": 0.7664040923118591, "learning_rate": 3.180124223602485e-05, "loss": 0.0312, "step": 1370 }, { "epoch": 12.694444444444445, "grad_norm": 0.7717130780220032, "learning_rate": 3.167701863354037e-05, "loss": 0.0352, "step": 1371 }, { "epoch": 12.703703703703704, "grad_norm": 3.366727590560913, "learning_rate": 3.15527950310559e-05, "loss": 0.1035, "step": 1372 }, { "epoch": 12.712962962962964, "grad_norm": 0.723676323890686, "learning_rate": 3.142857142857143e-05, "loss": 0.0251, "step": 1373 }, { "epoch": 12.722222222222221, "grad_norm": 0.7675268650054932, "learning_rate": 3.130434782608696e-05, "loss": 0.0313, "step": 1374 }, { "epoch": 12.731481481481481, "grad_norm": 0.6413992643356323, "learning_rate": 3.1180124223602484e-05, "loss": 0.0287, "step": 1375 }, { "epoch": 12.74074074074074, "grad_norm": 0.8057203888893127, "learning_rate": 3.1055900621118014e-05, "loss": 0.0252, "step": 1376 }, { "epoch": 12.75, "grad_norm": 0.9161250591278076, "learning_rate": 3.0931677018633543e-05, "loss": 0.0329, "step": 1377 }, { "epoch": 12.75925925925926, "grad_norm": 0.7578794956207275, "learning_rate": 3.0807453416149067e-05, "loss": 0.031, "step": 1378 }, { "epoch": 12.768518518518519, "grad_norm": 0.7553548216819763, "learning_rate": 3.0683229813664596e-05, "loss": 0.0279, "step": 1379 }, { "epoch": 12.777777777777779, "grad_norm": 0.9717443585395813, "learning_rate": 3.0559006211180126e-05, "loss": 0.0333, "step": 1380 }, { "epoch": 12.787037037037036, "grad_norm": 0.7987967133522034, "learning_rate": 3.0434782608695656e-05, "loss": 0.0317, "step": 1381 }, { "epoch": 12.796296296296296, "grad_norm": 1.3905160427093506, "learning_rate": 3.031055900621118e-05, "loss": 0.0558, "step": 1382 }, { "epoch": 12.805555555555555, "grad_norm": 0.7816559076309204, "learning_rate": 3.0186335403726713e-05, "loss": 0.036, "step": 1383 }, { "epoch": 12.814814814814815, "grad_norm": 0.6633138060569763, "learning_rate": 3.006211180124224e-05, "loss": 0.0287, "step": 1384 }, { "epoch": 12.824074074074074, "grad_norm": 0.6967059373855591, "learning_rate": 2.9937888198757762e-05, "loss": 0.0237, "step": 1385 }, { "epoch": 12.833333333333334, "grad_norm": 0.9407333135604858, "learning_rate": 2.9813664596273296e-05, "loss": 0.0354, "step": 1386 }, { "epoch": 12.842592592592592, "grad_norm": 1.1632462739944458, "learning_rate": 2.968944099378882e-05, "loss": 0.0468, "step": 1387 }, { "epoch": 12.851851851851851, "grad_norm": 0.5927537679672241, "learning_rate": 2.9565217391304352e-05, "loss": 0.0226, "step": 1388 }, { "epoch": 12.86111111111111, "grad_norm": 0.7235986590385437, "learning_rate": 2.9440993788819875e-05, "loss": 0.0229, "step": 1389 }, { "epoch": 12.87037037037037, "grad_norm": 0.7975889444351196, "learning_rate": 2.931677018633541e-05, "loss": 0.0293, "step": 1390 }, { "epoch": 12.87962962962963, "grad_norm": 0.8041378259658813, "learning_rate": 2.919254658385093e-05, "loss": 0.0276, "step": 1391 }, { "epoch": 12.88888888888889, "grad_norm": 0.5750847458839417, "learning_rate": 2.9068322981366465e-05, "loss": 0.024, "step": 1392 }, { "epoch": 12.898148148148149, "grad_norm": 0.8676141500473022, "learning_rate": 2.8944099378881988e-05, "loss": 0.0302, "step": 1393 }, { "epoch": 12.907407407407408, "grad_norm": 0.9506133794784546, "learning_rate": 2.8819875776397514e-05, "loss": 0.0317, "step": 1394 }, { "epoch": 12.916666666666666, "grad_norm": 1.1184026002883911, "learning_rate": 2.8695652173913044e-05, "loss": 0.0448, "step": 1395 }, { "epoch": 12.925925925925926, "grad_norm": 1.0374799966812134, "learning_rate": 2.857142857142857e-05, "loss": 0.0428, "step": 1396 }, { "epoch": 12.935185185185185, "grad_norm": 1.3991471529006958, "learning_rate": 2.84472049689441e-05, "loss": 0.0761, "step": 1397 }, { "epoch": 12.944444444444445, "grad_norm": 1.7948813438415527, "learning_rate": 2.8322981366459627e-05, "loss": 0.0657, "step": 1398 }, { "epoch": 12.953703703703704, "grad_norm": 0.7981337308883667, "learning_rate": 2.8198757763975157e-05, "loss": 0.0256, "step": 1399 }, { "epoch": 12.962962962962964, "grad_norm": 1.0698860883712769, "learning_rate": 2.8074534161490684e-05, "loss": 0.0428, "step": 1400 }, { "epoch": 12.972222222222221, "grad_norm": 1.00368332862854, "learning_rate": 2.795031055900621e-05, "loss": 0.0353, "step": 1401 }, { "epoch": 12.981481481481481, "grad_norm": 0.9316563606262207, "learning_rate": 2.782608695652174e-05, "loss": 0.0312, "step": 1402 }, { "epoch": 12.99074074074074, "grad_norm": 0.9837682843208313, "learning_rate": 2.7701863354037267e-05, "loss": 0.039, "step": 1403 }, { "epoch": 13.0, "grad_norm": 1.0201270580291748, "learning_rate": 2.7577639751552796e-05, "loss": 0.0391, "step": 1404 }, { "epoch": 13.00925925925926, "grad_norm": 0.7964289784431458, "learning_rate": 2.7453416149068323e-05, "loss": 0.0294, "step": 1405 }, { "epoch": 13.018518518518519, "grad_norm": 0.5249894261360168, "learning_rate": 2.7329192546583853e-05, "loss": 0.0237, "step": 1406 }, { "epoch": 13.027777777777779, "grad_norm": 0.9201458096504211, "learning_rate": 2.720496894409938e-05, "loss": 0.0257, "step": 1407 }, { "epoch": 13.037037037037036, "grad_norm": 0.6795076131820679, "learning_rate": 2.708074534161491e-05, "loss": 0.0207, "step": 1408 }, { "epoch": 13.046296296296296, "grad_norm": 1.0227335691452026, "learning_rate": 2.6956521739130436e-05, "loss": 0.044, "step": 1409 }, { "epoch": 13.055555555555555, "grad_norm": 0.9504062533378601, "learning_rate": 2.6832298136645962e-05, "loss": 0.029, "step": 1410 }, { "epoch": 13.064814814814815, "grad_norm": 0.8904616832733154, "learning_rate": 2.6708074534161492e-05, "loss": 0.0286, "step": 1411 }, { "epoch": 13.074074074074074, "grad_norm": 1.118471384048462, "learning_rate": 2.658385093167702e-05, "loss": 0.0352, "step": 1412 }, { "epoch": 13.083333333333334, "grad_norm": 0.6861169934272766, "learning_rate": 2.645962732919255e-05, "loss": 0.0218, "step": 1413 }, { "epoch": 13.092592592592593, "grad_norm": 0.802804172039032, "learning_rate": 2.6335403726708075e-05, "loss": 0.0302, "step": 1414 }, { "epoch": 13.101851851851851, "grad_norm": 0.7278661727905273, "learning_rate": 2.6211180124223605e-05, "loss": 0.0251, "step": 1415 }, { "epoch": 13.11111111111111, "grad_norm": 0.9800047874450684, "learning_rate": 2.608695652173913e-05, "loss": 0.0274, "step": 1416 }, { "epoch": 13.12037037037037, "grad_norm": 0.8877214193344116, "learning_rate": 2.596273291925466e-05, "loss": 0.0292, "step": 1417 }, { "epoch": 13.12962962962963, "grad_norm": 0.8059203624725342, "learning_rate": 2.5838509316770188e-05, "loss": 0.0197, "step": 1418 }, { "epoch": 13.13888888888889, "grad_norm": 0.5480836033821106, "learning_rate": 2.5714285714285714e-05, "loss": 0.0181, "step": 1419 }, { "epoch": 13.148148148148149, "grad_norm": 0.649368166923523, "learning_rate": 2.5590062111801244e-05, "loss": 0.0182, "step": 1420 }, { "epoch": 13.157407407407407, "grad_norm": 0.8103269934654236, "learning_rate": 2.546583850931677e-05, "loss": 0.0229, "step": 1421 }, { "epoch": 13.166666666666666, "grad_norm": 0.8214313983917236, "learning_rate": 2.53416149068323e-05, "loss": 0.0293, "step": 1422 }, { "epoch": 13.175925925925926, "grad_norm": 0.7193138003349304, "learning_rate": 2.5217391304347827e-05, "loss": 0.0253, "step": 1423 }, { "epoch": 13.185185185185185, "grad_norm": 0.7240740656852722, "learning_rate": 2.5093167701863357e-05, "loss": 0.0251, "step": 1424 }, { "epoch": 13.194444444444445, "grad_norm": 0.720065176486969, "learning_rate": 2.4968944099378884e-05, "loss": 0.0207, "step": 1425 }, { "epoch": 13.203703703703704, "grad_norm": 0.49787089228630066, "learning_rate": 2.484472049689441e-05, "loss": 0.0184, "step": 1426 }, { "epoch": 13.212962962962964, "grad_norm": 0.836642861366272, "learning_rate": 2.472049689440994e-05, "loss": 0.0318, "step": 1427 }, { "epoch": 13.222222222222221, "grad_norm": 0.7410364151000977, "learning_rate": 2.4596273291925467e-05, "loss": 0.0254, "step": 1428 }, { "epoch": 13.231481481481481, "grad_norm": 0.7820467352867126, "learning_rate": 2.4472049689440996e-05, "loss": 0.0288, "step": 1429 }, { "epoch": 13.24074074074074, "grad_norm": 0.8425258994102478, "learning_rate": 2.4347826086956523e-05, "loss": 0.0349, "step": 1430 }, { "epoch": 13.25, "grad_norm": 0.891117513179779, "learning_rate": 2.4223602484472053e-05, "loss": 0.0277, "step": 1431 }, { "epoch": 13.25925925925926, "grad_norm": 0.9912608861923218, "learning_rate": 2.4099378881987576e-05, "loss": 0.0338, "step": 1432 }, { "epoch": 13.268518518518519, "grad_norm": 0.8655397295951843, "learning_rate": 2.3975155279503106e-05, "loss": 0.0341, "step": 1433 }, { "epoch": 13.277777777777779, "grad_norm": 0.8703702688217163, "learning_rate": 2.3850931677018636e-05, "loss": 0.0301, "step": 1434 }, { "epoch": 13.287037037037036, "grad_norm": 0.816727340221405, "learning_rate": 2.3726708074534162e-05, "loss": 0.0268, "step": 1435 }, { "epoch": 13.296296296296296, "grad_norm": 0.6376839280128479, "learning_rate": 2.3602484472049692e-05, "loss": 0.0251, "step": 1436 }, { "epoch": 13.305555555555555, "grad_norm": 1.3493186235427856, "learning_rate": 2.347826086956522e-05, "loss": 0.046, "step": 1437 }, { "epoch": 13.314814814814815, "grad_norm": 1.1794747114181519, "learning_rate": 2.335403726708075e-05, "loss": 0.0458, "step": 1438 }, { "epoch": 13.324074074074074, "grad_norm": 0.8929194211959839, "learning_rate": 2.3229813664596275e-05, "loss": 0.0345, "step": 1439 }, { "epoch": 13.333333333333334, "grad_norm": 0.7308538556098938, "learning_rate": 2.31055900621118e-05, "loss": 0.0246, "step": 1440 }, { "epoch": 13.342592592592593, "grad_norm": 0.701432466506958, "learning_rate": 2.2981366459627328e-05, "loss": 0.0233, "step": 1441 }, { "epoch": 13.351851851851851, "grad_norm": 0.7073217630386353, "learning_rate": 2.2857142857142858e-05, "loss": 0.0255, "step": 1442 }, { "epoch": 13.36111111111111, "grad_norm": 0.8080083727836609, "learning_rate": 2.2732919254658385e-05, "loss": 0.0322, "step": 1443 }, { "epoch": 13.37037037037037, "grad_norm": 0.8558483719825745, "learning_rate": 2.2608695652173914e-05, "loss": 0.0289, "step": 1444 }, { "epoch": 13.37962962962963, "grad_norm": 0.7872486710548401, "learning_rate": 2.248447204968944e-05, "loss": 0.0253, "step": 1445 }, { "epoch": 13.38888888888889, "grad_norm": 0.6285867094993591, "learning_rate": 2.236024844720497e-05, "loss": 0.0205, "step": 1446 }, { "epoch": 13.398148148148149, "grad_norm": 0.9971258044242859, "learning_rate": 2.2236024844720497e-05, "loss": 0.0332, "step": 1447 }, { "epoch": 13.407407407407407, "grad_norm": 1.0363770723342896, "learning_rate": 2.2111801242236027e-05, "loss": 0.0383, "step": 1448 }, { "epoch": 13.416666666666666, "grad_norm": 1.03330397605896, "learning_rate": 2.1987577639751554e-05, "loss": 0.0288, "step": 1449 }, { "epoch": 13.425925925925926, "grad_norm": 0.7237390279769897, "learning_rate": 2.186335403726708e-05, "loss": 0.0265, "step": 1450 }, { "epoch": 13.435185185185185, "grad_norm": 0.8273149132728577, "learning_rate": 2.173913043478261e-05, "loss": 0.024, "step": 1451 }, { "epoch": 13.444444444444445, "grad_norm": 0.7203170657157898, "learning_rate": 2.1614906832298137e-05, "loss": 0.0253, "step": 1452 }, { "epoch": 13.453703703703704, "grad_norm": 0.970585823059082, "learning_rate": 2.1490683229813667e-05, "loss": 0.0339, "step": 1453 }, { "epoch": 13.462962962962964, "grad_norm": 0.783500075340271, "learning_rate": 2.1366459627329193e-05, "loss": 0.024, "step": 1454 }, { "epoch": 13.472222222222221, "grad_norm": 0.7558131814002991, "learning_rate": 2.1242236024844723e-05, "loss": 0.0225, "step": 1455 }, { "epoch": 13.481481481481481, "grad_norm": 1.1136420965194702, "learning_rate": 2.111801242236025e-05, "loss": 0.0336, "step": 1456 }, { "epoch": 13.49074074074074, "grad_norm": 0.7491280436515808, "learning_rate": 2.099378881987578e-05, "loss": 0.0282, "step": 1457 }, { "epoch": 13.5, "grad_norm": 0.8197646737098694, "learning_rate": 2.0869565217391303e-05, "loss": 0.0323, "step": 1458 }, { "epoch": 13.50925925925926, "grad_norm": 1.0424060821533203, "learning_rate": 2.0745341614906832e-05, "loss": 0.0343, "step": 1459 }, { "epoch": 13.518518518518519, "grad_norm": 0.8803132176399231, "learning_rate": 2.0621118012422362e-05, "loss": 0.0321, "step": 1460 }, { "epoch": 13.527777777777779, "grad_norm": 0.8206008672714233, "learning_rate": 2.049689440993789e-05, "loss": 0.0266, "step": 1461 }, { "epoch": 13.537037037037036, "grad_norm": 0.9388163685798645, "learning_rate": 2.037267080745342e-05, "loss": 0.034, "step": 1462 }, { "epoch": 13.546296296296296, "grad_norm": 0.7545292973518372, "learning_rate": 2.0248447204968945e-05, "loss": 0.0308, "step": 1463 }, { "epoch": 13.555555555555555, "grad_norm": 0.9023415446281433, "learning_rate": 2.0124223602484475e-05, "loss": 0.0329, "step": 1464 }, { "epoch": 13.564814814814815, "grad_norm": 0.8173019886016846, "learning_rate": 2e-05, "loss": 0.0303, "step": 1465 }, { "epoch": 13.574074074074074, "grad_norm": 1.1476924419403076, "learning_rate": 1.9875776397515528e-05, "loss": 0.0428, "step": 1466 }, { "epoch": 13.583333333333334, "grad_norm": 1.10149085521698, "learning_rate": 1.9751552795031055e-05, "loss": 0.0377, "step": 1467 }, { "epoch": 13.592592592592592, "grad_norm": 0.8061996102333069, "learning_rate": 1.9627329192546585e-05, "loss": 0.031, "step": 1468 }, { "epoch": 13.601851851851851, "grad_norm": 0.982791543006897, "learning_rate": 1.950310559006211e-05, "loss": 0.0341, "step": 1469 }, { "epoch": 13.61111111111111, "grad_norm": 0.7261083126068115, "learning_rate": 1.937888198757764e-05, "loss": 0.0269, "step": 1470 }, { "epoch": 13.62037037037037, "grad_norm": 1.003458857536316, "learning_rate": 1.9254658385093167e-05, "loss": 0.0344, "step": 1471 }, { "epoch": 13.62962962962963, "grad_norm": 0.9957963228225708, "learning_rate": 1.9130434782608697e-05, "loss": 0.0359, "step": 1472 }, { "epoch": 13.63888888888889, "grad_norm": 0.9977306127548218, "learning_rate": 1.9006211180124224e-05, "loss": 0.0398, "step": 1473 }, { "epoch": 13.648148148148149, "grad_norm": 0.6655192375183105, "learning_rate": 1.8881987577639754e-05, "loss": 0.0252, "step": 1474 }, { "epoch": 13.657407407407408, "grad_norm": 0.9726771712303162, "learning_rate": 1.875776397515528e-05, "loss": 0.0345, "step": 1475 }, { "epoch": 13.666666666666666, "grad_norm": 0.9408605694770813, "learning_rate": 1.8633540372670807e-05, "loss": 0.0295, "step": 1476 }, { "epoch": 13.675925925925926, "grad_norm": 0.9074751138687134, "learning_rate": 1.8509316770186337e-05, "loss": 0.0294, "step": 1477 }, { "epoch": 13.685185185185185, "grad_norm": 0.5638766884803772, "learning_rate": 1.8385093167701863e-05, "loss": 0.0215, "step": 1478 }, { "epoch": 13.694444444444445, "grad_norm": 1.9929472208023071, "learning_rate": 1.8260869565217393e-05, "loss": 0.0977, "step": 1479 }, { "epoch": 13.703703703703704, "grad_norm": 0.645876944065094, "learning_rate": 1.813664596273292e-05, "loss": 0.0319, "step": 1480 }, { "epoch": 13.712962962962964, "grad_norm": 0.8106043338775635, "learning_rate": 1.801242236024845e-05, "loss": 0.0333, "step": 1481 }, { "epoch": 13.722222222222221, "grad_norm": 0.9891499280929565, "learning_rate": 1.7888198757763976e-05, "loss": 0.0348, "step": 1482 }, { "epoch": 13.731481481481481, "grad_norm": 0.7550816535949707, "learning_rate": 1.7763975155279506e-05, "loss": 0.0309, "step": 1483 }, { "epoch": 13.74074074074074, "grad_norm": 0.7451123595237732, "learning_rate": 1.7639751552795032e-05, "loss": 0.0278, "step": 1484 }, { "epoch": 13.75, "grad_norm": 0.8352628946304321, "learning_rate": 1.751552795031056e-05, "loss": 0.0274, "step": 1485 }, { "epoch": 13.75925925925926, "grad_norm": 0.8166201710700989, "learning_rate": 1.739130434782609e-05, "loss": 0.026, "step": 1486 }, { "epoch": 13.768518518518519, "grad_norm": 1.422154426574707, "learning_rate": 1.7267080745341615e-05, "loss": 0.0392, "step": 1487 }, { "epoch": 13.777777777777779, "grad_norm": 0.8086047172546387, "learning_rate": 1.7142857142857145e-05, "loss": 0.0286, "step": 1488 }, { "epoch": 13.787037037037036, "grad_norm": 0.9935398101806641, "learning_rate": 1.7018633540372672e-05, "loss": 0.0326, "step": 1489 }, { "epoch": 13.796296296296296, "grad_norm": 0.8330230712890625, "learning_rate": 1.68944099378882e-05, "loss": 0.0268, "step": 1490 }, { "epoch": 13.805555555555555, "grad_norm": 0.8333765268325806, "learning_rate": 1.6770186335403728e-05, "loss": 0.0286, "step": 1491 }, { "epoch": 13.814814814814815, "grad_norm": 0.641851007938385, "learning_rate": 1.6645962732919255e-05, "loss": 0.0285, "step": 1492 }, { "epoch": 13.824074074074074, "grad_norm": 0.789412260055542, "learning_rate": 1.652173913043478e-05, "loss": 0.0288, "step": 1493 }, { "epoch": 13.833333333333334, "grad_norm": 0.9055834412574768, "learning_rate": 1.639751552795031e-05, "loss": 0.0278, "step": 1494 }, { "epoch": 13.842592592592592, "grad_norm": 0.7134174108505249, "learning_rate": 1.6273291925465838e-05, "loss": 0.0292, "step": 1495 }, { "epoch": 13.851851851851851, "grad_norm": 1.001650333404541, "learning_rate": 1.6149068322981367e-05, "loss": 0.0345, "step": 1496 }, { "epoch": 13.86111111111111, "grad_norm": 0.800206184387207, "learning_rate": 1.6024844720496894e-05, "loss": 0.0304, "step": 1497 }, { "epoch": 13.87037037037037, "grad_norm": 0.8903914093971252, "learning_rate": 1.5900621118012424e-05, "loss": 0.0362, "step": 1498 }, { "epoch": 13.87962962962963, "grad_norm": 0.7213857769966125, "learning_rate": 1.577639751552795e-05, "loss": 0.0317, "step": 1499 }, { "epoch": 13.88888888888889, "grad_norm": 0.7966362237930298, "learning_rate": 1.565217391304348e-05, "loss": 0.0238, "step": 1500 }, { "epoch": 13.898148148148149, "grad_norm": 0.877825140953064, "learning_rate": 1.5527950310559007e-05, "loss": 0.0293, "step": 1501 }, { "epoch": 13.907407407407408, "grad_norm": 1.0245945453643799, "learning_rate": 1.5403726708074533e-05, "loss": 0.038, "step": 1502 }, { "epoch": 13.916666666666666, "grad_norm": 0.9006460309028625, "learning_rate": 1.5279503105590063e-05, "loss": 0.0375, "step": 1503 }, { "epoch": 13.925925925925926, "grad_norm": 1.0009137392044067, "learning_rate": 1.515527950310559e-05, "loss": 0.036, "step": 1504 }, { "epoch": 13.935185185185185, "grad_norm": 1.1247698068618774, "learning_rate": 1.503105590062112e-05, "loss": 0.0462, "step": 1505 }, { "epoch": 13.944444444444445, "grad_norm": 1.1003761291503906, "learning_rate": 1.4906832298136648e-05, "loss": 0.0386, "step": 1506 }, { "epoch": 13.953703703703704, "grad_norm": 0.9153029918670654, "learning_rate": 1.4782608695652176e-05, "loss": 0.0305, "step": 1507 }, { "epoch": 13.962962962962964, "grad_norm": 0.6955792903900146, "learning_rate": 1.4658385093167704e-05, "loss": 0.0211, "step": 1508 }, { "epoch": 13.972222222222221, "grad_norm": 0.918514609336853, "learning_rate": 1.4534161490683232e-05, "loss": 0.0296, "step": 1509 }, { "epoch": 13.981481481481481, "grad_norm": 0.7454483509063721, "learning_rate": 1.4409937888198757e-05, "loss": 0.029, "step": 1510 }, { "epoch": 13.99074074074074, "grad_norm": 0.7576525211334229, "learning_rate": 1.4285714285714285e-05, "loss": 0.0293, "step": 1511 }, { "epoch": 14.0, "grad_norm": 0.9357626438140869, "learning_rate": 1.4161490683229814e-05, "loss": 0.0339, "step": 1512 }, { "epoch": 14.00925925925926, "grad_norm": 1.0074899196624756, "learning_rate": 1.4037267080745342e-05, "loss": 0.0338, "step": 1513 }, { "epoch": 14.018518518518519, "grad_norm": 0.5901816487312317, "learning_rate": 1.391304347826087e-05, "loss": 0.0201, "step": 1514 }, { "epoch": 14.027777777777779, "grad_norm": 0.6215850710868835, "learning_rate": 1.3788819875776398e-05, "loss": 0.0256, "step": 1515 }, { "epoch": 14.037037037037036, "grad_norm": 0.7354216575622559, "learning_rate": 1.3664596273291926e-05, "loss": 0.0289, "step": 1516 }, { "epoch": 14.046296296296296, "grad_norm": 0.5978808403015137, "learning_rate": 1.3540372670807455e-05, "loss": 0.0202, "step": 1517 }, { "epoch": 14.055555555555555, "grad_norm": 0.7752763628959656, "learning_rate": 1.3416149068322981e-05, "loss": 0.0275, "step": 1518 }, { "epoch": 14.064814814814815, "grad_norm": 0.8024547100067139, "learning_rate": 1.329192546583851e-05, "loss": 0.0254, "step": 1519 }, { "epoch": 14.074074074074074, "grad_norm": 0.7298763394355774, "learning_rate": 1.3167701863354038e-05, "loss": 0.028, "step": 1520 }, { "epoch": 14.083333333333334, "grad_norm": 0.4958341419696808, "learning_rate": 1.3043478260869566e-05, "loss": 0.0221, "step": 1521 }, { "epoch": 14.092592592592593, "grad_norm": 0.9592882990837097, "learning_rate": 1.2919254658385094e-05, "loss": 0.03, "step": 1522 }, { "epoch": 14.101851851851851, "grad_norm": 0.9992770552635193, "learning_rate": 1.2795031055900622e-05, "loss": 0.0291, "step": 1523 }, { "epoch": 14.11111111111111, "grad_norm": 0.7901219129562378, "learning_rate": 1.267080745341615e-05, "loss": 0.0263, "step": 1524 }, { "epoch": 14.12037037037037, "grad_norm": 0.7581226825714111, "learning_rate": 1.2546583850931679e-05, "loss": 0.0241, "step": 1525 }, { "epoch": 14.12962962962963, "grad_norm": 0.9516726732254028, "learning_rate": 1.2422360248447205e-05, "loss": 0.0302, "step": 1526 }, { "epoch": 14.13888888888889, "grad_norm": 0.791049063205719, "learning_rate": 1.2298136645962733e-05, "loss": 0.0224, "step": 1527 }, { "epoch": 14.148148148148149, "grad_norm": 0.8296195268630981, "learning_rate": 1.2173913043478261e-05, "loss": 0.0274, "step": 1528 }, { "epoch": 14.157407407407407, "grad_norm": 0.6744861602783203, "learning_rate": 1.2049689440993788e-05, "loss": 0.0216, "step": 1529 }, { "epoch": 14.166666666666666, "grad_norm": 0.7672494649887085, "learning_rate": 1.1925465838509318e-05, "loss": 0.0228, "step": 1530 }, { "epoch": 14.175925925925926, "grad_norm": 0.7237677574157715, "learning_rate": 1.1801242236024846e-05, "loss": 0.0228, "step": 1531 }, { "epoch": 14.185185185185185, "grad_norm": 1.0658645629882812, "learning_rate": 1.1677018633540374e-05, "loss": 0.0349, "step": 1532 }, { "epoch": 14.194444444444445, "grad_norm": 0.8038221001625061, "learning_rate": 1.15527950310559e-05, "loss": 0.0258, "step": 1533 }, { "epoch": 14.203703703703704, "grad_norm": 0.7512486577033997, "learning_rate": 1.1428571428571429e-05, "loss": 0.027, "step": 1534 }, { "epoch": 14.212962962962964, "grad_norm": 0.7751714587211609, "learning_rate": 1.1304347826086957e-05, "loss": 0.0275, "step": 1535 }, { "epoch": 14.222222222222221, "grad_norm": 0.8435266017913818, "learning_rate": 1.1180124223602485e-05, "loss": 0.0288, "step": 1536 }, { "epoch": 14.231481481481481, "grad_norm": 0.7947162985801697, "learning_rate": 1.1055900621118014e-05, "loss": 0.0274, "step": 1537 }, { "epoch": 14.24074074074074, "grad_norm": 0.6701507568359375, "learning_rate": 1.093167701863354e-05, "loss": 0.0198, "step": 1538 }, { "epoch": 14.25, "grad_norm": 0.6321646571159363, "learning_rate": 1.0807453416149068e-05, "loss": 0.0212, "step": 1539 }, { "epoch": 14.25925925925926, "grad_norm": 0.6891692280769348, "learning_rate": 1.0683229813664597e-05, "loss": 0.0232, "step": 1540 }, { "epoch": 14.268518518518519, "grad_norm": 0.9563750624656677, "learning_rate": 1.0559006211180125e-05, "loss": 0.0315, "step": 1541 }, { "epoch": 14.277777777777779, "grad_norm": 0.7779302597045898, "learning_rate": 1.0434782608695651e-05, "loss": 0.0244, "step": 1542 }, { "epoch": 14.287037037037036, "grad_norm": 0.8213071823120117, "learning_rate": 1.0310559006211181e-05, "loss": 0.0267, "step": 1543 }, { "epoch": 14.296296296296296, "grad_norm": 0.7205976843833923, "learning_rate": 1.018633540372671e-05, "loss": 0.0237, "step": 1544 }, { "epoch": 14.305555555555555, "grad_norm": 0.8843669295310974, "learning_rate": 1.0062111801242238e-05, "loss": 0.0264, "step": 1545 }, { "epoch": 14.314814814814815, "grad_norm": 1.0456665754318237, "learning_rate": 9.937888198757764e-06, "loss": 0.0357, "step": 1546 }, { "epoch": 14.324074074074074, "grad_norm": 1.090004324913025, "learning_rate": 9.813664596273292e-06, "loss": 0.0351, "step": 1547 }, { "epoch": 14.333333333333334, "grad_norm": 0.9663504958152771, "learning_rate": 9.68944099378882e-06, "loss": 0.0285, "step": 1548 }, { "epoch": 14.342592592592593, "grad_norm": 0.6734511256217957, "learning_rate": 9.565217391304349e-06, "loss": 0.023, "step": 1549 }, { "epoch": 14.351851851851851, "grad_norm": 1.1340980529785156, "learning_rate": 9.440993788819877e-06, "loss": 0.035, "step": 1550 }, { "epoch": 14.36111111111111, "grad_norm": 0.6549594402313232, "learning_rate": 9.316770186335403e-06, "loss": 0.0184, "step": 1551 }, { "epoch": 14.37037037037037, "grad_norm": 0.732521116733551, "learning_rate": 9.192546583850932e-06, "loss": 0.0247, "step": 1552 }, { "epoch": 14.37962962962963, "grad_norm": 0.8055844306945801, "learning_rate": 9.06832298136646e-06, "loss": 0.0307, "step": 1553 }, { "epoch": 14.38888888888889, "grad_norm": 0.8509539365768433, "learning_rate": 8.944099378881988e-06, "loss": 0.0283, "step": 1554 }, { "epoch": 14.398148148148149, "grad_norm": 0.6740679144859314, "learning_rate": 8.819875776397516e-06, "loss": 0.0235, "step": 1555 }, { "epoch": 14.407407407407407, "grad_norm": 0.8333950638771057, "learning_rate": 8.695652173913044e-06, "loss": 0.0296, "step": 1556 }, { "epoch": 14.416666666666666, "grad_norm": 1.1533918380737305, "learning_rate": 8.571428571428573e-06, "loss": 0.0308, "step": 1557 }, { "epoch": 14.425925925925926, "grad_norm": 0.7060422301292419, "learning_rate": 8.4472049689441e-06, "loss": 0.0244, "step": 1558 }, { "epoch": 14.435185185185185, "grad_norm": 0.9376555681228638, "learning_rate": 8.322981366459627e-06, "loss": 0.027, "step": 1559 }, { "epoch": 14.444444444444445, "grad_norm": 1.0092798471450806, "learning_rate": 8.198757763975156e-06, "loss": 0.0291, "step": 1560 }, { "epoch": 14.453703703703704, "grad_norm": 0.8119834661483765, "learning_rate": 8.074534161490684e-06, "loss": 0.0254, "step": 1561 }, { "epoch": 14.462962962962964, "grad_norm": 1.0415433645248413, "learning_rate": 7.950310559006212e-06, "loss": 0.032, "step": 1562 }, { "epoch": 14.472222222222221, "grad_norm": 0.9587968587875366, "learning_rate": 7.82608695652174e-06, "loss": 0.0321, "step": 1563 }, { "epoch": 14.481481481481481, "grad_norm": 0.9727386832237244, "learning_rate": 7.701863354037267e-06, "loss": 0.0344, "step": 1564 }, { "epoch": 14.49074074074074, "grad_norm": 0.8076003193855286, "learning_rate": 7.577639751552795e-06, "loss": 0.0255, "step": 1565 }, { "epoch": 14.5, "grad_norm": 0.7089889049530029, "learning_rate": 7.453416149068324e-06, "loss": 0.0239, "step": 1566 }, { "epoch": 14.50925925925926, "grad_norm": 0.6875848174095154, "learning_rate": 7.329192546583852e-06, "loss": 0.0257, "step": 1567 }, { "epoch": 14.518518518518519, "grad_norm": 0.7025448679924011, "learning_rate": 7.204968944099379e-06, "loss": 0.0218, "step": 1568 }, { "epoch": 14.527777777777779, "grad_norm": 0.686010479927063, "learning_rate": 7.080745341614907e-06, "loss": 0.0204, "step": 1569 }, { "epoch": 14.537037037037036, "grad_norm": 0.8934335708618164, "learning_rate": 6.956521739130435e-06, "loss": 0.0291, "step": 1570 }, { "epoch": 14.546296296296296, "grad_norm": 0.7375936508178711, "learning_rate": 6.832298136645963e-06, "loss": 0.024, "step": 1571 }, { "epoch": 14.555555555555555, "grad_norm": 0.865662693977356, "learning_rate": 6.708074534161491e-06, "loss": 0.0263, "step": 1572 }, { "epoch": 14.564814814814815, "grad_norm": 0.6800843477249146, "learning_rate": 6.583850931677019e-06, "loss": 0.0247, "step": 1573 }, { "epoch": 14.574074074074074, "grad_norm": 0.8272568583488464, "learning_rate": 6.459627329192547e-06, "loss": 0.0316, "step": 1574 }, { "epoch": 14.583333333333334, "grad_norm": 0.5861930251121521, "learning_rate": 6.335403726708075e-06, "loss": 0.0217, "step": 1575 }, { "epoch": 14.592592592592592, "grad_norm": 0.6214504241943359, "learning_rate": 6.2111801242236025e-06, "loss": 0.0212, "step": 1576 }, { "epoch": 14.601851851851851, "grad_norm": 0.5459976196289062, "learning_rate": 6.086956521739131e-06, "loss": 0.0189, "step": 1577 }, { "epoch": 14.61111111111111, "grad_norm": 1.0046416521072388, "learning_rate": 5.962732919254659e-06, "loss": 0.0337, "step": 1578 }, { "epoch": 14.62037037037037, "grad_norm": 0.7500995993614197, "learning_rate": 5.838509316770187e-06, "loss": 0.0259, "step": 1579 }, { "epoch": 14.62962962962963, "grad_norm": 0.8355032801628113, "learning_rate": 5.7142857142857145e-06, "loss": 0.0255, "step": 1580 }, { "epoch": 14.63888888888889, "grad_norm": 2.543530225753784, "learning_rate": 5.590062111801243e-06, "loss": 0.0551, "step": 1581 }, { "epoch": 14.648148148148149, "grad_norm": 0.7404892444610596, "learning_rate": 5.46583850931677e-06, "loss": 0.0234, "step": 1582 }, { "epoch": 14.657407407407408, "grad_norm": 0.8910460472106934, "learning_rate": 5.341614906832298e-06, "loss": 0.026, "step": 1583 }, { "epoch": 14.666666666666666, "grad_norm": 0.7894750237464905, "learning_rate": 5.217391304347826e-06, "loss": 0.031, "step": 1584 }, { "epoch": 14.675925925925926, "grad_norm": 0.719997763633728, "learning_rate": 5.093167701863355e-06, "loss": 0.0236, "step": 1585 }, { "epoch": 14.685185185185185, "grad_norm": 0.8099305629730225, "learning_rate": 4.968944099378882e-06, "loss": 0.0272, "step": 1586 }, { "epoch": 14.694444444444445, "grad_norm": 0.9355957508087158, "learning_rate": 4.84472049689441e-06, "loss": 0.0279, "step": 1587 }, { "epoch": 14.703703703703704, "grad_norm": 0.9016175270080566, "learning_rate": 4.7204968944099384e-06, "loss": 0.0275, "step": 1588 }, { "epoch": 14.712962962962964, "grad_norm": 0.8064716458320618, "learning_rate": 4.596273291925466e-06, "loss": 0.0275, "step": 1589 }, { "epoch": 14.722222222222221, "grad_norm": 0.7592180967330933, "learning_rate": 4.472049689440994e-06, "loss": 0.0246, "step": 1590 }, { "epoch": 14.731481481481481, "grad_norm": 0.8269107341766357, "learning_rate": 4.347826086956522e-06, "loss": 0.0275, "step": 1591 }, { "epoch": 14.74074074074074, "grad_norm": 0.8514775633811951, "learning_rate": 4.22360248447205e-06, "loss": 0.0282, "step": 1592 }, { "epoch": 14.75, "grad_norm": 0.7846024632453918, "learning_rate": 4.099378881987578e-06, "loss": 0.0235, "step": 1593 }, { "epoch": 14.75925925925926, "grad_norm": 0.8372007012367249, "learning_rate": 3.975155279503106e-06, "loss": 0.0286, "step": 1594 }, { "epoch": 14.768518518518519, "grad_norm": 0.965676486492157, "learning_rate": 3.850931677018633e-06, "loss": 0.0316, "step": 1595 }, { "epoch": 14.777777777777779, "grad_norm": 0.925922691822052, "learning_rate": 3.726708074534162e-06, "loss": 0.0281, "step": 1596 }, { "epoch": 14.787037037037036, "grad_norm": 1.0515544414520264, "learning_rate": 3.6024844720496893e-06, "loss": 0.031, "step": 1597 }, { "epoch": 14.796296296296296, "grad_norm": 0.88392573595047, "learning_rate": 3.4782608695652175e-06, "loss": 0.0282, "step": 1598 }, { "epoch": 14.805555555555555, "grad_norm": 0.984674334526062, "learning_rate": 3.3540372670807453e-06, "loss": 0.0273, "step": 1599 }, { "epoch": 14.814814814814815, "grad_norm": 0.7821362018585205, "learning_rate": 3.2298136645962735e-06, "loss": 0.0295, "step": 1600 }, { "epoch": 14.824074074074074, "grad_norm": 0.9542562961578369, "learning_rate": 3.1055900621118013e-06, "loss": 0.0296, "step": 1601 }, { "epoch": 14.833333333333334, "grad_norm": 1.00540292263031, "learning_rate": 2.9813664596273295e-06, "loss": 0.0302, "step": 1602 }, { "epoch": 14.842592592592592, "grad_norm": 0.8111598491668701, "learning_rate": 2.8571428571428573e-06, "loss": 0.0231, "step": 1603 }, { "epoch": 14.851851851851851, "grad_norm": 0.7784263491630554, "learning_rate": 2.732919254658385e-06, "loss": 0.0276, "step": 1604 }, { "epoch": 14.86111111111111, "grad_norm": 0.6490851044654846, "learning_rate": 2.608695652173913e-06, "loss": 0.0196, "step": 1605 }, { "epoch": 14.87037037037037, "grad_norm": 0.8383249640464783, "learning_rate": 2.484472049689441e-06, "loss": 0.0244, "step": 1606 }, { "epoch": 14.87962962962963, "grad_norm": 0.7719290852546692, "learning_rate": 2.3602484472049692e-06, "loss": 0.0288, "step": 1607 }, { "epoch": 14.88888888888889, "grad_norm": 0.9684656858444214, "learning_rate": 2.236024844720497e-06, "loss": 0.0296, "step": 1608 }, { "epoch": 14.898148148148149, "grad_norm": 0.8234147429466248, "learning_rate": 2.111801242236025e-06, "loss": 0.0237, "step": 1609 }, { "epoch": 14.907407407407408, "grad_norm": 0.7515906095504761, "learning_rate": 1.987577639751553e-06, "loss": 0.0281, "step": 1610 }, { "epoch": 14.916666666666666, "grad_norm": 0.7840167284011841, "learning_rate": 1.863354037267081e-06, "loss": 0.0239, "step": 1611 }, { "epoch": 14.925925925925926, "grad_norm": 0.720740020275116, "learning_rate": 1.7391304347826088e-06, "loss": 0.0249, "step": 1612 }, { "epoch": 14.935185185185185, "grad_norm": 0.9448007941246033, "learning_rate": 1.6149068322981367e-06, "loss": 0.0301, "step": 1613 }, { "epoch": 14.944444444444445, "grad_norm": 0.5331466197967529, "learning_rate": 1.4906832298136647e-06, "loss": 0.0187, "step": 1614 }, { "epoch": 14.953703703703704, "grad_norm": 0.8185423016548157, "learning_rate": 1.3664596273291925e-06, "loss": 0.0309, "step": 1615 }, { "epoch": 14.962962962962964, "grad_norm": 0.684502124786377, "learning_rate": 1.2422360248447205e-06, "loss": 0.0211, "step": 1616 }, { "epoch": 14.972222222222221, "grad_norm": 0.8936087489128113, "learning_rate": 1.1180124223602485e-06, "loss": 0.0252, "step": 1617 }, { "epoch": 14.981481481481481, "grad_norm": 1.2246485948562622, "learning_rate": 9.937888198757765e-07, "loss": 0.0386, "step": 1618 }, { "epoch": 14.99074074074074, "grad_norm": 0.8689442873001099, "learning_rate": 8.695652173913044e-07, "loss": 0.0265, "step": 1619 }, { "epoch": 15.0, "grad_norm": 0.6490731239318848, "learning_rate": 7.453416149068324e-07, "loss": 0.0299, "step": 1620 } ], "logging_steps": 1, "max_steps": 1620, "num_input_tokens_seen": 0, "num_train_epochs": 15, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.598422110437376e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }