{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 5198, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00019238168526356292, "grad_norm": 3.7442002323252477, "learning_rate": 1.282051282051282e-06, "loss": 1.5495, "step": 1 }, { "epoch": 0.00038476337052712584, "grad_norm": 4.682249840607647, "learning_rate": 2.564102564102564e-06, "loss": 1.7634, "step": 2 }, { "epoch": 0.0005771450557906887, "grad_norm": 3.019913924739427, "learning_rate": 3.846153846153847e-06, "loss": 1.5305, "step": 3 }, { "epoch": 0.0007695267410542517, "grad_norm": 3.6308296366578214, "learning_rate": 5.128205128205128e-06, "loss": 1.5337, "step": 4 }, { "epoch": 0.0009619084263178145, "grad_norm": 3.423254283495366, "learning_rate": 6.41025641025641e-06, "loss": 1.537, "step": 5 }, { "epoch": 0.0011542901115813775, "grad_norm": 3.3775224852097008, "learning_rate": 7.692307692307694e-06, "loss": 1.5101, "step": 6 }, { "epoch": 0.0013466717968449403, "grad_norm": 2.5826408975649695, "learning_rate": 8.974358974358976e-06, "loss": 1.4474, "step": 7 }, { "epoch": 0.0015390534821085034, "grad_norm": 1.878800621928004, "learning_rate": 1.0256410256410256e-05, "loss": 1.3364, "step": 8 }, { "epoch": 0.0017314351673720662, "grad_norm": 1.396826828578569, "learning_rate": 1.153846153846154e-05, "loss": 1.3731, "step": 9 }, { "epoch": 0.001923816852635629, "grad_norm": 1.280362318277793, "learning_rate": 1.282051282051282e-05, "loss": 1.2524, "step": 10 }, { "epoch": 0.002116198537899192, "grad_norm": 1.1572836561609374, "learning_rate": 1.4102564102564104e-05, "loss": 1.1698, "step": 11 }, { "epoch": 0.002308580223162755, "grad_norm": 1.3299288924704236, "learning_rate": 1.5384615384615387e-05, "loss": 1.1991, "step": 12 }, { "epoch": 0.002500961908426318, "grad_norm": 0.7533802534077734, "learning_rate": 1.6666666666666667e-05, "loss": 1.1956, "step": 13 }, { "epoch": 0.0026933435936898806, "grad_norm": 0.6165549136921542, "learning_rate": 1.794871794871795e-05, "loss": 1.0832, "step": 14 }, { "epoch": 0.0028857252789534437, "grad_norm": 0.5866888743768268, "learning_rate": 1.923076923076923e-05, "loss": 1.141, "step": 15 }, { "epoch": 0.0030781069642170067, "grad_norm": 0.5188530346376825, "learning_rate": 2.0512820512820512e-05, "loss": 1.1689, "step": 16 }, { "epoch": 0.0032704886494805694, "grad_norm": 0.5018665590625723, "learning_rate": 2.1794871794871795e-05, "loss": 0.6965, "step": 17 }, { "epoch": 0.0034628703347441324, "grad_norm": 0.43296881965049566, "learning_rate": 2.307692307692308e-05, "loss": 1.0603, "step": 18 }, { "epoch": 0.0036552520200076955, "grad_norm": 0.4376159086534617, "learning_rate": 2.435897435897436e-05, "loss": 1.0841, "step": 19 }, { "epoch": 0.003847633705271258, "grad_norm": 0.4546329914972269, "learning_rate": 2.564102564102564e-05, "loss": 1.1068, "step": 20 }, { "epoch": 0.004040015390534821, "grad_norm": 0.4522795547049394, "learning_rate": 2.6923076923076923e-05, "loss": 1.0668, "step": 21 }, { "epoch": 0.004232397075798384, "grad_norm": 0.423894584816149, "learning_rate": 2.8205128205128207e-05, "loss": 1.1154, "step": 22 }, { "epoch": 0.004424778761061947, "grad_norm": 0.36865490224745834, "learning_rate": 2.948717948717949e-05, "loss": 0.9922, "step": 23 }, { "epoch": 0.00461716044632551, "grad_norm": 0.3343564778967025, "learning_rate": 3.0769230769230774e-05, "loss": 1.0684, "step": 24 }, { "epoch": 0.004809542131589073, "grad_norm": 0.29555526085445666, "learning_rate": 3.205128205128206e-05, "loss": 1.0478, "step": 25 }, { "epoch": 0.005001923816852636, "grad_norm": 0.29716140329316293, "learning_rate": 3.3333333333333335e-05, "loss": 1.0288, "step": 26 }, { "epoch": 0.005194305502116198, "grad_norm": 0.3301422747445203, "learning_rate": 3.461538461538462e-05, "loss": 1.0714, "step": 27 }, { "epoch": 0.005386687187379761, "grad_norm": 0.24848513029305336, "learning_rate": 3.58974358974359e-05, "loss": 1.0392, "step": 28 }, { "epoch": 0.005579068872643324, "grad_norm": 0.2749027806159161, "learning_rate": 3.717948717948718e-05, "loss": 1.006, "step": 29 }, { "epoch": 0.005771450557906887, "grad_norm": 0.2904854653367682, "learning_rate": 3.846153846153846e-05, "loss": 1.0518, "step": 30 }, { "epoch": 0.00596383224317045, "grad_norm": 0.3254054030699705, "learning_rate": 3.974358974358974e-05, "loss": 0.9948, "step": 31 }, { "epoch": 0.0061562139284340135, "grad_norm": 0.3049940911442011, "learning_rate": 4.1025641025641023e-05, "loss": 1.0782, "step": 32 }, { "epoch": 0.006348595613697576, "grad_norm": 0.2889991725046085, "learning_rate": 4.230769230769231e-05, "loss": 1.0116, "step": 33 }, { "epoch": 0.006540977298961139, "grad_norm": 0.26592474448691966, "learning_rate": 4.358974358974359e-05, "loss": 0.9752, "step": 34 }, { "epoch": 0.006733358984224702, "grad_norm": 0.25540910863777455, "learning_rate": 4.4871794871794874e-05, "loss": 0.9809, "step": 35 }, { "epoch": 0.006925740669488265, "grad_norm": 0.2453475555509423, "learning_rate": 4.615384615384616e-05, "loss": 0.9607, "step": 36 }, { "epoch": 0.007118122354751828, "grad_norm": 0.2598801685032486, "learning_rate": 4.7435897435897435e-05, "loss": 1.0164, "step": 37 }, { "epoch": 0.007310504040015391, "grad_norm": 0.263327875849137, "learning_rate": 4.871794871794872e-05, "loss": 0.9658, "step": 38 }, { "epoch": 0.007502885725278953, "grad_norm": 0.24281790534726377, "learning_rate": 5e-05, "loss": 1.0031, "step": 39 }, { "epoch": 0.007695267410542516, "grad_norm": 0.272568573070101, "learning_rate": 5.128205128205128e-05, "loss": 0.9565, "step": 40 }, { "epoch": 0.007887649095806078, "grad_norm": 0.24716341249726012, "learning_rate": 5.256410256410257e-05, "loss": 0.951, "step": 41 }, { "epoch": 0.008080030781069641, "grad_norm": 0.30553144338883054, "learning_rate": 5.384615384615385e-05, "loss": 0.9547, "step": 42 }, { "epoch": 0.008272412466333205, "grad_norm": 0.21716142359312623, "learning_rate": 5.512820512820514e-05, "loss": 0.9626, "step": 43 }, { "epoch": 0.008464794151596768, "grad_norm": 0.22786751373010689, "learning_rate": 5.6410256410256414e-05, "loss": 0.9413, "step": 44 }, { "epoch": 0.00865717583686033, "grad_norm": 0.2438350555798463, "learning_rate": 5.769230769230769e-05, "loss": 0.9466, "step": 45 }, { "epoch": 0.008849557522123894, "grad_norm": 0.2431349885260068, "learning_rate": 5.897435897435898e-05, "loss": 0.9857, "step": 46 }, { "epoch": 0.009041939207387457, "grad_norm": 0.22607808772093818, "learning_rate": 6.025641025641026e-05, "loss": 1.0205, "step": 47 }, { "epoch": 0.00923432089265102, "grad_norm": 0.24510433342202054, "learning_rate": 6.153846153846155e-05, "loss": 0.9118, "step": 48 }, { "epoch": 0.009426702577914583, "grad_norm": 0.23573785883560075, "learning_rate": 6.282051282051282e-05, "loss": 1.0052, "step": 49 }, { "epoch": 0.009619084263178146, "grad_norm": 0.22938462969391735, "learning_rate": 6.410256410256412e-05, "loss": 0.9841, "step": 50 }, { "epoch": 0.009811465948441709, "grad_norm": 0.24112023064506677, "learning_rate": 6.538461538461539e-05, "loss": 1.0006, "step": 51 }, { "epoch": 0.010003847633705272, "grad_norm": 0.2338679775808676, "learning_rate": 6.666666666666667e-05, "loss": 0.9382, "step": 52 }, { "epoch": 0.010196229318968833, "grad_norm": 0.23309066072277476, "learning_rate": 6.794871794871795e-05, "loss": 0.9425, "step": 53 }, { "epoch": 0.010388611004232396, "grad_norm": 0.22235674692063212, "learning_rate": 6.923076923076924e-05, "loss": 0.9449, "step": 54 }, { "epoch": 0.01058099268949596, "grad_norm": 0.26974907813619775, "learning_rate": 7.051282051282052e-05, "loss": 1.042, "step": 55 }, { "epoch": 0.010773374374759523, "grad_norm": 0.21848974913606237, "learning_rate": 7.17948717948718e-05, "loss": 0.9918, "step": 56 }, { "epoch": 0.010965756060023086, "grad_norm": 0.22323209045694725, "learning_rate": 7.307692307692307e-05, "loss": 0.964, "step": 57 }, { "epoch": 0.011158137745286649, "grad_norm": 0.2306507774985283, "learning_rate": 7.435897435897436e-05, "loss": 0.9792, "step": 58 }, { "epoch": 0.011350519430550212, "grad_norm": 0.2251483357295499, "learning_rate": 7.564102564102564e-05, "loss": 0.9631, "step": 59 }, { "epoch": 0.011542901115813775, "grad_norm": 0.23177565224939617, "learning_rate": 7.692307692307693e-05, "loss": 0.9403, "step": 60 }, { "epoch": 0.011735282801077338, "grad_norm": 0.2326475501748066, "learning_rate": 7.820512820512821e-05, "loss": 0.9568, "step": 61 }, { "epoch": 0.0119276644863409, "grad_norm": 0.20754597985609727, "learning_rate": 7.948717948717948e-05, "loss": 0.9964, "step": 62 }, { "epoch": 0.012120046171604464, "grad_norm": 0.2463298775842452, "learning_rate": 8.076923076923078e-05, "loss": 1.0076, "step": 63 }, { "epoch": 0.012312427856868027, "grad_norm": 0.24698587601028157, "learning_rate": 8.205128205128205e-05, "loss": 0.9507, "step": 64 }, { "epoch": 0.012504809542131588, "grad_norm": 0.23789053887034137, "learning_rate": 8.333333333333334e-05, "loss": 0.919, "step": 65 }, { "epoch": 0.012697191227395151, "grad_norm": 0.26330268459775147, "learning_rate": 8.461538461538461e-05, "loss": 0.9696, "step": 66 }, { "epoch": 0.012889572912658714, "grad_norm": 0.23309919247509012, "learning_rate": 8.58974358974359e-05, "loss": 0.9429, "step": 67 }, { "epoch": 0.013081954597922277, "grad_norm": 0.2296064470392471, "learning_rate": 8.717948717948718e-05, "loss": 0.9351, "step": 68 }, { "epoch": 0.01327433628318584, "grad_norm": 0.24761897643999844, "learning_rate": 8.846153846153847e-05, "loss": 1.0037, "step": 69 }, { "epoch": 0.013466717968449404, "grad_norm": 0.24959028959704216, "learning_rate": 8.974358974358975e-05, "loss": 0.9222, "step": 70 }, { "epoch": 0.013659099653712967, "grad_norm": 0.25433435376442104, "learning_rate": 9.102564102564103e-05, "loss": 0.9058, "step": 71 }, { "epoch": 0.01385148133897653, "grad_norm": 0.22964487592806523, "learning_rate": 9.230769230769232e-05, "loss": 0.9585, "step": 72 }, { "epoch": 0.014043863024240093, "grad_norm": 0.23396463836875264, "learning_rate": 9.35897435897436e-05, "loss": 0.9616, "step": 73 }, { "epoch": 0.014236244709503656, "grad_norm": 0.2368520873190386, "learning_rate": 9.487179487179487e-05, "loss": 0.9451, "step": 74 }, { "epoch": 0.014428626394767219, "grad_norm": 0.2557728790586544, "learning_rate": 9.615384615384617e-05, "loss": 0.9345, "step": 75 }, { "epoch": 0.014621008080030782, "grad_norm": 0.24845743271571732, "learning_rate": 9.743589743589744e-05, "loss": 0.9414, "step": 76 }, { "epoch": 0.014813389765294343, "grad_norm": 0.21717920521945372, "learning_rate": 9.871794871794872e-05, "loss": 0.9105, "step": 77 }, { "epoch": 0.015005771450557906, "grad_norm": 0.2024395633476369, "learning_rate": 0.0001, "loss": 0.9712, "step": 78 }, { "epoch": 0.01519815313582147, "grad_norm": 0.23413765736700232, "learning_rate": 0.00010128205128205129, "loss": 0.94, "step": 79 }, { "epoch": 0.015390534821085032, "grad_norm": 0.24830956775868968, "learning_rate": 0.00010256410256410256, "loss": 0.9034, "step": 80 }, { "epoch": 0.015582916506348595, "grad_norm": 0.2505189248649216, "learning_rate": 0.00010384615384615386, "loss": 0.9476, "step": 81 }, { "epoch": 0.015775298191612157, "grad_norm": 0.23994287846472534, "learning_rate": 0.00010512820512820514, "loss": 0.9095, "step": 82 }, { "epoch": 0.01596767987687572, "grad_norm": 0.23762649506838984, "learning_rate": 0.00010641025641025641, "loss": 0.901, "step": 83 }, { "epoch": 0.016160061562139283, "grad_norm": 0.24678212970803592, "learning_rate": 0.0001076923076923077, "loss": 0.9338, "step": 84 }, { "epoch": 0.016352443247402846, "grad_norm": 0.3542583340041871, "learning_rate": 0.00010897435897435896, "loss": 0.6755, "step": 85 }, { "epoch": 0.01654482493266641, "grad_norm": 0.2790684996874095, "learning_rate": 0.00011025641025641027, "loss": 0.9272, "step": 86 }, { "epoch": 0.016737206617929972, "grad_norm": 0.22974257950278487, "learning_rate": 0.00011153846153846154, "loss": 0.9439, "step": 87 }, { "epoch": 0.016929588303193535, "grad_norm": 0.2198436588397564, "learning_rate": 0.00011282051282051283, "loss": 0.9306, "step": 88 }, { "epoch": 0.017121969988457098, "grad_norm": 0.22111159264598154, "learning_rate": 0.0001141025641025641, "loss": 0.9517, "step": 89 }, { "epoch": 0.01731435167372066, "grad_norm": 0.2529480181887484, "learning_rate": 0.00011538461538461538, "loss": 0.9722, "step": 90 }, { "epoch": 0.017506733358984224, "grad_norm": 0.22507212184586003, "learning_rate": 0.00011666666666666668, "loss": 0.9386, "step": 91 }, { "epoch": 0.017699115044247787, "grad_norm": 0.23984196360645318, "learning_rate": 0.00011794871794871796, "loss": 0.9635, "step": 92 }, { "epoch": 0.01789149672951135, "grad_norm": 0.24250889290000271, "learning_rate": 0.00011923076923076923, "loss": 0.9258, "step": 93 }, { "epoch": 0.018083878414774913, "grad_norm": 0.2668742654805047, "learning_rate": 0.00012051282051282052, "loss": 0.8882, "step": 94 }, { "epoch": 0.018276260100038477, "grad_norm": 0.24199991778332866, "learning_rate": 0.00012179487179487179, "loss": 0.9363, "step": 95 }, { "epoch": 0.01846864178530204, "grad_norm": 0.24127007790878796, "learning_rate": 0.0001230769230769231, "loss": 0.9467, "step": 96 }, { "epoch": 0.018661023470565603, "grad_norm": 0.20976799289683673, "learning_rate": 0.00012435897435897437, "loss": 0.9064, "step": 97 }, { "epoch": 0.018853405155829166, "grad_norm": 0.2297634704152219, "learning_rate": 0.00012564102564102564, "loss": 0.9172, "step": 98 }, { "epoch": 0.01904578684109273, "grad_norm": 0.21260916524253945, "learning_rate": 0.00012692307692307693, "loss": 0.9103, "step": 99 }, { "epoch": 0.019238168526356292, "grad_norm": 0.21760700541738703, "learning_rate": 0.00012820512820512823, "loss": 0.9525, "step": 100 }, { "epoch": 0.019430550211619855, "grad_norm": 0.24145953209957519, "learning_rate": 0.0001294871794871795, "loss": 1.0071, "step": 101 }, { "epoch": 0.019622931896883418, "grad_norm": 0.2297438610079206, "learning_rate": 0.00013076923076923077, "loss": 0.9321, "step": 102 }, { "epoch": 0.01981531358214698, "grad_norm": 0.22497133280561546, "learning_rate": 0.00013205128205128204, "loss": 0.9346, "step": 103 }, { "epoch": 0.020007695267410544, "grad_norm": 0.20108979467015967, "learning_rate": 0.00013333333333333334, "loss": 0.8741, "step": 104 }, { "epoch": 0.020200076952674107, "grad_norm": 0.21658697660384524, "learning_rate": 0.00013461538461538464, "loss": 0.9606, "step": 105 }, { "epoch": 0.020392458637937667, "grad_norm": 0.21595066163304966, "learning_rate": 0.0001358974358974359, "loss": 0.9249, "step": 106 }, { "epoch": 0.02058484032320123, "grad_norm": 0.21005067463476118, "learning_rate": 0.00013717948717948718, "loss": 0.9592, "step": 107 }, { "epoch": 0.020777222008464793, "grad_norm": 0.21253071766631587, "learning_rate": 0.00013846153846153847, "loss": 0.8751, "step": 108 }, { "epoch": 0.020969603693728356, "grad_norm": 0.2202287777922542, "learning_rate": 0.00013974358974358974, "loss": 0.9498, "step": 109 }, { "epoch": 0.02116198537899192, "grad_norm": 0.21819756753665479, "learning_rate": 0.00014102564102564104, "loss": 0.9206, "step": 110 }, { "epoch": 0.021354367064255482, "grad_norm": 0.24624607391479672, "learning_rate": 0.0001423076923076923, "loss": 0.9696, "step": 111 }, { "epoch": 0.021546748749519045, "grad_norm": 0.18865179690372727, "learning_rate": 0.0001435897435897436, "loss": 0.9147, "step": 112 }, { "epoch": 0.021739130434782608, "grad_norm": 0.2495103874355497, "learning_rate": 0.00014487179487179488, "loss": 0.9065, "step": 113 }, { "epoch": 0.02193151212004617, "grad_norm": 0.23084298244418824, "learning_rate": 0.00014615384615384615, "loss": 0.9473, "step": 114 }, { "epoch": 0.022123893805309734, "grad_norm": 0.2101915248982665, "learning_rate": 0.00014743589743589745, "loss": 0.8994, "step": 115 }, { "epoch": 0.022316275490573297, "grad_norm": 0.22023203510790845, "learning_rate": 0.00014871794871794872, "loss": 0.9219, "step": 116 }, { "epoch": 0.02250865717583686, "grad_norm": 0.22504504086954197, "learning_rate": 0.00015000000000000001, "loss": 0.8837, "step": 117 }, { "epoch": 0.022701038861100423, "grad_norm": 0.22534021680414154, "learning_rate": 0.00015128205128205128, "loss": 0.8931, "step": 118 }, { "epoch": 0.022893420546363986, "grad_norm": 0.22659300772380703, "learning_rate": 0.00015256410256410255, "loss": 0.947, "step": 119 }, { "epoch": 0.02308580223162755, "grad_norm": 0.20730549526360512, "learning_rate": 0.00015384615384615385, "loss": 0.9165, "step": 120 }, { "epoch": 0.023278183916891113, "grad_norm": 0.2133396725726003, "learning_rate": 0.00015512820512820515, "loss": 0.9451, "step": 121 }, { "epoch": 0.023470565602154676, "grad_norm": 0.21818099461200743, "learning_rate": 0.00015641025641025642, "loss": 0.9034, "step": 122 }, { "epoch": 0.02366294728741824, "grad_norm": 0.23989972990658884, "learning_rate": 0.0001576923076923077, "loss": 0.9229, "step": 123 }, { "epoch": 0.0238553289726818, "grad_norm": 0.21392375106138103, "learning_rate": 0.00015897435897435896, "loss": 0.9511, "step": 124 }, { "epoch": 0.024047710657945365, "grad_norm": 0.301934549241715, "learning_rate": 0.00016025641025641028, "loss": 0.9121, "step": 125 }, { "epoch": 0.024240092343208928, "grad_norm": 0.22878361011565618, "learning_rate": 0.00016153846153846155, "loss": 0.9629, "step": 126 }, { "epoch": 0.02443247402847249, "grad_norm": 0.20707903018440324, "learning_rate": 0.00016282051282051282, "loss": 0.9064, "step": 127 }, { "epoch": 0.024624855713736054, "grad_norm": 0.20987818842592798, "learning_rate": 0.0001641025641025641, "loss": 0.8837, "step": 128 }, { "epoch": 0.024817237398999614, "grad_norm": 0.2175404068886466, "learning_rate": 0.0001653846153846154, "loss": 0.9049, "step": 129 }, { "epoch": 0.025009619084263177, "grad_norm": 0.23742760680074443, "learning_rate": 0.0001666666666666667, "loss": 0.932, "step": 130 }, { "epoch": 0.02520200076952674, "grad_norm": 0.21426958015509057, "learning_rate": 0.00016794871794871796, "loss": 0.8686, "step": 131 }, { "epoch": 0.025394382454790303, "grad_norm": 0.2043003972692777, "learning_rate": 0.00016923076923076923, "loss": 0.9504, "step": 132 }, { "epoch": 0.025586764140053866, "grad_norm": 0.20896068310729182, "learning_rate": 0.00017051282051282053, "loss": 0.9308, "step": 133 }, { "epoch": 0.02577914582531743, "grad_norm": 0.2143922640174859, "learning_rate": 0.0001717948717948718, "loss": 0.9386, "step": 134 }, { "epoch": 0.025971527510580992, "grad_norm": 0.2150257917262387, "learning_rate": 0.0001730769230769231, "loss": 0.9521, "step": 135 }, { "epoch": 0.026163909195844555, "grad_norm": 0.19222083515426933, "learning_rate": 0.00017435897435897436, "loss": 0.69, "step": 136 }, { "epoch": 0.026356290881108118, "grad_norm": 0.16285866459468973, "learning_rate": 0.00017564102564102566, "loss": 0.6694, "step": 137 }, { "epoch": 0.02654867256637168, "grad_norm": 0.18056027121662155, "learning_rate": 0.00017692307692307693, "loss": 0.9277, "step": 138 }, { "epoch": 0.026741054251635244, "grad_norm": 0.21481007412934222, "learning_rate": 0.00017820512820512823, "loss": 0.9458, "step": 139 }, { "epoch": 0.026933435936898807, "grad_norm": 0.21575264968234362, "learning_rate": 0.0001794871794871795, "loss": 0.9406, "step": 140 }, { "epoch": 0.02712581762216237, "grad_norm": 0.2029631801486215, "learning_rate": 0.00018076923076923077, "loss": 0.8802, "step": 141 }, { "epoch": 0.027318199307425933, "grad_norm": 0.23690874785006671, "learning_rate": 0.00018205128205128207, "loss": 0.9304, "step": 142 }, { "epoch": 0.027510580992689496, "grad_norm": 0.2811900445973516, "learning_rate": 0.00018333333333333334, "loss": 0.8782, "step": 143 }, { "epoch": 0.02770296267795306, "grad_norm": 0.2134149412410721, "learning_rate": 0.00018461538461538463, "loss": 0.8785, "step": 144 }, { "epoch": 0.027895344363216622, "grad_norm": 0.23212806441089945, "learning_rate": 0.0001858974358974359, "loss": 0.8333, "step": 145 }, { "epoch": 0.028087726048480186, "grad_norm": 0.20792044083186376, "learning_rate": 0.0001871794871794872, "loss": 0.8631, "step": 146 }, { "epoch": 0.02828010773374375, "grad_norm": 0.1902399112515953, "learning_rate": 0.00018846153846153847, "loss": 0.9328, "step": 147 }, { "epoch": 0.02847248941900731, "grad_norm": 0.22486004470666543, "learning_rate": 0.00018974358974358974, "loss": 0.8557, "step": 148 }, { "epoch": 0.028664871104270875, "grad_norm": 0.20524133157808722, "learning_rate": 0.00019102564102564104, "loss": 0.9708, "step": 149 }, { "epoch": 0.028857252789534438, "grad_norm": 0.24433060821544642, "learning_rate": 0.00019230769230769233, "loss": 0.9443, "step": 150 }, { "epoch": 0.029049634474798, "grad_norm": 0.22813234191136403, "learning_rate": 0.0001935897435897436, "loss": 0.8785, "step": 151 }, { "epoch": 0.029242016160061564, "grad_norm": 0.20018294423170738, "learning_rate": 0.00019487179487179487, "loss": 0.8756, "step": 152 }, { "epoch": 0.029434397845325123, "grad_norm": 0.20604245049646633, "learning_rate": 0.00019615384615384615, "loss": 0.9248, "step": 153 }, { "epoch": 0.029626779530588686, "grad_norm": 0.24734568263826395, "learning_rate": 0.00019743589743589744, "loss": 0.9102, "step": 154 }, { "epoch": 0.02981916121585225, "grad_norm": 0.20389311199439714, "learning_rate": 0.00019871794871794874, "loss": 0.9057, "step": 155 }, { "epoch": 0.030011542901115813, "grad_norm": 0.21399041172386252, "learning_rate": 0.0002, "loss": 0.9201, "step": 156 }, { "epoch": 0.030203924586379376, "grad_norm": 0.20506519174051277, "learning_rate": 0.00019999998058827846, "loss": 0.8985, "step": 157 }, { "epoch": 0.03039630627164294, "grad_norm": 0.20463991103729828, "learning_rate": 0.00019999992235312134, "loss": 0.8696, "step": 158 }, { "epoch": 0.030588687956906502, "grad_norm": 0.21209668523123304, "learning_rate": 0.00019999982529455128, "loss": 0.864, "step": 159 }, { "epoch": 0.030781069642170065, "grad_norm": 0.21261113550305322, "learning_rate": 0.00019999968941260594, "loss": 0.9427, "step": 160 }, { "epoch": 0.030973451327433628, "grad_norm": 0.18036132150668382, "learning_rate": 0.00019999951470733808, "loss": 0.85, "step": 161 }, { "epoch": 0.03116583301269719, "grad_norm": 0.23667641488093075, "learning_rate": 0.0001999993011788155, "loss": 0.8359, "step": 162 }, { "epoch": 0.031358214697960754, "grad_norm": 0.22645581309229176, "learning_rate": 0.00019999904882712116, "loss": 0.9017, "step": 163 }, { "epoch": 0.031550596383224314, "grad_norm": 0.22458925766125126, "learning_rate": 0.000199998757652353, "loss": 0.9233, "step": 164 }, { "epoch": 0.03174297806848788, "grad_norm": 0.2042215805706502, "learning_rate": 0.000199998427654624, "loss": 0.9203, "step": 165 }, { "epoch": 0.03193535975375144, "grad_norm": 0.1998765238145995, "learning_rate": 0.0001999980588340624, "loss": 0.8741, "step": 166 }, { "epoch": 0.032127741439015006, "grad_norm": 0.20722606034842078, "learning_rate": 0.0001999976511908113, "loss": 0.9176, "step": 167 }, { "epoch": 0.032320123124278566, "grad_norm": 0.18657233851627578, "learning_rate": 0.00019999720472502902, "loss": 0.9125, "step": 168 }, { "epoch": 0.03251250480954213, "grad_norm": 0.1784981034252015, "learning_rate": 0.00019999671943688883, "loss": 0.6756, "step": 169 }, { "epoch": 0.03270488649480569, "grad_norm": 0.21199980062818183, "learning_rate": 0.00019999619532657916, "loss": 0.91, "step": 170 }, { "epoch": 0.03289726818006926, "grad_norm": 0.2195532730688087, "learning_rate": 0.00019999563239430352, "loss": 0.9013, "step": 171 }, { "epoch": 0.03308964986533282, "grad_norm": 0.20497199040971095, "learning_rate": 0.00019999503064028042, "loss": 0.9136, "step": 172 }, { "epoch": 0.033282031550596385, "grad_norm": 0.2058889510251947, "learning_rate": 0.0001999943900647435, "loss": 0.8843, "step": 173 }, { "epoch": 0.033474413235859944, "grad_norm": 0.2216514087862708, "learning_rate": 0.00019999371066794147, "loss": 0.9611, "step": 174 }, { "epoch": 0.03366679492112351, "grad_norm": 0.20588585232591008, "learning_rate": 0.00019999299245013804, "loss": 0.9062, "step": 175 }, { "epoch": 0.03385917660638707, "grad_norm": 0.2075386181527988, "learning_rate": 0.0001999922354116121, "loss": 0.9147, "step": 176 }, { "epoch": 0.03405155829165064, "grad_norm": 0.211007948026382, "learning_rate": 0.0001999914395526575, "loss": 0.8907, "step": 177 }, { "epoch": 0.034243939976914196, "grad_norm": 0.20285040346081426, "learning_rate": 0.0001999906048735833, "loss": 0.9399, "step": 178 }, { "epoch": 0.03443632166217776, "grad_norm": 0.1768340024678368, "learning_rate": 0.00019998973137471353, "loss": 0.8879, "step": 179 }, { "epoch": 0.03462870334744132, "grad_norm": 0.19623723484670935, "learning_rate": 0.00019998881905638726, "loss": 0.9185, "step": 180 }, { "epoch": 0.03482108503270489, "grad_norm": 0.19030681478138944, "learning_rate": 0.0001999878679189587, "loss": 0.8962, "step": 181 }, { "epoch": 0.03501346671796845, "grad_norm": 0.20298901632567304, "learning_rate": 0.00019998687796279717, "loss": 0.8957, "step": 182 }, { "epoch": 0.035205848403232015, "grad_norm": 0.19812961668976592, "learning_rate": 0.00019998584918828695, "loss": 0.9165, "step": 183 }, { "epoch": 0.035398230088495575, "grad_norm": 0.18491949372380864, "learning_rate": 0.00019998478159582748, "loss": 0.9655, "step": 184 }, { "epoch": 0.03559061177375914, "grad_norm": 0.1942075922283302, "learning_rate": 0.00019998367518583317, "loss": 0.8863, "step": 185 }, { "epoch": 0.0357829934590227, "grad_norm": 0.23881479837643246, "learning_rate": 0.00019998252995873366, "loss": 0.9102, "step": 186 }, { "epoch": 0.03597537514428627, "grad_norm": 0.22442789638870772, "learning_rate": 0.0001999813459149735, "loss": 0.9439, "step": 187 }, { "epoch": 0.03616775682954983, "grad_norm": 0.20620688786442298, "learning_rate": 0.0001999801230550124, "loss": 0.8864, "step": 188 }, { "epoch": 0.03636013851481339, "grad_norm": 0.22677089511354462, "learning_rate": 0.00019997886137932512, "loss": 0.9343, "step": 189 }, { "epoch": 0.03655252020007695, "grad_norm": 0.22963410257006117, "learning_rate": 0.00019997756088840148, "loss": 0.9013, "step": 190 }, { "epoch": 0.03674490188534051, "grad_norm": 0.19541153408073714, "learning_rate": 0.00019997622158274636, "loss": 0.8819, "step": 191 }, { "epoch": 0.03693728357060408, "grad_norm": 0.19167331393692535, "learning_rate": 0.00019997484346287973, "loss": 0.9073, "step": 192 }, { "epoch": 0.03712966525586764, "grad_norm": 0.2068073021730728, "learning_rate": 0.00019997342652933666, "loss": 0.8898, "step": 193 }, { "epoch": 0.037322046941131205, "grad_norm": 0.22231919047315185, "learning_rate": 0.00019997197078266724, "loss": 0.8891, "step": 194 }, { "epoch": 0.037514428626394765, "grad_norm": 0.18917929955521462, "learning_rate": 0.00019997047622343658, "loss": 0.9189, "step": 195 }, { "epoch": 0.03770681031165833, "grad_norm": 0.21996694881026388, "learning_rate": 0.00019996894285222497, "loss": 0.8946, "step": 196 }, { "epoch": 0.03789919199692189, "grad_norm": 0.21662136334239734, "learning_rate": 0.00019996737066962775, "loss": 0.8479, "step": 197 }, { "epoch": 0.03809157368218546, "grad_norm": 0.19587116641722593, "learning_rate": 0.00019996575967625523, "loss": 0.9094, "step": 198 }, { "epoch": 0.03828395536744902, "grad_norm": 0.19130369566595284, "learning_rate": 0.0001999641098727329, "loss": 0.9242, "step": 199 }, { "epoch": 0.038476337052712584, "grad_norm": 0.220494436178227, "learning_rate": 0.00019996242125970126, "loss": 0.8515, "step": 200 }, { "epoch": 0.03866871873797614, "grad_norm": 0.1913514786362502, "learning_rate": 0.00019996069383781587, "loss": 0.8905, "step": 201 }, { "epoch": 0.03886110042323971, "grad_norm": 0.20045927443341371, "learning_rate": 0.00019995892760774737, "loss": 0.9115, "step": 202 }, { "epoch": 0.03905348210850327, "grad_norm": 0.22325817142138868, "learning_rate": 0.0001999571225701815, "loss": 0.9524, "step": 203 }, { "epoch": 0.039245863793766836, "grad_norm": 0.21274162090011756, "learning_rate": 0.00019995527872581902, "loss": 0.8874, "step": 204 }, { "epoch": 0.039438245479030395, "grad_norm": 0.22482262581290757, "learning_rate": 0.00019995339607537579, "loss": 0.9164, "step": 205 }, { "epoch": 0.03963062716429396, "grad_norm": 0.1941863016894808, "learning_rate": 0.00019995147461958267, "loss": 0.8999, "step": 206 }, { "epoch": 0.03982300884955752, "grad_norm": 0.17301384290582592, "learning_rate": 0.0001999495143591857, "loss": 0.6713, "step": 207 }, { "epoch": 0.04001539053482109, "grad_norm": 0.21120506857816843, "learning_rate": 0.0001999475152949459, "loss": 0.868, "step": 208 }, { "epoch": 0.04020777222008465, "grad_norm": 0.19725713329103817, "learning_rate": 0.00019994547742763934, "loss": 0.884, "step": 209 }, { "epoch": 0.040400153905348214, "grad_norm": 0.1865500969601041, "learning_rate": 0.00019994340075805725, "loss": 0.9284, "step": 210 }, { "epoch": 0.040592535590611774, "grad_norm": 0.21558684062173636, "learning_rate": 0.00019994128528700583, "loss": 0.96, "step": 211 }, { "epoch": 0.04078491727587533, "grad_norm": 0.18999270987418607, "learning_rate": 0.00019993913101530635, "loss": 0.8784, "step": 212 }, { "epoch": 0.0409772989611389, "grad_norm": 0.19294379138583292, "learning_rate": 0.00019993693794379525, "loss": 0.836, "step": 213 }, { "epoch": 0.04116968064640246, "grad_norm": 0.19033120694306557, "learning_rate": 0.0001999347060733239, "loss": 0.9174, "step": 214 }, { "epoch": 0.041362062331666026, "grad_norm": 0.20290330884835467, "learning_rate": 0.00019993243540475877, "loss": 0.9337, "step": 215 }, { "epoch": 0.041554444016929586, "grad_norm": 0.1916346103691421, "learning_rate": 0.00019993012593898146, "loss": 0.8899, "step": 216 }, { "epoch": 0.04174682570219315, "grad_norm": 0.19822799460682916, "learning_rate": 0.00019992777767688854, "loss": 0.8688, "step": 217 }, { "epoch": 0.04193920738745671, "grad_norm": 0.19433173302518736, "learning_rate": 0.00019992539061939175, "loss": 0.9312, "step": 218 }, { "epoch": 0.04213158907272028, "grad_norm": 0.2236217542558508, "learning_rate": 0.00019992296476741776, "loss": 0.8878, "step": 219 }, { "epoch": 0.04232397075798384, "grad_norm": 0.19052129552099334, "learning_rate": 0.00019992050012190843, "loss": 0.8731, "step": 220 }, { "epoch": 0.042516352443247404, "grad_norm": 0.1096591283882888, "learning_rate": 0.00019991799668382058, "loss": 0.6672, "step": 221 }, { "epoch": 0.042708734128510964, "grad_norm": 0.1891844841465431, "learning_rate": 0.00019991545445412613, "loss": 0.8878, "step": 222 }, { "epoch": 0.04290111581377453, "grad_norm": 0.20113662998340062, "learning_rate": 0.0001999128734338121, "loss": 0.9868, "step": 223 }, { "epoch": 0.04309349749903809, "grad_norm": 0.21184293996629627, "learning_rate": 0.00019991025362388044, "loss": 0.8822, "step": 224 }, { "epoch": 0.04328587918430166, "grad_norm": 0.17982596739822404, "learning_rate": 0.00019990759502534834, "loss": 0.8923, "step": 225 }, { "epoch": 0.043478260869565216, "grad_norm": 0.1807577662556677, "learning_rate": 0.00019990489763924797, "loss": 0.8788, "step": 226 }, { "epoch": 0.04367064255482878, "grad_norm": 0.185153394710195, "learning_rate": 0.00019990216146662646, "loss": 0.8943, "step": 227 }, { "epoch": 0.04386302424009234, "grad_norm": 0.18033034386021188, "learning_rate": 0.00019989938650854616, "loss": 0.9271, "step": 228 }, { "epoch": 0.04405540592535591, "grad_norm": 0.19025520683429847, "learning_rate": 0.00019989657276608437, "loss": 0.8716, "step": 229 }, { "epoch": 0.04424778761061947, "grad_norm": 0.2159317979361635, "learning_rate": 0.00019989372024033352, "loss": 0.8994, "step": 230 }, { "epoch": 0.044440169295883035, "grad_norm": 0.2062108746302333, "learning_rate": 0.000199890828932401, "loss": 0.9125, "step": 231 }, { "epoch": 0.044632550981146595, "grad_norm": 0.1982429055058165, "learning_rate": 0.00019988789884340936, "loss": 0.831, "step": 232 }, { "epoch": 0.04482493266641016, "grad_norm": 0.18587713449064217, "learning_rate": 0.00019988492997449615, "loss": 0.8535, "step": 233 }, { "epoch": 0.04501731435167372, "grad_norm": 0.2105743188409605, "learning_rate": 0.00019988192232681397, "loss": 0.8558, "step": 234 }, { "epoch": 0.04520969603693728, "grad_norm": 0.19599638393846205, "learning_rate": 0.00019987887590153055, "loss": 0.9119, "step": 235 }, { "epoch": 0.04540207772220085, "grad_norm": 0.2207194679193591, "learning_rate": 0.00019987579069982855, "loss": 0.9013, "step": 236 }, { "epoch": 0.045594459407464406, "grad_norm": 0.18281705413655833, "learning_rate": 0.00019987266672290575, "loss": 0.8792, "step": 237 }, { "epoch": 0.04578684109272797, "grad_norm": 0.19235819241728627, "learning_rate": 0.00019986950397197503, "loss": 0.92, "step": 238 }, { "epoch": 0.04597922277799153, "grad_norm": 0.20708476125844436, "learning_rate": 0.00019986630244826426, "loss": 0.8723, "step": 239 }, { "epoch": 0.0461716044632551, "grad_norm": 0.19084221789616373, "learning_rate": 0.00019986306215301637, "loss": 0.8613, "step": 240 }, { "epoch": 0.04636398614851866, "grad_norm": 0.1962039956582299, "learning_rate": 0.00019985978308748937, "loss": 0.857, "step": 241 }, { "epoch": 0.046556367833782225, "grad_norm": 0.27799694098264915, "learning_rate": 0.00019985646525295632, "loss": 0.9389, "step": 242 }, { "epoch": 0.046748749519045785, "grad_norm": 0.19762593123417097, "learning_rate": 0.0001998531086507053, "loss": 0.8913, "step": 243 }, { "epoch": 0.04694113120430935, "grad_norm": 0.19355698390676038, "learning_rate": 0.00019984971328203946, "loss": 0.914, "step": 244 }, { "epoch": 0.04713351288957291, "grad_norm": 0.19048657908463293, "learning_rate": 0.00019984627914827697, "loss": 0.921, "step": 245 }, { "epoch": 0.04732589457483648, "grad_norm": 0.19523840725638852, "learning_rate": 0.00019984280625075113, "loss": 0.8854, "step": 246 }, { "epoch": 0.04751827626010004, "grad_norm": 0.18999802713615838, "learning_rate": 0.0001998392945908102, "loss": 0.8449, "step": 247 }, { "epoch": 0.0477106579453636, "grad_norm": 0.09561684433777591, "learning_rate": 0.00019983574416981755, "loss": 0.674, "step": 248 }, { "epoch": 0.04790303963062716, "grad_norm": 0.19372935033006206, "learning_rate": 0.00019983215498915158, "loss": 0.9065, "step": 249 }, { "epoch": 0.04809542131589073, "grad_norm": 0.16612228636391105, "learning_rate": 0.0001998285270502057, "loss": 0.9015, "step": 250 }, { "epoch": 0.04828780300115429, "grad_norm": 0.19481277018020898, "learning_rate": 0.00019982486035438846, "loss": 0.9059, "step": 251 }, { "epoch": 0.048480184686417856, "grad_norm": 0.17407730757768297, "learning_rate": 0.00019982115490312332, "loss": 0.8614, "step": 252 }, { "epoch": 0.048672566371681415, "grad_norm": 0.09736222607908981, "learning_rate": 0.0001998174106978489, "loss": 0.6903, "step": 253 }, { "epoch": 0.04886494805694498, "grad_norm": 0.19729226228890967, "learning_rate": 0.00019981362774001887, "loss": 0.9306, "step": 254 }, { "epoch": 0.04905732974220854, "grad_norm": 0.17289912812614408, "learning_rate": 0.00019980980603110185, "loss": 0.8996, "step": 255 }, { "epoch": 0.04924971142747211, "grad_norm": 0.11327522498032354, "learning_rate": 0.00019980594557258155, "loss": 0.6447, "step": 256 }, { "epoch": 0.04944209311273567, "grad_norm": 0.09024381396796899, "learning_rate": 0.00019980204636595682, "loss": 0.656, "step": 257 }, { "epoch": 0.04963447479799923, "grad_norm": 0.08209158767776666, "learning_rate": 0.00019979810841274132, "loss": 0.5737, "step": 258 }, { "epoch": 0.049826856483262794, "grad_norm": 0.18959969111237968, "learning_rate": 0.00019979413171446403, "loss": 0.846, "step": 259 }, { "epoch": 0.05001923816852635, "grad_norm": 0.19119069988343895, "learning_rate": 0.00019979011627266883, "loss": 0.9022, "step": 260 }, { "epoch": 0.05021161985378992, "grad_norm": 0.18029662003814056, "learning_rate": 0.00019978606208891458, "loss": 0.9011, "step": 261 }, { "epoch": 0.05040400153905348, "grad_norm": 0.2016495809077164, "learning_rate": 0.0001997819691647753, "loss": 0.8809, "step": 262 }, { "epoch": 0.050596383224317046, "grad_norm": 0.19980133745402204, "learning_rate": 0.00019977783750183998, "loss": 0.8732, "step": 263 }, { "epoch": 0.050788764909580605, "grad_norm": 0.16157042961469673, "learning_rate": 0.00019977366710171272, "loss": 0.8276, "step": 264 }, { "epoch": 0.05098114659484417, "grad_norm": 0.16921343111638015, "learning_rate": 0.00019976945796601258, "loss": 0.9049, "step": 265 }, { "epoch": 0.05117352828010773, "grad_norm": 0.18286364439836783, "learning_rate": 0.00019976521009637364, "loss": 0.8396, "step": 266 }, { "epoch": 0.0513659099653713, "grad_norm": 0.18446403616066462, "learning_rate": 0.0001997609234944452, "loss": 0.9125, "step": 267 }, { "epoch": 0.05155829165063486, "grad_norm": 0.2088078355307454, "learning_rate": 0.00019975659816189137, "loss": 0.8273, "step": 268 }, { "epoch": 0.051750673335898424, "grad_norm": 0.23031017954392535, "learning_rate": 0.0001997522341003914, "loss": 0.885, "step": 269 }, { "epoch": 0.051943055021161984, "grad_norm": 0.179462355274658, "learning_rate": 0.00019974783131163957, "loss": 0.9026, "step": 270 }, { "epoch": 0.05213543670642555, "grad_norm": 0.17098772896902728, "learning_rate": 0.00019974338979734523, "loss": 0.8336, "step": 271 }, { "epoch": 0.05232781839168911, "grad_norm": 0.21663183799081726, "learning_rate": 0.00019973890955923268, "loss": 0.8681, "step": 272 }, { "epoch": 0.052520200076952676, "grad_norm": 0.176759481573498, "learning_rate": 0.00019973439059904133, "loss": 0.9184, "step": 273 }, { "epoch": 0.052712581762216236, "grad_norm": 0.17966828974839788, "learning_rate": 0.00019972983291852563, "loss": 0.9053, "step": 274 }, { "epoch": 0.0529049634474798, "grad_norm": 0.20374021306243364, "learning_rate": 0.00019972523651945497, "loss": 0.8691, "step": 275 }, { "epoch": 0.05309734513274336, "grad_norm": 0.20291773438247993, "learning_rate": 0.00019972060140361383, "loss": 0.9024, "step": 276 }, { "epoch": 0.05328972681800693, "grad_norm": 0.17339492307957002, "learning_rate": 0.00019971592757280177, "loss": 0.8671, "step": 277 }, { "epoch": 0.05348210850327049, "grad_norm": 0.31033825489322214, "learning_rate": 0.0001997112150288333, "loss": 0.8693, "step": 278 }, { "epoch": 0.053674490188534055, "grad_norm": 0.19953496610943305, "learning_rate": 0.000199706463773538, "loss": 0.8761, "step": 279 }, { "epoch": 0.053866871873797614, "grad_norm": 0.23968897407953313, "learning_rate": 0.0001997016738087605, "loss": 0.9045, "step": 280 }, { "epoch": 0.054059253559061174, "grad_norm": 0.17276087374471324, "learning_rate": 0.00019969684513636033, "loss": 0.8283, "step": 281 }, { "epoch": 0.05425163524432474, "grad_norm": 0.19679297344300253, "learning_rate": 0.00019969197775821225, "loss": 0.8905, "step": 282 }, { "epoch": 0.0544440169295883, "grad_norm": 0.18599944370763985, "learning_rate": 0.00019968707167620592, "loss": 0.8717, "step": 283 }, { "epoch": 0.05463639861485187, "grad_norm": 0.1906065146179007, "learning_rate": 0.00019968212689224602, "loss": 0.8507, "step": 284 }, { "epoch": 0.054828780300115426, "grad_norm": 0.18905427736371028, "learning_rate": 0.00019967714340825233, "loss": 0.8469, "step": 285 }, { "epoch": 0.05502116198537899, "grad_norm": 0.17539564883789113, "learning_rate": 0.00019967212122615956, "loss": 0.8675, "step": 286 }, { "epoch": 0.05521354367064255, "grad_norm": 0.08406652969026258, "learning_rate": 0.00019966706034791752, "loss": 0.5611, "step": 287 }, { "epoch": 0.05540592535590612, "grad_norm": 0.19118244801710516, "learning_rate": 0.00019966196077549105, "loss": 0.9105, "step": 288 }, { "epoch": 0.05559830704116968, "grad_norm": 0.10173187847696259, "learning_rate": 0.0001996568225108599, "loss": 0.6307, "step": 289 }, { "epoch": 0.055790688726433245, "grad_norm": 0.19041655258976897, "learning_rate": 0.00019965164555601898, "loss": 0.8781, "step": 290 }, { "epoch": 0.055983070411696804, "grad_norm": 0.19518005809390263, "learning_rate": 0.00019964642991297815, "loss": 0.9302, "step": 291 }, { "epoch": 0.05617545209696037, "grad_norm": 0.18515815739742666, "learning_rate": 0.0001996411755837623, "loss": 0.8792, "step": 292 }, { "epoch": 0.05636783378222393, "grad_norm": 0.18093184969759443, "learning_rate": 0.00019963588257041135, "loss": 0.9441, "step": 293 }, { "epoch": 0.0565602154674875, "grad_norm": 0.1937757245627149, "learning_rate": 0.00019963055087498022, "loss": 0.9072, "step": 294 }, { "epoch": 0.05675259715275106, "grad_norm": 0.17840291866385963, "learning_rate": 0.00019962518049953886, "loss": 0.8749, "step": 295 }, { "epoch": 0.05694497883801462, "grad_norm": 0.19803410603366375, "learning_rate": 0.00019961977144617224, "loss": 0.851, "step": 296 }, { "epoch": 0.05713736052327818, "grad_norm": 0.1919160892443671, "learning_rate": 0.0001996143237169803, "loss": 0.8656, "step": 297 }, { "epoch": 0.05732974220854175, "grad_norm": 0.17178196432146303, "learning_rate": 0.00019960883731407812, "loss": 0.8448, "step": 298 }, { "epoch": 0.05752212389380531, "grad_norm": 0.1813771787504376, "learning_rate": 0.00019960331223959564, "loss": 0.9082, "step": 299 }, { "epoch": 0.057714505579068875, "grad_norm": 0.17540667116169248, "learning_rate": 0.0001995977484956779, "loss": 0.8432, "step": 300 }, { "epoch": 0.057906887264332435, "grad_norm": 0.18881932206706717, "learning_rate": 0.00019959214608448496, "loss": 0.8458, "step": 301 }, { "epoch": 0.058099268949596, "grad_norm": 0.19612674373671216, "learning_rate": 0.00019958650500819184, "loss": 0.8761, "step": 302 }, { "epoch": 0.05829165063485956, "grad_norm": 0.2104230310666131, "learning_rate": 0.0001995808252689886, "loss": 0.9158, "step": 303 }, { "epoch": 0.05848403232012313, "grad_norm": 0.19082068482955877, "learning_rate": 0.00019957510686908033, "loss": 0.868, "step": 304 }, { "epoch": 0.05867641400538669, "grad_norm": 0.19235941777486248, "learning_rate": 0.00019956934981068712, "loss": 0.8875, "step": 305 }, { "epoch": 0.05886879569065025, "grad_norm": 0.1911659155010832, "learning_rate": 0.00019956355409604402, "loss": 0.8836, "step": 306 }, { "epoch": 0.05906117737591381, "grad_norm": 0.18749994229076553, "learning_rate": 0.00019955771972740116, "loss": 0.885, "step": 307 }, { "epoch": 0.05925355906117737, "grad_norm": 0.18112891587339938, "learning_rate": 0.0001995518467070236, "loss": 0.876, "step": 308 }, { "epoch": 0.05944594074644094, "grad_norm": 0.19228625743467925, "learning_rate": 0.00019954593503719153, "loss": 0.9025, "step": 309 }, { "epoch": 0.0596383224317045, "grad_norm": 0.17481295754860954, "learning_rate": 0.00019953998472019995, "loss": 0.9437, "step": 310 }, { "epoch": 0.059830704116968066, "grad_norm": 0.1952383630496437, "learning_rate": 0.0001995339957583591, "loss": 0.9301, "step": 311 }, { "epoch": 0.060023085802231625, "grad_norm": 0.09621529772643021, "learning_rate": 0.00019952796815399403, "loss": 0.6246, "step": 312 }, { "epoch": 0.06021546748749519, "grad_norm": 0.19192907122161232, "learning_rate": 0.00019952190190944484, "loss": 0.8513, "step": 313 }, { "epoch": 0.06040784917275875, "grad_norm": 0.20957066052404938, "learning_rate": 0.00019951579702706668, "loss": 0.8716, "step": 314 }, { "epoch": 0.06060023085802232, "grad_norm": 0.1823777585176578, "learning_rate": 0.00019950965350922975, "loss": 0.8939, "step": 315 }, { "epoch": 0.06079261254328588, "grad_norm": 0.18814178108429216, "learning_rate": 0.00019950347135831907, "loss": 0.8717, "step": 316 }, { "epoch": 0.060984994228549444, "grad_norm": 0.20326681583981235, "learning_rate": 0.0001994972505767348, "loss": 0.7949, "step": 317 }, { "epoch": 0.061177375913813004, "grad_norm": 0.17766571106097073, "learning_rate": 0.00019949099116689208, "loss": 0.87, "step": 318 }, { "epoch": 0.06136975759907657, "grad_norm": 0.1898471270464398, "learning_rate": 0.000199484693131221, "loss": 0.9234, "step": 319 }, { "epoch": 0.06156213928434013, "grad_norm": 0.21004792267764574, "learning_rate": 0.00019947835647216668, "loss": 0.9137, "step": 320 }, { "epoch": 0.061754520969603696, "grad_norm": 0.19704879652686477, "learning_rate": 0.00019947198119218924, "loss": 0.8877, "step": 321 }, { "epoch": 0.061946902654867256, "grad_norm": 0.19614423328725517, "learning_rate": 0.0001994655672937638, "loss": 0.8552, "step": 322 }, { "epoch": 0.06213928434013082, "grad_norm": 0.25389822973994264, "learning_rate": 0.00019945911477938042, "loss": 0.8874, "step": 323 }, { "epoch": 0.06233166602539438, "grad_norm": 0.28517429429271357, "learning_rate": 0.0001994526236515442, "loss": 0.6778, "step": 324 }, { "epoch": 0.06252404771065795, "grad_norm": 0.09619432482625646, "learning_rate": 0.00019944609391277528, "loss": 0.653, "step": 325 }, { "epoch": 0.06271642939592151, "grad_norm": 0.17843374659801664, "learning_rate": 0.0001994395255656086, "loss": 0.8524, "step": 326 }, { "epoch": 0.06290881108118507, "grad_norm": 0.17968122098769812, "learning_rate": 0.00019943291861259431, "loss": 0.8009, "step": 327 }, { "epoch": 0.06310119276644863, "grad_norm": 0.18488991128923568, "learning_rate": 0.00019942627305629746, "loss": 0.9147, "step": 328 }, { "epoch": 0.0632935744517122, "grad_norm": 0.19387662966508662, "learning_rate": 0.00019941958889929806, "loss": 0.9053, "step": 329 }, { "epoch": 0.06348595613697576, "grad_norm": 0.18878256593060205, "learning_rate": 0.0001994128661441911, "loss": 0.8872, "step": 330 }, { "epoch": 0.06367833782223932, "grad_norm": 0.19120690659966955, "learning_rate": 0.00019940610479358669, "loss": 0.8549, "step": 331 }, { "epoch": 0.06387071950750288, "grad_norm": 0.23654168903060382, "learning_rate": 0.00019939930485010968, "loss": 0.9093, "step": 332 }, { "epoch": 0.06406310119276645, "grad_norm": 0.19769021839098547, "learning_rate": 0.00019939246631640013, "loss": 0.8775, "step": 333 }, { "epoch": 0.06425548287803001, "grad_norm": 0.1960876942748798, "learning_rate": 0.00019938558919511298, "loss": 0.8895, "step": 334 }, { "epoch": 0.06444786456329357, "grad_norm": 0.18542632818164972, "learning_rate": 0.00019937867348891815, "loss": 0.851, "step": 335 }, { "epoch": 0.06464024624855713, "grad_norm": 0.17370081148738498, "learning_rate": 0.00019937171920050056, "loss": 0.8784, "step": 336 }, { "epoch": 0.0648326279338207, "grad_norm": 0.09451350432594514, "learning_rate": 0.00019936472633256012, "loss": 0.6209, "step": 337 }, { "epoch": 0.06502500961908426, "grad_norm": 0.18332106088644326, "learning_rate": 0.00019935769488781168, "loss": 0.8662, "step": 338 }, { "epoch": 0.06521739130434782, "grad_norm": 0.2123660515507496, "learning_rate": 0.0001993506248689851, "loss": 0.8493, "step": 339 }, { "epoch": 0.06540977298961138, "grad_norm": 0.20659415718482005, "learning_rate": 0.0001993435162788252, "loss": 0.9215, "step": 340 }, { "epoch": 0.06560215467487496, "grad_norm": 0.19743622321208454, "learning_rate": 0.0001993363691200918, "loss": 0.9405, "step": 341 }, { "epoch": 0.06579453636013852, "grad_norm": 0.18204537229427828, "learning_rate": 0.00019932918339555963, "loss": 0.8693, "step": 342 }, { "epoch": 0.06598691804540208, "grad_norm": 0.2073662781204831, "learning_rate": 0.00019932195910801848, "loss": 0.9702, "step": 343 }, { "epoch": 0.06617929973066564, "grad_norm": 0.19538553876964476, "learning_rate": 0.00019931469626027304, "loss": 0.8909, "step": 344 }, { "epoch": 0.06637168141592921, "grad_norm": 0.18277265793213177, "learning_rate": 0.00019930739485514302, "loss": 0.8999, "step": 345 }, { "epoch": 0.06656406310119277, "grad_norm": 0.20164534863315217, "learning_rate": 0.00019930005489546306, "loss": 0.9021, "step": 346 }, { "epoch": 0.06675644478645633, "grad_norm": 0.1928483144906915, "learning_rate": 0.00019929267638408278, "loss": 0.8638, "step": 347 }, { "epoch": 0.06694882647171989, "grad_norm": 0.3703322200739051, "learning_rate": 0.00019928525932386677, "loss": 0.6967, "step": 348 }, { "epoch": 0.06714120815698346, "grad_norm": 0.09637561801586485, "learning_rate": 0.00019927780371769462, "loss": 0.6389, "step": 349 }, { "epoch": 0.06733358984224702, "grad_norm": 0.1949760615937191, "learning_rate": 0.00019927030956846083, "loss": 0.8681, "step": 350 }, { "epoch": 0.06752597152751058, "grad_norm": 0.203084735569454, "learning_rate": 0.0001992627768790749, "loss": 0.8354, "step": 351 }, { "epoch": 0.06771835321277414, "grad_norm": 0.2007002915307087, "learning_rate": 0.00019925520565246122, "loss": 0.9083, "step": 352 }, { "epoch": 0.0679107348980377, "grad_norm": 0.2066931223659682, "learning_rate": 0.0001992475958915593, "loss": 0.9223, "step": 353 }, { "epoch": 0.06810311658330127, "grad_norm": 0.16713684935441364, "learning_rate": 0.00019923994759932343, "loss": 0.8516, "step": 354 }, { "epoch": 0.06829549826856483, "grad_norm": 0.19263921879625334, "learning_rate": 0.00019923226077872298, "loss": 0.9103, "step": 355 }, { "epoch": 0.06848787995382839, "grad_norm": 0.17523672458389067, "learning_rate": 0.0001992245354327422, "loss": 0.8739, "step": 356 }, { "epoch": 0.06868026163909195, "grad_norm": 0.1936619558920389, "learning_rate": 0.00019921677156438045, "loss": 0.9111, "step": 357 }, { "epoch": 0.06887264332435553, "grad_norm": 0.1857812208376336, "learning_rate": 0.00019920896917665178, "loss": 0.8911, "step": 358 }, { "epoch": 0.06906502500961909, "grad_norm": 0.17863864876167782, "learning_rate": 0.0001992011282725854, "loss": 0.8725, "step": 359 }, { "epoch": 0.06925740669488265, "grad_norm": 0.1970635964798975, "learning_rate": 0.00019919324885522545, "loss": 0.9318, "step": 360 }, { "epoch": 0.0694497883801462, "grad_norm": 0.18597537007896572, "learning_rate": 0.000199185330927631, "loss": 0.8507, "step": 361 }, { "epoch": 0.06964217006540978, "grad_norm": 0.1947972228510841, "learning_rate": 0.00019917737449287598, "loss": 0.8677, "step": 362 }, { "epoch": 0.06983455175067334, "grad_norm": 0.20146958829624786, "learning_rate": 0.00019916937955404942, "loss": 0.9296, "step": 363 }, { "epoch": 0.0700269334359369, "grad_norm": 0.1056954413208648, "learning_rate": 0.00019916134611425523, "loss": 0.6381, "step": 364 }, { "epoch": 0.07021931512120046, "grad_norm": 0.20617591560812562, "learning_rate": 0.00019915327417661225, "loss": 0.842, "step": 365 }, { "epoch": 0.07041169680646403, "grad_norm": 0.21175435183603175, "learning_rate": 0.00019914516374425429, "loss": 0.8399, "step": 366 }, { "epoch": 0.07060407849172759, "grad_norm": 0.09794325660813924, "learning_rate": 0.00019913701482033008, "loss": 0.6532, "step": 367 }, { "epoch": 0.07079646017699115, "grad_norm": 0.19767087734113198, "learning_rate": 0.00019912882740800333, "loss": 0.8412, "step": 368 }, { "epoch": 0.07098884186225471, "grad_norm": 0.19351649882661115, "learning_rate": 0.0001991206015104527, "loss": 0.8722, "step": 369 }, { "epoch": 0.07118122354751828, "grad_norm": 0.2195148413976203, "learning_rate": 0.0001991123371308717, "loss": 0.893, "step": 370 }, { "epoch": 0.07137360523278184, "grad_norm": 0.1885664898873106, "learning_rate": 0.00019910403427246895, "loss": 0.82, "step": 371 }, { "epoch": 0.0715659869180454, "grad_norm": 0.09880724181071018, "learning_rate": 0.0001990956929384678, "loss": 0.648, "step": 372 }, { "epoch": 0.07175836860330896, "grad_norm": 0.18050266925163447, "learning_rate": 0.0001990873131321067, "loss": 0.9523, "step": 373 }, { "epoch": 0.07195075028857253, "grad_norm": 0.17485103031356333, "learning_rate": 0.00019907889485663895, "loss": 0.8707, "step": 374 }, { "epoch": 0.0721431319738361, "grad_norm": 0.19231242081214625, "learning_rate": 0.00019907043811533283, "loss": 0.9096, "step": 375 }, { "epoch": 0.07233551365909965, "grad_norm": 0.18421887341523557, "learning_rate": 0.00019906194291147156, "loss": 0.9148, "step": 376 }, { "epoch": 0.07252789534436321, "grad_norm": 0.20108756006321887, "learning_rate": 0.00019905340924835323, "loss": 0.89, "step": 377 }, { "epoch": 0.07272027702962677, "grad_norm": 0.1818812405732916, "learning_rate": 0.00019904483712929094, "loss": 0.8739, "step": 378 }, { "epoch": 0.07291265871489035, "grad_norm": 0.19750763848357253, "learning_rate": 0.00019903622655761266, "loss": 0.9289, "step": 379 }, { "epoch": 0.0731050404001539, "grad_norm": 0.17950044795745854, "learning_rate": 0.0001990275775366613, "loss": 0.8751, "step": 380 }, { "epoch": 0.07329742208541747, "grad_norm": 0.20659211770033106, "learning_rate": 0.00019901889006979472, "loss": 0.915, "step": 381 }, { "epoch": 0.07348980377068103, "grad_norm": 0.2044981692887046, "learning_rate": 0.0001990101641603857, "loss": 0.8829, "step": 382 }, { "epoch": 0.0736821854559446, "grad_norm": 0.19542712458171305, "learning_rate": 0.00019900139981182194, "loss": 0.9447, "step": 383 }, { "epoch": 0.07387456714120816, "grad_norm": 0.09747258592897488, "learning_rate": 0.00019899259702750604, "loss": 0.6162, "step": 384 }, { "epoch": 0.07406694882647172, "grad_norm": 0.18403584475941925, "learning_rate": 0.00019898375581085557, "loss": 0.8283, "step": 385 }, { "epoch": 0.07425933051173528, "grad_norm": 0.18553191213190195, "learning_rate": 0.00019897487616530297, "loss": 0.8255, "step": 386 }, { "epoch": 0.07445171219699885, "grad_norm": 0.18208354057713144, "learning_rate": 0.00019896595809429566, "loss": 0.9139, "step": 387 }, { "epoch": 0.07464409388226241, "grad_norm": 0.17499442863587839, "learning_rate": 0.0001989570016012959, "loss": 0.8461, "step": 388 }, { "epoch": 0.07483647556752597, "grad_norm": 0.17065310259548697, "learning_rate": 0.00019894800668978095, "loss": 0.8962, "step": 389 }, { "epoch": 0.07502885725278953, "grad_norm": 0.18661727267680733, "learning_rate": 0.0001989389733632429, "loss": 0.8103, "step": 390 }, { "epoch": 0.0752212389380531, "grad_norm": 0.16821451389054812, "learning_rate": 0.00019892990162518882, "loss": 0.8406, "step": 391 }, { "epoch": 0.07541362062331666, "grad_norm": 0.2092452957352391, "learning_rate": 0.00019892079147914072, "loss": 0.8717, "step": 392 }, { "epoch": 0.07560600230858022, "grad_norm": 0.18895167110197925, "learning_rate": 0.00019891164292863537, "loss": 0.8555, "step": 393 }, { "epoch": 0.07579838399384378, "grad_norm": 0.18263150704420822, "learning_rate": 0.00019890245597722463, "loss": 0.8792, "step": 394 }, { "epoch": 0.07599076567910736, "grad_norm": 0.1653737008063093, "learning_rate": 0.00019889323062847515, "loss": 0.8878, "step": 395 }, { "epoch": 0.07618314736437092, "grad_norm": 0.18993611839008892, "learning_rate": 0.0001988839668859686, "loss": 0.8647, "step": 396 }, { "epoch": 0.07637552904963447, "grad_norm": 0.22716506674786516, "learning_rate": 0.00019887466475330138, "loss": 0.8849, "step": 397 }, { "epoch": 0.07656791073489803, "grad_norm": 0.16493171787780397, "learning_rate": 0.00019886532423408495, "loss": 0.9018, "step": 398 }, { "epoch": 0.0767602924201616, "grad_norm": 0.18386908339078828, "learning_rate": 0.00019885594533194562, "loss": 0.9191, "step": 399 }, { "epoch": 0.07695267410542517, "grad_norm": 0.19685890674937195, "learning_rate": 0.00019884652805052464, "loss": 0.8689, "step": 400 }, { "epoch": 0.07714505579068873, "grad_norm": 0.17522223126047687, "learning_rate": 0.00019883707239347803, "loss": 0.8751, "step": 401 }, { "epoch": 0.07733743747595229, "grad_norm": 0.2059081592677902, "learning_rate": 0.0001988275783644769, "loss": 0.9056, "step": 402 }, { "epoch": 0.07752981916121585, "grad_norm": 0.16942297963312258, "learning_rate": 0.0001988180459672071, "loss": 0.8343, "step": 403 }, { "epoch": 0.07772220084647942, "grad_norm": 0.20239789337292607, "learning_rate": 0.00019880847520536946, "loss": 0.9255, "step": 404 }, { "epoch": 0.07791458253174298, "grad_norm": 0.18172979340517512, "learning_rate": 0.00019879886608267967, "loss": 0.8458, "step": 405 }, { "epoch": 0.07810696421700654, "grad_norm": 0.19409730231274164, "learning_rate": 0.0001987892186028683, "loss": 0.9309, "step": 406 }, { "epoch": 0.0782993459022701, "grad_norm": 0.1879866004327857, "learning_rate": 0.00019877953276968086, "loss": 0.8803, "step": 407 }, { "epoch": 0.07849172758753367, "grad_norm": 0.2075267093831644, "learning_rate": 0.00019876980858687775, "loss": 0.8615, "step": 408 }, { "epoch": 0.07868410927279723, "grad_norm": 0.1761946903871106, "learning_rate": 0.00019876004605823419, "loss": 0.8683, "step": 409 }, { "epoch": 0.07887649095806079, "grad_norm": 0.20860018264067454, "learning_rate": 0.0001987502451875403, "loss": 0.8773, "step": 410 }, { "epoch": 0.07906887264332435, "grad_norm": 0.18639085021054794, "learning_rate": 0.0001987404059786012, "loss": 0.9069, "step": 411 }, { "epoch": 0.07926125432858792, "grad_norm": 0.22355582767554252, "learning_rate": 0.00019873052843523677, "loss": 0.8703, "step": 412 }, { "epoch": 0.07945363601385148, "grad_norm": 0.19207050888242877, "learning_rate": 0.0001987206125612818, "loss": 0.9204, "step": 413 }, { "epoch": 0.07964601769911504, "grad_norm": 0.20987313956486742, "learning_rate": 0.00019871065836058596, "loss": 0.9091, "step": 414 }, { "epoch": 0.0798383993843786, "grad_norm": 0.1854089085196169, "learning_rate": 0.00019870066583701387, "loss": 0.8712, "step": 415 }, { "epoch": 0.08003078106964218, "grad_norm": 0.2008598964099777, "learning_rate": 0.00019869063499444495, "loss": 0.8988, "step": 416 }, { "epoch": 0.08022316275490574, "grad_norm": 0.18712393410417158, "learning_rate": 0.00019868056583677345, "loss": 0.8699, "step": 417 }, { "epoch": 0.0804155444401693, "grad_norm": 0.19370996501270193, "learning_rate": 0.00019867045836790867, "loss": 0.8502, "step": 418 }, { "epoch": 0.08060792612543285, "grad_norm": 0.0954584050610073, "learning_rate": 0.00019866031259177462, "loss": 0.6282, "step": 419 }, { "epoch": 0.08080030781069643, "grad_norm": 0.22093117591033742, "learning_rate": 0.00019865012851231023, "loss": 0.9555, "step": 420 }, { "epoch": 0.08099268949595999, "grad_norm": 0.20010730971538157, "learning_rate": 0.00019863990613346933, "loss": 0.8725, "step": 421 }, { "epoch": 0.08118507118122355, "grad_norm": 0.20662614746587746, "learning_rate": 0.0001986296454592206, "loss": 0.9061, "step": 422 }, { "epoch": 0.08137745286648711, "grad_norm": 0.18374313561642347, "learning_rate": 0.00019861934649354762, "loss": 0.8216, "step": 423 }, { "epoch": 0.08156983455175067, "grad_norm": 0.2129932989870255, "learning_rate": 0.00019860900924044872, "loss": 0.8671, "step": 424 }, { "epoch": 0.08176221623701424, "grad_norm": 0.13255598527998488, "learning_rate": 0.00019859863370393725, "loss": 0.6675, "step": 425 }, { "epoch": 0.0819545979222778, "grad_norm": 0.23123445848153626, "learning_rate": 0.00019858821988804132, "loss": 0.8448, "step": 426 }, { "epoch": 0.08214697960754136, "grad_norm": 0.18322732343103434, "learning_rate": 0.00019857776779680394, "loss": 0.9045, "step": 427 }, { "epoch": 0.08233936129280492, "grad_norm": 0.1945981321534598, "learning_rate": 0.00019856727743428296, "loss": 0.9168, "step": 428 }, { "epoch": 0.08253174297806849, "grad_norm": 0.16565919627586587, "learning_rate": 0.00019855674880455112, "loss": 0.8624, "step": 429 }, { "epoch": 0.08272412466333205, "grad_norm": 0.17945116216313675, "learning_rate": 0.00019854618191169599, "loss": 0.859, "step": 430 }, { "epoch": 0.08291650634859561, "grad_norm": 0.30773877946659656, "learning_rate": 0.00019853557675982, "loss": 0.9047, "step": 431 }, { "epoch": 0.08310888803385917, "grad_norm": 0.2045905688374686, "learning_rate": 0.0001985249333530404, "loss": 0.9215, "step": 432 }, { "epoch": 0.08330126971912274, "grad_norm": 0.19246782555720815, "learning_rate": 0.00019851425169548938, "loss": 0.8995, "step": 433 }, { "epoch": 0.0834936514043863, "grad_norm": 0.18642856735623464, "learning_rate": 0.0001985035317913139, "loss": 0.8508, "step": 434 }, { "epoch": 0.08368603308964986, "grad_norm": 0.18581382270624958, "learning_rate": 0.0001984927736446758, "loss": 0.9132, "step": 435 }, { "epoch": 0.08387841477491342, "grad_norm": 0.09522835053530213, "learning_rate": 0.0001984819772597518, "loss": 0.6616, "step": 436 }, { "epoch": 0.084070796460177, "grad_norm": 0.20140352446188517, "learning_rate": 0.00019847114264073338, "loss": 0.8343, "step": 437 }, { "epoch": 0.08426317814544056, "grad_norm": 0.1738846210031312, "learning_rate": 0.0001984602697918269, "loss": 0.8824, "step": 438 }, { "epoch": 0.08445555983070412, "grad_norm": 0.19348453077920266, "learning_rate": 0.0001984493587172536, "loss": 0.8661, "step": 439 }, { "epoch": 0.08464794151596768, "grad_norm": 0.1767314485639461, "learning_rate": 0.00019843840942124956, "loss": 0.8987, "step": 440 }, { "epoch": 0.08484032320123125, "grad_norm": 0.18018795596619924, "learning_rate": 0.00019842742190806566, "loss": 0.8485, "step": 441 }, { "epoch": 0.08503270488649481, "grad_norm": 0.1819691219419223, "learning_rate": 0.00019841639618196758, "loss": 0.9076, "step": 442 }, { "epoch": 0.08522508657175837, "grad_norm": 0.16689396017102867, "learning_rate": 0.00019840533224723593, "loss": 0.8513, "step": 443 }, { "epoch": 0.08541746825702193, "grad_norm": 0.1954775295496425, "learning_rate": 0.00019839423010816615, "loss": 0.8725, "step": 444 }, { "epoch": 0.08560984994228549, "grad_norm": 0.19160302057083756, "learning_rate": 0.0001983830897690684, "loss": 0.8326, "step": 445 }, { "epoch": 0.08580223162754906, "grad_norm": 0.18516954244544032, "learning_rate": 0.00019837191123426776, "loss": 0.922, "step": 446 }, { "epoch": 0.08599461331281262, "grad_norm": 0.21487741306997044, "learning_rate": 0.00019836069450810418, "loss": 0.9455, "step": 447 }, { "epoch": 0.08618699499807618, "grad_norm": 0.1879598597425328, "learning_rate": 0.0001983494395949323, "loss": 0.9069, "step": 448 }, { "epoch": 0.08637937668333974, "grad_norm": 0.17908633489094755, "learning_rate": 0.0001983381464991217, "loss": 0.9018, "step": 449 }, { "epoch": 0.08657175836860331, "grad_norm": 0.17264100923788495, "learning_rate": 0.00019832681522505676, "loss": 0.8353, "step": 450 }, { "epoch": 0.08676414005386687, "grad_norm": 0.1779437696462019, "learning_rate": 0.00019831544577713663, "loss": 0.8421, "step": 451 }, { "epoch": 0.08695652173913043, "grad_norm": 0.16300083323946868, "learning_rate": 0.0001983040381597754, "loss": 0.8511, "step": 452 }, { "epoch": 0.08714890342439399, "grad_norm": 0.22685818372994046, "learning_rate": 0.00019829259237740177, "loss": 0.9242, "step": 453 }, { "epoch": 0.08734128510965757, "grad_norm": 0.09220034059072824, "learning_rate": 0.00019828110843445954, "loss": 0.6312, "step": 454 }, { "epoch": 0.08753366679492113, "grad_norm": 0.09670617481150318, "learning_rate": 0.00019826958633540706, "loss": 0.6712, "step": 455 }, { "epoch": 0.08772604848018468, "grad_norm": 0.20169186256113777, "learning_rate": 0.00019825802608471765, "loss": 0.8762, "step": 456 }, { "epoch": 0.08791843016544824, "grad_norm": 0.1835246322326427, "learning_rate": 0.0001982464276868794, "loss": 0.8295, "step": 457 }, { "epoch": 0.08811081185071182, "grad_norm": 0.20698392861994616, "learning_rate": 0.0001982347911463952, "loss": 0.8821, "step": 458 }, { "epoch": 0.08830319353597538, "grad_norm": 0.19244226132575382, "learning_rate": 0.00019822311646778277, "loss": 0.9558, "step": 459 }, { "epoch": 0.08849557522123894, "grad_norm": 0.17914344100611682, "learning_rate": 0.00019821140365557457, "loss": 0.9094, "step": 460 }, { "epoch": 0.0886879569065025, "grad_norm": 0.20260344984571785, "learning_rate": 0.00019819965271431796, "loss": 0.888, "step": 461 }, { "epoch": 0.08888033859176607, "grad_norm": 0.19127181285122039, "learning_rate": 0.00019818786364857505, "loss": 0.882, "step": 462 }, { "epoch": 0.08907272027702963, "grad_norm": 0.20804089376188187, "learning_rate": 0.00019817603646292276, "loss": 0.8546, "step": 463 }, { "epoch": 0.08926510196229319, "grad_norm": 0.1800219784819821, "learning_rate": 0.00019816417116195285, "loss": 0.8574, "step": 464 }, { "epoch": 0.08945748364755675, "grad_norm": 0.19231098786334022, "learning_rate": 0.0001981522677502718, "loss": 0.8676, "step": 465 }, { "epoch": 0.08964986533282032, "grad_norm": 0.08700303154304541, "learning_rate": 0.00019814032623250092, "loss": 0.6113, "step": 466 }, { "epoch": 0.08984224701808388, "grad_norm": 0.19031595737944432, "learning_rate": 0.0001981283466132763, "loss": 0.8987, "step": 467 }, { "epoch": 0.09003462870334744, "grad_norm": 0.19483306365839861, "learning_rate": 0.00019811632889724887, "loss": 0.8609, "step": 468 }, { "epoch": 0.090227010388611, "grad_norm": 0.1795007304503445, "learning_rate": 0.00019810427308908436, "loss": 0.8427, "step": 469 }, { "epoch": 0.09041939207387456, "grad_norm": 0.18478547691615088, "learning_rate": 0.00019809217919346315, "loss": 0.9011, "step": 470 }, { "epoch": 0.09061177375913813, "grad_norm": 0.22164423688169008, "learning_rate": 0.0001980800472150806, "loss": 0.883, "step": 471 }, { "epoch": 0.0908041554444017, "grad_norm": 0.19216216986885568, "learning_rate": 0.00019806787715864673, "loss": 0.8899, "step": 472 }, { "epoch": 0.09099653712966525, "grad_norm": 0.18956173010356578, "learning_rate": 0.00019805566902888634, "loss": 0.8564, "step": 473 }, { "epoch": 0.09118891881492881, "grad_norm": 0.1872957281658341, "learning_rate": 0.00019804342283053915, "loss": 0.904, "step": 474 }, { "epoch": 0.09138130050019239, "grad_norm": 0.18528747933589718, "learning_rate": 0.00019803113856835938, "loss": 0.7958, "step": 475 }, { "epoch": 0.09157368218545595, "grad_norm": 0.18223359156239913, "learning_rate": 0.00019801881624711637, "loss": 0.9009, "step": 476 }, { "epoch": 0.0917660638707195, "grad_norm": 0.19825874066276383, "learning_rate": 0.000198006455871594, "loss": 0.8609, "step": 477 }, { "epoch": 0.09195844555598306, "grad_norm": 0.20766353835132875, "learning_rate": 0.000197994057446591, "loss": 0.8847, "step": 478 }, { "epoch": 0.09215082724124664, "grad_norm": 0.20586797880522203, "learning_rate": 0.00019798162097692089, "loss": 0.8741, "step": 479 }, { "epoch": 0.0923432089265102, "grad_norm": 0.22081774482654654, "learning_rate": 0.00019796914646741186, "loss": 0.9088, "step": 480 }, { "epoch": 0.09253559061177376, "grad_norm": 0.20917646439647275, "learning_rate": 0.000197956633922907, "loss": 0.906, "step": 481 }, { "epoch": 0.09272797229703732, "grad_norm": 0.09436167272623047, "learning_rate": 0.00019794408334826414, "loss": 0.6444, "step": 482 }, { "epoch": 0.09292035398230089, "grad_norm": 0.20159047850379128, "learning_rate": 0.0001979314947483558, "loss": 0.878, "step": 483 }, { "epoch": 0.09311273566756445, "grad_norm": 0.17732927277749957, "learning_rate": 0.0001979188681280693, "loss": 0.8751, "step": 484 }, { "epoch": 0.09330511735282801, "grad_norm": 0.19202032357815257, "learning_rate": 0.00019790620349230674, "loss": 0.8107, "step": 485 }, { "epoch": 0.09349749903809157, "grad_norm": 0.08598321372051802, "learning_rate": 0.000197893500845985, "loss": 0.6184, "step": 486 }, { "epoch": 0.09368988072335514, "grad_norm": 0.18499360263669415, "learning_rate": 0.00019788076019403565, "loss": 0.8732, "step": 487 }, { "epoch": 0.0938822624086187, "grad_norm": 0.19048390049180755, "learning_rate": 0.00019786798154140507, "loss": 0.8366, "step": 488 }, { "epoch": 0.09407464409388226, "grad_norm": 0.1756754710211176, "learning_rate": 0.00019785516489305435, "loss": 0.8465, "step": 489 }, { "epoch": 0.09426702577914582, "grad_norm": 0.20020202693913303, "learning_rate": 0.00019784231025395937, "loss": 0.8849, "step": 490 }, { "epoch": 0.0944594074644094, "grad_norm": 0.17059353238347863, "learning_rate": 0.00019782941762911074, "loss": 0.9071, "step": 491 }, { "epoch": 0.09465178914967295, "grad_norm": 0.19356155944689796, "learning_rate": 0.00019781648702351382, "loss": 0.8659, "step": 492 }, { "epoch": 0.09484417083493651, "grad_norm": 0.23433654479253369, "learning_rate": 0.0001978035184421887, "loss": 0.9155, "step": 493 }, { "epoch": 0.09503655252020007, "grad_norm": 0.18861569440879955, "learning_rate": 0.0001977905118901703, "loss": 0.9038, "step": 494 }, { "epoch": 0.09522893420546363, "grad_norm": 0.17178646218591218, "learning_rate": 0.00019777746737250807, "loss": 0.8659, "step": 495 }, { "epoch": 0.0954213158907272, "grad_norm": 0.18289596801156197, "learning_rate": 0.0001977643848942665, "loss": 0.8979, "step": 496 }, { "epoch": 0.09561369757599077, "grad_norm": 0.20397231407172278, "learning_rate": 0.00019775126446052458, "loss": 0.8857, "step": 497 }, { "epoch": 0.09580607926125433, "grad_norm": 0.0940089857248865, "learning_rate": 0.0001977381060763761, "loss": 0.6364, "step": 498 }, { "epoch": 0.09599846094651789, "grad_norm": 0.17927682669691095, "learning_rate": 0.00019772490974692962, "loss": 0.8893, "step": 499 }, { "epoch": 0.09619084263178146, "grad_norm": 0.21581921723464423, "learning_rate": 0.00019771167547730844, "loss": 0.8886, "step": 500 }, { "epoch": 0.09638322431704502, "grad_norm": 0.19152379356216426, "learning_rate": 0.0001976984032726505, "loss": 0.8892, "step": 501 }, { "epoch": 0.09657560600230858, "grad_norm": 0.21346405905019236, "learning_rate": 0.0001976850931381086, "loss": 0.8989, "step": 502 }, { "epoch": 0.09676798768757214, "grad_norm": 0.1821084707801725, "learning_rate": 0.00019767174507885008, "loss": 0.8741, "step": 503 }, { "epoch": 0.09696036937283571, "grad_norm": 0.19529625210374907, "learning_rate": 0.00019765835910005722, "loss": 0.9352, "step": 504 }, { "epoch": 0.09715275105809927, "grad_norm": 0.19779310076276915, "learning_rate": 0.00019764493520692687, "loss": 0.8554, "step": 505 }, { "epoch": 0.09734513274336283, "grad_norm": 0.209869773870797, "learning_rate": 0.00019763147340467067, "loss": 0.8841, "step": 506 }, { "epoch": 0.09753751442862639, "grad_norm": 0.1868131023296359, "learning_rate": 0.00019761797369851497, "loss": 0.8568, "step": 507 }, { "epoch": 0.09772989611388996, "grad_norm": 0.18988346835536754, "learning_rate": 0.00019760443609370074, "loss": 0.8438, "step": 508 }, { "epoch": 0.09792227779915352, "grad_norm": 0.10162377677180714, "learning_rate": 0.00019759086059548383, "loss": 0.6588, "step": 509 }, { "epoch": 0.09811465948441708, "grad_norm": 0.20374305446788454, "learning_rate": 0.00019757724720913467, "loss": 0.8972, "step": 510 }, { "epoch": 0.09830704116968064, "grad_norm": 0.18826777933396566, "learning_rate": 0.00019756359593993845, "loss": 0.8523, "step": 511 }, { "epoch": 0.09849942285494422, "grad_norm": 0.16972618025153477, "learning_rate": 0.0001975499067931951, "loss": 0.8682, "step": 512 }, { "epoch": 0.09869180454020778, "grad_norm": 0.19278847047770384, "learning_rate": 0.0001975361797742192, "loss": 0.8324, "step": 513 }, { "epoch": 0.09888418622547133, "grad_norm": 0.2143323195086593, "learning_rate": 0.00019752241488834002, "loss": 0.8301, "step": 514 }, { "epoch": 0.0990765679107349, "grad_norm": 0.24020912158383426, "learning_rate": 0.0001975086121409016, "loss": 0.9195, "step": 515 }, { "epoch": 0.09926894959599845, "grad_norm": 0.33890476868043795, "learning_rate": 0.0001974947715372626, "loss": 0.8831, "step": 516 }, { "epoch": 0.09946133128126203, "grad_norm": 0.20597176556774563, "learning_rate": 0.00019748089308279648, "loss": 0.8843, "step": 517 }, { "epoch": 0.09965371296652559, "grad_norm": 0.19717879837619873, "learning_rate": 0.0001974669767828913, "loss": 0.8753, "step": 518 }, { "epoch": 0.09984609465178915, "grad_norm": 0.20209832671838074, "learning_rate": 0.0001974530226429498, "loss": 0.925, "step": 519 }, { "epoch": 0.1000384763370527, "grad_norm": 0.21251747380928387, "learning_rate": 0.00019743903066838953, "loss": 0.9079, "step": 520 }, { "epoch": 0.10023085802231628, "grad_norm": 0.19770980478743208, "learning_rate": 0.00019742500086464263, "loss": 0.8336, "step": 521 }, { "epoch": 0.10042323970757984, "grad_norm": 0.19847868741912178, "learning_rate": 0.00019741093323715595, "loss": 0.822, "step": 522 }, { "epoch": 0.1006156213928434, "grad_norm": 0.19320995895731008, "learning_rate": 0.00019739682779139106, "loss": 0.8925, "step": 523 }, { "epoch": 0.10080800307810696, "grad_norm": 0.1965254375441096, "learning_rate": 0.0001973826845328241, "loss": 0.8456, "step": 524 }, { "epoch": 0.10100038476337053, "grad_norm": 0.18103005683922732, "learning_rate": 0.0001973685034669461, "loss": 0.8349, "step": 525 }, { "epoch": 0.10119276644863409, "grad_norm": 0.17910074756162067, "learning_rate": 0.00019735428459926249, "loss": 0.9185, "step": 526 }, { "epoch": 0.10138514813389765, "grad_norm": 0.20292708863416178, "learning_rate": 0.0001973400279352936, "loss": 0.9038, "step": 527 }, { "epoch": 0.10157752981916121, "grad_norm": 0.16997600235449278, "learning_rate": 0.00019732573348057436, "loss": 0.8618, "step": 528 }, { "epoch": 0.10176991150442478, "grad_norm": 0.18674805629725447, "learning_rate": 0.00019731140124065437, "loss": 0.8353, "step": 529 }, { "epoch": 0.10196229318968834, "grad_norm": 0.17097198292133842, "learning_rate": 0.00019729703122109787, "loss": 0.8856, "step": 530 }, { "epoch": 0.1021546748749519, "grad_norm": 0.1867761424541795, "learning_rate": 0.00019728262342748383, "loss": 0.8743, "step": 531 }, { "epoch": 0.10234705656021546, "grad_norm": 0.17853204811052653, "learning_rate": 0.00019726817786540584, "loss": 0.8588, "step": 532 }, { "epoch": 0.10253943824547904, "grad_norm": 0.2044900311106747, "learning_rate": 0.00019725369454047214, "loss": 0.8279, "step": 533 }, { "epoch": 0.1027318199307426, "grad_norm": 0.32059411485566347, "learning_rate": 0.00019723917345830568, "loss": 0.853, "step": 534 }, { "epoch": 0.10292420161600616, "grad_norm": 0.20381100856827913, "learning_rate": 0.00019722461462454404, "loss": 0.871, "step": 535 }, { "epoch": 0.10311658330126972, "grad_norm": 0.1905296361783215, "learning_rate": 0.00019721001804483948, "loss": 0.8616, "step": 536 }, { "epoch": 0.10330896498653329, "grad_norm": 0.18610269675342317, "learning_rate": 0.00019719538372485885, "loss": 0.9147, "step": 537 }, { "epoch": 0.10350134667179685, "grad_norm": 0.19220580177339214, "learning_rate": 0.00019718071167028375, "loss": 0.8497, "step": 538 }, { "epoch": 0.10369372835706041, "grad_norm": 0.2035477817652282, "learning_rate": 0.00019716600188681037, "loss": 0.8444, "step": 539 }, { "epoch": 0.10388611004232397, "grad_norm": 0.1750848508679971, "learning_rate": 0.00019715125438014947, "loss": 0.8258, "step": 540 }, { "epoch": 0.10407849172758753, "grad_norm": 0.18324560547753413, "learning_rate": 0.00019713646915602663, "loss": 0.8433, "step": 541 }, { "epoch": 0.1042708734128511, "grad_norm": 0.1976662737549662, "learning_rate": 0.00019712164622018197, "loss": 0.9065, "step": 542 }, { "epoch": 0.10446325509811466, "grad_norm": 0.17409821265306055, "learning_rate": 0.00019710678557837024, "loss": 0.9168, "step": 543 }, { "epoch": 0.10465563678337822, "grad_norm": 0.19692441411079603, "learning_rate": 0.00019709188723636086, "loss": 0.8942, "step": 544 }, { "epoch": 0.10484801846864178, "grad_norm": 0.20105695763354683, "learning_rate": 0.00019707695119993793, "loss": 0.8896, "step": 545 }, { "epoch": 0.10504040015390535, "grad_norm": 0.1898915103559077, "learning_rate": 0.00019706197747490001, "loss": 0.878, "step": 546 }, { "epoch": 0.10523278183916891, "grad_norm": 0.19238888622683417, "learning_rate": 0.00019704696606706053, "loss": 0.8657, "step": 547 }, { "epoch": 0.10542516352443247, "grad_norm": 0.18132283351400338, "learning_rate": 0.00019703191698224742, "loss": 0.7986, "step": 548 }, { "epoch": 0.10561754520969603, "grad_norm": 0.20449200227495248, "learning_rate": 0.0001970168302263032, "loss": 0.8695, "step": 549 }, { "epoch": 0.1058099268949596, "grad_norm": 0.17451352977961487, "learning_rate": 0.00019700170580508512, "loss": 0.8756, "step": 550 }, { "epoch": 0.10600230858022316, "grad_norm": 0.17074861703372451, "learning_rate": 0.00019698654372446493, "loss": 0.8482, "step": 551 }, { "epoch": 0.10619469026548672, "grad_norm": 0.17679407029445057, "learning_rate": 0.0001969713439903292, "loss": 0.9154, "step": 552 }, { "epoch": 0.10638707195075028, "grad_norm": 0.19153766257714486, "learning_rate": 0.00019695610660857887, "loss": 0.8959, "step": 553 }, { "epoch": 0.10657945363601386, "grad_norm": 0.20424936013272207, "learning_rate": 0.00019694083158512964, "loss": 0.91, "step": 554 }, { "epoch": 0.10677183532127742, "grad_norm": 0.1778245201436538, "learning_rate": 0.00019692551892591185, "loss": 0.909, "step": 555 }, { "epoch": 0.10696421700654098, "grad_norm": 0.1662179145638787, "learning_rate": 0.00019691016863687035, "loss": 0.8395, "step": 556 }, { "epoch": 0.10715659869180454, "grad_norm": 0.19031137052967098, "learning_rate": 0.00019689478072396467, "loss": 0.8527, "step": 557 }, { "epoch": 0.10734898037706811, "grad_norm": 0.19716439941487746, "learning_rate": 0.00019687935519316897, "loss": 0.9226, "step": 558 }, { "epoch": 0.10754136206233167, "grad_norm": 0.09081537183879476, "learning_rate": 0.00019686389205047188, "loss": 0.5833, "step": 559 }, { "epoch": 0.10773374374759523, "grad_norm": 0.17735405514217828, "learning_rate": 0.0001968483913018768, "loss": 0.92, "step": 560 }, { "epoch": 0.10792612543285879, "grad_norm": 0.21096478848481914, "learning_rate": 0.0001968328529534016, "loss": 0.8788, "step": 561 }, { "epoch": 0.10811850711812235, "grad_norm": 0.20561044880372717, "learning_rate": 0.00019681727701107885, "loss": 0.8271, "step": 562 }, { "epoch": 0.10831088880338592, "grad_norm": 0.20715018186045298, "learning_rate": 0.00019680166348095564, "loss": 0.8824, "step": 563 }, { "epoch": 0.10850327048864948, "grad_norm": 0.18211570899354107, "learning_rate": 0.00019678601236909371, "loss": 0.8424, "step": 564 }, { "epoch": 0.10869565217391304, "grad_norm": 0.1724948992744619, "learning_rate": 0.00019677032368156931, "loss": 0.885, "step": 565 }, { "epoch": 0.1088880338591766, "grad_norm": 0.18543427538104898, "learning_rate": 0.00019675459742447338, "loss": 0.8545, "step": 566 }, { "epoch": 0.10908041554444017, "grad_norm": 0.18367006168108005, "learning_rate": 0.00019673883360391136, "loss": 0.8572, "step": 567 }, { "epoch": 0.10927279722970373, "grad_norm": 0.19932852376107163, "learning_rate": 0.00019672303222600333, "loss": 0.8066, "step": 568 }, { "epoch": 0.10946517891496729, "grad_norm": 0.17861299385352536, "learning_rate": 0.00019670719329688388, "loss": 0.9114, "step": 569 }, { "epoch": 0.10965756060023085, "grad_norm": 0.10593130118023346, "learning_rate": 0.0001966913168227023, "loss": 0.6609, "step": 570 }, { "epoch": 0.10984994228549443, "grad_norm": 0.19614744276761206, "learning_rate": 0.00019667540280962235, "loss": 0.9055, "step": 571 }, { "epoch": 0.11004232397075799, "grad_norm": 0.17938842421032292, "learning_rate": 0.0001966594512638224, "loss": 0.8049, "step": 572 }, { "epoch": 0.11023470565602154, "grad_norm": 0.1956775755537338, "learning_rate": 0.0001966434621914954, "loss": 0.8568, "step": 573 }, { "epoch": 0.1104270873412851, "grad_norm": 0.19408412261829142, "learning_rate": 0.0001966274355988488, "loss": 0.8791, "step": 574 }, { "epoch": 0.11061946902654868, "grad_norm": 0.16341781740809622, "learning_rate": 0.00019661137149210473, "loss": 0.8307, "step": 575 }, { "epoch": 0.11081185071181224, "grad_norm": 0.19545978634313405, "learning_rate": 0.00019659526987749986, "loss": 0.8521, "step": 576 }, { "epoch": 0.1110042323970758, "grad_norm": 0.1692528849805835, "learning_rate": 0.00019657913076128532, "loss": 0.8942, "step": 577 }, { "epoch": 0.11119661408233936, "grad_norm": 0.20497424298937766, "learning_rate": 0.0001965629541497269, "loss": 0.8753, "step": 578 }, { "epoch": 0.11138899576760293, "grad_norm": 0.20394437018958905, "learning_rate": 0.00019654674004910492, "loss": 0.8435, "step": 579 }, { "epoch": 0.11158137745286649, "grad_norm": 0.18195601036381653, "learning_rate": 0.00019653048846571426, "loss": 0.8662, "step": 580 }, { "epoch": 0.11177375913813005, "grad_norm": 0.18836336931405456, "learning_rate": 0.00019651419940586436, "loss": 0.8645, "step": 581 }, { "epoch": 0.11196614082339361, "grad_norm": 0.1867926686566879, "learning_rate": 0.00019649787287587912, "loss": 0.9289, "step": 582 }, { "epoch": 0.11215852250865718, "grad_norm": 0.191350021621172, "learning_rate": 0.00019648150888209714, "loss": 0.8597, "step": 583 }, { "epoch": 0.11235090419392074, "grad_norm": 0.17879739545451537, "learning_rate": 0.00019646510743087143, "loss": 0.9037, "step": 584 }, { "epoch": 0.1125432858791843, "grad_norm": 0.1910311807498745, "learning_rate": 0.00019644866852856966, "loss": 0.9096, "step": 585 }, { "epoch": 0.11273566756444786, "grad_norm": 0.19817203398926528, "learning_rate": 0.00019643219218157392, "loss": 0.8344, "step": 586 }, { "epoch": 0.11292804924971142, "grad_norm": 0.1708294683740298, "learning_rate": 0.00019641567839628093, "loss": 0.8602, "step": 587 }, { "epoch": 0.113120430934975, "grad_norm": 0.19740660572351437, "learning_rate": 0.0001963991271791019, "loss": 0.8019, "step": 588 }, { "epoch": 0.11331281262023855, "grad_norm": 0.20215353080289666, "learning_rate": 0.00019638253853646257, "loss": 0.809, "step": 589 }, { "epoch": 0.11350519430550211, "grad_norm": 0.18309939099321662, "learning_rate": 0.00019636591247480323, "loss": 0.8394, "step": 590 }, { "epoch": 0.11369757599076567, "grad_norm": 0.21100530082144625, "learning_rate": 0.00019634924900057866, "loss": 0.9364, "step": 591 }, { "epoch": 0.11388995767602925, "grad_norm": 0.19261698723568313, "learning_rate": 0.0001963325481202583, "loss": 0.8334, "step": 592 }, { "epoch": 0.1140823393612928, "grad_norm": 0.1924693242345748, "learning_rate": 0.0001963158098403259, "loss": 0.8283, "step": 593 }, { "epoch": 0.11427472104655637, "grad_norm": 0.18854118940646158, "learning_rate": 0.00019629903416727986, "loss": 0.8402, "step": 594 }, { "epoch": 0.11446710273181993, "grad_norm": 0.09572575163901947, "learning_rate": 0.0001962822211076331, "loss": 0.6429, "step": 595 }, { "epoch": 0.1146594844170835, "grad_norm": 0.19730956004646336, "learning_rate": 0.00019626537066791297, "loss": 0.8568, "step": 596 }, { "epoch": 0.11485186610234706, "grad_norm": 0.20539351858416616, "learning_rate": 0.00019624848285466147, "loss": 0.8431, "step": 597 }, { "epoch": 0.11504424778761062, "grad_norm": 0.17200784658201845, "learning_rate": 0.00019623155767443498, "loss": 0.8452, "step": 598 }, { "epoch": 0.11523662947287418, "grad_norm": 0.1842396269383066, "learning_rate": 0.00019621459513380444, "loss": 0.8665, "step": 599 }, { "epoch": 0.11542901115813775, "grad_norm": 0.18533408597254378, "learning_rate": 0.00019619759523935532, "loss": 0.8986, "step": 600 }, { "epoch": 0.11562139284340131, "grad_norm": 0.2035711071207875, "learning_rate": 0.00019618055799768756, "loss": 0.8157, "step": 601 }, { "epoch": 0.11581377452866487, "grad_norm": 0.17857936641070654, "learning_rate": 0.00019616348341541557, "loss": 0.8572, "step": 602 }, { "epoch": 0.11600615621392843, "grad_norm": 0.09949855018720144, "learning_rate": 0.00019614637149916836, "loss": 0.696, "step": 603 }, { "epoch": 0.116198537899192, "grad_norm": 0.1970935872023361, "learning_rate": 0.00019612922225558925, "loss": 0.9095, "step": 604 }, { "epoch": 0.11639091958445556, "grad_norm": 0.21752312013905556, "learning_rate": 0.00019611203569133627, "loss": 0.8485, "step": 605 }, { "epoch": 0.11658330126971912, "grad_norm": 0.1878700636657265, "learning_rate": 0.00019609481181308177, "loss": 0.9261, "step": 606 }, { "epoch": 0.11677568295498268, "grad_norm": 0.198202604442119, "learning_rate": 0.00019607755062751273, "loss": 0.8776, "step": 607 }, { "epoch": 0.11696806464024626, "grad_norm": 0.18565451858480228, "learning_rate": 0.00019606025214133046, "loss": 0.8756, "step": 608 }, { "epoch": 0.11716044632550981, "grad_norm": 0.17652521899480011, "learning_rate": 0.00019604291636125085, "loss": 0.8732, "step": 609 }, { "epoch": 0.11735282801077337, "grad_norm": 0.19775986018103875, "learning_rate": 0.00019602554329400428, "loss": 0.8849, "step": 610 }, { "epoch": 0.11754520969603693, "grad_norm": 0.21613828458168172, "learning_rate": 0.00019600813294633554, "loss": 0.8717, "step": 611 }, { "epoch": 0.1177375913813005, "grad_norm": 0.09831399357292449, "learning_rate": 0.00019599068532500394, "loss": 0.6239, "step": 612 }, { "epoch": 0.11792997306656407, "grad_norm": 0.2141334145597178, "learning_rate": 0.00019597320043678322, "loss": 0.8586, "step": 613 }, { "epoch": 0.11812235475182763, "grad_norm": 0.2057797171476571, "learning_rate": 0.00019595567828846163, "loss": 0.8696, "step": 614 }, { "epoch": 0.11831473643709119, "grad_norm": 0.19611677630917335, "learning_rate": 0.0001959381188868419, "loss": 0.8152, "step": 615 }, { "epoch": 0.11850711812235475, "grad_norm": 0.17418915191012563, "learning_rate": 0.00019592052223874117, "loss": 0.8622, "step": 616 }, { "epoch": 0.11869949980761832, "grad_norm": 0.1795298911245062, "learning_rate": 0.00019590288835099107, "loss": 0.853, "step": 617 }, { "epoch": 0.11889188149288188, "grad_norm": 0.19889619706118306, "learning_rate": 0.00019588521723043766, "loss": 0.8791, "step": 618 }, { "epoch": 0.11908426317814544, "grad_norm": 0.2365446704393795, "learning_rate": 0.0001958675088839415, "loss": 0.8489, "step": 619 }, { "epoch": 0.119276644863409, "grad_norm": 0.19024115995677332, "learning_rate": 0.00019584976331837758, "loss": 0.8325, "step": 620 }, { "epoch": 0.11946902654867257, "grad_norm": 0.1993353870667499, "learning_rate": 0.00019583198054063533, "loss": 0.8473, "step": 621 }, { "epoch": 0.11966140823393613, "grad_norm": 0.1950510206417347, "learning_rate": 0.00019581416055761865, "loss": 0.8369, "step": 622 }, { "epoch": 0.11985378991919969, "grad_norm": 0.09132905389495975, "learning_rate": 0.00019579630337624587, "loss": 0.6429, "step": 623 }, { "epoch": 0.12004617160446325, "grad_norm": 0.179460039622916, "learning_rate": 0.00019577840900344976, "loss": 0.9036, "step": 624 }, { "epoch": 0.12023855328972682, "grad_norm": 0.1912034450536512, "learning_rate": 0.00019576047744617752, "loss": 0.8743, "step": 625 }, { "epoch": 0.12043093497499038, "grad_norm": 0.09316849035002464, "learning_rate": 0.0001957425087113908, "loss": 0.6021, "step": 626 }, { "epoch": 0.12062331666025394, "grad_norm": 0.21752849101996047, "learning_rate": 0.0001957245028060657, "loss": 0.9156, "step": 627 }, { "epoch": 0.1208156983455175, "grad_norm": 0.18754957120840593, "learning_rate": 0.0001957064597371927, "loss": 0.8816, "step": 628 }, { "epoch": 0.12100808003078108, "grad_norm": 0.21013074565237244, "learning_rate": 0.00019568837951177676, "loss": 0.9037, "step": 629 }, { "epoch": 0.12120046171604464, "grad_norm": 0.19350595889930136, "learning_rate": 0.00019567026213683725, "loss": 0.9018, "step": 630 }, { "epoch": 0.1213928434013082, "grad_norm": 0.08350016520775827, "learning_rate": 0.00019565210761940798, "loss": 0.6179, "step": 631 }, { "epoch": 0.12158522508657175, "grad_norm": 0.20469267029184338, "learning_rate": 0.00019563391596653712, "loss": 0.8494, "step": 632 }, { "epoch": 0.12177760677183531, "grad_norm": 0.18784216868056577, "learning_rate": 0.0001956156871852873, "loss": 0.8492, "step": 633 }, { "epoch": 0.12196998845709889, "grad_norm": 0.09313146661555792, "learning_rate": 0.00019559742128273558, "loss": 0.6034, "step": 634 }, { "epoch": 0.12216237014236245, "grad_norm": 0.18127294345829484, "learning_rate": 0.00019557911826597338, "loss": 0.8498, "step": 635 }, { "epoch": 0.12235475182762601, "grad_norm": 0.19318911616760148, "learning_rate": 0.0001955607781421066, "loss": 0.844, "step": 636 }, { "epoch": 0.12254713351288957, "grad_norm": 0.1818225970390642, "learning_rate": 0.00019554240091825548, "loss": 0.9129, "step": 637 }, { "epoch": 0.12273951519815314, "grad_norm": 0.18410398351160917, "learning_rate": 0.0001955239866015547, "loss": 0.8721, "step": 638 }, { "epoch": 0.1229318968834167, "grad_norm": 0.20484668593881644, "learning_rate": 0.00019550553519915337, "loss": 0.9025, "step": 639 }, { "epoch": 0.12312427856868026, "grad_norm": 0.21918999329751482, "learning_rate": 0.00019548704671821493, "loss": 0.8739, "step": 640 }, { "epoch": 0.12331666025394382, "grad_norm": 0.1736874905793036, "learning_rate": 0.0001954685211659172, "loss": 0.9116, "step": 641 }, { "epoch": 0.12350904193920739, "grad_norm": 0.1964982843116399, "learning_rate": 0.00019544995854945248, "loss": 0.8412, "step": 642 }, { "epoch": 0.12370142362447095, "grad_norm": 0.20305644486411412, "learning_rate": 0.0001954313588760274, "loss": 0.83, "step": 643 }, { "epoch": 0.12389380530973451, "grad_norm": 0.18707825800692277, "learning_rate": 0.00019541272215286304, "loss": 0.8598, "step": 644 }, { "epoch": 0.12408618699499807, "grad_norm": 0.18239593197246673, "learning_rate": 0.00019539404838719477, "loss": 0.8323, "step": 645 }, { "epoch": 0.12427856868026164, "grad_norm": 0.1769471815116936, "learning_rate": 0.0001953753375862724, "loss": 0.8613, "step": 646 }, { "epoch": 0.1244709503655252, "grad_norm": 0.1989030301211667, "learning_rate": 0.0001953565897573601, "loss": 0.8254, "step": 647 }, { "epoch": 0.12466333205078876, "grad_norm": 0.19684873292328478, "learning_rate": 0.00019533780490773643, "loss": 0.8595, "step": 648 }, { "epoch": 0.12485571373605232, "grad_norm": 0.19981021544645033, "learning_rate": 0.00019531898304469433, "loss": 0.8403, "step": 649 }, { "epoch": 0.1250480954213159, "grad_norm": 0.09812081971027593, "learning_rate": 0.0001953001241755411, "loss": 0.6242, "step": 650 }, { "epoch": 0.12524047710657946, "grad_norm": 0.17343968682489194, "learning_rate": 0.0001952812283075984, "loss": 0.7904, "step": 651 }, { "epoch": 0.12543285879184302, "grad_norm": 0.21084574348341265, "learning_rate": 0.00019526229544820217, "loss": 0.8604, "step": 652 }, { "epoch": 0.12562524047710658, "grad_norm": 0.1720929295215633, "learning_rate": 0.00019524332560470292, "loss": 0.8442, "step": 653 }, { "epoch": 0.12581762216237014, "grad_norm": 0.18177499895472135, "learning_rate": 0.00019522431878446535, "loss": 0.8455, "step": 654 }, { "epoch": 0.1260100038476337, "grad_norm": 0.19740979636278116, "learning_rate": 0.00019520527499486856, "loss": 0.8303, "step": 655 }, { "epoch": 0.12620238553289725, "grad_norm": 0.18565314282734238, "learning_rate": 0.00019518619424330598, "loss": 0.8588, "step": 656 }, { "epoch": 0.12639476721816084, "grad_norm": 0.17294679254422318, "learning_rate": 0.00019516707653718544, "loss": 0.892, "step": 657 }, { "epoch": 0.1265871489034244, "grad_norm": 0.19835839726194143, "learning_rate": 0.00019514792188392912, "loss": 0.8617, "step": 658 }, { "epoch": 0.12677953058868796, "grad_norm": 0.16948841409941368, "learning_rate": 0.00019512873029097345, "loss": 0.8132, "step": 659 }, { "epoch": 0.12697191227395152, "grad_norm": 0.1704489653691407, "learning_rate": 0.00019510950176576932, "loss": 0.855, "step": 660 }, { "epoch": 0.12716429395921508, "grad_norm": 0.19125036539328166, "learning_rate": 0.0001950902363157819, "loss": 0.8728, "step": 661 }, { "epoch": 0.12735667564447864, "grad_norm": 0.2030323874540624, "learning_rate": 0.00019507093394849066, "loss": 0.834, "step": 662 }, { "epoch": 0.1275490573297422, "grad_norm": 0.19273601980736568, "learning_rate": 0.0001950515946713895, "loss": 0.8592, "step": 663 }, { "epoch": 0.12774143901500576, "grad_norm": 0.09901206747966566, "learning_rate": 0.00019503221849198656, "loss": 0.6048, "step": 664 }, { "epoch": 0.12793382070026935, "grad_norm": 0.20633863078187822, "learning_rate": 0.00019501280541780433, "loss": 0.8628, "step": 665 }, { "epoch": 0.1281262023855329, "grad_norm": 0.20113528727533073, "learning_rate": 0.00019499335545637966, "loss": 0.9218, "step": 666 }, { "epoch": 0.12831858407079647, "grad_norm": 0.08885380902196117, "learning_rate": 0.0001949738686152637, "loss": 0.6189, "step": 667 }, { "epoch": 0.12851096575606002, "grad_norm": 0.202567738923743, "learning_rate": 0.00019495434490202188, "loss": 0.9095, "step": 668 }, { "epoch": 0.12870334744132358, "grad_norm": 0.18515283703844146, "learning_rate": 0.000194934784324234, "loss": 0.8108, "step": 669 }, { "epoch": 0.12889572912658714, "grad_norm": 0.16969298074128933, "learning_rate": 0.00019491518688949416, "loss": 0.8743, "step": 670 }, { "epoch": 0.1290881108118507, "grad_norm": 0.20686176896030273, "learning_rate": 0.00019489555260541073, "loss": 0.868, "step": 671 }, { "epoch": 0.12928049249711426, "grad_norm": 0.21045003591159003, "learning_rate": 0.0001948758814796064, "loss": 0.9148, "step": 672 }, { "epoch": 0.12947287418237785, "grad_norm": 0.1724941745477603, "learning_rate": 0.00019485617351971827, "loss": 0.7945, "step": 673 }, { "epoch": 0.1296652558676414, "grad_norm": 0.19002193031001857, "learning_rate": 0.00019483642873339751, "loss": 0.8612, "step": 674 }, { "epoch": 0.12985763755290497, "grad_norm": 0.20999064667317818, "learning_rate": 0.00019481664712830984, "loss": 0.8848, "step": 675 }, { "epoch": 0.13005001923816853, "grad_norm": 0.21494791718497042, "learning_rate": 0.00019479682871213514, "loss": 0.869, "step": 676 }, { "epoch": 0.1302424009234321, "grad_norm": 0.17583452804853758, "learning_rate": 0.00019477697349256756, "loss": 0.8671, "step": 677 }, { "epoch": 0.13043478260869565, "grad_norm": 0.19159198763918278, "learning_rate": 0.0001947570814773156, "loss": 0.8534, "step": 678 }, { "epoch": 0.1306271642939592, "grad_norm": 0.17277265428475921, "learning_rate": 0.00019473715267410207, "loss": 0.8443, "step": 679 }, { "epoch": 0.13081954597922277, "grad_norm": 0.16937521409010697, "learning_rate": 0.00019471718709066394, "loss": 0.9029, "step": 680 }, { "epoch": 0.13101192766448633, "grad_norm": 0.2884716387660774, "learning_rate": 0.00019469718473475256, "loss": 0.889, "step": 681 }, { "epoch": 0.13120430934974991, "grad_norm": 0.18991786454975165, "learning_rate": 0.00019467714561413358, "loss": 0.899, "step": 682 }, { "epoch": 0.13139669103501347, "grad_norm": 0.18348755266189998, "learning_rate": 0.00019465706973658683, "loss": 0.8017, "step": 683 }, { "epoch": 0.13158907272027703, "grad_norm": 0.10294053330487155, "learning_rate": 0.00019463695710990645, "loss": 0.6255, "step": 684 }, { "epoch": 0.1317814544055406, "grad_norm": 0.18227639738178916, "learning_rate": 0.00019461680774190092, "loss": 0.8245, "step": 685 }, { "epoch": 0.13197383609080415, "grad_norm": 0.19688028887209932, "learning_rate": 0.00019459662164039284, "loss": 0.8777, "step": 686 }, { "epoch": 0.1321662177760677, "grad_norm": 0.1751368666967072, "learning_rate": 0.00019457639881321917, "loss": 0.8476, "step": 687 }, { "epoch": 0.13235859946133127, "grad_norm": 0.19962943172583897, "learning_rate": 0.00019455613926823114, "loss": 0.8164, "step": 688 }, { "epoch": 0.13255098114659483, "grad_norm": 0.19835743851523086, "learning_rate": 0.0001945358430132942, "loss": 0.8401, "step": 689 }, { "epoch": 0.13274336283185842, "grad_norm": 0.21701062797553375, "learning_rate": 0.000194515510056288, "loss": 0.9544, "step": 690 }, { "epoch": 0.13293574451712198, "grad_norm": 0.19281503443429837, "learning_rate": 0.00019449514040510655, "loss": 0.8253, "step": 691 }, { "epoch": 0.13312812620238554, "grad_norm": 0.20318297966212653, "learning_rate": 0.00019447473406765804, "loss": 0.8447, "step": 692 }, { "epoch": 0.1333205078876491, "grad_norm": 0.18196752417922507, "learning_rate": 0.00019445429105186487, "loss": 0.8038, "step": 693 }, { "epoch": 0.13351288957291266, "grad_norm": 0.20008684958424672, "learning_rate": 0.0001944338113656638, "loss": 0.911, "step": 694 }, { "epoch": 0.13370527125817622, "grad_norm": 0.1899245358723444, "learning_rate": 0.00019441329501700567, "loss": 0.8454, "step": 695 }, { "epoch": 0.13389765294343978, "grad_norm": 0.20873927714848506, "learning_rate": 0.00019439274201385568, "loss": 0.9006, "step": 696 }, { "epoch": 0.13409003462870334, "grad_norm": 0.185910877593038, "learning_rate": 0.0001943721523641932, "loss": 0.7996, "step": 697 }, { "epoch": 0.13428241631396692, "grad_norm": 0.17856862092359593, "learning_rate": 0.00019435152607601188, "loss": 0.8563, "step": 698 }, { "epoch": 0.13447479799923048, "grad_norm": 0.23258510333005325, "learning_rate": 0.0001943308631573195, "loss": 0.8365, "step": 699 }, { "epoch": 0.13466717968449404, "grad_norm": 0.18806989268536997, "learning_rate": 0.00019431016361613814, "loss": 0.8145, "step": 700 }, { "epoch": 0.1348595613697576, "grad_norm": 0.17821803465809657, "learning_rate": 0.00019428942746050406, "loss": 0.8764, "step": 701 }, { "epoch": 0.13505194305502116, "grad_norm": 0.19008549325556579, "learning_rate": 0.00019426865469846774, "loss": 0.8572, "step": 702 }, { "epoch": 0.13524432474028472, "grad_norm": 0.19426789971701644, "learning_rate": 0.00019424784533809392, "loss": 0.8461, "step": 703 }, { "epoch": 0.13543670642554828, "grad_norm": 0.19510328648749464, "learning_rate": 0.00019422699938746152, "loss": 0.8294, "step": 704 }, { "epoch": 0.13562908811081184, "grad_norm": 0.20128687300949813, "learning_rate": 0.00019420611685466358, "loss": 0.7799, "step": 705 }, { "epoch": 0.1358214697960754, "grad_norm": 0.20525111774307744, "learning_rate": 0.00019418519774780748, "loss": 0.8703, "step": 706 }, { "epoch": 0.136013851481339, "grad_norm": 0.21170100817205287, "learning_rate": 0.00019416424207501474, "loss": 0.88, "step": 707 }, { "epoch": 0.13620623316660255, "grad_norm": 0.1964986416229853, "learning_rate": 0.00019414324984442103, "loss": 0.845, "step": 708 }, { "epoch": 0.1363986148518661, "grad_norm": 0.21971927262099394, "learning_rate": 0.0001941222210641763, "loss": 0.8632, "step": 709 }, { "epoch": 0.13659099653712967, "grad_norm": 0.19285920913222568, "learning_rate": 0.00019410115574244462, "loss": 0.8343, "step": 710 }, { "epoch": 0.13678337822239323, "grad_norm": 0.19009033257979552, "learning_rate": 0.00019408005388740432, "loss": 0.8789, "step": 711 }, { "epoch": 0.13697575990765679, "grad_norm": 0.18827082517416271, "learning_rate": 0.00019405891550724778, "loss": 0.8378, "step": 712 }, { "epoch": 0.13716814159292035, "grad_norm": 0.20019737076386182, "learning_rate": 0.00019403774061018173, "loss": 0.8447, "step": 713 }, { "epoch": 0.1373605232781839, "grad_norm": 0.19972068841254492, "learning_rate": 0.00019401652920442692, "loss": 0.8879, "step": 714 }, { "epoch": 0.1375529049634475, "grad_norm": 0.20621100102503395, "learning_rate": 0.00019399528129821842, "loss": 0.9001, "step": 715 }, { "epoch": 0.13774528664871105, "grad_norm": 0.19332298237375034, "learning_rate": 0.0001939739968998054, "loss": 0.8376, "step": 716 }, { "epoch": 0.1379376683339746, "grad_norm": 0.16563469198898, "learning_rate": 0.0001939526760174511, "loss": 0.8539, "step": 717 }, { "epoch": 0.13813005001923817, "grad_norm": 0.19860471317565717, "learning_rate": 0.0001939313186594331, "loss": 0.8814, "step": 718 }, { "epoch": 0.13832243170450173, "grad_norm": 0.18520788383479594, "learning_rate": 0.00019390992483404306, "loss": 0.8731, "step": 719 }, { "epoch": 0.1385148133897653, "grad_norm": 0.1985988281625428, "learning_rate": 0.0001938884945495868, "loss": 0.8398, "step": 720 }, { "epoch": 0.13870719507502885, "grad_norm": 0.20354762768647075, "learning_rate": 0.00019386702781438425, "loss": 0.8689, "step": 721 }, { "epoch": 0.1388995767602924, "grad_norm": 0.17491052022914474, "learning_rate": 0.0001938455246367696, "loss": 0.8856, "step": 722 }, { "epoch": 0.139091958445556, "grad_norm": 0.18509927851133265, "learning_rate": 0.00019382398502509107, "loss": 0.8817, "step": 723 }, { "epoch": 0.13928434013081956, "grad_norm": 0.1831571578373972, "learning_rate": 0.0001938024089877111, "loss": 0.8392, "step": 724 }, { "epoch": 0.13947672181608312, "grad_norm": 0.18418589573023522, "learning_rate": 0.00019378079653300625, "loss": 0.8902, "step": 725 }, { "epoch": 0.13966910350134668, "grad_norm": 0.1200796298154663, "learning_rate": 0.0001937591476693672, "loss": 0.6338, "step": 726 }, { "epoch": 0.13986148518661023, "grad_norm": 0.22245543503598164, "learning_rate": 0.00019373746240519883, "loss": 0.861, "step": 727 }, { "epoch": 0.1400538668718738, "grad_norm": 0.17639585596213142, "learning_rate": 0.00019371574074892006, "loss": 0.8508, "step": 728 }, { "epoch": 0.14024624855713735, "grad_norm": 0.2107245117171115, "learning_rate": 0.00019369398270896403, "loss": 0.8453, "step": 729 }, { "epoch": 0.1404386302424009, "grad_norm": 0.17507224570977795, "learning_rate": 0.0001936721882937779, "loss": 0.8989, "step": 730 }, { "epoch": 0.14063101192766447, "grad_norm": 0.19284665221819489, "learning_rate": 0.0001936503575118231, "loss": 0.8424, "step": 731 }, { "epoch": 0.14082339361292806, "grad_norm": 0.19400251388860948, "learning_rate": 0.00019362849037157497, "loss": 0.843, "step": 732 }, { "epoch": 0.14101577529819162, "grad_norm": 0.17896096537902945, "learning_rate": 0.0001936065868815232, "loss": 0.8613, "step": 733 }, { "epoch": 0.14120815698345518, "grad_norm": 0.20249250139972677, "learning_rate": 0.00019358464705017143, "loss": 0.7978, "step": 734 }, { "epoch": 0.14140053866871874, "grad_norm": 0.19045480708375653, "learning_rate": 0.00019356267088603744, "loss": 0.7795, "step": 735 }, { "epoch": 0.1415929203539823, "grad_norm": 0.15971982656810316, "learning_rate": 0.00019354065839765315, "loss": 0.8959, "step": 736 }, { "epoch": 0.14178530203924586, "grad_norm": 0.24106699833981043, "learning_rate": 0.0001935186095935646, "loss": 0.8774, "step": 737 }, { "epoch": 0.14197768372450942, "grad_norm": 0.19406025616747116, "learning_rate": 0.00019349652448233187, "loss": 0.844, "step": 738 }, { "epoch": 0.14217006540977298, "grad_norm": 0.1824177758834242, "learning_rate": 0.00019347440307252911, "loss": 0.843, "step": 739 }, { "epoch": 0.14236244709503657, "grad_norm": 0.19358201492740218, "learning_rate": 0.0001934522453727447, "loss": 0.8419, "step": 740 }, { "epoch": 0.14255482878030012, "grad_norm": 0.252529488917006, "learning_rate": 0.00019343005139158096, "loss": 0.8334, "step": 741 }, { "epoch": 0.14274721046556368, "grad_norm": 0.1880401105122422, "learning_rate": 0.00019340782113765439, "loss": 0.8739, "step": 742 }, { "epoch": 0.14293959215082724, "grad_norm": 0.1862385440487931, "learning_rate": 0.00019338555461959552, "loss": 0.8651, "step": 743 }, { "epoch": 0.1431319738360908, "grad_norm": 0.18635622887656617, "learning_rate": 0.000193363251846049, "loss": 0.8219, "step": 744 }, { "epoch": 0.14332435552135436, "grad_norm": 0.1949683359910251, "learning_rate": 0.00019334091282567352, "loss": 0.8403, "step": 745 }, { "epoch": 0.14351673720661792, "grad_norm": 0.22340020679001282, "learning_rate": 0.00019331853756714184, "loss": 0.9066, "step": 746 }, { "epoch": 0.14370911889188148, "grad_norm": 0.19889423217226582, "learning_rate": 0.00019329612607914086, "loss": 0.8888, "step": 747 }, { "epoch": 0.14390150057714507, "grad_norm": 0.2041052279653166, "learning_rate": 0.00019327367837037142, "loss": 0.8605, "step": 748 }, { "epoch": 0.14409388226240863, "grad_norm": 0.19406268473144186, "learning_rate": 0.00019325119444954855, "loss": 0.827, "step": 749 }, { "epoch": 0.1442862639476722, "grad_norm": 0.09501263231953187, "learning_rate": 0.00019322867432540125, "loss": 0.6203, "step": 750 }, { "epoch": 0.14447864563293575, "grad_norm": 0.18678032703883835, "learning_rate": 0.00019320611800667265, "loss": 0.861, "step": 751 }, { "epoch": 0.1446710273181993, "grad_norm": 0.2362633366473677, "learning_rate": 0.00019318352550211986, "loss": 0.8023, "step": 752 }, { "epoch": 0.14486340900346287, "grad_norm": 0.1907953453015115, "learning_rate": 0.00019316089682051404, "loss": 0.8342, "step": 753 }, { "epoch": 0.14505579068872643, "grad_norm": 0.20218413979820718, "learning_rate": 0.00019313823197064043, "loss": 0.7934, "step": 754 }, { "epoch": 0.14524817237399, "grad_norm": 0.18633714791123296, "learning_rate": 0.00019311553096129836, "loss": 0.8813, "step": 755 }, { "epoch": 0.14544055405925355, "grad_norm": 0.18814477831293194, "learning_rate": 0.0001930927938013011, "loss": 0.8622, "step": 756 }, { "epoch": 0.14563293574451713, "grad_norm": 0.20125649724850708, "learning_rate": 0.000193070020499476, "loss": 0.8588, "step": 757 }, { "epoch": 0.1458253174297807, "grad_norm": 0.2134794408808386, "learning_rate": 0.00019304721106466448, "loss": 0.8061, "step": 758 }, { "epoch": 0.14601769911504425, "grad_norm": 0.09500218282155276, "learning_rate": 0.00019302436550572188, "loss": 0.5925, "step": 759 }, { "epoch": 0.1462100808003078, "grad_norm": 0.1807580248718071, "learning_rate": 0.00019300148383151767, "loss": 0.7964, "step": 760 }, { "epoch": 0.14640246248557137, "grad_norm": 0.19286167350660913, "learning_rate": 0.00019297856605093533, "loss": 0.8426, "step": 761 }, { "epoch": 0.14659484417083493, "grad_norm": 0.19019602134904146, "learning_rate": 0.00019295561217287226, "loss": 0.834, "step": 762 }, { "epoch": 0.1467872258560985, "grad_norm": 0.20269322328130907, "learning_rate": 0.00019293262220624002, "loss": 0.8567, "step": 763 }, { "epoch": 0.14697960754136205, "grad_norm": 0.17802512295764997, "learning_rate": 0.00019290959615996407, "loss": 0.8616, "step": 764 }, { "epoch": 0.14717198922662564, "grad_norm": 0.1944864069864121, "learning_rate": 0.0001928865340429839, "loss": 0.8958, "step": 765 }, { "epoch": 0.1473643709118892, "grad_norm": 0.20326513490507495, "learning_rate": 0.00019286343586425307, "loss": 0.8379, "step": 766 }, { "epoch": 0.14755675259715276, "grad_norm": 0.21816672145120433, "learning_rate": 0.00019284030163273905, "loss": 0.8821, "step": 767 }, { "epoch": 0.14774913428241632, "grad_norm": 0.20116500167914284, "learning_rate": 0.00019281713135742333, "loss": 0.8432, "step": 768 }, { "epoch": 0.14794151596767988, "grad_norm": 0.1951747273744955, "learning_rate": 0.00019279392504730145, "loss": 0.8665, "step": 769 }, { "epoch": 0.14813389765294344, "grad_norm": 0.18446561129057154, "learning_rate": 0.00019277068271138288, "loss": 0.8019, "step": 770 }, { "epoch": 0.148326279338207, "grad_norm": 0.18879841596189806, "learning_rate": 0.00019274740435869106, "loss": 0.8805, "step": 771 }, { "epoch": 0.14851866102347056, "grad_norm": 0.1772748063505939, "learning_rate": 0.0001927240899982635, "loss": 0.8089, "step": 772 }, { "epoch": 0.14871104270873411, "grad_norm": 0.16973990658209812, "learning_rate": 0.00019270073963915162, "loss": 0.8634, "step": 773 }, { "epoch": 0.1489034243939977, "grad_norm": 0.18552040962308114, "learning_rate": 0.00019267735329042086, "loss": 0.8881, "step": 774 }, { "epoch": 0.14909580607926126, "grad_norm": 0.2110611091382782, "learning_rate": 0.00019265393096115055, "loss": 0.8448, "step": 775 }, { "epoch": 0.14928818776452482, "grad_norm": 0.20392012524745246, "learning_rate": 0.00019263047266043407, "loss": 0.8318, "step": 776 }, { "epoch": 0.14948056944978838, "grad_norm": 0.19410667918845298, "learning_rate": 0.00019260697839737875, "loss": 0.8624, "step": 777 }, { "epoch": 0.14967295113505194, "grad_norm": 0.22161132440505016, "learning_rate": 0.00019258344818110588, "loss": 0.7922, "step": 778 }, { "epoch": 0.1498653328203155, "grad_norm": 0.1767567265929312, "learning_rate": 0.00019255988202075065, "loss": 0.8517, "step": 779 }, { "epoch": 0.15005771450557906, "grad_norm": 0.24269093915558948, "learning_rate": 0.0001925362799254623, "loss": 0.8713, "step": 780 }, { "epoch": 0.15025009619084262, "grad_norm": 0.19013778991790758, "learning_rate": 0.00019251264190440397, "loss": 0.8675, "step": 781 }, { "epoch": 0.1504424778761062, "grad_norm": 0.2009813726010405, "learning_rate": 0.00019248896796675275, "loss": 0.8534, "step": 782 }, { "epoch": 0.15063485956136977, "grad_norm": 0.18082515022030993, "learning_rate": 0.0001924652581216997, "loss": 0.8629, "step": 783 }, { "epoch": 0.15082724124663333, "grad_norm": 0.18267301861079085, "learning_rate": 0.00019244151237844974, "loss": 0.8781, "step": 784 }, { "epoch": 0.15101962293189689, "grad_norm": 0.18897658978174228, "learning_rate": 0.00019241773074622182, "loss": 0.852, "step": 785 }, { "epoch": 0.15121200461716044, "grad_norm": 0.198156151670771, "learning_rate": 0.00019239391323424881, "loss": 0.8441, "step": 786 }, { "epoch": 0.151404386302424, "grad_norm": 0.19523621596918914, "learning_rate": 0.00019237005985177748, "loss": 0.8941, "step": 787 }, { "epoch": 0.15159676798768756, "grad_norm": 0.21446764092972764, "learning_rate": 0.00019234617060806848, "loss": 0.8526, "step": 788 }, { "epoch": 0.15178914967295112, "grad_norm": 0.1805019789534979, "learning_rate": 0.0001923222455123965, "loss": 0.8524, "step": 789 }, { "epoch": 0.1519815313582147, "grad_norm": 0.1657178804370895, "learning_rate": 0.00019229828457405007, "loss": 0.8153, "step": 790 }, { "epoch": 0.15217391304347827, "grad_norm": 0.1902614733085851, "learning_rate": 0.00019227428780233165, "loss": 0.8772, "step": 791 }, { "epoch": 0.15236629472874183, "grad_norm": 0.1906052327508976, "learning_rate": 0.00019225025520655758, "loss": 0.8357, "step": 792 }, { "epoch": 0.1525586764140054, "grad_norm": 0.20403086477787843, "learning_rate": 0.0001922261867960582, "loss": 0.825, "step": 793 }, { "epoch": 0.15275105809926895, "grad_norm": 0.1819919169447467, "learning_rate": 0.00019220208258017763, "loss": 0.8293, "step": 794 }, { "epoch": 0.1529434397845325, "grad_norm": 0.20914429622776604, "learning_rate": 0.00019217794256827403, "loss": 0.8581, "step": 795 }, { "epoch": 0.15313582146979607, "grad_norm": 0.17643198649142083, "learning_rate": 0.0001921537667697193, "loss": 0.8441, "step": 796 }, { "epoch": 0.15332820315505963, "grad_norm": 0.18193616001749152, "learning_rate": 0.0001921295551938994, "loss": 0.8771, "step": 797 }, { "epoch": 0.1535205848403232, "grad_norm": 0.20868441474372254, "learning_rate": 0.00019210530785021405, "loss": 0.8632, "step": 798 }, { "epoch": 0.15371296652558677, "grad_norm": 0.18968863441344883, "learning_rate": 0.0001920810247480769, "loss": 0.8871, "step": 799 }, { "epoch": 0.15390534821085033, "grad_norm": 0.19550750456596996, "learning_rate": 0.0001920567058969155, "loss": 0.8343, "step": 800 }, { "epoch": 0.1540977298961139, "grad_norm": 0.1980977703815738, "learning_rate": 0.0001920323513061713, "loss": 0.8912, "step": 801 }, { "epoch": 0.15429011158137745, "grad_norm": 0.18886997351279441, "learning_rate": 0.00019200796098529956, "loss": 0.8464, "step": 802 }, { "epoch": 0.154482493266641, "grad_norm": 0.20138723461471894, "learning_rate": 0.00019198353494376938, "loss": 0.8919, "step": 803 }, { "epoch": 0.15467487495190457, "grad_norm": 0.18638275656760447, "learning_rate": 0.00019195907319106392, "loss": 0.8065, "step": 804 }, { "epoch": 0.15486725663716813, "grad_norm": 0.20700282658547856, "learning_rate": 0.00019193457573667995, "loss": 0.883, "step": 805 }, { "epoch": 0.1550596383224317, "grad_norm": 0.1943118744116374, "learning_rate": 0.0001919100425901283, "loss": 0.8816, "step": 806 }, { "epoch": 0.15525202000769528, "grad_norm": 0.19264587974702024, "learning_rate": 0.00019188547376093354, "loss": 0.8042, "step": 807 }, { "epoch": 0.15544440169295884, "grad_norm": 0.1943972983292274, "learning_rate": 0.0001918608692586342, "loss": 0.8373, "step": 808 }, { "epoch": 0.1556367833782224, "grad_norm": 0.2030718733431976, "learning_rate": 0.0001918362290927825, "loss": 0.8482, "step": 809 }, { "epoch": 0.15582916506348596, "grad_norm": 0.18783146901966682, "learning_rate": 0.0001918115532729447, "loss": 0.8616, "step": 810 }, { "epoch": 0.15602154674874952, "grad_norm": 0.18277640401352632, "learning_rate": 0.00019178684180870073, "loss": 0.8378, "step": 811 }, { "epoch": 0.15621392843401308, "grad_norm": 0.19416190900912275, "learning_rate": 0.00019176209470964445, "loss": 0.8421, "step": 812 }, { "epoch": 0.15640631011927664, "grad_norm": 0.20709093124615793, "learning_rate": 0.00019173731198538353, "loss": 0.8221, "step": 813 }, { "epoch": 0.1565986918045402, "grad_norm": 0.19275857807573435, "learning_rate": 0.00019171249364553954, "loss": 0.8556, "step": 814 }, { "epoch": 0.15679107348980378, "grad_norm": 0.20625419983629692, "learning_rate": 0.00019168763969974772, "loss": 0.8462, "step": 815 }, { "epoch": 0.15698345517506734, "grad_norm": 0.1765647756817115, "learning_rate": 0.0001916627501576573, "loss": 0.8811, "step": 816 }, { "epoch": 0.1571758368603309, "grad_norm": 0.19216153515358553, "learning_rate": 0.0001916378250289312, "loss": 0.7842, "step": 817 }, { "epoch": 0.15736821854559446, "grad_norm": 0.1985050366823736, "learning_rate": 0.00019161286432324624, "loss": 0.8586, "step": 818 }, { "epoch": 0.15756060023085802, "grad_norm": 0.21017588564133535, "learning_rate": 0.00019158786805029306, "loss": 0.7864, "step": 819 }, { "epoch": 0.15775298191612158, "grad_norm": 0.20454693805869534, "learning_rate": 0.00019156283621977603, "loss": 0.8789, "step": 820 }, { "epoch": 0.15794536360138514, "grad_norm": 0.19853107718316684, "learning_rate": 0.00019153776884141336, "loss": 0.887, "step": 821 }, { "epoch": 0.1581377452866487, "grad_norm": 0.2086910790100048, "learning_rate": 0.00019151266592493712, "loss": 0.8983, "step": 822 }, { "epoch": 0.15833012697191226, "grad_norm": 0.19278241263287835, "learning_rate": 0.00019148752748009306, "loss": 0.8127, "step": 823 }, { "epoch": 0.15852250865717585, "grad_norm": 0.19575686331816736, "learning_rate": 0.0001914623535166408, "loss": 0.7869, "step": 824 }, { "epoch": 0.1587148903424394, "grad_norm": 0.09936878619895137, "learning_rate": 0.00019143714404435381, "loss": 0.673, "step": 825 }, { "epoch": 0.15890727202770297, "grad_norm": 0.23079114300602538, "learning_rate": 0.00019141189907301921, "loss": 0.8828, "step": 826 }, { "epoch": 0.15909965371296653, "grad_norm": 0.1991930983721062, "learning_rate": 0.000191386618612438, "loss": 0.7902, "step": 827 }, { "epoch": 0.1592920353982301, "grad_norm": 0.20013153176390616, "learning_rate": 0.00019136130267242488, "loss": 0.8658, "step": 828 }, { "epoch": 0.15948441708349365, "grad_norm": 0.1985886493515215, "learning_rate": 0.00019133595126280846, "loss": 0.8328, "step": 829 }, { "epoch": 0.1596767987687572, "grad_norm": 0.18490754326709055, "learning_rate": 0.00019131056439343093, "loss": 0.8639, "step": 830 }, { "epoch": 0.15986918045402077, "grad_norm": 0.17256565015400802, "learning_rate": 0.0001912851420741484, "loss": 0.8151, "step": 831 }, { "epoch": 0.16006156213928435, "grad_norm": 0.20163883859705192, "learning_rate": 0.00019125968431483065, "loss": 0.8763, "step": 832 }, { "epoch": 0.1602539438245479, "grad_norm": 0.20349979902519247, "learning_rate": 0.0001912341911253613, "loss": 0.8213, "step": 833 }, { "epoch": 0.16044632550981147, "grad_norm": 0.1858441377438849, "learning_rate": 0.0001912086625156377, "loss": 0.9132, "step": 834 }, { "epoch": 0.16063870719507503, "grad_norm": 0.2014300074826958, "learning_rate": 0.00019118309849557087, "loss": 0.9, "step": 835 }, { "epoch": 0.1608310888803386, "grad_norm": 0.20690426919236365, "learning_rate": 0.0001911574990750857, "loss": 0.8404, "step": 836 }, { "epoch": 0.16102347056560215, "grad_norm": 0.20125848580926814, "learning_rate": 0.00019113186426412073, "loss": 0.8164, "step": 837 }, { "epoch": 0.1612158522508657, "grad_norm": 0.19769170274619763, "learning_rate": 0.0001911061940726283, "loss": 0.8502, "step": 838 }, { "epoch": 0.16140823393612927, "grad_norm": 0.20239546249808762, "learning_rate": 0.00019108048851057446, "loss": 0.8493, "step": 839 }, { "epoch": 0.16160061562139286, "grad_norm": 0.18860897665429288, "learning_rate": 0.00019105474758793895, "loss": 0.8652, "step": 840 }, { "epoch": 0.16179299730665642, "grad_norm": 0.19745268619695608, "learning_rate": 0.00019102897131471536, "loss": 0.8603, "step": 841 }, { "epoch": 0.16198537899191998, "grad_norm": 0.18949840175158086, "learning_rate": 0.00019100315970091086, "loss": 0.8253, "step": 842 }, { "epoch": 0.16217776067718354, "grad_norm": 0.18203444493625526, "learning_rate": 0.00019097731275654643, "loss": 0.8595, "step": 843 }, { "epoch": 0.1623701423624471, "grad_norm": 0.2109163479025291, "learning_rate": 0.00019095143049165678, "loss": 0.8669, "step": 844 }, { "epoch": 0.16256252404771065, "grad_norm": 0.20560215707928367, "learning_rate": 0.00019092551291629027, "loss": 0.887, "step": 845 }, { "epoch": 0.16275490573297421, "grad_norm": 0.21837409627985654, "learning_rate": 0.00019089956004050894, "loss": 0.8669, "step": 846 }, { "epoch": 0.16294728741823777, "grad_norm": 0.17916713579029278, "learning_rate": 0.00019087357187438868, "loss": 0.8382, "step": 847 }, { "epoch": 0.16313966910350133, "grad_norm": 0.16358269171874906, "learning_rate": 0.0001908475484280189, "loss": 0.8463, "step": 848 }, { "epoch": 0.16333205078876492, "grad_norm": 0.19168737458704568, "learning_rate": 0.0001908214897115029, "loss": 0.8209, "step": 849 }, { "epoch": 0.16352443247402848, "grad_norm": 0.18345570150997023, "learning_rate": 0.00019079539573495748, "loss": 0.8984, "step": 850 }, { "epoch": 0.16371681415929204, "grad_norm": 0.17577631511592998, "learning_rate": 0.0001907692665085133, "loss": 0.7906, "step": 851 }, { "epoch": 0.1639091958445556, "grad_norm": 0.18922608812394653, "learning_rate": 0.00019074310204231457, "loss": 0.8403, "step": 852 }, { "epoch": 0.16410157752981916, "grad_norm": 0.1850155030751691, "learning_rate": 0.0001907169023465192, "loss": 0.8205, "step": 853 }, { "epoch": 0.16429395921508272, "grad_norm": 0.16359885875023145, "learning_rate": 0.00019069066743129893, "loss": 0.9105, "step": 854 }, { "epoch": 0.16448634090034628, "grad_norm": 0.20475938146675354, "learning_rate": 0.000190664397306839, "loss": 0.8643, "step": 855 }, { "epoch": 0.16467872258560984, "grad_norm": 0.19281773349589107, "learning_rate": 0.00019063809198333832, "loss": 0.8385, "step": 856 }, { "epoch": 0.16487110427087343, "grad_norm": 0.20013354100115036, "learning_rate": 0.0001906117514710096, "loss": 0.8212, "step": 857 }, { "epoch": 0.16506348595613698, "grad_norm": 0.10287550371360259, "learning_rate": 0.00019058537578007907, "loss": 0.6253, "step": 858 }, { "epoch": 0.16525586764140054, "grad_norm": 0.2060644230621728, "learning_rate": 0.00019055896492078676, "loss": 0.8853, "step": 859 }, { "epoch": 0.1654482493266641, "grad_norm": 0.20351920377074276, "learning_rate": 0.0001905325189033862, "loss": 0.85, "step": 860 }, { "epoch": 0.16564063101192766, "grad_norm": 0.1875628190974467, "learning_rate": 0.00019050603773814468, "loss": 0.8642, "step": 861 }, { "epoch": 0.16583301269719122, "grad_norm": 0.1881532053420717, "learning_rate": 0.0001904795214353431, "loss": 0.8619, "step": 862 }, { "epoch": 0.16602539438245478, "grad_norm": 0.19707555463186377, "learning_rate": 0.000190452970005276, "loss": 0.8567, "step": 863 }, { "epoch": 0.16621777606771834, "grad_norm": 0.18812668640944, "learning_rate": 0.00019042638345825156, "loss": 0.8744, "step": 864 }, { "epoch": 0.16641015775298193, "grad_norm": 0.21001102657933435, "learning_rate": 0.00019039976180459158, "loss": 0.9186, "step": 865 }, { "epoch": 0.1666025394382455, "grad_norm": 0.19237802680378058, "learning_rate": 0.00019037310505463153, "loss": 0.8538, "step": 866 }, { "epoch": 0.16679492112350905, "grad_norm": 0.17619316984268374, "learning_rate": 0.00019034641321872045, "loss": 0.8727, "step": 867 }, { "epoch": 0.1669873028087726, "grad_norm": 0.19731993916462276, "learning_rate": 0.00019031968630722103, "loss": 0.8092, "step": 868 }, { "epoch": 0.16717968449403617, "grad_norm": 0.1961218617894081, "learning_rate": 0.0001902929243305096, "loss": 0.8352, "step": 869 }, { "epoch": 0.16737206617929973, "grad_norm": 0.18185280626300698, "learning_rate": 0.00019026612729897607, "loss": 0.7669, "step": 870 }, { "epoch": 0.1675644478645633, "grad_norm": 0.26879419382376474, "learning_rate": 0.00019023929522302396, "loss": 0.8683, "step": 871 }, { "epoch": 0.16775682954982685, "grad_norm": 0.18963517434487326, "learning_rate": 0.00019021242811307043, "loss": 0.8832, "step": 872 }, { "epoch": 0.1679492112350904, "grad_norm": 0.1961489977893349, "learning_rate": 0.0001901855259795462, "loss": 0.8236, "step": 873 }, { "epoch": 0.168141592920354, "grad_norm": 0.1821108537266276, "learning_rate": 0.00019015858883289556, "loss": 0.8044, "step": 874 }, { "epoch": 0.16833397460561755, "grad_norm": 0.19169805147105376, "learning_rate": 0.00019013161668357653, "loss": 0.846, "step": 875 }, { "epoch": 0.1685263562908811, "grad_norm": 0.19190084157911047, "learning_rate": 0.00019010460954206056, "loss": 0.8913, "step": 876 }, { "epoch": 0.16871873797614467, "grad_norm": 0.20646246953128874, "learning_rate": 0.00019007756741883282, "loss": 0.8018, "step": 877 }, { "epoch": 0.16891111966140823, "grad_norm": 0.20981141001806994, "learning_rate": 0.0001900504903243919, "loss": 0.8638, "step": 878 }, { "epoch": 0.1691035013466718, "grad_norm": 0.18842391865249086, "learning_rate": 0.00019002337826925012, "loss": 0.8723, "step": 879 }, { "epoch": 0.16929588303193535, "grad_norm": 0.10417101478033719, "learning_rate": 0.0001899962312639333, "loss": 0.6529, "step": 880 }, { "epoch": 0.1694882647171989, "grad_norm": 0.19037838364379017, "learning_rate": 0.00018996904931898082, "loss": 0.8267, "step": 881 }, { "epoch": 0.1696806464024625, "grad_norm": 0.1815735418357112, "learning_rate": 0.0001899418324449457, "loss": 0.8716, "step": 882 }, { "epoch": 0.16987302808772606, "grad_norm": 0.20593425382366073, "learning_rate": 0.00018991458065239445, "loss": 0.9127, "step": 883 }, { "epoch": 0.17006540977298962, "grad_norm": 0.2036942220870153, "learning_rate": 0.0001898872939519071, "loss": 0.8189, "step": 884 }, { "epoch": 0.17025779145825318, "grad_norm": 0.19672661502123862, "learning_rate": 0.00018985997235407734, "loss": 0.866, "step": 885 }, { "epoch": 0.17045017314351674, "grad_norm": 0.2158496836401602, "learning_rate": 0.00018983261586951237, "loss": 0.8559, "step": 886 }, { "epoch": 0.1706425548287803, "grad_norm": 0.27345237555875945, "learning_rate": 0.00018980522450883286, "loss": 0.8205, "step": 887 }, { "epoch": 0.17083493651404386, "grad_norm": 0.18098888051943687, "learning_rate": 0.00018977779828267311, "loss": 0.8715, "step": 888 }, { "epoch": 0.17102731819930742, "grad_norm": 0.18057574948780106, "learning_rate": 0.00018975033720168094, "loss": 0.8593, "step": 889 }, { "epoch": 0.17121969988457098, "grad_norm": 0.20258010719509148, "learning_rate": 0.00018972284127651768, "loss": 0.846, "step": 890 }, { "epoch": 0.17141208156983456, "grad_norm": 0.09148221872082807, "learning_rate": 0.00018969531051785818, "loss": 0.5987, "step": 891 }, { "epoch": 0.17160446325509812, "grad_norm": 0.20285238255146426, "learning_rate": 0.00018966774493639084, "loss": 0.8726, "step": 892 }, { "epoch": 0.17179684494036168, "grad_norm": 0.10095456270676138, "learning_rate": 0.00018964014454281758, "loss": 0.65, "step": 893 }, { "epoch": 0.17198922662562524, "grad_norm": 0.18765014459016893, "learning_rate": 0.00018961250934785377, "loss": 0.8686, "step": 894 }, { "epoch": 0.1721816083108888, "grad_norm": 0.11650868629322009, "learning_rate": 0.00018958483936222842, "loss": 0.6425, "step": 895 }, { "epoch": 0.17237398999615236, "grad_norm": 0.18503805596228148, "learning_rate": 0.0001895571345966839, "loss": 0.8469, "step": 896 }, { "epoch": 0.17256637168141592, "grad_norm": 0.20916577287960178, "learning_rate": 0.0001895293950619762, "loss": 0.887, "step": 897 }, { "epoch": 0.17275875336667948, "grad_norm": 0.18532710031931407, "learning_rate": 0.00018950162076887477, "loss": 0.8318, "step": 898 }, { "epoch": 0.17295113505194307, "grad_norm": 0.1857014680101479, "learning_rate": 0.00018947381172816248, "loss": 0.8376, "step": 899 }, { "epoch": 0.17314351673720663, "grad_norm": 0.1995976172300551, "learning_rate": 0.00018944596795063582, "loss": 0.8417, "step": 900 }, { "epoch": 0.17333589842247019, "grad_norm": 0.2230212546283908, "learning_rate": 0.0001894180894471047, "loss": 0.843, "step": 901 }, { "epoch": 0.17352828010773375, "grad_norm": 0.2004578181855843, "learning_rate": 0.00018939017622839253, "loss": 0.7863, "step": 902 }, { "epoch": 0.1737206617929973, "grad_norm": 0.10665977611781467, "learning_rate": 0.0001893622283053361, "loss": 0.5823, "step": 903 }, { "epoch": 0.17391304347826086, "grad_norm": 0.2036213107711315, "learning_rate": 0.00018933424568878584, "loss": 0.8225, "step": 904 }, { "epoch": 0.17410542516352442, "grad_norm": 0.244910566526374, "learning_rate": 0.00018930622838960555, "loss": 0.8764, "step": 905 }, { "epoch": 0.17429780684878798, "grad_norm": 0.19062521353349374, "learning_rate": 0.00018927817641867245, "loss": 0.8714, "step": 906 }, { "epoch": 0.17449018853405157, "grad_norm": 0.19781904564364228, "learning_rate": 0.00018925008978687736, "loss": 0.817, "step": 907 }, { "epoch": 0.17468257021931513, "grad_norm": 0.21476638235533455, "learning_rate": 0.00018922196850512446, "loss": 0.8711, "step": 908 }, { "epoch": 0.1748749519045787, "grad_norm": 0.17844470580002467, "learning_rate": 0.00018919381258433133, "loss": 0.8355, "step": 909 }, { "epoch": 0.17506733358984225, "grad_norm": 0.1882287331333208, "learning_rate": 0.00018916562203542917, "loss": 0.8509, "step": 910 }, { "epoch": 0.1752597152751058, "grad_norm": 0.18546342737074906, "learning_rate": 0.00018913739686936246, "loss": 0.815, "step": 911 }, { "epoch": 0.17545209696036937, "grad_norm": 0.21841384389166785, "learning_rate": 0.00018910913709708918, "loss": 0.7656, "step": 912 }, { "epoch": 0.17564447864563293, "grad_norm": 0.22499714017824934, "learning_rate": 0.00018908084272958078, "loss": 0.8285, "step": 913 }, { "epoch": 0.1758368603308965, "grad_norm": 0.2022279041659363, "learning_rate": 0.00018905251377782206, "loss": 0.8069, "step": 914 }, { "epoch": 0.17602924201616005, "grad_norm": 0.11614010858896158, "learning_rate": 0.00018902415025281134, "loss": 0.6577, "step": 915 }, { "epoch": 0.17622162370142364, "grad_norm": 0.20609286503897417, "learning_rate": 0.00018899575216556032, "loss": 0.899, "step": 916 }, { "epoch": 0.1764140053866872, "grad_norm": 0.1977643002511966, "learning_rate": 0.00018896731952709408, "loss": 0.8981, "step": 917 }, { "epoch": 0.17660638707195075, "grad_norm": 0.25653594548931336, "learning_rate": 0.00018893885234845116, "loss": 0.8789, "step": 918 }, { "epoch": 0.17679876875721431, "grad_norm": 0.19370986497735082, "learning_rate": 0.0001889103506406835, "loss": 0.8329, "step": 919 }, { "epoch": 0.17699115044247787, "grad_norm": 0.21024479139650926, "learning_rate": 0.00018888181441485647, "loss": 0.8605, "step": 920 }, { "epoch": 0.17718353212774143, "grad_norm": 0.32030747848795027, "learning_rate": 0.0001888532436820488, "loss": 0.8505, "step": 921 }, { "epoch": 0.177375913813005, "grad_norm": 0.17910373545398395, "learning_rate": 0.0001888246384533526, "loss": 0.833, "step": 922 }, { "epoch": 0.17756829549826855, "grad_norm": 0.17582856937614907, "learning_rate": 0.00018879599873987345, "loss": 0.8172, "step": 923 }, { "epoch": 0.17776067718353214, "grad_norm": 0.19484751287496266, "learning_rate": 0.00018876732455273024, "loss": 0.8464, "step": 924 }, { "epoch": 0.1779530588687957, "grad_norm": 0.19016119488627414, "learning_rate": 0.00018873861590305526, "loss": 0.8469, "step": 925 }, { "epoch": 0.17814544055405926, "grad_norm": 0.1814205832177741, "learning_rate": 0.00018870987280199426, "loss": 0.8417, "step": 926 }, { "epoch": 0.17833782223932282, "grad_norm": 0.22009475254926916, "learning_rate": 0.00018868109526070628, "loss": 0.8215, "step": 927 }, { "epoch": 0.17853020392458638, "grad_norm": 0.16308661329864182, "learning_rate": 0.00018865228329036372, "loss": 0.8206, "step": 928 }, { "epoch": 0.17872258560984994, "grad_norm": 0.23505423182876387, "learning_rate": 0.0001886234369021524, "loss": 0.8185, "step": 929 }, { "epoch": 0.1789149672951135, "grad_norm": 0.10382846564446144, "learning_rate": 0.0001885945561072715, "loss": 0.6568, "step": 930 }, { "epoch": 0.17910734898037706, "grad_norm": 0.18793876442087137, "learning_rate": 0.0001885656409169335, "loss": 0.8652, "step": 931 }, { "epoch": 0.17929973066564064, "grad_norm": 0.17456380864688023, "learning_rate": 0.0001885366913423643, "loss": 0.8581, "step": 932 }, { "epoch": 0.1794921123509042, "grad_norm": 0.17791700009288652, "learning_rate": 0.0001885077073948031, "loss": 0.7975, "step": 933 }, { "epoch": 0.17968449403616776, "grad_norm": 0.19177274673495343, "learning_rate": 0.0001884786890855025, "loss": 0.8488, "step": 934 }, { "epoch": 0.17987687572143132, "grad_norm": 0.19825128726748206, "learning_rate": 0.00018844963642572837, "loss": 0.8957, "step": 935 }, { "epoch": 0.18006925740669488, "grad_norm": 0.1943092180541477, "learning_rate": 0.00018842054942675996, "loss": 0.8623, "step": 936 }, { "epoch": 0.18026163909195844, "grad_norm": 0.2661994627240527, "learning_rate": 0.00018839142809988987, "loss": 0.8083, "step": 937 }, { "epoch": 0.180454020777222, "grad_norm": 0.19374694023428304, "learning_rate": 0.00018836227245642398, "loss": 0.8846, "step": 938 }, { "epoch": 0.18064640246248556, "grad_norm": 0.1904527322682366, "learning_rate": 0.00018833308250768154, "loss": 0.8392, "step": 939 }, { "epoch": 0.18083878414774912, "grad_norm": 0.10599940676287511, "learning_rate": 0.00018830385826499506, "loss": 0.6419, "step": 940 }, { "epoch": 0.1810311658330127, "grad_norm": 0.18504264767377404, "learning_rate": 0.0001882745997397104, "loss": 0.8124, "step": 941 }, { "epoch": 0.18122354751827627, "grad_norm": 0.21520018462913654, "learning_rate": 0.00018824530694318674, "loss": 0.8, "step": 942 }, { "epoch": 0.18141592920353983, "grad_norm": 0.19637074834742, "learning_rate": 0.00018821597988679658, "loss": 0.8359, "step": 943 }, { "epoch": 0.1816083108888034, "grad_norm": 0.17823149815332773, "learning_rate": 0.0001881866185819256, "loss": 0.8521, "step": 944 }, { "epoch": 0.18180069257406695, "grad_norm": 0.1943429505464618, "learning_rate": 0.000188157223039973, "loss": 0.8185, "step": 945 }, { "epoch": 0.1819930742593305, "grad_norm": 0.19838187050716302, "learning_rate": 0.00018812779327235103, "loss": 0.84, "step": 946 }, { "epoch": 0.18218545594459407, "grad_norm": 0.2044011281581252, "learning_rate": 0.0001880983292904854, "loss": 0.841, "step": 947 }, { "epoch": 0.18237783762985763, "grad_norm": 0.18062606400087927, "learning_rate": 0.000188068831105815, "loss": 0.8852, "step": 948 }, { "epoch": 0.1825702193151212, "grad_norm": 0.20492883215990929, "learning_rate": 0.00018803929872979213, "loss": 0.8479, "step": 949 }, { "epoch": 0.18276260100038477, "grad_norm": 0.1995369346238575, "learning_rate": 0.00018800973217388215, "loss": 0.8302, "step": 950 }, { "epoch": 0.18295498268564833, "grad_norm": 0.18004703498261465, "learning_rate": 0.0001879801314495639, "loss": 0.855, "step": 951 }, { "epoch": 0.1831473643709119, "grad_norm": 0.21978724366837313, "learning_rate": 0.00018795049656832936, "loss": 0.8748, "step": 952 }, { "epoch": 0.18333974605617545, "grad_norm": 0.18990355332278228, "learning_rate": 0.00018792082754168386, "loss": 0.9255, "step": 953 }, { "epoch": 0.183532127741439, "grad_norm": 0.1898516451655455, "learning_rate": 0.00018789112438114588, "loss": 0.855, "step": 954 }, { "epoch": 0.18372450942670257, "grad_norm": 0.18645646547478065, "learning_rate": 0.00018786138709824725, "loss": 0.8068, "step": 955 }, { "epoch": 0.18391689111196613, "grad_norm": 0.2068085742376573, "learning_rate": 0.00018783161570453297, "loss": 0.8017, "step": 956 }, { "epoch": 0.18410927279722972, "grad_norm": 0.10069460885318508, "learning_rate": 0.0001878018102115614, "loss": 0.6505, "step": 957 }, { "epoch": 0.18430165448249328, "grad_norm": 0.20541847931639812, "learning_rate": 0.00018777197063090393, "loss": 0.8598, "step": 958 }, { "epoch": 0.18449403616775684, "grad_norm": 0.21096285088247535, "learning_rate": 0.0001877420969741454, "loss": 0.833, "step": 959 }, { "epoch": 0.1846864178530204, "grad_norm": 0.2906054301735772, "learning_rate": 0.0001877121892528838, "loss": 0.8389, "step": 960 }, { "epoch": 0.18487879953828396, "grad_norm": 0.20806111123422363, "learning_rate": 0.0001876822474787303, "loss": 0.8258, "step": 961 }, { "epoch": 0.18507118122354752, "grad_norm": 0.1907266301875969, "learning_rate": 0.00018765227166330935, "loss": 0.8931, "step": 962 }, { "epoch": 0.18526356290881107, "grad_norm": 0.20969938248394698, "learning_rate": 0.00018762226181825856, "loss": 0.8944, "step": 963 }, { "epoch": 0.18545594459407463, "grad_norm": 0.19352931262023032, "learning_rate": 0.0001875922179552288, "loss": 0.8839, "step": 964 }, { "epoch": 0.1856483262793382, "grad_norm": 0.19104561868226583, "learning_rate": 0.00018756214008588417, "loss": 0.8071, "step": 965 }, { "epoch": 0.18584070796460178, "grad_norm": 0.18182236499061016, "learning_rate": 0.0001875320282219019, "loss": 0.8868, "step": 966 }, { "epoch": 0.18603308964986534, "grad_norm": 0.18500841778897095, "learning_rate": 0.00018750188237497245, "loss": 0.8536, "step": 967 }, { "epoch": 0.1862254713351289, "grad_norm": 0.19538869421718108, "learning_rate": 0.0001874717025567995, "loss": 0.89, "step": 968 }, { "epoch": 0.18641785302039246, "grad_norm": 0.19283394296633277, "learning_rate": 0.00018744148877909987, "loss": 0.8904, "step": 969 }, { "epoch": 0.18661023470565602, "grad_norm": 0.18723975040376076, "learning_rate": 0.0001874112410536036, "loss": 0.8938, "step": 970 }, { "epoch": 0.18680261639091958, "grad_norm": 0.18574319182327298, "learning_rate": 0.0001873809593920539, "loss": 0.8782, "step": 971 }, { "epoch": 0.18699499807618314, "grad_norm": 0.19761947940992702, "learning_rate": 0.00018735064380620716, "loss": 0.8404, "step": 972 }, { "epoch": 0.1871873797614467, "grad_norm": 0.1996537255345759, "learning_rate": 0.0001873202943078329, "loss": 0.865, "step": 973 }, { "epoch": 0.18737976144671029, "grad_norm": 0.16537554948618288, "learning_rate": 0.00018728991090871387, "loss": 0.8065, "step": 974 }, { "epoch": 0.18757214313197385, "grad_norm": 0.18188093531513244, "learning_rate": 0.00018725949362064594, "loss": 0.862, "step": 975 }, { "epoch": 0.1877645248172374, "grad_norm": 0.1991741259157709, "learning_rate": 0.00018722904245543816, "loss": 0.8701, "step": 976 }, { "epoch": 0.18795690650250096, "grad_norm": 0.20537968861080952, "learning_rate": 0.0001871985574249127, "loss": 0.8107, "step": 977 }, { "epoch": 0.18814928818776452, "grad_norm": 0.20676043996016907, "learning_rate": 0.00018716803854090493, "loss": 0.8426, "step": 978 }, { "epoch": 0.18834166987302808, "grad_norm": 0.18794416655889865, "learning_rate": 0.00018713748581526332, "loss": 0.8612, "step": 979 }, { "epoch": 0.18853405155829164, "grad_norm": 0.18885573461122324, "learning_rate": 0.00018710689925984948, "loss": 0.8589, "step": 980 }, { "epoch": 0.1887264332435552, "grad_norm": 0.18511024234203963, "learning_rate": 0.00018707627888653815, "loss": 0.8505, "step": 981 }, { "epoch": 0.1889188149288188, "grad_norm": 0.19098240624738802, "learning_rate": 0.00018704562470721727, "loss": 0.8112, "step": 982 }, { "epoch": 0.18911119661408235, "grad_norm": 0.1727980586992357, "learning_rate": 0.00018701493673378778, "loss": 0.8318, "step": 983 }, { "epoch": 0.1893035782993459, "grad_norm": 0.1327849753099699, "learning_rate": 0.00018698421497816386, "loss": 0.6694, "step": 984 }, { "epoch": 0.18949595998460947, "grad_norm": 0.20378680652243342, "learning_rate": 0.0001869534594522727, "loss": 0.8155, "step": 985 }, { "epoch": 0.18968834166987303, "grad_norm": 0.1752157415656886, "learning_rate": 0.00018692267016805471, "loss": 0.8276, "step": 986 }, { "epoch": 0.1898807233551366, "grad_norm": 0.1948511162385625, "learning_rate": 0.00018689184713746332, "loss": 0.8169, "step": 987 }, { "epoch": 0.19007310504040015, "grad_norm": 0.1843639194605135, "learning_rate": 0.0001868609903724651, "loss": 0.7653, "step": 988 }, { "epoch": 0.1902654867256637, "grad_norm": 0.1979691000163968, "learning_rate": 0.0001868300998850397, "loss": 0.8266, "step": 989 }, { "epoch": 0.19045786841092727, "grad_norm": 0.0881971362892459, "learning_rate": 0.0001867991756871799, "loss": 0.5935, "step": 990 }, { "epoch": 0.19065025009619085, "grad_norm": 0.17765391921437956, "learning_rate": 0.0001867682177908915, "loss": 0.7999, "step": 991 }, { "epoch": 0.1908426317814544, "grad_norm": 0.18566149466513088, "learning_rate": 0.0001867372262081934, "loss": 0.8708, "step": 992 }, { "epoch": 0.19103501346671797, "grad_norm": 0.17610424511094622, "learning_rate": 0.00018670620095111767, "loss": 0.8708, "step": 993 }, { "epoch": 0.19122739515198153, "grad_norm": 0.19506979772432634, "learning_rate": 0.00018667514203170936, "loss": 0.8691, "step": 994 }, { "epoch": 0.1914197768372451, "grad_norm": 0.1887624130153998, "learning_rate": 0.00018664404946202657, "loss": 0.9067, "step": 995 }, { "epoch": 0.19161215852250865, "grad_norm": 0.19171400433491634, "learning_rate": 0.00018661292325414057, "loss": 0.824, "step": 996 }, { "epoch": 0.1918045402077722, "grad_norm": 0.18419059592102707, "learning_rate": 0.00018658176342013557, "loss": 0.8277, "step": 997 }, { "epoch": 0.19199692189303577, "grad_norm": 0.195597231134138, "learning_rate": 0.0001865505699721089, "loss": 0.8351, "step": 998 }, { "epoch": 0.19218930357829936, "grad_norm": 0.21062239084004244, "learning_rate": 0.00018651934292217097, "loss": 0.872, "step": 999 }, { "epoch": 0.19238168526356292, "grad_norm": 0.19778703944581325, "learning_rate": 0.00018648808228244516, "loss": 0.8456, "step": 1000 }, { "epoch": 0.19257406694882648, "grad_norm": 0.18253786623307736, "learning_rate": 0.00018645678806506795, "loss": 0.8404, "step": 1001 }, { "epoch": 0.19276644863409004, "grad_norm": 0.19483347691624248, "learning_rate": 0.0001864254602821888, "loss": 0.8255, "step": 1002 }, { "epoch": 0.1929588303193536, "grad_norm": 0.17607756490817805, "learning_rate": 0.00018639409894597025, "loss": 0.823, "step": 1003 }, { "epoch": 0.19315121200461716, "grad_norm": 0.17839291521674244, "learning_rate": 0.00018636270406858786, "loss": 0.8563, "step": 1004 }, { "epoch": 0.19334359368988072, "grad_norm": 0.17530905447657658, "learning_rate": 0.00018633127566223021, "loss": 0.8517, "step": 1005 }, { "epoch": 0.19353597537514428, "grad_norm": 0.18329697115575586, "learning_rate": 0.00018629981373909885, "loss": 0.8071, "step": 1006 }, { "epoch": 0.19372835706040784, "grad_norm": 0.17489791686296496, "learning_rate": 0.00018626831831140844, "loss": 0.7901, "step": 1007 }, { "epoch": 0.19392073874567142, "grad_norm": 0.20075012910645382, "learning_rate": 0.0001862367893913865, "loss": 0.8389, "step": 1008 }, { "epoch": 0.19411312043093498, "grad_norm": 0.19462597498962597, "learning_rate": 0.00018620522699127373, "loss": 0.855, "step": 1009 }, { "epoch": 0.19430550211619854, "grad_norm": 0.17384800256250107, "learning_rate": 0.00018617363112332375, "loss": 0.8331, "step": 1010 }, { "epoch": 0.1944978838014621, "grad_norm": 0.2152364457345958, "learning_rate": 0.00018614200179980307, "loss": 0.8478, "step": 1011 }, { "epoch": 0.19469026548672566, "grad_norm": 0.19638494258803937, "learning_rate": 0.00018611033903299134, "loss": 0.8217, "step": 1012 }, { "epoch": 0.19488264717198922, "grad_norm": 0.1731210884039701, "learning_rate": 0.00018607864283518114, "loss": 0.841, "step": 1013 }, { "epoch": 0.19507502885725278, "grad_norm": 0.20247499248495832, "learning_rate": 0.00018604691321867805, "loss": 0.8131, "step": 1014 }, { "epoch": 0.19526741054251634, "grad_norm": 0.1780954587298043, "learning_rate": 0.00018601515019580053, "loss": 0.83, "step": 1015 }, { "epoch": 0.19545979222777993, "grad_norm": 0.2123271033803798, "learning_rate": 0.00018598335377888013, "loss": 0.8898, "step": 1016 }, { "epoch": 0.1956521739130435, "grad_norm": 0.20393701877481696, "learning_rate": 0.0001859515239802613, "loss": 0.8496, "step": 1017 }, { "epoch": 0.19584455559830705, "grad_norm": 0.09652588374756206, "learning_rate": 0.00018591966081230144, "loss": 0.6001, "step": 1018 }, { "epoch": 0.1960369372835706, "grad_norm": 0.10100458453566588, "learning_rate": 0.00018588776428737097, "loss": 0.6355, "step": 1019 }, { "epoch": 0.19622931896883417, "grad_norm": 0.1792982016408341, "learning_rate": 0.0001858558344178532, "loss": 0.838, "step": 1020 }, { "epoch": 0.19642170065409773, "grad_norm": 0.1986232138117339, "learning_rate": 0.00018582387121614437, "loss": 0.8666, "step": 1021 }, { "epoch": 0.19661408233936128, "grad_norm": 0.19849354240466277, "learning_rate": 0.00018579187469465377, "loss": 0.8547, "step": 1022 }, { "epoch": 0.19680646402462484, "grad_norm": 0.18980660336125432, "learning_rate": 0.0001857598448658035, "loss": 0.9195, "step": 1023 }, { "epoch": 0.19699884570988843, "grad_norm": 0.19304967400325349, "learning_rate": 0.0001857277817420287, "loss": 0.8712, "step": 1024 }, { "epoch": 0.197191227395152, "grad_norm": 0.18870180218645877, "learning_rate": 0.00018569568533577727, "loss": 0.864, "step": 1025 }, { "epoch": 0.19738360908041555, "grad_norm": 0.1853069275141225, "learning_rate": 0.00018566355565951024, "loss": 0.8949, "step": 1026 }, { "epoch": 0.1975759907656791, "grad_norm": 0.22340298677774675, "learning_rate": 0.0001856313927257014, "loss": 0.8632, "step": 1027 }, { "epoch": 0.19776837245094267, "grad_norm": 0.19763885469308093, "learning_rate": 0.00018559919654683756, "loss": 0.8626, "step": 1028 }, { "epoch": 0.19796075413620623, "grad_norm": 0.19335869750529888, "learning_rate": 0.00018556696713541832, "loss": 0.8588, "step": 1029 }, { "epoch": 0.1981531358214698, "grad_norm": 0.18136859374121625, "learning_rate": 0.0001855347045039563, "loss": 0.83, "step": 1030 }, { "epoch": 0.19834551750673335, "grad_norm": 0.18193489450698644, "learning_rate": 0.00018550240866497695, "loss": 0.8265, "step": 1031 }, { "epoch": 0.1985378991919969, "grad_norm": 0.18376473881083233, "learning_rate": 0.0001854700796310186, "loss": 0.8335, "step": 1032 }, { "epoch": 0.1987302808772605, "grad_norm": 0.1982361835304817, "learning_rate": 0.00018543771741463255, "loss": 0.8479, "step": 1033 }, { "epoch": 0.19892266256252406, "grad_norm": 0.19824904366066298, "learning_rate": 0.00018540532202838286, "loss": 0.8634, "step": 1034 }, { "epoch": 0.19911504424778761, "grad_norm": 0.1833397584434087, "learning_rate": 0.00018537289348484657, "loss": 0.8267, "step": 1035 }, { "epoch": 0.19930742593305117, "grad_norm": 0.17695151473857604, "learning_rate": 0.00018534043179661356, "loss": 0.7944, "step": 1036 }, { "epoch": 0.19949980761831473, "grad_norm": 0.13506261349994883, "learning_rate": 0.00018530793697628658, "loss": 0.619, "step": 1037 }, { "epoch": 0.1996921893035783, "grad_norm": 0.21384098009136587, "learning_rate": 0.00018527540903648122, "loss": 0.9124, "step": 1038 }, { "epoch": 0.19988457098884185, "grad_norm": 0.19392452537077354, "learning_rate": 0.00018524284798982594, "loss": 0.874, "step": 1039 }, { "epoch": 0.2000769526741054, "grad_norm": 0.19854990371131154, "learning_rate": 0.00018521025384896208, "loss": 0.8529, "step": 1040 }, { "epoch": 0.200269334359369, "grad_norm": 0.19510079920558315, "learning_rate": 0.0001851776266265438, "loss": 0.8272, "step": 1041 }, { "epoch": 0.20046171604463256, "grad_norm": 0.10417682005650666, "learning_rate": 0.00018514496633523813, "loss": 0.6437, "step": 1042 }, { "epoch": 0.20065409772989612, "grad_norm": 0.20258468230778057, "learning_rate": 0.0001851122729877249, "loss": 0.7969, "step": 1043 }, { "epoch": 0.20084647941515968, "grad_norm": 0.19142455918355364, "learning_rate": 0.00018507954659669677, "loss": 0.7808, "step": 1044 }, { "epoch": 0.20103886110042324, "grad_norm": 0.20327706948894292, "learning_rate": 0.0001850467871748593, "loss": 0.8466, "step": 1045 }, { "epoch": 0.2012312427856868, "grad_norm": 0.1905114538708152, "learning_rate": 0.0001850139947349308, "loss": 0.8132, "step": 1046 }, { "epoch": 0.20142362447095036, "grad_norm": 0.1704693485656684, "learning_rate": 0.0001849811692896424, "loss": 0.8666, "step": 1047 }, { "epoch": 0.20161600615621392, "grad_norm": 0.19599932421786095, "learning_rate": 0.0001849483108517381, "loss": 0.8344, "step": 1048 }, { "epoch": 0.2018083878414775, "grad_norm": 0.17161869401494706, "learning_rate": 0.0001849154194339747, "loss": 0.8189, "step": 1049 }, { "epoch": 0.20200076952674106, "grad_norm": 0.2036541122139959, "learning_rate": 0.00018488249504912174, "loss": 0.8661, "step": 1050 }, { "epoch": 0.20219315121200462, "grad_norm": 0.1723915702103837, "learning_rate": 0.00018484953770996163, "loss": 0.8273, "step": 1051 }, { "epoch": 0.20238553289726818, "grad_norm": 0.18177432255338133, "learning_rate": 0.00018481654742928948, "loss": 0.7862, "step": 1052 }, { "epoch": 0.20257791458253174, "grad_norm": 0.19434260578732968, "learning_rate": 0.00018478352421991335, "loss": 0.8433, "step": 1053 }, { "epoch": 0.2027702962677953, "grad_norm": 0.18869351303238052, "learning_rate": 0.0001847504680946539, "loss": 0.9139, "step": 1054 }, { "epoch": 0.20296267795305886, "grad_norm": 0.1953627026901257, "learning_rate": 0.0001847173790663447, "loss": 0.8268, "step": 1055 }, { "epoch": 0.20315505963832242, "grad_norm": 0.1889637125588925, "learning_rate": 0.00018468425714783204, "loss": 0.8381, "step": 1056 }, { "epoch": 0.20334744132358598, "grad_norm": 0.16983548639923043, "learning_rate": 0.000184651102351975, "loss": 0.8669, "step": 1057 }, { "epoch": 0.20353982300884957, "grad_norm": 0.18081688556240455, "learning_rate": 0.0001846179146916454, "loss": 0.8343, "step": 1058 }, { "epoch": 0.20373220469411313, "grad_norm": 0.18763043944100954, "learning_rate": 0.00018458469417972784, "loss": 0.834, "step": 1059 }, { "epoch": 0.2039245863793767, "grad_norm": 0.11124867591156906, "learning_rate": 0.00018455144082911966, "loss": 0.6731, "step": 1060 }, { "epoch": 0.20411696806464025, "grad_norm": 0.18395766060858312, "learning_rate": 0.00018451815465273094, "loss": 0.8249, "step": 1061 }, { "epoch": 0.2043093497499038, "grad_norm": 0.18190171997252993, "learning_rate": 0.00018448483566348457, "loss": 0.8426, "step": 1062 }, { "epoch": 0.20450173143516737, "grad_norm": 0.18413384444002814, "learning_rate": 0.00018445148387431605, "loss": 0.8317, "step": 1063 }, { "epoch": 0.20469411312043093, "grad_norm": 0.19266929172816374, "learning_rate": 0.0001844180992981738, "loss": 0.8371, "step": 1064 }, { "epoch": 0.20488649480569449, "grad_norm": 0.17312149033511606, "learning_rate": 0.00018438468194801875, "loss": 0.8762, "step": 1065 }, { "epoch": 0.20507887649095807, "grad_norm": 0.183013602275573, "learning_rate": 0.00018435123183682475, "loss": 0.8101, "step": 1066 }, { "epoch": 0.20527125817622163, "grad_norm": 0.214787271505764, "learning_rate": 0.00018431774897757826, "loss": 0.7886, "step": 1067 }, { "epoch": 0.2054636398614852, "grad_norm": 0.16980799192513502, "learning_rate": 0.00018428423338327844, "loss": 0.8152, "step": 1068 }, { "epoch": 0.20565602154674875, "grad_norm": 0.1919906343977681, "learning_rate": 0.00018425068506693727, "loss": 0.9078, "step": 1069 }, { "epoch": 0.2058484032320123, "grad_norm": 0.19479508038034798, "learning_rate": 0.00018421710404157928, "loss": 0.8173, "step": 1070 }, { "epoch": 0.20604078491727587, "grad_norm": 0.19440499679349368, "learning_rate": 0.00018418349032024183, "loss": 0.8376, "step": 1071 }, { "epoch": 0.20623316660253943, "grad_norm": 0.19056410394866394, "learning_rate": 0.0001841498439159749, "loss": 0.8861, "step": 1072 }, { "epoch": 0.206425548287803, "grad_norm": 0.1814072925812284, "learning_rate": 0.00018411616484184125, "loss": 0.831, "step": 1073 }, { "epoch": 0.20661792997306658, "grad_norm": 0.22974073524493308, "learning_rate": 0.00018408245311091616, "loss": 0.8088, "step": 1074 }, { "epoch": 0.20681031165833014, "grad_norm": 0.18384049427072102, "learning_rate": 0.00018404870873628775, "loss": 0.7821, "step": 1075 }, { "epoch": 0.2070026933435937, "grad_norm": 0.1861991075636456, "learning_rate": 0.00018401493173105673, "loss": 0.8699, "step": 1076 }, { "epoch": 0.20719507502885726, "grad_norm": 0.18351977538193665, "learning_rate": 0.0001839811221083365, "loss": 0.8176, "step": 1077 }, { "epoch": 0.20738745671412082, "grad_norm": 0.19468961005227983, "learning_rate": 0.00018394727988125308, "loss": 0.8367, "step": 1078 }, { "epoch": 0.20757983839938438, "grad_norm": 0.174865541199393, "learning_rate": 0.00018391340506294522, "loss": 0.8217, "step": 1079 }, { "epoch": 0.20777222008464794, "grad_norm": 0.18654499442076528, "learning_rate": 0.00018387949766656434, "loss": 0.7926, "step": 1080 }, { "epoch": 0.2079646017699115, "grad_norm": 0.19107697282095396, "learning_rate": 0.00018384555770527435, "loss": 0.8482, "step": 1081 }, { "epoch": 0.20815698345517505, "grad_norm": 0.19518319286517472, "learning_rate": 0.00018381158519225203, "loss": 0.8423, "step": 1082 }, { "epoch": 0.20834936514043864, "grad_norm": 0.1907546619363961, "learning_rate": 0.0001837775801406866, "loss": 0.8454, "step": 1083 }, { "epoch": 0.2085417468257022, "grad_norm": 0.25044185412489245, "learning_rate": 0.00018374354256378, "loss": 0.8737, "step": 1084 }, { "epoch": 0.20873412851096576, "grad_norm": 0.10415407574105295, "learning_rate": 0.0001837094724747468, "loss": 0.646, "step": 1085 }, { "epoch": 0.20892651019622932, "grad_norm": 0.20513937430249618, "learning_rate": 0.0001836753698868142, "loss": 0.8404, "step": 1086 }, { "epoch": 0.20911889188149288, "grad_norm": 0.22579510763815844, "learning_rate": 0.00018364123481322197, "loss": 0.8806, "step": 1087 }, { "epoch": 0.20931127356675644, "grad_norm": 0.17728947377360482, "learning_rate": 0.00018360706726722252, "loss": 0.8294, "step": 1088 }, { "epoch": 0.20950365525202, "grad_norm": 0.2223110097935046, "learning_rate": 0.0001835728672620809, "loss": 0.8086, "step": 1089 }, { "epoch": 0.20969603693728356, "grad_norm": 0.214319131435939, "learning_rate": 0.0001835386348110747, "loss": 0.8907, "step": 1090 }, { "epoch": 0.20988841862254715, "grad_norm": 0.190699265142797, "learning_rate": 0.00018350436992749418, "loss": 0.8143, "step": 1091 }, { "epoch": 0.2100808003078107, "grad_norm": 0.202906079187464, "learning_rate": 0.00018347007262464206, "loss": 0.8409, "step": 1092 }, { "epoch": 0.21027318199307427, "grad_norm": 0.22474703048088907, "learning_rate": 0.00018343574291583383, "loss": 0.8097, "step": 1093 }, { "epoch": 0.21046556367833782, "grad_norm": 0.21666752748334286, "learning_rate": 0.0001834013808143974, "loss": 0.7969, "step": 1094 }, { "epoch": 0.21065794536360138, "grad_norm": 0.19177432290318064, "learning_rate": 0.00018336698633367337, "loss": 0.8457, "step": 1095 }, { "epoch": 0.21085032704886494, "grad_norm": 0.18836762505075014, "learning_rate": 0.00018333255948701482, "loss": 0.8069, "step": 1096 }, { "epoch": 0.2110427087341285, "grad_norm": 0.20815971888913384, "learning_rate": 0.00018329810028778747, "loss": 0.8753, "step": 1097 }, { "epoch": 0.21123509041939206, "grad_norm": 0.19853365360288086, "learning_rate": 0.0001832636087493695, "loss": 0.8448, "step": 1098 }, { "epoch": 0.21142747210465565, "grad_norm": 0.20280374317917316, "learning_rate": 0.0001832290848851518, "loss": 0.8401, "step": 1099 }, { "epoch": 0.2116198537899192, "grad_norm": 0.20789397903197054, "learning_rate": 0.0001831945287085377, "loss": 0.8843, "step": 1100 }, { "epoch": 0.21181223547518277, "grad_norm": 0.1995253230786853, "learning_rate": 0.00018315994023294306, "loss": 0.8805, "step": 1101 }, { "epoch": 0.21200461716044633, "grad_norm": 0.1972203902767824, "learning_rate": 0.00018312531947179634, "loss": 0.8428, "step": 1102 }, { "epoch": 0.2121969988457099, "grad_norm": 0.18003108783431698, "learning_rate": 0.00018309066643853853, "loss": 0.8771, "step": 1103 }, { "epoch": 0.21238938053097345, "grad_norm": 0.20255166613480957, "learning_rate": 0.00018305598114662312, "loss": 0.8579, "step": 1104 }, { "epoch": 0.212581762216237, "grad_norm": 0.19194837961325695, "learning_rate": 0.00018302126360951611, "loss": 0.8795, "step": 1105 }, { "epoch": 0.21277414390150057, "grad_norm": 0.19813097472140395, "learning_rate": 0.00018298651384069604, "loss": 0.8127, "step": 1106 }, { "epoch": 0.21296652558676413, "grad_norm": 0.11329944122019521, "learning_rate": 0.00018295173185365403, "loss": 0.655, "step": 1107 }, { "epoch": 0.21315890727202771, "grad_norm": 0.20741702912045198, "learning_rate": 0.00018291691766189358, "loss": 0.8585, "step": 1108 }, { "epoch": 0.21335128895729127, "grad_norm": 0.1992657887297054, "learning_rate": 0.0001828820712789308, "loss": 0.8331, "step": 1109 }, { "epoch": 0.21354367064255483, "grad_norm": 0.18118615218224002, "learning_rate": 0.0001828471927182942, "loss": 0.8218, "step": 1110 }, { "epoch": 0.2137360523278184, "grad_norm": 0.2125353871898771, "learning_rate": 0.0001828122819935249, "loss": 0.8428, "step": 1111 }, { "epoch": 0.21392843401308195, "grad_norm": 0.2019083346698719, "learning_rate": 0.0001827773391181764, "loss": 0.8037, "step": 1112 }, { "epoch": 0.2141208156983455, "grad_norm": 0.1973848710026425, "learning_rate": 0.0001827423641058148, "loss": 0.7977, "step": 1113 }, { "epoch": 0.21431319738360907, "grad_norm": 0.20103368154805112, "learning_rate": 0.00018270735697001848, "loss": 0.8762, "step": 1114 }, { "epoch": 0.21450557906887263, "grad_norm": 0.23856478237295636, "learning_rate": 0.00018267231772437853, "loss": 0.8509, "step": 1115 }, { "epoch": 0.21469796075413622, "grad_norm": 0.21470324008589212, "learning_rate": 0.00018263724638249834, "loss": 0.8691, "step": 1116 }, { "epoch": 0.21489034243939978, "grad_norm": 0.1924635588955638, "learning_rate": 0.00018260214295799383, "loss": 0.8174, "step": 1117 }, { "epoch": 0.21508272412466334, "grad_norm": 0.1857047950015689, "learning_rate": 0.0001825670074644933, "loss": 0.8291, "step": 1118 }, { "epoch": 0.2152751058099269, "grad_norm": 0.1060268032459395, "learning_rate": 0.00018253183991563768, "loss": 0.6678, "step": 1119 }, { "epoch": 0.21546748749519046, "grad_norm": 0.18065465653996554, "learning_rate": 0.0001824966403250801, "loss": 0.8313, "step": 1120 }, { "epoch": 0.21565986918045402, "grad_norm": 0.19337442572800864, "learning_rate": 0.00018246140870648633, "loss": 0.8636, "step": 1121 }, { "epoch": 0.21585225086571758, "grad_norm": 0.09464721678397328, "learning_rate": 0.00018242614507353445, "loss": 0.656, "step": 1122 }, { "epoch": 0.21604463255098114, "grad_norm": 0.18102079198504373, "learning_rate": 0.00018239084943991505, "loss": 0.8481, "step": 1123 }, { "epoch": 0.2162370142362447, "grad_norm": 0.1854449289915446, "learning_rate": 0.00018235552181933108, "loss": 0.8022, "step": 1124 }, { "epoch": 0.21642939592150828, "grad_norm": 0.21307560157549055, "learning_rate": 0.00018232016222549796, "loss": 0.8503, "step": 1125 }, { "epoch": 0.21662177760677184, "grad_norm": 0.09411868729318804, "learning_rate": 0.0001822847706721435, "loss": 0.6368, "step": 1126 }, { "epoch": 0.2168141592920354, "grad_norm": 0.20134459874486654, "learning_rate": 0.00018224934717300792, "loss": 0.8346, "step": 1127 }, { "epoch": 0.21700654097729896, "grad_norm": 0.1921050696003894, "learning_rate": 0.00018221389174184386, "loss": 0.8515, "step": 1128 }, { "epoch": 0.21719892266256252, "grad_norm": 0.21062809795586995, "learning_rate": 0.00018217840439241633, "loss": 0.7911, "step": 1129 }, { "epoch": 0.21739130434782608, "grad_norm": 0.17812355789018652, "learning_rate": 0.00018214288513850266, "loss": 0.8737, "step": 1130 }, { "epoch": 0.21758368603308964, "grad_norm": 0.215296769903897, "learning_rate": 0.00018210733399389277, "loss": 0.8154, "step": 1131 }, { "epoch": 0.2177760677183532, "grad_norm": 0.1729464936585431, "learning_rate": 0.00018207175097238878, "loss": 0.7871, "step": 1132 }, { "epoch": 0.2179684494036168, "grad_norm": 0.18323730762003632, "learning_rate": 0.00018203613608780523, "loss": 0.8438, "step": 1133 }, { "epoch": 0.21816083108888035, "grad_norm": 0.1992257269498531, "learning_rate": 0.00018200048935396908, "loss": 0.8832, "step": 1134 }, { "epoch": 0.2183532127741439, "grad_norm": 0.20746122222110364, "learning_rate": 0.00018196481078471962, "loss": 0.8118, "step": 1135 }, { "epoch": 0.21854559445940747, "grad_norm": 0.18630747751654342, "learning_rate": 0.00018192910039390844, "loss": 0.8507, "step": 1136 }, { "epoch": 0.21873797614467103, "grad_norm": 0.18335574017518125, "learning_rate": 0.0001818933581953996, "loss": 0.865, "step": 1137 }, { "epoch": 0.21893035782993459, "grad_norm": 0.16768926536393258, "learning_rate": 0.00018185758420306947, "loss": 0.8033, "step": 1138 }, { "epoch": 0.21912273951519814, "grad_norm": 0.18077517616428326, "learning_rate": 0.0001818217784308067, "loss": 0.8653, "step": 1139 }, { "epoch": 0.2193151212004617, "grad_norm": 0.10774968353120094, "learning_rate": 0.0001817859408925123, "loss": 0.6251, "step": 1140 }, { "epoch": 0.2195075028857253, "grad_norm": 0.21895489996600379, "learning_rate": 0.00018175007160209968, "loss": 0.8321, "step": 1141 }, { "epoch": 0.21969988457098885, "grad_norm": 0.18801195154083244, "learning_rate": 0.00018171417057349457, "loss": 0.7766, "step": 1142 }, { "epoch": 0.2198922662562524, "grad_norm": 0.2532489760342155, "learning_rate": 0.00018167823782063488, "loss": 0.8402, "step": 1143 }, { "epoch": 0.22008464794151597, "grad_norm": 0.2444238719140528, "learning_rate": 0.00018164227335747105, "loss": 0.8915, "step": 1144 }, { "epoch": 0.22027702962677953, "grad_norm": 0.22055075747171865, "learning_rate": 0.00018160627719796568, "loss": 0.8885, "step": 1145 }, { "epoch": 0.2204694113120431, "grad_norm": 0.19328771892804947, "learning_rate": 0.0001815702493560937, "loss": 0.8888, "step": 1146 }, { "epoch": 0.22066179299730665, "grad_norm": 0.1818468245172146, "learning_rate": 0.00018153418984584238, "loss": 0.8728, "step": 1147 }, { "epoch": 0.2208541746825702, "grad_norm": 0.19436569070960588, "learning_rate": 0.00018149809868121125, "loss": 0.861, "step": 1148 }, { "epoch": 0.22104655636783377, "grad_norm": 0.20349439575204736, "learning_rate": 0.00018146197587621217, "loss": 0.8231, "step": 1149 }, { "epoch": 0.22123893805309736, "grad_norm": 0.19101276435919182, "learning_rate": 0.0001814258214448692, "loss": 0.8697, "step": 1150 }, { "epoch": 0.22143131973836092, "grad_norm": 0.17429261034187732, "learning_rate": 0.0001813896354012188, "loss": 0.8597, "step": 1151 }, { "epoch": 0.22162370142362448, "grad_norm": 0.18841591881379624, "learning_rate": 0.0001813534177593096, "loss": 0.8264, "step": 1152 }, { "epoch": 0.22181608310888803, "grad_norm": 0.18422178481649198, "learning_rate": 0.00018131716853320254, "loss": 0.8513, "step": 1153 }, { "epoch": 0.2220084647941516, "grad_norm": 0.21168569586198063, "learning_rate": 0.00018128088773697085, "loss": 0.8254, "step": 1154 }, { "epoch": 0.22220084647941515, "grad_norm": 0.21843904720383475, "learning_rate": 0.00018124457538469996, "loss": 0.8629, "step": 1155 }, { "epoch": 0.2223932281646787, "grad_norm": 0.1960946988314738, "learning_rate": 0.00018120823149048754, "loss": 0.7945, "step": 1156 }, { "epoch": 0.22258560984994227, "grad_norm": 0.20733286186892963, "learning_rate": 0.00018117185606844358, "loss": 0.8627, "step": 1157 }, { "epoch": 0.22277799153520586, "grad_norm": 0.09770229223288349, "learning_rate": 0.00018113544913269026, "loss": 0.6504, "step": 1158 }, { "epoch": 0.22297037322046942, "grad_norm": 0.18144795234158057, "learning_rate": 0.00018109901069736202, "loss": 0.8728, "step": 1159 }, { "epoch": 0.22316275490573298, "grad_norm": 0.1881139826923896, "learning_rate": 0.00018106254077660552, "loss": 0.8301, "step": 1160 }, { "epoch": 0.22335513659099654, "grad_norm": 0.2012960188541156, "learning_rate": 0.0001810260393845796, "loss": 0.8489, "step": 1161 }, { "epoch": 0.2235475182762601, "grad_norm": 0.1838571869124727, "learning_rate": 0.0001809895065354554, "loss": 0.8422, "step": 1162 }, { "epoch": 0.22373989996152366, "grad_norm": 0.18969706735854863, "learning_rate": 0.00018095294224341622, "loss": 0.8201, "step": 1163 }, { "epoch": 0.22393228164678722, "grad_norm": 0.19814818320519376, "learning_rate": 0.00018091634652265753, "loss": 0.8575, "step": 1164 }, { "epoch": 0.22412466333205078, "grad_norm": 0.19279700901449293, "learning_rate": 0.00018087971938738714, "loss": 0.8107, "step": 1165 }, { "epoch": 0.22431704501731436, "grad_norm": 0.18287022056632476, "learning_rate": 0.00018084306085182488, "loss": 0.8272, "step": 1166 }, { "epoch": 0.22450942670257792, "grad_norm": 0.18224533184337746, "learning_rate": 0.00018080637093020288, "loss": 0.8394, "step": 1167 }, { "epoch": 0.22470180838784148, "grad_norm": 0.1847498854826778, "learning_rate": 0.0001807696496367655, "loss": 0.8258, "step": 1168 }, { "epoch": 0.22489419007310504, "grad_norm": 0.17214422377615113, "learning_rate": 0.00018073289698576913, "loss": 0.8535, "step": 1169 }, { "epoch": 0.2250865717583686, "grad_norm": 0.18814549794809202, "learning_rate": 0.00018069611299148237, "loss": 0.8299, "step": 1170 }, { "epoch": 0.22527895344363216, "grad_norm": 0.1860501780264038, "learning_rate": 0.00018065929766818617, "loss": 0.8196, "step": 1171 }, { "epoch": 0.22547133512889572, "grad_norm": 0.09185221845223619, "learning_rate": 0.0001806224510301734, "loss": 0.5691, "step": 1172 }, { "epoch": 0.22566371681415928, "grad_norm": 0.19686155960827115, "learning_rate": 0.00018058557309174925, "loss": 0.8541, "step": 1173 }, { "epoch": 0.22585609849942284, "grad_norm": 0.1886467265183384, "learning_rate": 0.00018054866386723096, "loss": 0.8021, "step": 1174 }, { "epoch": 0.22604848018468643, "grad_norm": 0.2024758432900088, "learning_rate": 0.000180511723370948, "loss": 0.8574, "step": 1175 }, { "epoch": 0.22624086186995, "grad_norm": 0.20348117669728624, "learning_rate": 0.0001804747516172419, "loss": 0.8915, "step": 1176 }, { "epoch": 0.22643324355521355, "grad_norm": 0.17361195922113512, "learning_rate": 0.00018043774862046643, "loss": 0.814, "step": 1177 }, { "epoch": 0.2266256252404771, "grad_norm": 0.18987991690076358, "learning_rate": 0.0001804007143949874, "loss": 0.8494, "step": 1178 }, { "epoch": 0.22681800692574067, "grad_norm": 0.201851938559913, "learning_rate": 0.00018036364895518273, "loss": 0.7544, "step": 1179 }, { "epoch": 0.22701038861100423, "grad_norm": 0.19848906535387717, "learning_rate": 0.00018032655231544253, "loss": 0.7829, "step": 1180 }, { "epoch": 0.2272027702962678, "grad_norm": 0.209741554330773, "learning_rate": 0.00018028942449016904, "loss": 0.8559, "step": 1181 }, { "epoch": 0.22739515198153135, "grad_norm": 0.2251540424025956, "learning_rate": 0.00018025226549377648, "loss": 0.8136, "step": 1182 }, { "epoch": 0.22758753366679493, "grad_norm": 0.21130294784060627, "learning_rate": 0.0001802150753406913, "loss": 0.8573, "step": 1183 }, { "epoch": 0.2277799153520585, "grad_norm": 0.18413739935487672, "learning_rate": 0.000180177854045352, "loss": 0.7757, "step": 1184 }, { "epoch": 0.22797229703732205, "grad_norm": 0.2272497824313606, "learning_rate": 0.00018014060162220913, "loss": 0.8753, "step": 1185 }, { "epoch": 0.2281646787225856, "grad_norm": 0.21448637241255616, "learning_rate": 0.0001801033180857254, "loss": 0.8293, "step": 1186 }, { "epoch": 0.22835706040784917, "grad_norm": 0.1800385783777638, "learning_rate": 0.00018006600345037556, "loss": 0.8631, "step": 1187 }, { "epoch": 0.22854944209311273, "grad_norm": 0.1981354487866229, "learning_rate": 0.00018002865773064644, "loss": 0.8242, "step": 1188 }, { "epoch": 0.2287418237783763, "grad_norm": 0.18955874043015433, "learning_rate": 0.00017999128094103688, "loss": 0.8352, "step": 1189 }, { "epoch": 0.22893420546363985, "grad_norm": 0.18578046252181907, "learning_rate": 0.00017995387309605788, "loss": 0.836, "step": 1190 }, { "epoch": 0.22912658714890344, "grad_norm": 0.20758538973727853, "learning_rate": 0.0001799164342102325, "loss": 0.8395, "step": 1191 }, { "epoch": 0.229318968834167, "grad_norm": 0.20196884499822101, "learning_rate": 0.0001798789642980957, "loss": 0.838, "step": 1192 }, { "epoch": 0.22951135051943056, "grad_norm": 0.10067583848490018, "learning_rate": 0.00017984146337419467, "loss": 0.6231, "step": 1193 }, { "epoch": 0.22970373220469412, "grad_norm": 0.20630825545484535, "learning_rate": 0.00017980393145308857, "loss": 0.7621, "step": 1194 }, { "epoch": 0.22989611388995768, "grad_norm": 0.18626602553154611, "learning_rate": 0.00017976636854934852, "loss": 0.8642, "step": 1195 }, { "epoch": 0.23008849557522124, "grad_norm": 0.20462446123384298, "learning_rate": 0.00017972877467755776, "loss": 0.8526, "step": 1196 }, { "epoch": 0.2302808772604848, "grad_norm": 0.17503497179132185, "learning_rate": 0.00017969114985231152, "loss": 0.8128, "step": 1197 }, { "epoch": 0.23047325894574835, "grad_norm": 0.18885152581997622, "learning_rate": 0.00017965349408821709, "loss": 0.8851, "step": 1198 }, { "epoch": 0.23066564063101191, "grad_norm": 0.1833396189721438, "learning_rate": 0.00017961580739989365, "loss": 0.7571, "step": 1199 }, { "epoch": 0.2308580223162755, "grad_norm": 0.17700962581409185, "learning_rate": 0.0001795780898019726, "loss": 0.8556, "step": 1200 }, { "epoch": 0.23105040400153906, "grad_norm": 0.17488714230226698, "learning_rate": 0.00017954034130909707, "loss": 0.8751, "step": 1201 }, { "epoch": 0.23124278568680262, "grad_norm": 0.18063408365317485, "learning_rate": 0.00017950256193592244, "loss": 0.7859, "step": 1202 }, { "epoch": 0.23143516737206618, "grad_norm": 0.19718291833969406, "learning_rate": 0.00017946475169711588, "loss": 0.8366, "step": 1203 }, { "epoch": 0.23162754905732974, "grad_norm": 0.18845213991055698, "learning_rate": 0.00017942691060735666, "loss": 0.8009, "step": 1204 }, { "epoch": 0.2318199307425933, "grad_norm": 0.18615580607939153, "learning_rate": 0.00017938903868133598, "loss": 0.8297, "step": 1205 }, { "epoch": 0.23201231242785686, "grad_norm": 0.21763436232780026, "learning_rate": 0.00017935113593375707, "loss": 0.7871, "step": 1206 }, { "epoch": 0.23220469411312042, "grad_norm": 0.18833529453685147, "learning_rate": 0.00017931320237933503, "loss": 0.8132, "step": 1207 }, { "epoch": 0.232397075798384, "grad_norm": 0.17290179891239774, "learning_rate": 0.000179275238032797, "loss": 0.8799, "step": 1208 }, { "epoch": 0.23258945748364757, "grad_norm": 0.1992353997523004, "learning_rate": 0.00017923724290888206, "loss": 0.8277, "step": 1209 }, { "epoch": 0.23278183916891113, "grad_norm": 0.19824506135669845, "learning_rate": 0.00017919921702234118, "loss": 0.8363, "step": 1210 }, { "epoch": 0.23297422085417469, "grad_norm": 0.1992368508904255, "learning_rate": 0.0001791611603879374, "loss": 0.8253, "step": 1211 }, { "epoch": 0.23316660253943824, "grad_norm": 0.1965998612994148, "learning_rate": 0.00017912307302044548, "loss": 0.8643, "step": 1212 }, { "epoch": 0.2333589842247018, "grad_norm": 0.18588585626876203, "learning_rate": 0.00017908495493465236, "loss": 0.8188, "step": 1213 }, { "epoch": 0.23355136590996536, "grad_norm": 0.17634360653696368, "learning_rate": 0.00017904680614535675, "loss": 0.8908, "step": 1214 }, { "epoch": 0.23374374759522892, "grad_norm": 0.19788219590439685, "learning_rate": 0.00017900862666736934, "loss": 0.8072, "step": 1215 }, { "epoch": 0.2339361292804925, "grad_norm": 0.19987332577397027, "learning_rate": 0.0001789704165155127, "loss": 0.8433, "step": 1216 }, { "epoch": 0.23412851096575607, "grad_norm": 0.1919183117451716, "learning_rate": 0.00017893217570462134, "loss": 0.8163, "step": 1217 }, { "epoch": 0.23432089265101963, "grad_norm": 0.1822737941576912, "learning_rate": 0.00017889390424954166, "loss": 0.8257, "step": 1218 }, { "epoch": 0.2345132743362832, "grad_norm": 0.18147704411594257, "learning_rate": 0.00017885560216513197, "loss": 0.8346, "step": 1219 }, { "epoch": 0.23470565602154675, "grad_norm": 0.6100849411666639, "learning_rate": 0.00017881726946626243, "loss": 0.8375, "step": 1220 }, { "epoch": 0.2348980377068103, "grad_norm": 0.24145106600084554, "learning_rate": 0.00017877890616781512, "loss": 0.844, "step": 1221 }, { "epoch": 0.23509041939207387, "grad_norm": 0.18145477268950275, "learning_rate": 0.000178740512284684, "loss": 0.8474, "step": 1222 }, { "epoch": 0.23528280107733743, "grad_norm": 0.18162783946402236, "learning_rate": 0.0001787020878317749, "loss": 0.8317, "step": 1223 }, { "epoch": 0.235475182762601, "grad_norm": 0.17679648678121326, "learning_rate": 0.00017866363282400554, "loss": 0.8168, "step": 1224 }, { "epoch": 0.23566756444786457, "grad_norm": 0.19998815508490705, "learning_rate": 0.00017862514727630543, "loss": 0.8543, "step": 1225 }, { "epoch": 0.23585994613312813, "grad_norm": 0.18621908101988846, "learning_rate": 0.00017858663120361596, "loss": 0.883, "step": 1226 }, { "epoch": 0.2360523278183917, "grad_norm": 0.184857668119696, "learning_rate": 0.00017854808462089048, "loss": 0.8883, "step": 1227 }, { "epoch": 0.23624470950365525, "grad_norm": 0.1828817793516205, "learning_rate": 0.00017850950754309407, "loss": 0.8057, "step": 1228 }, { "epoch": 0.2364370911889188, "grad_norm": 0.18264036677274553, "learning_rate": 0.00017847089998520365, "loss": 0.8073, "step": 1229 }, { "epoch": 0.23662947287418237, "grad_norm": 0.20210559932256983, "learning_rate": 0.00017843226196220802, "loss": 0.8994, "step": 1230 }, { "epoch": 0.23682185455944593, "grad_norm": 0.1953412952255688, "learning_rate": 0.0001783935934891078, "loss": 0.81, "step": 1231 }, { "epoch": 0.2370142362447095, "grad_norm": 0.16634209472082243, "learning_rate": 0.00017835489458091543, "loss": 0.8606, "step": 1232 }, { "epoch": 0.23720661792997308, "grad_norm": 0.18906485740761123, "learning_rate": 0.00017831616525265513, "loss": 0.8225, "step": 1233 }, { "epoch": 0.23739899961523664, "grad_norm": 0.24446298388833718, "learning_rate": 0.00017827740551936295, "loss": 0.9044, "step": 1234 }, { "epoch": 0.2375913813005002, "grad_norm": 0.18350262968645165, "learning_rate": 0.00017823861539608684, "loss": 0.8602, "step": 1235 }, { "epoch": 0.23778376298576376, "grad_norm": 0.176638248003139, "learning_rate": 0.00017819979489788637, "loss": 0.8905, "step": 1236 }, { "epoch": 0.23797614467102732, "grad_norm": 0.21274999403738282, "learning_rate": 0.00017816094403983298, "loss": 0.7983, "step": 1237 }, { "epoch": 0.23816852635629088, "grad_norm": 0.1908508946045328, "learning_rate": 0.00017812206283701002, "loss": 0.8516, "step": 1238 }, { "epoch": 0.23836090804155444, "grad_norm": 0.20076528758180684, "learning_rate": 0.00017808315130451243, "loss": 0.8498, "step": 1239 }, { "epoch": 0.238553289726818, "grad_norm": 0.20848277427956174, "learning_rate": 0.000178044209457447, "loss": 0.802, "step": 1240 }, { "epoch": 0.23874567141208156, "grad_norm": 0.1855755632974751, "learning_rate": 0.00017800523731093234, "loss": 0.8906, "step": 1241 }, { "epoch": 0.23893805309734514, "grad_norm": 0.20810133919805296, "learning_rate": 0.00017796623488009874, "loss": 0.8431, "step": 1242 }, { "epoch": 0.2391304347826087, "grad_norm": 0.1794455818557291, "learning_rate": 0.00017792720218008827, "loss": 0.8517, "step": 1243 }, { "epoch": 0.23932281646787226, "grad_norm": 0.17829756557650056, "learning_rate": 0.00017788813922605488, "loss": 0.8163, "step": 1244 }, { "epoch": 0.23951519815313582, "grad_norm": 0.18715148236817628, "learning_rate": 0.000177849046033164, "loss": 0.8193, "step": 1245 }, { "epoch": 0.23970757983839938, "grad_norm": 0.18699089804925137, "learning_rate": 0.00017780992261659306, "loss": 0.8207, "step": 1246 }, { "epoch": 0.23989996152366294, "grad_norm": 0.19488929845632835, "learning_rate": 0.00017777076899153112, "loss": 0.8404, "step": 1247 }, { "epoch": 0.2400923432089265, "grad_norm": 0.1895779889712498, "learning_rate": 0.00017773158517317888, "loss": 0.8074, "step": 1248 }, { "epoch": 0.24028472489419006, "grad_norm": 0.17661430548118687, "learning_rate": 0.00017769237117674892, "loss": 0.8158, "step": 1249 }, { "epoch": 0.24047710657945365, "grad_norm": 0.19333068344712692, "learning_rate": 0.00017765312701746545, "loss": 0.8407, "step": 1250 }, { "epoch": 0.2406694882647172, "grad_norm": 0.21816640366072307, "learning_rate": 0.00017761385271056437, "loss": 0.8771, "step": 1251 }, { "epoch": 0.24086186994998077, "grad_norm": 0.19279028047376967, "learning_rate": 0.0001775745482712934, "loss": 0.8735, "step": 1252 }, { "epoch": 0.24105425163524433, "grad_norm": 0.11807050952489027, "learning_rate": 0.00017753521371491175, "loss": 0.6762, "step": 1253 }, { "epoch": 0.24124663332050789, "grad_norm": 0.19744328321959861, "learning_rate": 0.00017749584905669058, "loss": 0.7568, "step": 1254 }, { "epoch": 0.24143901500577145, "grad_norm": 0.21499128003083787, "learning_rate": 0.00017745645431191252, "loss": 0.8615, "step": 1255 }, { "epoch": 0.241631396691035, "grad_norm": 0.20198100772490601, "learning_rate": 0.00017741702949587194, "loss": 0.8575, "step": 1256 }, { "epoch": 0.24182377837629856, "grad_norm": 0.18839809505702373, "learning_rate": 0.00017737757462387506, "loss": 0.8692, "step": 1257 }, { "epoch": 0.24201616006156215, "grad_norm": 0.18012523668532518, "learning_rate": 0.00017733808971123947, "loss": 0.8832, "step": 1258 }, { "epoch": 0.2422085417468257, "grad_norm": 0.20163669497494732, "learning_rate": 0.00017729857477329463, "loss": 0.8161, "step": 1259 }, { "epoch": 0.24240092343208927, "grad_norm": 0.20921651088460896, "learning_rate": 0.00017725902982538162, "loss": 0.8161, "step": 1260 }, { "epoch": 0.24259330511735283, "grad_norm": 0.17365153697567903, "learning_rate": 0.0001772194548828531, "loss": 0.8222, "step": 1261 }, { "epoch": 0.2427856868026164, "grad_norm": 0.18614001164681854, "learning_rate": 0.00017717984996107344, "loss": 0.7988, "step": 1262 }, { "epoch": 0.24297806848787995, "grad_norm": 0.09805758060307666, "learning_rate": 0.00017714021507541866, "loss": 0.605, "step": 1263 }, { "epoch": 0.2431704501731435, "grad_norm": 0.18594046719396431, "learning_rate": 0.00017710055024127637, "loss": 0.8078, "step": 1264 }, { "epoch": 0.24336283185840707, "grad_norm": 0.20770554879056102, "learning_rate": 0.00017706085547404584, "loss": 0.8297, "step": 1265 }, { "epoch": 0.24355521354367063, "grad_norm": 0.17629805931098028, "learning_rate": 0.0001770211307891379, "loss": 0.8561, "step": 1266 }, { "epoch": 0.24374759522893422, "grad_norm": 0.18030057536114935, "learning_rate": 0.0001769813762019751, "loss": 0.8026, "step": 1267 }, { "epoch": 0.24393997691419778, "grad_norm": 0.20527110348958405, "learning_rate": 0.00017694159172799152, "loss": 0.8027, "step": 1268 }, { "epoch": 0.24413235859946134, "grad_norm": 0.19270948317148823, "learning_rate": 0.00017690177738263283, "loss": 0.845, "step": 1269 }, { "epoch": 0.2443247402847249, "grad_norm": 0.19498999121495433, "learning_rate": 0.00017686193318135635, "loss": 0.8765, "step": 1270 }, { "epoch": 0.24451712196998845, "grad_norm": 0.1692676994403039, "learning_rate": 0.00017682205913963103, "loss": 0.8408, "step": 1271 }, { "epoch": 0.24470950365525201, "grad_norm": 0.1884432678245104, "learning_rate": 0.00017678215527293722, "loss": 0.8409, "step": 1272 }, { "epoch": 0.24490188534051557, "grad_norm": 0.19223886019913167, "learning_rate": 0.00017674222159676708, "loss": 0.8708, "step": 1273 }, { "epoch": 0.24509426702577913, "grad_norm": 0.18804362913350048, "learning_rate": 0.0001767022581266242, "loss": 0.8838, "step": 1274 }, { "epoch": 0.24528664871104272, "grad_norm": 0.1806963234068854, "learning_rate": 0.0001766622648780238, "loss": 0.8995, "step": 1275 }, { "epoch": 0.24547903039630628, "grad_norm": 0.20609538040480252, "learning_rate": 0.00017662224186649258, "loss": 0.8025, "step": 1276 }, { "epoch": 0.24567141208156984, "grad_norm": 0.19358451567279245, "learning_rate": 0.00017658218910756893, "loss": 0.8622, "step": 1277 }, { "epoch": 0.2458637937668334, "grad_norm": 0.18351040280545136, "learning_rate": 0.00017654210661680262, "loss": 0.8437, "step": 1278 }, { "epoch": 0.24605617545209696, "grad_norm": 0.18007485763867395, "learning_rate": 0.0001765019944097551, "loss": 0.8749, "step": 1279 }, { "epoch": 0.24624855713736052, "grad_norm": 0.10389315899783678, "learning_rate": 0.00017646185250199935, "loss": 0.6285, "step": 1280 }, { "epoch": 0.24644093882262408, "grad_norm": 0.11162234490716226, "learning_rate": 0.00017642168090911973, "loss": 0.6552, "step": 1281 }, { "epoch": 0.24663332050788764, "grad_norm": 0.20024641070705135, "learning_rate": 0.00017638147964671238, "loss": 0.8756, "step": 1282 }, { "epoch": 0.24682570219315123, "grad_norm": 0.19685862348502461, "learning_rate": 0.0001763412487303847, "loss": 0.8335, "step": 1283 }, { "epoch": 0.24701808387841478, "grad_norm": 0.18555949683481512, "learning_rate": 0.00017630098817575577, "loss": 0.8724, "step": 1284 }, { "epoch": 0.24721046556367834, "grad_norm": 0.1955777384835772, "learning_rate": 0.0001762606979984561, "loss": 0.8334, "step": 1285 }, { "epoch": 0.2474028472489419, "grad_norm": 0.2021154025667084, "learning_rate": 0.00017622037821412776, "loss": 0.8737, "step": 1286 }, { "epoch": 0.24759522893420546, "grad_norm": 0.20824555629149147, "learning_rate": 0.00017618002883842423, "loss": 0.8133, "step": 1287 }, { "epoch": 0.24778761061946902, "grad_norm": 0.18940507454205685, "learning_rate": 0.00017613964988701057, "loss": 0.8395, "step": 1288 }, { "epoch": 0.24797999230473258, "grad_norm": 0.17935815614656953, "learning_rate": 0.00017609924137556324, "loss": 0.8741, "step": 1289 }, { "epoch": 0.24817237398999614, "grad_norm": 0.21093414915967196, "learning_rate": 0.00017605880331977023, "loss": 0.8018, "step": 1290 }, { "epoch": 0.2483647556752597, "grad_norm": 0.18183194611385192, "learning_rate": 0.000176018335735331, "loss": 0.8601, "step": 1291 }, { "epoch": 0.2485571373605233, "grad_norm": 0.18094034984251467, "learning_rate": 0.00017597783863795643, "loss": 0.8713, "step": 1292 }, { "epoch": 0.24874951904578685, "grad_norm": 0.24571073665857912, "learning_rate": 0.00017593731204336893, "loss": 0.8241, "step": 1293 }, { "epoch": 0.2489419007310504, "grad_norm": 0.18057541089081636, "learning_rate": 0.0001758967559673023, "loss": 0.8633, "step": 1294 }, { "epoch": 0.24913428241631397, "grad_norm": 0.17084154254253334, "learning_rate": 0.0001758561704255018, "loss": 0.8626, "step": 1295 }, { "epoch": 0.24932666410157753, "grad_norm": 0.176455318725715, "learning_rate": 0.00017581555543372412, "loss": 0.8586, "step": 1296 }, { "epoch": 0.2495190457868411, "grad_norm": 0.23970417618061732, "learning_rate": 0.00017577491100773744, "loss": 0.8365, "step": 1297 }, { "epoch": 0.24971142747210465, "grad_norm": 0.18782028737890225, "learning_rate": 0.00017573423716332127, "loss": 0.7943, "step": 1298 }, { "epoch": 0.2499038091573682, "grad_norm": 0.1888722456367085, "learning_rate": 0.00017569353391626663, "loss": 0.813, "step": 1299 }, { "epoch": 0.2500961908426318, "grad_norm": 0.19305640327131135, "learning_rate": 0.00017565280128237595, "loss": 0.839, "step": 1300 }, { "epoch": 0.2502885725278953, "grad_norm": 0.10174212273912497, "learning_rate": 0.000175612039277463, "loss": 0.6358, "step": 1301 }, { "epoch": 0.2504809542131589, "grad_norm": 0.1985349348105587, "learning_rate": 0.00017557124791735298, "loss": 0.8476, "step": 1302 }, { "epoch": 0.25067333589842244, "grad_norm": 0.19080226695459274, "learning_rate": 0.00017553042721788254, "loss": 0.8196, "step": 1303 }, { "epoch": 0.25086571758368603, "grad_norm": 0.09995207437460432, "learning_rate": 0.00017548957719489968, "loss": 0.6494, "step": 1304 }, { "epoch": 0.2510580992689496, "grad_norm": 0.21281305144877377, "learning_rate": 0.00017544869786426372, "loss": 0.8805, "step": 1305 }, { "epoch": 0.25125048095421315, "grad_norm": 0.1876493128137456, "learning_rate": 0.00017540778924184553, "loss": 0.82, "step": 1306 }, { "epoch": 0.25144286263947674, "grad_norm": 0.19691245442278918, "learning_rate": 0.00017536685134352716, "loss": 0.8413, "step": 1307 }, { "epoch": 0.25163524432474027, "grad_norm": 0.18871873647062345, "learning_rate": 0.00017532588418520215, "loss": 0.8812, "step": 1308 }, { "epoch": 0.25182762601000386, "grad_norm": 0.21263217457422248, "learning_rate": 0.00017528488778277535, "loss": 0.9012, "step": 1309 }, { "epoch": 0.2520200076952674, "grad_norm": 0.21011455297493833, "learning_rate": 0.000175243862152163, "loss": 0.7996, "step": 1310 }, { "epoch": 0.252212389380531, "grad_norm": 0.1881848687464559, "learning_rate": 0.0001752028073092926, "loss": 0.791, "step": 1311 }, { "epoch": 0.2524047710657945, "grad_norm": 0.1994188285206938, "learning_rate": 0.0001751617232701031, "loss": 0.8595, "step": 1312 }, { "epoch": 0.2525971527510581, "grad_norm": 0.18680460524240303, "learning_rate": 0.0001751206100505448, "loss": 0.8648, "step": 1313 }, { "epoch": 0.2527895344363217, "grad_norm": 0.1980939830458383, "learning_rate": 0.00017507946766657912, "loss": 0.8904, "step": 1314 }, { "epoch": 0.2529819161215852, "grad_norm": 0.2229159722819279, "learning_rate": 0.00017503829613417905, "loss": 0.8152, "step": 1315 }, { "epoch": 0.2531742978068488, "grad_norm": 0.19560994385159267, "learning_rate": 0.0001749970954693288, "loss": 0.8084, "step": 1316 }, { "epoch": 0.25336667949211233, "grad_norm": 0.19554323228955559, "learning_rate": 0.00017495586568802384, "loss": 0.8271, "step": 1317 }, { "epoch": 0.2535590611773759, "grad_norm": 0.19984736094394312, "learning_rate": 0.00017491460680627103, "loss": 0.8427, "step": 1318 }, { "epoch": 0.25375144286263945, "grad_norm": 0.19947609893826124, "learning_rate": 0.00017487331884008845, "loss": 0.8205, "step": 1319 }, { "epoch": 0.25394382454790304, "grad_norm": 0.3340111343920411, "learning_rate": 0.00017483200180550554, "loss": 0.9064, "step": 1320 }, { "epoch": 0.25413620623316663, "grad_norm": 0.18475872152303008, "learning_rate": 0.00017479065571856302, "loss": 0.8326, "step": 1321 }, { "epoch": 0.25432858791843016, "grad_norm": 0.11501265454743305, "learning_rate": 0.0001747492805953128, "loss": 0.6121, "step": 1322 }, { "epoch": 0.25452096960369375, "grad_norm": 0.19915548075370823, "learning_rate": 0.00017470787645181815, "loss": 0.82, "step": 1323 }, { "epoch": 0.2547133512889573, "grad_norm": 0.20668555638597974, "learning_rate": 0.00017466644330415364, "loss": 0.8607, "step": 1324 }, { "epoch": 0.25490573297422087, "grad_norm": 0.1820175096769794, "learning_rate": 0.00017462498116840495, "loss": 0.8099, "step": 1325 }, { "epoch": 0.2550981146594844, "grad_norm": 0.2054536834037318, "learning_rate": 0.00017458349006066917, "loss": 0.8909, "step": 1326 }, { "epoch": 0.255290496344748, "grad_norm": 0.22152984867544828, "learning_rate": 0.00017454196999705456, "loss": 0.8188, "step": 1327 }, { "epoch": 0.2554828780300115, "grad_norm": 0.1883703561790491, "learning_rate": 0.00017450042099368066, "loss": 0.8682, "step": 1328 }, { "epoch": 0.2556752597152751, "grad_norm": 0.18120561498949583, "learning_rate": 0.0001744588430666782, "loss": 0.8261, "step": 1329 }, { "epoch": 0.2558676414005387, "grad_norm": 0.20268420044252083, "learning_rate": 0.00017441723623218915, "loss": 0.8075, "step": 1330 }, { "epoch": 0.2560600230858022, "grad_norm": 0.18840349708508194, "learning_rate": 0.00017437560050636677, "loss": 0.8105, "step": 1331 }, { "epoch": 0.2562524047710658, "grad_norm": 0.2051860983431174, "learning_rate": 0.0001743339359053754, "loss": 0.8095, "step": 1332 }, { "epoch": 0.25644478645632934, "grad_norm": 0.19024760110688332, "learning_rate": 0.00017429224244539077, "loss": 0.8365, "step": 1333 }, { "epoch": 0.25663716814159293, "grad_norm": 0.1999408866220327, "learning_rate": 0.00017425052014259963, "loss": 0.879, "step": 1334 }, { "epoch": 0.25682954982685646, "grad_norm": 0.1944222246894097, "learning_rate": 0.00017420876901320007, "loss": 0.8826, "step": 1335 }, { "epoch": 0.25702193151212005, "grad_norm": 0.22237679656767917, "learning_rate": 0.00017416698907340126, "loss": 0.8619, "step": 1336 }, { "epoch": 0.2572143131973836, "grad_norm": 0.19655347190749425, "learning_rate": 0.00017412518033942372, "loss": 0.8945, "step": 1337 }, { "epoch": 0.25740669488264717, "grad_norm": 0.217032468820287, "learning_rate": 0.0001740833428274989, "loss": 0.8018, "step": 1338 }, { "epoch": 0.25759907656791076, "grad_norm": 0.18409661178499903, "learning_rate": 0.00017404147655386963, "loss": 0.8491, "step": 1339 }, { "epoch": 0.2577914582531743, "grad_norm": 0.21490763616484668, "learning_rate": 0.0001739995815347899, "loss": 0.8732, "step": 1340 }, { "epoch": 0.2579838399384379, "grad_norm": 0.19949283171565052, "learning_rate": 0.0001739576577865247, "loss": 0.8693, "step": 1341 }, { "epoch": 0.2581762216237014, "grad_norm": 0.19696113386536, "learning_rate": 0.0001739157053253503, "loss": 0.8403, "step": 1342 }, { "epoch": 0.258368603308965, "grad_norm": 0.19761100836604129, "learning_rate": 0.0001738737241675541, "loss": 0.8227, "step": 1343 }, { "epoch": 0.2585609849942285, "grad_norm": 0.18035254075682186, "learning_rate": 0.00017383171432943466, "loss": 0.8583, "step": 1344 }, { "epoch": 0.2587533666794921, "grad_norm": 0.18778647148081018, "learning_rate": 0.00017378967582730158, "loss": 0.8409, "step": 1345 }, { "epoch": 0.2589457483647557, "grad_norm": 0.18935130558934177, "learning_rate": 0.00017374760867747572, "loss": 0.8825, "step": 1346 }, { "epoch": 0.25913813005001923, "grad_norm": 0.21364543250553464, "learning_rate": 0.00017370551289628896, "loss": 0.8112, "step": 1347 }, { "epoch": 0.2593305117352828, "grad_norm": 0.20217200369617125, "learning_rate": 0.00017366338850008432, "loss": 0.8276, "step": 1348 }, { "epoch": 0.25952289342054635, "grad_norm": 0.22830273626297712, "learning_rate": 0.000173621235505216, "loss": 0.8186, "step": 1349 }, { "epoch": 0.25971527510580994, "grad_norm": 0.19762601180219233, "learning_rate": 0.00017357905392804916, "loss": 0.9078, "step": 1350 }, { "epoch": 0.25990765679107347, "grad_norm": 0.17383344734806985, "learning_rate": 0.0001735368437849602, "loss": 0.8457, "step": 1351 }, { "epoch": 0.26010003847633706, "grad_norm": 0.13859084175434988, "learning_rate": 0.00017349460509233654, "loss": 0.6571, "step": 1352 }, { "epoch": 0.2602924201616006, "grad_norm": 0.16906498069649567, "learning_rate": 0.0001734523378665767, "loss": 0.8343, "step": 1353 }, { "epoch": 0.2604848018468642, "grad_norm": 0.1890371714803512, "learning_rate": 0.00017341004212409025, "loss": 0.8937, "step": 1354 }, { "epoch": 0.26067718353212777, "grad_norm": 0.19497884218363723, "learning_rate": 0.00017336771788129785, "loss": 0.8054, "step": 1355 }, { "epoch": 0.2608695652173913, "grad_norm": 0.1823888505588386, "learning_rate": 0.00017332536515463125, "loss": 0.8666, "step": 1356 }, { "epoch": 0.2610619469026549, "grad_norm": 0.1971723693192003, "learning_rate": 0.00017328298396053325, "loss": 0.8404, "step": 1357 }, { "epoch": 0.2612543285879184, "grad_norm": 0.19859666947827256, "learning_rate": 0.00017324057431545765, "loss": 0.8672, "step": 1358 }, { "epoch": 0.261446710273182, "grad_norm": 0.1980304304547438, "learning_rate": 0.00017319813623586935, "loss": 0.8425, "step": 1359 }, { "epoch": 0.26163909195844554, "grad_norm": 0.19297117934434277, "learning_rate": 0.0001731556697382443, "loss": 0.8856, "step": 1360 }, { "epoch": 0.2618314736437091, "grad_norm": 0.19535309122018713, "learning_rate": 0.00017311317483906943, "loss": 0.8775, "step": 1361 }, { "epoch": 0.26202385532897265, "grad_norm": 0.16605831730832354, "learning_rate": 0.00017307065155484273, "loss": 0.8608, "step": 1362 }, { "epoch": 0.26221623701423624, "grad_norm": 0.181804758728873, "learning_rate": 0.00017302809990207317, "loss": 0.8405, "step": 1363 }, { "epoch": 0.26240861869949983, "grad_norm": 0.1815787686594028, "learning_rate": 0.00017298551989728084, "loss": 0.8382, "step": 1364 }, { "epoch": 0.26260100038476336, "grad_norm": 0.17783679087459875, "learning_rate": 0.0001729429115569967, "loss": 0.8231, "step": 1365 }, { "epoch": 0.26279338207002695, "grad_norm": 0.16112605446909314, "learning_rate": 0.00017290027489776278, "loss": 0.8065, "step": 1366 }, { "epoch": 0.2629857637552905, "grad_norm": 0.17830174970922902, "learning_rate": 0.00017285760993613215, "loss": 0.8517, "step": 1367 }, { "epoch": 0.26317814544055407, "grad_norm": 0.18485150543005896, "learning_rate": 0.00017281491668866874, "loss": 0.8248, "step": 1368 }, { "epoch": 0.2633705271258176, "grad_norm": 0.20158952136025315, "learning_rate": 0.0001727721951719476, "loss": 0.8462, "step": 1369 }, { "epoch": 0.2635629088110812, "grad_norm": 0.19649797569334887, "learning_rate": 0.0001727294454025547, "loss": 0.8559, "step": 1370 }, { "epoch": 0.2637552904963448, "grad_norm": 0.2008414966709754, "learning_rate": 0.0001726866673970869, "loss": 0.8747, "step": 1371 }, { "epoch": 0.2639476721816083, "grad_norm": 0.18839790372615756, "learning_rate": 0.00017264386117215216, "loss": 0.8581, "step": 1372 }, { "epoch": 0.2641400538668719, "grad_norm": 0.20387888859695716, "learning_rate": 0.00017260102674436932, "loss": 0.7957, "step": 1373 }, { "epoch": 0.2643324355521354, "grad_norm": 0.19167424111439135, "learning_rate": 0.0001725581641303682, "loss": 0.8018, "step": 1374 }, { "epoch": 0.264524817237399, "grad_norm": 0.11541887620849418, "learning_rate": 0.00017251527334678944, "loss": 0.6002, "step": 1375 }, { "epoch": 0.26471719892266254, "grad_norm": 0.183467403938036, "learning_rate": 0.00017247235441028485, "loss": 0.8746, "step": 1376 }, { "epoch": 0.26490958060792613, "grad_norm": 0.19200637249103833, "learning_rate": 0.00017242940733751695, "loss": 0.8274, "step": 1377 }, { "epoch": 0.26510196229318966, "grad_norm": 0.1869770584055729, "learning_rate": 0.00017238643214515932, "loss": 0.8575, "step": 1378 }, { "epoch": 0.26529434397845325, "grad_norm": 0.18303240518583785, "learning_rate": 0.00017234342884989642, "loss": 0.7948, "step": 1379 }, { "epoch": 0.26548672566371684, "grad_norm": 0.2012802004331186, "learning_rate": 0.00017230039746842353, "loss": 0.8791, "step": 1380 }, { "epoch": 0.26567910734898037, "grad_norm": 0.19371713876796873, "learning_rate": 0.000172257338017447, "loss": 0.8359, "step": 1381 }, { "epoch": 0.26587148903424396, "grad_norm": 0.19482751948222543, "learning_rate": 0.00017221425051368395, "loss": 0.8651, "step": 1382 }, { "epoch": 0.2660638707195075, "grad_norm": 0.21139718076960243, "learning_rate": 0.00017217113497386242, "loss": 0.8396, "step": 1383 }, { "epoch": 0.2662562524047711, "grad_norm": 0.17795546457160305, "learning_rate": 0.0001721279914147214, "loss": 0.8407, "step": 1384 }, { "epoch": 0.2664486340900346, "grad_norm": 0.19074668573372022, "learning_rate": 0.00017208481985301065, "loss": 0.8558, "step": 1385 }, { "epoch": 0.2666410157752982, "grad_norm": 0.17846091071255443, "learning_rate": 0.00017204162030549092, "loss": 0.8664, "step": 1386 }, { "epoch": 0.2668333974605617, "grad_norm": 0.1678359700519566, "learning_rate": 0.00017199839278893368, "loss": 0.8043, "step": 1387 }, { "epoch": 0.2670257791458253, "grad_norm": 0.20407133049543308, "learning_rate": 0.0001719551373201214, "loss": 0.7763, "step": 1388 }, { "epoch": 0.2672181608310889, "grad_norm": 0.20909369489845894, "learning_rate": 0.00017191185391584733, "loss": 0.851, "step": 1389 }, { "epoch": 0.26741054251635243, "grad_norm": 0.19782240364663756, "learning_rate": 0.00017186854259291557, "loss": 0.8981, "step": 1390 }, { "epoch": 0.267602924201616, "grad_norm": 0.19054128527745345, "learning_rate": 0.00017182520336814106, "loss": 0.8318, "step": 1391 }, { "epoch": 0.26779530588687955, "grad_norm": 0.16506301757127223, "learning_rate": 0.00017178183625834957, "loss": 0.8031, "step": 1392 }, { "epoch": 0.26798768757214314, "grad_norm": 0.19958375767445605, "learning_rate": 0.00017173844128037775, "loss": 0.8303, "step": 1393 }, { "epoch": 0.2681800692574067, "grad_norm": 0.19187397297632258, "learning_rate": 0.000171695018451073, "loss": 0.8676, "step": 1394 }, { "epoch": 0.26837245094267026, "grad_norm": 0.17161155690194696, "learning_rate": 0.00017165156778729354, "loss": 0.8198, "step": 1395 }, { "epoch": 0.26856483262793385, "grad_norm": 0.16700523755313834, "learning_rate": 0.00017160808930590844, "loss": 0.8849, "step": 1396 }, { "epoch": 0.2687572143131974, "grad_norm": 0.18973109679577985, "learning_rate": 0.00017156458302379753, "loss": 0.8562, "step": 1397 }, { "epoch": 0.26894959599846097, "grad_norm": 0.18990365735384307, "learning_rate": 0.00017152104895785147, "loss": 0.7953, "step": 1398 }, { "epoch": 0.2691419776837245, "grad_norm": 0.18028671551319017, "learning_rate": 0.00017147748712497164, "loss": 0.834, "step": 1399 }, { "epoch": 0.2693343593689881, "grad_norm": 0.16689495994835765, "learning_rate": 0.00017143389754207026, "loss": 0.8609, "step": 1400 }, { "epoch": 0.2695267410542516, "grad_norm": 0.19019918940688116, "learning_rate": 0.00017139028022607032, "loss": 0.8552, "step": 1401 }, { "epoch": 0.2697191227395152, "grad_norm": 0.19561886544617557, "learning_rate": 0.00017134663519390557, "loss": 0.8682, "step": 1402 }, { "epoch": 0.26991150442477874, "grad_norm": 0.18268730459717678, "learning_rate": 0.00017130296246252046, "loss": 0.8405, "step": 1403 }, { "epoch": 0.2701038861100423, "grad_norm": 0.16875372176631362, "learning_rate": 0.00017125926204887034, "loss": 0.8416, "step": 1404 }, { "epoch": 0.2702962677953059, "grad_norm": 0.20699644054427085, "learning_rate": 0.00017121553396992113, "loss": 0.8363, "step": 1405 }, { "epoch": 0.27048864948056944, "grad_norm": 0.19045321787595068, "learning_rate": 0.00017117177824264962, "loss": 0.8469, "step": 1406 }, { "epoch": 0.27068103116583303, "grad_norm": 0.10189055368529247, "learning_rate": 0.00017112799488404326, "loss": 0.6262, "step": 1407 }, { "epoch": 0.27087341285109656, "grad_norm": 0.17631701266369326, "learning_rate": 0.00017108418391110033, "loss": 0.8059, "step": 1408 }, { "epoch": 0.27106579453636015, "grad_norm": 0.18686048259560156, "learning_rate": 0.00017104034534082968, "loss": 0.8657, "step": 1409 }, { "epoch": 0.2712581762216237, "grad_norm": 0.17017954725253173, "learning_rate": 0.00017099647919025096, "loss": 0.8036, "step": 1410 }, { "epoch": 0.27145055790688727, "grad_norm": 0.1760546782166992, "learning_rate": 0.00017095258547639454, "loss": 0.8448, "step": 1411 }, { "epoch": 0.2716429395921508, "grad_norm": 0.20523276454466557, "learning_rate": 0.0001709086642163015, "loss": 0.8168, "step": 1412 }, { "epoch": 0.2718353212774144, "grad_norm": 0.1775378828269356, "learning_rate": 0.00017086471542702354, "loss": 0.8464, "step": 1413 }, { "epoch": 0.272027702962678, "grad_norm": 0.16559720250101012, "learning_rate": 0.00017082073912562309, "loss": 0.8151, "step": 1414 }, { "epoch": 0.2722200846479415, "grad_norm": 0.19845038898342604, "learning_rate": 0.0001707767353291733, "loss": 0.7986, "step": 1415 }, { "epoch": 0.2724124663332051, "grad_norm": 0.20259598121500896, "learning_rate": 0.00017073270405475795, "loss": 0.8246, "step": 1416 }, { "epoch": 0.2726048480184686, "grad_norm": 0.20875248246355388, "learning_rate": 0.00017068864531947147, "loss": 0.7981, "step": 1417 }, { "epoch": 0.2727972297037322, "grad_norm": 0.2117256331968699, "learning_rate": 0.00017064455914041896, "loss": 0.8578, "step": 1418 }, { "epoch": 0.27298961138899575, "grad_norm": 0.18194469260016244, "learning_rate": 0.00017060044553471627, "loss": 0.8116, "step": 1419 }, { "epoch": 0.27318199307425933, "grad_norm": 0.19038734083510678, "learning_rate": 0.0001705563045194898, "loss": 0.842, "step": 1420 }, { "epoch": 0.2733743747595229, "grad_norm": 0.2109383245203454, "learning_rate": 0.00017051213611187654, "loss": 0.8529, "step": 1421 }, { "epoch": 0.27356675644478645, "grad_norm": 0.20640193986902913, "learning_rate": 0.0001704679403290243, "loss": 0.7966, "step": 1422 }, { "epoch": 0.27375913813005004, "grad_norm": 0.1749807000867148, "learning_rate": 0.0001704237171880913, "loss": 0.849, "step": 1423 }, { "epoch": 0.27395151981531357, "grad_norm": 0.1775214940634306, "learning_rate": 0.00017037946670624653, "loss": 0.875, "step": 1424 }, { "epoch": 0.27414390150057716, "grad_norm": 0.20675875249979236, "learning_rate": 0.00017033518890066955, "loss": 0.8176, "step": 1425 }, { "epoch": 0.2743362831858407, "grad_norm": 0.18498904938930935, "learning_rate": 0.00017029088378855053, "loss": 0.8267, "step": 1426 }, { "epoch": 0.2745286648711043, "grad_norm": 0.18193134071586956, "learning_rate": 0.00017024655138709024, "loss": 0.8471, "step": 1427 }, { "epoch": 0.2747210465563678, "grad_norm": 0.19184116046858554, "learning_rate": 0.00017020219171350002, "loss": 0.8081, "step": 1428 }, { "epoch": 0.2749134282416314, "grad_norm": 0.10863676532904212, "learning_rate": 0.00017015780478500187, "loss": 0.6667, "step": 1429 }, { "epoch": 0.275105809926895, "grad_norm": 0.2103939299855128, "learning_rate": 0.0001701133906188283, "loss": 0.7914, "step": 1430 }, { "epoch": 0.2752981916121585, "grad_norm": 0.19405052853584073, "learning_rate": 0.00017006894923222242, "loss": 0.812, "step": 1431 }, { "epoch": 0.2754905732974221, "grad_norm": 0.1918786468978619, "learning_rate": 0.0001700244806424379, "loss": 0.858, "step": 1432 }, { "epoch": 0.27568295498268564, "grad_norm": 0.18590288966291382, "learning_rate": 0.00016997998486673895, "loss": 0.797, "step": 1433 }, { "epoch": 0.2758753366679492, "grad_norm": 0.19154436857426071, "learning_rate": 0.00016993546192240041, "loss": 0.8476, "step": 1434 }, { "epoch": 0.27606771835321275, "grad_norm": 0.20661290255764955, "learning_rate": 0.00016989091182670763, "loss": 0.8266, "step": 1435 }, { "epoch": 0.27626010003847634, "grad_norm": 0.16088044112632055, "learning_rate": 0.00016984633459695645, "loss": 0.8326, "step": 1436 }, { "epoch": 0.2764524817237399, "grad_norm": 0.17914925442745128, "learning_rate": 0.00016980173025045328, "loss": 0.7758, "step": 1437 }, { "epoch": 0.27664486340900346, "grad_norm": 0.19408252803823953, "learning_rate": 0.0001697570988045151, "loss": 0.8771, "step": 1438 }, { "epoch": 0.27683724509426705, "grad_norm": 0.19744709521269357, "learning_rate": 0.00016971244027646937, "loss": 0.8761, "step": 1439 }, { "epoch": 0.2770296267795306, "grad_norm": 0.21636208769446294, "learning_rate": 0.000169667754683654, "loss": 0.8125, "step": 1440 }, { "epoch": 0.27722200846479417, "grad_norm": 0.19495507704059623, "learning_rate": 0.0001696230420434176, "loss": 0.807, "step": 1441 }, { "epoch": 0.2774143901500577, "grad_norm": 0.1883455283466202, "learning_rate": 0.00016957830237311904, "loss": 0.7918, "step": 1442 }, { "epoch": 0.2776067718353213, "grad_norm": 0.18248300290069322, "learning_rate": 0.00016953353569012784, "loss": 0.8307, "step": 1443 }, { "epoch": 0.2777991535205848, "grad_norm": 0.21088474889677022, "learning_rate": 0.00016948874201182403, "loss": 0.8138, "step": 1444 }, { "epoch": 0.2779915352058484, "grad_norm": 0.18625938834634082, "learning_rate": 0.00016944392135559796, "loss": 0.8297, "step": 1445 }, { "epoch": 0.278183916891112, "grad_norm": 0.2306343435138208, "learning_rate": 0.00016939907373885062, "loss": 0.8313, "step": 1446 }, { "epoch": 0.2783762985763755, "grad_norm": 0.20303227759301473, "learning_rate": 0.00016935419917899335, "loss": 0.8155, "step": 1447 }, { "epoch": 0.2785686802616391, "grad_norm": 0.20295628102812946, "learning_rate": 0.00016930929769344805, "loss": 0.8362, "step": 1448 }, { "epoch": 0.27876106194690264, "grad_norm": 0.18085389422713538, "learning_rate": 0.00016926436929964699, "loss": 0.8306, "step": 1449 }, { "epoch": 0.27895344363216623, "grad_norm": 0.19149369589761367, "learning_rate": 0.0001692194140150329, "loss": 0.8498, "step": 1450 }, { "epoch": 0.27914582531742976, "grad_norm": 0.17961656600131107, "learning_rate": 0.000169174431857059, "loss": 0.8511, "step": 1451 }, { "epoch": 0.27933820700269335, "grad_norm": 0.09991868430791222, "learning_rate": 0.00016912942284318894, "loss": 0.6347, "step": 1452 }, { "epoch": 0.2795305886879569, "grad_norm": 0.1933374404369002, "learning_rate": 0.00016908438699089672, "loss": 0.8481, "step": 1453 }, { "epoch": 0.27972297037322047, "grad_norm": 0.18816640173138324, "learning_rate": 0.0001690393243176668, "loss": 0.8012, "step": 1454 }, { "epoch": 0.27991535205848406, "grad_norm": 0.19042228546068649, "learning_rate": 0.00016899423484099412, "loss": 0.7434, "step": 1455 }, { "epoch": 0.2801077337437476, "grad_norm": 0.20440574154911728, "learning_rate": 0.00016894911857838393, "loss": 0.8121, "step": 1456 }, { "epoch": 0.2803001154290112, "grad_norm": 0.1866088594259107, "learning_rate": 0.00016890397554735192, "loss": 0.8576, "step": 1457 }, { "epoch": 0.2804924971142747, "grad_norm": 0.19684678583811305, "learning_rate": 0.00016885880576542417, "loss": 0.78, "step": 1458 }, { "epoch": 0.2806848787995383, "grad_norm": 0.18005153767239387, "learning_rate": 0.00016881360925013713, "loss": 0.8433, "step": 1459 }, { "epoch": 0.2808772604848018, "grad_norm": 0.18215359572208184, "learning_rate": 0.00016876838601903767, "loss": 0.8393, "step": 1460 }, { "epoch": 0.2810696421700654, "grad_norm": 0.18700610657926728, "learning_rate": 0.00016872313608968294, "loss": 0.8464, "step": 1461 }, { "epoch": 0.28126202385532895, "grad_norm": 0.20659555989102993, "learning_rate": 0.00016867785947964064, "loss": 0.8551, "step": 1462 }, { "epoch": 0.28145440554059253, "grad_norm": 0.21769034686307193, "learning_rate": 0.00016863255620648863, "loss": 0.8378, "step": 1463 }, { "epoch": 0.2816467872258561, "grad_norm": 0.18988715777430829, "learning_rate": 0.0001685872262878152, "loss": 0.795, "step": 1464 }, { "epoch": 0.28183916891111965, "grad_norm": 0.18471193208814757, "learning_rate": 0.000168541869741219, "loss": 0.8622, "step": 1465 }, { "epoch": 0.28203155059638324, "grad_norm": 0.16152861032762394, "learning_rate": 0.000168496486584309, "loss": 0.8286, "step": 1466 }, { "epoch": 0.28222393228164677, "grad_norm": 0.20010040338227708, "learning_rate": 0.00016845107683470452, "loss": 0.8548, "step": 1467 }, { "epoch": 0.28241631396691036, "grad_norm": 0.1886919322907406, "learning_rate": 0.00016840564051003514, "loss": 0.8227, "step": 1468 }, { "epoch": 0.2826086956521739, "grad_norm": 0.18651472587511642, "learning_rate": 0.00016836017762794086, "loss": 0.8402, "step": 1469 }, { "epoch": 0.2828010773374375, "grad_norm": 0.19386135818926553, "learning_rate": 0.0001683146882060719, "loss": 0.8099, "step": 1470 }, { "epoch": 0.28299345902270107, "grad_norm": 0.1796251815492587, "learning_rate": 0.00016826917226208887, "loss": 0.8788, "step": 1471 }, { "epoch": 0.2831858407079646, "grad_norm": 0.17731644215521358, "learning_rate": 0.00016822362981366257, "loss": 0.8133, "step": 1472 }, { "epoch": 0.2833782223932282, "grad_norm": 0.18751233548316837, "learning_rate": 0.00016817806087847417, "loss": 0.8108, "step": 1473 }, { "epoch": 0.2835706040784917, "grad_norm": 0.20532143700579691, "learning_rate": 0.0001681324654742151, "loss": 0.796, "step": 1474 }, { "epoch": 0.2837629857637553, "grad_norm": 0.19431610547558448, "learning_rate": 0.00016808684361858703, "loss": 0.8205, "step": 1475 }, { "epoch": 0.28395536744901884, "grad_norm": 0.19532087559493583, "learning_rate": 0.000168041195329302, "loss": 0.7949, "step": 1476 }, { "epoch": 0.2841477491342824, "grad_norm": 0.19494046323007228, "learning_rate": 0.00016799552062408225, "loss": 0.8053, "step": 1477 }, { "epoch": 0.28434013081954596, "grad_norm": 0.20562572789083972, "learning_rate": 0.00016794981952066018, "loss": 0.7954, "step": 1478 }, { "epoch": 0.28453251250480954, "grad_norm": 0.10625825476981952, "learning_rate": 0.0001679040920367786, "loss": 0.6257, "step": 1479 }, { "epoch": 0.28472489419007313, "grad_norm": 0.19659069950572966, "learning_rate": 0.0001678583381901905, "loss": 0.8237, "step": 1480 }, { "epoch": 0.28491727587533666, "grad_norm": 0.18518123149271204, "learning_rate": 0.0001678125579986591, "loss": 0.8229, "step": 1481 }, { "epoch": 0.28510965756060025, "grad_norm": 0.1881339084818998, "learning_rate": 0.0001677667514799578, "loss": 0.8559, "step": 1482 }, { "epoch": 0.2853020392458638, "grad_norm": 0.200237382736596, "learning_rate": 0.0001677209186518703, "loss": 0.767, "step": 1483 }, { "epoch": 0.28549442093112737, "grad_norm": 0.18458485513689327, "learning_rate": 0.00016767505953219048, "loss": 0.8446, "step": 1484 }, { "epoch": 0.2856868026163909, "grad_norm": 0.18866498776711246, "learning_rate": 0.00016762917413872246, "loss": 0.7824, "step": 1485 }, { "epoch": 0.2858791843016545, "grad_norm": 0.2104492016166846, "learning_rate": 0.00016758326248928049, "loss": 0.8502, "step": 1486 }, { "epoch": 0.286071565986918, "grad_norm": 0.20695167350957894, "learning_rate": 0.00016753732460168905, "loss": 0.7547, "step": 1487 }, { "epoch": 0.2862639476721816, "grad_norm": 0.2009004891419446, "learning_rate": 0.0001674913604937828, "loss": 0.8298, "step": 1488 }, { "epoch": 0.2864563293574452, "grad_norm": 0.2056083345777036, "learning_rate": 0.0001674453701834066, "loss": 0.8284, "step": 1489 }, { "epoch": 0.2866487110427087, "grad_norm": 0.20082922980998277, "learning_rate": 0.00016739935368841554, "loss": 0.8479, "step": 1490 }, { "epoch": 0.2868410927279723, "grad_norm": 0.18456074013414203, "learning_rate": 0.00016735331102667473, "loss": 0.8258, "step": 1491 }, { "epoch": 0.28703347441323585, "grad_norm": 0.20998607767384903, "learning_rate": 0.00016730724221605952, "loss": 0.846, "step": 1492 }, { "epoch": 0.28722585609849943, "grad_norm": 0.1960965505223357, "learning_rate": 0.00016726114727445544, "loss": 0.8108, "step": 1493 }, { "epoch": 0.28741823778376296, "grad_norm": 0.21334027078230003, "learning_rate": 0.00016721502621975813, "loss": 0.8343, "step": 1494 }, { "epoch": 0.28761061946902655, "grad_norm": 0.1791829517895399, "learning_rate": 0.00016716887906987333, "loss": 0.8383, "step": 1495 }, { "epoch": 0.28780300115429014, "grad_norm": 0.19299007324362896, "learning_rate": 0.000167122705842717, "loss": 0.8553, "step": 1496 }, { "epoch": 0.28799538283955367, "grad_norm": 0.2164732074909183, "learning_rate": 0.00016707650655621516, "loss": 0.8903, "step": 1497 }, { "epoch": 0.28818776452481726, "grad_norm": 0.19272321558820013, "learning_rate": 0.000167030281228304, "loss": 0.7999, "step": 1498 }, { "epoch": 0.2883801462100808, "grad_norm": 0.19230157578618162, "learning_rate": 0.00016698402987692966, "loss": 0.852, "step": 1499 }, { "epoch": 0.2885725278953444, "grad_norm": 0.20653309528365024, "learning_rate": 0.00016693775252004867, "loss": 0.867, "step": 1500 }, { "epoch": 0.2887649095806079, "grad_norm": 0.17291021425717443, "learning_rate": 0.0001668914491756274, "loss": 0.8292, "step": 1501 }, { "epoch": 0.2889572912658715, "grad_norm": 0.1820001892690797, "learning_rate": 0.0001668451198616424, "loss": 0.795, "step": 1502 }, { "epoch": 0.28914967295113503, "grad_norm": 0.19159349750036764, "learning_rate": 0.00016679876459608034, "loss": 0.8227, "step": 1503 }, { "epoch": 0.2893420546363986, "grad_norm": 0.10735694187958733, "learning_rate": 0.00016675238339693791, "loss": 0.5983, "step": 1504 }, { "epoch": 0.2895344363216622, "grad_norm": 0.19964997023072753, "learning_rate": 0.00016670597628222188, "loss": 0.8301, "step": 1505 }, { "epoch": 0.28972681800692573, "grad_norm": 0.17693341300659038, "learning_rate": 0.0001666595432699491, "loss": 0.8586, "step": 1506 }, { "epoch": 0.2899191996921893, "grad_norm": 0.1831984444508707, "learning_rate": 0.0001666130843781465, "loss": 0.8085, "step": 1507 }, { "epoch": 0.29011158137745285, "grad_norm": 0.17351826345932966, "learning_rate": 0.00016656659962485098, "loss": 0.7957, "step": 1508 }, { "epoch": 0.29030396306271644, "grad_norm": 0.1951874288465188, "learning_rate": 0.0001665200890281095, "loss": 0.8581, "step": 1509 }, { "epoch": 0.29049634474798, "grad_norm": 0.19220831061264573, "learning_rate": 0.00016647355260597914, "loss": 0.7968, "step": 1510 }, { "epoch": 0.29068872643324356, "grad_norm": 0.19729764884590598, "learning_rate": 0.00016642699037652692, "loss": 0.8397, "step": 1511 }, { "epoch": 0.2908811081185071, "grad_norm": 0.20574399103557284, "learning_rate": 0.00016638040235782982, "loss": 0.9214, "step": 1512 }, { "epoch": 0.2910734898037707, "grad_norm": 0.19253522614255425, "learning_rate": 0.00016633378856797503, "loss": 0.8622, "step": 1513 }, { "epoch": 0.29126587148903427, "grad_norm": 0.17598937563000255, "learning_rate": 0.00016628714902505956, "loss": 0.8574, "step": 1514 }, { "epoch": 0.2914582531742978, "grad_norm": 0.18991466199598783, "learning_rate": 0.00016624048374719053, "loss": 0.8553, "step": 1515 }, { "epoch": 0.2916506348595614, "grad_norm": 0.17706086465301105, "learning_rate": 0.000166193792752485, "loss": 0.8471, "step": 1516 }, { "epoch": 0.2918430165448249, "grad_norm": 0.17943543991784966, "learning_rate": 0.00016614707605906996, "loss": 0.8201, "step": 1517 }, { "epoch": 0.2920353982300885, "grad_norm": 0.18634796341093496, "learning_rate": 0.0001661003336850825, "loss": 0.8455, "step": 1518 }, { "epoch": 0.29222777991535204, "grad_norm": 0.18724443202326266, "learning_rate": 0.00016605356564866958, "loss": 0.8528, "step": 1519 }, { "epoch": 0.2924201616006156, "grad_norm": 0.5162573973344791, "learning_rate": 0.00016600677196798821, "loss": 0.8283, "step": 1520 }, { "epoch": 0.2926125432858792, "grad_norm": 0.19409851916815885, "learning_rate": 0.00016595995266120527, "loss": 0.9193, "step": 1521 }, { "epoch": 0.29280492497114274, "grad_norm": 0.1912896171277791, "learning_rate": 0.00016591310774649766, "loss": 0.8304, "step": 1522 }, { "epoch": 0.29299730665640633, "grad_norm": 0.19710445171650015, "learning_rate": 0.00016586623724205215, "loss": 0.8209, "step": 1523 }, { "epoch": 0.29318968834166986, "grad_norm": 0.1970944600111447, "learning_rate": 0.00016581934116606552, "loss": 0.8657, "step": 1524 }, { "epoch": 0.29338207002693345, "grad_norm": 0.2028319438812835, "learning_rate": 0.00016577241953674439, "loss": 0.8458, "step": 1525 }, { "epoch": 0.293574451712197, "grad_norm": 0.18527715469088285, "learning_rate": 0.0001657254723723054, "loss": 0.7896, "step": 1526 }, { "epoch": 0.29376683339746057, "grad_norm": 0.2014596003743171, "learning_rate": 0.00016567849969097503, "loss": 0.8196, "step": 1527 }, { "epoch": 0.2939592150827241, "grad_norm": 0.18725028885518769, "learning_rate": 0.0001656315015109897, "loss": 0.8203, "step": 1528 }, { "epoch": 0.2941515967679877, "grad_norm": 0.11938654137178392, "learning_rate": 0.00016558447785059577, "loss": 0.6316, "step": 1529 }, { "epoch": 0.2943439784532513, "grad_norm": 0.21721352449806486, "learning_rate": 0.00016553742872804937, "loss": 0.8881, "step": 1530 }, { "epoch": 0.2945363601385148, "grad_norm": 0.20780151564314897, "learning_rate": 0.00016549035416161664, "loss": 0.8702, "step": 1531 }, { "epoch": 0.2947287418237784, "grad_norm": 0.18443631936656565, "learning_rate": 0.0001654432541695735, "loss": 0.8085, "step": 1532 }, { "epoch": 0.2949211235090419, "grad_norm": 0.18230156134675257, "learning_rate": 0.00016539612877020582, "loss": 0.8456, "step": 1533 }, { "epoch": 0.2951135051943055, "grad_norm": 0.18854675587108624, "learning_rate": 0.00016534897798180928, "loss": 0.8202, "step": 1534 }, { "epoch": 0.29530588687956905, "grad_norm": 0.20048117551147931, "learning_rate": 0.00016530180182268946, "loss": 0.7832, "step": 1535 }, { "epoch": 0.29549826856483263, "grad_norm": 0.1667727745068663, "learning_rate": 0.00016525460031116177, "loss": 0.7769, "step": 1536 }, { "epoch": 0.29569065025009617, "grad_norm": 0.180110416393928, "learning_rate": 0.00016520737346555149, "loss": 0.7931, "step": 1537 }, { "epoch": 0.29588303193535975, "grad_norm": 0.18203803169512328, "learning_rate": 0.00016516012130419366, "loss": 0.8142, "step": 1538 }, { "epoch": 0.29607541362062334, "grad_norm": 0.1849928949717766, "learning_rate": 0.00016511284384543317, "loss": 0.7875, "step": 1539 }, { "epoch": 0.29626779530588687, "grad_norm": 0.18701175872473866, "learning_rate": 0.00016506554110762482, "loss": 0.819, "step": 1540 }, { "epoch": 0.29646017699115046, "grad_norm": 0.2029976071478923, "learning_rate": 0.00016501821310913316, "loss": 0.8491, "step": 1541 }, { "epoch": 0.296652558676414, "grad_norm": 0.17977490124982978, "learning_rate": 0.00016497085986833253, "loss": 0.818, "step": 1542 }, { "epoch": 0.2968449403616776, "grad_norm": 0.1922057474372565, "learning_rate": 0.00016492348140360706, "loss": 0.8184, "step": 1543 }, { "epoch": 0.2970373220469411, "grad_norm": 0.2223534630426283, "learning_rate": 0.00016487607773335073, "loss": 0.8187, "step": 1544 }, { "epoch": 0.2972297037322047, "grad_norm": 0.18897866785071366, "learning_rate": 0.0001648286488759673, "loss": 0.7681, "step": 1545 }, { "epoch": 0.29742208541746823, "grad_norm": 0.1702597658065217, "learning_rate": 0.00016478119484987025, "loss": 0.8346, "step": 1546 }, { "epoch": 0.2976144671027318, "grad_norm": 0.1875153605723152, "learning_rate": 0.00016473371567348285, "loss": 0.8517, "step": 1547 }, { "epoch": 0.2978068487879954, "grad_norm": 0.12311092984729861, "learning_rate": 0.00016468621136523823, "loss": 0.6207, "step": 1548 }, { "epoch": 0.29799923047325894, "grad_norm": 0.19215840154215327, "learning_rate": 0.0001646386819435791, "loss": 0.804, "step": 1549 }, { "epoch": 0.2981916121585225, "grad_norm": 0.18723263675451887, "learning_rate": 0.00016459112742695807, "loss": 0.8263, "step": 1550 }, { "epoch": 0.29838399384378606, "grad_norm": 0.2272132240558357, "learning_rate": 0.00016454354783383747, "loss": 0.7829, "step": 1551 }, { "epoch": 0.29857637552904964, "grad_norm": 0.19906788487941196, "learning_rate": 0.00016449594318268927, "loss": 0.8706, "step": 1552 }, { "epoch": 0.2987687572143132, "grad_norm": 0.179250081488488, "learning_rate": 0.00016444831349199528, "loss": 0.8113, "step": 1553 }, { "epoch": 0.29896113889957676, "grad_norm": 0.19804342930521382, "learning_rate": 0.00016440065878024697, "loss": 0.7978, "step": 1554 }, { "epoch": 0.29915352058484035, "grad_norm": 0.17506864645900996, "learning_rate": 0.00016435297906594555, "loss": 0.8577, "step": 1555 }, { "epoch": 0.2993459022701039, "grad_norm": 0.1996738815095168, "learning_rate": 0.00016430527436760191, "loss": 0.7554, "step": 1556 }, { "epoch": 0.29953828395536747, "grad_norm": 0.20118208423628478, "learning_rate": 0.00016425754470373666, "loss": 0.8197, "step": 1557 }, { "epoch": 0.299730665640631, "grad_norm": 0.12656245163589241, "learning_rate": 0.0001642097900928801, "loss": 0.6494, "step": 1558 }, { "epoch": 0.2999230473258946, "grad_norm": 0.2117000168755742, "learning_rate": 0.00016416201055357224, "loss": 0.7958, "step": 1559 }, { "epoch": 0.3001154290111581, "grad_norm": 0.20470493268115864, "learning_rate": 0.0001641142061043627, "loss": 0.7739, "step": 1560 }, { "epoch": 0.3003078106964217, "grad_norm": 0.17671849759070457, "learning_rate": 0.00016406637676381083, "loss": 0.8796, "step": 1561 }, { "epoch": 0.30050019238168524, "grad_norm": 0.2005364943443463, "learning_rate": 0.00016401852255048564, "loss": 0.837, "step": 1562 }, { "epoch": 0.3006925740669488, "grad_norm": 0.19368561317233574, "learning_rate": 0.00016397064348296577, "loss": 0.82, "step": 1563 }, { "epoch": 0.3008849557522124, "grad_norm": 0.11343267496390116, "learning_rate": 0.00016392273957983956, "loss": 0.6839, "step": 1564 }, { "epoch": 0.30107733743747594, "grad_norm": 0.23236962616750204, "learning_rate": 0.0001638748108597049, "loss": 0.874, "step": 1565 }, { "epoch": 0.30126971912273953, "grad_norm": 0.1894757891318984, "learning_rate": 0.00016382685734116933, "loss": 0.8351, "step": 1566 }, { "epoch": 0.30146210080800306, "grad_norm": 0.1915875463587414, "learning_rate": 0.00016377887904285018, "loss": 0.8076, "step": 1567 }, { "epoch": 0.30165448249326665, "grad_norm": 0.19115534319292582, "learning_rate": 0.00016373087598337422, "loss": 0.808, "step": 1568 }, { "epoch": 0.3018468641785302, "grad_norm": 0.1902145142135365, "learning_rate": 0.00016368284818137787, "loss": 0.874, "step": 1569 }, { "epoch": 0.30203924586379377, "grad_norm": 0.17655503221479243, "learning_rate": 0.00016363479565550718, "loss": 0.8443, "step": 1570 }, { "epoch": 0.3022316275490573, "grad_norm": 0.19813030668728152, "learning_rate": 0.0001635867184244178, "loss": 0.8274, "step": 1571 }, { "epoch": 0.3024240092343209, "grad_norm": 0.18837714540542513, "learning_rate": 0.000163538616506775, "loss": 0.8072, "step": 1572 }, { "epoch": 0.3026163909195845, "grad_norm": 0.20755503837478664, "learning_rate": 0.00016349048992125357, "loss": 0.8709, "step": 1573 }, { "epoch": 0.302808772604848, "grad_norm": 0.19509155808094433, "learning_rate": 0.00016344233868653788, "loss": 0.8314, "step": 1574 }, { "epoch": 0.3030011542901116, "grad_norm": 0.19297309439824775, "learning_rate": 0.00016339416282132196, "loss": 0.8509, "step": 1575 }, { "epoch": 0.30319353597537513, "grad_norm": 0.19727154919148376, "learning_rate": 0.0001633459623443093, "loss": 0.8968, "step": 1576 }, { "epoch": 0.3033859176606387, "grad_norm": 0.19556094153885212, "learning_rate": 0.00016329773727421297, "loss": 0.8598, "step": 1577 }, { "epoch": 0.30357829934590225, "grad_norm": 0.16829567855465763, "learning_rate": 0.00016324948762975566, "loss": 0.8253, "step": 1578 }, { "epoch": 0.30377068103116583, "grad_norm": 0.19551968129962263, "learning_rate": 0.0001632012134296695, "loss": 0.8213, "step": 1579 }, { "epoch": 0.3039630627164294, "grad_norm": 0.21196026320559885, "learning_rate": 0.0001631529146926962, "loss": 0.8194, "step": 1580 }, { "epoch": 0.30415544440169295, "grad_norm": 0.11150653835224815, "learning_rate": 0.000163104591437587, "loss": 0.6346, "step": 1581 }, { "epoch": 0.30434782608695654, "grad_norm": 0.10421315858596283, "learning_rate": 0.00016305624368310263, "loss": 0.6095, "step": 1582 }, { "epoch": 0.3045402077722201, "grad_norm": 0.20440136554621716, "learning_rate": 0.0001630078714480134, "loss": 0.8195, "step": 1583 }, { "epoch": 0.30473258945748366, "grad_norm": 0.20089281999426892, "learning_rate": 0.00016295947475109904, "loss": 0.826, "step": 1584 }, { "epoch": 0.3049249711427472, "grad_norm": 0.20529368953919974, "learning_rate": 0.0001629110536111488, "loss": 0.849, "step": 1585 }, { "epoch": 0.3051173528280108, "grad_norm": 0.1997130556724391, "learning_rate": 0.0001628626080469615, "loss": 0.8593, "step": 1586 }, { "epoch": 0.3053097345132743, "grad_norm": 0.1971779580928274, "learning_rate": 0.0001628141380773453, "loss": 0.7802, "step": 1587 }, { "epoch": 0.3055021161985379, "grad_norm": 0.20591794730662216, "learning_rate": 0.00016276564372111796, "loss": 0.8651, "step": 1588 }, { "epoch": 0.3056944978838015, "grad_norm": 0.1806691221163487, "learning_rate": 0.00016271712499710663, "loss": 0.7953, "step": 1589 }, { "epoch": 0.305886879569065, "grad_norm": 0.2347351922404659, "learning_rate": 0.00016266858192414798, "loss": 0.877, "step": 1590 }, { "epoch": 0.3060792612543286, "grad_norm": 0.10613612480931052, "learning_rate": 0.00016262001452108807, "loss": 0.6194, "step": 1591 }, { "epoch": 0.30627164293959214, "grad_norm": 0.17330008503467595, "learning_rate": 0.00016257142280678245, "loss": 0.8421, "step": 1592 }, { "epoch": 0.3064640246248557, "grad_norm": 0.18553536057548561, "learning_rate": 0.00016252280680009612, "loss": 0.8575, "step": 1593 }, { "epoch": 0.30665640631011926, "grad_norm": 0.1880074369085204, "learning_rate": 0.00016247416651990343, "loss": 0.8153, "step": 1594 }, { "epoch": 0.30684878799538284, "grad_norm": 0.18154961572162792, "learning_rate": 0.0001624255019850883, "loss": 0.7805, "step": 1595 }, { "epoch": 0.3070411696806464, "grad_norm": 0.1720765401575581, "learning_rate": 0.00016237681321454387, "loss": 0.8593, "step": 1596 }, { "epoch": 0.30723355136590996, "grad_norm": 0.1792696528872783, "learning_rate": 0.0001623281002271729, "loss": 0.8555, "step": 1597 }, { "epoch": 0.30742593305117355, "grad_norm": 0.18222303821453661, "learning_rate": 0.0001622793630418874, "loss": 0.7842, "step": 1598 }, { "epoch": 0.3076183147364371, "grad_norm": 0.17640567967736864, "learning_rate": 0.0001622306016776088, "loss": 0.7981, "step": 1599 }, { "epoch": 0.30781069642170067, "grad_norm": 0.18990762068379075, "learning_rate": 0.00016218181615326795, "loss": 0.8413, "step": 1600 }, { "epoch": 0.3080030781069642, "grad_norm": 0.18716367552729138, "learning_rate": 0.00016213300648780513, "loss": 0.867, "step": 1601 }, { "epoch": 0.3081954597922278, "grad_norm": 0.19686819240869638, "learning_rate": 0.0001620841727001699, "loss": 0.7985, "step": 1602 }, { "epoch": 0.3083878414774913, "grad_norm": 0.19090819080302943, "learning_rate": 0.00016203531480932115, "loss": 0.8088, "step": 1603 }, { "epoch": 0.3085802231627549, "grad_norm": 0.19860243194636282, "learning_rate": 0.00016198643283422729, "loss": 0.8265, "step": 1604 }, { "epoch": 0.3087726048480185, "grad_norm": 0.17311689232811323, "learning_rate": 0.00016193752679386593, "loss": 0.8185, "step": 1605 }, { "epoch": 0.308964986533282, "grad_norm": 0.18682424637187706, "learning_rate": 0.0001618885967072241, "loss": 0.8013, "step": 1606 }, { "epoch": 0.3091573682185456, "grad_norm": 0.184194840086558, "learning_rate": 0.00016183964259329815, "loss": 0.8299, "step": 1607 }, { "epoch": 0.30934974990380915, "grad_norm": 0.17032329846434752, "learning_rate": 0.00016179066447109377, "loss": 0.8036, "step": 1608 }, { "epoch": 0.30954213158907273, "grad_norm": 0.19878807119340186, "learning_rate": 0.00016174166235962586, "loss": 0.8252, "step": 1609 }, { "epoch": 0.30973451327433627, "grad_norm": 0.1871060480719126, "learning_rate": 0.00016169263627791887, "loss": 0.8588, "step": 1610 }, { "epoch": 0.30992689495959985, "grad_norm": 0.20397726479800096, "learning_rate": 0.0001616435862450063, "loss": 0.818, "step": 1611 }, { "epoch": 0.3101192766448634, "grad_norm": 0.11053921243705896, "learning_rate": 0.0001615945122799311, "loss": 0.6536, "step": 1612 }, { "epoch": 0.31031165833012697, "grad_norm": 0.10777193948379875, "learning_rate": 0.00016154541440174548, "loss": 0.6328, "step": 1613 }, { "epoch": 0.31050404001539056, "grad_norm": 0.1964512908070893, "learning_rate": 0.0001614962926295109, "loss": 0.7941, "step": 1614 }, { "epoch": 0.3106964217006541, "grad_norm": 0.22453000034059806, "learning_rate": 0.00016144714698229816, "loss": 0.7754, "step": 1615 }, { "epoch": 0.3108888033859177, "grad_norm": 0.1970862696472047, "learning_rate": 0.00016139797747918725, "loss": 0.8588, "step": 1616 }, { "epoch": 0.3110811850711812, "grad_norm": 0.17953001065968557, "learning_rate": 0.0001613487841392675, "loss": 0.8404, "step": 1617 }, { "epoch": 0.3112735667564448, "grad_norm": 0.2120917066249377, "learning_rate": 0.00016129956698163745, "loss": 0.8429, "step": 1618 }, { "epoch": 0.31146594844170833, "grad_norm": 0.18218461717088105, "learning_rate": 0.0001612503260254049, "loss": 0.8442, "step": 1619 }, { "epoch": 0.3116583301269719, "grad_norm": 0.1720946660638647, "learning_rate": 0.00016120106128968687, "loss": 0.8607, "step": 1620 }, { "epoch": 0.31185071181223545, "grad_norm": 0.1881955179530893, "learning_rate": 0.00016115177279360965, "loss": 0.8438, "step": 1621 }, { "epoch": 0.31204309349749904, "grad_norm": 0.20119984340159094, "learning_rate": 0.00016110246055630868, "loss": 0.7962, "step": 1622 }, { "epoch": 0.3122354751827626, "grad_norm": 0.19295353901046636, "learning_rate": 0.00016105312459692873, "loss": 0.8239, "step": 1623 }, { "epoch": 0.31242785686802615, "grad_norm": 0.1896046980916223, "learning_rate": 0.00016100376493462368, "loss": 0.8822, "step": 1624 }, { "epoch": 0.31262023855328974, "grad_norm": 0.17550198098183728, "learning_rate": 0.00016095438158855667, "loss": 0.7814, "step": 1625 }, { "epoch": 0.3128126202385533, "grad_norm": 0.1140288960505512, "learning_rate": 0.00016090497457789998, "loss": 0.5951, "step": 1626 }, { "epoch": 0.31300500192381686, "grad_norm": 0.1895769799937163, "learning_rate": 0.00016085554392183517, "loss": 0.8429, "step": 1627 }, { "epoch": 0.3131973836090804, "grad_norm": 0.17900655017425124, "learning_rate": 0.00016080608963955288, "loss": 0.7934, "step": 1628 }, { "epoch": 0.313389765294344, "grad_norm": 0.19006649348502222, "learning_rate": 0.00016075661175025298, "loss": 0.8526, "step": 1629 }, { "epoch": 0.31358214697960757, "grad_norm": 0.1976173679216062, "learning_rate": 0.00016070711027314446, "loss": 0.7725, "step": 1630 }, { "epoch": 0.3137745286648711, "grad_norm": 0.19248573719058945, "learning_rate": 0.00016065758522744558, "loss": 0.9051, "step": 1631 }, { "epoch": 0.3139669103501347, "grad_norm": 0.11295426599291601, "learning_rate": 0.0001606080366323836, "loss": 0.6653, "step": 1632 }, { "epoch": 0.3141592920353982, "grad_norm": 0.18578545256739348, "learning_rate": 0.00016055846450719496, "loss": 0.8442, "step": 1633 }, { "epoch": 0.3143516737206618, "grad_norm": 0.19830779365170959, "learning_rate": 0.00016050886887112532, "loss": 0.8456, "step": 1634 }, { "epoch": 0.31454405540592534, "grad_norm": 0.17612118710597585, "learning_rate": 0.00016045924974342942, "loss": 0.8493, "step": 1635 }, { "epoch": 0.3147364370911889, "grad_norm": 0.1885329756587949, "learning_rate": 0.0001604096071433711, "loss": 0.7735, "step": 1636 }, { "epoch": 0.31492881877645246, "grad_norm": 0.16923273373956205, "learning_rate": 0.00016035994109022333, "loss": 0.8445, "step": 1637 }, { "epoch": 0.31512120046171604, "grad_norm": 0.18541179246189055, "learning_rate": 0.00016031025160326813, "loss": 0.8352, "step": 1638 }, { "epoch": 0.31531358214697963, "grad_norm": 0.18833179238895065, "learning_rate": 0.00016026053870179675, "loss": 0.7852, "step": 1639 }, { "epoch": 0.31550596383224316, "grad_norm": 0.18659534761550597, "learning_rate": 0.00016021080240510944, "loss": 0.8199, "step": 1640 }, { "epoch": 0.31569834551750675, "grad_norm": 0.20050355859208635, "learning_rate": 0.0001601610427325155, "loss": 0.8432, "step": 1641 }, { "epoch": 0.3158907272027703, "grad_norm": 0.17215928398419322, "learning_rate": 0.00016011125970333333, "loss": 0.8421, "step": 1642 }, { "epoch": 0.31608310888803387, "grad_norm": 0.16508515819136377, "learning_rate": 0.00016006145333689049, "loss": 0.8193, "step": 1643 }, { "epoch": 0.3162754905732974, "grad_norm": 0.19550645040966116, "learning_rate": 0.00016001162365252347, "loss": 0.7952, "step": 1644 }, { "epoch": 0.316467872258561, "grad_norm": 0.19081694490724027, "learning_rate": 0.00015996177066957787, "loss": 0.8339, "step": 1645 }, { "epoch": 0.3166602539438245, "grad_norm": 0.21440183717967093, "learning_rate": 0.0001599118944074084, "loss": 0.8225, "step": 1646 }, { "epoch": 0.3168526356290881, "grad_norm": 0.17001628690864226, "learning_rate": 0.00015986199488537866, "loss": 0.8499, "step": 1647 }, { "epoch": 0.3170450173143517, "grad_norm": 0.19214317629296337, "learning_rate": 0.0001598120721228614, "loss": 0.8462, "step": 1648 }, { "epoch": 0.3172373989996152, "grad_norm": 0.18759112994011629, "learning_rate": 0.00015976212613923836, "loss": 0.8198, "step": 1649 }, { "epoch": 0.3174297806848788, "grad_norm": 0.1988399864621667, "learning_rate": 0.00015971215695390027, "loss": 0.8215, "step": 1650 }, { "epoch": 0.31762216237014235, "grad_norm": 0.1803841337817245, "learning_rate": 0.00015966216458624692, "loss": 0.8046, "step": 1651 }, { "epoch": 0.31781454405540593, "grad_norm": 0.18896757566330177, "learning_rate": 0.00015961214905568704, "loss": 0.8329, "step": 1652 }, { "epoch": 0.31800692574066947, "grad_norm": 0.19074678084274052, "learning_rate": 0.0001595621103816384, "loss": 0.8838, "step": 1653 }, { "epoch": 0.31819930742593305, "grad_norm": 0.20351653152334623, "learning_rate": 0.00015951204858352772, "loss": 0.8691, "step": 1654 }, { "epoch": 0.31839168911119664, "grad_norm": 0.18361129244603486, "learning_rate": 0.0001594619636807907, "loss": 0.7979, "step": 1655 }, { "epoch": 0.3185840707964602, "grad_norm": 0.18249499159757246, "learning_rate": 0.00015941185569287204, "loss": 0.8569, "step": 1656 }, { "epoch": 0.31877645248172376, "grad_norm": 0.2001048284441986, "learning_rate": 0.00015936172463922542, "loss": 0.8198, "step": 1657 }, { "epoch": 0.3189688341669873, "grad_norm": 0.20568325011049476, "learning_rate": 0.0001593115705393134, "loss": 0.8093, "step": 1658 }, { "epoch": 0.3191612158522509, "grad_norm": 0.17919680215187478, "learning_rate": 0.00015926139341260756, "loss": 0.8423, "step": 1659 }, { "epoch": 0.3193535975375144, "grad_norm": 0.1831588916693923, "learning_rate": 0.00015921119327858834, "loss": 0.8543, "step": 1660 }, { "epoch": 0.319545979222778, "grad_norm": 0.17610172111959183, "learning_rate": 0.00015916097015674518, "loss": 0.8499, "step": 1661 }, { "epoch": 0.31973836090804153, "grad_norm": 0.20621013908648456, "learning_rate": 0.00015911072406657646, "loss": 0.8506, "step": 1662 }, { "epoch": 0.3199307425933051, "grad_norm": 0.18462295328625386, "learning_rate": 0.0001590604550275894, "loss": 0.8118, "step": 1663 }, { "epoch": 0.3201231242785687, "grad_norm": 0.19990633332364507, "learning_rate": 0.0001590101630593002, "loss": 0.8013, "step": 1664 }, { "epoch": 0.32031550596383224, "grad_norm": 0.1824659818256553, "learning_rate": 0.0001589598481812339, "loss": 0.894, "step": 1665 }, { "epoch": 0.3205078876490958, "grad_norm": 0.18065845319605373, "learning_rate": 0.0001589095104129245, "loss": 0.8402, "step": 1666 }, { "epoch": 0.32070026933435936, "grad_norm": 0.2127571210295558, "learning_rate": 0.00015885914977391486, "loss": 0.822, "step": 1667 }, { "epoch": 0.32089265101962294, "grad_norm": 0.1909392749036821, "learning_rate": 0.00015880876628375666, "loss": 0.7883, "step": 1668 }, { "epoch": 0.3210850327048865, "grad_norm": 0.18209212272086575, "learning_rate": 0.00015875835996201058, "loss": 0.7896, "step": 1669 }, { "epoch": 0.32127741439015006, "grad_norm": 0.2030720199573047, "learning_rate": 0.00015870793082824604, "loss": 0.7643, "step": 1670 }, { "epoch": 0.3214697960754136, "grad_norm": 0.20004257291642583, "learning_rate": 0.00015865747890204138, "loss": 0.8341, "step": 1671 }, { "epoch": 0.3216621777606772, "grad_norm": 0.10468713405485196, "learning_rate": 0.00015860700420298378, "loss": 0.6197, "step": 1672 }, { "epoch": 0.32185455944594077, "grad_norm": 0.1805151632660655, "learning_rate": 0.00015855650675066925, "loss": 0.8403, "step": 1673 }, { "epoch": 0.3220469411312043, "grad_norm": 0.2159129319240224, "learning_rate": 0.00015850598656470263, "loss": 0.8272, "step": 1674 }, { "epoch": 0.3222393228164679, "grad_norm": 0.20498742768186562, "learning_rate": 0.0001584554436646976, "loss": 0.8266, "step": 1675 }, { "epoch": 0.3224317045017314, "grad_norm": 0.2100552180652674, "learning_rate": 0.00015840487807027665, "loss": 0.8381, "step": 1676 }, { "epoch": 0.322624086186995, "grad_norm": 0.20511316636239102, "learning_rate": 0.00015835428980107112, "loss": 0.8125, "step": 1677 }, { "epoch": 0.32281646787225854, "grad_norm": 0.17529614531030013, "learning_rate": 0.0001583036788767211, "loss": 0.8365, "step": 1678 }, { "epoch": 0.3230088495575221, "grad_norm": 0.19573175210277385, "learning_rate": 0.0001582530453168755, "loss": 0.8162, "step": 1679 }, { "epoch": 0.3232012312427857, "grad_norm": 0.18116709881066548, "learning_rate": 0.00015820238914119195, "loss": 0.789, "step": 1680 }, { "epoch": 0.32339361292804925, "grad_norm": 0.20731161802263856, "learning_rate": 0.000158151710369337, "loss": 0.7792, "step": 1681 }, { "epoch": 0.32358599461331283, "grad_norm": 0.18612656836866545, "learning_rate": 0.00015810100902098582, "loss": 0.8762, "step": 1682 }, { "epoch": 0.32377837629857636, "grad_norm": 0.19342974846296251, "learning_rate": 0.0001580502851158225, "loss": 0.8507, "step": 1683 }, { "epoch": 0.32397075798383995, "grad_norm": 0.1821427026118213, "learning_rate": 0.00015799953867353975, "loss": 0.8469, "step": 1684 }, { "epoch": 0.3241631396691035, "grad_norm": 0.1915146089413501, "learning_rate": 0.0001579487697138391, "loss": 0.8484, "step": 1685 }, { "epoch": 0.32435552135436707, "grad_norm": 0.20327160112606016, "learning_rate": 0.00015789797825643085, "loss": 0.8071, "step": 1686 }, { "epoch": 0.3245479030396306, "grad_norm": 0.18045595466035017, "learning_rate": 0.00015784716432103393, "loss": 0.8386, "step": 1687 }, { "epoch": 0.3247402847248942, "grad_norm": 0.18824280425833487, "learning_rate": 0.00015779632792737607, "loss": 0.8962, "step": 1688 }, { "epoch": 0.3249326664101578, "grad_norm": 0.19204259347019714, "learning_rate": 0.00015774546909519375, "loss": 0.8258, "step": 1689 }, { "epoch": 0.3251250480954213, "grad_norm": 0.19453608684563753, "learning_rate": 0.00015769458784423206, "loss": 0.8037, "step": 1690 }, { "epoch": 0.3253174297806849, "grad_norm": 0.2200266422136802, "learning_rate": 0.0001576436841942449, "loss": 0.8436, "step": 1691 }, { "epoch": 0.32550981146594843, "grad_norm": 0.17690105989752353, "learning_rate": 0.0001575927581649948, "loss": 0.8336, "step": 1692 }, { "epoch": 0.325702193151212, "grad_norm": 0.18765039846758466, "learning_rate": 0.00015754180977625303, "loss": 0.8721, "step": 1693 }, { "epoch": 0.32589457483647555, "grad_norm": 0.1986815182500631, "learning_rate": 0.00015749083904779946, "loss": 0.8041, "step": 1694 }, { "epoch": 0.32608695652173914, "grad_norm": 0.17824229378306303, "learning_rate": 0.00015743984599942272, "loss": 0.8324, "step": 1695 }, { "epoch": 0.32627933820700267, "grad_norm": 0.191011181153868, "learning_rate": 0.00015738883065092004, "loss": 0.8246, "step": 1696 }, { "epoch": 0.32647171989226625, "grad_norm": 0.1916833148037735, "learning_rate": 0.00015733779302209736, "loss": 0.7773, "step": 1697 }, { "epoch": 0.32666410157752984, "grad_norm": 0.2114362533526981, "learning_rate": 0.0001572867331327692, "loss": 0.8042, "step": 1698 }, { "epoch": 0.3268564832627934, "grad_norm": 0.18582872704069298, "learning_rate": 0.00015723565100275884, "loss": 0.8416, "step": 1699 }, { "epoch": 0.32704886494805696, "grad_norm": 0.19086538554428295, "learning_rate": 0.00015718454665189806, "loss": 0.8462, "step": 1700 }, { "epoch": 0.3272412466333205, "grad_norm": 0.21348940808725497, "learning_rate": 0.00015713342010002733, "loss": 0.8284, "step": 1701 }, { "epoch": 0.3274336283185841, "grad_norm": 0.19485073634248212, "learning_rate": 0.00015708227136699575, "loss": 0.8069, "step": 1702 }, { "epoch": 0.3276260100038476, "grad_norm": 0.1856424450406405, "learning_rate": 0.00015703110047266106, "loss": 0.7746, "step": 1703 }, { "epoch": 0.3278183916891112, "grad_norm": 0.1978978523585134, "learning_rate": 0.0001569799074368895, "loss": 0.8047, "step": 1704 }, { "epoch": 0.3280107733743748, "grad_norm": 0.18918012160092473, "learning_rate": 0.00015692869227955603, "loss": 0.7784, "step": 1705 }, { "epoch": 0.3282031550596383, "grad_norm": 0.20406065293744266, "learning_rate": 0.00015687745502054407, "loss": 0.7676, "step": 1706 }, { "epoch": 0.3283955367449019, "grad_norm": 0.19280142982295842, "learning_rate": 0.00015682619567974576, "loss": 0.8513, "step": 1707 }, { "epoch": 0.32858791843016544, "grad_norm": 0.20051826786009533, "learning_rate": 0.00015677491427706168, "loss": 0.8717, "step": 1708 }, { "epoch": 0.328780300115429, "grad_norm": 0.18881598545140066, "learning_rate": 0.00015672361083240107, "loss": 0.8162, "step": 1709 }, { "epoch": 0.32897268180069256, "grad_norm": 0.19704941968824555, "learning_rate": 0.00015667228536568167, "loss": 0.8397, "step": 1710 }, { "epoch": 0.32916506348595614, "grad_norm": 0.10682700224495083, "learning_rate": 0.00015662093789682978, "loss": 0.6045, "step": 1711 }, { "epoch": 0.3293574451712197, "grad_norm": 0.1933163517666145, "learning_rate": 0.0001565695684457803, "loss": 0.7604, "step": 1712 }, { "epoch": 0.32954982685648326, "grad_norm": 0.18690888818181559, "learning_rate": 0.00015651817703247667, "loss": 0.8721, "step": 1713 }, { "epoch": 0.32974220854174685, "grad_norm": 0.19299346744278864, "learning_rate": 0.00015646676367687067, "loss": 0.8268, "step": 1714 }, { "epoch": 0.3299345902270104, "grad_norm": 0.194578616977826, "learning_rate": 0.0001564153283989228, "loss": 0.792, "step": 1715 }, { "epoch": 0.33012697191227397, "grad_norm": 0.12439442410735901, "learning_rate": 0.00015636387121860206, "loss": 0.6815, "step": 1716 }, { "epoch": 0.3303193535975375, "grad_norm": 0.19173094078407688, "learning_rate": 0.00015631239215588578, "loss": 0.7351, "step": 1717 }, { "epoch": 0.3305117352828011, "grad_norm": 0.22010966807675164, "learning_rate": 0.00015626089123076, "loss": 0.8262, "step": 1718 }, { "epoch": 0.3307041169680646, "grad_norm": 0.18980125487917496, "learning_rate": 0.00015620936846321917, "loss": 0.853, "step": 1719 }, { "epoch": 0.3308964986533282, "grad_norm": 0.17628886599402893, "learning_rate": 0.0001561578238732661, "loss": 0.8004, "step": 1720 }, { "epoch": 0.33108888033859174, "grad_norm": 0.1972957763249144, "learning_rate": 0.00015610625748091228, "loss": 0.8025, "step": 1721 }, { "epoch": 0.3312812620238553, "grad_norm": 0.19650798978702294, "learning_rate": 0.00015605466930617746, "loss": 0.844, "step": 1722 }, { "epoch": 0.3314736437091189, "grad_norm": 0.17485036875198348, "learning_rate": 0.00015600305936909004, "loss": 0.8099, "step": 1723 }, { "epoch": 0.33166602539438245, "grad_norm": 0.16950463281456782, "learning_rate": 0.00015595142768968668, "loss": 0.8742, "step": 1724 }, { "epoch": 0.33185840707964603, "grad_norm": 0.20636679963975402, "learning_rate": 0.00015589977428801267, "loss": 0.8833, "step": 1725 }, { "epoch": 0.33205078876490957, "grad_norm": 0.19505015728667935, "learning_rate": 0.00015584809918412157, "loss": 0.8317, "step": 1726 }, { "epoch": 0.33224317045017315, "grad_norm": 0.1826441267439801, "learning_rate": 0.00015579640239807548, "loss": 0.7976, "step": 1727 }, { "epoch": 0.3324355521354367, "grad_norm": 0.11944263181340453, "learning_rate": 0.00015574468394994486, "loss": 0.6582, "step": 1728 }, { "epoch": 0.3326279338207003, "grad_norm": 0.18819083881429968, "learning_rate": 0.00015569294385980856, "loss": 0.8171, "step": 1729 }, { "epoch": 0.33282031550596386, "grad_norm": 0.18276978863772123, "learning_rate": 0.0001556411821477539, "loss": 0.7823, "step": 1730 }, { "epoch": 0.3330126971912274, "grad_norm": 0.20046896114268264, "learning_rate": 0.00015558939883387655, "loss": 0.8476, "step": 1731 }, { "epoch": 0.333205078876491, "grad_norm": 0.22388470792073412, "learning_rate": 0.00015553759393828058, "loss": 0.7528, "step": 1732 }, { "epoch": 0.3333974605617545, "grad_norm": 0.1864346420367414, "learning_rate": 0.00015548576748107838, "loss": 0.8677, "step": 1733 }, { "epoch": 0.3335898422470181, "grad_norm": 0.2013215233548761, "learning_rate": 0.00015543391948239088, "loss": 0.8467, "step": 1734 }, { "epoch": 0.33378222393228163, "grad_norm": 0.1935663683388311, "learning_rate": 0.00015538204996234716, "loss": 0.7407, "step": 1735 }, { "epoch": 0.3339746056175452, "grad_norm": 0.19862986350454273, "learning_rate": 0.0001553301589410848, "loss": 0.7769, "step": 1736 }, { "epoch": 0.33416698730280875, "grad_norm": 0.19047382623015013, "learning_rate": 0.00015527824643874966, "loss": 0.8204, "step": 1737 }, { "epoch": 0.33435936898807234, "grad_norm": 0.20408966385216362, "learning_rate": 0.00015522631247549598, "loss": 0.8054, "step": 1738 }, { "epoch": 0.3345517506733359, "grad_norm": 0.1943982535261032, "learning_rate": 0.00015517435707148626, "loss": 0.8412, "step": 1739 }, { "epoch": 0.33474413235859946, "grad_norm": 0.18185876727379158, "learning_rate": 0.00015512238024689142, "loss": 0.879, "step": 1740 }, { "epoch": 0.33493651404386304, "grad_norm": 0.18904678925167936, "learning_rate": 0.0001550703820218907, "loss": 0.8289, "step": 1741 }, { "epoch": 0.3351288957291266, "grad_norm": 0.19781373208561548, "learning_rate": 0.0001550183624166715, "loss": 0.8346, "step": 1742 }, { "epoch": 0.33532127741439016, "grad_norm": 0.20210814992446136, "learning_rate": 0.0001549663214514297, "loss": 0.8518, "step": 1743 }, { "epoch": 0.3355136590996537, "grad_norm": 0.18966839780682698, "learning_rate": 0.00015491425914636935, "loss": 0.7682, "step": 1744 }, { "epoch": 0.3357060407849173, "grad_norm": 0.164931098426725, "learning_rate": 0.00015486217552170283, "loss": 0.8336, "step": 1745 }, { "epoch": 0.3358984224701808, "grad_norm": 0.18778821851790317, "learning_rate": 0.00015481007059765082, "loss": 0.8293, "step": 1746 }, { "epoch": 0.3360908041554444, "grad_norm": 0.19073551000082092, "learning_rate": 0.00015475794439444226, "loss": 0.8916, "step": 1747 }, { "epoch": 0.336283185840708, "grad_norm": 0.18623139548943088, "learning_rate": 0.00015470579693231432, "loss": 0.7879, "step": 1748 }, { "epoch": 0.3364755675259715, "grad_norm": 0.1940590510443446, "learning_rate": 0.00015465362823151245, "loss": 0.8071, "step": 1749 }, { "epoch": 0.3366679492112351, "grad_norm": 0.18311436140380274, "learning_rate": 0.00015460143831229025, "loss": 0.8475, "step": 1750 }, { "epoch": 0.33686033089649864, "grad_norm": 0.2187015741041999, "learning_rate": 0.00015454922719490978, "loss": 0.8686, "step": 1751 }, { "epoch": 0.3370527125817622, "grad_norm": 0.1819456730450867, "learning_rate": 0.0001544969948996411, "loss": 0.8336, "step": 1752 }, { "epoch": 0.33724509426702576, "grad_norm": 0.19025731475645186, "learning_rate": 0.0001544447414467626, "loss": 0.7841, "step": 1753 }, { "epoch": 0.33743747595228935, "grad_norm": 0.21338674358867166, "learning_rate": 0.00015439246685656092, "loss": 0.8369, "step": 1754 }, { "epoch": 0.33762985763755293, "grad_norm": 0.18554744902352385, "learning_rate": 0.0001543401711493308, "loss": 0.8062, "step": 1755 }, { "epoch": 0.33782223932281646, "grad_norm": 0.1879680247117677, "learning_rate": 0.00015428785434537526, "loss": 0.8551, "step": 1756 }, { "epoch": 0.33801462100808005, "grad_norm": 0.1926600614459196, "learning_rate": 0.00015423551646500546, "loss": 0.8334, "step": 1757 }, { "epoch": 0.3382070026933436, "grad_norm": 0.1906057371410584, "learning_rate": 0.0001541831575285408, "loss": 0.8051, "step": 1758 }, { "epoch": 0.33839938437860717, "grad_norm": 0.19103356467499302, "learning_rate": 0.00015413077755630877, "loss": 0.8453, "step": 1759 }, { "epoch": 0.3385917660638707, "grad_norm": 0.17669441217393111, "learning_rate": 0.0001540783765686452, "loss": 0.8064, "step": 1760 }, { "epoch": 0.3387841477491343, "grad_norm": 0.19649578460021175, "learning_rate": 0.0001540259545858938, "loss": 0.8624, "step": 1761 }, { "epoch": 0.3389765294343978, "grad_norm": 0.20837712021296526, "learning_rate": 0.0001539735116284067, "loss": 0.8556, "step": 1762 }, { "epoch": 0.3391689111196614, "grad_norm": 0.1906803360317098, "learning_rate": 0.000153921047716544, "loss": 0.7768, "step": 1763 }, { "epoch": 0.339361292804925, "grad_norm": 0.20800669908423944, "learning_rate": 0.00015386856287067402, "loss": 0.7755, "step": 1764 }, { "epoch": 0.33955367449018853, "grad_norm": 0.20024237347537363, "learning_rate": 0.00015381605711117316, "loss": 0.8296, "step": 1765 }, { "epoch": 0.3397460561754521, "grad_norm": 0.11265264765598482, "learning_rate": 0.00015376353045842603, "loss": 0.627, "step": 1766 }, { "epoch": 0.33993843786071565, "grad_norm": 0.20511330634464367, "learning_rate": 0.00015371098293282524, "loss": 0.8759, "step": 1767 }, { "epoch": 0.34013081954597923, "grad_norm": 0.1951374712711241, "learning_rate": 0.00015365841455477157, "loss": 0.8607, "step": 1768 }, { "epoch": 0.34032320123124277, "grad_norm": 0.1937278466364481, "learning_rate": 0.0001536058253446738, "loss": 0.8137, "step": 1769 }, { "epoch": 0.34051558291650635, "grad_norm": 0.21352348598731902, "learning_rate": 0.00015355321532294896, "loss": 0.8096, "step": 1770 }, { "epoch": 0.3407079646017699, "grad_norm": 0.09416484555235052, "learning_rate": 0.00015350058451002205, "loss": 0.5869, "step": 1771 }, { "epoch": 0.3409003462870335, "grad_norm": 0.2300452963181672, "learning_rate": 0.00015344793292632614, "loss": 0.777, "step": 1772 }, { "epoch": 0.34109272797229706, "grad_norm": 0.19450105619400693, "learning_rate": 0.0001533952605923024, "loss": 0.787, "step": 1773 }, { "epoch": 0.3412851096575606, "grad_norm": 0.1868547169181129, "learning_rate": 0.00015334256752840006, "loss": 0.7949, "step": 1774 }, { "epoch": 0.3414774913428242, "grad_norm": 0.1964768758969962, "learning_rate": 0.00015328985375507638, "loss": 0.8329, "step": 1775 }, { "epoch": 0.3416698730280877, "grad_norm": 0.1897109559909842, "learning_rate": 0.0001532371192927966, "loss": 0.7992, "step": 1776 }, { "epoch": 0.3418622547133513, "grad_norm": 0.18016931728214863, "learning_rate": 0.00015318436416203413, "loss": 0.7574, "step": 1777 }, { "epoch": 0.34205463639861483, "grad_norm": 0.1848540232822341, "learning_rate": 0.0001531315883832703, "loss": 0.8232, "step": 1778 }, { "epoch": 0.3422470180838784, "grad_norm": 0.19066388904464585, "learning_rate": 0.0001530787919769945, "loss": 0.7529, "step": 1779 }, { "epoch": 0.34243939976914195, "grad_norm": 0.10414219287272304, "learning_rate": 0.00015302597496370408, "loss": 0.6657, "step": 1780 }, { "epoch": 0.34263178145440554, "grad_norm": 0.17724750130180095, "learning_rate": 0.00015297313736390447, "loss": 0.7827, "step": 1781 }, { "epoch": 0.3428241631396691, "grad_norm": 0.18595985470427073, "learning_rate": 0.00015292027919810896, "loss": 0.8109, "step": 1782 }, { "epoch": 0.34301654482493266, "grad_norm": 0.19934360593988018, "learning_rate": 0.00015286740048683898, "loss": 0.7928, "step": 1783 }, { "epoch": 0.34320892651019624, "grad_norm": 0.17726543010286355, "learning_rate": 0.0001528145012506239, "loss": 0.8074, "step": 1784 }, { "epoch": 0.3434013081954598, "grad_norm": 0.20669561761136135, "learning_rate": 0.00015276158151000094, "loss": 0.8384, "step": 1785 }, { "epoch": 0.34359368988072336, "grad_norm": 0.10535739409404726, "learning_rate": 0.00015270864128551543, "loss": 0.6184, "step": 1786 }, { "epoch": 0.3437860715659869, "grad_norm": 0.18350805782290922, "learning_rate": 0.00015265568059772054, "loss": 0.8043, "step": 1787 }, { "epoch": 0.3439784532512505, "grad_norm": 0.1795099990606464, "learning_rate": 0.00015260269946717746, "loss": 0.8337, "step": 1788 }, { "epoch": 0.34417083493651407, "grad_norm": 0.19905391653202376, "learning_rate": 0.00015254969791445526, "loss": 0.8517, "step": 1789 }, { "epoch": 0.3443632166217776, "grad_norm": 0.18610903074940838, "learning_rate": 0.000152496675960131, "loss": 0.7788, "step": 1790 }, { "epoch": 0.3445555983070412, "grad_norm": 0.19275383142725466, "learning_rate": 0.00015244363362478967, "loss": 0.8153, "step": 1791 }, { "epoch": 0.3447479799923047, "grad_norm": 0.17253110683340245, "learning_rate": 0.00015239057092902405, "loss": 0.8216, "step": 1792 }, { "epoch": 0.3449403616775683, "grad_norm": 0.1841499610095359, "learning_rate": 0.00015233748789343488, "loss": 0.824, "step": 1793 }, { "epoch": 0.34513274336283184, "grad_norm": 0.1984124626143295, "learning_rate": 0.00015228438453863095, "loss": 0.8258, "step": 1794 }, { "epoch": 0.3453251250480954, "grad_norm": 0.1866118430081041, "learning_rate": 0.0001522312608852287, "loss": 0.8163, "step": 1795 }, { "epoch": 0.34551750673335896, "grad_norm": 0.19523677837893272, "learning_rate": 0.00015217811695385263, "loss": 0.8443, "step": 1796 }, { "epoch": 0.34570988841862255, "grad_norm": 0.17233779536014404, "learning_rate": 0.000152124952765135, "loss": 0.8394, "step": 1797 }, { "epoch": 0.34590227010388613, "grad_norm": 0.10616872214704556, "learning_rate": 0.00015207176833971598, "loss": 0.6603, "step": 1798 }, { "epoch": 0.34609465178914967, "grad_norm": 0.097218272649807, "learning_rate": 0.0001520185636982436, "loss": 0.6347, "step": 1799 }, { "epoch": 0.34628703347441325, "grad_norm": 0.2110012519920593, "learning_rate": 0.00015196533886137377, "loss": 0.8384, "step": 1800 }, { "epoch": 0.3464794151596768, "grad_norm": 0.20176111826684043, "learning_rate": 0.00015191209384977014, "loss": 0.8138, "step": 1801 }, { "epoch": 0.34667179684494037, "grad_norm": 0.19023496729248732, "learning_rate": 0.0001518588286841043, "loss": 0.8162, "step": 1802 }, { "epoch": 0.3468641785302039, "grad_norm": 0.10580048583544063, "learning_rate": 0.00015180554338505565, "loss": 0.6039, "step": 1803 }, { "epoch": 0.3470565602154675, "grad_norm": 0.22107083185639387, "learning_rate": 0.0001517522379733113, "loss": 0.7849, "step": 1804 }, { "epoch": 0.347248941900731, "grad_norm": 0.1926014116401741, "learning_rate": 0.0001516989124695663, "loss": 0.7956, "step": 1805 }, { "epoch": 0.3474413235859946, "grad_norm": 0.20459299196953698, "learning_rate": 0.00015164556689452346, "loss": 0.7837, "step": 1806 }, { "epoch": 0.3476337052712582, "grad_norm": 0.20730796147473546, "learning_rate": 0.0001515922012688933, "loss": 0.7581, "step": 1807 }, { "epoch": 0.34782608695652173, "grad_norm": 0.21129441038484498, "learning_rate": 0.00015153881561339426, "loss": 0.8538, "step": 1808 }, { "epoch": 0.3480184686417853, "grad_norm": 0.21514572689017722, "learning_rate": 0.0001514854099487524, "loss": 0.8265, "step": 1809 }, { "epoch": 0.34821085032704885, "grad_norm": 0.18642960582552015, "learning_rate": 0.0001514319842957018, "loss": 0.8523, "step": 1810 }, { "epoch": 0.34840323201231244, "grad_norm": 0.09617604523441817, "learning_rate": 0.000151378538674984, "loss": 0.5818, "step": 1811 }, { "epoch": 0.34859561369757597, "grad_norm": 0.1944138047229798, "learning_rate": 0.00015132507310734847, "loss": 0.8515, "step": 1812 }, { "epoch": 0.34878799538283956, "grad_norm": 0.18656174507171697, "learning_rate": 0.0001512715876135524, "loss": 0.87, "step": 1813 }, { "epoch": 0.34898037706810314, "grad_norm": 0.19781659459426543, "learning_rate": 0.0001512180822143607, "loss": 0.8069, "step": 1814 }, { "epoch": 0.3491727587533667, "grad_norm": 0.1865912016271153, "learning_rate": 0.00015116455693054594, "loss": 0.8203, "step": 1815 }, { "epoch": 0.34936514043863026, "grad_norm": 0.20493712991774385, "learning_rate": 0.00015111101178288857, "loss": 0.8146, "step": 1816 }, { "epoch": 0.3495575221238938, "grad_norm": 0.18076150529955531, "learning_rate": 0.0001510574467921766, "loss": 0.8067, "step": 1817 }, { "epoch": 0.3497499038091574, "grad_norm": 0.1925842085731863, "learning_rate": 0.00015100386197920584, "loss": 0.7624, "step": 1818 }, { "epoch": 0.3499422854944209, "grad_norm": 0.19884067323434876, "learning_rate": 0.00015095025736477978, "loss": 0.8374, "step": 1819 }, { "epoch": 0.3501346671796845, "grad_norm": 0.18549997671663837, "learning_rate": 0.0001508966329697095, "loss": 0.8042, "step": 1820 }, { "epoch": 0.35032704886494803, "grad_norm": 0.19928066175770293, "learning_rate": 0.00015084298881481387, "loss": 0.7729, "step": 1821 }, { "epoch": 0.3505194305502116, "grad_norm": 0.20721866279664444, "learning_rate": 0.00015078932492091943, "loss": 0.8192, "step": 1822 }, { "epoch": 0.3507118122354752, "grad_norm": 0.19961445966315458, "learning_rate": 0.0001507356413088603, "loss": 0.8644, "step": 1823 }, { "epoch": 0.35090419392073874, "grad_norm": 0.20710169996679864, "learning_rate": 0.00015068193799947839, "loss": 0.7299, "step": 1824 }, { "epoch": 0.3510965756060023, "grad_norm": 0.18489054144395833, "learning_rate": 0.00015062821501362307, "loss": 0.7926, "step": 1825 }, { "epoch": 0.35128895729126586, "grad_norm": 0.18656100935783632, "learning_rate": 0.0001505744723721515, "loss": 0.8401, "step": 1826 }, { "epoch": 0.35148133897652944, "grad_norm": 0.19335017892348053, "learning_rate": 0.00015052071009592844, "loss": 0.8033, "step": 1827 }, { "epoch": 0.351673720661793, "grad_norm": 0.19595453533647536, "learning_rate": 0.00015046692820582625, "loss": 0.8343, "step": 1828 }, { "epoch": 0.35186610234705656, "grad_norm": 0.19311524049149822, "learning_rate": 0.0001504131267227249, "loss": 0.8256, "step": 1829 }, { "epoch": 0.3520584840323201, "grad_norm": 0.20096129056845102, "learning_rate": 0.00015035930566751197, "loss": 0.8562, "step": 1830 }, { "epoch": 0.3522508657175837, "grad_norm": 0.18351215385303182, "learning_rate": 0.00015030546506108268, "loss": 0.8679, "step": 1831 }, { "epoch": 0.35244324740284727, "grad_norm": 0.18182493867368346, "learning_rate": 0.00015025160492433977, "loss": 0.7959, "step": 1832 }, { "epoch": 0.3526356290881108, "grad_norm": 0.19654814611663687, "learning_rate": 0.0001501977252781936, "loss": 0.818, "step": 1833 }, { "epoch": 0.3528280107733744, "grad_norm": 0.2055489151650067, "learning_rate": 0.00015014382614356213, "loss": 0.8385, "step": 1834 }, { "epoch": 0.3530203924586379, "grad_norm": 0.20241311447987584, "learning_rate": 0.00015008990754137086, "loss": 0.717, "step": 1835 }, { "epoch": 0.3532127741439015, "grad_norm": 0.19998750874918353, "learning_rate": 0.00015003596949255283, "loss": 0.7879, "step": 1836 }, { "epoch": 0.35340515582916504, "grad_norm": 0.18648744461211342, "learning_rate": 0.00014998201201804867, "loss": 0.8205, "step": 1837 }, { "epoch": 0.35359753751442863, "grad_norm": 0.18986197812627809, "learning_rate": 0.0001499280351388065, "loss": 0.8136, "step": 1838 }, { "epoch": 0.3537899191996922, "grad_norm": 0.19900120411528008, "learning_rate": 0.000149874038875782, "loss": 0.7876, "step": 1839 }, { "epoch": 0.35398230088495575, "grad_norm": 0.2062420304680895, "learning_rate": 0.0001498200232499384, "loss": 0.798, "step": 1840 }, { "epoch": 0.35417468257021933, "grad_norm": 0.2149584210092314, "learning_rate": 0.0001497659882822464, "loss": 0.8455, "step": 1841 }, { "epoch": 0.35436706425548287, "grad_norm": 0.1923081065020319, "learning_rate": 0.0001497119339936843, "loss": 0.76, "step": 1842 }, { "epoch": 0.35455944594074645, "grad_norm": 0.18721306936578977, "learning_rate": 0.0001496578604052378, "loss": 0.8234, "step": 1843 }, { "epoch": 0.35475182762601, "grad_norm": 0.19637454925373923, "learning_rate": 0.0001496037675379001, "loss": 0.8336, "step": 1844 }, { "epoch": 0.3549442093112736, "grad_norm": 0.10851193739466082, "learning_rate": 0.00014954965541267193, "loss": 0.6063, "step": 1845 }, { "epoch": 0.3551365909965371, "grad_norm": 0.20752056564544022, "learning_rate": 0.0001494955240505615, "loss": 0.7719, "step": 1846 }, { "epoch": 0.3553289726818007, "grad_norm": 0.19956864249701636, "learning_rate": 0.0001494413734725844, "loss": 0.7902, "step": 1847 }, { "epoch": 0.3555213543670643, "grad_norm": 0.19203423735497713, "learning_rate": 0.00014938720369976386, "loss": 0.8188, "step": 1848 }, { "epoch": 0.3557137360523278, "grad_norm": 0.20335928853270613, "learning_rate": 0.00014933301475313036, "loss": 0.8461, "step": 1849 }, { "epoch": 0.3559061177375914, "grad_norm": 0.19529562905742653, "learning_rate": 0.00014927880665372196, "loss": 0.8718, "step": 1850 }, { "epoch": 0.35609849942285493, "grad_norm": 0.187275734202214, "learning_rate": 0.0001492245794225841, "loss": 0.8188, "step": 1851 }, { "epoch": 0.3562908811081185, "grad_norm": 0.2050493367248302, "learning_rate": 0.00014917033308076965, "loss": 0.8413, "step": 1852 }, { "epoch": 0.35648326279338205, "grad_norm": 0.19446938126158383, "learning_rate": 0.00014911606764933893, "loss": 0.8384, "step": 1853 }, { "epoch": 0.35667564447864564, "grad_norm": 0.17526036093216882, "learning_rate": 0.0001490617831493596, "loss": 0.8383, "step": 1854 }, { "epoch": 0.35686802616390917, "grad_norm": 0.17908951664551603, "learning_rate": 0.00014900747960190682, "loss": 0.794, "step": 1855 }, { "epoch": 0.35706040784917276, "grad_norm": 0.1857918302999117, "learning_rate": 0.0001489531570280631, "loss": 0.8441, "step": 1856 }, { "epoch": 0.35725278953443634, "grad_norm": 0.20504332708246298, "learning_rate": 0.0001488988154489183, "loss": 0.7742, "step": 1857 }, { "epoch": 0.3574451712196999, "grad_norm": 0.18169661192543457, "learning_rate": 0.0001488444548855697, "loss": 0.7822, "step": 1858 }, { "epoch": 0.35763755290496346, "grad_norm": 0.17431292857275585, "learning_rate": 0.00014879007535912198, "loss": 0.8263, "step": 1859 }, { "epoch": 0.357829934590227, "grad_norm": 0.11450647272326933, "learning_rate": 0.00014873567689068707, "loss": 0.6584, "step": 1860 }, { "epoch": 0.3580223162754906, "grad_norm": 0.16649234115528158, "learning_rate": 0.00014868125950138442, "loss": 0.8317, "step": 1861 }, { "epoch": 0.3582146979607541, "grad_norm": 0.18617570508678663, "learning_rate": 0.00014862682321234062, "loss": 0.8187, "step": 1862 }, { "epoch": 0.3584070796460177, "grad_norm": 0.16978973510814596, "learning_rate": 0.00014857236804468983, "loss": 0.8487, "step": 1863 }, { "epoch": 0.3585994613312813, "grad_norm": 0.18407840935206038, "learning_rate": 0.00014851789401957338, "loss": 0.7833, "step": 1864 }, { "epoch": 0.3587918430165448, "grad_norm": 0.19014584975357213, "learning_rate": 0.00014846340115813993, "loss": 0.7826, "step": 1865 }, { "epoch": 0.3589842247018084, "grad_norm": 0.19126622817669237, "learning_rate": 0.00014840888948154549, "loss": 0.8358, "step": 1866 }, { "epoch": 0.35917660638707194, "grad_norm": 0.196546172387649, "learning_rate": 0.0001483543590109534, "loss": 0.886, "step": 1867 }, { "epoch": 0.3593689880723355, "grad_norm": 0.10743406782856424, "learning_rate": 0.00014829980976753426, "loss": 0.625, "step": 1868 }, { "epoch": 0.35956136975759906, "grad_norm": 0.17962910224887618, "learning_rate": 0.00014824524177246598, "loss": 0.8256, "step": 1869 }, { "epoch": 0.35975375144286265, "grad_norm": 0.19293490502246657, "learning_rate": 0.00014819065504693364, "loss": 0.7732, "step": 1870 }, { "epoch": 0.3599461331281262, "grad_norm": 0.17751965779162718, "learning_rate": 0.00014813604961212983, "loss": 0.8606, "step": 1871 }, { "epoch": 0.36013851481338977, "grad_norm": 0.20096227127891061, "learning_rate": 0.00014808142548925418, "loss": 0.8765, "step": 1872 }, { "epoch": 0.36033089649865335, "grad_norm": 0.1947820307619621, "learning_rate": 0.00014802678269951365, "loss": 0.8186, "step": 1873 }, { "epoch": 0.3605232781839169, "grad_norm": 0.20939036649278472, "learning_rate": 0.00014797212126412243, "loss": 0.834, "step": 1874 }, { "epoch": 0.36071565986918047, "grad_norm": 0.21825701683697352, "learning_rate": 0.00014791744120430203, "loss": 0.773, "step": 1875 }, { "epoch": 0.360908041554444, "grad_norm": 0.1907671360153622, "learning_rate": 0.0001478627425412811, "loss": 0.8578, "step": 1876 }, { "epoch": 0.3611004232397076, "grad_norm": 0.19532119889003818, "learning_rate": 0.00014780802529629558, "loss": 0.8323, "step": 1877 }, { "epoch": 0.3612928049249711, "grad_norm": 0.2228365046476096, "learning_rate": 0.00014775328949058854, "loss": 0.8319, "step": 1878 }, { "epoch": 0.3614851866102347, "grad_norm": 0.19456043173663898, "learning_rate": 0.00014769853514541036, "loss": 0.8895, "step": 1879 }, { "epoch": 0.36167756829549824, "grad_norm": 0.18056467782326693, "learning_rate": 0.0001476437622820185, "loss": 0.7698, "step": 1880 }, { "epoch": 0.36186994998076183, "grad_norm": 0.18090028768698777, "learning_rate": 0.0001475889709216777, "loss": 0.7996, "step": 1881 }, { "epoch": 0.3620623316660254, "grad_norm": 0.1784749993309769, "learning_rate": 0.00014753416108565984, "loss": 0.8536, "step": 1882 }, { "epoch": 0.36225471335128895, "grad_norm": 0.17737261708011298, "learning_rate": 0.000147479332795244, "loss": 0.8128, "step": 1883 }, { "epoch": 0.36244709503655254, "grad_norm": 0.19595978322856414, "learning_rate": 0.00014742448607171644, "loss": 0.8592, "step": 1884 }, { "epoch": 0.36263947672181607, "grad_norm": 0.14149693641092254, "learning_rate": 0.0001473696209363705, "loss": 0.6783, "step": 1885 }, { "epoch": 0.36283185840707965, "grad_norm": 0.18889713033587413, "learning_rate": 0.00014731473741050671, "loss": 0.7926, "step": 1886 }, { "epoch": 0.3630242400923432, "grad_norm": 0.2081646527871777, "learning_rate": 0.00014725983551543278, "loss": 0.8101, "step": 1887 }, { "epoch": 0.3632166217776068, "grad_norm": 0.21005756764873237, "learning_rate": 0.0001472049152724635, "loss": 0.8076, "step": 1888 }, { "epoch": 0.36340900346287036, "grad_norm": 0.18539489873983636, "learning_rate": 0.00014714997670292078, "loss": 0.8724, "step": 1889 }, { "epoch": 0.3636013851481339, "grad_norm": 0.19550445373277522, "learning_rate": 0.00014709501982813368, "loss": 0.8465, "step": 1890 }, { "epoch": 0.3637937668333975, "grad_norm": 0.21187962007492772, "learning_rate": 0.0001470400446694384, "loss": 0.7785, "step": 1891 }, { "epoch": 0.363986148518661, "grad_norm": 0.21044679190661073, "learning_rate": 0.00014698505124817812, "loss": 0.7856, "step": 1892 }, { "epoch": 0.3641785302039246, "grad_norm": 0.18380026917565095, "learning_rate": 0.00014693003958570318, "loss": 0.7849, "step": 1893 }, { "epoch": 0.36437091188918813, "grad_norm": 0.11631027156237951, "learning_rate": 0.00014687500970337103, "loss": 0.6725, "step": 1894 }, { "epoch": 0.3645632935744517, "grad_norm": 0.1878648885964501, "learning_rate": 0.00014681996162254618, "loss": 0.8424, "step": 1895 }, { "epoch": 0.36475567525971525, "grad_norm": 0.188343029763607, "learning_rate": 0.00014676489536460014, "loss": 0.8789, "step": 1896 }, { "epoch": 0.36494805694497884, "grad_norm": 0.18783484827182048, "learning_rate": 0.0001467098109509116, "loss": 0.8537, "step": 1897 }, { "epoch": 0.3651404386302424, "grad_norm": 0.17705562113204906, "learning_rate": 0.00014665470840286615, "loss": 0.7945, "step": 1898 }, { "epoch": 0.36533282031550596, "grad_norm": 0.20334625718742802, "learning_rate": 0.00014659958774185654, "loss": 0.8385, "step": 1899 }, { "epoch": 0.36552520200076954, "grad_norm": 0.15907319390704466, "learning_rate": 0.0001465444489892825, "loss": 0.8112, "step": 1900 }, { "epoch": 0.3657175836860331, "grad_norm": 0.17924957648447495, "learning_rate": 0.00014648929216655077, "loss": 0.8519, "step": 1901 }, { "epoch": 0.36590996537129666, "grad_norm": 0.19464996318954686, "learning_rate": 0.00014643411729507516, "loss": 0.8581, "step": 1902 }, { "epoch": 0.3661023470565602, "grad_norm": 0.17483220303303595, "learning_rate": 0.00014637892439627642, "loss": 0.8204, "step": 1903 }, { "epoch": 0.3662947287418238, "grad_norm": 0.2034142206470253, "learning_rate": 0.0001463237134915824, "loss": 0.8745, "step": 1904 }, { "epoch": 0.3664871104270873, "grad_norm": 0.1929542477083979, "learning_rate": 0.0001462684846024278, "loss": 0.8389, "step": 1905 }, { "epoch": 0.3666794921123509, "grad_norm": 0.18055937546522993, "learning_rate": 0.00014621323775025445, "loss": 0.8015, "step": 1906 }, { "epoch": 0.3668718737976145, "grad_norm": 0.19875760639323883, "learning_rate": 0.000146157972956511, "loss": 0.8132, "step": 1907 }, { "epoch": 0.367064255482878, "grad_norm": 0.19409525236287795, "learning_rate": 0.00014610269024265316, "loss": 0.863, "step": 1908 }, { "epoch": 0.3672566371681416, "grad_norm": 0.21031762214827826, "learning_rate": 0.00014604738963014364, "loss": 0.8117, "step": 1909 }, { "epoch": 0.36744901885340514, "grad_norm": 0.10560023876152874, "learning_rate": 0.00014599207114045202, "loss": 0.6594, "step": 1910 }, { "epoch": 0.36764140053866873, "grad_norm": 0.1921432895246697, "learning_rate": 0.00014593673479505482, "loss": 0.8302, "step": 1911 }, { "epoch": 0.36783378222393226, "grad_norm": 0.1648248501763874, "learning_rate": 0.0001458813806154355, "loss": 0.7985, "step": 1912 }, { "epoch": 0.36802616390919585, "grad_norm": 0.19286689849524033, "learning_rate": 0.00014582600862308447, "loss": 0.8797, "step": 1913 }, { "epoch": 0.36821854559445943, "grad_norm": 0.09726591349795057, "learning_rate": 0.00014577061883949913, "loss": 0.5721, "step": 1914 }, { "epoch": 0.36841092727972297, "grad_norm": 0.18610107170224663, "learning_rate": 0.00014571521128618357, "loss": 0.784, "step": 1915 }, { "epoch": 0.36860330896498655, "grad_norm": 0.193534833147241, "learning_rate": 0.00014565978598464894, "loss": 0.8372, "step": 1916 }, { "epoch": 0.3687956906502501, "grad_norm": 0.20886586709982, "learning_rate": 0.00014560434295641337, "loss": 0.8226, "step": 1917 }, { "epoch": 0.3689880723355137, "grad_norm": 0.18354207895129288, "learning_rate": 0.0001455488822230016, "loss": 0.7684, "step": 1918 }, { "epoch": 0.3691804540207772, "grad_norm": 0.1908970523853877, "learning_rate": 0.00014549340380594545, "loss": 0.8152, "step": 1919 }, { "epoch": 0.3693728357060408, "grad_norm": 0.1878713587216557, "learning_rate": 0.00014543790772678357, "loss": 0.7978, "step": 1920 }, { "epoch": 0.3695652173913043, "grad_norm": 0.19004053822723505, "learning_rate": 0.00014538239400706146, "loss": 0.8073, "step": 1921 }, { "epoch": 0.3697575990765679, "grad_norm": 0.18097191716386565, "learning_rate": 0.00014532686266833142, "loss": 0.8254, "step": 1922 }, { "epoch": 0.3699499807618315, "grad_norm": 0.10486146830834761, "learning_rate": 0.00014527131373215263, "loss": 0.6299, "step": 1923 }, { "epoch": 0.37014236244709503, "grad_norm": 0.21732073810314176, "learning_rate": 0.00014521574722009114, "loss": 0.8116, "step": 1924 }, { "epoch": 0.3703347441323586, "grad_norm": 0.0956497920247047, "learning_rate": 0.00014516016315371975, "loss": 0.5858, "step": 1925 }, { "epoch": 0.37052712581762215, "grad_norm": 0.18860062046900528, "learning_rate": 0.00014510456155461806, "loss": 0.8458, "step": 1926 }, { "epoch": 0.37071950750288574, "grad_norm": 0.19169713848926678, "learning_rate": 0.00014504894244437265, "loss": 0.792, "step": 1927 }, { "epoch": 0.37091188918814927, "grad_norm": 0.2104045971300326, "learning_rate": 0.00014499330584457666, "loss": 0.8546, "step": 1928 }, { "epoch": 0.37110427087341286, "grad_norm": 0.19743093054988484, "learning_rate": 0.00014493765177683016, "loss": 0.8181, "step": 1929 }, { "epoch": 0.3712966525586764, "grad_norm": 0.1742298863832253, "learning_rate": 0.00014488198026274006, "loss": 0.8357, "step": 1930 }, { "epoch": 0.37148903424394, "grad_norm": 0.18967929379021023, "learning_rate": 0.00014482629132391985, "loss": 0.8548, "step": 1931 }, { "epoch": 0.37168141592920356, "grad_norm": 0.1919666476113764, "learning_rate": 0.00014477058498198992, "loss": 0.8323, "step": 1932 }, { "epoch": 0.3718737976144671, "grad_norm": 0.11285146107299893, "learning_rate": 0.00014471486125857742, "loss": 0.6313, "step": 1933 }, { "epoch": 0.3720661792997307, "grad_norm": 0.1849234717564937, "learning_rate": 0.0001446591201753162, "loss": 0.7716, "step": 1934 }, { "epoch": 0.3722585609849942, "grad_norm": 0.19321564297661317, "learning_rate": 0.0001446033617538469, "loss": 0.8441, "step": 1935 }, { "epoch": 0.3724509426702578, "grad_norm": 0.20099678669324564, "learning_rate": 0.00014454758601581676, "loss": 0.8136, "step": 1936 }, { "epoch": 0.37264332435552133, "grad_norm": 0.18854722871371687, "learning_rate": 0.00014449179298287997, "loss": 0.83, "step": 1937 }, { "epoch": 0.3728357060407849, "grad_norm": 0.2052968960609972, "learning_rate": 0.00014443598267669723, "loss": 0.787, "step": 1938 }, { "epoch": 0.3730280877260485, "grad_norm": 0.21035278135411398, "learning_rate": 0.000144380155118936, "loss": 0.785, "step": 1939 }, { "epoch": 0.37322046941131204, "grad_norm": 0.1939545545225855, "learning_rate": 0.00014432431033127056, "loss": 0.876, "step": 1940 }, { "epoch": 0.3734128510965756, "grad_norm": 0.20365975419545396, "learning_rate": 0.0001442684483353817, "loss": 0.8139, "step": 1941 }, { "epoch": 0.37360523278183916, "grad_norm": 0.17974859806986296, "learning_rate": 0.00014421256915295695, "loss": 0.7273, "step": 1942 }, { "epoch": 0.37379761446710275, "grad_norm": 0.19777627038637158, "learning_rate": 0.00014415667280569063, "loss": 0.7714, "step": 1943 }, { "epoch": 0.3739899961523663, "grad_norm": 0.2636171390381799, "learning_rate": 0.00014410075931528356, "loss": 0.8494, "step": 1944 }, { "epoch": 0.37418237783762986, "grad_norm": 0.19491477304269125, "learning_rate": 0.00014404482870344323, "loss": 0.8277, "step": 1945 }, { "epoch": 0.3743747595228934, "grad_norm": 0.19162766690487806, "learning_rate": 0.00014398888099188394, "loss": 0.8629, "step": 1946 }, { "epoch": 0.374567141208157, "grad_norm": 0.18863055635037818, "learning_rate": 0.00014393291620232645, "loss": 0.8025, "step": 1947 }, { "epoch": 0.37475952289342057, "grad_norm": 0.19410207786112507, "learning_rate": 0.00014387693435649826, "loss": 0.8261, "step": 1948 }, { "epoch": 0.3749519045786841, "grad_norm": 0.18956390298600753, "learning_rate": 0.00014382093547613337, "loss": 0.8545, "step": 1949 }, { "epoch": 0.3751442862639477, "grad_norm": 0.2090352833806316, "learning_rate": 0.00014376491958297262, "loss": 0.7997, "step": 1950 }, { "epoch": 0.3753366679492112, "grad_norm": 0.19408904425853973, "learning_rate": 0.00014370888669876317, "loss": 0.8531, "step": 1951 }, { "epoch": 0.3755290496344748, "grad_norm": 0.2063819867073976, "learning_rate": 0.00014365283684525894, "loss": 0.8227, "step": 1952 }, { "epoch": 0.37572143131973834, "grad_norm": 0.185194380512611, "learning_rate": 0.00014359677004422046, "loss": 0.7576, "step": 1953 }, { "epoch": 0.37591381300500193, "grad_norm": 0.1859951886627826, "learning_rate": 0.00014354068631741476, "loss": 0.8448, "step": 1954 }, { "epoch": 0.37610619469026546, "grad_norm": 0.2026836259905788, "learning_rate": 0.00014348458568661548, "loss": 0.8965, "step": 1955 }, { "epoch": 0.37629857637552905, "grad_norm": 0.17895767890125047, "learning_rate": 0.0001434284681736028, "loss": 0.8378, "step": 1956 }, { "epoch": 0.37649095806079264, "grad_norm": 0.20286230345841605, "learning_rate": 0.00014337233380016352, "loss": 0.8732, "step": 1957 }, { "epoch": 0.37668333974605617, "grad_norm": 0.18943498167376266, "learning_rate": 0.0001433161825880909, "loss": 0.8262, "step": 1958 }, { "epoch": 0.37687572143131975, "grad_norm": 0.19132268067365504, "learning_rate": 0.00014326001455918478, "loss": 0.8274, "step": 1959 }, { "epoch": 0.3770681031165833, "grad_norm": 0.10948577708959333, "learning_rate": 0.0001432038297352515, "loss": 0.6328, "step": 1960 }, { "epoch": 0.3772604848018469, "grad_norm": 0.20134567903897052, "learning_rate": 0.000143147628138104, "loss": 0.8242, "step": 1961 }, { "epoch": 0.3774528664871104, "grad_norm": 0.18529890837900348, "learning_rate": 0.0001430914097895616, "loss": 0.8563, "step": 1962 }, { "epoch": 0.377645248172374, "grad_norm": 0.18865332511813085, "learning_rate": 0.0001430351747114503, "loss": 0.8357, "step": 1963 }, { "epoch": 0.3778376298576376, "grad_norm": 0.1981568890803525, "learning_rate": 0.0001429789229256024, "loss": 0.7722, "step": 1964 }, { "epoch": 0.3780300115429011, "grad_norm": 0.10743136744523804, "learning_rate": 0.0001429226544538568, "loss": 0.5769, "step": 1965 }, { "epoch": 0.3782223932281647, "grad_norm": 0.11633805234653162, "learning_rate": 0.00014286636931805887, "loss": 0.5955, "step": 1966 }, { "epoch": 0.37841477491342823, "grad_norm": 0.1838699874413837, "learning_rate": 0.00014281006754006045, "loss": 0.8153, "step": 1967 }, { "epoch": 0.3786071565986918, "grad_norm": 0.18529946627282498, "learning_rate": 0.0001427537491417198, "loss": 0.8221, "step": 1968 }, { "epoch": 0.37879953828395535, "grad_norm": 0.1740722626500716, "learning_rate": 0.0001426974141449017, "loss": 0.7543, "step": 1969 }, { "epoch": 0.37899191996921894, "grad_norm": 0.1897376527998149, "learning_rate": 0.00014264106257147732, "loss": 0.8213, "step": 1970 }, { "epoch": 0.37918430165448247, "grad_norm": 0.20610829760058502, "learning_rate": 0.00014258469444332423, "loss": 0.8272, "step": 1971 }, { "epoch": 0.37937668333974606, "grad_norm": 0.18400015028047492, "learning_rate": 0.00014252830978232656, "loss": 0.8853, "step": 1972 }, { "epoch": 0.37956906502500964, "grad_norm": 0.18391796837179789, "learning_rate": 0.00014247190861037474, "loss": 0.7923, "step": 1973 }, { "epoch": 0.3797614467102732, "grad_norm": 0.20255650443146625, "learning_rate": 0.00014241549094936567, "loss": 0.8468, "step": 1974 }, { "epoch": 0.37995382839553676, "grad_norm": 0.18719252809323614, "learning_rate": 0.00014235905682120255, "loss": 0.7941, "step": 1975 }, { "epoch": 0.3801462100808003, "grad_norm": 0.19331381951702725, "learning_rate": 0.00014230260624779512, "loss": 0.8174, "step": 1976 }, { "epoch": 0.3803385917660639, "grad_norm": 0.19035019547546614, "learning_rate": 0.00014224613925105947, "loss": 0.7668, "step": 1977 }, { "epoch": 0.3805309734513274, "grad_norm": 0.1839747212319997, "learning_rate": 0.00014218965585291793, "loss": 0.826, "step": 1978 }, { "epoch": 0.380723355136591, "grad_norm": 0.18940137042996208, "learning_rate": 0.00014213315607529938, "loss": 0.8206, "step": 1979 }, { "epoch": 0.38091573682185453, "grad_norm": 0.1955880189135607, "learning_rate": 0.00014207663994013895, "loss": 0.7987, "step": 1980 }, { "epoch": 0.3811081185071181, "grad_norm": 0.18299227556986122, "learning_rate": 0.00014202010746937815, "loss": 0.831, "step": 1981 }, { "epoch": 0.3813005001923817, "grad_norm": 0.195572919878127, "learning_rate": 0.00014196355868496485, "loss": 0.8386, "step": 1982 }, { "epoch": 0.38149288187764524, "grad_norm": 0.1898600652229428, "learning_rate": 0.00014190699360885323, "loss": 0.8249, "step": 1983 }, { "epoch": 0.3816852635629088, "grad_norm": 0.216878991252128, "learning_rate": 0.00014185041226300374, "loss": 0.7588, "step": 1984 }, { "epoch": 0.38187764524817236, "grad_norm": 0.18559492357238982, "learning_rate": 0.0001417938146693833, "loss": 0.7846, "step": 1985 }, { "epoch": 0.38207002693343595, "grad_norm": 0.21280594713178863, "learning_rate": 0.000141737200849965, "loss": 0.8085, "step": 1986 }, { "epoch": 0.3822624086186995, "grad_norm": 0.19152162223376373, "learning_rate": 0.00014168057082672826, "loss": 0.887, "step": 1987 }, { "epoch": 0.38245479030396307, "grad_norm": 0.17809414012196687, "learning_rate": 0.00014162392462165884, "loss": 0.8286, "step": 1988 }, { "epoch": 0.38264717198922665, "grad_norm": 0.18343479261064766, "learning_rate": 0.00014156726225674873, "loss": 0.7852, "step": 1989 }, { "epoch": 0.3828395536744902, "grad_norm": 0.1960927655256895, "learning_rate": 0.0001415105837539962, "loss": 0.8337, "step": 1990 }, { "epoch": 0.3830319353597538, "grad_norm": 0.19356948074865094, "learning_rate": 0.0001414538891354058, "loss": 0.8134, "step": 1991 }, { "epoch": 0.3832243170450173, "grad_norm": 0.18602349619876016, "learning_rate": 0.00014139717842298835, "loss": 0.8237, "step": 1992 }, { "epoch": 0.3834166987302809, "grad_norm": 0.17937512457320018, "learning_rate": 0.00014134045163876087, "loss": 0.8401, "step": 1993 }, { "epoch": 0.3836090804155444, "grad_norm": 0.20129196054068402, "learning_rate": 0.00014128370880474666, "loss": 0.8254, "step": 1994 }, { "epoch": 0.383801462100808, "grad_norm": 0.18870899733260835, "learning_rate": 0.0001412269499429753, "loss": 0.8205, "step": 1995 }, { "epoch": 0.38399384378607154, "grad_norm": 0.1953859974774167, "learning_rate": 0.00014117017507548244, "loss": 0.8317, "step": 1996 }, { "epoch": 0.38418622547133513, "grad_norm": 0.19705963943226953, "learning_rate": 0.00014111338422431014, "loss": 0.8309, "step": 1997 }, { "epoch": 0.3843786071565987, "grad_norm": 0.21251903177080292, "learning_rate": 0.00014105657741150647, "loss": 0.7558, "step": 1998 }, { "epoch": 0.38457098884186225, "grad_norm": 0.19573815047658588, "learning_rate": 0.00014099975465912582, "loss": 0.8405, "step": 1999 }, { "epoch": 0.38476337052712584, "grad_norm": 0.16966959794680872, "learning_rate": 0.00014094291598922877, "loss": 0.8139, "step": 2000 }, { "epoch": 0.38495575221238937, "grad_norm": 0.19264732283407196, "learning_rate": 0.00014088606142388202, "loss": 0.8089, "step": 2001 }, { "epoch": 0.38514813389765296, "grad_norm": 0.18313453737738666, "learning_rate": 0.00014082919098515846, "loss": 0.808, "step": 2002 }, { "epoch": 0.3853405155829165, "grad_norm": 0.18738333792295495, "learning_rate": 0.0001407723046951372, "loss": 0.7552, "step": 2003 }, { "epoch": 0.3855328972681801, "grad_norm": 0.1833372069148496, "learning_rate": 0.0001407154025759034, "loss": 0.813, "step": 2004 }, { "epoch": 0.3857252789534436, "grad_norm": 0.1964132198956144, "learning_rate": 0.00014065848464954846, "loss": 0.8564, "step": 2005 }, { "epoch": 0.3859176606387072, "grad_norm": 0.1787400609764939, "learning_rate": 0.00014060155093816986, "loss": 0.7865, "step": 2006 }, { "epoch": 0.3861100423239708, "grad_norm": 0.1244212565990295, "learning_rate": 0.00014054460146387124, "loss": 0.6677, "step": 2007 }, { "epoch": 0.3863024240092343, "grad_norm": 0.19652322891694549, "learning_rate": 0.00014048763624876235, "loss": 0.7653, "step": 2008 }, { "epoch": 0.3864948056944979, "grad_norm": 0.23154713358790088, "learning_rate": 0.00014043065531495902, "loss": 0.8104, "step": 2009 }, { "epoch": 0.38668718737976143, "grad_norm": 0.19818095910902267, "learning_rate": 0.00014037365868458326, "loss": 0.7784, "step": 2010 }, { "epoch": 0.386879569065025, "grad_norm": 0.19593665197510335, "learning_rate": 0.00014031664637976306, "loss": 0.7976, "step": 2011 }, { "epoch": 0.38707195075028855, "grad_norm": 0.10208525848504363, "learning_rate": 0.0001402596184226326, "loss": 0.5989, "step": 2012 }, { "epoch": 0.38726433243555214, "grad_norm": 0.19193381878154137, "learning_rate": 0.00014020257483533208, "loss": 0.8174, "step": 2013 }, { "epoch": 0.38745671412081567, "grad_norm": 0.21197070909692842, "learning_rate": 0.0001401455156400078, "loss": 0.8237, "step": 2014 }, { "epoch": 0.38764909580607926, "grad_norm": 0.19748079651796896, "learning_rate": 0.0001400884408588121, "loss": 0.8221, "step": 2015 }, { "epoch": 0.38784147749134285, "grad_norm": 0.20593262011093835, "learning_rate": 0.0001400313505139034, "loss": 0.8582, "step": 2016 }, { "epoch": 0.3880338591766064, "grad_norm": 0.17120770015766984, "learning_rate": 0.00013997424462744605, "loss": 0.801, "step": 2017 }, { "epoch": 0.38822624086186996, "grad_norm": 0.1894210454415694, "learning_rate": 0.00013991712322161063, "loss": 0.8351, "step": 2018 }, { "epoch": 0.3884186225471335, "grad_norm": 0.20480007083316226, "learning_rate": 0.00013985998631857357, "loss": 0.8073, "step": 2019 }, { "epoch": 0.3886110042323971, "grad_norm": 0.18485795980065517, "learning_rate": 0.0001398028339405174, "loss": 0.8319, "step": 2020 }, { "epoch": 0.3888033859176606, "grad_norm": 0.11272465342466802, "learning_rate": 0.00013974566610963068, "loss": 0.6227, "step": 2021 }, { "epoch": 0.3889957676029242, "grad_norm": 0.20537197710736402, "learning_rate": 0.00013968848284810786, "loss": 0.8022, "step": 2022 }, { "epoch": 0.3891881492881878, "grad_norm": 0.19270344110385784, "learning_rate": 0.0001396312841781495, "loss": 0.8007, "step": 2023 }, { "epoch": 0.3893805309734513, "grad_norm": 0.19703326591897555, "learning_rate": 0.00013957407012196203, "loss": 0.8806, "step": 2024 }, { "epoch": 0.3895729126587149, "grad_norm": 0.21187689851441155, "learning_rate": 0.000139516840701758, "loss": 0.8603, "step": 2025 }, { "epoch": 0.38976529434397844, "grad_norm": 0.19384311507528357, "learning_rate": 0.0001394595959397558, "loss": 0.8506, "step": 2026 }, { "epoch": 0.38995767602924203, "grad_norm": 0.17017365816618973, "learning_rate": 0.00013940233585817984, "loss": 0.7967, "step": 2027 }, { "epoch": 0.39015005771450556, "grad_norm": 0.19005190533615857, "learning_rate": 0.0001393450604792604, "loss": 0.7835, "step": 2028 }, { "epoch": 0.39034243939976915, "grad_norm": 0.18509203701871843, "learning_rate": 0.00013928776982523384, "loss": 0.8356, "step": 2029 }, { "epoch": 0.3905348210850327, "grad_norm": 0.18811850200977834, "learning_rate": 0.00013923046391834228, "loss": 0.8521, "step": 2030 }, { "epoch": 0.39072720277029627, "grad_norm": 0.19500663134838636, "learning_rate": 0.0001391731427808339, "loss": 0.754, "step": 2031 }, { "epoch": 0.39091958445555985, "grad_norm": 0.10540680835672976, "learning_rate": 0.0001391158064349627, "loss": 0.6172, "step": 2032 }, { "epoch": 0.3911119661408234, "grad_norm": 0.20827377531987615, "learning_rate": 0.00013905845490298866, "loss": 0.7703, "step": 2033 }, { "epoch": 0.391304347826087, "grad_norm": 0.20359694767831515, "learning_rate": 0.0001390010882071776, "loss": 0.8566, "step": 2034 }, { "epoch": 0.3914967295113505, "grad_norm": 0.10464646711999277, "learning_rate": 0.00013894370636980126, "loss": 0.6064, "step": 2035 }, { "epoch": 0.3916891111966141, "grad_norm": 0.18140883738899385, "learning_rate": 0.00013888630941313727, "loss": 0.7906, "step": 2036 }, { "epoch": 0.3918814928818776, "grad_norm": 0.21989092812578492, "learning_rate": 0.000138828897359469, "loss": 0.8888, "step": 2037 }, { "epoch": 0.3920738745671412, "grad_norm": 0.11387066635477756, "learning_rate": 0.0001387714702310859, "loss": 0.6128, "step": 2038 }, { "epoch": 0.39226625625240474, "grad_norm": 0.19038647658633198, "learning_rate": 0.00013871402805028314, "loss": 0.864, "step": 2039 }, { "epoch": 0.39245863793766833, "grad_norm": 0.09979860794912114, "learning_rate": 0.00013865657083936164, "loss": 0.5816, "step": 2040 }, { "epoch": 0.3926510196229319, "grad_norm": 0.2025749370865704, "learning_rate": 0.00013859909862062842, "loss": 0.842, "step": 2041 }, { "epoch": 0.39284340130819545, "grad_norm": 0.18215220534459042, "learning_rate": 0.0001385416114163961, "loss": 0.8564, "step": 2042 }, { "epoch": 0.39303578299345904, "grad_norm": 0.20608182980734802, "learning_rate": 0.0001384841092489832, "loss": 0.7841, "step": 2043 }, { "epoch": 0.39322816467872257, "grad_norm": 0.1895727840643053, "learning_rate": 0.00013842659214071404, "loss": 0.7978, "step": 2044 }, { "epoch": 0.39342054636398616, "grad_norm": 0.1790605254656987, "learning_rate": 0.00013836906011391878, "loss": 0.8586, "step": 2045 }, { "epoch": 0.3936129280492497, "grad_norm": 0.17009674315122855, "learning_rate": 0.00013831151319093323, "loss": 0.8251, "step": 2046 }, { "epoch": 0.3938053097345133, "grad_norm": 0.1921573161409087, "learning_rate": 0.0001382539513940992, "loss": 0.7702, "step": 2047 }, { "epoch": 0.39399769141977686, "grad_norm": 0.19798892696251835, "learning_rate": 0.0001381963747457641, "loss": 0.8607, "step": 2048 }, { "epoch": 0.3941900731050404, "grad_norm": 0.18605696505060249, "learning_rate": 0.0001381387832682812, "loss": 0.7806, "step": 2049 }, { "epoch": 0.394382454790304, "grad_norm": 0.1780450498838225, "learning_rate": 0.00013808117698400947, "loss": 0.7965, "step": 2050 }, { "epoch": 0.3945748364755675, "grad_norm": 0.1923290548863538, "learning_rate": 0.00013802355591531365, "loss": 0.8253, "step": 2051 }, { "epoch": 0.3947672181608311, "grad_norm": 0.20004998240214336, "learning_rate": 0.00013796592008456427, "loss": 0.7952, "step": 2052 }, { "epoch": 0.39495959984609463, "grad_norm": 0.10544687407138997, "learning_rate": 0.00013790826951413747, "loss": 0.6176, "step": 2053 }, { "epoch": 0.3951519815313582, "grad_norm": 0.19639601444430194, "learning_rate": 0.00013785060422641526, "loss": 0.7821, "step": 2054 }, { "epoch": 0.39534436321662175, "grad_norm": 0.2039615743252326, "learning_rate": 0.0001377929242437852, "loss": 0.804, "step": 2055 }, { "epoch": 0.39553674490188534, "grad_norm": 0.1971336601881658, "learning_rate": 0.00013773522958864076, "loss": 0.7758, "step": 2056 }, { "epoch": 0.3957291265871489, "grad_norm": 0.19940490300240254, "learning_rate": 0.0001376775202833809, "loss": 0.8236, "step": 2057 }, { "epoch": 0.39592150827241246, "grad_norm": 0.19973337561243484, "learning_rate": 0.0001376197963504104, "loss": 0.8749, "step": 2058 }, { "epoch": 0.39611388995767605, "grad_norm": 0.19862197484017913, "learning_rate": 0.00013756205781213963, "loss": 0.7865, "step": 2059 }, { "epoch": 0.3963062716429396, "grad_norm": 0.20667347972478525, "learning_rate": 0.00013750430469098478, "loss": 0.7924, "step": 2060 }, { "epoch": 0.39649865332820317, "grad_norm": 0.19866984639754987, "learning_rate": 0.00013744653700936753, "loss": 0.822, "step": 2061 }, { "epoch": 0.3966910350134667, "grad_norm": 0.2070665479418815, "learning_rate": 0.00013738875478971525, "loss": 0.7721, "step": 2062 }, { "epoch": 0.3968834166987303, "grad_norm": 0.1842639342520598, "learning_rate": 0.00013733095805446106, "loss": 0.8958, "step": 2063 }, { "epoch": 0.3970757983839938, "grad_norm": 0.20444591466267553, "learning_rate": 0.0001372731468260436, "loss": 0.7782, "step": 2064 }, { "epoch": 0.3972681800692574, "grad_norm": 0.1953044665362584, "learning_rate": 0.0001372153211269072, "loss": 0.8189, "step": 2065 }, { "epoch": 0.397460561754521, "grad_norm": 0.18603989999010528, "learning_rate": 0.00013715748097950175, "loss": 0.8149, "step": 2066 }, { "epoch": 0.3976529434397845, "grad_norm": 0.196321292860681, "learning_rate": 0.00013709962640628284, "loss": 0.8321, "step": 2067 }, { "epoch": 0.3978453251250481, "grad_norm": 0.2023896216318482, "learning_rate": 0.00013704175742971158, "loss": 0.8441, "step": 2068 }, { "epoch": 0.39803770681031164, "grad_norm": 0.1985186654572538, "learning_rate": 0.0001369838740722547, "loss": 0.7895, "step": 2069 }, { "epoch": 0.39823008849557523, "grad_norm": 0.18322845290391132, "learning_rate": 0.0001369259763563845, "loss": 0.7486, "step": 2070 }, { "epoch": 0.39842247018083876, "grad_norm": 0.20912575549445195, "learning_rate": 0.0001368680643045789, "loss": 0.798, "step": 2071 }, { "epoch": 0.39861485186610235, "grad_norm": 0.18411097302728102, "learning_rate": 0.0001368101379393213, "loss": 0.788, "step": 2072 }, { "epoch": 0.39880723355136594, "grad_norm": 0.20077637783905486, "learning_rate": 0.00013675219728310077, "loss": 0.8017, "step": 2073 }, { "epoch": 0.39899961523662947, "grad_norm": 0.20296950973867794, "learning_rate": 0.00013669424235841185, "loss": 0.805, "step": 2074 }, { "epoch": 0.39919199692189306, "grad_norm": 0.19543214385477387, "learning_rate": 0.00013663627318775457, "loss": 0.8228, "step": 2075 }, { "epoch": 0.3993843786071566, "grad_norm": 0.1899960704107546, "learning_rate": 0.00013657828979363467, "loss": 0.8267, "step": 2076 }, { "epoch": 0.3995767602924202, "grad_norm": 0.19760696103802325, "learning_rate": 0.00013652029219856322, "loss": 0.8236, "step": 2077 }, { "epoch": 0.3997691419776837, "grad_norm": 0.11023095539741078, "learning_rate": 0.00013646228042505694, "loss": 0.6701, "step": 2078 }, { "epoch": 0.3999615236629473, "grad_norm": 0.20016245657987147, "learning_rate": 0.0001364042544956379, "loss": 0.7808, "step": 2079 }, { "epoch": 0.4001539053482108, "grad_norm": 0.19507928415255588, "learning_rate": 0.00013634621443283387, "loss": 0.8257, "step": 2080 }, { "epoch": 0.4003462870334744, "grad_norm": 0.21076808560457228, "learning_rate": 0.00013628816025917798, "loss": 0.8165, "step": 2081 }, { "epoch": 0.400538668718738, "grad_norm": 0.19557683724863084, "learning_rate": 0.00013623009199720883, "loss": 0.7979, "step": 2082 }, { "epoch": 0.40073105040400153, "grad_norm": 0.17531365456664186, "learning_rate": 0.00013617200966947052, "loss": 0.8359, "step": 2083 }, { "epoch": 0.4009234320892651, "grad_norm": 0.17376006355009666, "learning_rate": 0.00013611391329851263, "loss": 0.8421, "step": 2084 }, { "epoch": 0.40111581377452865, "grad_norm": 0.1979506390747952, "learning_rate": 0.00013605580290689014, "loss": 0.8214, "step": 2085 }, { "epoch": 0.40130819545979224, "grad_norm": 0.19434964272587868, "learning_rate": 0.00013599767851716352, "loss": 0.8572, "step": 2086 }, { "epoch": 0.40150057714505577, "grad_norm": 0.2137720644488875, "learning_rate": 0.00013593954015189867, "loss": 0.8342, "step": 2087 }, { "epoch": 0.40169295883031936, "grad_norm": 0.19133001964879467, "learning_rate": 0.0001358813878336669, "loss": 0.8344, "step": 2088 }, { "epoch": 0.4018853405155829, "grad_norm": 0.22345112196955072, "learning_rate": 0.00013582322158504494, "loss": 0.7862, "step": 2089 }, { "epoch": 0.4020777222008465, "grad_norm": 0.19826991876370392, "learning_rate": 0.00013576504142861495, "loss": 0.8156, "step": 2090 }, { "epoch": 0.40227010388611006, "grad_norm": 0.1904782289181967, "learning_rate": 0.00013570684738696445, "loss": 0.851, "step": 2091 }, { "epoch": 0.4024624855713736, "grad_norm": 0.21004884376975427, "learning_rate": 0.0001356486394826863, "loss": 0.7789, "step": 2092 }, { "epoch": 0.4026548672566372, "grad_norm": 0.18906199792909312, "learning_rate": 0.00013559041773837898, "loss": 0.8753, "step": 2093 }, { "epoch": 0.4028472489419007, "grad_norm": 0.18774810942754963, "learning_rate": 0.00013553218217664603, "loss": 0.7912, "step": 2094 }, { "epoch": 0.4030396306271643, "grad_norm": 0.18635133847364213, "learning_rate": 0.00013547393282009655, "loss": 0.8167, "step": 2095 }, { "epoch": 0.40323201231242783, "grad_norm": 0.20661999925940688, "learning_rate": 0.00013541566969134497, "loss": 0.7495, "step": 2096 }, { "epoch": 0.4034243939976914, "grad_norm": 0.18260335946694758, "learning_rate": 0.000135357392813011, "loss": 0.8294, "step": 2097 }, { "epoch": 0.403616775682955, "grad_norm": 0.2032897647222455, "learning_rate": 0.00013529910220771977, "loss": 0.7711, "step": 2098 }, { "epoch": 0.40380915736821854, "grad_norm": 0.18913628662293364, "learning_rate": 0.00013524079789810163, "loss": 0.771, "step": 2099 }, { "epoch": 0.40400153905348213, "grad_norm": 0.21711813466284147, "learning_rate": 0.00013518247990679241, "loss": 0.7851, "step": 2100 }, { "epoch": 0.40419392073874566, "grad_norm": 0.2874863588805058, "learning_rate": 0.00013512414825643312, "loss": 0.8236, "step": 2101 }, { "epoch": 0.40438630242400925, "grad_norm": 0.21173582246121936, "learning_rate": 0.0001350658029696701, "loss": 0.7886, "step": 2102 }, { "epoch": 0.4045786841092728, "grad_norm": 0.18847185198526403, "learning_rate": 0.00013500744406915504, "loss": 0.7644, "step": 2103 }, { "epoch": 0.40477106579453637, "grad_norm": 0.18620282172872146, "learning_rate": 0.00013494907157754485, "loss": 0.8379, "step": 2104 }, { "epoch": 0.4049634474797999, "grad_norm": 0.21783073979454784, "learning_rate": 0.0001348906855175017, "loss": 0.8179, "step": 2105 }, { "epoch": 0.4051558291650635, "grad_norm": 0.1796432710608437, "learning_rate": 0.00013483228591169315, "loss": 0.8182, "step": 2106 }, { "epoch": 0.4053482108503271, "grad_norm": 0.1833400315363875, "learning_rate": 0.0001347738727827919, "loss": 0.7565, "step": 2107 }, { "epoch": 0.4055405925355906, "grad_norm": 0.1899173344230086, "learning_rate": 0.0001347154461534759, "loss": 0.7546, "step": 2108 }, { "epoch": 0.4057329742208542, "grad_norm": 0.18732642401833796, "learning_rate": 0.00013465700604642846, "loss": 0.8032, "step": 2109 }, { "epoch": 0.4059253559061177, "grad_norm": 0.19042629595755922, "learning_rate": 0.000134598552484338, "loss": 0.8201, "step": 2110 }, { "epoch": 0.4061177375913813, "grad_norm": 0.09685078585720215, "learning_rate": 0.00013454008548989817, "loss": 0.5776, "step": 2111 }, { "epoch": 0.40631011927664484, "grad_norm": 0.19642674930301754, "learning_rate": 0.0001344816050858079, "loss": 0.7669, "step": 2112 }, { "epoch": 0.40650250096190843, "grad_norm": 0.2098155359775701, "learning_rate": 0.00013442311129477133, "loss": 0.8708, "step": 2113 }, { "epoch": 0.40669488264717196, "grad_norm": 0.11021513473314591, "learning_rate": 0.0001343646041394977, "loss": 0.6203, "step": 2114 }, { "epoch": 0.40688726433243555, "grad_norm": 0.19677867612735583, "learning_rate": 0.00013430608364270157, "loss": 0.8274, "step": 2115 }, { "epoch": 0.40707964601769914, "grad_norm": 0.17763361613135822, "learning_rate": 0.00013424754982710254, "loss": 0.8562, "step": 2116 }, { "epoch": 0.40727202770296267, "grad_norm": 0.17315400501411007, "learning_rate": 0.0001341890027154255, "loss": 0.7891, "step": 2117 }, { "epoch": 0.40746440938822626, "grad_norm": 0.17972006561940979, "learning_rate": 0.00013413044233040043, "loss": 0.829, "step": 2118 }, { "epoch": 0.4076567910734898, "grad_norm": 0.20007464933338268, "learning_rate": 0.00013407186869476254, "loss": 0.8142, "step": 2119 }, { "epoch": 0.4078491727587534, "grad_norm": 0.1775620348233157, "learning_rate": 0.00013401328183125207, "loss": 0.7595, "step": 2120 }, { "epoch": 0.4080415544440169, "grad_norm": 0.185990630592936, "learning_rate": 0.0001339546817626145, "loss": 0.8239, "step": 2121 }, { "epoch": 0.4082339361292805, "grad_norm": 0.20666844095301154, "learning_rate": 0.00013389606851160036, "loss": 0.8223, "step": 2122 }, { "epoch": 0.4084263178145441, "grad_norm": 0.2046393164424845, "learning_rate": 0.00013383744210096536, "loss": 0.8323, "step": 2123 }, { "epoch": 0.4086186994998076, "grad_norm": 0.18216933351953968, "learning_rate": 0.00013377880255347027, "loss": 0.8199, "step": 2124 }, { "epoch": 0.4088110811850712, "grad_norm": 0.17624039731512475, "learning_rate": 0.00013372014989188099, "loss": 0.8281, "step": 2125 }, { "epoch": 0.40900346287033473, "grad_norm": 0.2097943047705052, "learning_rate": 0.0001336614841389685, "loss": 0.7803, "step": 2126 }, { "epoch": 0.4091958445555983, "grad_norm": 0.18092725702765178, "learning_rate": 0.00013360280531750886, "loss": 0.7914, "step": 2127 }, { "epoch": 0.40938822624086185, "grad_norm": 0.1940064249967171, "learning_rate": 0.00013354411345028325, "loss": 0.9342, "step": 2128 }, { "epoch": 0.40958060792612544, "grad_norm": 0.1701891168147079, "learning_rate": 0.0001334854085600778, "loss": 0.8621, "step": 2129 }, { "epoch": 0.40977298961138897, "grad_norm": 0.1873741362931039, "learning_rate": 0.00013342669066968385, "loss": 0.8105, "step": 2130 }, { "epoch": 0.40996537129665256, "grad_norm": 0.18660486365810225, "learning_rate": 0.0001333679598018976, "loss": 0.8154, "step": 2131 }, { "epoch": 0.41015775298191615, "grad_norm": 0.1867222104979636, "learning_rate": 0.00013330921597952056, "loss": 0.8155, "step": 2132 }, { "epoch": 0.4103501346671797, "grad_norm": 0.1823332061407613, "learning_rate": 0.00013325045922535895, "loss": 0.7936, "step": 2133 }, { "epoch": 0.41054251635244327, "grad_norm": 0.17729590110073198, "learning_rate": 0.00013319168956222422, "loss": 0.8425, "step": 2134 }, { "epoch": 0.4107348980377068, "grad_norm": 0.2042095189524555, "learning_rate": 0.0001331329070129328, "loss": 0.8104, "step": 2135 }, { "epoch": 0.4109272797229704, "grad_norm": 0.18685960643185295, "learning_rate": 0.00013307411160030607, "loss": 0.8137, "step": 2136 }, { "epoch": 0.4111196614082339, "grad_norm": 0.1883994431776281, "learning_rate": 0.00013301530334717046, "loss": 0.826, "step": 2137 }, { "epoch": 0.4113120430934975, "grad_norm": 0.20223057437282432, "learning_rate": 0.0001329564822763573, "loss": 0.8737, "step": 2138 }, { "epoch": 0.41150442477876104, "grad_norm": 0.1748349625350118, "learning_rate": 0.000132897648410703, "loss": 0.7878, "step": 2139 }, { "epoch": 0.4116968064640246, "grad_norm": 0.17341117974714193, "learning_rate": 0.00013283880177304892, "loss": 0.8133, "step": 2140 }, { "epoch": 0.4118891881492882, "grad_norm": 0.1836375723511798, "learning_rate": 0.0001327799423862413, "loss": 0.7946, "step": 2141 }, { "epoch": 0.41208156983455174, "grad_norm": 0.10335411060222716, "learning_rate": 0.00013272107027313142, "loss": 0.6079, "step": 2142 }, { "epoch": 0.41227395151981533, "grad_norm": 0.19679993508930851, "learning_rate": 0.00013266218545657541, "loss": 0.8294, "step": 2143 }, { "epoch": 0.41246633320507886, "grad_norm": 0.19116865244428632, "learning_rate": 0.0001326032879594344, "loss": 0.8423, "step": 2144 }, { "epoch": 0.41265871489034245, "grad_norm": 0.19809914819934799, "learning_rate": 0.00013254437780457447, "loss": 0.7929, "step": 2145 }, { "epoch": 0.412851096575606, "grad_norm": 0.20000026554469208, "learning_rate": 0.00013248545501486653, "loss": 0.83, "step": 2146 }, { "epoch": 0.41304347826086957, "grad_norm": 0.18781659036399595, "learning_rate": 0.00013242651961318646, "loss": 0.8102, "step": 2147 }, { "epoch": 0.41323585994613315, "grad_norm": 0.1918292514375016, "learning_rate": 0.000132367571622415, "loss": 0.7902, "step": 2148 }, { "epoch": 0.4134282416313967, "grad_norm": 0.18340771820484952, "learning_rate": 0.00013230861106543776, "loss": 0.87, "step": 2149 }, { "epoch": 0.4136206233166603, "grad_norm": 0.18075449356346493, "learning_rate": 0.0001322496379651453, "loss": 0.8249, "step": 2150 }, { "epoch": 0.4138130050019238, "grad_norm": 0.19773787285700814, "learning_rate": 0.000132190652344433, "loss": 0.8237, "step": 2151 }, { "epoch": 0.4140053866871874, "grad_norm": 0.1852116580249238, "learning_rate": 0.0001321316542262011, "loss": 0.8005, "step": 2152 }, { "epoch": 0.4141977683724509, "grad_norm": 0.18110187273369943, "learning_rate": 0.0001320726436333547, "loss": 0.8038, "step": 2153 }, { "epoch": 0.4143901500577145, "grad_norm": 0.11761165505723885, "learning_rate": 0.00013201362058880376, "loss": 0.6629, "step": 2154 }, { "epoch": 0.41458253174297804, "grad_norm": 0.20065286987606168, "learning_rate": 0.00013195458511546307, "loss": 0.801, "step": 2155 }, { "epoch": 0.41477491342824163, "grad_norm": 0.18160877979041598, "learning_rate": 0.00013189553723625216, "loss": 0.7993, "step": 2156 }, { "epoch": 0.4149672951135052, "grad_norm": 0.18705703311275043, "learning_rate": 0.0001318364769740955, "loss": 0.8465, "step": 2157 }, { "epoch": 0.41515967679876875, "grad_norm": 0.17436892430652293, "learning_rate": 0.00013177740435192233, "loss": 0.7874, "step": 2158 }, { "epoch": 0.41535205848403234, "grad_norm": 0.20171438164070166, "learning_rate": 0.00013171831939266668, "loss": 0.8347, "step": 2159 }, { "epoch": 0.41554444016929587, "grad_norm": 0.20538888944759864, "learning_rate": 0.00013165922211926732, "loss": 0.8622, "step": 2160 }, { "epoch": 0.41573682185455946, "grad_norm": 0.19355314866328302, "learning_rate": 0.0001316001125546679, "loss": 0.7617, "step": 2161 }, { "epoch": 0.415929203539823, "grad_norm": 0.18899665280611921, "learning_rate": 0.00013154099072181676, "loss": 0.7934, "step": 2162 }, { "epoch": 0.4161215852250866, "grad_norm": 0.20753659619271803, "learning_rate": 0.00013148185664366702, "loss": 0.8187, "step": 2163 }, { "epoch": 0.4163139669103501, "grad_norm": 0.18772137088378818, "learning_rate": 0.0001314227103431766, "loss": 0.7976, "step": 2164 }, { "epoch": 0.4165063485956137, "grad_norm": 0.12354591670950481, "learning_rate": 0.0001313635518433081, "loss": 0.6252, "step": 2165 }, { "epoch": 0.4166987302808773, "grad_norm": 0.23259567462584793, "learning_rate": 0.00013130438116702888, "loss": 0.8186, "step": 2166 }, { "epoch": 0.4168911119661408, "grad_norm": 0.19240385825109935, "learning_rate": 0.00013124519833731106, "loss": 0.7992, "step": 2167 }, { "epoch": 0.4170834936514044, "grad_norm": 0.18705684022733315, "learning_rate": 0.00013118600337713146, "loss": 0.761, "step": 2168 }, { "epoch": 0.41727587533666793, "grad_norm": 0.18269716713138323, "learning_rate": 0.00013112679630947156, "loss": 0.828, "step": 2169 }, { "epoch": 0.4174682570219315, "grad_norm": 0.1921098800472185, "learning_rate": 0.00013106757715731758, "loss": 0.8464, "step": 2170 }, { "epoch": 0.41766063870719505, "grad_norm": 0.20765862377205566, "learning_rate": 0.0001310083459436605, "loss": 0.8193, "step": 2171 }, { "epoch": 0.41785302039245864, "grad_norm": 0.1959329157814698, "learning_rate": 0.00013094910269149587, "loss": 0.7807, "step": 2172 }, { "epoch": 0.41804540207772223, "grad_norm": 0.19424757642196386, "learning_rate": 0.00013088984742382396, "loss": 0.8106, "step": 2173 }, { "epoch": 0.41823778376298576, "grad_norm": 0.18303482704581836, "learning_rate": 0.0001308305801636497, "loss": 0.8029, "step": 2174 }, { "epoch": 0.41843016544824935, "grad_norm": 0.1937397271417068, "learning_rate": 0.00013077130093398274, "loss": 0.8461, "step": 2175 }, { "epoch": 0.4186225471335129, "grad_norm": 0.19764884522344495, "learning_rate": 0.00013071200975783723, "loss": 0.792, "step": 2176 }, { "epoch": 0.41881492881877647, "grad_norm": 0.19712528345800687, "learning_rate": 0.00013065270665823205, "loss": 0.8574, "step": 2177 }, { "epoch": 0.41900731050404, "grad_norm": 0.18068009302674073, "learning_rate": 0.00013059339165819083, "loss": 0.775, "step": 2178 }, { "epoch": 0.4191996921893036, "grad_norm": 0.20079453817571657, "learning_rate": 0.00013053406478074156, "loss": 0.8365, "step": 2179 }, { "epoch": 0.4193920738745671, "grad_norm": 0.19312055595416874, "learning_rate": 0.00013047472604891703, "loss": 0.8377, "step": 2180 }, { "epoch": 0.4195844555598307, "grad_norm": 0.19360186851912536, "learning_rate": 0.00013041537548575457, "loss": 0.7528, "step": 2181 }, { "epoch": 0.4197768372450943, "grad_norm": 0.19182014499521918, "learning_rate": 0.00013035601311429612, "loss": 0.8302, "step": 2182 }, { "epoch": 0.4199692189303578, "grad_norm": 0.21564073565102104, "learning_rate": 0.00013029663895758813, "loss": 0.8319, "step": 2183 }, { "epoch": 0.4201616006156214, "grad_norm": 0.19836452698218213, "learning_rate": 0.0001302372530386818, "loss": 0.7664, "step": 2184 }, { "epoch": 0.42035398230088494, "grad_norm": 0.19790598430799552, "learning_rate": 0.00013017785538063276, "loss": 0.7988, "step": 2185 }, { "epoch": 0.42054636398614853, "grad_norm": 0.20967452405085246, "learning_rate": 0.0001301184460065012, "loss": 0.8435, "step": 2186 }, { "epoch": 0.42073874567141206, "grad_norm": 0.19599561778349223, "learning_rate": 0.0001300590249393519, "loss": 0.8106, "step": 2187 }, { "epoch": 0.42093112735667565, "grad_norm": 0.17033913162065414, "learning_rate": 0.00012999959220225414, "loss": 0.84, "step": 2188 }, { "epoch": 0.4211235090419392, "grad_norm": 0.17901821275988733, "learning_rate": 0.00012994014781828178, "loss": 0.8247, "step": 2189 }, { "epoch": 0.42131589072720277, "grad_norm": 0.10753307550105522, "learning_rate": 0.00012988069181051314, "loss": 0.6009, "step": 2190 }, { "epoch": 0.42150827241246636, "grad_norm": 0.18756849481427218, "learning_rate": 0.00012982122420203114, "loss": 0.8349, "step": 2191 }, { "epoch": 0.4217006540977299, "grad_norm": 0.18669218456403772, "learning_rate": 0.00012976174501592313, "loss": 0.8115, "step": 2192 }, { "epoch": 0.4218930357829935, "grad_norm": 0.16789952286646176, "learning_rate": 0.00012970225427528097, "loss": 0.7905, "step": 2193 }, { "epoch": 0.422085417468257, "grad_norm": 0.18160132409010526, "learning_rate": 0.00012964275200320105, "loss": 0.8194, "step": 2194 }, { "epoch": 0.4222777991535206, "grad_norm": 0.18377727382776096, "learning_rate": 0.00012958323822278412, "loss": 0.8213, "step": 2195 }, { "epoch": 0.4224701808387841, "grad_norm": 0.2259457832818271, "learning_rate": 0.00012952371295713556, "loss": 0.8399, "step": 2196 }, { "epoch": 0.4226625625240477, "grad_norm": 0.20101966456160922, "learning_rate": 0.00012946417622936512, "loss": 0.7949, "step": 2197 }, { "epoch": 0.4228549442093113, "grad_norm": 0.21227104015775775, "learning_rate": 0.00012940462806258695, "loss": 0.82, "step": 2198 }, { "epoch": 0.42304732589457483, "grad_norm": 0.18218246146278225, "learning_rate": 0.00012934506847991976, "loss": 0.794, "step": 2199 }, { "epoch": 0.4232397075798384, "grad_norm": 0.20306362947970114, "learning_rate": 0.0001292854975044866, "loss": 0.7759, "step": 2200 }, { "epoch": 0.42343208926510195, "grad_norm": 0.18751902527421718, "learning_rate": 0.00012922591515941498, "loss": 0.8155, "step": 2201 }, { "epoch": 0.42362447095036554, "grad_norm": 0.2028259973332848, "learning_rate": 0.00012916632146783684, "loss": 0.7916, "step": 2202 }, { "epoch": 0.42381685263562907, "grad_norm": 0.1924319929766732, "learning_rate": 0.0001291067164528884, "loss": 0.8634, "step": 2203 }, { "epoch": 0.42400923432089266, "grad_norm": 0.2059749661123503, "learning_rate": 0.00012904710013771054, "loss": 0.7585, "step": 2204 }, { "epoch": 0.4242016160061562, "grad_norm": 0.1992264708709048, "learning_rate": 0.00012898747254544825, "loss": 0.846, "step": 2205 }, { "epoch": 0.4243939976914198, "grad_norm": 0.1934354417456004, "learning_rate": 0.00012892783369925105, "loss": 0.7971, "step": 2206 }, { "epoch": 0.42458637937668336, "grad_norm": 0.18429783928807805, "learning_rate": 0.00012886818362227282, "loss": 0.7999, "step": 2207 }, { "epoch": 0.4247787610619469, "grad_norm": 0.1896839952366966, "learning_rate": 0.00012880852233767174, "loss": 0.8254, "step": 2208 }, { "epoch": 0.4249711427472105, "grad_norm": 0.18877811203912276, "learning_rate": 0.0001287488498686104, "loss": 0.8379, "step": 2209 }, { "epoch": 0.425163524432474, "grad_norm": 0.20213648194214265, "learning_rate": 0.0001286891662382556, "loss": 0.8373, "step": 2210 }, { "epoch": 0.4253559061177376, "grad_norm": 0.19695980597092924, "learning_rate": 0.00012862947146977875, "loss": 0.798, "step": 2211 }, { "epoch": 0.42554828780300114, "grad_norm": 0.19312913256145608, "learning_rate": 0.0001285697655863553, "loss": 0.8182, "step": 2212 }, { "epoch": 0.4257406694882647, "grad_norm": 0.18379735869541616, "learning_rate": 0.00012851004861116517, "loss": 0.8543, "step": 2213 }, { "epoch": 0.42593305117352825, "grad_norm": 0.1881708271834793, "learning_rate": 0.00012845032056739257, "loss": 0.8177, "step": 2214 }, { "epoch": 0.42612543285879184, "grad_norm": 0.18069616397000673, "learning_rate": 0.00012839058147822593, "loss": 0.8121, "step": 2215 }, { "epoch": 0.42631781454405543, "grad_norm": 0.18145154045044248, "learning_rate": 0.00012833083136685802, "loss": 0.8174, "step": 2216 }, { "epoch": 0.42651019622931896, "grad_norm": 0.16912011885315512, "learning_rate": 0.00012827107025648594, "loss": 0.8101, "step": 2217 }, { "epoch": 0.42670257791458255, "grad_norm": 0.18961968405421042, "learning_rate": 0.000128211298170311, "loss": 0.8209, "step": 2218 }, { "epoch": 0.4268949595998461, "grad_norm": 0.19793893803284224, "learning_rate": 0.00012815151513153873, "loss": 0.8321, "step": 2219 }, { "epoch": 0.42708734128510967, "grad_norm": 0.18451060727793472, "learning_rate": 0.00012809172116337903, "loss": 0.8432, "step": 2220 }, { "epoch": 0.4272797229703732, "grad_norm": 0.19015621925406112, "learning_rate": 0.00012803191628904593, "loss": 0.8292, "step": 2221 }, { "epoch": 0.4274721046556368, "grad_norm": 0.18240944988292312, "learning_rate": 0.0001279721005317578, "loss": 0.7793, "step": 2222 }, { "epoch": 0.4276644863409004, "grad_norm": 0.18420944014805127, "learning_rate": 0.00012791227391473705, "loss": 0.8231, "step": 2223 }, { "epoch": 0.4278568680261639, "grad_norm": 0.1963847798952816, "learning_rate": 0.0001278524364612106, "loss": 0.7853, "step": 2224 }, { "epoch": 0.4280492497114275, "grad_norm": 0.2079258182857159, "learning_rate": 0.0001277925881944093, "loss": 0.8275, "step": 2225 }, { "epoch": 0.428241631396691, "grad_norm": 0.1976088373278361, "learning_rate": 0.00012773272913756833, "loss": 0.8531, "step": 2226 }, { "epoch": 0.4284340130819546, "grad_norm": 0.1969214733920069, "learning_rate": 0.00012767285931392704, "loss": 0.8848, "step": 2227 }, { "epoch": 0.42862639476721814, "grad_norm": 0.1965477354530095, "learning_rate": 0.000127612978746729, "loss": 0.8116, "step": 2228 }, { "epoch": 0.42881877645248173, "grad_norm": 0.19184410258397708, "learning_rate": 0.0001275530874592218, "loss": 0.8524, "step": 2229 }, { "epoch": 0.42901115813774526, "grad_norm": 0.18613482631950187, "learning_rate": 0.0001274931854746574, "loss": 0.8164, "step": 2230 }, { "epoch": 0.42920353982300885, "grad_norm": 0.2069048037434274, "learning_rate": 0.0001274332728162918, "loss": 0.776, "step": 2231 }, { "epoch": 0.42939592150827244, "grad_norm": 0.20106751475149115, "learning_rate": 0.0001273733495073851, "loss": 0.8214, "step": 2232 }, { "epoch": 0.42958830319353597, "grad_norm": 0.11015623717420696, "learning_rate": 0.00012731341557120168, "loss": 0.6453, "step": 2233 }, { "epoch": 0.42978068487879956, "grad_norm": 0.11853969826991993, "learning_rate": 0.00012725347103100987, "loss": 0.6764, "step": 2234 }, { "epoch": 0.4299730665640631, "grad_norm": 0.17113519680190345, "learning_rate": 0.0001271935159100823, "loss": 0.8316, "step": 2235 }, { "epoch": 0.4301654482493267, "grad_norm": 0.1804233693727038, "learning_rate": 0.00012713355023169547, "loss": 0.7208, "step": 2236 }, { "epoch": 0.4303578299345902, "grad_norm": 0.2011619217110054, "learning_rate": 0.00012707357401913022, "loss": 0.8416, "step": 2237 }, { "epoch": 0.4305502116198538, "grad_norm": 0.18038546828869725, "learning_rate": 0.0001270135872956714, "loss": 0.8157, "step": 2238 }, { "epoch": 0.4307425933051173, "grad_norm": 0.18586242021268679, "learning_rate": 0.00012695359008460785, "loss": 0.8277, "step": 2239 }, { "epoch": 0.4309349749903809, "grad_norm": 0.1947434885364023, "learning_rate": 0.00012689358240923262, "loss": 0.7336, "step": 2240 }, { "epoch": 0.4311273566756445, "grad_norm": 0.1905499983207008, "learning_rate": 0.00012683356429284273, "loss": 0.7727, "step": 2241 }, { "epoch": 0.43131973836090803, "grad_norm": 0.1838574102729289, "learning_rate": 0.00012677353575873924, "loss": 0.8081, "step": 2242 }, { "epoch": 0.4315121200461716, "grad_norm": 0.20076527982566986, "learning_rate": 0.00012671349683022735, "loss": 0.8194, "step": 2243 }, { "epoch": 0.43170450173143515, "grad_norm": 0.11445855409541339, "learning_rate": 0.0001266534475306162, "loss": 0.6494, "step": 2244 }, { "epoch": 0.43189688341669874, "grad_norm": 0.11074492136189407, "learning_rate": 0.00012659338788321905, "loss": 0.6492, "step": 2245 }, { "epoch": 0.43208926510196227, "grad_norm": 0.20888890261417475, "learning_rate": 0.00012653331791135308, "loss": 0.8071, "step": 2246 }, { "epoch": 0.43228164678722586, "grad_norm": 0.2113752905480993, "learning_rate": 0.0001264732376383395, "loss": 0.8443, "step": 2247 }, { "epoch": 0.4324740284724894, "grad_norm": 0.20148872099697024, "learning_rate": 0.0001264131470875036, "loss": 0.7973, "step": 2248 }, { "epoch": 0.432666410157753, "grad_norm": 0.20173087151389812, "learning_rate": 0.00012635304628217452, "loss": 0.8482, "step": 2249 }, { "epoch": 0.43285879184301657, "grad_norm": 0.2079505778954155, "learning_rate": 0.00012629293524568555, "loss": 0.8811, "step": 2250 }, { "epoch": 0.4330511735282801, "grad_norm": 0.10464127241843635, "learning_rate": 0.00012623281400137384, "loss": 0.6013, "step": 2251 }, { "epoch": 0.4332435552135437, "grad_norm": 0.17885371056146634, "learning_rate": 0.0001261726825725805, "loss": 0.7466, "step": 2252 }, { "epoch": 0.4334359368988072, "grad_norm": 0.19918721836363482, "learning_rate": 0.0001261125409826506, "loss": 0.8494, "step": 2253 }, { "epoch": 0.4336283185840708, "grad_norm": 0.17770333847974715, "learning_rate": 0.00012605238925493325, "loss": 0.827, "step": 2254 }, { "epoch": 0.43382070026933434, "grad_norm": 0.18184294175521923, "learning_rate": 0.00012599222741278135, "loss": 0.8655, "step": 2255 }, { "epoch": 0.4340130819545979, "grad_norm": 0.19981320494532892, "learning_rate": 0.00012593205547955185, "loss": 0.7743, "step": 2256 }, { "epoch": 0.4342054636398615, "grad_norm": 0.1782945144286798, "learning_rate": 0.00012587187347860553, "loss": 0.8367, "step": 2257 }, { "epoch": 0.43439784532512504, "grad_norm": 0.18965744655205308, "learning_rate": 0.00012581168143330714, "loss": 0.7509, "step": 2258 }, { "epoch": 0.43459022701038863, "grad_norm": 0.18612452344737834, "learning_rate": 0.0001257514793670253, "loss": 0.7995, "step": 2259 }, { "epoch": 0.43478260869565216, "grad_norm": 0.19721012087895076, "learning_rate": 0.00012569126730313255, "loss": 0.8057, "step": 2260 }, { "epoch": 0.43497499038091575, "grad_norm": 0.20425425135408876, "learning_rate": 0.00012563104526500522, "loss": 0.8379, "step": 2261 }, { "epoch": 0.4351673720661793, "grad_norm": 0.10604273217755139, "learning_rate": 0.0001255708132760236, "loss": 0.6574, "step": 2262 }, { "epoch": 0.43535975375144287, "grad_norm": 0.20256895825742435, "learning_rate": 0.00012551057135957186, "loss": 0.781, "step": 2263 }, { "epoch": 0.4355521354367064, "grad_norm": 0.21718489018759418, "learning_rate": 0.00012545031953903796, "loss": 0.8097, "step": 2264 }, { "epoch": 0.43574451712197, "grad_norm": 0.18105493450783483, "learning_rate": 0.0001253900578378137, "loss": 0.7732, "step": 2265 }, { "epoch": 0.4359368988072336, "grad_norm": 0.18733234322008344, "learning_rate": 0.00012532978627929485, "loss": 0.8511, "step": 2266 }, { "epoch": 0.4361292804924971, "grad_norm": 0.18329327666469952, "learning_rate": 0.00012526950488688082, "loss": 0.8357, "step": 2267 }, { "epoch": 0.4363216621777607, "grad_norm": 0.22909818178629865, "learning_rate": 0.0001252092136839749, "loss": 0.8257, "step": 2268 }, { "epoch": 0.4365140438630242, "grad_norm": 0.20155811605402635, "learning_rate": 0.00012514891269398427, "loss": 0.8064, "step": 2269 }, { "epoch": 0.4367064255482878, "grad_norm": 0.20658703811810342, "learning_rate": 0.00012508860194031986, "loss": 0.8138, "step": 2270 }, { "epoch": 0.43689880723355135, "grad_norm": 0.10891320348773663, "learning_rate": 0.00012502828144639628, "loss": 0.6386, "step": 2271 }, { "epoch": 0.43709118891881493, "grad_norm": 0.17710446956508044, "learning_rate": 0.0001249679512356322, "loss": 0.809, "step": 2272 }, { "epoch": 0.43728357060407846, "grad_norm": 0.22664206814036789, "learning_rate": 0.0001249076113314497, "loss": 0.8274, "step": 2273 }, { "epoch": 0.43747595228934205, "grad_norm": 0.18688560674731686, "learning_rate": 0.0001248472617572749, "loss": 0.7348, "step": 2274 }, { "epoch": 0.43766833397460564, "grad_norm": 0.1950857735280968, "learning_rate": 0.00012478690253653755, "loss": 0.8509, "step": 2275 }, { "epoch": 0.43786071565986917, "grad_norm": 0.20383291737894452, "learning_rate": 0.00012472653369267122, "loss": 0.8103, "step": 2276 }, { "epoch": 0.43805309734513276, "grad_norm": 0.21574130355298443, "learning_rate": 0.00012466615524911315, "loss": 0.7352, "step": 2277 }, { "epoch": 0.4382454790303963, "grad_norm": 0.1933818860385151, "learning_rate": 0.0001246057672293043, "loss": 0.8068, "step": 2278 }, { "epoch": 0.4384378607156599, "grad_norm": 0.20117612339117896, "learning_rate": 0.00012454536965668948, "loss": 0.8249, "step": 2279 }, { "epoch": 0.4386302424009234, "grad_norm": 0.18933303219287842, "learning_rate": 0.000124484962554717, "loss": 0.8479, "step": 2280 }, { "epoch": 0.438822624086187, "grad_norm": 0.2121571768247874, "learning_rate": 0.000124424545946839, "loss": 0.7737, "step": 2281 }, { "epoch": 0.4390150057714506, "grad_norm": 0.1866569406835597, "learning_rate": 0.00012436411985651132, "loss": 0.8535, "step": 2282 }, { "epoch": 0.4392073874567141, "grad_norm": 0.19319091523964685, "learning_rate": 0.0001243036843071934, "loss": 0.7893, "step": 2283 }, { "epoch": 0.4393997691419777, "grad_norm": 0.20140350267635698, "learning_rate": 0.0001242432393223485, "loss": 0.8364, "step": 2284 }, { "epoch": 0.43959215082724123, "grad_norm": 0.19973151514089418, "learning_rate": 0.0001241827849254433, "loss": 0.7993, "step": 2285 }, { "epoch": 0.4397845325125048, "grad_norm": 0.2023062444137733, "learning_rate": 0.0001241223211399484, "loss": 0.7991, "step": 2286 }, { "epoch": 0.43997691419776835, "grad_norm": 0.19043347010350378, "learning_rate": 0.00012406184798933786, "loss": 0.8165, "step": 2287 }, { "epoch": 0.44016929588303194, "grad_norm": 0.16529173214738727, "learning_rate": 0.00012400136549708946, "loss": 0.8129, "step": 2288 }, { "epoch": 0.4403616775682955, "grad_norm": 0.16719512885898308, "learning_rate": 0.00012394087368668459, "loss": 0.8222, "step": 2289 }, { "epoch": 0.44055405925355906, "grad_norm": 0.1928116599656064, "learning_rate": 0.00012388037258160824, "loss": 0.779, "step": 2290 }, { "epoch": 0.44074644093882265, "grad_norm": 0.1857632331913081, "learning_rate": 0.000123819862205349, "loss": 0.7809, "step": 2291 }, { "epoch": 0.4409388226240862, "grad_norm": 0.18460264890299785, "learning_rate": 0.00012375934258139915, "loss": 0.8102, "step": 2292 }, { "epoch": 0.44113120430934977, "grad_norm": 0.17885123746899773, "learning_rate": 0.00012369881373325447, "loss": 0.8396, "step": 2293 }, { "epoch": 0.4413235859946133, "grad_norm": 0.20386567339755987, "learning_rate": 0.0001236382756844143, "loss": 0.8789, "step": 2294 }, { "epoch": 0.4415159676798769, "grad_norm": 0.12037105033986455, "learning_rate": 0.00012357772845838157, "loss": 0.6688, "step": 2295 }, { "epoch": 0.4417083493651404, "grad_norm": 0.17716359590266412, "learning_rate": 0.00012351717207866291, "loss": 0.7744, "step": 2296 }, { "epoch": 0.441900731050404, "grad_norm": 0.183551774979298, "learning_rate": 0.0001234566065687683, "loss": 0.7807, "step": 2297 }, { "epoch": 0.44209311273566754, "grad_norm": 0.19216416898549957, "learning_rate": 0.0001233960319522114, "loss": 0.8353, "step": 2298 }, { "epoch": 0.4422854944209311, "grad_norm": 0.19862687799857184, "learning_rate": 0.00012333544825250938, "loss": 0.8526, "step": 2299 }, { "epoch": 0.4424778761061947, "grad_norm": 0.19367881650056049, "learning_rate": 0.00012327485549318284, "loss": 0.7904, "step": 2300 }, { "epoch": 0.44267025779145824, "grad_norm": 0.1950877759562291, "learning_rate": 0.00012321425369775602, "loss": 0.8212, "step": 2301 }, { "epoch": 0.44286263947672183, "grad_norm": 0.19431097211542578, "learning_rate": 0.00012315364288975663, "loss": 0.7958, "step": 2302 }, { "epoch": 0.44305502116198536, "grad_norm": 0.18697417345995568, "learning_rate": 0.00012309302309271587, "loss": 0.8001, "step": 2303 }, { "epoch": 0.44324740284724895, "grad_norm": 0.20513921612616984, "learning_rate": 0.00012303239433016842, "loss": 0.8222, "step": 2304 }, { "epoch": 0.4434397845325125, "grad_norm": 0.19760672865666631, "learning_rate": 0.00012297175662565247, "loss": 0.8273, "step": 2305 }, { "epoch": 0.44363216621777607, "grad_norm": 0.12676507651991215, "learning_rate": 0.0001229111100027097, "loss": 0.6525, "step": 2306 }, { "epoch": 0.44382454790303966, "grad_norm": 0.21777691823306158, "learning_rate": 0.00012285045448488511, "loss": 0.857, "step": 2307 }, { "epoch": 0.4440169295883032, "grad_norm": 0.1045705317743851, "learning_rate": 0.00012278979009572736, "loss": 0.6203, "step": 2308 }, { "epoch": 0.4442093112735668, "grad_norm": 0.19563568597243308, "learning_rate": 0.0001227291168587884, "loss": 0.7506, "step": 2309 }, { "epoch": 0.4444016929588303, "grad_norm": 0.19579658459998403, "learning_rate": 0.0001226684347976237, "loss": 0.8083, "step": 2310 }, { "epoch": 0.4445940746440939, "grad_norm": 0.1944008159776391, "learning_rate": 0.0001226077439357921, "loss": 0.7965, "step": 2311 }, { "epoch": 0.4447864563293574, "grad_norm": 0.17501174681772827, "learning_rate": 0.00012254704429685593, "loss": 0.8212, "step": 2312 }, { "epoch": 0.444978838014621, "grad_norm": 0.19697978060674165, "learning_rate": 0.00012248633590438083, "loss": 0.8304, "step": 2313 }, { "epoch": 0.44517121969988455, "grad_norm": 0.17717771739883412, "learning_rate": 0.00012242561878193588, "loss": 0.7906, "step": 2314 }, { "epoch": 0.44536360138514813, "grad_norm": 0.20328403376016121, "learning_rate": 0.00012236489295309362, "loss": 0.8464, "step": 2315 }, { "epoch": 0.4455559830704117, "grad_norm": 0.1230477478094978, "learning_rate": 0.00012230415844142984, "loss": 0.6517, "step": 2316 }, { "epoch": 0.44574836475567525, "grad_norm": 0.20159233396296553, "learning_rate": 0.0001222434152705238, "loss": 0.8319, "step": 2317 }, { "epoch": 0.44594074644093884, "grad_norm": 0.18334875922484, "learning_rate": 0.0001221826634639581, "loss": 0.8086, "step": 2318 }, { "epoch": 0.44613312812620237, "grad_norm": 0.2046748252496384, "learning_rate": 0.0001221219030453187, "loss": 0.8346, "step": 2319 }, { "epoch": 0.44632550981146596, "grad_norm": 0.21549721400921135, "learning_rate": 0.00012206113403819479, "loss": 0.8238, "step": 2320 }, { "epoch": 0.4465178914967295, "grad_norm": 0.16811411480777716, "learning_rate": 0.00012200035646617911, "loss": 0.798, "step": 2321 }, { "epoch": 0.4467102731819931, "grad_norm": 0.21173168847069232, "learning_rate": 0.00012193957035286755, "loss": 0.7728, "step": 2322 }, { "epoch": 0.4469026548672566, "grad_norm": 0.19536812222331604, "learning_rate": 0.00012187877572185937, "loss": 0.806, "step": 2323 }, { "epoch": 0.4470950365525202, "grad_norm": 0.2118652238813527, "learning_rate": 0.00012181797259675714, "loss": 0.8749, "step": 2324 }, { "epoch": 0.4472874182377838, "grad_norm": 0.1924270254860103, "learning_rate": 0.00012175716100116676, "loss": 0.8447, "step": 2325 }, { "epoch": 0.4474797999230473, "grad_norm": 0.1932857544278746, "learning_rate": 0.00012169634095869735, "loss": 0.7995, "step": 2326 }, { "epoch": 0.4476721816083109, "grad_norm": 0.20190595098470593, "learning_rate": 0.00012163551249296133, "loss": 0.8692, "step": 2327 }, { "epoch": 0.44786456329357444, "grad_norm": 0.21466017226813947, "learning_rate": 0.00012157467562757443, "loss": 0.7995, "step": 2328 }, { "epoch": 0.448056944978838, "grad_norm": 0.18473307244259843, "learning_rate": 0.00012151383038615563, "loss": 0.7925, "step": 2329 }, { "epoch": 0.44824932666410156, "grad_norm": 0.1906771332225023, "learning_rate": 0.0001214529767923271, "loss": 0.7871, "step": 2330 }, { "epoch": 0.44844170834936514, "grad_norm": 0.17922837787137919, "learning_rate": 0.00012139211486971436, "loss": 0.8132, "step": 2331 }, { "epoch": 0.44863409003462873, "grad_norm": 0.1855200245071817, "learning_rate": 0.0001213312446419461, "loss": 0.8151, "step": 2332 }, { "epoch": 0.44882647171989226, "grad_norm": 0.1777711630830202, "learning_rate": 0.00012127036613265417, "loss": 0.8325, "step": 2333 }, { "epoch": 0.44901885340515585, "grad_norm": 0.20280545531117733, "learning_rate": 0.00012120947936547376, "loss": 0.8026, "step": 2334 }, { "epoch": 0.4492112350904194, "grad_norm": 0.12041788116668357, "learning_rate": 0.00012114858436404321, "loss": 0.6288, "step": 2335 }, { "epoch": 0.44940361677568297, "grad_norm": 0.1119311274365922, "learning_rate": 0.00012108768115200404, "loss": 0.6595, "step": 2336 }, { "epoch": 0.4495959984609465, "grad_norm": 0.20564189200865401, "learning_rate": 0.00012102676975300094, "loss": 0.8526, "step": 2337 }, { "epoch": 0.4497883801462101, "grad_norm": 0.1995570718721733, "learning_rate": 0.0001209658501906819, "loss": 0.8037, "step": 2338 }, { "epoch": 0.4499807618314736, "grad_norm": 0.18759647176430008, "learning_rate": 0.00012090492248869793, "loss": 0.7498, "step": 2339 }, { "epoch": 0.4501731435167372, "grad_norm": 0.18747497433751228, "learning_rate": 0.00012084398667070325, "loss": 0.8448, "step": 2340 }, { "epoch": 0.4503655252020008, "grad_norm": 0.17348804706667917, "learning_rate": 0.00012078304276035526, "loss": 0.7995, "step": 2341 }, { "epoch": 0.4505579068872643, "grad_norm": 0.19109039989519652, "learning_rate": 0.00012072209078131451, "loss": 0.8129, "step": 2342 }, { "epoch": 0.4507502885725279, "grad_norm": 0.18988263593034788, "learning_rate": 0.0001206611307572446, "loss": 0.8176, "step": 2343 }, { "epoch": 0.45094267025779144, "grad_norm": 0.19342909579396883, "learning_rate": 0.0001206001627118124, "loss": 0.8292, "step": 2344 }, { "epoch": 0.45113505194305503, "grad_norm": 0.17757555645449272, "learning_rate": 0.00012053918666868776, "loss": 0.78, "step": 2345 }, { "epoch": 0.45132743362831856, "grad_norm": 0.16830197192209928, "learning_rate": 0.00012047820265154361, "loss": 0.7494, "step": 2346 }, { "epoch": 0.45151981531358215, "grad_norm": 0.20148071411985433, "learning_rate": 0.00012041721068405613, "loss": 0.824, "step": 2347 }, { "epoch": 0.4517121969988457, "grad_norm": 0.17907881678410886, "learning_rate": 0.00012035621078990449, "loss": 0.7784, "step": 2348 }, { "epoch": 0.45190457868410927, "grad_norm": 0.18380479224042973, "learning_rate": 0.00012029520299277095, "loss": 0.7924, "step": 2349 }, { "epoch": 0.45209696036937286, "grad_norm": 0.196529410589154, "learning_rate": 0.00012023418731634077, "loss": 0.8652, "step": 2350 }, { "epoch": 0.4522893420546364, "grad_norm": 0.20933079805807817, "learning_rate": 0.00012017316378430243, "loss": 0.8142, "step": 2351 }, { "epoch": 0.4524817237399, "grad_norm": 0.20568490854098082, "learning_rate": 0.00012011213242034733, "loss": 0.7915, "step": 2352 }, { "epoch": 0.4526741054251635, "grad_norm": 0.18545101369684813, "learning_rate": 0.00012005109324816992, "loss": 0.7731, "step": 2353 }, { "epoch": 0.4528664871104271, "grad_norm": 0.210623796742014, "learning_rate": 0.00011999004629146775, "loss": 0.886, "step": 2354 }, { "epoch": 0.45305886879569063, "grad_norm": 0.19345398519067095, "learning_rate": 0.00011992899157394133, "loss": 0.845, "step": 2355 }, { "epoch": 0.4532512504809542, "grad_norm": 0.18816756824447214, "learning_rate": 0.00011986792911929419, "loss": 0.7807, "step": 2356 }, { "epoch": 0.4534436321662178, "grad_norm": 0.18382963782554762, "learning_rate": 0.00011980685895123289, "loss": 0.7777, "step": 2357 }, { "epoch": 0.45363601385148133, "grad_norm": 0.20096358153891095, "learning_rate": 0.00011974578109346702, "loss": 0.7776, "step": 2358 }, { "epoch": 0.4538283955367449, "grad_norm": 0.1064614297521542, "learning_rate": 0.00011968469556970905, "loss": 0.6184, "step": 2359 }, { "epoch": 0.45402077722200845, "grad_norm": 0.2026460134539475, "learning_rate": 0.00011962360240367445, "loss": 0.8768, "step": 2360 }, { "epoch": 0.45421315890727204, "grad_norm": 0.1754333363358491, "learning_rate": 0.00011956250161908178, "loss": 0.8168, "step": 2361 }, { "epoch": 0.4544055405925356, "grad_norm": 0.20733174048377753, "learning_rate": 0.0001195013932396524, "loss": 0.7511, "step": 2362 }, { "epoch": 0.45459792227779916, "grad_norm": 0.21841469529092353, "learning_rate": 0.00011944027728911071, "loss": 0.8305, "step": 2363 }, { "epoch": 0.4547903039630627, "grad_norm": 0.19219962692833206, "learning_rate": 0.00011937915379118405, "loss": 0.8213, "step": 2364 }, { "epoch": 0.4549826856483263, "grad_norm": 0.20116427318029834, "learning_rate": 0.00011931802276960265, "loss": 0.8568, "step": 2365 }, { "epoch": 0.45517506733358987, "grad_norm": 0.1887775841791557, "learning_rate": 0.00011925688424809964, "loss": 0.8404, "step": 2366 }, { "epoch": 0.4553674490188534, "grad_norm": 0.18295412642041403, "learning_rate": 0.00011919573825041115, "loss": 0.8399, "step": 2367 }, { "epoch": 0.455559830704117, "grad_norm": 0.17881812307662712, "learning_rate": 0.00011913458480027613, "loss": 0.8223, "step": 2368 }, { "epoch": 0.4557522123893805, "grad_norm": 0.17680163650584307, "learning_rate": 0.00011907342392143645, "loss": 0.8815, "step": 2369 }, { "epoch": 0.4559445940746441, "grad_norm": 0.18973110930113593, "learning_rate": 0.00011901225563763693, "loss": 0.7792, "step": 2370 }, { "epoch": 0.45613697575990764, "grad_norm": 0.20269927307035437, "learning_rate": 0.00011895107997262515, "loss": 0.7397, "step": 2371 }, { "epoch": 0.4563293574451712, "grad_norm": 0.18920301558935407, "learning_rate": 0.00011888989695015165, "loss": 0.7652, "step": 2372 }, { "epoch": 0.45652173913043476, "grad_norm": 0.195213785396656, "learning_rate": 0.00011882870659396968, "loss": 0.8057, "step": 2373 }, { "epoch": 0.45671412081569834, "grad_norm": 0.19072330231481685, "learning_rate": 0.00011876750892783557, "loss": 0.8013, "step": 2374 }, { "epoch": 0.45690650250096193, "grad_norm": 0.19803848061318283, "learning_rate": 0.00011870630397550831, "loss": 0.8573, "step": 2375 }, { "epoch": 0.45709888418622546, "grad_norm": 0.19001559909965057, "learning_rate": 0.00011864509176074974, "loss": 0.8447, "step": 2376 }, { "epoch": 0.45729126587148905, "grad_norm": 0.19483547029929274, "learning_rate": 0.0001185838723073246, "loss": 0.845, "step": 2377 }, { "epoch": 0.4574836475567526, "grad_norm": 0.18499965311473202, "learning_rate": 0.00011852264563900038, "loss": 0.8603, "step": 2378 }, { "epoch": 0.45767602924201617, "grad_norm": 0.18719465247008577, "learning_rate": 0.00011846141177954733, "loss": 0.8188, "step": 2379 }, { "epoch": 0.4578684109272797, "grad_norm": 0.18410996390614706, "learning_rate": 0.00011840017075273861, "loss": 0.8171, "step": 2380 }, { "epoch": 0.4580607926125433, "grad_norm": 0.18022363522508483, "learning_rate": 0.00011833892258235007, "loss": 0.833, "step": 2381 }, { "epoch": 0.4582531742978069, "grad_norm": 0.19074928004623865, "learning_rate": 0.00011827766729216035, "loss": 0.7885, "step": 2382 }, { "epoch": 0.4584455559830704, "grad_norm": 0.18630087639502893, "learning_rate": 0.00011821640490595084, "loss": 0.8475, "step": 2383 }, { "epoch": 0.458637937668334, "grad_norm": 0.20025357494364981, "learning_rate": 0.00011815513544750579, "loss": 0.7775, "step": 2384 }, { "epoch": 0.4588303193535975, "grad_norm": 0.19338709668505444, "learning_rate": 0.00011809385894061205, "loss": 0.7767, "step": 2385 }, { "epoch": 0.4590227010388611, "grad_norm": 0.20943459638728584, "learning_rate": 0.00011803257540905925, "loss": 0.8678, "step": 2386 }, { "epoch": 0.45921508272412465, "grad_norm": 0.1950984076753448, "learning_rate": 0.00011797128487663981, "loss": 0.8371, "step": 2387 }, { "epoch": 0.45940746440938823, "grad_norm": 0.19494095712039838, "learning_rate": 0.00011790998736714883, "loss": 0.7962, "step": 2388 }, { "epoch": 0.45959984609465176, "grad_norm": 0.18625371617593822, "learning_rate": 0.00011784868290438404, "loss": 0.7738, "step": 2389 }, { "epoch": 0.45979222777991535, "grad_norm": 0.2132689931641498, "learning_rate": 0.00011778737151214605, "loss": 0.7993, "step": 2390 }, { "epoch": 0.45998460946517894, "grad_norm": 0.1897418327628487, "learning_rate": 0.00011772605321423799, "loss": 0.7629, "step": 2391 }, { "epoch": 0.46017699115044247, "grad_norm": 0.21369705586940868, "learning_rate": 0.00011766472803446575, "loss": 0.7737, "step": 2392 }, { "epoch": 0.46036937283570606, "grad_norm": 0.18644773335436354, "learning_rate": 0.00011760339599663787, "loss": 0.8164, "step": 2393 }, { "epoch": 0.4605617545209696, "grad_norm": 0.1883275100545801, "learning_rate": 0.00011754205712456557, "loss": 0.7967, "step": 2394 }, { "epoch": 0.4607541362062332, "grad_norm": 0.20176244574474242, "learning_rate": 0.00011748071144206266, "loss": 0.8219, "step": 2395 }, { "epoch": 0.4609465178914967, "grad_norm": 0.22729975516391243, "learning_rate": 0.00011741935897294571, "loss": 0.8004, "step": 2396 }, { "epoch": 0.4611388995767603, "grad_norm": 0.09228160177267355, "learning_rate": 0.00011735799974103388, "loss": 0.5848, "step": 2397 }, { "epoch": 0.46133128126202383, "grad_norm": 0.2669672026932527, "learning_rate": 0.00011729663377014888, "loss": 0.8217, "step": 2398 }, { "epoch": 0.4615236629472874, "grad_norm": 0.1882102197770994, "learning_rate": 0.00011723526108411509, "loss": 0.7361, "step": 2399 }, { "epoch": 0.461716044632551, "grad_norm": 0.18526382262040575, "learning_rate": 0.00011717388170675954, "loss": 0.7957, "step": 2400 }, { "epoch": 0.46190842631781454, "grad_norm": 0.19664301824100874, "learning_rate": 0.00011711249566191178, "loss": 0.8059, "step": 2401 }, { "epoch": 0.4621008080030781, "grad_norm": 0.18735306730511928, "learning_rate": 0.00011705110297340397, "loss": 0.8137, "step": 2402 }, { "epoch": 0.46229318968834165, "grad_norm": 0.1928018174265508, "learning_rate": 0.00011698970366507096, "loss": 0.7741, "step": 2403 }, { "epoch": 0.46248557137360524, "grad_norm": 0.17903351790617786, "learning_rate": 0.00011692829776074998, "loss": 0.8402, "step": 2404 }, { "epoch": 0.4626779530588688, "grad_norm": 0.18141979652865786, "learning_rate": 0.00011686688528428098, "loss": 0.7913, "step": 2405 }, { "epoch": 0.46287033474413236, "grad_norm": 0.1814809787983582, "learning_rate": 0.00011680546625950635, "loss": 0.8263, "step": 2406 }, { "epoch": 0.46306271642939595, "grad_norm": 0.19584776459005399, "learning_rate": 0.0001167440407102711, "loss": 0.8722, "step": 2407 }, { "epoch": 0.4632550981146595, "grad_norm": 0.18752621151193283, "learning_rate": 0.00011668260866042271, "loss": 0.7872, "step": 2408 }, { "epoch": 0.46344747979992307, "grad_norm": 0.18158786925070386, "learning_rate": 0.00011662117013381126, "loss": 0.714, "step": 2409 }, { "epoch": 0.4636398614851866, "grad_norm": 0.1872794535068407, "learning_rate": 0.00011655972515428928, "loss": 0.825, "step": 2410 }, { "epoch": 0.4638322431704502, "grad_norm": 0.20238636080178274, "learning_rate": 0.0001164982737457118, "loss": 0.8479, "step": 2411 }, { "epoch": 0.4640246248557137, "grad_norm": 0.18598056710350494, "learning_rate": 0.00011643681593193643, "loss": 0.7827, "step": 2412 }, { "epoch": 0.4642170065409773, "grad_norm": 0.1928291809408523, "learning_rate": 0.00011637535173682316, "loss": 0.782, "step": 2413 }, { "epoch": 0.46440938822624084, "grad_norm": 0.17874041106179522, "learning_rate": 0.00011631388118423456, "loss": 0.8304, "step": 2414 }, { "epoch": 0.4646017699115044, "grad_norm": 0.11103723809664154, "learning_rate": 0.00011625240429803552, "loss": 0.6665, "step": 2415 }, { "epoch": 0.464794151596768, "grad_norm": 0.10100991405903938, "learning_rate": 0.0001161909211020936, "loss": 0.6263, "step": 2416 }, { "epoch": 0.46498653328203154, "grad_norm": 0.19622396204826648, "learning_rate": 0.00011612943162027863, "loss": 0.8371, "step": 2417 }, { "epoch": 0.46517891496729513, "grad_norm": 0.20655666257838115, "learning_rate": 0.00011606793587646294, "loss": 0.7716, "step": 2418 }, { "epoch": 0.46537129665255866, "grad_norm": 0.2052836193402877, "learning_rate": 0.0001160064338945213, "loss": 0.8357, "step": 2419 }, { "epoch": 0.46556367833782225, "grad_norm": 0.24044659296797732, "learning_rate": 0.00011594492569833094, "loss": 0.8398, "step": 2420 }, { "epoch": 0.4657560600230858, "grad_norm": 0.0952589005854049, "learning_rate": 0.00011588341131177136, "loss": 0.6184, "step": 2421 }, { "epoch": 0.46594844170834937, "grad_norm": 0.20646895557645897, "learning_rate": 0.00011582189075872468, "loss": 0.8527, "step": 2422 }, { "epoch": 0.4661408233936129, "grad_norm": 0.20610574614838625, "learning_rate": 0.00011576036406307522, "loss": 0.8048, "step": 2423 }, { "epoch": 0.4663332050788765, "grad_norm": 0.19660380884898246, "learning_rate": 0.00011569883124870979, "loss": 0.8203, "step": 2424 }, { "epoch": 0.4665255867641401, "grad_norm": 0.17742981043204706, "learning_rate": 0.00011563729233951756, "loss": 0.8514, "step": 2425 }, { "epoch": 0.4667179684494036, "grad_norm": 0.19243183877195502, "learning_rate": 0.00011557574735939004, "loss": 0.8152, "step": 2426 }, { "epoch": 0.4669103501346672, "grad_norm": 0.18734614188717996, "learning_rate": 0.00011551419633222108, "loss": 0.783, "step": 2427 }, { "epoch": 0.4671027318199307, "grad_norm": 0.19133487832798346, "learning_rate": 0.00011545263928190693, "loss": 0.8165, "step": 2428 }, { "epoch": 0.4672951135051943, "grad_norm": 0.1846941956853475, "learning_rate": 0.00011539107623234617, "loss": 0.8271, "step": 2429 }, { "epoch": 0.46748749519045785, "grad_norm": 0.19608002946414316, "learning_rate": 0.0001153295072074397, "loss": 0.8343, "step": 2430 }, { "epoch": 0.46767987687572143, "grad_norm": 0.19467746941967007, "learning_rate": 0.00011526793223109071, "loss": 0.8174, "step": 2431 }, { "epoch": 0.467872258560985, "grad_norm": 0.20710382138547437, "learning_rate": 0.00011520635132720474, "loss": 0.7883, "step": 2432 }, { "epoch": 0.46806464024624855, "grad_norm": 0.19386078897556044, "learning_rate": 0.0001151447645196896, "loss": 0.8324, "step": 2433 }, { "epoch": 0.46825702193151214, "grad_norm": 0.18538965257191453, "learning_rate": 0.00011508317183245545, "loss": 0.7787, "step": 2434 }, { "epoch": 0.4684494036167757, "grad_norm": 0.20293992729331076, "learning_rate": 0.00011502157328941465, "loss": 0.7901, "step": 2435 }, { "epoch": 0.46864178530203926, "grad_norm": 0.19865596759143955, "learning_rate": 0.00011495996891448189, "loss": 0.8289, "step": 2436 }, { "epoch": 0.4688341669873028, "grad_norm": 0.19594293992373782, "learning_rate": 0.00011489835873157413, "loss": 0.8671, "step": 2437 }, { "epoch": 0.4690265486725664, "grad_norm": 0.18362847110462363, "learning_rate": 0.00011483674276461052, "loss": 0.811, "step": 2438 }, { "epoch": 0.4692189303578299, "grad_norm": 0.19404360119304773, "learning_rate": 0.00011477512103751253, "loss": 0.8373, "step": 2439 }, { "epoch": 0.4694113120430935, "grad_norm": 0.19462991775358193, "learning_rate": 0.00011471349357420384, "loss": 0.7987, "step": 2440 }, { "epoch": 0.4696036937283571, "grad_norm": 0.21245083593317074, "learning_rate": 0.00011465186039861033, "loss": 0.81, "step": 2441 }, { "epoch": 0.4697960754136206, "grad_norm": 0.2119224372819171, "learning_rate": 0.00011459022153466015, "loss": 0.8497, "step": 2442 }, { "epoch": 0.4699884570988842, "grad_norm": 0.21022792903897713, "learning_rate": 0.00011452857700628361, "loss": 0.8256, "step": 2443 }, { "epoch": 0.47018083878414774, "grad_norm": 0.19148174907245769, "learning_rate": 0.00011446692683741325, "loss": 0.7883, "step": 2444 }, { "epoch": 0.4703732204694113, "grad_norm": 0.20111029077747822, "learning_rate": 0.00011440527105198378, "loss": 0.8322, "step": 2445 }, { "epoch": 0.47056560215467486, "grad_norm": 0.1879507051946582, "learning_rate": 0.0001143436096739321, "loss": 0.7782, "step": 2446 }, { "epoch": 0.47075798383993844, "grad_norm": 0.18931722896605196, "learning_rate": 0.00011428194272719728, "loss": 0.7998, "step": 2447 }, { "epoch": 0.470950365525202, "grad_norm": 0.18450803344606156, "learning_rate": 0.00011422027023572051, "loss": 0.8375, "step": 2448 }, { "epoch": 0.47114274721046556, "grad_norm": 0.1809306463755825, "learning_rate": 0.00011415859222344525, "loss": 0.8111, "step": 2449 }, { "epoch": 0.47133512889572915, "grad_norm": 0.19486902442989934, "learning_rate": 0.000114096908714317, "loss": 0.8153, "step": 2450 }, { "epoch": 0.4715275105809927, "grad_norm": 0.19063031563526284, "learning_rate": 0.00011403521973228341, "loss": 0.7946, "step": 2451 }, { "epoch": 0.47171989226625627, "grad_norm": 0.18693258961900316, "learning_rate": 0.00011397352530129429, "loss": 0.8452, "step": 2452 }, { "epoch": 0.4719122739515198, "grad_norm": 0.21727569425798315, "learning_rate": 0.0001139118254453015, "loss": 0.8552, "step": 2453 }, { "epoch": 0.4721046556367834, "grad_norm": 0.18730916811552528, "learning_rate": 0.00011385012018825907, "loss": 0.7885, "step": 2454 }, { "epoch": 0.4722970373220469, "grad_norm": 0.10988406459273335, "learning_rate": 0.00011378840955412313, "loss": 0.632, "step": 2455 }, { "epoch": 0.4724894190073105, "grad_norm": 0.1880879806343264, "learning_rate": 0.00011372669356685183, "loss": 0.862, "step": 2456 }, { "epoch": 0.4726818006925741, "grad_norm": 0.19318427444511604, "learning_rate": 0.00011366497225040548, "loss": 0.7878, "step": 2457 }, { "epoch": 0.4728741823778376, "grad_norm": 0.23022152823299816, "learning_rate": 0.00011360324562874643, "loss": 0.8128, "step": 2458 }, { "epoch": 0.4730665640631012, "grad_norm": 0.1925342936191478, "learning_rate": 0.00011354151372583901, "loss": 0.834, "step": 2459 }, { "epoch": 0.47325894574836475, "grad_norm": 0.18104989722122877, "learning_rate": 0.00011347977656564975, "loss": 0.8019, "step": 2460 }, { "epoch": 0.47345132743362833, "grad_norm": 0.18987807701671225, "learning_rate": 0.00011341803417214705, "loss": 0.7831, "step": 2461 }, { "epoch": 0.47364370911889186, "grad_norm": 0.19971412234479852, "learning_rate": 0.00011335628656930152, "loss": 0.8055, "step": 2462 }, { "epoch": 0.47383609080415545, "grad_norm": 0.21455867273258006, "learning_rate": 0.0001132945337810857, "loss": 0.784, "step": 2463 }, { "epoch": 0.474028472489419, "grad_norm": 0.18335194794099233, "learning_rate": 0.00011323277583147409, "loss": 0.7804, "step": 2464 }, { "epoch": 0.47422085417468257, "grad_norm": 0.1908002198485425, "learning_rate": 0.00011317101274444328, "loss": 0.8471, "step": 2465 }, { "epoch": 0.47441323585994616, "grad_norm": 0.1858949612132562, "learning_rate": 0.00011310924454397187, "loss": 0.7971, "step": 2466 }, { "epoch": 0.4746056175452097, "grad_norm": 0.18240943575796034, "learning_rate": 0.00011304747125404031, "loss": 0.8158, "step": 2467 }, { "epoch": 0.4747979992304733, "grad_norm": 0.19552573684591698, "learning_rate": 0.00011298569289863119, "loss": 0.8513, "step": 2468 }, { "epoch": 0.4749903809157368, "grad_norm": 0.18963706998015908, "learning_rate": 0.00011292390950172899, "loss": 0.7798, "step": 2469 }, { "epoch": 0.4751827626010004, "grad_norm": 0.21180065839329626, "learning_rate": 0.00011286212108732015, "loss": 0.8114, "step": 2470 }, { "epoch": 0.47537514428626393, "grad_norm": 0.18330681408942598, "learning_rate": 0.00011280032767939301, "loss": 0.7718, "step": 2471 }, { "epoch": 0.4755675259715275, "grad_norm": 0.19863196562595928, "learning_rate": 0.00011273852930193798, "loss": 0.7916, "step": 2472 }, { "epoch": 0.47575990765679105, "grad_norm": 0.2022294366037505, "learning_rate": 0.00011267672597894725, "loss": 0.8498, "step": 2473 }, { "epoch": 0.47595228934205464, "grad_norm": 0.2010402092168082, "learning_rate": 0.000112614917734415, "loss": 0.6904, "step": 2474 }, { "epoch": 0.4761446710273182, "grad_norm": 0.18567785383844043, "learning_rate": 0.00011255310459233737, "loss": 0.8055, "step": 2475 }, { "epoch": 0.47633705271258175, "grad_norm": 0.1838034764614967, "learning_rate": 0.00011249128657671233, "loss": 0.8658, "step": 2476 }, { "epoch": 0.47652943439784534, "grad_norm": 0.18281962017598452, "learning_rate": 0.00011242946371153972, "loss": 0.8156, "step": 2477 }, { "epoch": 0.4767218160831089, "grad_norm": 0.20256668299181774, "learning_rate": 0.00011236763602082137, "loss": 0.8482, "step": 2478 }, { "epoch": 0.47691419776837246, "grad_norm": 0.19669185275904685, "learning_rate": 0.00011230580352856088, "loss": 0.8039, "step": 2479 }, { "epoch": 0.477106579453636, "grad_norm": 0.1814468893028477, "learning_rate": 0.00011224396625876375, "loss": 0.8242, "step": 2480 }, { "epoch": 0.4772989611388996, "grad_norm": 0.17323353244046136, "learning_rate": 0.00011218212423543734, "loss": 0.7701, "step": 2481 }, { "epoch": 0.4774913428241631, "grad_norm": 0.18842657971425206, "learning_rate": 0.00011212027748259086, "loss": 0.8658, "step": 2482 }, { "epoch": 0.4776837245094267, "grad_norm": 0.18109133189570473, "learning_rate": 0.00011205842602423537, "loss": 0.7745, "step": 2483 }, { "epoch": 0.4778761061946903, "grad_norm": 0.19542720467945304, "learning_rate": 0.00011199656988438372, "loss": 0.7926, "step": 2484 }, { "epoch": 0.4780684878799538, "grad_norm": 0.19788455047715073, "learning_rate": 0.00011193470908705055, "loss": 0.8715, "step": 2485 }, { "epoch": 0.4782608695652174, "grad_norm": 0.20368101562565175, "learning_rate": 0.00011187284365625242, "loss": 0.8199, "step": 2486 }, { "epoch": 0.47845325125048094, "grad_norm": 0.19788904805720486, "learning_rate": 0.00011181097361600755, "loss": 0.805, "step": 2487 }, { "epoch": 0.4786456329357445, "grad_norm": 0.19238273288934593, "learning_rate": 0.00011174909899033607, "loss": 0.7346, "step": 2488 }, { "epoch": 0.47883801462100806, "grad_norm": 0.20521548115613253, "learning_rate": 0.00011168721980325987, "loss": 0.8513, "step": 2489 }, { "epoch": 0.47903039630627164, "grad_norm": 0.186713057626199, "learning_rate": 0.0001116253360788025, "loss": 0.8186, "step": 2490 }, { "epoch": 0.47922277799153523, "grad_norm": 0.1833689930768899, "learning_rate": 0.00011156344784098942, "loss": 0.8419, "step": 2491 }, { "epoch": 0.47941515967679876, "grad_norm": 0.2078431768992956, "learning_rate": 0.00011150155511384773, "loss": 0.8278, "step": 2492 }, { "epoch": 0.47960754136206235, "grad_norm": 0.18259789019793254, "learning_rate": 0.00011143965792140632, "loss": 0.8328, "step": 2493 }, { "epoch": 0.4797999230473259, "grad_norm": 0.2126109788773165, "learning_rate": 0.00011137775628769584, "loss": 0.8676, "step": 2494 }, { "epoch": 0.47999230473258947, "grad_norm": 0.19290618878644597, "learning_rate": 0.00011131585023674861, "loss": 0.8393, "step": 2495 }, { "epoch": 0.480184686417853, "grad_norm": 0.20438412163686343, "learning_rate": 0.00011125393979259872, "loss": 0.7743, "step": 2496 }, { "epoch": 0.4803770681031166, "grad_norm": 0.1824871440258919, "learning_rate": 0.00011119202497928192, "loss": 0.7855, "step": 2497 }, { "epoch": 0.4805694497883801, "grad_norm": 0.18974169141341993, "learning_rate": 0.00011113010582083567, "loss": 0.8287, "step": 2498 }, { "epoch": 0.4807618314736437, "grad_norm": 0.18365945609573128, "learning_rate": 0.00011106818234129912, "loss": 0.8105, "step": 2499 }, { "epoch": 0.4809542131589073, "grad_norm": 0.2053830224525449, "learning_rate": 0.00011100625456471307, "loss": 0.8176, "step": 2500 }, { "epoch": 0.4811465948441708, "grad_norm": 0.20585986614779525, "learning_rate": 0.00011094432251512005, "loss": 0.8002, "step": 2501 }, { "epoch": 0.4813389765294344, "grad_norm": 0.18651414110106393, "learning_rate": 0.00011088238621656421, "loss": 0.7959, "step": 2502 }, { "epoch": 0.48153135821469795, "grad_norm": 0.19409274372004492, "learning_rate": 0.00011082044569309137, "loss": 0.8395, "step": 2503 }, { "epoch": 0.48172373989996153, "grad_norm": 0.20170296066419632, "learning_rate": 0.00011075850096874893, "loss": 0.8227, "step": 2504 }, { "epoch": 0.48191612158522507, "grad_norm": 0.10764254609250187, "learning_rate": 0.00011069655206758602, "loss": 0.6228, "step": 2505 }, { "epoch": 0.48210850327048865, "grad_norm": 0.18634194846958255, "learning_rate": 0.00011063459901365325, "loss": 0.854, "step": 2506 }, { "epoch": 0.4823008849557522, "grad_norm": 0.2831034747692125, "learning_rate": 0.00011057264183100302, "loss": 0.8234, "step": 2507 }, { "epoch": 0.48249326664101577, "grad_norm": 0.19298043604978063, "learning_rate": 0.0001105106805436892, "loss": 0.8, "step": 2508 }, { "epoch": 0.48268564832627936, "grad_norm": 0.17350239019618588, "learning_rate": 0.00011044871517576729, "loss": 0.7976, "step": 2509 }, { "epoch": 0.4828780300115429, "grad_norm": 0.2040742510625555, "learning_rate": 0.00011038674575129441, "loss": 0.7586, "step": 2510 }, { "epoch": 0.4830704116968065, "grad_norm": 0.21303844578066303, "learning_rate": 0.0001103247722943292, "loss": 0.8257, "step": 2511 }, { "epoch": 0.48326279338207, "grad_norm": 0.11071480217869777, "learning_rate": 0.00011026279482893187, "loss": 0.6536, "step": 2512 }, { "epoch": 0.4834551750673336, "grad_norm": 0.18038525289373303, "learning_rate": 0.00011020081337916424, "loss": 0.8263, "step": 2513 }, { "epoch": 0.48364755675259713, "grad_norm": 0.09899813465397421, "learning_rate": 0.00011013882796908963, "loss": 0.6097, "step": 2514 }, { "epoch": 0.4838399384378607, "grad_norm": 0.20794640806498635, "learning_rate": 0.00011007683862277292, "loss": 0.8011, "step": 2515 }, { "epoch": 0.4840323201231243, "grad_norm": 0.20090748902161262, "learning_rate": 0.00011001484536428051, "loss": 0.7925, "step": 2516 }, { "epoch": 0.48422470180838784, "grad_norm": 0.18040633531342884, "learning_rate": 0.00010995284821768028, "loss": 0.8139, "step": 2517 }, { "epoch": 0.4844170834936514, "grad_norm": 0.18688767402489245, "learning_rate": 0.0001098908472070417, "loss": 0.8194, "step": 2518 }, { "epoch": 0.48460946517891496, "grad_norm": 0.1959945029270076, "learning_rate": 0.00010982884235643567, "loss": 0.836, "step": 2519 }, { "epoch": 0.48480184686417854, "grad_norm": 0.20986487881955437, "learning_rate": 0.00010976683368993464, "loss": 0.8402, "step": 2520 }, { "epoch": 0.4849942285494421, "grad_norm": 0.1898504572788511, "learning_rate": 0.00010970482123161248, "loss": 0.8938, "step": 2521 }, { "epoch": 0.48518661023470566, "grad_norm": 0.1822475091175196, "learning_rate": 0.00010964280500554458, "loss": 0.8195, "step": 2522 }, { "epoch": 0.4853789919199692, "grad_norm": 0.19947735242014505, "learning_rate": 0.00010958078503580775, "loss": 0.8493, "step": 2523 }, { "epoch": 0.4855713736052328, "grad_norm": 0.20514636713993623, "learning_rate": 0.00010951876134648032, "loss": 0.8301, "step": 2524 }, { "epoch": 0.48576375529049637, "grad_norm": 0.20393909701653426, "learning_rate": 0.00010945673396164198, "loss": 0.8082, "step": 2525 }, { "epoch": 0.4859561369757599, "grad_norm": 0.18054474104340923, "learning_rate": 0.00010939470290537388, "loss": 0.819, "step": 2526 }, { "epoch": 0.4861485186610235, "grad_norm": 0.19977675326457045, "learning_rate": 0.00010933266820175867, "loss": 0.7709, "step": 2527 }, { "epoch": 0.486340900346287, "grad_norm": 0.10317500869656634, "learning_rate": 0.00010927062987488035, "loss": 0.616, "step": 2528 }, { "epoch": 0.4865332820315506, "grad_norm": 0.18057706998377804, "learning_rate": 0.00010920858794882429, "loss": 0.8331, "step": 2529 }, { "epoch": 0.48672566371681414, "grad_norm": 0.1838734453126098, "learning_rate": 0.00010914654244767734, "loss": 0.8462, "step": 2530 }, { "epoch": 0.4869180454020777, "grad_norm": 0.10459992969664375, "learning_rate": 0.00010908449339552768, "loss": 0.5966, "step": 2531 }, { "epoch": 0.48711042708734126, "grad_norm": 0.1967210787688675, "learning_rate": 0.00010902244081646489, "loss": 0.7424, "step": 2532 }, { "epoch": 0.48730280877260485, "grad_norm": 0.1883372737835181, "learning_rate": 0.00010896038473457993, "loss": 0.7465, "step": 2533 }, { "epoch": 0.48749519045786843, "grad_norm": 0.1931696736956314, "learning_rate": 0.00010889832517396509, "loss": 0.8121, "step": 2534 }, { "epoch": 0.48768757214313196, "grad_norm": 0.18283211457239143, "learning_rate": 0.00010883626215871408, "loss": 0.8346, "step": 2535 }, { "epoch": 0.48787995382839555, "grad_norm": 0.19215525804164765, "learning_rate": 0.00010877419571292182, "loss": 0.8291, "step": 2536 }, { "epoch": 0.4880723355136591, "grad_norm": 0.19581749270771892, "learning_rate": 0.00010871212586068469, "loss": 0.841, "step": 2537 }, { "epoch": 0.48826471719892267, "grad_norm": 0.20538864070533003, "learning_rate": 0.00010865005262610033, "loss": 0.8219, "step": 2538 }, { "epoch": 0.4884570988841862, "grad_norm": 0.16748632458097507, "learning_rate": 0.00010858797603326769, "loss": 0.7892, "step": 2539 }, { "epoch": 0.4886494805694498, "grad_norm": 0.2046249783256248, "learning_rate": 0.0001085258961062871, "loss": 0.8004, "step": 2540 }, { "epoch": 0.4888418622547134, "grad_norm": 0.1969353519539599, "learning_rate": 0.00010846381286926006, "loss": 0.7612, "step": 2541 }, { "epoch": 0.4890342439399769, "grad_norm": 0.19335735025424175, "learning_rate": 0.00010840172634628946, "loss": 0.8025, "step": 2542 }, { "epoch": 0.4892266256252405, "grad_norm": 0.1956757122383245, "learning_rate": 0.00010833963656147943, "loss": 0.7743, "step": 2543 }, { "epoch": 0.48941900731050403, "grad_norm": 0.10995089476963113, "learning_rate": 0.00010827754353893532, "loss": 0.6198, "step": 2544 }, { "epoch": 0.4896113889957676, "grad_norm": 0.21372798520703284, "learning_rate": 0.00010821544730276378, "loss": 0.8191, "step": 2545 }, { "epoch": 0.48980377068103115, "grad_norm": 0.18351348747775342, "learning_rate": 0.00010815334787707276, "loss": 0.7953, "step": 2546 }, { "epoch": 0.48999615236629473, "grad_norm": 0.1819996191343118, "learning_rate": 0.00010809124528597139, "loss": 0.8102, "step": 2547 }, { "epoch": 0.49018853405155827, "grad_norm": 0.1805864583785493, "learning_rate": 0.00010802913955356997, "loss": 0.8057, "step": 2548 }, { "epoch": 0.49038091573682185, "grad_norm": 0.20425629894557382, "learning_rate": 0.00010796703070398015, "loss": 0.7636, "step": 2549 }, { "epoch": 0.49057329742208544, "grad_norm": 0.19771988862328513, "learning_rate": 0.0001079049187613147, "loss": 0.812, "step": 2550 }, { "epoch": 0.490765679107349, "grad_norm": 0.215770945109068, "learning_rate": 0.0001078428037496876, "loss": 0.8126, "step": 2551 }, { "epoch": 0.49095806079261256, "grad_norm": 0.18254017413610696, "learning_rate": 0.00010778068569321403, "loss": 0.8662, "step": 2552 }, { "epoch": 0.4911504424778761, "grad_norm": 0.21370887466444932, "learning_rate": 0.0001077185646160104, "loss": 0.8148, "step": 2553 }, { "epoch": 0.4913428241631397, "grad_norm": 0.20973082276627347, "learning_rate": 0.00010765644054219421, "loss": 0.8235, "step": 2554 }, { "epoch": 0.4915352058484032, "grad_norm": 0.20223107716066807, "learning_rate": 0.0001075943134958842, "loss": 0.7976, "step": 2555 }, { "epoch": 0.4917275875336668, "grad_norm": 0.199537728033698, "learning_rate": 0.00010753218350120022, "loss": 0.853, "step": 2556 }, { "epoch": 0.49191996921893033, "grad_norm": 0.19025322859736596, "learning_rate": 0.00010747005058226324, "loss": 0.794, "step": 2557 }, { "epoch": 0.4921123509041939, "grad_norm": 0.1848406459003397, "learning_rate": 0.00010740791476319543, "loss": 0.7799, "step": 2558 }, { "epoch": 0.4923047325894575, "grad_norm": 0.21276599366032262, "learning_rate": 0.00010734577606812007, "loss": 0.7912, "step": 2559 }, { "epoch": 0.49249711427472104, "grad_norm": 0.1993418279855315, "learning_rate": 0.00010728363452116148, "loss": 0.7846, "step": 2560 }, { "epoch": 0.4926894959599846, "grad_norm": 0.2004757943908398, "learning_rate": 0.00010722149014644522, "loss": 0.7954, "step": 2561 }, { "epoch": 0.49288187764524816, "grad_norm": 0.19656349529355094, "learning_rate": 0.00010715934296809782, "loss": 0.7934, "step": 2562 }, { "epoch": 0.49307425933051174, "grad_norm": 0.20442221012430214, "learning_rate": 0.000107097193010247, "loss": 0.8161, "step": 2563 }, { "epoch": 0.4932666410157753, "grad_norm": 0.20864627792991297, "learning_rate": 0.00010703504029702148, "loss": 0.8462, "step": 2564 }, { "epoch": 0.49345902270103886, "grad_norm": 0.20398483376269927, "learning_rate": 0.00010697288485255106, "loss": 0.8202, "step": 2565 }, { "epoch": 0.49365140438630245, "grad_norm": 0.11432487043802712, "learning_rate": 0.00010691072670096669, "loss": 0.6389, "step": 2566 }, { "epoch": 0.493843786071566, "grad_norm": 0.18639721828929676, "learning_rate": 0.00010684856586640026, "loss": 0.7828, "step": 2567 }, { "epoch": 0.49403616775682957, "grad_norm": 0.1984736458916416, "learning_rate": 0.00010678640237298476, "loss": 0.8153, "step": 2568 }, { "epoch": 0.4942285494420931, "grad_norm": 0.18064293930687256, "learning_rate": 0.00010672423624485421, "loss": 0.8231, "step": 2569 }, { "epoch": 0.4944209311273567, "grad_norm": 0.19630679880540547, "learning_rate": 0.00010666206750614362, "loss": 0.795, "step": 2570 }, { "epoch": 0.4946133128126202, "grad_norm": 0.1890696793577531, "learning_rate": 0.00010659989618098905, "loss": 0.8327, "step": 2571 }, { "epoch": 0.4948056944978838, "grad_norm": 0.18544487964289647, "learning_rate": 0.00010653772229352749, "loss": 0.8397, "step": 2572 }, { "epoch": 0.49499807618314734, "grad_norm": 0.19518176961359826, "learning_rate": 0.00010647554586789708, "loss": 0.8574, "step": 2573 }, { "epoch": 0.4951904578684109, "grad_norm": 0.18709246464516727, "learning_rate": 0.0001064133669282368, "loss": 0.7918, "step": 2574 }, { "epoch": 0.4953828395536745, "grad_norm": 0.18684307597297978, "learning_rate": 0.00010635118549868668, "loss": 0.819, "step": 2575 }, { "epoch": 0.49557522123893805, "grad_norm": 0.19596399048484084, "learning_rate": 0.00010628900160338764, "loss": 0.7692, "step": 2576 }, { "epoch": 0.49576760292420163, "grad_norm": 0.19602700995009742, "learning_rate": 0.00010622681526648167, "loss": 0.7815, "step": 2577 }, { "epoch": 0.49595998460946517, "grad_norm": 0.19419852933922468, "learning_rate": 0.00010616462651211157, "loss": 0.8021, "step": 2578 }, { "epoch": 0.49615236629472875, "grad_norm": 0.17147427281332114, "learning_rate": 0.00010610243536442124, "loss": 0.7467, "step": 2579 }, { "epoch": 0.4963447479799923, "grad_norm": 0.10726136429569945, "learning_rate": 0.00010604024184755539, "loss": 0.6114, "step": 2580 }, { "epoch": 0.49653712966525587, "grad_norm": 0.19984202672997217, "learning_rate": 0.00010597804598565968, "loss": 0.877, "step": 2581 }, { "epoch": 0.4967295113505194, "grad_norm": 0.2022242361807275, "learning_rate": 0.00010591584780288068, "loss": 0.812, "step": 2582 }, { "epoch": 0.496921893035783, "grad_norm": 0.19361954790089675, "learning_rate": 0.00010585364732336587, "loss": 0.8067, "step": 2583 }, { "epoch": 0.4971142747210466, "grad_norm": 0.18337714313346806, "learning_rate": 0.00010579144457126364, "loss": 0.7726, "step": 2584 }, { "epoch": 0.4973066564063101, "grad_norm": 0.18783952021911446, "learning_rate": 0.00010572923957072321, "loss": 0.7982, "step": 2585 }, { "epoch": 0.4974990380915737, "grad_norm": 0.19549215234545936, "learning_rate": 0.0001056670323458947, "loss": 0.8229, "step": 2586 }, { "epoch": 0.49769141977683723, "grad_norm": 0.10435550984897518, "learning_rate": 0.00010560482292092913, "loss": 0.6381, "step": 2587 }, { "epoch": 0.4978838014621008, "grad_norm": 0.1896648460440543, "learning_rate": 0.00010554261131997834, "loss": 0.8389, "step": 2588 }, { "epoch": 0.49807618314736435, "grad_norm": 0.18282247195506063, "learning_rate": 0.00010548039756719497, "loss": 0.8005, "step": 2589 }, { "epoch": 0.49826856483262794, "grad_norm": 0.19808063084351493, "learning_rate": 0.0001054181816867326, "loss": 0.8364, "step": 2590 }, { "epoch": 0.4984609465178915, "grad_norm": 0.19670865657810924, "learning_rate": 0.0001053559637027455, "loss": 0.8555, "step": 2591 }, { "epoch": 0.49865332820315506, "grad_norm": 0.19931153655381337, "learning_rate": 0.00010529374363938889, "loss": 0.8764, "step": 2592 }, { "epoch": 0.49884570988841864, "grad_norm": 0.20092342088253656, "learning_rate": 0.00010523152152081874, "loss": 0.7363, "step": 2593 }, { "epoch": 0.4990380915736822, "grad_norm": 0.19595144951527138, "learning_rate": 0.00010516929737119181, "loss": 0.8216, "step": 2594 }, { "epoch": 0.49923047325894576, "grad_norm": 0.18578429993179868, "learning_rate": 0.00010510707121466567, "loss": 0.8545, "step": 2595 }, { "epoch": 0.4994228549442093, "grad_norm": 0.11189241424529181, "learning_rate": 0.00010504484307539863, "loss": 0.6188, "step": 2596 }, { "epoch": 0.4996152366294729, "grad_norm": 0.18327581947929292, "learning_rate": 0.00010498261297754983, "loss": 0.8053, "step": 2597 }, { "epoch": 0.4998076183147364, "grad_norm": 0.2119051896985304, "learning_rate": 0.00010492038094527907, "loss": 0.8302, "step": 2598 }, { "epoch": 0.5, "grad_norm": 0.2050863936143546, "learning_rate": 0.00010485814700274706, "loss": 0.7933, "step": 2599 }, { "epoch": 0.5001923816852636, "grad_norm": 0.21558042207519698, "learning_rate": 0.0001047959111741151, "loss": 0.756, "step": 2600 }, { "epoch": 0.5003847633705272, "grad_norm": 0.18615551268538244, "learning_rate": 0.00010473367348354527, "loss": 0.8599, "step": 2601 }, { "epoch": 0.5005771450557907, "grad_norm": 0.19069491142820938, "learning_rate": 0.00010467143395520043, "loss": 0.8018, "step": 2602 }, { "epoch": 0.5007695267410542, "grad_norm": 0.19536998202223754, "learning_rate": 0.0001046091926132441, "loss": 0.7972, "step": 2603 }, { "epoch": 0.5009619084263178, "grad_norm": 0.2081126178609065, "learning_rate": 0.00010454694948184045, "loss": 0.7952, "step": 2604 }, { "epoch": 0.5011542901115814, "grad_norm": 0.17459256540965437, "learning_rate": 0.00010448470458515449, "loss": 0.8179, "step": 2605 }, { "epoch": 0.5013466717968449, "grad_norm": 0.18427943229236576, "learning_rate": 0.00010442245794735179, "loss": 0.8394, "step": 2606 }, { "epoch": 0.5015390534821085, "grad_norm": 0.20576258464601582, "learning_rate": 0.00010436020959259862, "loss": 0.7983, "step": 2607 }, { "epoch": 0.5017314351673721, "grad_norm": 0.19935893979951566, "learning_rate": 0.00010429795954506203, "loss": 0.7847, "step": 2608 }, { "epoch": 0.5019238168526357, "grad_norm": 0.17551852034789714, "learning_rate": 0.0001042357078289095, "loss": 0.8099, "step": 2609 }, { "epoch": 0.5021161985378992, "grad_norm": 0.19624794901527398, "learning_rate": 0.00010417345446830937, "loss": 0.7687, "step": 2610 }, { "epoch": 0.5023085802231627, "grad_norm": 0.19220056724755216, "learning_rate": 0.00010411119948743051, "loss": 0.7725, "step": 2611 }, { "epoch": 0.5025009619084263, "grad_norm": 0.18803518916749956, "learning_rate": 0.00010404894291044246, "loss": 0.7912, "step": 2612 }, { "epoch": 0.5026933435936899, "grad_norm": 0.18118386420512864, "learning_rate": 0.00010398668476151538, "loss": 0.7704, "step": 2613 }, { "epoch": 0.5028857252789535, "grad_norm": 0.18264460790646966, "learning_rate": 0.00010392442506481999, "loss": 0.8753, "step": 2614 }, { "epoch": 0.503078106964217, "grad_norm": 0.19518873311357798, "learning_rate": 0.0001038621638445277, "loss": 0.807, "step": 2615 }, { "epoch": 0.5032704886494805, "grad_norm": 0.19042446902878943, "learning_rate": 0.00010379990112481039, "loss": 0.8458, "step": 2616 }, { "epoch": 0.5034628703347441, "grad_norm": 0.17437891134253403, "learning_rate": 0.00010373763692984062, "loss": 0.805, "step": 2617 }, { "epoch": 0.5036552520200077, "grad_norm": 0.18303545445377375, "learning_rate": 0.00010367537128379154, "loss": 0.8057, "step": 2618 }, { "epoch": 0.5038476337052713, "grad_norm": 0.1950917733232443, "learning_rate": 0.00010361310421083676, "loss": 0.7908, "step": 2619 }, { "epoch": 0.5040400153905348, "grad_norm": 0.20653754003577465, "learning_rate": 0.00010355083573515051, "loss": 0.8042, "step": 2620 }, { "epoch": 0.5042323970757984, "grad_norm": 0.1863296432246938, "learning_rate": 0.00010348856588090763, "loss": 0.7942, "step": 2621 }, { "epoch": 0.504424778761062, "grad_norm": 0.181985311815763, "learning_rate": 0.00010342629467228331, "loss": 0.8436, "step": 2622 }, { "epoch": 0.5046171604463255, "grad_norm": 0.2069171207170514, "learning_rate": 0.00010336402213345344, "loss": 0.7499, "step": 2623 }, { "epoch": 0.504809542131589, "grad_norm": 0.19909211089464973, "learning_rate": 0.00010330174828859433, "loss": 0.8157, "step": 2624 }, { "epoch": 0.5050019238168526, "grad_norm": 0.1914908077343828, "learning_rate": 0.00010323947316188288, "loss": 0.8051, "step": 2625 }, { "epoch": 0.5051943055021162, "grad_norm": 0.1777472341883536, "learning_rate": 0.0001031771967774964, "loss": 0.7842, "step": 2626 }, { "epoch": 0.5053866871873798, "grad_norm": 0.19413020497741285, "learning_rate": 0.00010311491915961271, "loss": 0.8173, "step": 2627 }, { "epoch": 0.5055790688726434, "grad_norm": 0.18517759390264882, "learning_rate": 0.0001030526403324102, "loss": 0.8411, "step": 2628 }, { "epoch": 0.5057714505579068, "grad_norm": 0.1785656375103651, "learning_rate": 0.0001029903603200676, "loss": 0.7996, "step": 2629 }, { "epoch": 0.5059638322431704, "grad_norm": 0.2016573283483471, "learning_rate": 0.00010292807914676412, "loss": 0.7955, "step": 2630 }, { "epoch": 0.506156213928434, "grad_norm": 0.19635734545165545, "learning_rate": 0.00010286579683667952, "loss": 0.8029, "step": 2631 }, { "epoch": 0.5063485956136976, "grad_norm": 0.21459886517798726, "learning_rate": 0.00010280351341399392, "loss": 0.7948, "step": 2632 }, { "epoch": 0.5065409772989612, "grad_norm": 0.1717662782197644, "learning_rate": 0.0001027412289028879, "loss": 0.8006, "step": 2633 }, { "epoch": 0.5067333589842247, "grad_norm": 0.20400373848061024, "learning_rate": 0.00010267894332754242, "loss": 0.8383, "step": 2634 }, { "epoch": 0.5069257406694883, "grad_norm": 0.10667256841796197, "learning_rate": 0.00010261665671213891, "loss": 0.6418, "step": 2635 }, { "epoch": 0.5071181223547518, "grad_norm": 0.18537023633477007, "learning_rate": 0.00010255436908085919, "loss": 0.772, "step": 2636 }, { "epoch": 0.5073105040400154, "grad_norm": 0.18842117762597754, "learning_rate": 0.00010249208045788538, "loss": 0.8241, "step": 2637 }, { "epoch": 0.5075028857252789, "grad_norm": 0.1060759431538148, "learning_rate": 0.00010242979086740019, "loss": 0.6165, "step": 2638 }, { "epoch": 0.5076952674105425, "grad_norm": 0.18153275131124744, "learning_rate": 0.00010236750033358648, "loss": 0.7963, "step": 2639 }, { "epoch": 0.5078876490958061, "grad_norm": 0.217129436267987, "learning_rate": 0.00010230520888062764, "loss": 0.7824, "step": 2640 }, { "epoch": 0.5080800307810697, "grad_norm": 0.11772064629940823, "learning_rate": 0.00010224291653270738, "loss": 0.6408, "step": 2641 }, { "epoch": 0.5082724124663333, "grad_norm": 0.20610666179587986, "learning_rate": 0.00010218062331400969, "loss": 0.8539, "step": 2642 }, { "epoch": 0.5084647941515967, "grad_norm": 0.20333971660874656, "learning_rate": 0.00010211832924871889, "loss": 0.8108, "step": 2643 }, { "epoch": 0.5086571758368603, "grad_norm": 0.19833455520586543, "learning_rate": 0.00010205603436101978, "loss": 0.8109, "step": 2644 }, { "epoch": 0.5088495575221239, "grad_norm": 0.1934989230682319, "learning_rate": 0.00010199373867509734, "loss": 0.8618, "step": 2645 }, { "epoch": 0.5090419392073875, "grad_norm": 0.1930314601556391, "learning_rate": 0.0001019314422151369, "loss": 0.8085, "step": 2646 }, { "epoch": 0.509234320892651, "grad_norm": 0.18682598823431498, "learning_rate": 0.00010186914500532407, "loss": 0.8201, "step": 2647 }, { "epoch": 0.5094267025779146, "grad_norm": 0.2078536050901647, "learning_rate": 0.00010180684706984483, "loss": 0.8043, "step": 2648 }, { "epoch": 0.5096190842631781, "grad_norm": 0.18504176551655052, "learning_rate": 0.00010174454843288532, "loss": 0.7856, "step": 2649 }, { "epoch": 0.5098114659484417, "grad_norm": 0.19284204084407044, "learning_rate": 0.00010168224911863205, "loss": 0.8404, "step": 2650 }, { "epoch": 0.5100038476337053, "grad_norm": 0.21185354525929564, "learning_rate": 0.00010161994915127173, "loss": 0.8295, "step": 2651 }, { "epoch": 0.5101962293189688, "grad_norm": 0.2207345728149037, "learning_rate": 0.00010155764855499139, "loss": 0.7597, "step": 2652 }, { "epoch": 0.5103886110042324, "grad_norm": 0.20955470402442242, "learning_rate": 0.00010149534735397823, "loss": 0.817, "step": 2653 }, { "epoch": 0.510580992689496, "grad_norm": 0.1845890691022372, "learning_rate": 0.0001014330455724198, "loss": 0.8064, "step": 2654 }, { "epoch": 0.5107733743747596, "grad_norm": 0.19240894065570904, "learning_rate": 0.00010137074323450371, "loss": 0.8022, "step": 2655 }, { "epoch": 0.510965756060023, "grad_norm": 0.2156029218393447, "learning_rate": 0.00010130844036441786, "loss": 0.8226, "step": 2656 }, { "epoch": 0.5111581377452866, "grad_norm": 0.1963448649797697, "learning_rate": 0.00010124613698635043, "loss": 0.7828, "step": 2657 }, { "epoch": 0.5113505194305502, "grad_norm": 0.20113373506191298, "learning_rate": 0.00010118383312448974, "loss": 0.82, "step": 2658 }, { "epoch": 0.5115429011158138, "grad_norm": 0.18648943878654578, "learning_rate": 0.00010112152880302426, "loss": 0.8494, "step": 2659 }, { "epoch": 0.5117352828010774, "grad_norm": 0.11693503565028131, "learning_rate": 0.00010105922404614265, "loss": 0.6245, "step": 2660 }, { "epoch": 0.5119276644863409, "grad_norm": 0.16910213758342224, "learning_rate": 0.00010099691887803384, "loss": 0.8268, "step": 2661 }, { "epoch": 0.5121200461716044, "grad_norm": 0.18709897177521354, "learning_rate": 0.00010093461332288679, "loss": 0.851, "step": 2662 }, { "epoch": 0.512312427856868, "grad_norm": 0.17816698824384722, "learning_rate": 0.00010087230740489064, "loss": 0.8485, "step": 2663 }, { "epoch": 0.5125048095421316, "grad_norm": 0.11657860990269224, "learning_rate": 0.00010081000114823474, "loss": 0.6158, "step": 2664 }, { "epoch": 0.5126971912273951, "grad_norm": 0.17139816601876817, "learning_rate": 0.0001007476945771085, "loss": 0.8196, "step": 2665 }, { "epoch": 0.5128895729126587, "grad_norm": 0.20552407685803986, "learning_rate": 0.00010068538771570148, "loss": 0.8301, "step": 2666 }, { "epoch": 0.5130819545979223, "grad_norm": 0.2021196424274004, "learning_rate": 0.00010062308058820335, "loss": 0.8238, "step": 2667 }, { "epoch": 0.5132743362831859, "grad_norm": 0.18808768792745872, "learning_rate": 0.00010056077321880393, "loss": 0.8296, "step": 2668 }, { "epoch": 0.5134667179684494, "grad_norm": 0.18744741978510204, "learning_rate": 0.00010049846563169296, "loss": 0.8145, "step": 2669 }, { "epoch": 0.5136590996537129, "grad_norm": 0.17977377272563064, "learning_rate": 0.00010043615785106051, "loss": 0.7949, "step": 2670 }, { "epoch": 0.5138514813389765, "grad_norm": 0.18215915554061918, "learning_rate": 0.00010037384990109657, "loss": 0.8184, "step": 2671 }, { "epoch": 0.5140438630242401, "grad_norm": 0.1869783166870896, "learning_rate": 0.00010031154180599122, "loss": 0.7781, "step": 2672 }, { "epoch": 0.5142362447095037, "grad_norm": 0.18970761333356853, "learning_rate": 0.0001002492335899346, "loss": 0.83, "step": 2673 }, { "epoch": 0.5144286263947672, "grad_norm": 0.19492547544758895, "learning_rate": 0.00010018692527711695, "loss": 0.8452, "step": 2674 }, { "epoch": 0.5146210080800308, "grad_norm": 0.20889336952219165, "learning_rate": 0.00010012461689172845, "loss": 0.8015, "step": 2675 }, { "epoch": 0.5148133897652943, "grad_norm": 0.177479106340836, "learning_rate": 0.00010006230845795936, "loss": 0.7624, "step": 2676 }, { "epoch": 0.5150057714505579, "grad_norm": 0.191880273915257, "learning_rate": 0.0001, "loss": 0.8351, "step": 2677 }, { "epoch": 0.5151981531358215, "grad_norm": 0.19630049027061705, "learning_rate": 9.993769154204063e-05, "loss": 0.7593, "step": 2678 }, { "epoch": 0.515390534821085, "grad_norm": 0.18772444727343782, "learning_rate": 9.987538310827159e-05, "loss": 0.817, "step": 2679 }, { "epoch": 0.5155829165063486, "grad_norm": 0.18986700254061153, "learning_rate": 9.981307472288309e-05, "loss": 0.7862, "step": 2680 }, { "epoch": 0.5157752981916122, "grad_norm": 0.19094317754829185, "learning_rate": 9.975076641006542e-05, "loss": 0.8255, "step": 2681 }, { "epoch": 0.5159676798768758, "grad_norm": 0.18831374092497477, "learning_rate": 9.968845819400883e-05, "loss": 0.7982, "step": 2682 }, { "epoch": 0.5161600615621393, "grad_norm": 0.1957443066693424, "learning_rate": 9.962615009890346e-05, "loss": 0.8103, "step": 2683 }, { "epoch": 0.5163524432474028, "grad_norm": 0.19855841197405136, "learning_rate": 9.956384214893949e-05, "loss": 0.8713, "step": 2684 }, { "epoch": 0.5165448249326664, "grad_norm": 0.1930954636569386, "learning_rate": 9.950153436830706e-05, "loss": 0.8262, "step": 2685 }, { "epoch": 0.51673720661793, "grad_norm": 0.1918490625218525, "learning_rate": 9.94392267811961e-05, "loss": 0.8872, "step": 2686 }, { "epoch": 0.5169295883031936, "grad_norm": 0.1862491633345922, "learning_rate": 9.937691941179665e-05, "loss": 0.8073, "step": 2687 }, { "epoch": 0.517121969988457, "grad_norm": 0.20605962247325169, "learning_rate": 9.931461228429854e-05, "loss": 0.8178, "step": 2688 }, { "epoch": 0.5173143516737206, "grad_norm": 0.18493277751700765, "learning_rate": 9.925230542289152e-05, "loss": 0.8337, "step": 2689 }, { "epoch": 0.5175067333589842, "grad_norm": 0.2060250631612628, "learning_rate": 9.91899988517653e-05, "loss": 0.7992, "step": 2690 }, { "epoch": 0.5176991150442478, "grad_norm": 0.11769934773680162, "learning_rate": 9.912769259510938e-05, "loss": 0.6546, "step": 2691 }, { "epoch": 0.5178914967295114, "grad_norm": 0.18223854620445065, "learning_rate": 9.906538667711322e-05, "loss": 0.8229, "step": 2692 }, { "epoch": 0.5180838784147749, "grad_norm": 0.18762890779823452, "learning_rate": 9.900308112196619e-05, "loss": 0.805, "step": 2693 }, { "epoch": 0.5182762601000385, "grad_norm": 0.19603243732700634, "learning_rate": 9.894077595385737e-05, "loss": 0.819, "step": 2694 }, { "epoch": 0.518468641785302, "grad_norm": 0.1925324723307606, "learning_rate": 9.887847119697578e-05, "loss": 0.7572, "step": 2695 }, { "epoch": 0.5186610234705656, "grad_norm": 0.17511636919700435, "learning_rate": 9.88161668755103e-05, "loss": 0.8717, "step": 2696 }, { "epoch": 0.5188534051558291, "grad_norm": 0.19393980237687225, "learning_rate": 9.875386301364958e-05, "loss": 0.7955, "step": 2697 }, { "epoch": 0.5190457868410927, "grad_norm": 0.1846425831829329, "learning_rate": 9.869155963558215e-05, "loss": 0.7388, "step": 2698 }, { "epoch": 0.5192381685263563, "grad_norm": 0.17921039936376465, "learning_rate": 9.862925676549635e-05, "loss": 0.7905, "step": 2699 }, { "epoch": 0.5194305502116199, "grad_norm": 0.20448499256953542, "learning_rate": 9.856695442758023e-05, "loss": 0.815, "step": 2700 }, { "epoch": 0.5196229318968835, "grad_norm": 0.1844522081664276, "learning_rate": 9.850465264602176e-05, "loss": 0.788, "step": 2701 }, { "epoch": 0.5198153135821469, "grad_norm": 0.1882098141950047, "learning_rate": 9.844235144500864e-05, "loss": 0.7695, "step": 2702 }, { "epoch": 0.5200076952674105, "grad_norm": 0.18689876622387971, "learning_rate": 9.83800508487283e-05, "loss": 0.767, "step": 2703 }, { "epoch": 0.5202000769526741, "grad_norm": 0.20127788057319965, "learning_rate": 9.831775088136796e-05, "loss": 0.7793, "step": 2704 }, { "epoch": 0.5203924586379377, "grad_norm": 0.18460122729847359, "learning_rate": 9.825545156711472e-05, "loss": 0.7569, "step": 2705 }, { "epoch": 0.5205848403232012, "grad_norm": 0.18932125985971077, "learning_rate": 9.819315293015518e-05, "loss": 0.7955, "step": 2706 }, { "epoch": 0.5207772220084648, "grad_norm": 0.1849611310286402, "learning_rate": 9.813085499467594e-05, "loss": 0.8029, "step": 2707 }, { "epoch": 0.5209696036937284, "grad_norm": 0.17885901085172565, "learning_rate": 9.806855778486315e-05, "loss": 0.8249, "step": 2708 }, { "epoch": 0.5211619853789919, "grad_norm": 0.1908306715141879, "learning_rate": 9.800626132490269e-05, "loss": 0.8194, "step": 2709 }, { "epoch": 0.5213543670642555, "grad_norm": 0.1833162539432982, "learning_rate": 9.794396563898022e-05, "loss": 0.7813, "step": 2710 }, { "epoch": 0.521546748749519, "grad_norm": 0.19607688525479128, "learning_rate": 9.788167075128113e-05, "loss": 0.752, "step": 2711 }, { "epoch": 0.5217391304347826, "grad_norm": 0.18428053816497691, "learning_rate": 9.781937668599035e-05, "loss": 0.7737, "step": 2712 }, { "epoch": 0.5219315121200462, "grad_norm": 0.18726306587130553, "learning_rate": 9.775708346729263e-05, "loss": 0.7651, "step": 2713 }, { "epoch": 0.5221238938053098, "grad_norm": 0.18025883858246505, "learning_rate": 9.769479111937238e-05, "loss": 0.7853, "step": 2714 }, { "epoch": 0.5223162754905732, "grad_norm": 0.1315552418306608, "learning_rate": 9.763249966641353e-05, "loss": 0.6155, "step": 2715 }, { "epoch": 0.5225086571758368, "grad_norm": 0.18771132059884338, "learning_rate": 9.757020913259987e-05, "loss": 0.842, "step": 2716 }, { "epoch": 0.5227010388611004, "grad_norm": 0.20225843341588662, "learning_rate": 9.750791954211464e-05, "loss": 0.8327, "step": 2717 }, { "epoch": 0.522893420546364, "grad_norm": 0.1965484157581568, "learning_rate": 9.744563091914085e-05, "loss": 0.8083, "step": 2718 }, { "epoch": 0.5230858022316276, "grad_norm": 0.17784831248530333, "learning_rate": 9.738334328786112e-05, "loss": 0.8229, "step": 2719 }, { "epoch": 0.5232781839168911, "grad_norm": 0.1758995012316945, "learning_rate": 9.732105667245759e-05, "loss": 0.853, "step": 2720 }, { "epoch": 0.5234705656021547, "grad_norm": 0.17499289646782207, "learning_rate": 9.725877109711212e-05, "loss": 0.8523, "step": 2721 }, { "epoch": 0.5236629472874182, "grad_norm": 0.19669813630034608, "learning_rate": 9.71964865860061e-05, "loss": 0.8217, "step": 2722 }, { "epoch": 0.5238553289726818, "grad_norm": 0.187905242517817, "learning_rate": 9.713420316332049e-05, "loss": 0.7955, "step": 2723 }, { "epoch": 0.5240477106579453, "grad_norm": 0.19543427886876977, "learning_rate": 9.707192085323589e-05, "loss": 0.8393, "step": 2724 }, { "epoch": 0.5242400923432089, "grad_norm": 0.17351910012836985, "learning_rate": 9.700963967993246e-05, "loss": 0.7846, "step": 2725 }, { "epoch": 0.5244324740284725, "grad_norm": 0.1914318253192299, "learning_rate": 9.694735966758982e-05, "loss": 0.8322, "step": 2726 }, { "epoch": 0.5246248557137361, "grad_norm": 0.18452313253336833, "learning_rate": 9.688508084038729e-05, "loss": 0.7709, "step": 2727 }, { "epoch": 0.5248172373989997, "grad_norm": 0.19296690242588724, "learning_rate": 9.682280322250366e-05, "loss": 0.7662, "step": 2728 }, { "epoch": 0.5250096190842631, "grad_norm": 0.18230109786299423, "learning_rate": 9.676052683811715e-05, "loss": 0.8121, "step": 2729 }, { "epoch": 0.5252020007695267, "grad_norm": 0.18569379306260322, "learning_rate": 9.669825171140568e-05, "loss": 0.8166, "step": 2730 }, { "epoch": 0.5253943824547903, "grad_norm": 0.17229199465735287, "learning_rate": 9.66359778665466e-05, "loss": 0.7732, "step": 2731 }, { "epoch": 0.5255867641400539, "grad_norm": 0.18118936955815784, "learning_rate": 9.657370532771671e-05, "loss": 0.7961, "step": 2732 }, { "epoch": 0.5257791458253175, "grad_norm": 0.19375497381066148, "learning_rate": 9.65114341190924e-05, "loss": 0.8053, "step": 2733 }, { "epoch": 0.525971527510581, "grad_norm": 0.19574983060417517, "learning_rate": 9.64491642648495e-05, "loss": 0.8446, "step": 2734 }, { "epoch": 0.5261639091958445, "grad_norm": 0.17213051916571398, "learning_rate": 9.638689578916326e-05, "loss": 0.825, "step": 2735 }, { "epoch": 0.5263562908811081, "grad_norm": 0.19144460084399842, "learning_rate": 9.632462871620847e-05, "loss": 0.8473, "step": 2736 }, { "epoch": 0.5265486725663717, "grad_norm": 0.1989552369396223, "learning_rate": 9.62623630701594e-05, "loss": 0.8074, "step": 2737 }, { "epoch": 0.5267410542516352, "grad_norm": 0.11085296775691596, "learning_rate": 9.620009887518963e-05, "loss": 0.6741, "step": 2738 }, { "epoch": 0.5269334359368988, "grad_norm": 0.17983960281068517, "learning_rate": 9.613783615547233e-05, "loss": 0.8004, "step": 2739 }, { "epoch": 0.5271258176221624, "grad_norm": 0.19938778601070242, "learning_rate": 9.607557493518005e-05, "loss": 0.7862, "step": 2740 }, { "epoch": 0.527318199307426, "grad_norm": 0.19921404323445854, "learning_rate": 9.601331523848464e-05, "loss": 0.7943, "step": 2741 }, { "epoch": 0.5275105809926895, "grad_norm": 0.2022747720717196, "learning_rate": 9.595105708955757e-05, "loss": 0.7516, "step": 2742 }, { "epoch": 0.527702962677953, "grad_norm": 0.1905316557181197, "learning_rate": 9.588880051256951e-05, "loss": 0.849, "step": 2743 }, { "epoch": 0.5278953443632166, "grad_norm": 0.18483972172970695, "learning_rate": 9.582654553169064e-05, "loss": 0.775, "step": 2744 }, { "epoch": 0.5280877260484802, "grad_norm": 0.18960834162221904, "learning_rate": 9.576429217109053e-05, "loss": 0.7811, "step": 2745 }, { "epoch": 0.5282801077337438, "grad_norm": 0.10330382750698262, "learning_rate": 9.5702040454938e-05, "loss": 0.6018, "step": 2746 }, { "epoch": 0.5284724894190073, "grad_norm": 0.19660562237540713, "learning_rate": 9.563979040740138e-05, "loss": 0.776, "step": 2747 }, { "epoch": 0.5286648711042709, "grad_norm": 0.11045999958919307, "learning_rate": 9.557754205264826e-05, "loss": 0.6137, "step": 2748 }, { "epoch": 0.5288572527895344, "grad_norm": 0.19452369860119492, "learning_rate": 9.551529541484554e-05, "loss": 0.838, "step": 2749 }, { "epoch": 0.529049634474798, "grad_norm": 0.2169997693748785, "learning_rate": 9.545305051815956e-05, "loss": 0.8445, "step": 2750 }, { "epoch": 0.5292420161600616, "grad_norm": 0.20044706346331198, "learning_rate": 9.539080738675597e-05, "loss": 0.8647, "step": 2751 }, { "epoch": 0.5294343978453251, "grad_norm": 0.18921777307219503, "learning_rate": 9.53285660447996e-05, "loss": 0.8192, "step": 2752 }, { "epoch": 0.5296267795305887, "grad_norm": 0.20317700420309823, "learning_rate": 9.526632651645475e-05, "loss": 0.7988, "step": 2753 }, { "epoch": 0.5298191612158523, "grad_norm": 0.1944382206477693, "learning_rate": 9.520408882588497e-05, "loss": 0.8207, "step": 2754 }, { "epoch": 0.5300115429011159, "grad_norm": 0.20798099557004338, "learning_rate": 9.514185299725298e-05, "loss": 0.8314, "step": 2755 }, { "epoch": 0.5302039245863793, "grad_norm": 0.182064233521188, "learning_rate": 9.507961905472093e-05, "loss": 0.7774, "step": 2756 }, { "epoch": 0.5303963062716429, "grad_norm": 0.1828157589381928, "learning_rate": 9.501738702245022e-05, "loss": 0.7933, "step": 2757 }, { "epoch": 0.5305886879569065, "grad_norm": 0.2280655239166218, "learning_rate": 9.495515692460138e-05, "loss": 0.7972, "step": 2758 }, { "epoch": 0.5307810696421701, "grad_norm": 0.19193014013356313, "learning_rate": 9.489292878533434e-05, "loss": 0.7592, "step": 2759 }, { "epoch": 0.5309734513274337, "grad_norm": 0.18737516758836145, "learning_rate": 9.483070262880822e-05, "loss": 0.8903, "step": 2760 }, { "epoch": 0.5311658330126972, "grad_norm": 0.20169639358510166, "learning_rate": 9.476847847918127e-05, "loss": 0.8334, "step": 2761 }, { "epoch": 0.5313582146979607, "grad_norm": 0.1910134276802344, "learning_rate": 9.47062563606111e-05, "loss": 0.8514, "step": 2762 }, { "epoch": 0.5315505963832243, "grad_norm": 0.18378896124481184, "learning_rate": 9.464403629725453e-05, "loss": 0.7879, "step": 2763 }, { "epoch": 0.5317429780684879, "grad_norm": 0.18589030649354657, "learning_rate": 9.458181831326743e-05, "loss": 0.7916, "step": 2764 }, { "epoch": 0.5319353597537514, "grad_norm": 0.20819373882977385, "learning_rate": 9.451960243280506e-05, "loss": 0.8178, "step": 2765 }, { "epoch": 0.532127741439015, "grad_norm": 0.20372540522695892, "learning_rate": 9.44573886800217e-05, "loss": 0.8324, "step": 2766 }, { "epoch": 0.5323201231242786, "grad_norm": 0.1977552925644709, "learning_rate": 9.439517707907088e-05, "loss": 0.8303, "step": 2767 }, { "epoch": 0.5325125048095422, "grad_norm": 0.2352179432213637, "learning_rate": 9.433296765410534e-05, "loss": 0.8076, "step": 2768 }, { "epoch": 0.5327048864948057, "grad_norm": 0.11703264036313442, "learning_rate": 9.427076042927683e-05, "loss": 0.622, "step": 2769 }, { "epoch": 0.5328972681800692, "grad_norm": 0.2016852788620623, "learning_rate": 9.420855542873637e-05, "loss": 0.801, "step": 2770 }, { "epoch": 0.5330896498653328, "grad_norm": 0.2020955914381733, "learning_rate": 9.414635267663417e-05, "loss": 0.7718, "step": 2771 }, { "epoch": 0.5332820315505964, "grad_norm": 0.18621161557591626, "learning_rate": 9.408415219711934e-05, "loss": 0.7482, "step": 2772 }, { "epoch": 0.53347441323586, "grad_norm": 0.1912597044358707, "learning_rate": 9.402195401434036e-05, "loss": 0.7251, "step": 2773 }, { "epoch": 0.5336667949211235, "grad_norm": 0.2005681454103332, "learning_rate": 9.395975815244466e-05, "loss": 0.7876, "step": 2774 }, { "epoch": 0.533859176606387, "grad_norm": 0.1960996214104589, "learning_rate": 9.389756463557878e-05, "loss": 0.8231, "step": 2775 }, { "epoch": 0.5340515582916506, "grad_norm": 0.1809201459747362, "learning_rate": 9.383537348788843e-05, "loss": 0.8241, "step": 2776 }, { "epoch": 0.5342439399769142, "grad_norm": 0.11881940102518568, "learning_rate": 9.377318473351838e-05, "loss": 0.711, "step": 2777 }, { "epoch": 0.5344363216621778, "grad_norm": 0.11084496953727085, "learning_rate": 9.371099839661238e-05, "loss": 0.6386, "step": 2778 }, { "epoch": 0.5346287033474413, "grad_norm": 0.197048736566187, "learning_rate": 9.364881450131336e-05, "loss": 0.7742, "step": 2779 }, { "epoch": 0.5348210850327049, "grad_norm": 0.19445930498711758, "learning_rate": 9.358663307176323e-05, "loss": 0.7941, "step": 2780 }, { "epoch": 0.5350134667179685, "grad_norm": 0.18866680092475657, "learning_rate": 9.352445413210293e-05, "loss": 0.8287, "step": 2781 }, { "epoch": 0.535205848403232, "grad_norm": 0.18726043408879806, "learning_rate": 9.34622777064725e-05, "loss": 0.8241, "step": 2782 }, { "epoch": 0.5353982300884956, "grad_norm": 0.19862174560935425, "learning_rate": 9.340010381901101e-05, "loss": 0.7709, "step": 2783 }, { "epoch": 0.5355906117737591, "grad_norm": 0.19363490350974227, "learning_rate": 9.33379324938564e-05, "loss": 0.7975, "step": 2784 }, { "epoch": 0.5357829934590227, "grad_norm": 0.19255337196619093, "learning_rate": 9.327576375514581e-05, "loss": 0.8225, "step": 2785 }, { "epoch": 0.5359753751442863, "grad_norm": 0.18729749666591333, "learning_rate": 9.321359762701526e-05, "loss": 0.7998, "step": 2786 }, { "epoch": 0.5361677568295499, "grad_norm": 0.0990798285142627, "learning_rate": 9.315143413359976e-05, "loss": 0.6311, "step": 2787 }, { "epoch": 0.5363601385148133, "grad_norm": 0.19892256943185954, "learning_rate": 9.308927329903332e-05, "loss": 0.7591, "step": 2788 }, { "epoch": 0.5365525202000769, "grad_norm": 0.2416304530634409, "learning_rate": 9.302711514744896e-05, "loss": 0.7903, "step": 2789 }, { "epoch": 0.5367449018853405, "grad_norm": 0.18313332196884544, "learning_rate": 9.296495970297854e-05, "loss": 0.8258, "step": 2790 }, { "epoch": 0.5369372835706041, "grad_norm": 0.20056927659523816, "learning_rate": 9.290280698975306e-05, "loss": 0.7668, "step": 2791 }, { "epoch": 0.5371296652558677, "grad_norm": 0.18597215447844137, "learning_rate": 9.28406570319022e-05, "loss": 0.813, "step": 2792 }, { "epoch": 0.5373220469411312, "grad_norm": 0.1838635256801194, "learning_rate": 9.27785098535548e-05, "loss": 0.8082, "step": 2793 }, { "epoch": 0.5375144286263948, "grad_norm": 0.19151841454311444, "learning_rate": 9.271636547883856e-05, "loss": 0.75, "step": 2794 }, { "epoch": 0.5377068103116583, "grad_norm": 0.198692841500491, "learning_rate": 9.265422393187997e-05, "loss": 0.7856, "step": 2795 }, { "epoch": 0.5378991919969219, "grad_norm": 0.20435997412460163, "learning_rate": 9.259208523680458e-05, "loss": 0.8275, "step": 2796 }, { "epoch": 0.5380915736821854, "grad_norm": 0.18173277237619018, "learning_rate": 9.252994941773678e-05, "loss": 0.7568, "step": 2797 }, { "epoch": 0.538283955367449, "grad_norm": 0.19214192548607759, "learning_rate": 9.24678164987998e-05, "loss": 0.7947, "step": 2798 }, { "epoch": 0.5384763370527126, "grad_norm": 0.1992983159641856, "learning_rate": 9.240568650411581e-05, "loss": 0.7556, "step": 2799 }, { "epoch": 0.5386687187379762, "grad_norm": 0.19612393995649507, "learning_rate": 9.234355945780581e-05, "loss": 0.7929, "step": 2800 }, { "epoch": 0.5388611004232398, "grad_norm": 0.24222011112858413, "learning_rate": 9.228143538398962e-05, "loss": 0.8427, "step": 2801 }, { "epoch": 0.5390534821085032, "grad_norm": 0.1996294770526611, "learning_rate": 9.221931430678598e-05, "loss": 0.8245, "step": 2802 }, { "epoch": 0.5392458637937668, "grad_norm": 0.2111693144575189, "learning_rate": 9.215719625031245e-05, "loss": 0.8055, "step": 2803 }, { "epoch": 0.5394382454790304, "grad_norm": 0.21153569000962313, "learning_rate": 9.209508123868533e-05, "loss": 0.8501, "step": 2804 }, { "epoch": 0.539630627164294, "grad_norm": 0.19040863858031923, "learning_rate": 9.203296929601986e-05, "loss": 0.7671, "step": 2805 }, { "epoch": 0.5398230088495575, "grad_norm": 0.20147253024701905, "learning_rate": 9.197086044643004e-05, "loss": 0.7867, "step": 2806 }, { "epoch": 0.5400153905348211, "grad_norm": 0.18188211396659282, "learning_rate": 9.190875471402865e-05, "loss": 0.7656, "step": 2807 }, { "epoch": 0.5402077722200846, "grad_norm": 0.19917977600556802, "learning_rate": 9.184665212292722e-05, "loss": 0.8031, "step": 2808 }, { "epoch": 0.5404001539053482, "grad_norm": 0.18225219219250913, "learning_rate": 9.178455269723623e-05, "loss": 0.7862, "step": 2809 }, { "epoch": 0.5405925355906118, "grad_norm": 0.10061274223361652, "learning_rate": 9.172245646106469e-05, "loss": 0.6296, "step": 2810 }, { "epoch": 0.5407849172758753, "grad_norm": 0.18559541727346615, "learning_rate": 9.16603634385206e-05, "loss": 0.7739, "step": 2811 }, { "epoch": 0.5409772989611389, "grad_norm": 0.11589432098686127, "learning_rate": 9.159827365371056e-05, "loss": 0.6872, "step": 2812 }, { "epoch": 0.5411696806464025, "grad_norm": 0.2018017957344978, "learning_rate": 9.153618713073995e-05, "loss": 0.8498, "step": 2813 }, { "epoch": 0.5413620623316661, "grad_norm": 0.19018115319357268, "learning_rate": 9.14741038937129e-05, "loss": 0.7931, "step": 2814 }, { "epoch": 0.5415544440169295, "grad_norm": 0.1920237322578562, "learning_rate": 9.141202396673232e-05, "loss": 0.8172, "step": 2815 }, { "epoch": 0.5417468257021931, "grad_norm": 0.2112416464396105, "learning_rate": 9.134994737389969e-05, "loss": 0.802, "step": 2816 }, { "epoch": 0.5419392073874567, "grad_norm": 0.18135558815135713, "learning_rate": 9.128787413931536e-05, "loss": 0.8217, "step": 2817 }, { "epoch": 0.5421315890727203, "grad_norm": 0.19276978624988508, "learning_rate": 9.12258042870782e-05, "loss": 0.7916, "step": 2818 }, { "epoch": 0.5423239707579839, "grad_norm": 0.19541874630413203, "learning_rate": 9.116373784128596e-05, "loss": 0.7765, "step": 2819 }, { "epoch": 0.5425163524432474, "grad_norm": 0.19010362042871268, "learning_rate": 9.110167482603494e-05, "loss": 0.7663, "step": 2820 }, { "epoch": 0.542708734128511, "grad_norm": 0.17883947726878707, "learning_rate": 9.10396152654201e-05, "loss": 0.8193, "step": 2821 }, { "epoch": 0.5429011158137745, "grad_norm": 0.18637249678613663, "learning_rate": 9.097755918353512e-05, "loss": 0.8108, "step": 2822 }, { "epoch": 0.5430934974990381, "grad_norm": 0.1990690480547995, "learning_rate": 9.091550660447235e-05, "loss": 0.8343, "step": 2823 }, { "epoch": 0.5432858791843016, "grad_norm": 0.2041087749865908, "learning_rate": 9.085345755232268e-05, "loss": 0.8364, "step": 2824 }, { "epoch": 0.5434782608695652, "grad_norm": 0.19180523091556778, "learning_rate": 9.079141205117572e-05, "loss": 0.7844, "step": 2825 }, { "epoch": 0.5436706425548288, "grad_norm": 0.2128168008279131, "learning_rate": 9.072937012511969e-05, "loss": 0.7926, "step": 2826 }, { "epoch": 0.5438630242400924, "grad_norm": 0.20321032230379193, "learning_rate": 9.066733179824134e-05, "loss": 0.8735, "step": 2827 }, { "epoch": 0.544055405925356, "grad_norm": 0.19528896611146637, "learning_rate": 9.060529709462611e-05, "loss": 0.7389, "step": 2828 }, { "epoch": 0.5442477876106194, "grad_norm": 0.18479248464861578, "learning_rate": 9.054326603835807e-05, "loss": 0.8227, "step": 2829 }, { "epoch": 0.544440169295883, "grad_norm": 0.18425172584241764, "learning_rate": 9.04812386535197e-05, "loss": 0.8367, "step": 2830 }, { "epoch": 0.5446325509811466, "grad_norm": 0.20897440430253544, "learning_rate": 9.041921496419226e-05, "loss": 0.8378, "step": 2831 }, { "epoch": 0.5448249326664102, "grad_norm": 0.19101157354320034, "learning_rate": 9.035719499445544e-05, "loss": 0.8067, "step": 2832 }, { "epoch": 0.5450173143516738, "grad_norm": 0.1987232020016017, "learning_rate": 9.029517876838755e-05, "loss": 0.7564, "step": 2833 }, { "epoch": 0.5452096960369373, "grad_norm": 0.1092123344528083, "learning_rate": 9.023316631006537e-05, "loss": 0.6237, "step": 2834 }, { "epoch": 0.5454020777222008, "grad_norm": 0.18614747494504144, "learning_rate": 9.017115764356435e-05, "loss": 0.8039, "step": 2835 }, { "epoch": 0.5455944594074644, "grad_norm": 0.10052398294664179, "learning_rate": 9.010915279295831e-05, "loss": 0.614, "step": 2836 }, { "epoch": 0.545786841092728, "grad_norm": 0.19279874857659432, "learning_rate": 9.004715178231975e-05, "loss": 0.7876, "step": 2837 }, { "epoch": 0.5459792227779915, "grad_norm": 0.1113678185542066, "learning_rate": 8.998515463571953e-05, "loss": 0.6306, "step": 2838 }, { "epoch": 0.5461716044632551, "grad_norm": 0.177730475798761, "learning_rate": 8.99231613772271e-05, "loss": 0.761, "step": 2839 }, { "epoch": 0.5463639861485187, "grad_norm": 0.20402606924400735, "learning_rate": 8.986117203091042e-05, "loss": 0.7763, "step": 2840 }, { "epoch": 0.5465563678337823, "grad_norm": 0.17044808926555652, "learning_rate": 8.97991866208358e-05, "loss": 0.8549, "step": 2841 }, { "epoch": 0.5467487495190458, "grad_norm": 0.20448758634000946, "learning_rate": 8.973720517106814e-05, "loss": 0.7603, "step": 2842 }, { "epoch": 0.5469411312043093, "grad_norm": 0.2006834350055488, "learning_rate": 8.967522770567084e-05, "loss": 0.8413, "step": 2843 }, { "epoch": 0.5471335128895729, "grad_norm": 0.20971034227378, "learning_rate": 8.961325424870561e-05, "loss": 0.7839, "step": 2844 }, { "epoch": 0.5473258945748365, "grad_norm": 0.19629074122817183, "learning_rate": 8.955128482423272e-05, "loss": 0.8526, "step": 2845 }, { "epoch": 0.5475182762601001, "grad_norm": 0.1981310875809713, "learning_rate": 8.948931945631082e-05, "loss": 0.8384, "step": 2846 }, { "epoch": 0.5477106579453636, "grad_norm": 0.18230644106446098, "learning_rate": 8.9427358168997e-05, "loss": 0.8453, "step": 2847 }, { "epoch": 0.5479030396306271, "grad_norm": 0.19364792145235618, "learning_rate": 8.936540098634674e-05, "loss": 0.7818, "step": 2848 }, { "epoch": 0.5480954213158907, "grad_norm": 0.1912014172765943, "learning_rate": 8.930344793241403e-05, "loss": 0.7796, "step": 2849 }, { "epoch": 0.5482878030011543, "grad_norm": 0.18263826617957238, "learning_rate": 8.924149903125108e-05, "loss": 0.7585, "step": 2850 }, { "epoch": 0.5484801846864179, "grad_norm": 0.19408681410848547, "learning_rate": 8.917955430690865e-05, "loss": 0.8224, "step": 2851 }, { "epoch": 0.5486725663716814, "grad_norm": 0.18313320729349966, "learning_rate": 8.91176137834358e-05, "loss": 0.8024, "step": 2852 }, { "epoch": 0.548864948056945, "grad_norm": 0.17240333580187597, "learning_rate": 8.905567748487996e-05, "loss": 0.7688, "step": 2853 }, { "epoch": 0.5490573297422086, "grad_norm": 0.1958969310167052, "learning_rate": 8.899374543528694e-05, "loss": 0.8114, "step": 2854 }, { "epoch": 0.5492497114274721, "grad_norm": 0.1138058775631987, "learning_rate": 8.893181765870094e-05, "loss": 0.6415, "step": 2855 }, { "epoch": 0.5494420931127356, "grad_norm": 0.19369081259462448, "learning_rate": 8.886989417916435e-05, "loss": 0.7888, "step": 2856 }, { "epoch": 0.5496344747979992, "grad_norm": 0.18522706442501982, "learning_rate": 8.880797502071809e-05, "loss": 0.7776, "step": 2857 }, { "epoch": 0.5498268564832628, "grad_norm": 0.20391434036105313, "learning_rate": 8.87460602074013e-05, "loss": 0.7845, "step": 2858 }, { "epoch": 0.5500192381685264, "grad_norm": 0.18968499230435407, "learning_rate": 8.86841497632514e-05, "loss": 0.7686, "step": 2859 }, { "epoch": 0.55021161985379, "grad_norm": 0.1969607332077735, "learning_rate": 8.862224371230417e-05, "loss": 0.7614, "step": 2860 }, { "epoch": 0.5504040015390534, "grad_norm": 0.18069010029190313, "learning_rate": 8.85603420785937e-05, "loss": 0.7661, "step": 2861 }, { "epoch": 0.550596383224317, "grad_norm": 0.18722206823467888, "learning_rate": 8.84984448861523e-05, "loss": 0.8471, "step": 2862 }, { "epoch": 0.5507887649095806, "grad_norm": 0.23160572409463234, "learning_rate": 8.843655215901059e-05, "loss": 0.7881, "step": 2863 }, { "epoch": 0.5509811465948442, "grad_norm": 0.17788850831315828, "learning_rate": 8.837466392119752e-05, "loss": 0.7184, "step": 2864 }, { "epoch": 0.5511735282801077, "grad_norm": 0.179201981202357, "learning_rate": 8.831278019674016e-05, "loss": 0.756, "step": 2865 }, { "epoch": 0.5513659099653713, "grad_norm": 0.18970026774360818, "learning_rate": 8.825090100966395e-05, "loss": 0.7775, "step": 2866 }, { "epoch": 0.5515582916506349, "grad_norm": 0.21183568637356048, "learning_rate": 8.818902638399246e-05, "loss": 0.7983, "step": 2867 }, { "epoch": 0.5517506733358984, "grad_norm": 0.19785706139304857, "learning_rate": 8.81271563437476e-05, "loss": 0.8109, "step": 2868 }, { "epoch": 0.551943055021162, "grad_norm": 0.1824542770632194, "learning_rate": 8.806529091294948e-05, "loss": 0.7811, "step": 2869 }, { "epoch": 0.5521354367064255, "grad_norm": 0.1778267009262026, "learning_rate": 8.800343011561632e-05, "loss": 0.8135, "step": 2870 }, { "epoch": 0.5523278183916891, "grad_norm": 0.18812738285814073, "learning_rate": 8.794157397576464e-05, "loss": 0.791, "step": 2871 }, { "epoch": 0.5525202000769527, "grad_norm": 0.11166398711141788, "learning_rate": 8.787972251740915e-05, "loss": 0.6595, "step": 2872 }, { "epoch": 0.5527125817622163, "grad_norm": 0.1962166992630005, "learning_rate": 8.781787576456268e-05, "loss": 0.7667, "step": 2873 }, { "epoch": 0.5529049634474797, "grad_norm": 0.1941315294221021, "learning_rate": 8.775603374123626e-05, "loss": 0.7986, "step": 2874 }, { "epoch": 0.5530973451327433, "grad_norm": 0.17971784167343277, "learning_rate": 8.769419647143917e-05, "loss": 0.7773, "step": 2875 }, { "epoch": 0.5532897268180069, "grad_norm": 0.20397503856112997, "learning_rate": 8.763236397917864e-05, "loss": 0.8009, "step": 2876 }, { "epoch": 0.5534821085032705, "grad_norm": 0.19789688210568576, "learning_rate": 8.757053628846029e-05, "loss": 0.805, "step": 2877 }, { "epoch": 0.5536744901885341, "grad_norm": 0.20584382260139358, "learning_rate": 8.750871342328771e-05, "loss": 0.7958, "step": 2878 }, { "epoch": 0.5538668718737976, "grad_norm": 0.18957593431729042, "learning_rate": 8.744689540766264e-05, "loss": 0.7904, "step": 2879 }, { "epoch": 0.5540592535590612, "grad_norm": 0.18931773985683517, "learning_rate": 8.7385082265585e-05, "loss": 0.7948, "step": 2880 }, { "epoch": 0.5542516352443247, "grad_norm": 0.1941371540489715, "learning_rate": 8.732327402105279e-05, "loss": 0.7951, "step": 2881 }, { "epoch": 0.5544440169295883, "grad_norm": 0.19382419761541597, "learning_rate": 8.726147069806205e-05, "loss": 0.807, "step": 2882 }, { "epoch": 0.5546363986148518, "grad_norm": 0.17454715952671832, "learning_rate": 8.719967232060697e-05, "loss": 0.7964, "step": 2883 }, { "epoch": 0.5548287803001154, "grad_norm": 0.19238086491747836, "learning_rate": 8.713787891267989e-05, "loss": 0.7425, "step": 2884 }, { "epoch": 0.555021161985379, "grad_norm": 0.19889743219121253, "learning_rate": 8.707609049827102e-05, "loss": 0.7899, "step": 2885 }, { "epoch": 0.5552135436706426, "grad_norm": 0.19669311114409072, "learning_rate": 8.70143071013688e-05, "loss": 0.7832, "step": 2886 }, { "epoch": 0.5554059253559062, "grad_norm": 0.2043669962119901, "learning_rate": 8.695252874595971e-05, "loss": 0.7594, "step": 2887 }, { "epoch": 0.5555983070411696, "grad_norm": 0.21395337185953814, "learning_rate": 8.689075545602815e-05, "loss": 0.7646, "step": 2888 }, { "epoch": 0.5557906887264332, "grad_norm": 0.10240496351171419, "learning_rate": 8.68289872555567e-05, "loss": 0.5979, "step": 2889 }, { "epoch": 0.5559830704116968, "grad_norm": 0.18915182170922284, "learning_rate": 8.676722416852594e-05, "loss": 0.7973, "step": 2890 }, { "epoch": 0.5561754520969604, "grad_norm": 0.17919239879419574, "learning_rate": 8.670546621891434e-05, "loss": 0.7859, "step": 2891 }, { "epoch": 0.556367833782224, "grad_norm": 0.18065377414524175, "learning_rate": 8.664371343069849e-05, "loss": 0.8066, "step": 2892 }, { "epoch": 0.5565602154674875, "grad_norm": 0.1747240922486361, "learning_rate": 8.658196582785296e-05, "loss": 0.7674, "step": 2893 }, { "epoch": 0.556752597152751, "grad_norm": 0.18645822769047724, "learning_rate": 8.652022343435028e-05, "loss": 0.8047, "step": 2894 }, { "epoch": 0.5569449788380146, "grad_norm": 0.19293620483957818, "learning_rate": 8.645848627416103e-05, "loss": 0.7921, "step": 2895 }, { "epoch": 0.5571373605232782, "grad_norm": 0.19866832065094073, "learning_rate": 8.63967543712536e-05, "loss": 0.8062, "step": 2896 }, { "epoch": 0.5573297422085417, "grad_norm": 0.19596688882773744, "learning_rate": 8.633502774959453e-05, "loss": 0.7987, "step": 2897 }, { "epoch": 0.5575221238938053, "grad_norm": 0.20092683024747562, "learning_rate": 8.627330643314818e-05, "loss": 0.7947, "step": 2898 }, { "epoch": 0.5577145055790689, "grad_norm": 0.18886793019122514, "learning_rate": 8.62115904458769e-05, "loss": 0.7641, "step": 2899 }, { "epoch": 0.5579068872643325, "grad_norm": 0.19832722051715154, "learning_rate": 8.614987981174094e-05, "loss": 0.7667, "step": 2900 }, { "epoch": 0.558099268949596, "grad_norm": 0.19953653217195924, "learning_rate": 8.608817455469854e-05, "loss": 0.7834, "step": 2901 }, { "epoch": 0.5582916506348595, "grad_norm": 0.18420609301418533, "learning_rate": 8.602647469870574e-05, "loss": 0.8509, "step": 2902 }, { "epoch": 0.5584840323201231, "grad_norm": 0.183452700316472, "learning_rate": 8.596478026771659e-05, "loss": 0.8376, "step": 2903 }, { "epoch": 0.5586764140053867, "grad_norm": 0.18995071463181662, "learning_rate": 8.590309128568303e-05, "loss": 0.8269, "step": 2904 }, { "epoch": 0.5588687956906503, "grad_norm": 0.18513944110439745, "learning_rate": 8.584140777655476e-05, "loss": 0.8358, "step": 2905 }, { "epoch": 0.5590611773759138, "grad_norm": 0.17676050480779995, "learning_rate": 8.577972976427949e-05, "loss": 0.7734, "step": 2906 }, { "epoch": 0.5592535590611774, "grad_norm": 0.17765286684323195, "learning_rate": 8.571805727280277e-05, "loss": 0.7884, "step": 2907 }, { "epoch": 0.5594459407464409, "grad_norm": 0.1775781492357062, "learning_rate": 8.565639032606794e-05, "loss": 0.8273, "step": 2908 }, { "epoch": 0.5596383224317045, "grad_norm": 0.20321595161436856, "learning_rate": 8.559472894801623e-05, "loss": 0.759, "step": 2909 }, { "epoch": 0.5598307041169681, "grad_norm": 0.18879706778153882, "learning_rate": 8.553307316258677e-05, "loss": 0.7453, "step": 2910 }, { "epoch": 0.5600230858022316, "grad_norm": 0.19214786872990308, "learning_rate": 8.547142299371641e-05, "loss": 0.7652, "step": 2911 }, { "epoch": 0.5602154674874952, "grad_norm": 0.18212117413396664, "learning_rate": 8.540977846533985e-05, "loss": 0.8007, "step": 2912 }, { "epoch": 0.5604078491727588, "grad_norm": 0.1888040859685135, "learning_rate": 8.534813960138969e-05, "loss": 0.8146, "step": 2913 }, { "epoch": 0.5606002308580224, "grad_norm": 0.17456483948428111, "learning_rate": 8.528650642579617e-05, "loss": 0.7635, "step": 2914 }, { "epoch": 0.5607926125432858, "grad_norm": 0.18971182299090725, "learning_rate": 8.522487896248749e-05, "loss": 0.8284, "step": 2915 }, { "epoch": 0.5609849942285494, "grad_norm": 0.20070621612274372, "learning_rate": 8.51632572353895e-05, "loss": 0.8691, "step": 2916 }, { "epoch": 0.561177375913813, "grad_norm": 0.1912216785734748, "learning_rate": 8.51016412684259e-05, "loss": 0.7794, "step": 2917 }, { "epoch": 0.5613697575990766, "grad_norm": 0.1855084030748606, "learning_rate": 8.504003108551814e-05, "loss": 0.8136, "step": 2918 }, { "epoch": 0.5615621392843402, "grad_norm": 0.18298998552954823, "learning_rate": 8.497842671058538e-05, "loss": 0.8471, "step": 2919 }, { "epoch": 0.5617545209696037, "grad_norm": 0.20304320304403997, "learning_rate": 8.491682816754456e-05, "loss": 0.7929, "step": 2920 }, { "epoch": 0.5619469026548672, "grad_norm": 0.17567917535874678, "learning_rate": 8.485523548031042e-05, "loss": 0.82, "step": 2921 }, { "epoch": 0.5621392843401308, "grad_norm": 0.19492815160815358, "learning_rate": 8.479364867279529e-05, "loss": 0.8221, "step": 2922 }, { "epoch": 0.5623316660253944, "grad_norm": 0.19983775262924272, "learning_rate": 8.47320677689093e-05, "loss": 0.8191, "step": 2923 }, { "epoch": 0.5625240477106579, "grad_norm": 0.1927651848586022, "learning_rate": 8.467049279256033e-05, "loss": 0.7961, "step": 2924 }, { "epoch": 0.5627164293959215, "grad_norm": 0.1942217147649761, "learning_rate": 8.460892376765386e-05, "loss": 0.8641, "step": 2925 }, { "epoch": 0.5629088110811851, "grad_norm": 0.19972465391207447, "learning_rate": 8.454736071809308e-05, "loss": 0.7668, "step": 2926 }, { "epoch": 0.5631011927664487, "grad_norm": 0.18783601858341445, "learning_rate": 8.448580366777897e-05, "loss": 0.8151, "step": 2927 }, { "epoch": 0.5632935744517122, "grad_norm": 0.1925055872879775, "learning_rate": 8.442425264061e-05, "loss": 0.8275, "step": 2928 }, { "epoch": 0.5634859561369757, "grad_norm": 0.20142758709949, "learning_rate": 8.436270766048245e-05, "loss": 0.7791, "step": 2929 }, { "epoch": 0.5636783378222393, "grad_norm": 0.19830991510044751, "learning_rate": 8.430116875129022e-05, "loss": 0.8172, "step": 2930 }, { "epoch": 0.5638707195075029, "grad_norm": 0.18533254652865672, "learning_rate": 8.42396359369248e-05, "loss": 0.7895, "step": 2931 }, { "epoch": 0.5640631011927665, "grad_norm": 0.2036757751531752, "learning_rate": 8.417810924127533e-05, "loss": 0.836, "step": 2932 }, { "epoch": 0.56425548287803, "grad_norm": 0.18463079258937068, "learning_rate": 8.411658868822866e-05, "loss": 0.8027, "step": 2933 }, { "epoch": 0.5644478645632935, "grad_norm": 0.17434968215400695, "learning_rate": 8.40550743016691e-05, "loss": 0.7397, "step": 2934 }, { "epoch": 0.5646402462485571, "grad_norm": 0.1979919370799327, "learning_rate": 8.399356610547869e-05, "loss": 0.7847, "step": 2935 }, { "epoch": 0.5648326279338207, "grad_norm": 0.18298585738920697, "learning_rate": 8.39320641235371e-05, "loss": 0.7922, "step": 2936 }, { "epoch": 0.5650250096190843, "grad_norm": 0.1826597737135554, "learning_rate": 8.38705683797214e-05, "loss": 0.807, "step": 2937 }, { "epoch": 0.5652173913043478, "grad_norm": 0.21368589049171155, "learning_rate": 8.380907889790641e-05, "loss": 0.7642, "step": 2938 }, { "epoch": 0.5654097729896114, "grad_norm": 0.19488542632046393, "learning_rate": 8.374759570196449e-05, "loss": 0.7389, "step": 2939 }, { "epoch": 0.565602154674875, "grad_norm": 0.19890942673320275, "learning_rate": 8.368611881576547e-05, "loss": 0.8151, "step": 2940 }, { "epoch": 0.5657945363601385, "grad_norm": 0.18665517432211548, "learning_rate": 8.362464826317686e-05, "loss": 0.7855, "step": 2941 }, { "epoch": 0.5659869180454021, "grad_norm": 0.20232355641127744, "learning_rate": 8.356318406806358e-05, "loss": 0.796, "step": 2942 }, { "epoch": 0.5661792997306656, "grad_norm": 0.18696423151140354, "learning_rate": 8.35017262542882e-05, "loss": 0.7971, "step": 2943 }, { "epoch": 0.5663716814159292, "grad_norm": 0.192051315839833, "learning_rate": 8.344027484571075e-05, "loss": 0.8096, "step": 2944 }, { "epoch": 0.5665640631011928, "grad_norm": 0.19173491092260883, "learning_rate": 8.337882986618876e-05, "loss": 0.7572, "step": 2945 }, { "epoch": 0.5667564447864564, "grad_norm": 0.2039164722554385, "learning_rate": 8.331739133957728e-05, "loss": 0.7857, "step": 2946 }, { "epoch": 0.5669488264717198, "grad_norm": 0.20551813986840803, "learning_rate": 8.325595928972894e-05, "loss": 0.7826, "step": 2947 }, { "epoch": 0.5671412081569834, "grad_norm": 0.17109361276384064, "learning_rate": 8.319453374049366e-05, "loss": 0.7765, "step": 2948 }, { "epoch": 0.567333589842247, "grad_norm": 0.18783908596642568, "learning_rate": 8.313311471571901e-05, "loss": 0.7622, "step": 2949 }, { "epoch": 0.5675259715275106, "grad_norm": 0.17808959806393754, "learning_rate": 8.307170223925003e-05, "loss": 0.7683, "step": 2950 }, { "epoch": 0.5677183532127742, "grad_norm": 0.20204840280146905, "learning_rate": 8.301029633492905e-05, "loss": 0.8134, "step": 2951 }, { "epoch": 0.5679107348980377, "grad_norm": 0.18325090913865819, "learning_rate": 8.294889702659602e-05, "loss": 0.8236, "step": 2952 }, { "epoch": 0.5681031165833013, "grad_norm": 0.1931172653213828, "learning_rate": 8.288750433808827e-05, "loss": 0.8283, "step": 2953 }, { "epoch": 0.5682954982685648, "grad_norm": 0.18466864312838657, "learning_rate": 8.28261182932405e-05, "loss": 0.823, "step": 2954 }, { "epoch": 0.5684878799538284, "grad_norm": 0.18729675279511424, "learning_rate": 8.276473891588492e-05, "loss": 0.7701, "step": 2955 }, { "epoch": 0.5686802616390919, "grad_norm": 0.20982262758123876, "learning_rate": 8.270336622985116e-05, "loss": 0.8177, "step": 2956 }, { "epoch": 0.5688726433243555, "grad_norm": 0.1821043905433762, "learning_rate": 8.264200025896616e-05, "loss": 0.8156, "step": 2957 }, { "epoch": 0.5690650250096191, "grad_norm": 0.1921802184498147, "learning_rate": 8.258064102705427e-05, "loss": 0.8051, "step": 2958 }, { "epoch": 0.5692574066948827, "grad_norm": 0.18238149385753635, "learning_rate": 8.251928855793735e-05, "loss": 0.8174, "step": 2959 }, { "epoch": 0.5694497883801463, "grad_norm": 0.1893300793996338, "learning_rate": 8.245794287543447e-05, "loss": 0.7592, "step": 2960 }, { "epoch": 0.5696421700654097, "grad_norm": 0.18046126536575688, "learning_rate": 8.239660400336214e-05, "loss": 0.7737, "step": 2961 }, { "epoch": 0.5698345517506733, "grad_norm": 0.19941203869354682, "learning_rate": 8.233527196553427e-05, "loss": 0.8557, "step": 2962 }, { "epoch": 0.5700269334359369, "grad_norm": 0.18292040342653224, "learning_rate": 8.227394678576203e-05, "loss": 0.7787, "step": 2963 }, { "epoch": 0.5702193151212005, "grad_norm": 0.18463543664759693, "learning_rate": 8.221262848785395e-05, "loss": 0.7859, "step": 2964 }, { "epoch": 0.570411696806464, "grad_norm": 0.20136985754738065, "learning_rate": 8.215131709561598e-05, "loss": 0.8561, "step": 2965 }, { "epoch": 0.5706040784917276, "grad_norm": 0.19678040950789688, "learning_rate": 8.20900126328512e-05, "loss": 0.7792, "step": 2966 }, { "epoch": 0.5707964601769911, "grad_norm": 0.21150852190784122, "learning_rate": 8.202871512336024e-05, "loss": 0.8023, "step": 2967 }, { "epoch": 0.5709888418622547, "grad_norm": 0.18569949246732906, "learning_rate": 8.196742459094077e-05, "loss": 0.8325, "step": 2968 }, { "epoch": 0.5711812235475183, "grad_norm": 0.10758202816489938, "learning_rate": 8.190614105938796e-05, "loss": 0.6344, "step": 2969 }, { "epoch": 0.5713736052327818, "grad_norm": 0.20610071253840775, "learning_rate": 8.184486455249424e-05, "loss": 0.7721, "step": 2970 }, { "epoch": 0.5715659869180454, "grad_norm": 0.18737283857631065, "learning_rate": 8.178359509404917e-05, "loss": 0.8101, "step": 2971 }, { "epoch": 0.571758368603309, "grad_norm": 0.18161845000109683, "learning_rate": 8.172233270783966e-05, "loss": 0.7661, "step": 2972 }, { "epoch": 0.5719507502885726, "grad_norm": 0.18161404032699746, "learning_rate": 8.166107741764997e-05, "loss": 0.7817, "step": 2973 }, { "epoch": 0.572143131973836, "grad_norm": 0.19427084486405824, "learning_rate": 8.159982924726141e-05, "loss": 0.7929, "step": 2974 }, { "epoch": 0.5723355136590996, "grad_norm": 0.2084448236372711, "learning_rate": 8.153858822045266e-05, "loss": 0.8116, "step": 2975 }, { "epoch": 0.5725278953443632, "grad_norm": 0.19593688511170268, "learning_rate": 8.147735436099967e-05, "loss": 0.774, "step": 2976 }, { "epoch": 0.5727202770296268, "grad_norm": 0.19501996204478647, "learning_rate": 8.141612769267544e-05, "loss": 0.7886, "step": 2977 }, { "epoch": 0.5729126587148904, "grad_norm": 0.18955456717535427, "learning_rate": 8.135490823925027e-05, "loss": 0.7759, "step": 2978 }, { "epoch": 0.5731050404001539, "grad_norm": 0.19454065457719644, "learning_rate": 8.129369602449175e-05, "loss": 0.7994, "step": 2979 }, { "epoch": 0.5732974220854175, "grad_norm": 0.18279188225238616, "learning_rate": 8.123249107216446e-05, "loss": 0.7772, "step": 2980 }, { "epoch": 0.573489803770681, "grad_norm": 0.18454680136906151, "learning_rate": 8.117129340603032e-05, "loss": 0.7864, "step": 2981 }, { "epoch": 0.5736821854559446, "grad_norm": 0.20104785823611213, "learning_rate": 8.111010304984841e-05, "loss": 0.7601, "step": 2982 }, { "epoch": 0.5738745671412081, "grad_norm": 0.19927643661398461, "learning_rate": 8.104892002737487e-05, "loss": 0.716, "step": 2983 }, { "epoch": 0.5740669488264717, "grad_norm": 0.17398451389074335, "learning_rate": 8.098774436236308e-05, "loss": 0.7783, "step": 2984 }, { "epoch": 0.5742593305117353, "grad_norm": 0.19789520661201285, "learning_rate": 8.092657607856356e-05, "loss": 0.8135, "step": 2985 }, { "epoch": 0.5744517121969989, "grad_norm": 0.12053015468042055, "learning_rate": 8.086541519972388e-05, "loss": 0.6181, "step": 2986 }, { "epoch": 0.5746440938822625, "grad_norm": 0.18887100055464873, "learning_rate": 8.080426174958886e-05, "loss": 0.7997, "step": 2987 }, { "epoch": 0.5748364755675259, "grad_norm": 0.196087357061998, "learning_rate": 8.074311575190038e-05, "loss": 0.8159, "step": 2988 }, { "epoch": 0.5750288572527895, "grad_norm": 0.1822556285294721, "learning_rate": 8.068197723039737e-05, "loss": 0.7724, "step": 2989 }, { "epoch": 0.5752212389380531, "grad_norm": 0.10801982941098758, "learning_rate": 8.062084620881597e-05, "loss": 0.6503, "step": 2990 }, { "epoch": 0.5754136206233167, "grad_norm": 1.0287700448602377, "learning_rate": 8.055972271088932e-05, "loss": 0.6511, "step": 2991 }, { "epoch": 0.5756060023085803, "grad_norm": 0.18195574771100503, "learning_rate": 8.049860676034761e-05, "loss": 0.7492, "step": 2992 }, { "epoch": 0.5757983839938438, "grad_norm": 0.19332728954967263, "learning_rate": 8.043749838091827e-05, "loss": 0.7518, "step": 2993 }, { "epoch": 0.5759907656791073, "grad_norm": 0.19211337393300693, "learning_rate": 8.037639759632556e-05, "loss": 0.7873, "step": 2994 }, { "epoch": 0.5761831473643709, "grad_norm": 0.18916191925246062, "learning_rate": 8.031530443029099e-05, "loss": 0.8195, "step": 2995 }, { "epoch": 0.5763755290496345, "grad_norm": 0.17162804737327772, "learning_rate": 8.025421890653302e-05, "loss": 0.7847, "step": 2996 }, { "epoch": 0.576567910734898, "grad_norm": 0.19699277952640584, "learning_rate": 8.019314104876712e-05, "loss": 0.782, "step": 2997 }, { "epoch": 0.5767602924201616, "grad_norm": 0.19175128262492674, "learning_rate": 8.013207088070582e-05, "loss": 0.8095, "step": 2998 }, { "epoch": 0.5769526741054252, "grad_norm": 0.1746683153153437, "learning_rate": 8.007100842605872e-05, "loss": 0.6757, "step": 2999 }, { "epoch": 0.5771450557906888, "grad_norm": 0.21329775994968192, "learning_rate": 8.000995370853227e-05, "loss": 0.7514, "step": 3000 }, { "epoch": 0.5773374374759523, "grad_norm": 0.1913284721174967, "learning_rate": 7.994890675183008e-05, "loss": 0.7816, "step": 3001 }, { "epoch": 0.5775298191612158, "grad_norm": 0.18558099679436516, "learning_rate": 7.98878675796527e-05, "loss": 0.7233, "step": 3002 }, { "epoch": 0.5777222008464794, "grad_norm": 0.12430134167679127, "learning_rate": 7.98268362156976e-05, "loss": 0.6405, "step": 3003 }, { "epoch": 0.577914582531743, "grad_norm": 0.2092996443199445, "learning_rate": 7.976581268365923e-05, "loss": 0.8481, "step": 3004 }, { "epoch": 0.5781069642170066, "grad_norm": 0.19378888950690995, "learning_rate": 7.97047970072291e-05, "loss": 0.7854, "step": 3005 }, { "epoch": 0.5782993459022701, "grad_norm": 0.2005155066120069, "learning_rate": 7.964378921009552e-05, "loss": 0.8125, "step": 3006 }, { "epoch": 0.5784917275875336, "grad_norm": 0.2029540138209556, "learning_rate": 7.958278931594385e-05, "loss": 0.7722, "step": 3007 }, { "epoch": 0.5786841092727972, "grad_norm": 0.1878649207275347, "learning_rate": 7.952179734845641e-05, "loss": 0.8059, "step": 3008 }, { "epoch": 0.5788764909580608, "grad_norm": 0.17879791833771605, "learning_rate": 7.946081333131227e-05, "loss": 0.78, "step": 3009 }, { "epoch": 0.5790688726433244, "grad_norm": 0.18556944553430654, "learning_rate": 7.93998372881876e-05, "loss": 0.7924, "step": 3010 }, { "epoch": 0.5792612543285879, "grad_norm": 0.20604619282138809, "learning_rate": 7.93388692427554e-05, "loss": 0.8259, "step": 3011 }, { "epoch": 0.5794536360138515, "grad_norm": 0.18139521729471123, "learning_rate": 7.92779092186855e-05, "loss": 0.8025, "step": 3012 }, { "epoch": 0.5796460176991151, "grad_norm": 0.21838849808147598, "learning_rate": 7.921695723964472e-05, "loss": 0.8019, "step": 3013 }, { "epoch": 0.5798383993843786, "grad_norm": 0.17791047993918938, "learning_rate": 7.915601332929679e-05, "loss": 0.7954, "step": 3014 }, { "epoch": 0.5800307810696421, "grad_norm": 0.1873255252437285, "learning_rate": 7.909507751130209e-05, "loss": 0.7862, "step": 3015 }, { "epoch": 0.5802231627549057, "grad_norm": 0.1822713512040725, "learning_rate": 7.903414980931814e-05, "loss": 0.7893, "step": 3016 }, { "epoch": 0.5804155444401693, "grad_norm": 0.18590835786440102, "learning_rate": 7.897323024699907e-05, "loss": 0.8206, "step": 3017 }, { "epoch": 0.5806079261254329, "grad_norm": 0.19457824897679646, "learning_rate": 7.891231884799599e-05, "loss": 0.8212, "step": 3018 }, { "epoch": 0.5808003078106965, "grad_norm": 0.1836132962887353, "learning_rate": 7.885141563595684e-05, "loss": 0.8566, "step": 3019 }, { "epoch": 0.58099268949596, "grad_norm": 0.18821078641431666, "learning_rate": 7.879052063452625e-05, "loss": 0.796, "step": 3020 }, { "epoch": 0.5811850711812235, "grad_norm": 0.19340566427276523, "learning_rate": 7.872963386734583e-05, "loss": 0.8078, "step": 3021 }, { "epoch": 0.5813774528664871, "grad_norm": 0.21620603096224297, "learning_rate": 7.866875535805394e-05, "loss": 0.8573, "step": 3022 }, { "epoch": 0.5815698345517507, "grad_norm": 0.20321317224264593, "learning_rate": 7.860788513028565e-05, "loss": 0.7556, "step": 3023 }, { "epoch": 0.5817622162370142, "grad_norm": 0.18171753000155047, "learning_rate": 7.854702320767289e-05, "loss": 0.7404, "step": 3024 }, { "epoch": 0.5819545979222778, "grad_norm": 0.18316902431852516, "learning_rate": 7.848616961384442e-05, "loss": 0.739, "step": 3025 }, { "epoch": 0.5821469796075414, "grad_norm": 0.1919554582278712, "learning_rate": 7.842532437242559e-05, "loss": 0.7932, "step": 3026 }, { "epoch": 0.582339361292805, "grad_norm": 0.17784286068436791, "learning_rate": 7.836448750703868e-05, "loss": 0.7406, "step": 3027 }, { "epoch": 0.5825317429780685, "grad_norm": 0.1853090248153634, "learning_rate": 7.83036590413027e-05, "loss": 0.7569, "step": 3028 }, { "epoch": 0.582724124663332, "grad_norm": 0.19105214159801506, "learning_rate": 7.824283899883326e-05, "loss": 0.8238, "step": 3029 }, { "epoch": 0.5829165063485956, "grad_norm": 0.19970096349840083, "learning_rate": 7.818202740324286e-05, "loss": 0.7862, "step": 3030 }, { "epoch": 0.5831088880338592, "grad_norm": 0.10805793490964982, "learning_rate": 7.812122427814068e-05, "loss": 0.5764, "step": 3031 }, { "epoch": 0.5833012697191228, "grad_norm": 0.18643517784125224, "learning_rate": 7.806042964713247e-05, "loss": 0.7924, "step": 3032 }, { "epoch": 0.5834936514043862, "grad_norm": 0.18751030448823217, "learning_rate": 7.799964353382089e-05, "loss": 0.7998, "step": 3033 }, { "epoch": 0.5836860330896498, "grad_norm": 0.2022591490559963, "learning_rate": 7.793886596180522e-05, "loss": 0.8295, "step": 3034 }, { "epoch": 0.5838784147749134, "grad_norm": 0.11340414214385927, "learning_rate": 7.787809695468133e-05, "loss": 0.6675, "step": 3035 }, { "epoch": 0.584070796460177, "grad_norm": 0.19276147386670242, "learning_rate": 7.78173365360419e-05, "loss": 0.7787, "step": 3036 }, { "epoch": 0.5842631781454406, "grad_norm": 0.19153533308419807, "learning_rate": 7.775658472947623e-05, "loss": 0.8163, "step": 3037 }, { "epoch": 0.5844555598307041, "grad_norm": 0.21895902882960316, "learning_rate": 7.769584155857018e-05, "loss": 0.8394, "step": 3038 }, { "epoch": 0.5846479415159677, "grad_norm": 0.20080943512391367, "learning_rate": 7.763510704690644e-05, "loss": 0.7263, "step": 3039 }, { "epoch": 0.5848403232012312, "grad_norm": 0.19306474442193144, "learning_rate": 7.757438121806414e-05, "loss": 0.8431, "step": 3040 }, { "epoch": 0.5850327048864948, "grad_norm": 0.20791194949993635, "learning_rate": 7.75136640956192e-05, "loss": 0.8106, "step": 3041 }, { "epoch": 0.5852250865717584, "grad_norm": 0.20567873949170434, "learning_rate": 7.745295570314411e-05, "loss": 0.8208, "step": 3042 }, { "epoch": 0.5854174682570219, "grad_norm": 0.20258753844601904, "learning_rate": 7.739225606420793e-05, "loss": 0.7715, "step": 3043 }, { "epoch": 0.5856098499422855, "grad_norm": 0.19882396601152544, "learning_rate": 7.733156520237632e-05, "loss": 0.7983, "step": 3044 }, { "epoch": 0.5858022316275491, "grad_norm": 0.20312910628042202, "learning_rate": 7.727088314121165e-05, "loss": 0.8061, "step": 3045 }, { "epoch": 0.5859946133128127, "grad_norm": 0.17852682554359167, "learning_rate": 7.721020990427268e-05, "loss": 0.7973, "step": 3046 }, { "epoch": 0.5861869949980761, "grad_norm": 0.20198456276070684, "learning_rate": 7.71495455151149e-05, "loss": 0.7409, "step": 3047 }, { "epoch": 0.5863793766833397, "grad_norm": 0.17406355773595103, "learning_rate": 7.708888999729036e-05, "loss": 0.8012, "step": 3048 }, { "epoch": 0.5865717583686033, "grad_norm": 0.20570004488589486, "learning_rate": 7.702824337434755e-05, "loss": 0.7864, "step": 3049 }, { "epoch": 0.5867641400538669, "grad_norm": 0.2576206595618076, "learning_rate": 7.696760566983159e-05, "loss": 0.7746, "step": 3050 }, { "epoch": 0.5869565217391305, "grad_norm": 0.1908607123993983, "learning_rate": 7.690697690728416e-05, "loss": 0.7552, "step": 3051 }, { "epoch": 0.587148903424394, "grad_norm": 0.20030349342755108, "learning_rate": 7.68463571102434e-05, "loss": 0.8363, "step": 3052 }, { "epoch": 0.5873412851096576, "grad_norm": 0.19424328302953975, "learning_rate": 7.6785746302244e-05, "loss": 0.8237, "step": 3053 }, { "epoch": 0.5875336667949211, "grad_norm": 0.20477605254780937, "learning_rate": 7.67251445068172e-05, "loss": 0.7995, "step": 3054 }, { "epoch": 0.5877260484801847, "grad_norm": 0.18981520553615985, "learning_rate": 7.666455174749066e-05, "loss": 0.7941, "step": 3055 }, { "epoch": 0.5879184301654482, "grad_norm": 0.18177828143907793, "learning_rate": 7.66039680477886e-05, "loss": 0.7888, "step": 3056 }, { "epoch": 0.5881108118507118, "grad_norm": 0.10806869575507233, "learning_rate": 7.654339343123173e-05, "loss": 0.6359, "step": 3057 }, { "epoch": 0.5883031935359754, "grad_norm": 0.21692342823836358, "learning_rate": 7.648282792133711e-05, "loss": 0.7951, "step": 3058 }, { "epoch": 0.588495575221239, "grad_norm": 0.2050803102937488, "learning_rate": 7.642227154161841e-05, "loss": 0.7962, "step": 3059 }, { "epoch": 0.5886879569065026, "grad_norm": 0.20190548085062499, "learning_rate": 7.636172431558575e-05, "loss": 0.7941, "step": 3060 }, { "epoch": 0.588880338591766, "grad_norm": 0.19933795840983648, "learning_rate": 7.630118626674556e-05, "loss": 0.779, "step": 3061 }, { "epoch": 0.5890727202770296, "grad_norm": 0.17910775646248578, "learning_rate": 7.624065741860084e-05, "loss": 0.7545, "step": 3062 }, { "epoch": 0.5892651019622932, "grad_norm": 0.19352519224415893, "learning_rate": 7.618013779465101e-05, "loss": 0.7948, "step": 3063 }, { "epoch": 0.5894574836475568, "grad_norm": 0.21432966208799345, "learning_rate": 7.611962741839178e-05, "loss": 0.8475, "step": 3064 }, { "epoch": 0.5896498653328203, "grad_norm": 0.19483506917257737, "learning_rate": 7.605912631331546e-05, "loss": 0.8017, "step": 3065 }, { "epoch": 0.5898422470180839, "grad_norm": 0.20697383287591647, "learning_rate": 7.599863450291055e-05, "loss": 0.7705, "step": 3066 }, { "epoch": 0.5900346287033474, "grad_norm": 0.20103376724115032, "learning_rate": 7.593815201066215e-05, "loss": 0.8218, "step": 3067 }, { "epoch": 0.590227010388611, "grad_norm": 0.22369877057571558, "learning_rate": 7.587767886005163e-05, "loss": 0.8583, "step": 3068 }, { "epoch": 0.5904193920738746, "grad_norm": 0.19310023122136413, "learning_rate": 7.581721507455673e-05, "loss": 0.8168, "step": 3069 }, { "epoch": 0.5906117737591381, "grad_norm": 0.19083568110022214, "learning_rate": 7.575676067765154e-05, "loss": 0.7983, "step": 3070 }, { "epoch": 0.5908041554444017, "grad_norm": 0.19332222123610865, "learning_rate": 7.569631569280661e-05, "loss": 0.7807, "step": 3071 }, { "epoch": 0.5909965371296653, "grad_norm": 0.1833328015739037, "learning_rate": 7.56358801434887e-05, "loss": 0.7958, "step": 3072 }, { "epoch": 0.5911889188149289, "grad_norm": 0.1126808740740305, "learning_rate": 7.5575454053161e-05, "loss": 0.6132, "step": 3073 }, { "epoch": 0.5913813005001923, "grad_norm": 0.1748936375913734, "learning_rate": 7.551503744528304e-05, "loss": 0.7774, "step": 3074 }, { "epoch": 0.5915736821854559, "grad_norm": 0.17690335371139046, "learning_rate": 7.545463034331053e-05, "loss": 0.8289, "step": 3075 }, { "epoch": 0.5917660638707195, "grad_norm": 0.1923282428078678, "learning_rate": 7.539423277069568e-05, "loss": 0.8047, "step": 3076 }, { "epoch": 0.5919584455559831, "grad_norm": 0.22492661325716876, "learning_rate": 7.53338447508869e-05, "loss": 0.7561, "step": 3077 }, { "epoch": 0.5921508272412467, "grad_norm": 0.19905397257998356, "learning_rate": 7.52734663073288e-05, "loss": 0.7788, "step": 3078 }, { "epoch": 0.5923432089265102, "grad_norm": 0.19873465272209473, "learning_rate": 7.521309746346246e-05, "loss": 0.7705, "step": 3079 }, { "epoch": 0.5925355906117737, "grad_norm": 0.19054651016391203, "learning_rate": 7.515273824272514e-05, "loss": 0.7952, "step": 3080 }, { "epoch": 0.5927279722970373, "grad_norm": 0.18156994766361426, "learning_rate": 7.509238866855033e-05, "loss": 0.7864, "step": 3081 }, { "epoch": 0.5929203539823009, "grad_norm": 0.21621086752267696, "learning_rate": 7.503204876436785e-05, "loss": 0.7516, "step": 3082 }, { "epoch": 0.5931127356675644, "grad_norm": 0.2062931577422085, "learning_rate": 7.497171855360371e-05, "loss": 0.8041, "step": 3083 }, { "epoch": 0.593305117352828, "grad_norm": 0.18647051566278158, "learning_rate": 7.491139805968018e-05, "loss": 0.7921, "step": 3084 }, { "epoch": 0.5934974990380916, "grad_norm": 0.188904985881777, "learning_rate": 7.485108730601571e-05, "loss": 0.8071, "step": 3085 }, { "epoch": 0.5936898807233552, "grad_norm": 0.1860936127804006, "learning_rate": 7.479078631602512e-05, "loss": 0.7524, "step": 3086 }, { "epoch": 0.5938822624086187, "grad_norm": 0.19378981900254955, "learning_rate": 7.47304951131192e-05, "loss": 0.8823, "step": 3087 }, { "epoch": 0.5940746440938822, "grad_norm": 0.20704050155419865, "learning_rate": 7.467021372070516e-05, "loss": 0.7829, "step": 3088 }, { "epoch": 0.5942670257791458, "grad_norm": 0.1969277170282524, "learning_rate": 7.46099421621863e-05, "loss": 0.7921, "step": 3089 }, { "epoch": 0.5944594074644094, "grad_norm": 0.1849074956260702, "learning_rate": 7.454968046096204e-05, "loss": 0.7727, "step": 3090 }, { "epoch": 0.594651789149673, "grad_norm": 0.19758396683168747, "learning_rate": 7.448942864042818e-05, "loss": 0.749, "step": 3091 }, { "epoch": 0.5948441708349365, "grad_norm": 0.18040466292823157, "learning_rate": 7.442918672397642e-05, "loss": 0.8047, "step": 3092 }, { "epoch": 0.5950365525202, "grad_norm": 0.19613899149755218, "learning_rate": 7.43689547349948e-05, "loss": 0.7836, "step": 3093 }, { "epoch": 0.5952289342054636, "grad_norm": 0.1997181153405551, "learning_rate": 7.43087326968675e-05, "loss": 0.7808, "step": 3094 }, { "epoch": 0.5954213158907272, "grad_norm": 0.19137118989387852, "learning_rate": 7.424852063297469e-05, "loss": 0.792, "step": 3095 }, { "epoch": 0.5956136975759908, "grad_norm": 0.21756575560337216, "learning_rate": 7.418831856669286e-05, "loss": 0.8547, "step": 3096 }, { "epoch": 0.5958060792612543, "grad_norm": 0.20160074312387438, "learning_rate": 7.41281265213945e-05, "loss": 0.7653, "step": 3097 }, { "epoch": 0.5959984609465179, "grad_norm": 0.18865843105743088, "learning_rate": 7.406794452044816e-05, "loss": 0.7612, "step": 3098 }, { "epoch": 0.5961908426317815, "grad_norm": 0.10717917280883237, "learning_rate": 7.400777258721865e-05, "loss": 0.6054, "step": 3099 }, { "epoch": 0.596383224317045, "grad_norm": 0.19540386863738027, "learning_rate": 7.394761074506679e-05, "loss": 0.7978, "step": 3100 }, { "epoch": 0.5965756060023086, "grad_norm": 0.19293973084728933, "learning_rate": 7.38874590173494e-05, "loss": 0.7843, "step": 3101 }, { "epoch": 0.5967679876875721, "grad_norm": 0.19911063415561306, "learning_rate": 7.382731742741952e-05, "loss": 0.7359, "step": 3102 }, { "epoch": 0.5969603693728357, "grad_norm": 0.19370069439887222, "learning_rate": 7.376718599862621e-05, "loss": 0.8226, "step": 3103 }, { "epoch": 0.5971527510580993, "grad_norm": 0.19774943791304653, "learning_rate": 7.370706475431446e-05, "loss": 0.7599, "step": 3104 }, { "epoch": 0.5973451327433629, "grad_norm": 0.18589072086989636, "learning_rate": 7.364695371782546e-05, "loss": 0.7968, "step": 3105 }, { "epoch": 0.5975375144286263, "grad_norm": 0.19260351070678672, "learning_rate": 7.358685291249644e-05, "loss": 0.7693, "step": 3106 }, { "epoch": 0.5977298961138899, "grad_norm": 0.18283281134984733, "learning_rate": 7.352676236166051e-05, "loss": 0.7909, "step": 3107 }, { "epoch": 0.5979222777991535, "grad_norm": 0.20636898005842932, "learning_rate": 7.346668208864695e-05, "loss": 0.7631, "step": 3108 }, { "epoch": 0.5981146594844171, "grad_norm": 0.1977975189191447, "learning_rate": 7.340661211678099e-05, "loss": 0.7957, "step": 3109 }, { "epoch": 0.5983070411696807, "grad_norm": 0.19019359157708635, "learning_rate": 7.33465524693838e-05, "loss": 0.7359, "step": 3110 }, { "epoch": 0.5984994228549442, "grad_norm": 0.19545337948626684, "learning_rate": 7.328650316977264e-05, "loss": 0.7564, "step": 3111 }, { "epoch": 0.5986918045402078, "grad_norm": 0.1905904002938959, "learning_rate": 7.322646424126078e-05, "loss": 0.7863, "step": 3112 }, { "epoch": 0.5988841862254713, "grad_norm": 0.22565368178503914, "learning_rate": 7.31664357071573e-05, "loss": 0.7685, "step": 3113 }, { "epoch": 0.5990765679107349, "grad_norm": 0.19894031234119255, "learning_rate": 7.310641759076741e-05, "loss": 0.7527, "step": 3114 }, { "epoch": 0.5992689495959984, "grad_norm": 0.20880986895237538, "learning_rate": 7.304640991539215e-05, "loss": 0.8587, "step": 3115 }, { "epoch": 0.599461331281262, "grad_norm": 0.20081277640189887, "learning_rate": 7.298641270432862e-05, "loss": 0.7825, "step": 3116 }, { "epoch": 0.5996537129665256, "grad_norm": 0.17559052959621968, "learning_rate": 7.292642598086981e-05, "loss": 0.8134, "step": 3117 }, { "epoch": 0.5998460946517892, "grad_norm": 0.19539179553912897, "learning_rate": 7.286644976830457e-05, "loss": 0.7384, "step": 3118 }, { "epoch": 0.6000384763370528, "grad_norm": 0.2113922784886209, "learning_rate": 7.280648408991775e-05, "loss": 0.75, "step": 3119 }, { "epoch": 0.6002308580223162, "grad_norm": 0.2082998302445072, "learning_rate": 7.274652896899015e-05, "loss": 0.8892, "step": 3120 }, { "epoch": 0.6004232397075798, "grad_norm": 0.17647968878055603, "learning_rate": 7.268658442879833e-05, "loss": 0.8405, "step": 3121 }, { "epoch": 0.6006156213928434, "grad_norm": 0.19976365246539496, "learning_rate": 7.262665049261488e-05, "loss": 0.7359, "step": 3122 }, { "epoch": 0.600808003078107, "grad_norm": 0.18549278653940884, "learning_rate": 7.256672718370824e-05, "loss": 0.8394, "step": 3123 }, { "epoch": 0.6010003847633705, "grad_norm": 0.19060540908299528, "learning_rate": 7.25068145253426e-05, "loss": 0.7292, "step": 3124 }, { "epoch": 0.6011927664486341, "grad_norm": 0.18630773696212868, "learning_rate": 7.244691254077819e-05, "loss": 0.7842, "step": 3125 }, { "epoch": 0.6013851481338977, "grad_norm": 0.18514162153159394, "learning_rate": 7.238702125327106e-05, "loss": 0.7838, "step": 3126 }, { "epoch": 0.6015775298191612, "grad_norm": 0.192928291738047, "learning_rate": 7.232714068607295e-05, "loss": 0.7912, "step": 3127 }, { "epoch": 0.6017699115044248, "grad_norm": 0.1731585419376127, "learning_rate": 7.226727086243168e-05, "loss": 0.7553, "step": 3128 }, { "epoch": 0.6019622931896883, "grad_norm": 0.1914914380768392, "learning_rate": 7.220741180559073e-05, "loss": 0.8043, "step": 3129 }, { "epoch": 0.6021546748749519, "grad_norm": 0.2149285094341726, "learning_rate": 7.214756353878942e-05, "loss": 0.7758, "step": 3130 }, { "epoch": 0.6023470565602155, "grad_norm": 0.19746162959221722, "learning_rate": 7.208772608526293e-05, "loss": 0.7876, "step": 3131 }, { "epoch": 0.6025394382454791, "grad_norm": 0.1977988755681702, "learning_rate": 7.202789946824227e-05, "loss": 0.7309, "step": 3132 }, { "epoch": 0.6027318199307425, "grad_norm": 0.19074404848118617, "learning_rate": 7.196808371095408e-05, "loss": 0.7953, "step": 3133 }, { "epoch": 0.6029242016160061, "grad_norm": 0.10542677826568331, "learning_rate": 7.190827883662099e-05, "loss": 0.6331, "step": 3134 }, { "epoch": 0.6031165833012697, "grad_norm": 0.18669687232905086, "learning_rate": 7.184848486846128e-05, "loss": 0.7693, "step": 3135 }, { "epoch": 0.6033089649865333, "grad_norm": 0.1982058761127634, "learning_rate": 7.178870182968903e-05, "loss": 0.7553, "step": 3136 }, { "epoch": 0.6035013466717969, "grad_norm": 0.2043978448753592, "learning_rate": 7.172892974351406e-05, "loss": 0.7975, "step": 3137 }, { "epoch": 0.6036937283570604, "grad_norm": 0.1830818108238471, "learning_rate": 7.166916863314199e-05, "loss": 0.8196, "step": 3138 }, { "epoch": 0.603886110042324, "grad_norm": 0.23207989507511642, "learning_rate": 7.160941852177409e-05, "loss": 0.8053, "step": 3139 }, { "epoch": 0.6040784917275875, "grad_norm": 0.18411323854772438, "learning_rate": 7.154967943260748e-05, "loss": 0.8604, "step": 3140 }, { "epoch": 0.6042708734128511, "grad_norm": 0.18748635369708194, "learning_rate": 7.148995138883483e-05, "loss": 0.8132, "step": 3141 }, { "epoch": 0.6044632550981146, "grad_norm": 0.2063838310890917, "learning_rate": 7.143023441364472e-05, "loss": 0.7958, "step": 3142 }, { "epoch": 0.6046556367833782, "grad_norm": 0.20322383411791403, "learning_rate": 7.13705285302213e-05, "loss": 0.7783, "step": 3143 }, { "epoch": 0.6048480184686418, "grad_norm": 0.2006647736442709, "learning_rate": 7.131083376174441e-05, "loss": 0.8092, "step": 3144 }, { "epoch": 0.6050404001539054, "grad_norm": 0.1972242494088441, "learning_rate": 7.125115013138966e-05, "loss": 0.8195, "step": 3145 }, { "epoch": 0.605232781839169, "grad_norm": 0.18000239736192541, "learning_rate": 7.119147766232833e-05, "loss": 0.7514, "step": 3146 }, { "epoch": 0.6054251635244324, "grad_norm": 0.1856535488487341, "learning_rate": 7.113181637772721e-05, "loss": 0.8188, "step": 3147 }, { "epoch": 0.605617545209696, "grad_norm": 0.1787888813360119, "learning_rate": 7.107216630074896e-05, "loss": 0.7538, "step": 3148 }, { "epoch": 0.6058099268949596, "grad_norm": 0.10726854168968233, "learning_rate": 7.101252745455179e-05, "loss": 0.6035, "step": 3149 }, { "epoch": 0.6060023085802232, "grad_norm": 0.18214640768721202, "learning_rate": 7.09528998622895e-05, "loss": 0.8275, "step": 3150 }, { "epoch": 0.6061946902654868, "grad_norm": 0.18649420363299443, "learning_rate": 7.089328354711159e-05, "loss": 0.7537, "step": 3151 }, { "epoch": 0.6063870719507503, "grad_norm": 0.1921049931903939, "learning_rate": 7.083367853216323e-05, "loss": 0.835, "step": 3152 }, { "epoch": 0.6065794536360138, "grad_norm": 0.1927168778006588, "learning_rate": 7.077408484058505e-05, "loss": 0.7992, "step": 3153 }, { "epoch": 0.6067718353212774, "grad_norm": 0.19063782151946504, "learning_rate": 7.071450249551341e-05, "loss": 0.7477, "step": 3154 }, { "epoch": 0.606964217006541, "grad_norm": 0.18317161599849394, "learning_rate": 7.065493152008026e-05, "loss": 0.8302, "step": 3155 }, { "epoch": 0.6071565986918045, "grad_norm": 0.21235052342374247, "learning_rate": 7.059537193741306e-05, "loss": 0.817, "step": 3156 }, { "epoch": 0.6073489803770681, "grad_norm": 0.19992801046643585, "learning_rate": 7.053582377063489e-05, "loss": 0.794, "step": 3157 }, { "epoch": 0.6075413620623317, "grad_norm": 0.18560122387577815, "learning_rate": 7.047628704286446e-05, "loss": 0.7825, "step": 3158 }, { "epoch": 0.6077337437475953, "grad_norm": 0.20411087575223552, "learning_rate": 7.041676177721588e-05, "loss": 0.8772, "step": 3159 }, { "epoch": 0.6079261254328588, "grad_norm": 0.10919700404013309, "learning_rate": 7.035724799679898e-05, "loss": 0.6565, "step": 3160 }, { "epoch": 0.6081185071181223, "grad_norm": 0.1082850715742641, "learning_rate": 7.029774572471903e-05, "loss": 0.6355, "step": 3161 }, { "epoch": 0.6083108888033859, "grad_norm": 0.1930478735523793, "learning_rate": 7.023825498407689e-05, "loss": 0.7904, "step": 3162 }, { "epoch": 0.6085032704886495, "grad_norm": 0.19069658233112255, "learning_rate": 7.017877579796885e-05, "loss": 0.8012, "step": 3163 }, { "epoch": 0.6086956521739131, "grad_norm": 0.19887860710289515, "learning_rate": 7.011930818948688e-05, "loss": 0.7903, "step": 3164 }, { "epoch": 0.6088880338591766, "grad_norm": 0.21751527222467829, "learning_rate": 7.005985218171825e-05, "loss": 0.8269, "step": 3165 }, { "epoch": 0.6090804155444401, "grad_norm": 0.20989476426475068, "learning_rate": 7.000040779774591e-05, "loss": 0.805, "step": 3166 }, { "epoch": 0.6092727972297037, "grad_norm": 0.1963718201174866, "learning_rate": 6.994097506064812e-05, "loss": 0.8161, "step": 3167 }, { "epoch": 0.6094651789149673, "grad_norm": 0.1907482479401921, "learning_rate": 6.988155399349881e-05, "loss": 0.7737, "step": 3168 }, { "epoch": 0.6096575606002309, "grad_norm": 0.19306662557167298, "learning_rate": 6.982214461936727e-05, "loss": 0.8336, "step": 3169 }, { "epoch": 0.6098499422854944, "grad_norm": 0.1832193515646526, "learning_rate": 6.97627469613182e-05, "loss": 0.7823, "step": 3170 }, { "epoch": 0.610042323970758, "grad_norm": 0.22754453904342192, "learning_rate": 6.970336104241186e-05, "loss": 0.758, "step": 3171 }, { "epoch": 0.6102347056560216, "grad_norm": 0.20017400512021394, "learning_rate": 6.964398688570395e-05, "loss": 0.7745, "step": 3172 }, { "epoch": 0.6104270873412851, "grad_norm": 0.2105757667779702, "learning_rate": 6.958462451424547e-05, "loss": 0.8026, "step": 3173 }, { "epoch": 0.6106194690265486, "grad_norm": 0.20988892050515134, "learning_rate": 6.952527395108302e-05, "loss": 0.7795, "step": 3174 }, { "epoch": 0.6108118507118122, "grad_norm": 0.19995936035233286, "learning_rate": 6.946593521925851e-05, "loss": 0.7835, "step": 3175 }, { "epoch": 0.6110042323970758, "grad_norm": 0.11788298000817243, "learning_rate": 6.940660834180921e-05, "loss": 0.636, "step": 3176 }, { "epoch": 0.6111966140823394, "grad_norm": 0.20079141119481456, "learning_rate": 6.934729334176792e-05, "loss": 0.7744, "step": 3177 }, { "epoch": 0.611388995767603, "grad_norm": 0.10761484418267832, "learning_rate": 6.928799024216282e-05, "loss": 0.6158, "step": 3178 }, { "epoch": 0.6115813774528664, "grad_norm": 0.18966448308616768, "learning_rate": 6.92286990660173e-05, "loss": 0.783, "step": 3179 }, { "epoch": 0.61177375913813, "grad_norm": 0.1824544524610041, "learning_rate": 6.91694198363503e-05, "loss": 0.7678, "step": 3180 }, { "epoch": 0.6119661408233936, "grad_norm": 0.1997582464811362, "learning_rate": 6.911015257617605e-05, "loss": 0.8208, "step": 3181 }, { "epoch": 0.6121585225086572, "grad_norm": 0.1937773379541031, "learning_rate": 6.905089730850416e-05, "loss": 0.8144, "step": 3182 }, { "epoch": 0.6123509041939207, "grad_norm": 0.20038458634336076, "learning_rate": 6.89916540563395e-05, "loss": 0.7972, "step": 3183 }, { "epoch": 0.6125432858791843, "grad_norm": 0.20416839011040608, "learning_rate": 6.893242284268244e-05, "loss": 0.8011, "step": 3184 }, { "epoch": 0.6127356675644479, "grad_norm": 0.22292826750860545, "learning_rate": 6.887320369052848e-05, "loss": 0.8164, "step": 3185 }, { "epoch": 0.6129280492497114, "grad_norm": 0.1880015521470896, "learning_rate": 6.881399662286858e-05, "loss": 0.8268, "step": 3186 }, { "epoch": 0.613120430934975, "grad_norm": 0.18876156460771706, "learning_rate": 6.875480166268896e-05, "loss": 0.8007, "step": 3187 }, { "epoch": 0.6133128126202385, "grad_norm": 0.2162227453935147, "learning_rate": 6.869561883297116e-05, "loss": 0.7915, "step": 3188 }, { "epoch": 0.6135051943055021, "grad_norm": 0.192578363705759, "learning_rate": 6.863644815669197e-05, "loss": 0.7852, "step": 3189 }, { "epoch": 0.6136975759907657, "grad_norm": 0.20196523647053777, "learning_rate": 6.857728965682345e-05, "loss": 0.8233, "step": 3190 }, { "epoch": 0.6138899576760293, "grad_norm": 0.2080399153712548, "learning_rate": 6.851814335633297e-05, "loss": 0.7404, "step": 3191 }, { "epoch": 0.6140823393612928, "grad_norm": 0.18668561331464228, "learning_rate": 6.845900927818327e-05, "loss": 0.8003, "step": 3192 }, { "epoch": 0.6142747210465563, "grad_norm": 0.20602901715895014, "learning_rate": 6.839988744533211e-05, "loss": 0.7799, "step": 3193 }, { "epoch": 0.6144671027318199, "grad_norm": 0.1844691906379472, "learning_rate": 6.834077788073267e-05, "loss": 0.7831, "step": 3194 }, { "epoch": 0.6146594844170835, "grad_norm": 0.19157246746817427, "learning_rate": 6.828168060733337e-05, "loss": 0.7759, "step": 3195 }, { "epoch": 0.6148518661023471, "grad_norm": 0.19525206783527488, "learning_rate": 6.822259564807768e-05, "loss": 0.8172, "step": 3196 }, { "epoch": 0.6150442477876106, "grad_norm": 1.1996868890236017, "learning_rate": 6.816352302590449e-05, "loss": 0.8174, "step": 3197 }, { "epoch": 0.6152366294728742, "grad_norm": 0.19319701240028328, "learning_rate": 6.810446276374789e-05, "loss": 0.7454, "step": 3198 }, { "epoch": 0.6154290111581378, "grad_norm": 0.2064883899562962, "learning_rate": 6.804541488453697e-05, "loss": 0.8259, "step": 3199 }, { "epoch": 0.6156213928434013, "grad_norm": 0.17599594593770812, "learning_rate": 6.798637941119625e-05, "loss": 0.779, "step": 3200 }, { "epoch": 0.6158137745286649, "grad_norm": 0.18697017305650876, "learning_rate": 6.792735636664532e-05, "loss": 0.8046, "step": 3201 }, { "epoch": 0.6160061562139284, "grad_norm": 0.17337316119728996, "learning_rate": 6.786834577379892e-05, "loss": 0.7874, "step": 3202 }, { "epoch": 0.616198537899192, "grad_norm": 0.19374797496001914, "learning_rate": 6.780934765556701e-05, "loss": 0.7345, "step": 3203 }, { "epoch": 0.6163909195844556, "grad_norm": 0.18494838240006256, "learning_rate": 6.775036203485473e-05, "loss": 0.7738, "step": 3204 }, { "epoch": 0.6165833012697192, "grad_norm": 0.1988756100507543, "learning_rate": 6.769138893456225e-05, "loss": 0.7869, "step": 3205 }, { "epoch": 0.6167756829549826, "grad_norm": 0.19620866678851245, "learning_rate": 6.763242837758504e-05, "loss": 0.8093, "step": 3206 }, { "epoch": 0.6169680646402462, "grad_norm": 0.18908264328297786, "learning_rate": 6.757348038681357e-05, "loss": 0.8144, "step": 3207 }, { "epoch": 0.6171604463255098, "grad_norm": 0.19707963350423882, "learning_rate": 6.751454498513348e-05, "loss": 0.781, "step": 3208 }, { "epoch": 0.6173528280107734, "grad_norm": 0.17843310813699786, "learning_rate": 6.745562219542554e-05, "loss": 0.8391, "step": 3209 }, { "epoch": 0.617545209696037, "grad_norm": 0.19959936607475687, "learning_rate": 6.739671204056562e-05, "loss": 0.7288, "step": 3210 }, { "epoch": 0.6177375913813005, "grad_norm": 0.19153477131131336, "learning_rate": 6.733781454342462e-05, "loss": 0.7649, "step": 3211 }, { "epoch": 0.617929973066564, "grad_norm": 0.20679838857854604, "learning_rate": 6.72789297268686e-05, "loss": 0.7573, "step": 3212 }, { "epoch": 0.6181223547518276, "grad_norm": 0.11206828255798673, "learning_rate": 6.722005761375873e-05, "loss": 0.6507, "step": 3213 }, { "epoch": 0.6183147364370912, "grad_norm": 0.19326288475237902, "learning_rate": 6.71611982269511e-05, "loss": 0.7765, "step": 3214 }, { "epoch": 0.6185071181223547, "grad_norm": 0.19256597317256283, "learning_rate": 6.710235158929703e-05, "loss": 0.8273, "step": 3215 }, { "epoch": 0.6186994998076183, "grad_norm": 0.18023746598074125, "learning_rate": 6.704351772364274e-05, "loss": 0.7827, "step": 3216 }, { "epoch": 0.6188918814928819, "grad_norm": 0.1932163822969443, "learning_rate": 6.698469665282957e-05, "loss": 0.8462, "step": 3217 }, { "epoch": 0.6190842631781455, "grad_norm": 0.2099749968445328, "learning_rate": 6.692588839969397e-05, "loss": 0.7905, "step": 3218 }, { "epoch": 0.619276644863409, "grad_norm": 0.2136801113533711, "learning_rate": 6.686709298706722e-05, "loss": 0.8406, "step": 3219 }, { "epoch": 0.6194690265486725, "grad_norm": 0.2017371718867474, "learning_rate": 6.680831043777579e-05, "loss": 0.8082, "step": 3220 }, { "epoch": 0.6196614082339361, "grad_norm": 0.17708090463411114, "learning_rate": 6.674954077464107e-05, "loss": 0.8397, "step": 3221 }, { "epoch": 0.6198537899191997, "grad_norm": 0.18284718522842144, "learning_rate": 6.669078402047948e-05, "loss": 0.809, "step": 3222 }, { "epoch": 0.6200461716044633, "grad_norm": 0.20988916084119522, "learning_rate": 6.663204019810237e-05, "loss": 0.7932, "step": 3223 }, { "epoch": 0.6202385532897268, "grad_norm": 0.18275432676792794, "learning_rate": 6.657330933031619e-05, "loss": 0.7893, "step": 3224 }, { "epoch": 0.6204309349749904, "grad_norm": 0.20079906453617685, "learning_rate": 6.65145914399222e-05, "loss": 0.8597, "step": 3225 }, { "epoch": 0.6206233166602539, "grad_norm": 0.18370760025869193, "learning_rate": 6.645588654971678e-05, "loss": 0.7989, "step": 3226 }, { "epoch": 0.6208156983455175, "grad_norm": 0.18578509360953605, "learning_rate": 6.639719468249115e-05, "loss": 0.8361, "step": 3227 }, { "epoch": 0.6210080800307811, "grad_norm": 0.18734942104809124, "learning_rate": 6.633851586103153e-05, "loss": 0.7895, "step": 3228 }, { "epoch": 0.6212004617160446, "grad_norm": 0.1849056220364915, "learning_rate": 6.627985010811902e-05, "loss": 0.7768, "step": 3229 }, { "epoch": 0.6213928434013082, "grad_norm": 0.2060645746020671, "learning_rate": 6.622119744652978e-05, "loss": 0.7761, "step": 3230 }, { "epoch": 0.6215852250865718, "grad_norm": 0.18321279595345802, "learning_rate": 6.616255789903466e-05, "loss": 0.7633, "step": 3231 }, { "epoch": 0.6217776067718354, "grad_norm": 0.20890846832064858, "learning_rate": 6.610393148839964e-05, "loss": 0.8309, "step": 3232 }, { "epoch": 0.6219699884570988, "grad_norm": 0.18521229886433702, "learning_rate": 6.604531823738552e-05, "loss": 0.8651, "step": 3233 }, { "epoch": 0.6221623701423624, "grad_norm": 0.20565460038408534, "learning_rate": 6.598671816874794e-05, "loss": 0.8101, "step": 3234 }, { "epoch": 0.622354751827626, "grad_norm": 0.236023716581406, "learning_rate": 6.592813130523746e-05, "loss": 0.7689, "step": 3235 }, { "epoch": 0.6225471335128896, "grad_norm": 0.19359068198662688, "learning_rate": 6.586955766959957e-05, "loss": 0.82, "step": 3236 }, { "epoch": 0.6227395151981532, "grad_norm": 0.21309602137231842, "learning_rate": 6.581099728457451e-05, "loss": 0.7522, "step": 3237 }, { "epoch": 0.6229318968834167, "grad_norm": 0.1814521263835678, "learning_rate": 6.575245017289746e-05, "loss": 0.821, "step": 3238 }, { "epoch": 0.6231242785686802, "grad_norm": 0.21288446024877566, "learning_rate": 6.569391635729847e-05, "loss": 0.83, "step": 3239 }, { "epoch": 0.6233166602539438, "grad_norm": 0.23485871241769488, "learning_rate": 6.563539586050233e-05, "loss": 0.8082, "step": 3240 }, { "epoch": 0.6235090419392074, "grad_norm": 0.224472912994044, "learning_rate": 6.557688870522871e-05, "loss": 0.8635, "step": 3241 }, { "epoch": 0.6237014236244709, "grad_norm": 0.20782702948706863, "learning_rate": 6.551839491419213e-05, "loss": 0.774, "step": 3242 }, { "epoch": 0.6238938053097345, "grad_norm": 0.18523703423515442, "learning_rate": 6.545991451010185e-05, "loss": 0.7884, "step": 3243 }, { "epoch": 0.6240861869949981, "grad_norm": 0.20126822071716052, "learning_rate": 6.540144751566207e-05, "loss": 0.7742, "step": 3244 }, { "epoch": 0.6242785686802617, "grad_norm": 0.21307002101611397, "learning_rate": 6.534299395357155e-05, "loss": 0.7649, "step": 3245 }, { "epoch": 0.6244709503655252, "grad_norm": 0.19609156548293444, "learning_rate": 6.52845538465241e-05, "loss": 0.7298, "step": 3246 }, { "epoch": 0.6246633320507887, "grad_norm": 0.2002865994549744, "learning_rate": 6.522612721720814e-05, "loss": 0.8231, "step": 3247 }, { "epoch": 0.6248557137360523, "grad_norm": 0.19647922633118645, "learning_rate": 6.516771408830686e-05, "loss": 0.8031, "step": 3248 }, { "epoch": 0.6250480954213159, "grad_norm": 0.1817966228501918, "learning_rate": 6.51093144824983e-05, "loss": 0.781, "step": 3249 }, { "epoch": 0.6252404771065795, "grad_norm": 0.1876071109090203, "learning_rate": 6.505092842245519e-05, "loss": 0.7936, "step": 3250 }, { "epoch": 0.6254328587918431, "grad_norm": 0.19802255436453922, "learning_rate": 6.499255593084498e-05, "loss": 0.8311, "step": 3251 }, { "epoch": 0.6256252404771065, "grad_norm": 0.1863310767553361, "learning_rate": 6.49341970303299e-05, "loss": 0.8088, "step": 3252 }, { "epoch": 0.6258176221623701, "grad_norm": 0.2075987673107603, "learning_rate": 6.48758517435669e-05, "loss": 0.8128, "step": 3253 }, { "epoch": 0.6260100038476337, "grad_norm": 0.22675297216408144, "learning_rate": 6.481752009320761e-05, "loss": 0.8221, "step": 3254 }, { "epoch": 0.6262023855328973, "grad_norm": 0.19636133557898855, "learning_rate": 6.475920210189836e-05, "loss": 0.7591, "step": 3255 }, { "epoch": 0.6263947672181608, "grad_norm": 0.20098071911323087, "learning_rate": 6.470089779228028e-05, "loss": 0.8484, "step": 3256 }, { "epoch": 0.6265871489034244, "grad_norm": 0.18643699810344666, "learning_rate": 6.464260718698902e-05, "loss": 0.8061, "step": 3257 }, { "epoch": 0.626779530588688, "grad_norm": 0.19470841070831194, "learning_rate": 6.458433030865502e-05, "loss": 0.7812, "step": 3258 }, { "epoch": 0.6269719122739515, "grad_norm": 0.2100234057441616, "learning_rate": 6.452606717990346e-05, "loss": 0.763, "step": 3259 }, { "epoch": 0.6271642939592151, "grad_norm": 0.18015150172681976, "learning_rate": 6.4467817823354e-05, "loss": 0.8431, "step": 3260 }, { "epoch": 0.6273566756444786, "grad_norm": 0.18479108576928857, "learning_rate": 6.440958226162104e-05, "loss": 0.7822, "step": 3261 }, { "epoch": 0.6275490573297422, "grad_norm": 0.19565668589904642, "learning_rate": 6.43513605173137e-05, "loss": 0.7716, "step": 3262 }, { "epoch": 0.6277414390150058, "grad_norm": 0.18018582851169435, "learning_rate": 6.429315261303559e-05, "loss": 0.7675, "step": 3263 }, { "epoch": 0.6279338207002694, "grad_norm": 0.18724412837701937, "learning_rate": 6.423495857138509e-05, "loss": 0.7832, "step": 3264 }, { "epoch": 0.6281262023855329, "grad_norm": 0.17993302965117783, "learning_rate": 6.417677841495507e-05, "loss": 0.7818, "step": 3265 }, { "epoch": 0.6283185840707964, "grad_norm": 0.18489311029740452, "learning_rate": 6.411861216633313e-05, "loss": 0.7281, "step": 3266 }, { "epoch": 0.62851096575606, "grad_norm": 0.1858547719616412, "learning_rate": 6.406045984810135e-05, "loss": 0.7952, "step": 3267 }, { "epoch": 0.6287033474413236, "grad_norm": 0.19471632762974214, "learning_rate": 6.400232148283651e-05, "loss": 0.764, "step": 3268 }, { "epoch": 0.6288957291265872, "grad_norm": 0.20683045886111867, "learning_rate": 6.394419709310988e-05, "loss": 0.7841, "step": 3269 }, { "epoch": 0.6290881108118507, "grad_norm": 0.2021094992731049, "learning_rate": 6.388608670148741e-05, "loss": 0.7312, "step": 3270 }, { "epoch": 0.6292804924971143, "grad_norm": 0.20477916055760012, "learning_rate": 6.38279903305295e-05, "loss": 0.8545, "step": 3271 }, { "epoch": 0.6294728741823779, "grad_norm": 0.18629405931885282, "learning_rate": 6.376990800279119e-05, "loss": 0.7668, "step": 3272 }, { "epoch": 0.6296652558676414, "grad_norm": 0.20820360343645125, "learning_rate": 6.371183974082203e-05, "loss": 0.8141, "step": 3273 }, { "epoch": 0.6298576375529049, "grad_norm": 0.19819523700402333, "learning_rate": 6.365378556716612e-05, "loss": 0.8244, "step": 3274 }, { "epoch": 0.6300500192381685, "grad_norm": 0.19528895721424772, "learning_rate": 6.359574550436208e-05, "loss": 0.7561, "step": 3275 }, { "epoch": 0.6302424009234321, "grad_norm": 0.18478913592732604, "learning_rate": 6.35377195749431e-05, "loss": 0.8063, "step": 3276 }, { "epoch": 0.6304347826086957, "grad_norm": 0.19394790125338418, "learning_rate": 6.347970780143678e-05, "loss": 0.7837, "step": 3277 }, { "epoch": 0.6306271642939593, "grad_norm": 0.11258120814658928, "learning_rate": 6.342171020636533e-05, "loss": 0.6317, "step": 3278 }, { "epoch": 0.6308195459792227, "grad_norm": 0.20943856787385254, "learning_rate": 6.336372681224542e-05, "loss": 0.7745, "step": 3279 }, { "epoch": 0.6310119276644863, "grad_norm": 0.18927689406566997, "learning_rate": 6.33057576415882e-05, "loss": 0.8265, "step": 3280 }, { "epoch": 0.6312043093497499, "grad_norm": 0.2131343002700523, "learning_rate": 6.324780271689923e-05, "loss": 0.8185, "step": 3281 }, { "epoch": 0.6313966910350135, "grad_norm": 0.19810551834295778, "learning_rate": 6.318986206067872e-05, "loss": 0.8518, "step": 3282 }, { "epoch": 0.631589072720277, "grad_norm": 0.11786563607489417, "learning_rate": 6.313193569542113e-05, "loss": 0.6693, "step": 3283 }, { "epoch": 0.6317814544055406, "grad_norm": 0.20729992888786772, "learning_rate": 6.30740236436155e-05, "loss": 0.7974, "step": 3284 }, { "epoch": 0.6319738360908042, "grad_norm": 0.19816191145379758, "learning_rate": 6.301612592774533e-05, "loss": 0.7804, "step": 3285 }, { "epoch": 0.6321662177760677, "grad_norm": 0.19788397166774763, "learning_rate": 6.295824257028844e-05, "loss": 0.7709, "step": 3286 }, { "epoch": 0.6323585994613313, "grad_norm": 0.18482248893300324, "learning_rate": 6.290037359371717e-05, "loss": 0.8362, "step": 3287 }, { "epoch": 0.6325509811465948, "grad_norm": 0.17214810755703633, "learning_rate": 6.284251902049827e-05, "loss": 0.8284, "step": 3288 }, { "epoch": 0.6327433628318584, "grad_norm": 0.1840493980562653, "learning_rate": 6.278467887309283e-05, "loss": 0.7806, "step": 3289 }, { "epoch": 0.632935744517122, "grad_norm": 0.22285156286506133, "learning_rate": 6.272685317395643e-05, "loss": 0.8004, "step": 3290 }, { "epoch": 0.6331281262023856, "grad_norm": 0.18789108119619571, "learning_rate": 6.266904194553895e-05, "loss": 0.7988, "step": 3291 }, { "epoch": 0.633320507887649, "grad_norm": 0.19741989881404923, "learning_rate": 6.261124521028476e-05, "loss": 0.8153, "step": 3292 }, { "epoch": 0.6335128895729126, "grad_norm": 0.1775472589556908, "learning_rate": 6.255346299063252e-05, "loss": 0.7785, "step": 3293 }, { "epoch": 0.6337052712581762, "grad_norm": 0.19749947261308484, "learning_rate": 6.249569530901524e-05, "loss": 0.7998, "step": 3294 }, { "epoch": 0.6338976529434398, "grad_norm": 0.2195424502924516, "learning_rate": 6.243794218786035e-05, "loss": 0.7773, "step": 3295 }, { "epoch": 0.6340900346287034, "grad_norm": 0.20354186742314856, "learning_rate": 6.238020364958965e-05, "loss": 0.8473, "step": 3296 }, { "epoch": 0.6342824163139669, "grad_norm": 0.19789429332571362, "learning_rate": 6.232247971661912e-05, "loss": 0.7921, "step": 3297 }, { "epoch": 0.6344747979992305, "grad_norm": 0.19907420650825208, "learning_rate": 6.226477041135926e-05, "loss": 0.7948, "step": 3298 }, { "epoch": 0.634667179684494, "grad_norm": 0.2051213274782692, "learning_rate": 6.22070757562148e-05, "loss": 0.7907, "step": 3299 }, { "epoch": 0.6348595613697576, "grad_norm": 0.1978796929425776, "learning_rate": 6.214939577358478e-05, "loss": 0.7721, "step": 3300 }, { "epoch": 0.6350519430550212, "grad_norm": 0.19784924095104545, "learning_rate": 6.209173048586253e-05, "loss": 0.8191, "step": 3301 }, { "epoch": 0.6352443247402847, "grad_norm": 0.186282892946832, "learning_rate": 6.203407991543577e-05, "loss": 0.8015, "step": 3302 }, { "epoch": 0.6354367064255483, "grad_norm": 0.19836587627650032, "learning_rate": 6.197644408468635e-05, "loss": 0.8568, "step": 3303 }, { "epoch": 0.6356290881108119, "grad_norm": 0.18658657561367045, "learning_rate": 6.191882301599052e-05, "loss": 0.837, "step": 3304 }, { "epoch": 0.6358214697960755, "grad_norm": 0.19413566599023804, "learning_rate": 6.186121673171882e-05, "loss": 0.7745, "step": 3305 }, { "epoch": 0.6360138514813389, "grad_norm": 0.18450670456910767, "learning_rate": 6.180362525423591e-05, "loss": 0.8371, "step": 3306 }, { "epoch": 0.6362062331666025, "grad_norm": 0.2118340967934365, "learning_rate": 6.174604860590081e-05, "loss": 0.8163, "step": 3307 }, { "epoch": 0.6363986148518661, "grad_norm": 0.1974581660664363, "learning_rate": 6.168848680906678e-05, "loss": 0.7663, "step": 3308 }, { "epoch": 0.6365909965371297, "grad_norm": 0.2082364604195254, "learning_rate": 6.163093988608127e-05, "loss": 0.7793, "step": 3309 }, { "epoch": 0.6367833782223933, "grad_norm": 0.1925474134782682, "learning_rate": 6.157340785928595e-05, "loss": 0.7663, "step": 3310 }, { "epoch": 0.6369757599076568, "grad_norm": 0.18331268259535508, "learning_rate": 6.151589075101682e-05, "loss": 0.789, "step": 3311 }, { "epoch": 0.6371681415929203, "grad_norm": 0.2154177140999453, "learning_rate": 6.145838858360391e-05, "loss": 0.7762, "step": 3312 }, { "epoch": 0.6373605232781839, "grad_norm": 0.1990488400002381, "learning_rate": 6.140090137937158e-05, "loss": 0.7461, "step": 3313 }, { "epoch": 0.6375529049634475, "grad_norm": 0.18175026888892343, "learning_rate": 6.134342916063838e-05, "loss": 0.8191, "step": 3314 }, { "epoch": 0.637745286648711, "grad_norm": 0.2079369482663373, "learning_rate": 6.12859719497169e-05, "loss": 0.7982, "step": 3315 }, { "epoch": 0.6379376683339746, "grad_norm": 0.19492665844477855, "learning_rate": 6.122852976891414e-05, "loss": 0.796, "step": 3316 }, { "epoch": 0.6381300500192382, "grad_norm": 0.20420054051794173, "learning_rate": 6.1171102640531e-05, "loss": 0.7573, "step": 3317 }, { "epoch": 0.6383224317045018, "grad_norm": 0.21835438801179874, "learning_rate": 6.111369058686276e-05, "loss": 0.7481, "step": 3318 }, { "epoch": 0.6385148133897653, "grad_norm": 0.20582796241920642, "learning_rate": 6.105629363019875e-05, "loss": 0.761, "step": 3319 }, { "epoch": 0.6387071950750288, "grad_norm": 0.181694137685693, "learning_rate": 6.099891179282241e-05, "loss": 0.6866, "step": 3320 }, { "epoch": 0.6388995767602924, "grad_norm": 0.11434498729755332, "learning_rate": 6.094154509701133e-05, "loss": 0.6144, "step": 3321 }, { "epoch": 0.639091958445556, "grad_norm": 0.20774966177396806, "learning_rate": 6.088419356503732e-05, "loss": 0.7365, "step": 3322 }, { "epoch": 0.6392843401308196, "grad_norm": 0.18476018509146422, "learning_rate": 6.0826857219166124e-05, "loss": 0.8065, "step": 3323 }, { "epoch": 0.6394767218160831, "grad_norm": 0.20316939223411282, "learning_rate": 6.0769536081657717e-05, "loss": 0.8834, "step": 3324 }, { "epoch": 0.6396691035013466, "grad_norm": 0.20825994877526116, "learning_rate": 6.0712230174766194e-05, "loss": 0.8066, "step": 3325 }, { "epoch": 0.6398614851866102, "grad_norm": 0.21691906755629903, "learning_rate": 6.065493952073961e-05, "loss": 0.8165, "step": 3326 }, { "epoch": 0.6400538668718738, "grad_norm": 0.19862756474474036, "learning_rate": 6.0597664141820176e-05, "loss": 0.8243, "step": 3327 }, { "epoch": 0.6402462485571374, "grad_norm": 0.19713930043861616, "learning_rate": 6.0540404060244214e-05, "loss": 0.7388, "step": 3328 }, { "epoch": 0.6404386302424009, "grad_norm": 0.21149066330131724, "learning_rate": 6.0483159298242e-05, "loss": 0.8311, "step": 3329 }, { "epoch": 0.6406310119276645, "grad_norm": 0.19855681055324248, "learning_rate": 6.042592987803796e-05, "loss": 0.7572, "step": 3330 }, { "epoch": 0.6408233936129281, "grad_norm": 0.199465987752368, "learning_rate": 6.036871582185054e-05, "loss": 0.8172, "step": 3331 }, { "epoch": 0.6410157752981916, "grad_norm": 0.20081245256468308, "learning_rate": 6.031151715189217e-05, "loss": 0.7806, "step": 3332 }, { "epoch": 0.6412081569834551, "grad_norm": 0.19986509839591327, "learning_rate": 6.025433389036934e-05, "loss": 0.7994, "step": 3333 }, { "epoch": 0.6414005386687187, "grad_norm": 0.19584463393276214, "learning_rate": 6.019716605948261e-05, "loss": 0.8157, "step": 3334 }, { "epoch": 0.6415929203539823, "grad_norm": 0.2178863188176752, "learning_rate": 6.014001368142643e-05, "loss": 0.8475, "step": 3335 }, { "epoch": 0.6417853020392459, "grad_norm": 0.22572463608177845, "learning_rate": 6.0082876778389365e-05, "loss": 0.7198, "step": 3336 }, { "epoch": 0.6419776837245095, "grad_norm": 0.19858442846107263, "learning_rate": 6.002575537255395e-05, "loss": 0.8441, "step": 3337 }, { "epoch": 0.642170065409773, "grad_norm": 0.19184178737199617, "learning_rate": 5.996864948609662e-05, "loss": 0.8284, "step": 3338 }, { "epoch": 0.6423624470950365, "grad_norm": 0.1170277829111131, "learning_rate": 5.991155914118792e-05, "loss": 0.671, "step": 3339 }, { "epoch": 0.6425548287803001, "grad_norm": 0.18803742615075608, "learning_rate": 5.985448435999224e-05, "loss": 0.8445, "step": 3340 }, { "epoch": 0.6427472104655637, "grad_norm": 0.20310523578821427, "learning_rate": 5.979742516466793e-05, "loss": 0.8109, "step": 3341 }, { "epoch": 0.6429395921508272, "grad_norm": 0.2196977377288048, "learning_rate": 5.9740381577367455e-05, "loss": 0.8063, "step": 3342 }, { "epoch": 0.6431319738360908, "grad_norm": 0.1954153388914609, "learning_rate": 5.968335362023697e-05, "loss": 0.7806, "step": 3343 }, { "epoch": 0.6433243555213544, "grad_norm": 0.18484697201126832, "learning_rate": 5.9626341315416754e-05, "loss": 0.8349, "step": 3344 }, { "epoch": 0.643516737206618, "grad_norm": 0.22572119592248482, "learning_rate": 5.9569344685041006e-05, "loss": 0.7788, "step": 3345 }, { "epoch": 0.6437091188918815, "grad_norm": 0.21436509839203657, "learning_rate": 5.9512363751237677e-05, "loss": 0.8023, "step": 3346 }, { "epoch": 0.643901500577145, "grad_norm": 0.21228692984714667, "learning_rate": 5.945539853612876e-05, "loss": 0.803, "step": 3347 }, { "epoch": 0.6440938822624086, "grad_norm": 0.20673482763149081, "learning_rate": 5.9398449061830164e-05, "loss": 0.8538, "step": 3348 }, { "epoch": 0.6442862639476722, "grad_norm": 0.18465557247551997, "learning_rate": 5.934151535045156e-05, "loss": 0.8096, "step": 3349 }, { "epoch": 0.6444786456329358, "grad_norm": 0.195423581298625, "learning_rate": 5.92845974240966e-05, "loss": 0.7506, "step": 3350 }, { "epoch": 0.6446710273181993, "grad_norm": 0.19530996626495448, "learning_rate": 5.9227695304862826e-05, "loss": 0.806, "step": 3351 }, { "epoch": 0.6448634090034628, "grad_norm": 0.18556530267936552, "learning_rate": 5.917080901484155e-05, "loss": 0.759, "step": 3352 }, { "epoch": 0.6450557906887264, "grad_norm": 0.19105057054130348, "learning_rate": 5.9113938576118e-05, "loss": 0.7546, "step": 3353 }, { "epoch": 0.64524817237399, "grad_norm": 0.1919297642837818, "learning_rate": 5.905708401077128e-05, "loss": 0.8319, "step": 3354 }, { "epoch": 0.6454405540592536, "grad_norm": 0.1807103285153046, "learning_rate": 5.900024534087421e-05, "loss": 0.7906, "step": 3355 }, { "epoch": 0.6456329357445171, "grad_norm": 0.19684663588074539, "learning_rate": 5.8943422588493544e-05, "loss": 0.7982, "step": 3356 }, { "epoch": 0.6458253174297807, "grad_norm": 0.19356300743923188, "learning_rate": 5.8886615775689904e-05, "loss": 0.7875, "step": 3357 }, { "epoch": 0.6460176991150443, "grad_norm": 0.19107569377444997, "learning_rate": 5.882982492451757e-05, "loss": 0.8155, "step": 3358 }, { "epoch": 0.6462100808003078, "grad_norm": 0.1956305671381613, "learning_rate": 5.8773050057024716e-05, "loss": 0.7716, "step": 3359 }, { "epoch": 0.6464024624855714, "grad_norm": 0.20374492397864716, "learning_rate": 5.871629119525335e-05, "loss": 0.8405, "step": 3360 }, { "epoch": 0.6465948441708349, "grad_norm": 0.17593814615696668, "learning_rate": 5.865954836123915e-05, "loss": 0.7617, "step": 3361 }, { "epoch": 0.6467872258560985, "grad_norm": 0.2176198627061489, "learning_rate": 5.860282157701167e-05, "loss": 0.7736, "step": 3362 }, { "epoch": 0.6469796075413621, "grad_norm": 0.21355004087951757, "learning_rate": 5.854611086459423e-05, "loss": 0.7923, "step": 3363 }, { "epoch": 0.6471719892266257, "grad_norm": 0.20675488010195966, "learning_rate": 5.848941624600381e-05, "loss": 0.7789, "step": 3364 }, { "epoch": 0.6473643709118891, "grad_norm": 0.23564064874826773, "learning_rate": 5.843273774325131e-05, "loss": 0.808, "step": 3365 }, { "epoch": 0.6475567525971527, "grad_norm": 0.21256423049237275, "learning_rate": 5.837607537834119e-05, "loss": 0.8011, "step": 3366 }, { "epoch": 0.6477491342824163, "grad_norm": 0.2086904555244728, "learning_rate": 5.8319429173271725e-05, "loss": 0.7996, "step": 3367 }, { "epoch": 0.6479415159676799, "grad_norm": 0.1715932407361039, "learning_rate": 5.826279915003503e-05, "loss": 0.7862, "step": 3368 }, { "epoch": 0.6481338976529435, "grad_norm": 0.18264612946188508, "learning_rate": 5.820618533061672e-05, "loss": 0.8149, "step": 3369 }, { "epoch": 0.648326279338207, "grad_norm": 0.20218575183425497, "learning_rate": 5.8149587736996246e-05, "loss": 0.8133, "step": 3370 }, { "epoch": 0.6485186610234706, "grad_norm": 0.19284454526084843, "learning_rate": 5.809300639114683e-05, "loss": 0.7672, "step": 3371 }, { "epoch": 0.6487110427087341, "grad_norm": 0.22276829042561144, "learning_rate": 5.803644131503516e-05, "loss": 0.776, "step": 3372 }, { "epoch": 0.6489034243939977, "grad_norm": 0.18383627272265468, "learning_rate": 5.797989253062186e-05, "loss": 0.8205, "step": 3373 }, { "epoch": 0.6490958060792612, "grad_norm": 0.1960134432981417, "learning_rate": 5.792336005986105e-05, "loss": 0.8387, "step": 3374 }, { "epoch": 0.6492881877645248, "grad_norm": 0.19912159945449104, "learning_rate": 5.7866843924700634e-05, "loss": 0.737, "step": 3375 }, { "epoch": 0.6494805694497884, "grad_norm": 0.21542385753821736, "learning_rate": 5.7810344147082086e-05, "loss": 0.7925, "step": 3376 }, { "epoch": 0.649672951135052, "grad_norm": 0.19611668662615944, "learning_rate": 5.775386074894058e-05, "loss": 0.7855, "step": 3377 }, { "epoch": 0.6498653328203156, "grad_norm": 0.19277673009292842, "learning_rate": 5.7697393752204887e-05, "loss": 0.7941, "step": 3378 }, { "epoch": 0.650057714505579, "grad_norm": 0.22269390090607188, "learning_rate": 5.7640943178797444e-05, "loss": 0.7937, "step": 3379 }, { "epoch": 0.6502500961908426, "grad_norm": 0.21945733807079754, "learning_rate": 5.758450905063439e-05, "loss": 0.7457, "step": 3380 }, { "epoch": 0.6504424778761062, "grad_norm": 0.19210045892867336, "learning_rate": 5.7528091389625247e-05, "loss": 0.7789, "step": 3381 }, { "epoch": 0.6506348595613698, "grad_norm": 0.22087450135829076, "learning_rate": 5.747169021767343e-05, "loss": 0.7442, "step": 3382 }, { "epoch": 0.6508272412466333, "grad_norm": 0.1767417644695515, "learning_rate": 5.7415305556675805e-05, "loss": 0.8099, "step": 3383 }, { "epoch": 0.6510196229318969, "grad_norm": 0.1828707695017436, "learning_rate": 5.7358937428522694e-05, "loss": 0.7682, "step": 3384 }, { "epoch": 0.6512120046171604, "grad_norm": 0.20653134308063714, "learning_rate": 5.7302585855098315e-05, "loss": 0.7451, "step": 3385 }, { "epoch": 0.651404386302424, "grad_norm": 0.19012508630013017, "learning_rate": 5.7246250858280214e-05, "loss": 0.7702, "step": 3386 }, { "epoch": 0.6515967679876876, "grad_norm": 0.19748203340939793, "learning_rate": 5.7189932459939575e-05, "loss": 0.8272, "step": 3387 }, { "epoch": 0.6517891496729511, "grad_norm": 0.20016417850979917, "learning_rate": 5.713363068194115e-05, "loss": 0.7835, "step": 3388 }, { "epoch": 0.6519815313582147, "grad_norm": 0.20033403665942293, "learning_rate": 5.7077345546143235e-05, "loss": 0.7791, "step": 3389 }, { "epoch": 0.6521739130434783, "grad_norm": 0.18901154524047897, "learning_rate": 5.7021077074397654e-05, "loss": 0.7871, "step": 3390 }, { "epoch": 0.6523662947287419, "grad_norm": 0.18732415140034006, "learning_rate": 5.696482528854974e-05, "loss": 0.8229, "step": 3391 }, { "epoch": 0.6525586764140053, "grad_norm": 0.20271088742425517, "learning_rate": 5.6908590210438416e-05, "loss": 0.73, "step": 3392 }, { "epoch": 0.6527510580992689, "grad_norm": 0.1643231939125391, "learning_rate": 5.6852371861896e-05, "loss": 0.7559, "step": 3393 }, { "epoch": 0.6529434397845325, "grad_norm": 0.21335394246072414, "learning_rate": 5.6796170264748525e-05, "loss": 0.7596, "step": 3394 }, { "epoch": 0.6531358214697961, "grad_norm": 0.19605770132522038, "learning_rate": 5.6739985440815266e-05, "loss": 0.8012, "step": 3395 }, { "epoch": 0.6533282031550597, "grad_norm": 0.21988443864961188, "learning_rate": 5.6683817411909114e-05, "loss": 0.8069, "step": 3396 }, { "epoch": 0.6535205848403232, "grad_norm": 0.10870986010638514, "learning_rate": 5.6627666199836526e-05, "loss": 0.6268, "step": 3397 }, { "epoch": 0.6537129665255867, "grad_norm": 0.2044563745601063, "learning_rate": 5.657153182639721e-05, "loss": 0.7595, "step": 3398 }, { "epoch": 0.6539053482108503, "grad_norm": 0.2024217838977603, "learning_rate": 5.651541431338454e-05, "loss": 0.7835, "step": 3399 }, { "epoch": 0.6540977298961139, "grad_norm": 0.19072757313746974, "learning_rate": 5.645931368258527e-05, "loss": 0.7809, "step": 3400 }, { "epoch": 0.6542901115813774, "grad_norm": 0.2233008902568654, "learning_rate": 5.640322995577958e-05, "loss": 0.7322, "step": 3401 }, { "epoch": 0.654482493266641, "grad_norm": 0.17743738016932997, "learning_rate": 5.634716315474109e-05, "loss": 0.838, "step": 3402 }, { "epoch": 0.6546748749519046, "grad_norm": 0.18856519530010638, "learning_rate": 5.629111330123689e-05, "loss": 0.759, "step": 3403 }, { "epoch": 0.6548672566371682, "grad_norm": 0.19351151794663224, "learning_rate": 5.623508041702743e-05, "loss": 0.7507, "step": 3404 }, { "epoch": 0.6550596383224317, "grad_norm": 0.19603218314928236, "learning_rate": 5.617906452386659e-05, "loss": 0.81, "step": 3405 }, { "epoch": 0.6552520200076952, "grad_norm": 0.21210530109648595, "learning_rate": 5.612306564350178e-05, "loss": 0.8174, "step": 3406 }, { "epoch": 0.6554444016929588, "grad_norm": 0.1973036029224019, "learning_rate": 5.6067083797673536e-05, "loss": 0.7938, "step": 3407 }, { "epoch": 0.6556367833782224, "grad_norm": 0.17733980168743838, "learning_rate": 5.601111900811606e-05, "loss": 0.8145, "step": 3408 }, { "epoch": 0.655829165063486, "grad_norm": 0.19374699675714335, "learning_rate": 5.595517129655681e-05, "loss": 0.7324, "step": 3409 }, { "epoch": 0.6560215467487496, "grad_norm": 0.20781502207310934, "learning_rate": 5.589924068471648e-05, "loss": 0.7987, "step": 3410 }, { "epoch": 0.656213928434013, "grad_norm": 0.1943782485455082, "learning_rate": 5.5843327194309394e-05, "loss": 0.7082, "step": 3411 }, { "epoch": 0.6564063101192766, "grad_norm": 0.2130064442360513, "learning_rate": 5.578743084704305e-05, "loss": 0.7801, "step": 3412 }, { "epoch": 0.6565986918045402, "grad_norm": 0.20313822428235254, "learning_rate": 5.573155166461833e-05, "loss": 0.8039, "step": 3413 }, { "epoch": 0.6567910734898038, "grad_norm": 0.19582496814581724, "learning_rate": 5.567568966872947e-05, "loss": 0.8231, "step": 3414 }, { "epoch": 0.6569834551750673, "grad_norm": 0.19369510278527152, "learning_rate": 5.5619844881064e-05, "loss": 0.8284, "step": 3415 }, { "epoch": 0.6571758368603309, "grad_norm": 0.183399985811098, "learning_rate": 5.556401732330281e-05, "loss": 0.7694, "step": 3416 }, { "epoch": 0.6573682185455945, "grad_norm": 0.18743992185174146, "learning_rate": 5.550820701712007e-05, "loss": 0.8369, "step": 3417 }, { "epoch": 0.657560600230858, "grad_norm": 0.20461610773075686, "learning_rate": 5.545241398418326e-05, "loss": 0.7865, "step": 3418 }, { "epoch": 0.6577529819161216, "grad_norm": 0.19789972678267007, "learning_rate": 5.539663824615312e-05, "loss": 0.7761, "step": 3419 }, { "epoch": 0.6579453636013851, "grad_norm": 0.22358472827821116, "learning_rate": 5.5340879824683834e-05, "loss": 0.7815, "step": 3420 }, { "epoch": 0.6581377452866487, "grad_norm": 0.19723328467628345, "learning_rate": 5.528513874142262e-05, "loss": 0.8033, "step": 3421 }, { "epoch": 0.6583301269719123, "grad_norm": 0.19318613411831534, "learning_rate": 5.522941501801008e-05, "loss": 0.8182, "step": 3422 }, { "epoch": 0.6585225086571759, "grad_norm": 0.1913252297290458, "learning_rate": 5.51737086760802e-05, "loss": 0.7582, "step": 3423 }, { "epoch": 0.6587148903424394, "grad_norm": 0.20930141193023102, "learning_rate": 5.5118019737259966e-05, "loss": 0.7602, "step": 3424 }, { "epoch": 0.6589072720277029, "grad_norm": 0.19186617912487747, "learning_rate": 5.5062348223169825e-05, "loss": 0.7711, "step": 3425 }, { "epoch": 0.6590996537129665, "grad_norm": 0.18887729846791826, "learning_rate": 5.5006694155423356e-05, "loss": 0.8471, "step": 3426 }, { "epoch": 0.6592920353982301, "grad_norm": 0.1925635589477758, "learning_rate": 5.495105755562738e-05, "loss": 0.7804, "step": 3427 }, { "epoch": 0.6594844170834937, "grad_norm": 0.18988781628496298, "learning_rate": 5.489543844538194e-05, "loss": 0.786, "step": 3428 }, { "epoch": 0.6596767987687572, "grad_norm": 0.18581965312603282, "learning_rate": 5.48398368462803e-05, "loss": 0.7925, "step": 3429 }, { "epoch": 0.6598691804540208, "grad_norm": 0.20450516893781567, "learning_rate": 5.4784252779908905e-05, "loss": 0.8292, "step": 3430 }, { "epoch": 0.6600615621392844, "grad_norm": 0.20311449342774215, "learning_rate": 5.4728686267847354e-05, "loss": 0.7901, "step": 3431 }, { "epoch": 0.6602539438245479, "grad_norm": 0.19851769833245847, "learning_rate": 5.467313733166862e-05, "loss": 0.7988, "step": 3432 }, { "epoch": 0.6604463255098114, "grad_norm": 0.2127444633274374, "learning_rate": 5.461760599293855e-05, "loss": 0.799, "step": 3433 }, { "epoch": 0.660638707195075, "grad_norm": 0.19254829314743305, "learning_rate": 5.456209227321642e-05, "loss": 0.8165, "step": 3434 }, { "epoch": 0.6608310888803386, "grad_norm": 0.20284945459404347, "learning_rate": 5.4506596194054584e-05, "loss": 0.7918, "step": 3435 }, { "epoch": 0.6610234705656022, "grad_norm": 0.20644377640582443, "learning_rate": 5.445111777699842e-05, "loss": 0.7425, "step": 3436 }, { "epoch": 0.6612158522508658, "grad_norm": 0.1955105036345417, "learning_rate": 5.439565704358667e-05, "loss": 0.7577, "step": 3437 }, { "epoch": 0.6614082339361292, "grad_norm": 0.18416759441736733, "learning_rate": 5.4340214015351046e-05, "loss": 0.8139, "step": 3438 }, { "epoch": 0.6616006156213928, "grad_norm": 0.21742616893487313, "learning_rate": 5.428478871381646e-05, "loss": 0.732, "step": 3439 }, { "epoch": 0.6617929973066564, "grad_norm": 0.18046494776748567, "learning_rate": 5.422938116050091e-05, "loss": 0.813, "step": 3440 }, { "epoch": 0.66198537899192, "grad_norm": 0.19227703506766267, "learning_rate": 5.417399137691552e-05, "loss": 0.8305, "step": 3441 }, { "epoch": 0.6621777606771835, "grad_norm": 0.11333239226199517, "learning_rate": 5.411861938456453e-05, "loss": 0.6573, "step": 3442 }, { "epoch": 0.6623701423624471, "grad_norm": 0.20460115761185063, "learning_rate": 5.4063265204945224e-05, "loss": 0.7939, "step": 3443 }, { "epoch": 0.6625625240477107, "grad_norm": 0.24936150812220312, "learning_rate": 5.400792885954802e-05, "loss": 0.8005, "step": 3444 }, { "epoch": 0.6627549057329742, "grad_norm": 0.19417463953224143, "learning_rate": 5.395261036985635e-05, "loss": 0.8184, "step": 3445 }, { "epoch": 0.6629472874182378, "grad_norm": 0.22715930561194636, "learning_rate": 5.389730975734686e-05, "loss": 0.815, "step": 3446 }, { "epoch": 0.6631396691035013, "grad_norm": 0.19998389504322253, "learning_rate": 5.384202704348902e-05, "loss": 0.8099, "step": 3447 }, { "epoch": 0.6633320507887649, "grad_norm": 0.20263838024873254, "learning_rate": 5.378676224974557e-05, "loss": 0.8146, "step": 3448 }, { "epoch": 0.6635244324740285, "grad_norm": 0.2087695893478533, "learning_rate": 5.373151539757224e-05, "loss": 0.7735, "step": 3449 }, { "epoch": 0.6637168141592921, "grad_norm": 0.2004777280315202, "learning_rate": 5.36762865084176e-05, "loss": 0.7818, "step": 3450 }, { "epoch": 0.6639091958445555, "grad_norm": 0.19762667981325752, "learning_rate": 5.362107560372358e-05, "loss": 0.7643, "step": 3451 }, { "epoch": 0.6641015775298191, "grad_norm": 0.18923273772160673, "learning_rate": 5.356588270492487e-05, "loss": 0.7772, "step": 3452 }, { "epoch": 0.6642939592150827, "grad_norm": 0.20980107504646128, "learning_rate": 5.351070783344926e-05, "loss": 0.835, "step": 3453 }, { "epoch": 0.6644863409003463, "grad_norm": 0.20492933615259123, "learning_rate": 5.3455551010717545e-05, "loss": 0.8279, "step": 3454 }, { "epoch": 0.6646787225856099, "grad_norm": 0.19192704544614894, "learning_rate": 5.340041225814349e-05, "loss": 0.8257, "step": 3455 }, { "epoch": 0.6648711042708734, "grad_norm": 0.2080783729979962, "learning_rate": 5.334529159713388e-05, "loss": 0.8608, "step": 3456 }, { "epoch": 0.665063485956137, "grad_norm": 0.1941102920334524, "learning_rate": 5.3290189049088416e-05, "loss": 0.7674, "step": 3457 }, { "epoch": 0.6652558676414005, "grad_norm": 0.17006062690306906, "learning_rate": 5.323510463539989e-05, "loss": 0.7205, "step": 3458 }, { "epoch": 0.6654482493266641, "grad_norm": 0.1849897353413788, "learning_rate": 5.3180038377453826e-05, "loss": 0.8222, "step": 3459 }, { "epoch": 0.6656406310119277, "grad_norm": 0.1920511832724923, "learning_rate": 5.3124990296628976e-05, "loss": 0.7758, "step": 3460 }, { "epoch": 0.6658330126971912, "grad_norm": 0.1842650407821742, "learning_rate": 5.306996041429687e-05, "loss": 0.8464, "step": 3461 }, { "epoch": 0.6660253943824548, "grad_norm": 0.18738140795336067, "learning_rate": 5.3014948751821915e-05, "loss": 0.703, "step": 3462 }, { "epoch": 0.6662177760677184, "grad_norm": 0.11799451347953516, "learning_rate": 5.295995533056162e-05, "loss": 0.6848, "step": 3463 }, { "epoch": 0.666410157752982, "grad_norm": 0.1823377236798642, "learning_rate": 5.290498017186631e-05, "loss": 0.8342, "step": 3464 }, { "epoch": 0.6666025394382454, "grad_norm": 0.2014444840327735, "learning_rate": 5.285002329707923e-05, "loss": 0.8537, "step": 3465 }, { "epoch": 0.666794921123509, "grad_norm": 0.19413646697929926, "learning_rate": 5.2795084727536536e-05, "loss": 0.7995, "step": 3466 }, { "epoch": 0.6669873028087726, "grad_norm": 0.19863459393974006, "learning_rate": 5.274016448456724e-05, "loss": 0.8004, "step": 3467 }, { "epoch": 0.6671796844940362, "grad_norm": 0.1776561365985681, "learning_rate": 5.268526258949331e-05, "loss": 0.8099, "step": 3468 }, { "epoch": 0.6673720661792998, "grad_norm": 0.19071686898864923, "learning_rate": 5.2630379063629534e-05, "loss": 0.7778, "step": 3469 }, { "epoch": 0.6675644478645633, "grad_norm": 0.10283243725762735, "learning_rate": 5.2575513928283594e-05, "loss": 0.6205, "step": 3470 }, { "epoch": 0.6677568295498268, "grad_norm": 0.19397619420167367, "learning_rate": 5.252066720475597e-05, "loss": 0.7711, "step": 3471 }, { "epoch": 0.6679492112350904, "grad_norm": 0.19936271674696762, "learning_rate": 5.246583891434018e-05, "loss": 0.8325, "step": 3472 }, { "epoch": 0.668141592920354, "grad_norm": 0.16771103047294886, "learning_rate": 5.2411029078322314e-05, "loss": 0.7965, "step": 3473 }, { "epoch": 0.6683339746056175, "grad_norm": 0.18251961635279648, "learning_rate": 5.235623771798151e-05, "loss": 0.8033, "step": 3474 }, { "epoch": 0.6685263562908811, "grad_norm": 0.18034791636276917, "learning_rate": 5.2301464854589696e-05, "loss": 0.804, "step": 3475 }, { "epoch": 0.6687187379761447, "grad_norm": 0.18764326411190033, "learning_rate": 5.224671050941146e-05, "loss": 0.816, "step": 3476 }, { "epoch": 0.6689111196614083, "grad_norm": 0.2020038410086649, "learning_rate": 5.219197470370443e-05, "loss": 0.744, "step": 3477 }, { "epoch": 0.6691035013466718, "grad_norm": 0.18508786799534918, "learning_rate": 5.2137257458718894e-05, "loss": 0.7945, "step": 3478 }, { "epoch": 0.6692958830319353, "grad_norm": 0.1788566112673712, "learning_rate": 5.208255879569799e-05, "loss": 0.7304, "step": 3479 }, { "epoch": 0.6694882647171989, "grad_norm": 0.19347552908057225, "learning_rate": 5.20278787358776e-05, "loss": 0.7867, "step": 3480 }, { "epoch": 0.6696806464024625, "grad_norm": 0.17849133108342763, "learning_rate": 5.197321730048641e-05, "loss": 0.772, "step": 3481 }, { "epoch": 0.6698730280877261, "grad_norm": 0.17587757326051662, "learning_rate": 5.191857451074587e-05, "loss": 0.7587, "step": 3482 }, { "epoch": 0.6700654097729896, "grad_norm": 0.1962153196857849, "learning_rate": 5.1863950387870176e-05, "loss": 0.848, "step": 3483 }, { "epoch": 0.6702577914582531, "grad_norm": 0.18281044958650294, "learning_rate": 5.1809344953066374e-05, "loss": 0.7649, "step": 3484 }, { "epoch": 0.6704501731435167, "grad_norm": 0.18799980777191092, "learning_rate": 5.175475822753404e-05, "loss": 0.795, "step": 3485 }, { "epoch": 0.6706425548287803, "grad_norm": 0.19345667059034066, "learning_rate": 5.170019023246574e-05, "loss": 0.8112, "step": 3486 }, { "epoch": 0.6708349365140439, "grad_norm": 0.18803504530505277, "learning_rate": 5.1645640989046595e-05, "loss": 0.75, "step": 3487 }, { "epoch": 0.6710273181993074, "grad_norm": 0.19170834523068386, "learning_rate": 5.159111051845451e-05, "loss": 0.8154, "step": 3488 }, { "epoch": 0.671219699884571, "grad_norm": 0.16968277413501112, "learning_rate": 5.153659884186013e-05, "loss": 0.7994, "step": 3489 }, { "epoch": 0.6714120815698346, "grad_norm": 0.194001657870788, "learning_rate": 5.1482105980426643e-05, "loss": 0.8039, "step": 3490 }, { "epoch": 0.6716044632550981, "grad_norm": 0.19861523756701283, "learning_rate": 5.142763195531017e-05, "loss": 0.8096, "step": 3491 }, { "epoch": 0.6717968449403616, "grad_norm": 0.20548136290369295, "learning_rate": 5.1373176787659386e-05, "loss": 0.8227, "step": 3492 }, { "epoch": 0.6719892266256252, "grad_norm": 0.21759067249351788, "learning_rate": 5.1318740498615624e-05, "loss": 0.8108, "step": 3493 }, { "epoch": 0.6721816083108888, "grad_norm": 0.18262496023848687, "learning_rate": 5.126432310931295e-05, "loss": 0.8094, "step": 3494 }, { "epoch": 0.6723739899961524, "grad_norm": 0.20968211816723875, "learning_rate": 5.120992464087807e-05, "loss": 0.7805, "step": 3495 }, { "epoch": 0.672566371681416, "grad_norm": 0.18312039780345704, "learning_rate": 5.115554511443033e-05, "loss": 0.7932, "step": 3496 }, { "epoch": 0.6727587533666795, "grad_norm": 0.20442960392690251, "learning_rate": 5.11011845510817e-05, "loss": 0.7734, "step": 3497 }, { "epoch": 0.672951135051943, "grad_norm": 0.18879337113981906, "learning_rate": 5.1046842971936936e-05, "loss": 0.7485, "step": 3498 }, { "epoch": 0.6731435167372066, "grad_norm": 0.20137250708579982, "learning_rate": 5.099252039809317e-05, "loss": 0.8188, "step": 3499 }, { "epoch": 0.6733358984224702, "grad_norm": 0.1850597909371135, "learning_rate": 5.0938216850640396e-05, "loss": 0.814, "step": 3500 }, { "epoch": 0.6735282801077337, "grad_norm": 0.20421622605827033, "learning_rate": 5.0883932350661134e-05, "loss": 0.7469, "step": 3501 }, { "epoch": 0.6737206617929973, "grad_norm": 0.18207455333544578, "learning_rate": 5.082966691923037e-05, "loss": 0.7639, "step": 3502 }, { "epoch": 0.6739130434782609, "grad_norm": 0.1930014058062258, "learning_rate": 5.077542057741592e-05, "loss": 0.7581, "step": 3503 }, { "epoch": 0.6741054251635245, "grad_norm": 0.19524200887514723, "learning_rate": 5.072119334627806e-05, "loss": 0.7881, "step": 3504 }, { "epoch": 0.674297806848788, "grad_norm": 0.18816933410875214, "learning_rate": 5.066698524686966e-05, "loss": 0.8493, "step": 3505 }, { "epoch": 0.6744901885340515, "grad_norm": 0.20778405496873975, "learning_rate": 5.0612796300236175e-05, "loss": 0.7959, "step": 3506 }, { "epoch": 0.6746825702193151, "grad_norm": 0.2039222894395242, "learning_rate": 5.055862652741562e-05, "loss": 0.8122, "step": 3507 }, { "epoch": 0.6748749519045787, "grad_norm": 0.1987468671848671, "learning_rate": 5.0504475949438555e-05, "loss": 0.8003, "step": 3508 }, { "epoch": 0.6750673335898423, "grad_norm": 0.19806679029022045, "learning_rate": 5.045034458732808e-05, "loss": 0.7592, "step": 3509 }, { "epoch": 0.6752597152751059, "grad_norm": 0.10611878612048242, "learning_rate": 5.039623246209995e-05, "loss": 0.6316, "step": 3510 }, { "epoch": 0.6754520969603693, "grad_norm": 0.21442428312488093, "learning_rate": 5.0342139594762216e-05, "loss": 0.8234, "step": 3511 }, { "epoch": 0.6756444786456329, "grad_norm": 0.19635344870170562, "learning_rate": 5.028806600631569e-05, "loss": 0.782, "step": 3512 }, { "epoch": 0.6758368603308965, "grad_norm": 0.1934287909836615, "learning_rate": 5.023401171775357e-05, "loss": 0.7476, "step": 3513 }, { "epoch": 0.6760292420161601, "grad_norm": 0.1969713477458372, "learning_rate": 5.0179976750061605e-05, "loss": 0.8094, "step": 3514 }, { "epoch": 0.6762216237014236, "grad_norm": 0.20151948728219912, "learning_rate": 5.012596112421806e-05, "loss": 0.7751, "step": 3515 }, { "epoch": 0.6764140053866872, "grad_norm": 0.20236679748150518, "learning_rate": 5.007196486119354e-05, "loss": 0.7989, "step": 3516 }, { "epoch": 0.6766063870719508, "grad_norm": 0.20056803022658454, "learning_rate": 5.001798798195136e-05, "loss": 0.7667, "step": 3517 }, { "epoch": 0.6767987687572143, "grad_norm": 0.21756022770488725, "learning_rate": 4.996403050744719e-05, "loss": 0.8214, "step": 3518 }, { "epoch": 0.6769911504424779, "grad_norm": 0.21247047144208892, "learning_rate": 4.991009245862917e-05, "loss": 0.8127, "step": 3519 }, { "epoch": 0.6771835321277414, "grad_norm": 0.21834069587202998, "learning_rate": 4.985617385643789e-05, "loss": 0.8421, "step": 3520 }, { "epoch": 0.677375913813005, "grad_norm": 0.192503717179977, "learning_rate": 4.980227472180643e-05, "loss": 0.7336, "step": 3521 }, { "epoch": 0.6775682954982686, "grad_norm": 0.20037033738131083, "learning_rate": 4.974839507566027e-05, "loss": 0.779, "step": 3522 }, { "epoch": 0.6777606771835322, "grad_norm": 0.21510015044987837, "learning_rate": 4.969453493891732e-05, "loss": 0.7945, "step": 3523 }, { "epoch": 0.6779530588687956, "grad_norm": 0.18771059970483492, "learning_rate": 4.964069433248807e-05, "loss": 0.7461, "step": 3524 }, { "epoch": 0.6781454405540592, "grad_norm": 0.1948001649676463, "learning_rate": 4.95868732772751e-05, "loss": 0.8109, "step": 3525 }, { "epoch": 0.6783378222393228, "grad_norm": 0.1848347989219261, "learning_rate": 4.953307179417376e-05, "loss": 0.7302, "step": 3526 }, { "epoch": 0.6785302039245864, "grad_norm": 0.19921930093976972, "learning_rate": 4.947928990407156e-05, "loss": 0.7554, "step": 3527 }, { "epoch": 0.67872258560985, "grad_norm": 0.20248197360186232, "learning_rate": 4.94255276278485e-05, "loss": 0.8186, "step": 3528 }, { "epoch": 0.6789149672951135, "grad_norm": 0.10400545061023375, "learning_rate": 4.937178498637696e-05, "loss": 0.6048, "step": 3529 }, { "epoch": 0.6791073489803771, "grad_norm": 0.1902267562848652, "learning_rate": 4.9318062000521644e-05, "loss": 0.8074, "step": 3530 }, { "epoch": 0.6792997306656406, "grad_norm": 0.21339914453152742, "learning_rate": 4.926435869113971e-05, "loss": 0.7833, "step": 3531 }, { "epoch": 0.6794921123509042, "grad_norm": 0.1140788053085581, "learning_rate": 4.9210675079080607e-05, "loss": 0.6329, "step": 3532 }, { "epoch": 0.6796844940361677, "grad_norm": 0.21826033758703797, "learning_rate": 4.915701118518615e-05, "loss": 0.8067, "step": 3533 }, { "epoch": 0.6798768757214313, "grad_norm": 0.20377311385238073, "learning_rate": 4.9103367030290547e-05, "loss": 0.7902, "step": 3534 }, { "epoch": 0.6800692574066949, "grad_norm": 0.18911434388744433, "learning_rate": 4.9049742635220243e-05, "loss": 0.8296, "step": 3535 }, { "epoch": 0.6802616390919585, "grad_norm": 0.19582618107568223, "learning_rate": 4.899613802079419e-05, "loss": 0.7596, "step": 3536 }, { "epoch": 0.6804540207772221, "grad_norm": 0.1882037682014994, "learning_rate": 4.8942553207823395e-05, "loss": 0.8168, "step": 3537 }, { "epoch": 0.6806464024624855, "grad_norm": 0.212869175344217, "learning_rate": 4.8888988217111444e-05, "loss": 0.7601, "step": 3538 }, { "epoch": 0.6808387841477491, "grad_norm": 0.1761701215605617, "learning_rate": 4.883544306945407e-05, "loss": 0.8016, "step": 3539 }, { "epoch": 0.6810311658330127, "grad_norm": 0.220714057507758, "learning_rate": 4.878191778563934e-05, "loss": 0.7893, "step": 3540 }, { "epoch": 0.6812235475182763, "grad_norm": 0.19112997723315292, "learning_rate": 4.8728412386447654e-05, "loss": 0.7719, "step": 3541 }, { "epoch": 0.6814159292035398, "grad_norm": 0.1894405382793158, "learning_rate": 4.867492689265154e-05, "loss": 0.8034, "step": 3542 }, { "epoch": 0.6816083108888034, "grad_norm": 0.19076304352801596, "learning_rate": 4.862146132501602e-05, "loss": 0.7666, "step": 3543 }, { "epoch": 0.681800692574067, "grad_norm": 0.18653955678383272, "learning_rate": 4.856801570429822e-05, "loss": 0.8393, "step": 3544 }, { "epoch": 0.6819930742593305, "grad_norm": 0.18901828090640632, "learning_rate": 4.851459005124759e-05, "loss": 0.7913, "step": 3545 }, { "epoch": 0.6821854559445941, "grad_norm": 0.1829903933608884, "learning_rate": 4.8461184386605785e-05, "loss": 0.7708, "step": 3546 }, { "epoch": 0.6823778376298576, "grad_norm": 0.1969107308803075, "learning_rate": 4.8407798731106746e-05, "loss": 0.8109, "step": 3547 }, { "epoch": 0.6825702193151212, "grad_norm": 0.11919239575389515, "learning_rate": 4.8354433105476604e-05, "loss": 0.6195, "step": 3548 }, { "epoch": 0.6827626010003848, "grad_norm": 0.19714818189871888, "learning_rate": 4.83010875304337e-05, "loss": 0.8041, "step": 3549 }, { "epoch": 0.6829549826856484, "grad_norm": 0.20444283375417643, "learning_rate": 4.8247762026688746e-05, "loss": 0.805, "step": 3550 }, { "epoch": 0.6831473643709118, "grad_norm": 0.19101753054415577, "learning_rate": 4.8194456614944375e-05, "loss": 0.7604, "step": 3551 }, { "epoch": 0.6833397460561754, "grad_norm": 0.19368358413123507, "learning_rate": 4.814117131589569e-05, "loss": 0.7903, "step": 3552 }, { "epoch": 0.683532127741439, "grad_norm": 0.18148308890625803, "learning_rate": 4.8087906150229866e-05, "loss": 0.8201, "step": 3553 }, { "epoch": 0.6837245094267026, "grad_norm": 0.20478146940829867, "learning_rate": 4.8034661138626255e-05, "loss": 0.8527, "step": 3554 }, { "epoch": 0.6839168911119662, "grad_norm": 0.1991278710602984, "learning_rate": 4.798143630175642e-05, "loss": 0.722, "step": 3555 }, { "epoch": 0.6841092727972297, "grad_norm": 0.19904414005847318, "learning_rate": 4.7928231660284054e-05, "loss": 0.7987, "step": 3556 }, { "epoch": 0.6843016544824932, "grad_norm": 0.21608710103770354, "learning_rate": 4.7875047234865045e-05, "loss": 0.8307, "step": 3557 }, { "epoch": 0.6844940361677568, "grad_norm": 0.2023407771608046, "learning_rate": 4.782188304614741e-05, "loss": 0.8328, "step": 3558 }, { "epoch": 0.6846864178530204, "grad_norm": 0.19029839212727864, "learning_rate": 4.776873911477132e-05, "loss": 0.8225, "step": 3559 }, { "epoch": 0.6848787995382839, "grad_norm": 0.18952648021693816, "learning_rate": 4.771561546136908e-05, "loss": 0.7863, "step": 3560 }, { "epoch": 0.6850711812235475, "grad_norm": 0.2089770027637194, "learning_rate": 4.766251210656509e-05, "loss": 0.7834, "step": 3561 }, { "epoch": 0.6852635629088111, "grad_norm": 0.1984894628581875, "learning_rate": 4.7609429070976006e-05, "loss": 0.7687, "step": 3562 }, { "epoch": 0.6854559445940747, "grad_norm": 0.1277624520669126, "learning_rate": 4.7556366375210345e-05, "loss": 0.6744, "step": 3563 }, { "epoch": 0.6856483262793382, "grad_norm": 0.20323187663630982, "learning_rate": 4.750332403986901e-05, "loss": 0.784, "step": 3564 }, { "epoch": 0.6858407079646017, "grad_norm": 0.19449642235884818, "learning_rate": 4.745030208554473e-05, "loss": 0.7592, "step": 3565 }, { "epoch": 0.6860330896498653, "grad_norm": 0.17674973805043828, "learning_rate": 4.739730053282255e-05, "loss": 0.7823, "step": 3566 }, { "epoch": 0.6862254713351289, "grad_norm": 0.18723769835334625, "learning_rate": 4.734431940227951e-05, "loss": 0.792, "step": 3567 }, { "epoch": 0.6864178530203925, "grad_norm": 0.2066636275780325, "learning_rate": 4.729135871448459e-05, "loss": 0.8204, "step": 3568 }, { "epoch": 0.6866102347056561, "grad_norm": 0.20337658340651182, "learning_rate": 4.723841848999907e-05, "loss": 0.7992, "step": 3569 }, { "epoch": 0.6868026163909196, "grad_norm": 0.22493133595006487, "learning_rate": 4.7185498749376125e-05, "loss": 0.8222, "step": 3570 }, { "epoch": 0.6869949980761831, "grad_norm": 0.18701871222837624, "learning_rate": 4.713259951316103e-05, "loss": 0.7417, "step": 3571 }, { "epoch": 0.6871873797614467, "grad_norm": 0.200402827812252, "learning_rate": 4.7079720801891055e-05, "loss": 0.7556, "step": 3572 }, { "epoch": 0.6873797614467103, "grad_norm": 0.2061002194121459, "learning_rate": 4.702686263609559e-05, "loss": 0.7646, "step": 3573 }, { "epoch": 0.6875721431319738, "grad_norm": 0.20213824981063322, "learning_rate": 4.6974025036295954e-05, "loss": 0.736, "step": 3574 }, { "epoch": 0.6877645248172374, "grad_norm": 0.19944842706417118, "learning_rate": 4.6921208023005505e-05, "loss": 0.8206, "step": 3575 }, { "epoch": 0.687956906502501, "grad_norm": 0.1977833893852174, "learning_rate": 4.6868411616729734e-05, "loss": 0.8251, "step": 3576 }, { "epoch": 0.6881492881877646, "grad_norm": 0.20052612669118802, "learning_rate": 4.681563583796587e-05, "loss": 0.818, "step": 3577 }, { "epoch": 0.6883416698730281, "grad_norm": 0.18393407007359303, "learning_rate": 4.6762880707203404e-05, "loss": 0.7541, "step": 3578 }, { "epoch": 0.6885340515582916, "grad_norm": 0.20087459784571732, "learning_rate": 4.671014624492365e-05, "loss": 0.8251, "step": 3579 }, { "epoch": 0.6887264332435552, "grad_norm": 0.19081986523032557, "learning_rate": 4.665743247159995e-05, "loss": 0.7821, "step": 3580 }, { "epoch": 0.6889188149288188, "grad_norm": 0.1857436793642905, "learning_rate": 4.660473940769761e-05, "loss": 0.8031, "step": 3581 }, { "epoch": 0.6891111966140824, "grad_norm": 0.19634617095098136, "learning_rate": 4.6552067073673876e-05, "loss": 0.7673, "step": 3582 }, { "epoch": 0.6893035782993459, "grad_norm": 0.1808859423739587, "learning_rate": 4.649941548997797e-05, "loss": 0.8149, "step": 3583 }, { "epoch": 0.6894959599846094, "grad_norm": 0.19751296815931652, "learning_rate": 4.644678467705101e-05, "loss": 0.7784, "step": 3584 }, { "epoch": 0.689688341669873, "grad_norm": 0.18437314526286394, "learning_rate": 4.6394174655326216e-05, "loss": 0.7505, "step": 3585 }, { "epoch": 0.6898807233551366, "grad_norm": 0.1920214028266703, "learning_rate": 4.6341585445228485e-05, "loss": 0.7802, "step": 3586 }, { "epoch": 0.6900731050404002, "grad_norm": 0.1874346048007456, "learning_rate": 4.628901706717476e-05, "loss": 0.8043, "step": 3587 }, { "epoch": 0.6902654867256637, "grad_norm": 0.19966511041623775, "learning_rate": 4.623646954157399e-05, "loss": 0.7703, "step": 3588 }, { "epoch": 0.6904578684109273, "grad_norm": 0.19475391363488603, "learning_rate": 4.618394288882681e-05, "loss": 0.8209, "step": 3589 }, { "epoch": 0.6906502500961909, "grad_norm": 0.18575468788160568, "learning_rate": 4.6131437129326025e-05, "loss": 0.7872, "step": 3590 }, { "epoch": 0.6908426317814544, "grad_norm": 0.10494924773416361, "learning_rate": 4.6078952283456024e-05, "loss": 0.607, "step": 3591 }, { "epoch": 0.6910350134667179, "grad_norm": 0.17881790575202786, "learning_rate": 4.6026488371593325e-05, "loss": 0.7652, "step": 3592 }, { "epoch": 0.6912273951519815, "grad_norm": 0.11621513153196152, "learning_rate": 4.5974045414106216e-05, "loss": 0.653, "step": 3593 }, { "epoch": 0.6914197768372451, "grad_norm": 0.20743352045667168, "learning_rate": 4.592162343135483e-05, "loss": 0.7803, "step": 3594 }, { "epoch": 0.6916121585225087, "grad_norm": 0.1883771304942785, "learning_rate": 4.586922244369122e-05, "loss": 0.7794, "step": 3595 }, { "epoch": 0.6918045402077723, "grad_norm": 0.1931010864053503, "learning_rate": 4.581684247145922e-05, "loss": 0.7446, "step": 3596 }, { "epoch": 0.6919969218930357, "grad_norm": 0.1950677741622504, "learning_rate": 4.576448353499456e-05, "loss": 0.8154, "step": 3597 }, { "epoch": 0.6921893035782993, "grad_norm": 0.21558914793113096, "learning_rate": 4.5712145654624774e-05, "loss": 0.7493, "step": 3598 }, { "epoch": 0.6923816852635629, "grad_norm": 0.20438027603167186, "learning_rate": 4.565982885066923e-05, "loss": 0.7334, "step": 3599 }, { "epoch": 0.6925740669488265, "grad_norm": 0.2031153070639275, "learning_rate": 4.560753314343912e-05, "loss": 0.7798, "step": 3600 }, { "epoch": 0.69276644863409, "grad_norm": 0.17862933672810277, "learning_rate": 4.555525855323738e-05, "loss": 0.7382, "step": 3601 }, { "epoch": 0.6929588303193536, "grad_norm": 0.1836225504968393, "learning_rate": 4.5503005100358945e-05, "loss": 0.7366, "step": 3602 }, { "epoch": 0.6931512120046172, "grad_norm": 0.13953349500565573, "learning_rate": 4.5450772805090226e-05, "loss": 0.6296, "step": 3603 }, { "epoch": 0.6933435936898807, "grad_norm": 0.109047523093113, "learning_rate": 4.539856168770974e-05, "loss": 0.6175, "step": 3604 }, { "epoch": 0.6935359753751443, "grad_norm": 0.19713130313005883, "learning_rate": 4.5346371768487586e-05, "loss": 0.8011, "step": 3605 }, { "epoch": 0.6937283570604078, "grad_norm": 0.19571733917869832, "learning_rate": 4.529420306768569e-05, "loss": 0.8455, "step": 3606 }, { "epoch": 0.6939207387456714, "grad_norm": 0.20071544354343396, "learning_rate": 4.524205560555774e-05, "loss": 0.7249, "step": 3607 }, { "epoch": 0.694113120430935, "grad_norm": 0.1917275023037877, "learning_rate": 4.518992940234917e-05, "loss": 0.7576, "step": 3608 }, { "epoch": 0.6943055021161986, "grad_norm": 0.19681366896904515, "learning_rate": 4.5137824478297177e-05, "loss": 0.8121, "step": 3609 }, { "epoch": 0.694497883801462, "grad_norm": 0.1886349350952763, "learning_rate": 4.508574085363064e-05, "loss": 0.7794, "step": 3610 }, { "epoch": 0.6946902654867256, "grad_norm": 0.1979284907872457, "learning_rate": 4.5033678548570346e-05, "loss": 0.8189, "step": 3611 }, { "epoch": 0.6948826471719892, "grad_norm": 0.18055711778221828, "learning_rate": 4.498163758332853e-05, "loss": 0.7139, "step": 3612 }, { "epoch": 0.6950750288572528, "grad_norm": 0.18976018691905563, "learning_rate": 4.4929617978109315e-05, "loss": 0.8214, "step": 3613 }, { "epoch": 0.6952674105425164, "grad_norm": 0.19639257282136455, "learning_rate": 4.48776197531086e-05, "loss": 0.731, "step": 3614 }, { "epoch": 0.6954597922277799, "grad_norm": 0.19237671779075285, "learning_rate": 4.4825642928513744e-05, "loss": 0.7631, "step": 3615 }, { "epoch": 0.6956521739130435, "grad_norm": 0.20038886371311307, "learning_rate": 4.4773687524504084e-05, "loss": 0.794, "step": 3616 }, { "epoch": 0.695844555598307, "grad_norm": 0.1976961768941338, "learning_rate": 4.4721753561250354e-05, "loss": 0.7588, "step": 3617 }, { "epoch": 0.6960369372835706, "grad_norm": 0.1782449895843598, "learning_rate": 4.466984105891521e-05, "loss": 0.7533, "step": 3618 }, { "epoch": 0.6962293189688342, "grad_norm": 0.1735112894500305, "learning_rate": 4.4617950037652846e-05, "loss": 0.8381, "step": 3619 }, { "epoch": 0.6964217006540977, "grad_norm": 0.19653987856403438, "learning_rate": 4.4566080517609134e-05, "loss": 0.8066, "step": 3620 }, { "epoch": 0.6966140823393613, "grad_norm": 0.1839936907466742, "learning_rate": 4.4514232518921605e-05, "loss": 0.8016, "step": 3621 }, { "epoch": 0.6968064640246249, "grad_norm": 0.1853791081984738, "learning_rate": 4.4462406061719455e-05, "loss": 0.7863, "step": 3622 }, { "epoch": 0.6969988457098885, "grad_norm": 0.20750893986788316, "learning_rate": 4.441060116612348e-05, "loss": 0.8036, "step": 3623 }, { "epoch": 0.6971912273951519, "grad_norm": 0.11868243757695353, "learning_rate": 4.435881785224613e-05, "loss": 0.678, "step": 3624 }, { "epoch": 0.6973836090804155, "grad_norm": 0.20105956143729564, "learning_rate": 4.430705614019147e-05, "loss": 0.744, "step": 3625 }, { "epoch": 0.6975759907656791, "grad_norm": 0.20866183260417323, "learning_rate": 4.425531605005518e-05, "loss": 0.8421, "step": 3626 }, { "epoch": 0.6977683724509427, "grad_norm": 0.18803687624012894, "learning_rate": 4.420359760192452e-05, "loss": 0.8132, "step": 3627 }, { "epoch": 0.6979607541362063, "grad_norm": 0.20953648516577228, "learning_rate": 4.415190081587845e-05, "loss": 0.7499, "step": 3628 }, { "epoch": 0.6981531358214698, "grad_norm": 0.20697513307427332, "learning_rate": 4.4100225711987344e-05, "loss": 0.7953, "step": 3629 }, { "epoch": 0.6983455175067333, "grad_norm": 0.18529812436714407, "learning_rate": 4.404857231031332e-05, "loss": 0.7666, "step": 3630 }, { "epoch": 0.6985378991919969, "grad_norm": 0.19877009835366638, "learning_rate": 4.3996940630909985e-05, "loss": 0.8209, "step": 3631 }, { "epoch": 0.6987302808772605, "grad_norm": 0.1981890729350321, "learning_rate": 4.394533069382255e-05, "loss": 0.828, "step": 3632 }, { "epoch": 0.698922662562524, "grad_norm": 0.21517723317095386, "learning_rate": 4.3893742519087754e-05, "loss": 0.7552, "step": 3633 }, { "epoch": 0.6991150442477876, "grad_norm": 0.18449699309598838, "learning_rate": 4.3842176126733916e-05, "loss": 0.7945, "step": 3634 }, { "epoch": 0.6993074259330512, "grad_norm": 0.18670466612289363, "learning_rate": 4.379063153678087e-05, "loss": 0.7748, "step": 3635 }, { "epoch": 0.6994998076183148, "grad_norm": 0.1818673221325133, "learning_rate": 4.373910876923997e-05, "loss": 0.83, "step": 3636 }, { "epoch": 0.6996921893035783, "grad_norm": 0.1932506965062284, "learning_rate": 4.368760784411423e-05, "loss": 0.758, "step": 3637 }, { "epoch": 0.6998845709888418, "grad_norm": 0.20632445303316654, "learning_rate": 4.363612878139799e-05, "loss": 0.7769, "step": 3638 }, { "epoch": 0.7000769526741054, "grad_norm": 0.19474540932509865, "learning_rate": 4.358467160107722e-05, "loss": 0.8354, "step": 3639 }, { "epoch": 0.700269334359369, "grad_norm": 0.20295852925759178, "learning_rate": 4.353323632312938e-05, "loss": 0.8243, "step": 3640 }, { "epoch": 0.7004617160446326, "grad_norm": 0.19687118397440814, "learning_rate": 4.348182296752336e-05, "loss": 0.8277, "step": 3641 }, { "epoch": 0.7006540977298961, "grad_norm": 0.17779518958779628, "learning_rate": 4.343043155421971e-05, "loss": 0.8806, "step": 3642 }, { "epoch": 0.7008464794151597, "grad_norm": 0.19778136917232716, "learning_rate": 4.337906210317021e-05, "loss": 0.8026, "step": 3643 }, { "epoch": 0.7010388611004232, "grad_norm": 0.20156881131818966, "learning_rate": 4.332771463431836e-05, "loss": 0.777, "step": 3644 }, { "epoch": 0.7012312427856868, "grad_norm": 0.18573679556657507, "learning_rate": 4.3276389167598975e-05, "loss": 0.7782, "step": 3645 }, { "epoch": 0.7014236244709504, "grad_norm": 0.20829343672968859, "learning_rate": 4.322508572293835e-05, "loss": 0.7863, "step": 3646 }, { "epoch": 0.7016160061562139, "grad_norm": 0.19515801365863503, "learning_rate": 4.3173804320254275e-05, "loss": 0.786, "step": 3647 }, { "epoch": 0.7018083878414775, "grad_norm": 0.20970642784257232, "learning_rate": 4.312254497945595e-05, "loss": 0.6966, "step": 3648 }, { "epoch": 0.7020007695267411, "grad_norm": 0.18403772435885718, "learning_rate": 4.307130772044401e-05, "loss": 0.7674, "step": 3649 }, { "epoch": 0.7021931512120047, "grad_norm": 0.19592827053513853, "learning_rate": 4.302009256311048e-05, "loss": 0.7826, "step": 3650 }, { "epoch": 0.7023855328972681, "grad_norm": 0.21590822034568868, "learning_rate": 4.296889952733898e-05, "loss": 0.7898, "step": 3651 }, { "epoch": 0.7025779145825317, "grad_norm": 0.18133642147461754, "learning_rate": 4.291772863300427e-05, "loss": 0.7872, "step": 3652 }, { "epoch": 0.7027702962677953, "grad_norm": 0.20491654827006628, "learning_rate": 4.286657989997268e-05, "loss": 0.8401, "step": 3653 }, { "epoch": 0.7029626779530589, "grad_norm": 0.19978483285520535, "learning_rate": 4.281545334810201e-05, "loss": 0.7535, "step": 3654 }, { "epoch": 0.7031550596383225, "grad_norm": 0.19960335935963805, "learning_rate": 4.276434899724119e-05, "loss": 0.8327, "step": 3655 }, { "epoch": 0.703347441323586, "grad_norm": 0.19137356430003913, "learning_rate": 4.2713266867230804e-05, "loss": 0.749, "step": 3656 }, { "epoch": 0.7035398230088495, "grad_norm": 0.16216422902668154, "learning_rate": 4.2662206977902664e-05, "loss": 0.7884, "step": 3657 }, { "epoch": 0.7037322046941131, "grad_norm": 0.1853411941081993, "learning_rate": 4.261116934907998e-05, "loss": 0.8124, "step": 3658 }, { "epoch": 0.7039245863793767, "grad_norm": 0.1848096049669027, "learning_rate": 4.2560154000577304e-05, "loss": 0.8046, "step": 3659 }, { "epoch": 0.7041169680646402, "grad_norm": 0.20752097079837886, "learning_rate": 4.250916095220056e-05, "loss": 0.8117, "step": 3660 }, { "epoch": 0.7043093497499038, "grad_norm": 0.20931273206778098, "learning_rate": 4.2458190223747e-05, "loss": 0.7681, "step": 3661 }, { "epoch": 0.7045017314351674, "grad_norm": 0.1833463477594264, "learning_rate": 4.2407241835005176e-05, "loss": 0.8307, "step": 3662 }, { "epoch": 0.704694113120431, "grad_norm": 0.2171055000308932, "learning_rate": 4.235631580575513e-05, "loss": 0.8333, "step": 3663 }, { "epoch": 0.7048864948056945, "grad_norm": 0.1925724540582895, "learning_rate": 4.230541215576798e-05, "loss": 0.8232, "step": 3664 }, { "epoch": 0.705078876490958, "grad_norm": 0.1985974368485125, "learning_rate": 4.225453090480631e-05, "loss": 0.8298, "step": 3665 }, { "epoch": 0.7052712581762216, "grad_norm": 0.12469495354352797, "learning_rate": 4.220367207262398e-05, "loss": 0.679, "step": 3666 }, { "epoch": 0.7054636398614852, "grad_norm": 0.19238406853161358, "learning_rate": 4.2152835678966096e-05, "loss": 0.7896, "step": 3667 }, { "epoch": 0.7056560215467488, "grad_norm": 0.17967909315698302, "learning_rate": 4.210202174356922e-05, "loss": 0.8089, "step": 3668 }, { "epoch": 0.7058484032320124, "grad_norm": 0.1848025043930747, "learning_rate": 4.20512302861609e-05, "loss": 0.7569, "step": 3669 }, { "epoch": 0.7060407849172758, "grad_norm": 0.19623122505227159, "learning_rate": 4.200046132646027e-05, "loss": 0.7757, "step": 3670 }, { "epoch": 0.7062331666025394, "grad_norm": 0.20518408622416331, "learning_rate": 4.1949714884177525e-05, "loss": 0.7942, "step": 3671 }, { "epoch": 0.706425548287803, "grad_norm": 0.20362276418981676, "learning_rate": 4.1898990979014205e-05, "loss": 0.8479, "step": 3672 }, { "epoch": 0.7066179299730666, "grad_norm": 0.2029791882532862, "learning_rate": 4.184828963066305e-05, "loss": 0.8216, "step": 3673 }, { "epoch": 0.7068103116583301, "grad_norm": 0.20018001964534496, "learning_rate": 4.179761085880809e-05, "loss": 0.7819, "step": 3674 }, { "epoch": 0.7070026933435937, "grad_norm": 0.2018693299079706, "learning_rate": 4.174695468312456e-05, "loss": 0.7524, "step": 3675 }, { "epoch": 0.7071950750288573, "grad_norm": 0.19544296786473372, "learning_rate": 4.16963211232789e-05, "loss": 0.7664, "step": 3676 }, { "epoch": 0.7073874567141208, "grad_norm": 0.19038122702441085, "learning_rate": 4.16457101989289e-05, "loss": 0.8221, "step": 3677 }, { "epoch": 0.7075798383993844, "grad_norm": 0.19293763330269287, "learning_rate": 4.159512192972337e-05, "loss": 0.7995, "step": 3678 }, { "epoch": 0.7077722200846479, "grad_norm": 0.19273072391890753, "learning_rate": 4.15445563353024e-05, "loss": 0.7672, "step": 3679 }, { "epoch": 0.7079646017699115, "grad_norm": 0.10617334523563964, "learning_rate": 4.1494013435297416e-05, "loss": 0.6016, "step": 3680 }, { "epoch": 0.7081569834551751, "grad_norm": 0.18039706995080462, "learning_rate": 4.1443493249330766e-05, "loss": 0.7417, "step": 3681 }, { "epoch": 0.7083493651404387, "grad_norm": 0.17591675618388825, "learning_rate": 4.139299579701623e-05, "loss": 0.7576, "step": 3682 }, { "epoch": 0.7085417468257021, "grad_norm": 0.217572472862418, "learning_rate": 4.134252109795863e-05, "loss": 0.7612, "step": 3683 }, { "epoch": 0.7087341285109657, "grad_norm": 0.2057757309634087, "learning_rate": 4.129206917175397e-05, "loss": 0.8201, "step": 3684 }, { "epoch": 0.7089265101962293, "grad_norm": 0.17935178386002532, "learning_rate": 4.124164003798944e-05, "loss": 0.8119, "step": 3685 }, { "epoch": 0.7091188918814929, "grad_norm": 0.2127637823044462, "learning_rate": 4.119123371624335e-05, "loss": 0.8092, "step": 3686 }, { "epoch": 0.7093112735667565, "grad_norm": 0.18929615557340795, "learning_rate": 4.114085022608517e-05, "loss": 0.781, "step": 3687 }, { "epoch": 0.70950365525202, "grad_norm": 0.22314472685677658, "learning_rate": 4.109048958707552e-05, "loss": 0.8347, "step": 3688 }, { "epoch": 0.7096960369372836, "grad_norm": 0.21136850104913654, "learning_rate": 4.104015181876612e-05, "loss": 0.7873, "step": 3689 }, { "epoch": 0.7098884186225471, "grad_norm": 0.19098338914235333, "learning_rate": 4.09898369406998e-05, "loss": 0.7781, "step": 3690 }, { "epoch": 0.7100808003078107, "grad_norm": 0.20808642588551046, "learning_rate": 4.093954497241064e-05, "loss": 0.8217, "step": 3691 }, { "epoch": 0.7102731819930742, "grad_norm": 0.18246454945761367, "learning_rate": 4.088927593342358e-05, "loss": 0.7779, "step": 3692 }, { "epoch": 0.7104655636783378, "grad_norm": 0.18716039339665316, "learning_rate": 4.0839029843254814e-05, "loss": 0.7551, "step": 3693 }, { "epoch": 0.7106579453636014, "grad_norm": 0.1953733897834668, "learning_rate": 4.078880672141171e-05, "loss": 0.7444, "step": 3694 }, { "epoch": 0.710850327048865, "grad_norm": 0.19948667307144058, "learning_rate": 4.073860658739246e-05, "loss": 0.8483, "step": 3695 }, { "epoch": 0.7110427087341286, "grad_norm": 0.21219934156727435, "learning_rate": 4.0688429460686605e-05, "loss": 0.7535, "step": 3696 }, { "epoch": 0.711235090419392, "grad_norm": 0.17122738177159752, "learning_rate": 4.0638275360774594e-05, "loss": 0.7539, "step": 3697 }, { "epoch": 0.7114274721046556, "grad_norm": 0.22058561601873075, "learning_rate": 4.058814430712796e-05, "loss": 0.7715, "step": 3698 }, { "epoch": 0.7116198537899192, "grad_norm": 0.21024424933253888, "learning_rate": 4.053803631920933e-05, "loss": 0.8012, "step": 3699 }, { "epoch": 0.7118122354751828, "grad_norm": 0.1950939253569678, "learning_rate": 4.0487951416472326e-05, "loss": 0.7851, "step": 3700 }, { "epoch": 0.7120046171604463, "grad_norm": 0.20848838225247382, "learning_rate": 4.0437889618361635e-05, "loss": 0.8162, "step": 3701 }, { "epoch": 0.7121969988457099, "grad_norm": 0.19859331770179225, "learning_rate": 4.0387850944312957e-05, "loss": 0.8039, "step": 3702 }, { "epoch": 0.7123893805309734, "grad_norm": 0.10413567490920214, "learning_rate": 4.033783541375311e-05, "loss": 0.5838, "step": 3703 }, { "epoch": 0.712581762216237, "grad_norm": 0.19137222633928733, "learning_rate": 4.028784304609976e-05, "loss": 0.7917, "step": 3704 }, { "epoch": 0.7127741439015006, "grad_norm": 0.20365328209354863, "learning_rate": 4.0237873860761646e-05, "loss": 0.7931, "step": 3705 }, { "epoch": 0.7129665255867641, "grad_norm": 0.20575145290992564, "learning_rate": 4.018792787713864e-05, "loss": 0.7795, "step": 3706 }, { "epoch": 0.7131589072720277, "grad_norm": 0.19573322991300324, "learning_rate": 4.013800511462135e-05, "loss": 0.7871, "step": 3707 }, { "epoch": 0.7133512889572913, "grad_norm": 0.1960848988429678, "learning_rate": 4.0088105592591616e-05, "loss": 0.7342, "step": 3708 }, { "epoch": 0.7135436706425549, "grad_norm": 0.18602608211931032, "learning_rate": 4.003822933042213e-05, "loss": 0.7785, "step": 3709 }, { "epoch": 0.7137360523278183, "grad_norm": 0.19848077895198776, "learning_rate": 3.998837634747655e-05, "loss": 0.7698, "step": 3710 }, { "epoch": 0.7139284340130819, "grad_norm": 0.2048810884875228, "learning_rate": 3.9938546663109544e-05, "loss": 0.774, "step": 3711 }, { "epoch": 0.7141208156983455, "grad_norm": 0.18940125231908095, "learning_rate": 3.9888740296666696e-05, "loss": 0.8153, "step": 3712 }, { "epoch": 0.7143131973836091, "grad_norm": 0.19067573718116967, "learning_rate": 3.983895726748455e-05, "loss": 0.7638, "step": 3713 }, { "epoch": 0.7145055790688727, "grad_norm": 0.19246562889188773, "learning_rate": 3.9789197594890606e-05, "loss": 0.8094, "step": 3714 }, { "epoch": 0.7146979607541362, "grad_norm": 0.18278216131399658, "learning_rate": 3.973946129820326e-05, "loss": 0.7787, "step": 3715 }, { "epoch": 0.7148903424393998, "grad_norm": 0.2133787573113392, "learning_rate": 3.9689748396731854e-05, "loss": 0.6995, "step": 3716 }, { "epoch": 0.7150827241246633, "grad_norm": 0.19151362476595746, "learning_rate": 3.964005890977672e-05, "loss": 0.7974, "step": 3717 }, { "epoch": 0.7152751058099269, "grad_norm": 0.0998546135965482, "learning_rate": 3.959039285662894e-05, "loss": 0.5891, "step": 3718 }, { "epoch": 0.7154674874951905, "grad_norm": 0.19926110586431578, "learning_rate": 3.954075025657058e-05, "loss": 0.7598, "step": 3719 }, { "epoch": 0.715659869180454, "grad_norm": 0.18258097286821942, "learning_rate": 3.949113112887471e-05, "loss": 0.8112, "step": 3720 }, { "epoch": 0.7158522508657176, "grad_norm": 0.17430771176928744, "learning_rate": 3.944153549280506e-05, "loss": 0.8522, "step": 3721 }, { "epoch": 0.7160446325509812, "grad_norm": 0.17937513481915568, "learning_rate": 3.939196336761645e-05, "loss": 0.7495, "step": 3722 }, { "epoch": 0.7162370142362448, "grad_norm": 0.1892330487716907, "learning_rate": 3.934241477255445e-05, "loss": 0.8246, "step": 3723 }, { "epoch": 0.7164293959215082, "grad_norm": 0.19987593115792687, "learning_rate": 3.929288972685554e-05, "loss": 0.757, "step": 3724 }, { "epoch": 0.7166217776067718, "grad_norm": 0.19674507208466124, "learning_rate": 3.9243388249747046e-05, "loss": 0.8254, "step": 3725 }, { "epoch": 0.7168141592920354, "grad_norm": 0.10872908563410567, "learning_rate": 3.919391036044715e-05, "loss": 0.6132, "step": 3726 }, { "epoch": 0.717006540977299, "grad_norm": 0.18467227223572116, "learning_rate": 3.914445607816486e-05, "loss": 0.737, "step": 3727 }, { "epoch": 0.7171989226625626, "grad_norm": 0.17671681100931255, "learning_rate": 3.9095025422100003e-05, "loss": 0.7797, "step": 3728 }, { "epoch": 0.717391304347826, "grad_norm": 0.1862661247921268, "learning_rate": 3.9045618411443374e-05, "loss": 0.7646, "step": 3729 }, { "epoch": 0.7175836860330896, "grad_norm": 0.19736314229544477, "learning_rate": 3.8996235065376354e-05, "loss": 0.8064, "step": 3730 }, { "epoch": 0.7177760677183532, "grad_norm": 0.21524260714721472, "learning_rate": 3.894687540307127e-05, "loss": 0.7931, "step": 3731 }, { "epoch": 0.7179684494036168, "grad_norm": 0.18935043257921252, "learning_rate": 3.889753944369135e-05, "loss": 0.8583, "step": 3732 }, { "epoch": 0.7181608310888803, "grad_norm": 0.1907568031994056, "learning_rate": 3.8848227206390364e-05, "loss": 0.8167, "step": 3733 }, { "epoch": 0.7183532127741439, "grad_norm": 0.19976683898369427, "learning_rate": 3.879893871031314e-05, "loss": 0.7656, "step": 3734 }, { "epoch": 0.7185455944594075, "grad_norm": 0.21033613497727244, "learning_rate": 3.8749673974595105e-05, "loss": 0.7737, "step": 3735 }, { "epoch": 0.718737976144671, "grad_norm": 0.1877458888463767, "learning_rate": 3.870043301836256e-05, "loss": 0.7858, "step": 3736 }, { "epoch": 0.7189303578299346, "grad_norm": 0.20423222868807642, "learning_rate": 3.865121586073251e-05, "loss": 0.8253, "step": 3737 }, { "epoch": 0.7191227395151981, "grad_norm": 0.2127855269421176, "learning_rate": 3.8602022520812764e-05, "loss": 0.8181, "step": 3738 }, { "epoch": 0.7193151212004617, "grad_norm": 0.19202614020315092, "learning_rate": 3.855285301770187e-05, "loss": 0.7683, "step": 3739 }, { "epoch": 0.7195075028857253, "grad_norm": 0.18737768752026238, "learning_rate": 3.850370737048913e-05, "loss": 0.7628, "step": 3740 }, { "epoch": 0.7196998845709889, "grad_norm": 0.2066131253519582, "learning_rate": 3.845458559825457e-05, "loss": 0.7739, "step": 3741 }, { "epoch": 0.7198922662562524, "grad_norm": 0.20951379395418088, "learning_rate": 3.840548772006891e-05, "loss": 0.7633, "step": 3742 }, { "epoch": 0.7200846479415159, "grad_norm": 0.18228297591125714, "learning_rate": 3.835641375499375e-05, "loss": 0.7962, "step": 3743 }, { "epoch": 0.7202770296267795, "grad_norm": 0.16691524164121285, "learning_rate": 3.830736372208118e-05, "loss": 0.8, "step": 3744 }, { "epoch": 0.7204694113120431, "grad_norm": 0.2062081258227451, "learning_rate": 3.825833764037412e-05, "loss": 0.7822, "step": 3745 }, { "epoch": 0.7206617929973067, "grad_norm": 0.21407581929009306, "learning_rate": 3.820933552890629e-05, "loss": 0.7603, "step": 3746 }, { "epoch": 0.7208541746825702, "grad_norm": 0.185585401627532, "learning_rate": 3.816035740670185e-05, "loss": 0.7691, "step": 3747 }, { "epoch": 0.7210465563678338, "grad_norm": 0.19023644843178905, "learning_rate": 3.8111403292775905e-05, "loss": 0.7805, "step": 3748 }, { "epoch": 0.7212389380530974, "grad_norm": 0.11444488005286854, "learning_rate": 3.8062473206134083e-05, "loss": 0.6585, "step": 3749 }, { "epoch": 0.7214313197383609, "grad_norm": 0.19796969498248046, "learning_rate": 3.801356716577273e-05, "loss": 0.8265, "step": 3750 }, { "epoch": 0.7216237014236244, "grad_norm": 0.19307068451107193, "learning_rate": 3.796468519067887e-05, "loss": 0.787, "step": 3751 }, { "epoch": 0.721816083108888, "grad_norm": 0.19408622935308123, "learning_rate": 3.7915827299830155e-05, "loss": 0.7779, "step": 3752 }, { "epoch": 0.7220084647941516, "grad_norm": 0.20647514515324328, "learning_rate": 3.786699351219489e-05, "loss": 0.7708, "step": 3753 }, { "epoch": 0.7222008464794152, "grad_norm": 0.21431417466173744, "learning_rate": 3.7818183846732024e-05, "loss": 0.795, "step": 3754 }, { "epoch": 0.7223932281646788, "grad_norm": 0.19647440694616472, "learning_rate": 3.776939832239125e-05, "loss": 0.663, "step": 3755 }, { "epoch": 0.7225856098499422, "grad_norm": 0.19346019575170334, "learning_rate": 3.772063695811262e-05, "loss": 0.7424, "step": 3756 }, { "epoch": 0.7227779915352058, "grad_norm": 0.1936938422440698, "learning_rate": 3.767189977282711e-05, "loss": 0.7971, "step": 3757 }, { "epoch": 0.7229703732204694, "grad_norm": 0.20629221319305732, "learning_rate": 3.762318678545616e-05, "loss": 0.7463, "step": 3758 }, { "epoch": 0.723162754905733, "grad_norm": 0.18848134544641254, "learning_rate": 3.7574498014911716e-05, "loss": 0.7639, "step": 3759 }, { "epoch": 0.7233551365909965, "grad_norm": 0.19190239180331348, "learning_rate": 3.752583348009657e-05, "loss": 0.6872, "step": 3760 }, { "epoch": 0.7235475182762601, "grad_norm": 0.18503483748513172, "learning_rate": 3.7477193199903904e-05, "loss": 0.8188, "step": 3761 }, { "epoch": 0.7237398999615237, "grad_norm": 0.18779424306703235, "learning_rate": 3.742857719321756e-05, "loss": 0.8015, "step": 3762 }, { "epoch": 0.7239322816467872, "grad_norm": 0.17347466981062948, "learning_rate": 3.737998547891195e-05, "loss": 0.7903, "step": 3763 }, { "epoch": 0.7241246633320508, "grad_norm": 0.18197886149404308, "learning_rate": 3.7331418075852053e-05, "loss": 0.8136, "step": 3764 }, { "epoch": 0.7243170450173143, "grad_norm": 0.2044139984081875, "learning_rate": 3.728287500289339e-05, "loss": 0.8167, "step": 3765 }, { "epoch": 0.7245094267025779, "grad_norm": 0.20809630371332627, "learning_rate": 3.723435627888208e-05, "loss": 0.7829, "step": 3766 }, { "epoch": 0.7247018083878415, "grad_norm": 0.19186150739518076, "learning_rate": 3.718586192265473e-05, "loss": 0.751, "step": 3767 }, { "epoch": 0.7248941900731051, "grad_norm": 0.18854044506271297, "learning_rate": 3.7137391953038514e-05, "loss": 0.844, "step": 3768 }, { "epoch": 0.7250865717583687, "grad_norm": 0.22133387872849672, "learning_rate": 3.708894638885122e-05, "loss": 0.7704, "step": 3769 }, { "epoch": 0.7252789534436321, "grad_norm": 0.20198732416729767, "learning_rate": 3.7040525248901005e-05, "loss": 0.7402, "step": 3770 }, { "epoch": 0.7254713351288957, "grad_norm": 0.19862547668981392, "learning_rate": 3.6992128551986616e-05, "loss": 0.8383, "step": 3771 }, { "epoch": 0.7256637168141593, "grad_norm": 0.2027176284176866, "learning_rate": 3.6943756316897404e-05, "loss": 0.7945, "step": 3772 }, { "epoch": 0.7258560984994229, "grad_norm": 0.1720303811387619, "learning_rate": 3.6895408562413025e-05, "loss": 0.7967, "step": 3773 }, { "epoch": 0.7260484801846864, "grad_norm": 0.19511255175993222, "learning_rate": 3.684708530730382e-05, "loss": 0.7854, "step": 3774 }, { "epoch": 0.72624086186995, "grad_norm": 0.20199586326476793, "learning_rate": 3.6798786570330525e-05, "loss": 0.7683, "step": 3775 }, { "epoch": 0.7264332435552135, "grad_norm": 0.19790952207472814, "learning_rate": 3.675051237024436e-05, "loss": 0.7807, "step": 3776 }, { "epoch": 0.7266256252404771, "grad_norm": 0.21287160523296367, "learning_rate": 3.670226272578704e-05, "loss": 0.8018, "step": 3777 }, { "epoch": 0.7268180069257407, "grad_norm": 0.20844762250446836, "learning_rate": 3.665403765569073e-05, "loss": 0.7662, "step": 3778 }, { "epoch": 0.7270103886110042, "grad_norm": 0.1874978093682268, "learning_rate": 3.660583717867807e-05, "loss": 0.8278, "step": 3779 }, { "epoch": 0.7272027702962678, "grad_norm": 0.1881694478775408, "learning_rate": 3.655766131346211e-05, "loss": 0.8504, "step": 3780 }, { "epoch": 0.7273951519815314, "grad_norm": 0.1880076769824601, "learning_rate": 3.650951007874648e-05, "loss": 0.8315, "step": 3781 }, { "epoch": 0.727587533666795, "grad_norm": 0.1925487640544741, "learning_rate": 3.646138349322501e-05, "loss": 0.7958, "step": 3782 }, { "epoch": 0.7277799153520584, "grad_norm": 0.21444586645903974, "learning_rate": 3.641328157558219e-05, "loss": 0.7526, "step": 3783 }, { "epoch": 0.727972297037322, "grad_norm": 0.19671891277172937, "learning_rate": 3.6365204344492867e-05, "loss": 0.8216, "step": 3784 }, { "epoch": 0.7281646787225856, "grad_norm": 0.20739346743732345, "learning_rate": 3.631715181862215e-05, "loss": 0.74, "step": 3785 }, { "epoch": 0.7283570604078492, "grad_norm": 0.19227614844855148, "learning_rate": 3.62691240166258e-05, "loss": 0.7688, "step": 3786 }, { "epoch": 0.7285494420931128, "grad_norm": 0.19050369048835242, "learning_rate": 3.6221120957149825e-05, "loss": 0.8268, "step": 3787 }, { "epoch": 0.7287418237783763, "grad_norm": 0.2139215437147208, "learning_rate": 3.617314265883066e-05, "loss": 0.787, "step": 3788 }, { "epoch": 0.7289342054636399, "grad_norm": 0.21746119211541248, "learning_rate": 3.6125189140295145e-05, "loss": 0.8073, "step": 3789 }, { "epoch": 0.7291265871489034, "grad_norm": 0.18505274510306421, "learning_rate": 3.607726042016049e-05, "loss": 0.8075, "step": 3790 }, { "epoch": 0.729318968834167, "grad_norm": 0.19484443037995092, "learning_rate": 3.602935651703424e-05, "loss": 0.8013, "step": 3791 }, { "epoch": 0.7295113505194305, "grad_norm": 0.20844088361825067, "learning_rate": 3.598147744951438e-05, "loss": 0.7489, "step": 3792 }, { "epoch": 0.7297037322046941, "grad_norm": 0.1943491139812831, "learning_rate": 3.5933623236189195e-05, "loss": 0.7901, "step": 3793 }, { "epoch": 0.7298961138899577, "grad_norm": 0.19733377746798703, "learning_rate": 3.58857938956373e-05, "loss": 0.8057, "step": 3794 }, { "epoch": 0.7300884955752213, "grad_norm": 0.10712748193881662, "learning_rate": 3.58379894464278e-05, "loss": 0.6046, "step": 3795 }, { "epoch": 0.7302808772604849, "grad_norm": 0.2064137352881143, "learning_rate": 3.57902099071199e-05, "loss": 0.8114, "step": 3796 }, { "epoch": 0.7304732589457483, "grad_norm": 0.22975858637039254, "learning_rate": 3.574245529626334e-05, "loss": 0.8545, "step": 3797 }, { "epoch": 0.7306656406310119, "grad_norm": 0.1896487905516682, "learning_rate": 3.569472563239814e-05, "loss": 0.7998, "step": 3798 }, { "epoch": 0.7308580223162755, "grad_norm": 0.18863586328608054, "learning_rate": 3.5647020934054464e-05, "loss": 0.7074, "step": 3799 }, { "epoch": 0.7310504040015391, "grad_norm": 0.21127095065894588, "learning_rate": 3.559934121975304e-05, "loss": 0.7401, "step": 3800 }, { "epoch": 0.7312427856868026, "grad_norm": 0.18827862682193353, "learning_rate": 3.555168650800473e-05, "loss": 0.8537, "step": 3801 }, { "epoch": 0.7314351673720662, "grad_norm": 0.20080441174673877, "learning_rate": 3.5504056817310735e-05, "loss": 0.8149, "step": 3802 }, { "epoch": 0.7316275490573297, "grad_norm": 0.20017504889686916, "learning_rate": 3.5456452166162545e-05, "loss": 0.7977, "step": 3803 }, { "epoch": 0.7318199307425933, "grad_norm": 0.2109016110439203, "learning_rate": 3.540887257304193e-05, "loss": 0.782, "step": 3804 }, { "epoch": 0.7320123124278569, "grad_norm": 0.20206440789778474, "learning_rate": 3.5361318056420925e-05, "loss": 0.8079, "step": 3805 }, { "epoch": 0.7322046941131204, "grad_norm": 0.21423325929309586, "learning_rate": 3.531378863476178e-05, "loss": 0.7951, "step": 3806 }, { "epoch": 0.732397075798384, "grad_norm": 0.11069897224406688, "learning_rate": 3.526628432651716e-05, "loss": 0.6075, "step": 3807 }, { "epoch": 0.7325894574836476, "grad_norm": 0.21264314161535214, "learning_rate": 3.5218805150129755e-05, "loss": 0.8001, "step": 3808 }, { "epoch": 0.7327818391689112, "grad_norm": 0.1965195150326286, "learning_rate": 3.51713511240327e-05, "loss": 0.7722, "step": 3809 }, { "epoch": 0.7329742208541746, "grad_norm": 0.10493225057704626, "learning_rate": 3.5123922266649303e-05, "loss": 0.6174, "step": 3810 }, { "epoch": 0.7331666025394382, "grad_norm": 0.21032950644905918, "learning_rate": 3.507651859639295e-05, "loss": 0.8037, "step": 3811 }, { "epoch": 0.7333589842247018, "grad_norm": 0.19449411425123944, "learning_rate": 3.502914013166749e-05, "loss": 0.7951, "step": 3812 }, { "epoch": 0.7335513659099654, "grad_norm": 0.20247248524755043, "learning_rate": 3.4981786890866854e-05, "loss": 0.7637, "step": 3813 }, { "epoch": 0.733743747595229, "grad_norm": 0.19672193240029576, "learning_rate": 3.493445889237518e-05, "loss": 0.7719, "step": 3814 }, { "epoch": 0.7339361292804925, "grad_norm": 0.19151607811384136, "learning_rate": 3.488715615456685e-05, "loss": 0.7871, "step": 3815 }, { "epoch": 0.734128510965756, "grad_norm": 0.178884291238906, "learning_rate": 3.483987869580638e-05, "loss": 0.8087, "step": 3816 }, { "epoch": 0.7343208926510196, "grad_norm": 0.20556658917440826, "learning_rate": 3.4792626534448544e-05, "loss": 0.8043, "step": 3817 }, { "epoch": 0.7345132743362832, "grad_norm": 0.1952805467285591, "learning_rate": 3.474539968883824e-05, "loss": 0.8498, "step": 3818 }, { "epoch": 0.7347056560215467, "grad_norm": 0.19228062128952442, "learning_rate": 3.469819817731056e-05, "loss": 0.8084, "step": 3819 }, { "epoch": 0.7348980377068103, "grad_norm": 0.18041788466689287, "learning_rate": 3.4651022018190715e-05, "loss": 0.7664, "step": 3820 }, { "epoch": 0.7350904193920739, "grad_norm": 0.19610228010348363, "learning_rate": 3.460387122979423e-05, "loss": 0.7846, "step": 3821 }, { "epoch": 0.7352828010773375, "grad_norm": 0.17959968542285118, "learning_rate": 3.455674583042652e-05, "loss": 0.7836, "step": 3822 }, { "epoch": 0.735475182762601, "grad_norm": 0.10692488448855433, "learning_rate": 3.450964583838339e-05, "loss": 0.6524, "step": 3823 }, { "epoch": 0.7356675644478645, "grad_norm": 0.19646188319711722, "learning_rate": 3.446257127195067e-05, "loss": 0.7729, "step": 3824 }, { "epoch": 0.7358599461331281, "grad_norm": 0.20563181948301168, "learning_rate": 3.4415522149404235e-05, "loss": 0.7958, "step": 3825 }, { "epoch": 0.7360523278183917, "grad_norm": 0.17662404825460246, "learning_rate": 3.4368498489010283e-05, "loss": 0.8099, "step": 3826 }, { "epoch": 0.7362447095036553, "grad_norm": 0.18584842516801625, "learning_rate": 3.432150030902497e-05, "loss": 0.8237, "step": 3827 }, { "epoch": 0.7364370911889189, "grad_norm": 0.18175235252334127, "learning_rate": 3.427452762769462e-05, "loss": 0.7974, "step": 3828 }, { "epoch": 0.7366294728741823, "grad_norm": 0.21162416903774423, "learning_rate": 3.422758046325562e-05, "loss": 0.7854, "step": 3829 }, { "epoch": 0.7368218545594459, "grad_norm": 0.12302844779880338, "learning_rate": 3.418065883393452e-05, "loss": 0.6633, "step": 3830 }, { "epoch": 0.7370142362447095, "grad_norm": 0.20389217394404918, "learning_rate": 3.4133762757947874e-05, "loss": 0.7948, "step": 3831 }, { "epoch": 0.7372066179299731, "grad_norm": 0.18203064742628727, "learning_rate": 3.408689225350234e-05, "loss": 0.7912, "step": 3832 }, { "epoch": 0.7373989996152366, "grad_norm": 0.19265830593140912, "learning_rate": 3.4040047338794756e-05, "loss": 0.7263, "step": 3833 }, { "epoch": 0.7375913813005002, "grad_norm": 0.19282324649033988, "learning_rate": 3.399322803201178e-05, "loss": 0.7715, "step": 3834 }, { "epoch": 0.7377837629857638, "grad_norm": 0.18296348479138275, "learning_rate": 3.394643435133041e-05, "loss": 0.756, "step": 3835 }, { "epoch": 0.7379761446710273, "grad_norm": 0.1132889707669432, "learning_rate": 3.389966631491751e-05, "loss": 0.6484, "step": 3836 }, { "epoch": 0.7381685263562909, "grad_norm": 0.18630135231133843, "learning_rate": 3.385292394093006e-05, "loss": 0.7997, "step": 3837 }, { "epoch": 0.7383609080415544, "grad_norm": 0.17716420625503046, "learning_rate": 3.380620724751506e-05, "loss": 0.796, "step": 3838 }, { "epoch": 0.738553289726818, "grad_norm": 0.212380343712034, "learning_rate": 3.3759516252809477e-05, "loss": 0.7731, "step": 3839 }, { "epoch": 0.7387456714120816, "grad_norm": 0.22060170420317832, "learning_rate": 3.3712850974940435e-05, "loss": 0.7738, "step": 3840 }, { "epoch": 0.7389380530973452, "grad_norm": 0.19577909848275146, "learning_rate": 3.366621143202498e-05, "loss": 0.7563, "step": 3841 }, { "epoch": 0.7391304347826086, "grad_norm": 0.20564419156592087, "learning_rate": 3.361959764217017e-05, "loss": 0.815, "step": 3842 }, { "epoch": 0.7393228164678722, "grad_norm": 0.18757969374080957, "learning_rate": 3.3573009623473126e-05, "loss": 0.8443, "step": 3843 }, { "epoch": 0.7395151981531358, "grad_norm": 0.17763059869110723, "learning_rate": 3.352644739402089e-05, "loss": 0.7973, "step": 3844 }, { "epoch": 0.7397075798383994, "grad_norm": 0.1819793292926212, "learning_rate": 3.3479910971890515e-05, "loss": 0.8017, "step": 3845 }, { "epoch": 0.739899961523663, "grad_norm": 0.20285831057703818, "learning_rate": 3.3433400375149026e-05, "loss": 0.8125, "step": 3846 }, { "epoch": 0.7400923432089265, "grad_norm": 0.19150849243896922, "learning_rate": 3.338691562185353e-05, "loss": 0.8024, "step": 3847 }, { "epoch": 0.7402847248941901, "grad_norm": 0.21261073890686674, "learning_rate": 3.3340456730050884e-05, "loss": 0.8334, "step": 3848 }, { "epoch": 0.7404771065794536, "grad_norm": 0.20638729049831175, "learning_rate": 3.3294023717778125e-05, "loss": 0.7942, "step": 3849 }, { "epoch": 0.7406694882647172, "grad_norm": 0.21211936849527324, "learning_rate": 3.324761660306215e-05, "loss": 0.8266, "step": 3850 }, { "epoch": 0.7408618699499807, "grad_norm": 0.2100439328335658, "learning_rate": 3.320123540391968e-05, "loss": 0.7584, "step": 3851 }, { "epoch": 0.7410542516352443, "grad_norm": 0.1229532229784527, "learning_rate": 3.315488013835762e-05, "loss": 0.6591, "step": 3852 }, { "epoch": 0.7412466333205079, "grad_norm": 0.18826042816831695, "learning_rate": 3.310855082437263e-05, "loss": 0.7776, "step": 3853 }, { "epoch": 0.7414390150057715, "grad_norm": 0.21097450189204595, "learning_rate": 3.306224747995136e-05, "loss": 0.7764, "step": 3854 }, { "epoch": 0.7416313966910351, "grad_norm": 0.2040082407989687, "learning_rate": 3.301597012307034e-05, "loss": 0.7073, "step": 3855 }, { "epoch": 0.7418237783762985, "grad_norm": 0.2173724480997047, "learning_rate": 3.296971877169605e-05, "loss": 0.8136, "step": 3856 }, { "epoch": 0.7420161600615621, "grad_norm": 0.19866075279809983, "learning_rate": 3.292349344378486e-05, "loss": 0.7279, "step": 3857 }, { "epoch": 0.7422085417468257, "grad_norm": 0.21218599588659262, "learning_rate": 3.287729415728298e-05, "loss": 0.7671, "step": 3858 }, { "epoch": 0.7424009234320893, "grad_norm": 0.19010992808947164, "learning_rate": 3.283112093012669e-05, "loss": 0.7803, "step": 3859 }, { "epoch": 0.7425933051173528, "grad_norm": 0.2151205534753765, "learning_rate": 3.278497378024187e-05, "loss": 0.7449, "step": 3860 }, { "epoch": 0.7427856868026164, "grad_norm": 0.3168557936668191, "learning_rate": 3.2738852725544544e-05, "loss": 0.8327, "step": 3861 }, { "epoch": 0.74297806848788, "grad_norm": 0.18576188911554886, "learning_rate": 3.269275778394047e-05, "loss": 0.77, "step": 3862 }, { "epoch": 0.7431704501731435, "grad_norm": 0.18953853241599455, "learning_rate": 3.264668897332527e-05, "loss": 0.7768, "step": 3863 }, { "epoch": 0.7433628318584071, "grad_norm": 0.1922791147220438, "learning_rate": 3.260064631158449e-05, "loss": 0.7606, "step": 3864 }, { "epoch": 0.7435552135436706, "grad_norm": 0.19013109090366329, "learning_rate": 3.2554629816593376e-05, "loss": 0.8396, "step": 3865 }, { "epoch": 0.7437475952289342, "grad_norm": 0.20305090726301972, "learning_rate": 3.250863950621721e-05, "loss": 0.7147, "step": 3866 }, { "epoch": 0.7439399769141978, "grad_norm": 0.19570919432267941, "learning_rate": 3.2462675398310984e-05, "loss": 0.7338, "step": 3867 }, { "epoch": 0.7441323585994614, "grad_norm": 0.19706006058062447, "learning_rate": 3.2416737510719544e-05, "loss": 0.7546, "step": 3868 }, { "epoch": 0.7443247402847248, "grad_norm": 0.18423407957456503, "learning_rate": 3.237082586127757e-05, "loss": 0.7638, "step": 3869 }, { "epoch": 0.7445171219699884, "grad_norm": 0.20266528577094975, "learning_rate": 3.232494046780953e-05, "loss": 0.787, "step": 3870 }, { "epoch": 0.744709503655252, "grad_norm": 0.2052668199798755, "learning_rate": 3.227908134812972e-05, "loss": 0.782, "step": 3871 }, { "epoch": 0.7449018853405156, "grad_norm": 0.18548268171010968, "learning_rate": 3.223324852004219e-05, "loss": 0.7398, "step": 3872 }, { "epoch": 0.7450942670257792, "grad_norm": 0.193415481439763, "learning_rate": 3.218744200134094e-05, "loss": 0.8054, "step": 3873 }, { "epoch": 0.7452866487110427, "grad_norm": 0.2174075290385756, "learning_rate": 3.21416618098095e-05, "loss": 0.8097, "step": 3874 }, { "epoch": 0.7454790303963063, "grad_norm": 0.21831328875091885, "learning_rate": 3.2095907963221396e-05, "loss": 0.7652, "step": 3875 }, { "epoch": 0.7456714120815698, "grad_norm": 0.11306384305685255, "learning_rate": 3.205018047933987e-05, "loss": 0.6476, "step": 3876 }, { "epoch": 0.7458637937668334, "grad_norm": 0.19673154875017484, "learning_rate": 3.200447937591779e-05, "loss": 0.7754, "step": 3877 }, { "epoch": 0.746056175452097, "grad_norm": 0.10893292010180523, "learning_rate": 3.195880467069801e-05, "loss": 0.6331, "step": 3878 }, { "epoch": 0.7462485571373605, "grad_norm": 0.21198357856544722, "learning_rate": 3.191315638141297e-05, "loss": 0.7876, "step": 3879 }, { "epoch": 0.7464409388226241, "grad_norm": 0.2517492999130788, "learning_rate": 3.1867534525784934e-05, "loss": 0.7268, "step": 3880 }, { "epoch": 0.7466333205078877, "grad_norm": 0.21066591982584715, "learning_rate": 3.182193912152586e-05, "loss": 0.7832, "step": 3881 }, { "epoch": 0.7468257021931513, "grad_norm": 0.20096259924436438, "learning_rate": 3.177637018633746e-05, "loss": 0.7891, "step": 3882 }, { "epoch": 0.7470180838784147, "grad_norm": 0.2061816286986062, "learning_rate": 3.173082773791116e-05, "loss": 0.7667, "step": 3883 }, { "epoch": 0.7472104655636783, "grad_norm": 0.18601964195543746, "learning_rate": 3.168531179392808e-05, "loss": 0.7917, "step": 3884 }, { "epoch": 0.7474028472489419, "grad_norm": 0.20619771589272382, "learning_rate": 3.163982237205917e-05, "loss": 0.7064, "step": 3885 }, { "epoch": 0.7475952289342055, "grad_norm": 0.19086151696273856, "learning_rate": 3.159435948996485e-05, "loss": 0.8218, "step": 3886 }, { "epoch": 0.7477876106194691, "grad_norm": 0.17634637079160673, "learning_rate": 3.154892316529549e-05, "loss": 0.7512, "step": 3887 }, { "epoch": 0.7479799923047326, "grad_norm": 0.19865102194217868, "learning_rate": 3.150351341569101e-05, "loss": 0.7784, "step": 3888 }, { "epoch": 0.7481723739899961, "grad_norm": 0.1965737385835937, "learning_rate": 3.1458130258781005e-05, "loss": 0.8456, "step": 3889 }, { "epoch": 0.7483647556752597, "grad_norm": 0.22473608339742357, "learning_rate": 3.141277371218484e-05, "loss": 0.7779, "step": 3890 }, { "epoch": 0.7485571373605233, "grad_norm": 0.2061045433000486, "learning_rate": 3.1367443793511386e-05, "loss": 0.7713, "step": 3891 }, { "epoch": 0.7487495190457868, "grad_norm": 0.11580087482992106, "learning_rate": 3.132214052035937e-05, "loss": 0.6822, "step": 3892 }, { "epoch": 0.7489419007310504, "grad_norm": 0.19643136916014217, "learning_rate": 3.127686391031705e-05, "loss": 0.8001, "step": 3893 }, { "epoch": 0.749134282416314, "grad_norm": 0.1739315783349164, "learning_rate": 3.123161398096237e-05, "loss": 0.7616, "step": 3894 }, { "epoch": 0.7493266641015776, "grad_norm": 0.18827225734700087, "learning_rate": 3.11863907498629e-05, "loss": 0.8026, "step": 3895 }, { "epoch": 0.7495190457868411, "grad_norm": 0.23139645600078343, "learning_rate": 3.114119423457588e-05, "loss": 0.7692, "step": 3896 }, { "epoch": 0.7497114274721046, "grad_norm": 0.19980361466781763, "learning_rate": 3.109602445264812e-05, "loss": 0.7338, "step": 3897 }, { "epoch": 0.7499038091573682, "grad_norm": 0.1989077166034427, "learning_rate": 3.1050881421616076e-05, "loss": 0.8093, "step": 3898 }, { "epoch": 0.7500961908426318, "grad_norm": 0.18754247215872993, "learning_rate": 3.1005765159005904e-05, "loss": 0.7457, "step": 3899 }, { "epoch": 0.7502885725278954, "grad_norm": 0.19962018823357236, "learning_rate": 3.0960675682333186e-05, "loss": 0.8188, "step": 3900 }, { "epoch": 0.7504809542131589, "grad_norm": 0.2045593028531866, "learning_rate": 3.0915613009103295e-05, "loss": 0.798, "step": 3901 }, { "epoch": 0.7506733358984224, "grad_norm": 0.2059534627437552, "learning_rate": 3.087057715681107e-05, "loss": 0.7403, "step": 3902 }, { "epoch": 0.750865717583686, "grad_norm": 0.2124315333422216, "learning_rate": 3.0825568142940995e-05, "loss": 0.7748, "step": 3903 }, { "epoch": 0.7510580992689496, "grad_norm": 0.21763204873646477, "learning_rate": 3.078058598496711e-05, "loss": 0.7893, "step": 3904 }, { "epoch": 0.7512504809542132, "grad_norm": 0.19448417833337853, "learning_rate": 3.0735630700353044e-05, "loss": 0.8127, "step": 3905 }, { "epoch": 0.7514428626394767, "grad_norm": 0.20654191809659248, "learning_rate": 3.069070230655198e-05, "loss": 0.7928, "step": 3906 }, { "epoch": 0.7516352443247403, "grad_norm": 0.18575762315685826, "learning_rate": 3.0645800821006665e-05, "loss": 0.7613, "step": 3907 }, { "epoch": 0.7518276260100039, "grad_norm": 0.2031262068635905, "learning_rate": 3.060092626114941e-05, "loss": 0.7469, "step": 3908 }, { "epoch": 0.7520200076952674, "grad_norm": 0.17987320268809553, "learning_rate": 3.055607864440206e-05, "loss": 0.7825, "step": 3909 }, { "epoch": 0.7522123893805309, "grad_norm": 0.2007302897159342, "learning_rate": 3.0511257988175978e-05, "loss": 0.7476, "step": 3910 }, { "epoch": 0.7524047710657945, "grad_norm": 0.209338026381964, "learning_rate": 3.0466464309872166e-05, "loss": 0.7449, "step": 3911 }, { "epoch": 0.7525971527510581, "grad_norm": 0.2407845495980447, "learning_rate": 3.0421697626880962e-05, "loss": 0.7287, "step": 3912 }, { "epoch": 0.7527895344363217, "grad_norm": 0.18749723845381389, "learning_rate": 3.0376957956582452e-05, "loss": 0.8103, "step": 3913 }, { "epoch": 0.7529819161215853, "grad_norm": 0.19201249042103716, "learning_rate": 3.0332245316345996e-05, "loss": 0.8379, "step": 3914 }, { "epoch": 0.7531742978068487, "grad_norm": 0.18961993939913183, "learning_rate": 3.0287559723530667e-05, "loss": 0.8388, "step": 3915 }, { "epoch": 0.7533666794921123, "grad_norm": 0.18880571940712412, "learning_rate": 3.0242901195484953e-05, "loss": 0.7858, "step": 3916 }, { "epoch": 0.7535590611773759, "grad_norm": 0.18593207751174629, "learning_rate": 3.0198269749546737e-05, "loss": 0.7451, "step": 3917 }, { "epoch": 0.7537514428626395, "grad_norm": 0.1696989334215181, "learning_rate": 3.0153665403043586e-05, "loss": 0.7585, "step": 3918 }, { "epoch": 0.753943824547903, "grad_norm": 0.20189806565069995, "learning_rate": 3.01090881732924e-05, "loss": 0.7864, "step": 3919 }, { "epoch": 0.7541362062331666, "grad_norm": 0.20634170871230934, "learning_rate": 3.0064538077599603e-05, "loss": 0.7628, "step": 3920 }, { "epoch": 0.7543285879184302, "grad_norm": 0.18845707046847746, "learning_rate": 3.002001513326107e-05, "loss": 0.7267, "step": 3921 }, { "epoch": 0.7545209696036937, "grad_norm": 0.21121152710026658, "learning_rate": 2.9975519357562155e-05, "loss": 0.7705, "step": 3922 }, { "epoch": 0.7547133512889573, "grad_norm": 0.18915530010243056, "learning_rate": 2.9931050767777625e-05, "loss": 0.7918, "step": 3923 }, { "epoch": 0.7549057329742208, "grad_norm": 0.1833961832783914, "learning_rate": 2.9886609381171705e-05, "loss": 0.6975, "step": 3924 }, { "epoch": 0.7550981146594844, "grad_norm": 0.19540261214655774, "learning_rate": 2.9842195214998158e-05, "loss": 0.7401, "step": 3925 }, { "epoch": 0.755290496344748, "grad_norm": 0.19400838065456777, "learning_rate": 2.9797808286499973e-05, "loss": 0.7526, "step": 3926 }, { "epoch": 0.7554828780300116, "grad_norm": 0.11040460552761384, "learning_rate": 2.9753448612909775e-05, "loss": 0.5887, "step": 3927 }, { "epoch": 0.7556752597152752, "grad_norm": 0.18922191365062274, "learning_rate": 2.9709116211449483e-05, "loss": 0.7564, "step": 3928 }, { "epoch": 0.7558676414005386, "grad_norm": 0.17905719462616768, "learning_rate": 2.966481109933047e-05, "loss": 0.7953, "step": 3929 }, { "epoch": 0.7560600230858022, "grad_norm": 0.20314394780386721, "learning_rate": 2.9620533293753495e-05, "loss": 0.7536, "step": 3930 }, { "epoch": 0.7562524047710658, "grad_norm": 0.19753246801184843, "learning_rate": 2.9576282811908728e-05, "loss": 0.7387, "step": 3931 }, { "epoch": 0.7564447864563294, "grad_norm": 0.20120918649166095, "learning_rate": 2.9532059670975732e-05, "loss": 0.8044, "step": 3932 }, { "epoch": 0.7566371681415929, "grad_norm": 0.20771548412947374, "learning_rate": 2.9487863888123456e-05, "loss": 0.7541, "step": 3933 }, { "epoch": 0.7568295498268565, "grad_norm": 0.2014999524567445, "learning_rate": 2.9443695480510224e-05, "loss": 0.7532, "step": 3934 }, { "epoch": 0.75702193151212, "grad_norm": 0.10634214474708786, "learning_rate": 2.939955446528374e-05, "loss": 0.5709, "step": 3935 }, { "epoch": 0.7572143131973836, "grad_norm": 0.2284764112489402, "learning_rate": 2.9355440859581018e-05, "loss": 0.7625, "step": 3936 }, { "epoch": 0.7574066948826472, "grad_norm": 0.17883355677814913, "learning_rate": 2.931135468052858e-05, "loss": 0.808, "step": 3937 }, { "epoch": 0.7575990765679107, "grad_norm": 0.21375977590904144, "learning_rate": 2.9267295945242067e-05, "loss": 0.7954, "step": 3938 }, { "epoch": 0.7577914582531743, "grad_norm": 0.18584901476788973, "learning_rate": 2.9223264670826743e-05, "loss": 0.81, "step": 3939 }, { "epoch": 0.7579838399384379, "grad_norm": 0.18527185193054266, "learning_rate": 2.9179260874376913e-05, "loss": 0.7913, "step": 3940 }, { "epoch": 0.7581762216237015, "grad_norm": 0.21538318041976942, "learning_rate": 2.9135284572976486e-05, "loss": 0.7571, "step": 3941 }, { "epoch": 0.7583686033089649, "grad_norm": 0.20684786210167375, "learning_rate": 2.9091335783698515e-05, "loss": 0.7519, "step": 3942 }, { "epoch": 0.7585609849942285, "grad_norm": 0.18630698601852505, "learning_rate": 2.9047414523605465e-05, "loss": 0.7805, "step": 3943 }, { "epoch": 0.7587533666794921, "grad_norm": 0.19010197643513205, "learning_rate": 2.9003520809749053e-05, "loss": 0.7697, "step": 3944 }, { "epoch": 0.7589457483647557, "grad_norm": 0.19603296554684435, "learning_rate": 2.8959654659170354e-05, "loss": 0.8149, "step": 3945 }, { "epoch": 0.7591381300500193, "grad_norm": 0.1894514229874989, "learning_rate": 2.8915816088899695e-05, "loss": 0.8043, "step": 3946 }, { "epoch": 0.7593305117352828, "grad_norm": 0.17816571741162027, "learning_rate": 2.8872005115956745e-05, "loss": 0.7919, "step": 3947 }, { "epoch": 0.7595228934205464, "grad_norm": 0.10224257827038509, "learning_rate": 2.8828221757350404e-05, "loss": 0.6076, "step": 3948 }, { "epoch": 0.7597152751058099, "grad_norm": 0.19656095599067938, "learning_rate": 2.8784466030078905e-05, "loss": 0.7648, "step": 3949 }, { "epoch": 0.7599076567910735, "grad_norm": 0.20297767954947257, "learning_rate": 2.874073795112967e-05, "loss": 0.7439, "step": 3950 }, { "epoch": 0.760100038476337, "grad_norm": 0.2070667551025449, "learning_rate": 2.8697037537479564e-05, "loss": 0.7934, "step": 3951 }, { "epoch": 0.7602924201616006, "grad_norm": 0.10412846331906143, "learning_rate": 2.8653364806094453e-05, "loss": 0.6119, "step": 3952 }, { "epoch": 0.7604848018468642, "grad_norm": 0.18629058062486478, "learning_rate": 2.86097197739297e-05, "loss": 0.8278, "step": 3953 }, { "epoch": 0.7606771835321278, "grad_norm": 0.1832603916456802, "learning_rate": 2.8566102457929755e-05, "loss": 0.7988, "step": 3954 }, { "epoch": 0.7608695652173914, "grad_norm": 0.20054431667404105, "learning_rate": 2.8522512875028395e-05, "loss": 0.8014, "step": 3955 }, { "epoch": 0.7610619469026548, "grad_norm": 0.1923080819140142, "learning_rate": 2.847895104214856e-05, "loss": 0.7829, "step": 3956 }, { "epoch": 0.7612543285879184, "grad_norm": 0.19847325196365223, "learning_rate": 2.843541697620249e-05, "loss": 0.7473, "step": 3957 }, { "epoch": 0.761446710273182, "grad_norm": 0.20496491733005504, "learning_rate": 2.8391910694091583e-05, "loss": 0.7357, "step": 3958 }, { "epoch": 0.7616390919584456, "grad_norm": 0.19584023511634957, "learning_rate": 2.8348432212706445e-05, "loss": 0.7862, "step": 3959 }, { "epoch": 0.7618314736437091, "grad_norm": 0.20256078009727738, "learning_rate": 2.8304981548927022e-05, "loss": 0.736, "step": 3960 }, { "epoch": 0.7620238553289727, "grad_norm": 0.2183386413690921, "learning_rate": 2.826155871962227e-05, "loss": 0.8338, "step": 3961 }, { "epoch": 0.7622162370142362, "grad_norm": 0.20075851661262495, "learning_rate": 2.8218163741650416e-05, "loss": 0.825, "step": 3962 }, { "epoch": 0.7624086186994998, "grad_norm": 0.20425977981524843, "learning_rate": 2.8174796631858978e-05, "loss": 0.7878, "step": 3963 }, { "epoch": 0.7626010003847634, "grad_norm": 0.20507031064883716, "learning_rate": 2.813145740708445e-05, "loss": 0.7386, "step": 3964 }, { "epoch": 0.7627933820700269, "grad_norm": 0.19600655674497353, "learning_rate": 2.8088146084152712e-05, "loss": 0.8422, "step": 3965 }, { "epoch": 0.7629857637552905, "grad_norm": 0.17953776327988824, "learning_rate": 2.8044862679878604e-05, "loss": 0.8423, "step": 3966 }, { "epoch": 0.7631781454405541, "grad_norm": 0.207442900489436, "learning_rate": 2.800160721106633e-05, "loss": 0.7435, "step": 3967 }, { "epoch": 0.7633705271258177, "grad_norm": 0.22793376471659088, "learning_rate": 2.7958379694509106e-05, "loss": 0.7958, "step": 3968 }, { "epoch": 0.7635629088110811, "grad_norm": 0.1968359066066432, "learning_rate": 2.791518014698935e-05, "loss": 0.7786, "step": 3969 }, { "epoch": 0.7637552904963447, "grad_norm": 0.2197322000521214, "learning_rate": 2.787200858527862e-05, "loss": 0.7608, "step": 3970 }, { "epoch": 0.7639476721816083, "grad_norm": 0.11421539760987512, "learning_rate": 2.7828865026137585e-05, "loss": 0.5918, "step": 3971 }, { "epoch": 0.7641400538668719, "grad_norm": 0.1973330984421332, "learning_rate": 2.778574948631608e-05, "loss": 0.7789, "step": 3972 }, { "epoch": 0.7643324355521355, "grad_norm": 0.18006560041024014, "learning_rate": 2.774266198255303e-05, "loss": 0.7662, "step": 3973 }, { "epoch": 0.764524817237399, "grad_norm": 0.20257636953811028, "learning_rate": 2.7699602531576496e-05, "loss": 0.7785, "step": 3974 }, { "epoch": 0.7647171989226625, "grad_norm": 0.19078210744672794, "learning_rate": 2.765657115010364e-05, "loss": 0.773, "step": 3975 }, { "epoch": 0.7649095806079261, "grad_norm": 0.2033578048599448, "learning_rate": 2.7613567854840683e-05, "loss": 0.8073, "step": 3976 }, { "epoch": 0.7651019622931897, "grad_norm": 0.17195273925710697, "learning_rate": 2.7570592662483087e-05, "loss": 0.8035, "step": 3977 }, { "epoch": 0.7652943439784533, "grad_norm": 0.18861641316196034, "learning_rate": 2.7527645589715167e-05, "loss": 0.8365, "step": 3978 }, { "epoch": 0.7654867256637168, "grad_norm": 0.1912012867981506, "learning_rate": 2.7484726653210558e-05, "loss": 0.8494, "step": 3979 }, { "epoch": 0.7656791073489804, "grad_norm": 0.17506093984150525, "learning_rate": 2.744183586963185e-05, "loss": 0.7489, "step": 3980 }, { "epoch": 0.765871489034244, "grad_norm": 0.21453507004954864, "learning_rate": 2.739897325563069e-05, "loss": 0.7942, "step": 3981 }, { "epoch": 0.7660638707195075, "grad_norm": 0.19715538012067907, "learning_rate": 2.7356138827847855e-05, "loss": 0.7969, "step": 3982 }, { "epoch": 0.766256252404771, "grad_norm": 0.1927591149585348, "learning_rate": 2.731333260291311e-05, "loss": 0.7934, "step": 3983 }, { "epoch": 0.7664486340900346, "grad_norm": 0.1859372029578872, "learning_rate": 2.7270554597445342e-05, "loss": 0.7986, "step": 3984 }, { "epoch": 0.7666410157752982, "grad_norm": 0.19173908179735294, "learning_rate": 2.7227804828052384e-05, "loss": 0.7362, "step": 3985 }, { "epoch": 0.7668333974605618, "grad_norm": 0.19906745820967423, "learning_rate": 2.718508331133128e-05, "loss": 0.7945, "step": 3986 }, { "epoch": 0.7670257791458254, "grad_norm": 0.2108390596550865, "learning_rate": 2.7142390063867896e-05, "loss": 0.8173, "step": 3987 }, { "epoch": 0.7672181608310888, "grad_norm": 0.10914222888183624, "learning_rate": 2.709972510223725e-05, "loss": 0.6211, "step": 3988 }, { "epoch": 0.7674105425163524, "grad_norm": 0.2065137654681762, "learning_rate": 2.7057088443003343e-05, "loss": 0.8189, "step": 3989 }, { "epoch": 0.767602924201616, "grad_norm": 0.18629346387793003, "learning_rate": 2.7014480102719176e-05, "loss": 0.7764, "step": 3990 }, { "epoch": 0.7677953058868796, "grad_norm": 0.20122422029762466, "learning_rate": 2.697190009792685e-05, "loss": 0.8257, "step": 3991 }, { "epoch": 0.7679876875721431, "grad_norm": 0.21611890736732894, "learning_rate": 2.692934844515729e-05, "loss": 0.796, "step": 3992 }, { "epoch": 0.7681800692574067, "grad_norm": 0.21230250248499466, "learning_rate": 2.6886825160930585e-05, "loss": 0.8072, "step": 3993 }, { "epoch": 0.7683724509426703, "grad_norm": 0.18022570103151792, "learning_rate": 2.6844330261755714e-05, "loss": 0.7975, "step": 3994 }, { "epoch": 0.7685648326279338, "grad_norm": 0.20954199683057748, "learning_rate": 2.6801863764130652e-05, "loss": 0.767, "step": 3995 }, { "epoch": 0.7687572143131974, "grad_norm": 0.1866575874645636, "learning_rate": 2.675942568454236e-05, "loss": 0.7756, "step": 3996 }, { "epoch": 0.7689495959984609, "grad_norm": 0.11025558502316174, "learning_rate": 2.671701603946678e-05, "loss": 0.6165, "step": 3997 }, { "epoch": 0.7691419776837245, "grad_norm": 0.18680635108864319, "learning_rate": 2.667463484536876e-05, "loss": 0.7571, "step": 3998 }, { "epoch": 0.7693343593689881, "grad_norm": 0.19892027448369476, "learning_rate": 2.6632282118702147e-05, "loss": 0.8186, "step": 3999 }, { "epoch": 0.7695267410542517, "grad_norm": 0.18802154886570535, "learning_rate": 2.6589957875909798e-05, "loss": 0.808, "step": 4000 }, { "epoch": 0.7697191227395151, "grad_norm": 0.20516401837572704, "learning_rate": 2.6547662133423345e-05, "loss": 0.8058, "step": 4001 }, { "epoch": 0.7699115044247787, "grad_norm": 0.18370473166195214, "learning_rate": 2.650539490766346e-05, "loss": 0.7596, "step": 4002 }, { "epoch": 0.7701038861100423, "grad_norm": 0.18453025785466406, "learning_rate": 2.646315621503983e-05, "loss": 0.7792, "step": 4003 }, { "epoch": 0.7702962677953059, "grad_norm": 0.19621684580519586, "learning_rate": 2.6420946071950847e-05, "loss": 0.7859, "step": 4004 }, { "epoch": 0.7704886494805695, "grad_norm": 0.20785824121672877, "learning_rate": 2.6378764494784024e-05, "loss": 0.8201, "step": 4005 }, { "epoch": 0.770681031165833, "grad_norm": 0.1770705447554243, "learning_rate": 2.6336611499915686e-05, "loss": 0.798, "step": 4006 }, { "epoch": 0.7708734128510966, "grad_norm": 0.18804028382885574, "learning_rate": 2.6294487103711065e-05, "loss": 0.7853, "step": 4007 }, { "epoch": 0.7710657945363601, "grad_norm": 0.18714162981895302, "learning_rate": 2.6252391322524293e-05, "loss": 0.7724, "step": 4008 }, { "epoch": 0.7712581762216237, "grad_norm": 0.18666274448452985, "learning_rate": 2.621032417269843e-05, "loss": 0.8205, "step": 4009 }, { "epoch": 0.7714505579068872, "grad_norm": 0.18571075528048706, "learning_rate": 2.6168285670565373e-05, "loss": 0.8226, "step": 4010 }, { "epoch": 0.7716429395921508, "grad_norm": 0.11045276224411518, "learning_rate": 2.612627583244589e-05, "loss": 0.651, "step": 4011 }, { "epoch": 0.7718353212774144, "grad_norm": 0.19219839865167287, "learning_rate": 2.6084294674649735e-05, "loss": 0.8457, "step": 4012 }, { "epoch": 0.772027702962678, "grad_norm": 0.20687251894038394, "learning_rate": 2.6042342213475347e-05, "loss": 0.7848, "step": 4013 }, { "epoch": 0.7722200846479416, "grad_norm": 0.1823810599783604, "learning_rate": 2.600041846521014e-05, "loss": 0.7678, "step": 4014 }, { "epoch": 0.772412466333205, "grad_norm": 0.19172855377872833, "learning_rate": 2.5958523446130378e-05, "loss": 0.7638, "step": 4015 }, { "epoch": 0.7726048480184686, "grad_norm": 0.17797028063385495, "learning_rate": 2.5916657172501103e-05, "loss": 0.7443, "step": 4016 }, { "epoch": 0.7727972297037322, "grad_norm": 0.17544961569263337, "learning_rate": 2.5874819660576334e-05, "loss": 0.6645, "step": 4017 }, { "epoch": 0.7729896113889958, "grad_norm": 0.23287079259680196, "learning_rate": 2.5833010926598722e-05, "loss": 0.8138, "step": 4018 }, { "epoch": 0.7731819930742593, "grad_norm": 0.19590455154585523, "learning_rate": 2.579123098679994e-05, "loss": 0.7418, "step": 4019 }, { "epoch": 0.7733743747595229, "grad_norm": 0.21711014084285685, "learning_rate": 2.5749479857400384e-05, "loss": 0.7967, "step": 4020 }, { "epoch": 0.7735667564447865, "grad_norm": 0.21085017048029445, "learning_rate": 2.5707757554609246e-05, "loss": 0.7575, "step": 4021 }, { "epoch": 0.77375913813005, "grad_norm": 0.17894885473178748, "learning_rate": 2.5666064094624598e-05, "loss": 0.8004, "step": 4022 }, { "epoch": 0.7739515198153136, "grad_norm": 0.10330429604828334, "learning_rate": 2.5624399493633257e-05, "loss": 0.624, "step": 4023 }, { "epoch": 0.7741439015005771, "grad_norm": 0.1872237207231097, "learning_rate": 2.5582763767810857e-05, "loss": 0.7682, "step": 4024 }, { "epoch": 0.7743362831858407, "grad_norm": 0.18469320269156067, "learning_rate": 2.5541156933321798e-05, "loss": 0.8224, "step": 4025 }, { "epoch": 0.7745286648711043, "grad_norm": 0.17328528515487232, "learning_rate": 2.5499579006319362e-05, "loss": 0.7729, "step": 4026 }, { "epoch": 0.7747210465563679, "grad_norm": 0.1968805827095711, "learning_rate": 2.5458030002945453e-05, "loss": 0.7855, "step": 4027 }, { "epoch": 0.7749134282416313, "grad_norm": 0.18751733463292064, "learning_rate": 2.5416509939330835e-05, "loss": 0.7725, "step": 4028 }, { "epoch": 0.7751058099268949, "grad_norm": 0.21651737536312993, "learning_rate": 2.5375018831595087e-05, "loss": 0.8126, "step": 4029 }, { "epoch": 0.7752981916121585, "grad_norm": 0.19588210412543594, "learning_rate": 2.5333556695846383e-05, "loss": 0.7817, "step": 4030 }, { "epoch": 0.7754905732974221, "grad_norm": 0.16812545263401835, "learning_rate": 2.5292123548181845e-05, "loss": 0.7837, "step": 4031 }, { "epoch": 0.7756829549826857, "grad_norm": 0.19721193797651965, "learning_rate": 2.5250719404687216e-05, "loss": 0.8268, "step": 4032 }, { "epoch": 0.7758753366679492, "grad_norm": 0.20495662833246198, "learning_rate": 2.520934428143701e-05, "loss": 0.823, "step": 4033 }, { "epoch": 0.7760677183532128, "grad_norm": 0.1942444747483496, "learning_rate": 2.516799819449447e-05, "loss": 0.7418, "step": 4034 }, { "epoch": 0.7762601000384763, "grad_norm": 0.19638180136609829, "learning_rate": 2.512668115991156e-05, "loss": 0.7497, "step": 4035 }, { "epoch": 0.7764524817237399, "grad_norm": 0.19013469201938515, "learning_rate": 2.5085393193729e-05, "loss": 0.8096, "step": 4036 }, { "epoch": 0.7766448634090035, "grad_norm": 0.21116412291587272, "learning_rate": 2.5044134311976152e-05, "loss": 0.7842, "step": 4037 }, { "epoch": 0.776837245094267, "grad_norm": 0.20805510204006858, "learning_rate": 2.5002904530671236e-05, "loss": 0.7222, "step": 4038 }, { "epoch": 0.7770296267795306, "grad_norm": 0.18650239605438665, "learning_rate": 2.4961703865820975e-05, "loss": 0.7973, "step": 4039 }, { "epoch": 0.7772220084647942, "grad_norm": 0.18523226490003047, "learning_rate": 2.492053233342091e-05, "loss": 0.7909, "step": 4040 }, { "epoch": 0.7774143901500578, "grad_norm": 0.20009189399945954, "learning_rate": 2.4879389949455268e-05, "loss": 0.7438, "step": 4041 }, { "epoch": 0.7776067718353212, "grad_norm": 0.18527177264510275, "learning_rate": 2.483827672989688e-05, "loss": 0.7563, "step": 4042 }, { "epoch": 0.7777991535205848, "grad_norm": 0.1928469649579981, "learning_rate": 2.4797192690707428e-05, "loss": 0.7894, "step": 4043 }, { "epoch": 0.7779915352058484, "grad_norm": 0.19816785960863906, "learning_rate": 2.4756137847837025e-05, "loss": 0.801, "step": 4044 }, { "epoch": 0.778183916891112, "grad_norm": 0.20057904996503126, "learning_rate": 2.4715112217224655e-05, "loss": 0.7659, "step": 4045 }, { "epoch": 0.7783762985763756, "grad_norm": 0.21098479306206205, "learning_rate": 2.467411581479786e-05, "loss": 0.8031, "step": 4046 }, { "epoch": 0.7785686802616391, "grad_norm": 0.19943314252069066, "learning_rate": 2.463314865647286e-05, "loss": 0.7117, "step": 4047 }, { "epoch": 0.7787610619469026, "grad_norm": 0.20688535712290512, "learning_rate": 2.45922107581545e-05, "loss": 0.7542, "step": 4048 }, { "epoch": 0.7789534436321662, "grad_norm": 0.210131655797018, "learning_rate": 2.4551302135736286e-05, "loss": 0.8659, "step": 4049 }, { "epoch": 0.7791458253174298, "grad_norm": 0.18622715375925492, "learning_rate": 2.4510422805100363e-05, "loss": 0.7775, "step": 4050 }, { "epoch": 0.7793382070026933, "grad_norm": 0.20589946091266687, "learning_rate": 2.4469572782117457e-05, "loss": 0.7222, "step": 4051 }, { "epoch": 0.7795305886879569, "grad_norm": 0.21642785887718566, "learning_rate": 2.4428752082647045e-05, "loss": 0.8007, "step": 4052 }, { "epoch": 0.7797229703732205, "grad_norm": 0.20215208155783512, "learning_rate": 2.438796072253704e-05, "loss": 0.7694, "step": 4053 }, { "epoch": 0.7799153520584841, "grad_norm": 0.1934791134904462, "learning_rate": 2.4347198717624054e-05, "loss": 0.7473, "step": 4054 }, { "epoch": 0.7801077337437476, "grad_norm": 0.19994751046543094, "learning_rate": 2.4306466083733393e-05, "loss": 0.7893, "step": 4055 }, { "epoch": 0.7803001154290111, "grad_norm": 0.19152014618429056, "learning_rate": 2.426576283667873e-05, "loss": 0.8326, "step": 4056 }, { "epoch": 0.7804924971142747, "grad_norm": 0.10598001932854197, "learning_rate": 2.422508899226258e-05, "loss": 0.6352, "step": 4057 }, { "epoch": 0.7806848787995383, "grad_norm": 0.20078584903646954, "learning_rate": 2.4184444566275886e-05, "loss": 0.7397, "step": 4058 }, { "epoch": 0.7808772604848019, "grad_norm": 0.11045234683384529, "learning_rate": 2.4143829574498223e-05, "loss": 0.6314, "step": 4059 }, { "epoch": 0.7810696421700654, "grad_norm": 0.1818524603164669, "learning_rate": 2.4103244032697715e-05, "loss": 0.7938, "step": 4060 }, { "epoch": 0.781262023855329, "grad_norm": 0.2038198249761358, "learning_rate": 2.406268795663108e-05, "loss": 0.789, "step": 4061 }, { "epoch": 0.7814544055405925, "grad_norm": 0.19745219108201423, "learning_rate": 2.402216136204357e-05, "loss": 0.7919, "step": 4062 }, { "epoch": 0.7816467872258561, "grad_norm": 0.20060493772341637, "learning_rate": 2.3981664264669023e-05, "loss": 0.8349, "step": 4063 }, { "epoch": 0.7818391689111197, "grad_norm": 0.214462011988407, "learning_rate": 2.3941196680229794e-05, "loss": 0.7916, "step": 4064 }, { "epoch": 0.7820315505963832, "grad_norm": 0.2080384513056943, "learning_rate": 2.390075862443677e-05, "loss": 0.8101, "step": 4065 }, { "epoch": 0.7822239322816468, "grad_norm": 0.18068666370785372, "learning_rate": 2.386035011298947e-05, "loss": 0.8122, "step": 4066 }, { "epoch": 0.7824163139669104, "grad_norm": 0.1777475187282729, "learning_rate": 2.3819971161575803e-05, "loss": 0.724, "step": 4067 }, { "epoch": 0.782608695652174, "grad_norm": 0.19043284924895618, "learning_rate": 2.377962178587225e-05, "loss": 0.7783, "step": 4068 }, { "epoch": 0.7828010773374374, "grad_norm": 0.2016993483192696, "learning_rate": 2.3739302001543918e-05, "loss": 0.733, "step": 4069 }, { "epoch": 0.782993459022701, "grad_norm": 0.19370440508156317, "learning_rate": 2.3699011824244232e-05, "loss": 0.7915, "step": 4070 }, { "epoch": 0.7831858407079646, "grad_norm": 0.21753621084555158, "learning_rate": 2.365875126961531e-05, "loss": 0.7574, "step": 4071 }, { "epoch": 0.7833782223932282, "grad_norm": 0.17744720302703113, "learning_rate": 2.3618520353287643e-05, "loss": 0.7694, "step": 4072 }, { "epoch": 0.7835706040784918, "grad_norm": 0.21597690911282258, "learning_rate": 2.357831909088026e-05, "loss": 0.8066, "step": 4073 }, { "epoch": 0.7837629857637552, "grad_norm": 0.18815738498494688, "learning_rate": 2.3538147498000695e-05, "loss": 0.7886, "step": 4074 }, { "epoch": 0.7839553674490188, "grad_norm": 0.20677591211218727, "learning_rate": 2.349800559024492e-05, "loss": 0.7241, "step": 4075 }, { "epoch": 0.7841477491342824, "grad_norm": 0.18584239837928937, "learning_rate": 2.3457893383197416e-05, "loss": 0.7762, "step": 4076 }, { "epoch": 0.784340130819546, "grad_norm": 0.2144038358009858, "learning_rate": 2.3417810892431103e-05, "loss": 0.7337, "step": 4077 }, { "epoch": 0.7845325125048095, "grad_norm": 0.1917208361535439, "learning_rate": 2.3377758133507456e-05, "loss": 0.7654, "step": 4078 }, { "epoch": 0.7847248941900731, "grad_norm": 0.21508128784333067, "learning_rate": 2.3337735121976245e-05, "loss": 0.8376, "step": 4079 }, { "epoch": 0.7849172758753367, "grad_norm": 0.18925855948823508, "learning_rate": 2.3297741873375802e-05, "loss": 0.8055, "step": 4080 }, { "epoch": 0.7851096575606002, "grad_norm": 0.21219833375576164, "learning_rate": 2.3257778403232954e-05, "loss": 0.7505, "step": 4081 }, { "epoch": 0.7853020392458638, "grad_norm": 0.2005454577795445, "learning_rate": 2.321784472706279e-05, "loss": 0.7802, "step": 4082 }, { "epoch": 0.7854944209311273, "grad_norm": 0.22104843993605885, "learning_rate": 2.3177940860369007e-05, "loss": 0.7961, "step": 4083 }, { "epoch": 0.7856868026163909, "grad_norm": 0.17346726139498078, "learning_rate": 2.3138066818643644e-05, "loss": 0.7854, "step": 4084 }, { "epoch": 0.7858791843016545, "grad_norm": 0.19806098962409455, "learning_rate": 2.3098222617367182e-05, "loss": 0.6996, "step": 4085 }, { "epoch": 0.7860715659869181, "grad_norm": 0.21067385388549012, "learning_rate": 2.30584082720085e-05, "loss": 0.7682, "step": 4086 }, { "epoch": 0.7862639476721817, "grad_norm": 0.1960004139936808, "learning_rate": 2.3018623798024918e-05, "loss": 0.8123, "step": 4087 }, { "epoch": 0.7864563293574451, "grad_norm": 0.2168461194679321, "learning_rate": 2.297886921086211e-05, "loss": 0.7494, "step": 4088 }, { "epoch": 0.7866487110427087, "grad_norm": 0.20301702117126624, "learning_rate": 2.293914452595419e-05, "loss": 0.7818, "step": 4089 }, { "epoch": 0.7868410927279723, "grad_norm": 0.18719090049483145, "learning_rate": 2.2899449758723657e-05, "loss": 0.7466, "step": 4090 }, { "epoch": 0.7870334744132359, "grad_norm": 0.21358049457068468, "learning_rate": 2.285978492458134e-05, "loss": 0.7114, "step": 4091 }, { "epoch": 0.7872258560984994, "grad_norm": 0.18447608863517417, "learning_rate": 2.282015003892659e-05, "loss": 0.7815, "step": 4092 }, { "epoch": 0.787418237783763, "grad_norm": 0.18310132422599548, "learning_rate": 2.2780545117146947e-05, "loss": 0.7607, "step": 4093 }, { "epoch": 0.7876106194690266, "grad_norm": 0.17916234696537894, "learning_rate": 2.2740970174618403e-05, "loss": 0.7645, "step": 4094 }, { "epoch": 0.7878030011542901, "grad_norm": 0.2093632400993173, "learning_rate": 2.2701425226705408e-05, "loss": 0.7513, "step": 4095 }, { "epoch": 0.7879953828395537, "grad_norm": 0.2281326429103137, "learning_rate": 2.2661910288760547e-05, "loss": 0.7627, "step": 4096 }, { "epoch": 0.7881877645248172, "grad_norm": 0.20723151509306165, "learning_rate": 2.2622425376124967e-05, "loss": 0.8171, "step": 4097 }, { "epoch": 0.7883801462100808, "grad_norm": 0.20066516610944762, "learning_rate": 2.2582970504128042e-05, "loss": 0.7426, "step": 4098 }, { "epoch": 0.7885725278953444, "grad_norm": 0.1970107290566636, "learning_rate": 2.2543545688087518e-05, "loss": 0.8103, "step": 4099 }, { "epoch": 0.788764909580608, "grad_norm": 0.20966681197670106, "learning_rate": 2.2504150943309453e-05, "loss": 0.7209, "step": 4100 }, { "epoch": 0.7889572912658714, "grad_norm": 0.19957748969057745, "learning_rate": 2.2464786285088268e-05, "loss": 0.7411, "step": 4101 }, { "epoch": 0.789149672951135, "grad_norm": 0.1867872144458973, "learning_rate": 2.242545172870665e-05, "loss": 0.7936, "step": 4102 }, { "epoch": 0.7893420546363986, "grad_norm": 0.11429152065955896, "learning_rate": 2.238614728943561e-05, "loss": 0.6186, "step": 4103 }, { "epoch": 0.7895344363216622, "grad_norm": 0.22410146546891593, "learning_rate": 2.2346872982534584e-05, "loss": 0.7346, "step": 4104 }, { "epoch": 0.7897268180069258, "grad_norm": 0.19291386094905313, "learning_rate": 2.230762882325108e-05, "loss": 0.7868, "step": 4105 }, { "epoch": 0.7899191996921893, "grad_norm": 0.10024363969030978, "learning_rate": 2.2268414826821115e-05, "loss": 0.5769, "step": 4106 }, { "epoch": 0.7901115813774529, "grad_norm": 0.18619817809963626, "learning_rate": 2.2229231008468932e-05, "loss": 0.8111, "step": 4107 }, { "epoch": 0.7903039630627164, "grad_norm": 0.19287323383344715, "learning_rate": 2.2190077383406936e-05, "loss": 0.7618, "step": 4108 }, { "epoch": 0.79049634474798, "grad_norm": 0.2018012728914857, "learning_rate": 2.2150953966835996e-05, "loss": 0.7791, "step": 4109 }, { "epoch": 0.7906887264332435, "grad_norm": 0.1973430096870648, "learning_rate": 2.211186077394516e-05, "loss": 0.7422, "step": 4110 }, { "epoch": 0.7908811081185071, "grad_norm": 0.19732823838345678, "learning_rate": 2.2072797819911726e-05, "loss": 0.7581, "step": 4111 }, { "epoch": 0.7910734898037707, "grad_norm": 0.19317879227783538, "learning_rate": 2.2033765119901295e-05, "loss": 0.7672, "step": 4112 }, { "epoch": 0.7912658714890343, "grad_norm": 0.21287385321759442, "learning_rate": 2.1994762689067704e-05, "loss": 0.7772, "step": 4113 }, { "epoch": 0.7914582531742979, "grad_norm": 0.19028240825296452, "learning_rate": 2.1955790542553033e-05, "loss": 0.766, "step": 4114 }, { "epoch": 0.7916506348595613, "grad_norm": 0.19179160781691962, "learning_rate": 2.1916848695487614e-05, "loss": 0.7625, "step": 4115 }, { "epoch": 0.7918430165448249, "grad_norm": 0.1910657066911245, "learning_rate": 2.1877937162990014e-05, "loss": 0.771, "step": 4116 }, { "epoch": 0.7920353982300885, "grad_norm": 0.20279380091949148, "learning_rate": 2.1839055960166997e-05, "loss": 0.7535, "step": 4117 }, { "epoch": 0.7922277799153521, "grad_norm": 0.19494451359062773, "learning_rate": 2.180020510211367e-05, "loss": 0.725, "step": 4118 }, { "epoch": 0.7924201616006156, "grad_norm": 0.19219551848332816, "learning_rate": 2.1761384603913205e-05, "loss": 0.8145, "step": 4119 }, { "epoch": 0.7926125432858792, "grad_norm": 0.18559154918676526, "learning_rate": 2.1722594480637036e-05, "loss": 0.8251, "step": 4120 }, { "epoch": 0.7928049249711427, "grad_norm": 0.18722753576350787, "learning_rate": 2.168383474734491e-05, "loss": 0.8404, "step": 4121 }, { "epoch": 0.7929973066564063, "grad_norm": 0.20757405816776406, "learning_rate": 2.1645105419084588e-05, "loss": 0.7505, "step": 4122 }, { "epoch": 0.7931896883416699, "grad_norm": 0.19140795691777784, "learning_rate": 2.1606406510892208e-05, "loss": 0.821, "step": 4123 }, { "epoch": 0.7933820700269334, "grad_norm": 0.190730764268605, "learning_rate": 2.1567738037791995e-05, "loss": 0.798, "step": 4124 }, { "epoch": 0.793574451712197, "grad_norm": 0.19829759001731942, "learning_rate": 2.152910001479638e-05, "loss": 0.8056, "step": 4125 }, { "epoch": 0.7937668333974606, "grad_norm": 0.18500433995653598, "learning_rate": 2.1490492456905965e-05, "loss": 0.7749, "step": 4126 }, { "epoch": 0.7939592150827242, "grad_norm": 0.18544634727544715, "learning_rate": 2.1451915379109543e-05, "loss": 0.7857, "step": 4127 }, { "epoch": 0.7941515967679876, "grad_norm": 0.17576859326190292, "learning_rate": 2.141336879638406e-05, "loss": 0.8035, "step": 4128 }, { "epoch": 0.7943439784532512, "grad_norm": 0.1970558033522612, "learning_rate": 2.1374852723694592e-05, "loss": 0.788, "step": 4129 }, { "epoch": 0.7945363601385148, "grad_norm": 0.19145676027837358, "learning_rate": 2.1336367175994508e-05, "loss": 0.813, "step": 4130 }, { "epoch": 0.7947287418237784, "grad_norm": 0.1721252949864677, "learning_rate": 2.1297912168225088e-05, "loss": 0.7328, "step": 4131 }, { "epoch": 0.794921123509042, "grad_norm": 0.19900133257362704, "learning_rate": 2.1259487715315997e-05, "loss": 0.7713, "step": 4132 }, { "epoch": 0.7951135051943055, "grad_norm": 0.10802814123120044, "learning_rate": 2.12210938321849e-05, "loss": 0.5467, "step": 4133 }, { "epoch": 0.795305886879569, "grad_norm": 0.20698112240983893, "learning_rate": 2.1182730533737573e-05, "loss": 0.7909, "step": 4134 }, { "epoch": 0.7954982685648326, "grad_norm": 0.10766248704158474, "learning_rate": 2.1144397834868035e-05, "loss": 0.6151, "step": 4135 }, { "epoch": 0.7956906502500962, "grad_norm": 0.2201801452157244, "learning_rate": 2.110609575045833e-05, "loss": 0.7713, "step": 4136 }, { "epoch": 0.7958830319353598, "grad_norm": 0.18144593184253016, "learning_rate": 2.1067824295378658e-05, "loss": 0.8087, "step": 4137 }, { "epoch": 0.7960754136206233, "grad_norm": 0.18347877711924593, "learning_rate": 2.1029583484487314e-05, "loss": 0.7913, "step": 4138 }, { "epoch": 0.7962677953058869, "grad_norm": 0.19811991713339447, "learning_rate": 2.0991373332630682e-05, "loss": 0.7918, "step": 4139 }, { "epoch": 0.7964601769911505, "grad_norm": 0.20227839054772837, "learning_rate": 2.0953193854643273e-05, "loss": 0.7418, "step": 4140 }, { "epoch": 0.796652558676414, "grad_norm": 0.17533342813452849, "learning_rate": 2.0915045065347672e-05, "loss": 0.7646, "step": 4141 }, { "epoch": 0.7968449403616775, "grad_norm": 0.18698177007206876, "learning_rate": 2.0876926979554546e-05, "loss": 0.7883, "step": 4142 }, { "epoch": 0.7970373220469411, "grad_norm": 0.20040660388285078, "learning_rate": 2.0838839612062634e-05, "loss": 0.7466, "step": 4143 }, { "epoch": 0.7972297037322047, "grad_norm": 0.17937174096885372, "learning_rate": 2.0800782977658838e-05, "loss": 0.7929, "step": 4144 }, { "epoch": 0.7974220854174683, "grad_norm": 0.18985217185702843, "learning_rate": 2.0762757091117936e-05, "loss": 0.8236, "step": 4145 }, { "epoch": 0.7976144671027319, "grad_norm": 0.20594714986577708, "learning_rate": 2.0724761967202987e-05, "loss": 0.8035, "step": 4146 }, { "epoch": 0.7978068487879953, "grad_norm": 0.16160947965474248, "learning_rate": 2.068679762066499e-05, "loss": 0.759, "step": 4147 }, { "epoch": 0.7979992304732589, "grad_norm": 0.19444417386537166, "learning_rate": 2.0648864066242933e-05, "loss": 0.7989, "step": 4148 }, { "epoch": 0.7981916121585225, "grad_norm": 0.19634305664244905, "learning_rate": 2.0610961318664013e-05, "loss": 0.8023, "step": 4149 }, { "epoch": 0.7983839938437861, "grad_norm": 0.19544641736782073, "learning_rate": 2.0573089392643364e-05, "loss": 0.7889, "step": 4150 }, { "epoch": 0.7985763755290496, "grad_norm": 0.20423155740386598, "learning_rate": 2.0535248302884146e-05, "loss": 0.7664, "step": 4151 }, { "epoch": 0.7987687572143132, "grad_norm": 0.19400752986745862, "learning_rate": 2.04974380640776e-05, "loss": 0.7788, "step": 4152 }, { "epoch": 0.7989611388995768, "grad_norm": 0.20848827321578145, "learning_rate": 2.045965869090295e-05, "loss": 0.7796, "step": 4153 }, { "epoch": 0.7991535205848403, "grad_norm": 0.20910045528008933, "learning_rate": 2.042191019802745e-05, "loss": 0.7679, "step": 4154 }, { "epoch": 0.7993459022701039, "grad_norm": 0.22467654597869238, "learning_rate": 2.0384192600106335e-05, "loss": 0.7389, "step": 4155 }, { "epoch": 0.7995382839553674, "grad_norm": 0.17820405346943582, "learning_rate": 2.0346505911782954e-05, "loss": 0.8084, "step": 4156 }, { "epoch": 0.799730665640631, "grad_norm": 0.20553931877750328, "learning_rate": 2.0308850147688486e-05, "loss": 0.7038, "step": 4157 }, { "epoch": 0.7999230473258946, "grad_norm": 0.1919293890381869, "learning_rate": 2.0271225322442256e-05, "loss": 0.786, "step": 4158 }, { "epoch": 0.8001154290111582, "grad_norm": 0.22035703094058265, "learning_rate": 2.0233631450651523e-05, "loss": 0.7853, "step": 4159 }, { "epoch": 0.8003078106964217, "grad_norm": 0.18851187041171866, "learning_rate": 2.019606854691145e-05, "loss": 0.7706, "step": 4160 }, { "epoch": 0.8005001923816852, "grad_norm": 0.19089310517587252, "learning_rate": 2.0158536625805325e-05, "loss": 0.7629, "step": 4161 }, { "epoch": 0.8006925740669488, "grad_norm": 0.20006902120585232, "learning_rate": 2.01210357019043e-05, "loss": 0.7734, "step": 4162 }, { "epoch": 0.8008849557522124, "grad_norm": 0.20467364793100756, "learning_rate": 2.0083565789767523e-05, "loss": 0.7537, "step": 4163 }, { "epoch": 0.801077337437476, "grad_norm": 0.19447407585198365, "learning_rate": 2.004612690394212e-05, "loss": 0.8013, "step": 4164 }, { "epoch": 0.8012697191227395, "grad_norm": 0.20463364458436978, "learning_rate": 2.0008719058963145e-05, "loss": 0.7723, "step": 4165 }, { "epoch": 0.8014621008080031, "grad_norm": 0.2204712532033596, "learning_rate": 1.997134226935361e-05, "loss": 0.8092, "step": 4166 }, { "epoch": 0.8016544824932667, "grad_norm": 0.19810544032952537, "learning_rate": 1.993399654962447e-05, "loss": 0.7884, "step": 4167 }, { "epoch": 0.8018468641785302, "grad_norm": 0.22000576865962967, "learning_rate": 1.9896681914274617e-05, "loss": 0.788, "step": 4168 }, { "epoch": 0.8020392458637937, "grad_norm": 0.20166403029950197, "learning_rate": 1.9859398377790873e-05, "loss": 0.7609, "step": 4169 }, { "epoch": 0.8022316275490573, "grad_norm": 0.45572983251347166, "learning_rate": 1.982214595464804e-05, "loss": 0.6488, "step": 4170 }, { "epoch": 0.8024240092343209, "grad_norm": 0.1905944313994455, "learning_rate": 1.97849246593087e-05, "loss": 0.7886, "step": 4171 }, { "epoch": 0.8026163909195845, "grad_norm": 0.19935081630531895, "learning_rate": 1.9747734506223525e-05, "loss": 0.7675, "step": 4172 }, { "epoch": 0.8028087726048481, "grad_norm": 0.11427221952655961, "learning_rate": 1.9710575509831007e-05, "loss": 0.6595, "step": 4173 }, { "epoch": 0.8030011542901115, "grad_norm": 0.22173417972031526, "learning_rate": 1.9673447684557468e-05, "loss": 0.8527, "step": 4174 }, { "epoch": 0.8031935359753751, "grad_norm": 0.19556854966157064, "learning_rate": 1.9636351044817292e-05, "loss": 0.7446, "step": 4175 }, { "epoch": 0.8033859176606387, "grad_norm": 0.18153358618047818, "learning_rate": 1.9599285605012642e-05, "loss": 0.8093, "step": 4176 }, { "epoch": 0.8035782993459023, "grad_norm": 0.20263377327629606, "learning_rate": 1.956225137953359e-05, "loss": 0.7671, "step": 4177 }, { "epoch": 0.8037706810311658, "grad_norm": 0.20172477043388123, "learning_rate": 1.952524838275811e-05, "loss": 0.8032, "step": 4178 }, { "epoch": 0.8039630627164294, "grad_norm": 0.19421024304461507, "learning_rate": 1.9488276629052026e-05, "loss": 0.7521, "step": 4179 }, { "epoch": 0.804155444401693, "grad_norm": 0.10908015933854244, "learning_rate": 1.945133613276907e-05, "loss": 0.5919, "step": 4180 }, { "epoch": 0.8043478260869565, "grad_norm": 0.20767451169092968, "learning_rate": 1.941442690825076e-05, "loss": 0.7421, "step": 4181 }, { "epoch": 0.8045402077722201, "grad_norm": 0.18763639810299718, "learning_rate": 1.9377548969826633e-05, "loss": 0.7624, "step": 4182 }, { "epoch": 0.8047325894574836, "grad_norm": 0.18478291403911967, "learning_rate": 1.9340702331813842e-05, "loss": 0.7062, "step": 4183 }, { "epoch": 0.8049249711427472, "grad_norm": 0.20242283652595552, "learning_rate": 1.930388700851762e-05, "loss": 0.758, "step": 4184 }, { "epoch": 0.8051173528280108, "grad_norm": 0.18626866496861424, "learning_rate": 1.9267103014230935e-05, "loss": 0.7524, "step": 4185 }, { "epoch": 0.8053097345132744, "grad_norm": 0.1852298003951285, "learning_rate": 1.9230350363234516e-05, "loss": 0.7425, "step": 4186 }, { "epoch": 0.805502116198538, "grad_norm": 0.1964835439402861, "learning_rate": 1.9193629069797102e-05, "loss": 0.7792, "step": 4187 }, { "epoch": 0.8056944978838014, "grad_norm": 0.18460300422389786, "learning_rate": 1.9156939148175125e-05, "loss": 0.8289, "step": 4188 }, { "epoch": 0.805886879569065, "grad_norm": 0.1931453970105838, "learning_rate": 1.912028061261287e-05, "loss": 0.7399, "step": 4189 }, { "epoch": 0.8060792612543286, "grad_norm": 0.18881121015438115, "learning_rate": 1.9083653477342466e-05, "loss": 0.7753, "step": 4190 }, { "epoch": 0.8062716429395922, "grad_norm": 0.19778078851073405, "learning_rate": 1.904705775658381e-05, "loss": 0.7819, "step": 4191 }, { "epoch": 0.8064640246248557, "grad_norm": 0.2108820353053329, "learning_rate": 1.901049346454462e-05, "loss": 0.8039, "step": 4192 }, { "epoch": 0.8066564063101193, "grad_norm": 0.18420045046713734, "learning_rate": 1.8973960615420416e-05, "loss": 0.7851, "step": 4193 }, { "epoch": 0.8068487879953828, "grad_norm": 0.19464449844021386, "learning_rate": 1.8937459223394515e-05, "loss": 0.7814, "step": 4194 }, { "epoch": 0.8070411696806464, "grad_norm": 0.1833808759349085, "learning_rate": 1.8900989302637985e-05, "loss": 0.7778, "step": 4195 }, { "epoch": 0.80723355136591, "grad_norm": 0.17424642563825643, "learning_rate": 1.8864550867309773e-05, "loss": 0.7492, "step": 4196 }, { "epoch": 0.8074259330511735, "grad_norm": 0.20103716953580258, "learning_rate": 1.8828143931556442e-05, "loss": 0.7972, "step": 4197 }, { "epoch": 0.8076183147364371, "grad_norm": 0.17740737878952476, "learning_rate": 1.8791768509512487e-05, "loss": 0.7518, "step": 4198 }, { "epoch": 0.8078106964217007, "grad_norm": 0.18488221025047008, "learning_rate": 1.875542461530011e-05, "loss": 0.7643, "step": 4199 }, { "epoch": 0.8080030781069643, "grad_norm": 0.201666754580493, "learning_rate": 1.871911226302917e-05, "loss": 0.7904, "step": 4200 }, { "epoch": 0.8081954597922277, "grad_norm": 0.20357500945523693, "learning_rate": 1.8682831466797467e-05, "loss": 0.7712, "step": 4201 }, { "epoch": 0.8083878414774913, "grad_norm": 0.2055421683522817, "learning_rate": 1.8646582240690412e-05, "loss": 0.7203, "step": 4202 }, { "epoch": 0.8085802231627549, "grad_norm": 0.19612203002415837, "learning_rate": 1.8610364598781226e-05, "loss": 0.6921, "step": 4203 }, { "epoch": 0.8087726048480185, "grad_norm": 0.2062468107656611, "learning_rate": 1.8574178555130817e-05, "loss": 0.8447, "step": 4204 }, { "epoch": 0.8089649865332821, "grad_norm": 0.17555012204251774, "learning_rate": 1.8538024123787868e-05, "loss": 0.8258, "step": 4205 }, { "epoch": 0.8091573682185456, "grad_norm": 0.19213022267556873, "learning_rate": 1.8501901318788774e-05, "loss": 0.8069, "step": 4206 }, { "epoch": 0.8093497499038091, "grad_norm": 0.19234573181702042, "learning_rate": 1.8465810154157624e-05, "loss": 0.753, "step": 4207 }, { "epoch": 0.8095421315890727, "grad_norm": 0.20643171397447355, "learning_rate": 1.842975064390633e-05, "loss": 0.7242, "step": 4208 }, { "epoch": 0.8097345132743363, "grad_norm": 0.18632244279070348, "learning_rate": 1.839372280203433e-05, "loss": 0.7742, "step": 4209 }, { "epoch": 0.8099268949595998, "grad_norm": 0.19984957690847646, "learning_rate": 1.835772664252895e-05, "loss": 0.7651, "step": 4210 }, { "epoch": 0.8101192766448634, "grad_norm": 0.1855857561080202, "learning_rate": 1.832176217936511e-05, "loss": 0.7506, "step": 4211 }, { "epoch": 0.810311658330127, "grad_norm": 0.19875083058310367, "learning_rate": 1.8285829426505453e-05, "loss": 0.8241, "step": 4212 }, { "epoch": 0.8105040400153906, "grad_norm": 0.1888162178441013, "learning_rate": 1.824992839790035e-05, "loss": 0.8076, "step": 4213 }, { "epoch": 0.8106964217006541, "grad_norm": 0.221334297402037, "learning_rate": 1.8214059107487726e-05, "loss": 0.8149, "step": 4214 }, { "epoch": 0.8108888033859176, "grad_norm": 0.21200633806240576, "learning_rate": 1.8178221569193343e-05, "loss": 0.7874, "step": 4215 }, { "epoch": 0.8110811850711812, "grad_norm": 0.1887312793266369, "learning_rate": 1.8142415796930568e-05, "loss": 0.7878, "step": 4216 }, { "epoch": 0.8112735667564448, "grad_norm": 0.11332506611207961, "learning_rate": 1.810664180460041e-05, "loss": 0.6364, "step": 4217 }, { "epoch": 0.8114659484417084, "grad_norm": 0.20915129776252483, "learning_rate": 1.8070899606091583e-05, "loss": 0.7955, "step": 4218 }, { "epoch": 0.8116583301269719, "grad_norm": 0.1934772271377311, "learning_rate": 1.8035189215280423e-05, "loss": 0.7527, "step": 4219 }, { "epoch": 0.8118507118122354, "grad_norm": 0.19503421361912826, "learning_rate": 1.7999510646030948e-05, "loss": 0.7612, "step": 4220 }, { "epoch": 0.812043093497499, "grad_norm": 0.19114770870508727, "learning_rate": 1.7963863912194766e-05, "loss": 0.8349, "step": 4221 }, { "epoch": 0.8122354751827626, "grad_norm": 0.18191959379100725, "learning_rate": 1.7928249027611254e-05, "loss": 0.7845, "step": 4222 }, { "epoch": 0.8124278568680262, "grad_norm": 0.12118306174937425, "learning_rate": 1.789266600610724e-05, "loss": 0.6363, "step": 4223 }, { "epoch": 0.8126202385532897, "grad_norm": 0.20823765537814157, "learning_rate": 1.7857114861497336e-05, "loss": 0.7621, "step": 4224 }, { "epoch": 0.8128126202385533, "grad_norm": 0.10835733708591597, "learning_rate": 1.782159560758373e-05, "loss": 0.6288, "step": 4225 }, { "epoch": 0.8130050019238169, "grad_norm": 0.20867435152299094, "learning_rate": 1.7786108258156155e-05, "loss": 0.8567, "step": 4226 }, { "epoch": 0.8131973836090804, "grad_norm": 0.2096205414167459, "learning_rate": 1.7750652826992075e-05, "loss": 0.7598, "step": 4227 }, { "epoch": 0.8133897652943439, "grad_norm": 0.20377798742761044, "learning_rate": 1.7715229327856498e-05, "loss": 0.7846, "step": 4228 }, { "epoch": 0.8135821469796075, "grad_norm": 0.2143157474261449, "learning_rate": 1.767983777450205e-05, "loss": 0.7561, "step": 4229 }, { "epoch": 0.8137745286648711, "grad_norm": 0.18542165330376323, "learning_rate": 1.7644478180668943e-05, "loss": 0.7985, "step": 4230 }, { "epoch": 0.8139669103501347, "grad_norm": 0.20158776342979973, "learning_rate": 1.7609150560084985e-05, "loss": 0.719, "step": 4231 }, { "epoch": 0.8141592920353983, "grad_norm": 0.21409586431352873, "learning_rate": 1.757385492646558e-05, "loss": 0.794, "step": 4232 }, { "epoch": 0.8143516737206618, "grad_norm": 0.22330123240682648, "learning_rate": 1.7538591293513685e-05, "loss": 0.7765, "step": 4233 }, { "epoch": 0.8145440554059253, "grad_norm": 0.19394792575353104, "learning_rate": 1.750335967491993e-05, "loss": 0.7844, "step": 4234 }, { "epoch": 0.8147364370911889, "grad_norm": 0.18846170514424396, "learning_rate": 1.746816008436234e-05, "loss": 0.8208, "step": 4235 }, { "epoch": 0.8149288187764525, "grad_norm": 0.20944474921110603, "learning_rate": 1.7432992535506686e-05, "loss": 0.7537, "step": 4236 }, { "epoch": 0.8151212004617161, "grad_norm": 0.19663193561468256, "learning_rate": 1.7397857042006194e-05, "loss": 0.7823, "step": 4237 }, { "epoch": 0.8153135821469796, "grad_norm": 0.09877460517471198, "learning_rate": 1.736275361750167e-05, "loss": 0.6048, "step": 4238 }, { "epoch": 0.8155059638322432, "grad_norm": 0.20432711963730307, "learning_rate": 1.7327682275621503e-05, "loss": 0.7938, "step": 4239 }, { "epoch": 0.8156983455175068, "grad_norm": 0.21195701914548223, "learning_rate": 1.7292643029981526e-05, "loss": 0.8224, "step": 4240 }, { "epoch": 0.8158907272027703, "grad_norm": 0.19511322248977608, "learning_rate": 1.725763589418523e-05, "loss": 0.7853, "step": 4241 }, { "epoch": 0.8160831088880338, "grad_norm": 0.2233850491240068, "learning_rate": 1.7222660881823593e-05, "loss": 0.8085, "step": 4242 }, { "epoch": 0.8162754905732974, "grad_norm": 0.19626808634343085, "learning_rate": 1.7187718006475117e-05, "loss": 0.7601, "step": 4243 }, { "epoch": 0.816467872258561, "grad_norm": 0.10317309524157219, "learning_rate": 1.7152807281705806e-05, "loss": 0.5721, "step": 4244 }, { "epoch": 0.8166602539438246, "grad_norm": 0.19674878004048443, "learning_rate": 1.7117928721069232e-05, "loss": 0.8232, "step": 4245 }, { "epoch": 0.8168526356290882, "grad_norm": 0.10758664915436657, "learning_rate": 1.7083082338106438e-05, "loss": 0.5795, "step": 4246 }, { "epoch": 0.8170450173143516, "grad_norm": 0.17830543816589944, "learning_rate": 1.7048268146345968e-05, "loss": 0.804, "step": 4247 }, { "epoch": 0.8172373989996152, "grad_norm": 0.20293185257229407, "learning_rate": 1.701348615930397e-05, "loss": 0.7825, "step": 4248 }, { "epoch": 0.8174297806848788, "grad_norm": 0.18842779216780897, "learning_rate": 1.6978736390483895e-05, "loss": 0.7517, "step": 4249 }, { "epoch": 0.8176221623701424, "grad_norm": 0.18821138012800803, "learning_rate": 1.6944018853376897e-05, "loss": 0.7473, "step": 4250 }, { "epoch": 0.8178145440554059, "grad_norm": 0.2118570881973256, "learning_rate": 1.6909333561461472e-05, "loss": 0.7709, "step": 4251 }, { "epoch": 0.8180069257406695, "grad_norm": 0.20466055148367385, "learning_rate": 1.6874680528203656e-05, "loss": 0.8252, "step": 4252 }, { "epoch": 0.818199307425933, "grad_norm": 0.20877206029502318, "learning_rate": 1.684005976705695e-05, "loss": 0.8088, "step": 4253 }, { "epoch": 0.8183916891111966, "grad_norm": 0.1930317048288182, "learning_rate": 1.6805471291462317e-05, "loss": 0.7423, "step": 4254 }, { "epoch": 0.8185840707964602, "grad_norm": 0.24047451955882498, "learning_rate": 1.6770915114848194e-05, "loss": 0.7574, "step": 4255 }, { "epoch": 0.8187764524817237, "grad_norm": 0.17507734083493345, "learning_rate": 1.67363912506305e-05, "loss": 0.7552, "step": 4256 }, { "epoch": 0.8189688341669873, "grad_norm": 0.17658367100147562, "learning_rate": 1.6701899712212564e-05, "loss": 0.7659, "step": 4257 }, { "epoch": 0.8191612158522509, "grad_norm": 0.09817979954169305, "learning_rate": 1.66674405129852e-05, "loss": 0.5587, "step": 4258 }, { "epoch": 0.8193535975375145, "grad_norm": 0.19768398700489795, "learning_rate": 1.6633013666326637e-05, "loss": 0.8008, "step": 4259 }, { "epoch": 0.8195459792227779, "grad_norm": 0.18038663174621794, "learning_rate": 1.6598619185602614e-05, "loss": 0.8111, "step": 4260 }, { "epoch": 0.8197383609080415, "grad_norm": 0.1955198305140136, "learning_rate": 1.656425708416617e-05, "loss": 0.7762, "step": 4261 }, { "epoch": 0.8199307425933051, "grad_norm": 0.19004394091571192, "learning_rate": 1.6529927375357958e-05, "loss": 0.8013, "step": 4262 }, { "epoch": 0.8201231242785687, "grad_norm": 0.20304471857984435, "learning_rate": 1.649563007250584e-05, "loss": 0.8052, "step": 4263 }, { "epoch": 0.8203155059638323, "grad_norm": 0.18947697729405966, "learning_rate": 1.6461365188925304e-05, "loss": 0.797, "step": 4264 }, { "epoch": 0.8205078876490958, "grad_norm": 0.18439692667261318, "learning_rate": 1.6427132737919137e-05, "loss": 0.7528, "step": 4265 }, { "epoch": 0.8207002693343594, "grad_norm": 0.20472246790508805, "learning_rate": 1.6392932732777487e-05, "loss": 0.7394, "step": 4266 }, { "epoch": 0.8208926510196229, "grad_norm": 0.18263981824416964, "learning_rate": 1.6358765186778057e-05, "loss": 0.7666, "step": 4267 }, { "epoch": 0.8210850327048865, "grad_norm": 0.17404414569244653, "learning_rate": 1.6324630113185836e-05, "loss": 0.7799, "step": 4268 }, { "epoch": 0.82127741439015, "grad_norm": 0.19104498214032584, "learning_rate": 1.6290527525253228e-05, "loss": 0.7764, "step": 4269 }, { "epoch": 0.8214697960754136, "grad_norm": 0.18006495923053725, "learning_rate": 1.6256457436220028e-05, "loss": 0.798, "step": 4270 }, { "epoch": 0.8216621777606772, "grad_norm": 0.23974723012427643, "learning_rate": 1.6222419859313442e-05, "loss": 0.8368, "step": 4271 }, { "epoch": 0.8218545594459408, "grad_norm": 0.20152961368897238, "learning_rate": 1.6188414807747997e-05, "loss": 0.7833, "step": 4272 }, { "epoch": 0.8220469411312044, "grad_norm": 0.17521629078193235, "learning_rate": 1.6154442294725636e-05, "loss": 0.8142, "step": 4273 }, { "epoch": 0.8222393228164678, "grad_norm": 0.21177379230306387, "learning_rate": 1.6120502333435693e-05, "loss": 0.8484, "step": 4274 }, { "epoch": 0.8224317045017314, "grad_norm": 0.18779451896515748, "learning_rate": 1.6086594937054765e-05, "loss": 0.8251, "step": 4275 }, { "epoch": 0.822624086186995, "grad_norm": 0.19797986001202703, "learning_rate": 1.6052720118746922e-05, "loss": 0.7885, "step": 4276 }, { "epoch": 0.8228164678722586, "grad_norm": 0.181258875956189, "learning_rate": 1.601887789166352e-05, "loss": 0.75, "step": 4277 }, { "epoch": 0.8230088495575221, "grad_norm": 0.18491483741748155, "learning_rate": 1.598506826894328e-05, "loss": 0.7534, "step": 4278 }, { "epoch": 0.8232012312427857, "grad_norm": 0.2277065167475249, "learning_rate": 1.5951291263712252e-05, "loss": 0.7941, "step": 4279 }, { "epoch": 0.8233936129280492, "grad_norm": 0.17576297236098573, "learning_rate": 1.5917546889083834e-05, "loss": 0.7087, "step": 4280 }, { "epoch": 0.8235859946133128, "grad_norm": 0.17669122288082273, "learning_rate": 1.5883835158158765e-05, "loss": 0.7603, "step": 4281 }, { "epoch": 0.8237783762985764, "grad_norm": 0.18713526203684122, "learning_rate": 1.585015608402509e-05, "loss": 0.7837, "step": 4282 }, { "epoch": 0.8239707579838399, "grad_norm": 0.09363246142685404, "learning_rate": 1.5816509679758185e-05, "loss": 0.5984, "step": 4283 }, { "epoch": 0.8241631396691035, "grad_norm": 0.19640682366210144, "learning_rate": 1.578289595842074e-05, "loss": 0.7564, "step": 4284 }, { "epoch": 0.8243555213543671, "grad_norm": 0.11198641627552249, "learning_rate": 1.5749314933062753e-05, "loss": 0.6413, "step": 4285 }, { "epoch": 0.8245479030396307, "grad_norm": 0.17684246777488125, "learning_rate": 1.5715766616721582e-05, "loss": 0.7883, "step": 4286 }, { "epoch": 0.8247402847248941, "grad_norm": 0.19281314294153393, "learning_rate": 1.5682251022421758e-05, "loss": 0.7463, "step": 4287 }, { "epoch": 0.8249326664101577, "grad_norm": 0.17161333953282057, "learning_rate": 1.5648768163175277e-05, "loss": 0.7552, "step": 4288 }, { "epoch": 0.8251250480954213, "grad_norm": 0.20956669935415106, "learning_rate": 1.561531805198124e-05, "loss": 0.8261, "step": 4289 }, { "epoch": 0.8253174297806849, "grad_norm": 0.19748684622259818, "learning_rate": 1.5581900701826225e-05, "loss": 0.7453, "step": 4290 }, { "epoch": 0.8255098114659485, "grad_norm": 0.18006009427029754, "learning_rate": 1.5548516125683978e-05, "loss": 0.8344, "step": 4291 }, { "epoch": 0.825702193151212, "grad_norm": 0.18608758471504902, "learning_rate": 1.5515164336515465e-05, "loss": 0.741, "step": 4292 }, { "epoch": 0.8258945748364755, "grad_norm": 0.17294410958297446, "learning_rate": 1.5481845347269076e-05, "loss": 0.7681, "step": 4293 }, { "epoch": 0.8260869565217391, "grad_norm": 0.18960170391629272, "learning_rate": 1.5448559170880374e-05, "loss": 0.7546, "step": 4294 }, { "epoch": 0.8262793382070027, "grad_norm": 0.20285926890786682, "learning_rate": 1.5415305820272197e-05, "loss": 0.815, "step": 4295 }, { "epoch": 0.8264717198922663, "grad_norm": 0.2003771043550806, "learning_rate": 1.538208530835463e-05, "loss": 0.7999, "step": 4296 }, { "epoch": 0.8266641015775298, "grad_norm": 0.19442553131663032, "learning_rate": 1.534889764802503e-05, "loss": 0.7134, "step": 4297 }, { "epoch": 0.8268564832627934, "grad_norm": 0.17961433776754238, "learning_rate": 1.531574285216799e-05, "loss": 0.7919, "step": 4298 }, { "epoch": 0.827048864948057, "grad_norm": 0.19809722327739468, "learning_rate": 1.528262093365531e-05, "loss": 0.8019, "step": 4299 }, { "epoch": 0.8272412466333205, "grad_norm": 0.2016715416704754, "learning_rate": 1.5249531905346138e-05, "loss": 0.818, "step": 4300 }, { "epoch": 0.827433628318584, "grad_norm": 0.18635199376674746, "learning_rate": 1.5216475780086669e-05, "loss": 0.7777, "step": 4301 }, { "epoch": 0.8276260100038476, "grad_norm": 0.20241670596734887, "learning_rate": 1.5183452570710522e-05, "loss": 0.7557, "step": 4302 }, { "epoch": 0.8278183916891112, "grad_norm": 0.17819716359533352, "learning_rate": 1.5150462290038392e-05, "loss": 0.789, "step": 4303 }, { "epoch": 0.8280107733743748, "grad_norm": 0.18017648039507514, "learning_rate": 1.511750495087827e-05, "loss": 0.7396, "step": 4304 }, { "epoch": 0.8282031550596384, "grad_norm": 0.19960881532891483, "learning_rate": 1.5084580566025308e-05, "loss": 0.8026, "step": 4305 }, { "epoch": 0.8283955367449019, "grad_norm": 0.19757605381571056, "learning_rate": 1.5051689148261894e-05, "loss": 0.741, "step": 4306 }, { "epoch": 0.8285879184301654, "grad_norm": 0.19331147784847777, "learning_rate": 1.5018830710357611e-05, "loss": 0.7839, "step": 4307 }, { "epoch": 0.828780300115429, "grad_norm": 0.21667010184124696, "learning_rate": 1.4986005265069204e-05, "loss": 0.7544, "step": 4308 }, { "epoch": 0.8289726818006926, "grad_norm": 0.1959266856209126, "learning_rate": 1.4953212825140727e-05, "loss": 0.7896, "step": 4309 }, { "epoch": 0.8291650634859561, "grad_norm": 0.20299022405667824, "learning_rate": 1.4920453403303247e-05, "loss": 0.7841, "step": 4310 }, { "epoch": 0.8293574451712197, "grad_norm": 0.19382948261150984, "learning_rate": 1.488772701227511e-05, "loss": 0.7719, "step": 4311 }, { "epoch": 0.8295498268564833, "grad_norm": 0.18116696595110368, "learning_rate": 1.4855033664761897e-05, "loss": 0.7787, "step": 4312 }, { "epoch": 0.8297422085417469, "grad_norm": 0.17993965420805408, "learning_rate": 1.48223733734562e-05, "loss": 0.7321, "step": 4313 }, { "epoch": 0.8299345902270104, "grad_norm": 0.19255546422930134, "learning_rate": 1.4789746151037942e-05, "loss": 0.7407, "step": 4314 }, { "epoch": 0.8301269719122739, "grad_norm": 0.1976827943393197, "learning_rate": 1.4757152010174069e-05, "loss": 0.7897, "step": 4315 }, { "epoch": 0.8303193535975375, "grad_norm": 0.19401998198798776, "learning_rate": 1.4724590963518803e-05, "loss": 0.8181, "step": 4316 }, { "epoch": 0.8305117352828011, "grad_norm": 0.19536858230713422, "learning_rate": 1.4692063023713443e-05, "loss": 0.7982, "step": 4317 }, { "epoch": 0.8307041169680647, "grad_norm": 0.1928241533319026, "learning_rate": 1.4659568203386465e-05, "loss": 0.7444, "step": 4318 }, { "epoch": 0.8308964986533282, "grad_norm": 0.21173741186385245, "learning_rate": 1.4627106515153455e-05, "loss": 0.7869, "step": 4319 }, { "epoch": 0.8310888803385917, "grad_norm": 0.2036282141563532, "learning_rate": 1.4594677971617177e-05, "loss": 0.752, "step": 4320 }, { "epoch": 0.8312812620238553, "grad_norm": 0.1891103168695411, "learning_rate": 1.4562282585367492e-05, "loss": 0.7874, "step": 4321 }, { "epoch": 0.8314736437091189, "grad_norm": 0.1970711395121686, "learning_rate": 1.452992036898142e-05, "loss": 0.7044, "step": 4322 }, { "epoch": 0.8316660253943825, "grad_norm": 0.19644008573201238, "learning_rate": 1.4497591335023086e-05, "loss": 0.7867, "step": 4323 }, { "epoch": 0.831858407079646, "grad_norm": 0.1810958804360067, "learning_rate": 1.4465295496043729e-05, "loss": 0.7369, "step": 4324 }, { "epoch": 0.8320507887649096, "grad_norm": 0.20994932487436418, "learning_rate": 1.4433032864581686e-05, "loss": 0.794, "step": 4325 }, { "epoch": 0.8322431704501732, "grad_norm": 0.19576221377147338, "learning_rate": 1.4400803453162482e-05, "loss": 0.7807, "step": 4326 }, { "epoch": 0.8324355521354367, "grad_norm": 0.17874805984155154, "learning_rate": 1.4368607274298595e-05, "loss": 0.8063, "step": 4327 }, { "epoch": 0.8326279338207002, "grad_norm": 0.18158215837544261, "learning_rate": 1.4336444340489775e-05, "loss": 0.7746, "step": 4328 }, { "epoch": 0.8328203155059638, "grad_norm": 0.1879265863566571, "learning_rate": 1.4304314664222728e-05, "loss": 0.7785, "step": 4329 }, { "epoch": 0.8330126971912274, "grad_norm": 0.19262041030737778, "learning_rate": 1.4272218257971326e-05, "loss": 0.8229, "step": 4330 }, { "epoch": 0.833205078876491, "grad_norm": 0.19427401273738706, "learning_rate": 1.4240155134196499e-05, "loss": 0.8229, "step": 4331 }, { "epoch": 0.8333974605617546, "grad_norm": 0.1908225902193861, "learning_rate": 1.4208125305346232e-05, "loss": 0.7983, "step": 4332 }, { "epoch": 0.833589842247018, "grad_norm": 0.21391887058167303, "learning_rate": 1.4176128783855636e-05, "loss": 0.7918, "step": 4333 }, { "epoch": 0.8337822239322816, "grad_norm": 0.19339711217881853, "learning_rate": 1.4144165582146817e-05, "loss": 0.7899, "step": 4334 }, { "epoch": 0.8339746056175452, "grad_norm": 0.1973128339382198, "learning_rate": 1.4112235712629064e-05, "loss": 0.7897, "step": 4335 }, { "epoch": 0.8341669873028088, "grad_norm": 0.18995017446296464, "learning_rate": 1.40803391876986e-05, "loss": 0.8105, "step": 4336 }, { "epoch": 0.8343593689880723, "grad_norm": 0.2160195341570423, "learning_rate": 1.4048476019738755e-05, "loss": 0.8102, "step": 4337 }, { "epoch": 0.8345517506733359, "grad_norm": 0.19259327974685062, "learning_rate": 1.401664622111991e-05, "loss": 0.7404, "step": 4338 }, { "epoch": 0.8347441323585995, "grad_norm": 0.19561544081711219, "learning_rate": 1.3984849804199485e-05, "loss": 0.7251, "step": 4339 }, { "epoch": 0.834936514043863, "grad_norm": 0.18716581022437315, "learning_rate": 1.3953086781321988e-05, "loss": 0.7926, "step": 4340 }, { "epoch": 0.8351288957291266, "grad_norm": 0.20315578805647, "learning_rate": 1.3921357164818848e-05, "loss": 0.7611, "step": 4341 }, { "epoch": 0.8353212774143901, "grad_norm": 0.18592899443851563, "learning_rate": 1.3889660967008655e-05, "loss": 0.7524, "step": 4342 }, { "epoch": 0.8355136590996537, "grad_norm": 0.18678069573020983, "learning_rate": 1.3857998200196941e-05, "loss": 0.7676, "step": 4343 }, { "epoch": 0.8357060407849173, "grad_norm": 0.18798251065940702, "learning_rate": 1.3826368876676278e-05, "loss": 0.7959, "step": 4344 }, { "epoch": 0.8358984224701809, "grad_norm": 0.22277584386411106, "learning_rate": 1.3794773008726259e-05, "loss": 0.8091, "step": 4345 }, { "epoch": 0.8360908041554445, "grad_norm": 0.19405302979305544, "learning_rate": 1.3763210608613498e-05, "loss": 0.7814, "step": 4346 }, { "epoch": 0.8362831858407079, "grad_norm": 0.18244654278821062, "learning_rate": 1.3731681688591591e-05, "loss": 0.804, "step": 4347 }, { "epoch": 0.8364755675259715, "grad_norm": 0.22084472918912548, "learning_rate": 1.370018626090116e-05, "loss": 0.7677, "step": 4348 }, { "epoch": 0.8366679492112351, "grad_norm": 0.21524474950161632, "learning_rate": 1.3668724337769823e-05, "loss": 0.8361, "step": 4349 }, { "epoch": 0.8368603308964987, "grad_norm": 0.20554301038038894, "learning_rate": 1.3637295931412152e-05, "loss": 0.8308, "step": 4350 }, { "epoch": 0.8370527125817622, "grad_norm": 0.18660672595838113, "learning_rate": 1.3605901054029746e-05, "loss": 0.8141, "step": 4351 }, { "epoch": 0.8372450942670258, "grad_norm": 0.18104552415672706, "learning_rate": 1.357453971781123e-05, "loss": 0.7864, "step": 4352 }, { "epoch": 0.8374374759522893, "grad_norm": 0.18363420239798314, "learning_rate": 1.3543211934932065e-05, "loss": 0.7916, "step": 4353 }, { "epoch": 0.8376298576375529, "grad_norm": 0.21142662261910414, "learning_rate": 1.3511917717554846e-05, "loss": 0.7896, "step": 4354 }, { "epoch": 0.8378222393228165, "grad_norm": 0.1893405406187269, "learning_rate": 1.348065707782904e-05, "loss": 0.7569, "step": 4355 }, { "epoch": 0.83801462100808, "grad_norm": 0.1828290086162219, "learning_rate": 1.3449430027891097e-05, "loss": 0.7828, "step": 4356 }, { "epoch": 0.8382070026933436, "grad_norm": 0.18937693393834604, "learning_rate": 1.341823657986445e-05, "loss": 0.8529, "step": 4357 }, { "epoch": 0.8383993843786072, "grad_norm": 0.20957127006694537, "learning_rate": 1.338707674585945e-05, "loss": 0.7912, "step": 4358 }, { "epoch": 0.8385917660638708, "grad_norm": 0.20571587675495143, "learning_rate": 1.335595053797344e-05, "loss": 0.8301, "step": 4359 }, { "epoch": 0.8387841477491342, "grad_norm": 0.19905261737062047, "learning_rate": 1.332485796829065e-05, "loss": 0.8337, "step": 4360 }, { "epoch": 0.8389765294343978, "grad_norm": 0.10470391189890113, "learning_rate": 1.329379904888235e-05, "loss": 0.5924, "step": 4361 }, { "epoch": 0.8391689111196614, "grad_norm": 0.16908582757542376, "learning_rate": 1.3262773791806615e-05, "loss": 0.783, "step": 4362 }, { "epoch": 0.839361292804925, "grad_norm": 0.1968067613285661, "learning_rate": 1.3231782209108545e-05, "loss": 0.7612, "step": 4363 }, { "epoch": 0.8395536744901886, "grad_norm": 0.19287200870327315, "learning_rate": 1.3200824312820136e-05, "loss": 0.8005, "step": 4364 }, { "epoch": 0.8397460561754521, "grad_norm": 0.1945589874440059, "learning_rate": 1.3169900114960298e-05, "loss": 0.8082, "step": 4365 }, { "epoch": 0.8399384378607156, "grad_norm": 0.1973322075772525, "learning_rate": 1.3139009627534927e-05, "loss": 0.8634, "step": 4366 }, { "epoch": 0.8401308195459792, "grad_norm": 0.17614680915580103, "learning_rate": 1.3108152862536682e-05, "loss": 0.7811, "step": 4367 }, { "epoch": 0.8403232012312428, "grad_norm": 0.20124879925939546, "learning_rate": 1.3077329831945295e-05, "loss": 0.7754, "step": 4368 }, { "epoch": 0.8405155829165063, "grad_norm": 0.2155429482018364, "learning_rate": 1.3046540547727303e-05, "loss": 0.7417, "step": 4369 }, { "epoch": 0.8407079646017699, "grad_norm": 0.19632150655318334, "learning_rate": 1.3015785021836158e-05, "loss": 0.8419, "step": 4370 }, { "epoch": 0.8409003462870335, "grad_norm": 0.19411615354435832, "learning_rate": 1.2985063266212227e-05, "loss": 0.7817, "step": 4371 }, { "epoch": 0.8410927279722971, "grad_norm": 0.20876149256083718, "learning_rate": 1.295437529278275e-05, "loss": 0.7606, "step": 4372 }, { "epoch": 0.8412851096575606, "grad_norm": 0.10715765493033937, "learning_rate": 1.2923721113461851e-05, "loss": 0.6344, "step": 4373 }, { "epoch": 0.8414774913428241, "grad_norm": 0.19635573349667612, "learning_rate": 1.289310074015052e-05, "loss": 0.7523, "step": 4374 }, { "epoch": 0.8416698730280877, "grad_norm": 0.19725950432415534, "learning_rate": 1.2862514184736696e-05, "loss": 0.7433, "step": 4375 }, { "epoch": 0.8418622547133513, "grad_norm": 0.23592172301754785, "learning_rate": 1.2831961459095088e-05, "loss": 0.7669, "step": 4376 }, { "epoch": 0.8420546363986149, "grad_norm": 0.20090169934461785, "learning_rate": 1.2801442575087298e-05, "loss": 0.7021, "step": 4377 }, { "epoch": 0.8422470180838784, "grad_norm": 0.1974328693231534, "learning_rate": 1.2770957544561869e-05, "loss": 0.7893, "step": 4378 }, { "epoch": 0.842439399769142, "grad_norm": 0.18963180042428432, "learning_rate": 1.2740506379354077e-05, "loss": 0.7806, "step": 4379 }, { "epoch": 0.8426317814544055, "grad_norm": 0.18745028192197968, "learning_rate": 1.2710089091286149e-05, "loss": 0.8268, "step": 4380 }, { "epoch": 0.8428241631396691, "grad_norm": 0.21197365879530478, "learning_rate": 1.267970569216712e-05, "loss": 0.7444, "step": 4381 }, { "epoch": 0.8430165448249327, "grad_norm": 0.17425037973252064, "learning_rate": 1.2649356193792872e-05, "loss": 0.7587, "step": 4382 }, { "epoch": 0.8432089265101962, "grad_norm": 0.18845124905421337, "learning_rate": 1.2619040607946119e-05, "loss": 0.7599, "step": 4383 }, { "epoch": 0.8434013081954598, "grad_norm": 0.21023222493271992, "learning_rate": 1.2588758946396418e-05, "loss": 0.8071, "step": 4384 }, { "epoch": 0.8435936898807234, "grad_norm": 0.1989825671824409, "learning_rate": 1.2558511220900138e-05, "loss": 0.8146, "step": 4385 }, { "epoch": 0.843786071565987, "grad_norm": 0.20194264612425136, "learning_rate": 1.2528297443200488e-05, "loss": 0.7755, "step": 4386 }, { "epoch": 0.8439784532512504, "grad_norm": 0.11109368982806009, "learning_rate": 1.2498117625027561e-05, "loss": 0.5945, "step": 4387 }, { "epoch": 0.844170834936514, "grad_norm": 0.2181305023126604, "learning_rate": 1.246797177809812e-05, "loss": 0.7791, "step": 4388 }, { "epoch": 0.8443632166217776, "grad_norm": 0.20245718192618928, "learning_rate": 1.2437859914115845e-05, "loss": 0.768, "step": 4389 }, { "epoch": 0.8445555983070412, "grad_norm": 0.19141413635210253, "learning_rate": 1.2407782044771221e-05, "loss": 0.7631, "step": 4390 }, { "epoch": 0.8447479799923048, "grad_norm": 0.18220468311980043, "learning_rate": 1.237773818174146e-05, "loss": 0.6998, "step": 4391 }, { "epoch": 0.8449403616775683, "grad_norm": 0.20128056319201793, "learning_rate": 1.2347728336690701e-05, "loss": 0.7623, "step": 4392 }, { "epoch": 0.8451327433628318, "grad_norm": 0.19738234030138976, "learning_rate": 1.2317752521269722e-05, "loss": 0.6842, "step": 4393 }, { "epoch": 0.8453251250480954, "grad_norm": 0.2052465956987967, "learning_rate": 1.2287810747116224e-05, "loss": 0.7924, "step": 4394 }, { "epoch": 0.845517506733359, "grad_norm": 0.2258722011241276, "learning_rate": 1.2257903025854612e-05, "loss": 0.6834, "step": 4395 }, { "epoch": 0.8457098884186226, "grad_norm": 0.18886904254564785, "learning_rate": 1.2228029369096094e-05, "loss": 0.8003, "step": 4396 }, { "epoch": 0.8459022701038861, "grad_norm": 0.19807293417538885, "learning_rate": 1.2198189788438651e-05, "loss": 0.7645, "step": 4397 }, { "epoch": 0.8460946517891497, "grad_norm": 0.19150374523961708, "learning_rate": 1.2168384295467038e-05, "loss": 0.7157, "step": 4398 }, { "epoch": 0.8462870334744133, "grad_norm": 0.18195917949946994, "learning_rate": 1.2138612901752777e-05, "loss": 0.7586, "step": 4399 }, { "epoch": 0.8464794151596768, "grad_norm": 0.12046242694915295, "learning_rate": 1.2108875618854121e-05, "loss": 0.6704, "step": 4400 }, { "epoch": 0.8466717968449403, "grad_norm": 0.18901459277443952, "learning_rate": 1.2079172458316168e-05, "loss": 0.7511, "step": 4401 }, { "epoch": 0.8468641785302039, "grad_norm": 0.2079138446553504, "learning_rate": 1.204950343167065e-05, "loss": 0.8105, "step": 4402 }, { "epoch": 0.8470565602154675, "grad_norm": 0.19453827465604379, "learning_rate": 1.2019868550436098e-05, "loss": 0.8277, "step": 4403 }, { "epoch": 0.8472489419007311, "grad_norm": 0.21801169316398977, "learning_rate": 1.1990267826117873e-05, "loss": 0.81, "step": 4404 }, { "epoch": 0.8474413235859947, "grad_norm": 0.18611621317077495, "learning_rate": 1.1960701270207886e-05, "loss": 0.7621, "step": 4405 }, { "epoch": 0.8476337052712581, "grad_norm": 0.20215929776237285, "learning_rate": 1.1931168894184974e-05, "loss": 0.7716, "step": 4406 }, { "epoch": 0.8478260869565217, "grad_norm": 0.1785826441756633, "learning_rate": 1.19016707095146e-05, "loss": 0.7949, "step": 4407 }, { "epoch": 0.8480184686417853, "grad_norm": 0.20677235201604155, "learning_rate": 1.1872206727648972e-05, "loss": 0.7676, "step": 4408 }, { "epoch": 0.8482108503270489, "grad_norm": 0.18413870478275782, "learning_rate": 1.1842776960027013e-05, "loss": 0.7857, "step": 4409 }, { "epoch": 0.8484032320123124, "grad_norm": 0.1955054340092222, "learning_rate": 1.1813381418074388e-05, "loss": 0.7832, "step": 4410 }, { "epoch": 0.848595613697576, "grad_norm": 0.10719481105644921, "learning_rate": 1.1784020113203454e-05, "loss": 0.6246, "step": 4411 }, { "epoch": 0.8487879953828396, "grad_norm": 0.18606897265285488, "learning_rate": 1.175469305681327e-05, "loss": 0.7037, "step": 4412 }, { "epoch": 0.8489803770681031, "grad_norm": 0.18534208196224067, "learning_rate": 1.1725400260289621e-05, "loss": 0.7483, "step": 4413 }, { "epoch": 0.8491727587533667, "grad_norm": 0.1986264991927893, "learning_rate": 1.1696141735004939e-05, "loss": 0.7203, "step": 4414 }, { "epoch": 0.8493651404386302, "grad_norm": 0.20166706473804147, "learning_rate": 1.1666917492318485e-05, "loss": 0.737, "step": 4415 }, { "epoch": 0.8495575221238938, "grad_norm": 0.1988005181838045, "learning_rate": 1.1637727543576027e-05, "loss": 0.7831, "step": 4416 }, { "epoch": 0.8497499038091574, "grad_norm": 0.19329858269106945, "learning_rate": 1.160857190011012e-05, "loss": 0.7663, "step": 4417 }, { "epoch": 0.849942285494421, "grad_norm": 0.20879564652162763, "learning_rate": 1.1579450573240058e-05, "loss": 0.7505, "step": 4418 }, { "epoch": 0.8501346671796844, "grad_norm": 0.19320776273253867, "learning_rate": 1.1550363574271639e-05, "loss": 0.8092, "step": 4419 }, { "epoch": 0.850327048864948, "grad_norm": 0.18524910417077003, "learning_rate": 1.1521310914497519e-05, "loss": 0.7882, "step": 4420 }, { "epoch": 0.8505194305502116, "grad_norm": 0.1908909238172279, "learning_rate": 1.1492292605196907e-05, "loss": 0.7577, "step": 4421 }, { "epoch": 0.8507118122354752, "grad_norm": 0.19306641117985593, "learning_rate": 1.1463308657635719e-05, "loss": 0.8243, "step": 4422 }, { "epoch": 0.8509041939207388, "grad_norm": 0.18796278612731165, "learning_rate": 1.1434359083066515e-05, "loss": 0.801, "step": 4423 }, { "epoch": 0.8510965756060023, "grad_norm": 0.21495626814411112, "learning_rate": 1.1405443892728528e-05, "loss": 0.798, "step": 4424 }, { "epoch": 0.8512889572912659, "grad_norm": 0.20599926593584728, "learning_rate": 1.1376563097847614e-05, "loss": 0.731, "step": 4425 }, { "epoch": 0.8514813389765294, "grad_norm": 0.19456792341649012, "learning_rate": 1.134771670963628e-05, "loss": 0.7562, "step": 4426 }, { "epoch": 0.851673720661793, "grad_norm": 0.2045467275242301, "learning_rate": 1.1318904739293745e-05, "loss": 0.7391, "step": 4427 }, { "epoch": 0.8518661023470565, "grad_norm": 0.19856991358893672, "learning_rate": 1.1290127198005752e-05, "loss": 0.7711, "step": 4428 }, { "epoch": 0.8520584840323201, "grad_norm": 0.22449043909064825, "learning_rate": 1.1261384096944727e-05, "loss": 0.7408, "step": 4429 }, { "epoch": 0.8522508657175837, "grad_norm": 0.20493416238391085, "learning_rate": 1.1232675447269803e-05, "loss": 0.7413, "step": 4430 }, { "epoch": 0.8524432474028473, "grad_norm": 0.18224130571012392, "learning_rate": 1.1204001260126574e-05, "loss": 0.7633, "step": 4431 }, { "epoch": 0.8526356290881109, "grad_norm": 0.21862700575498448, "learning_rate": 1.1175361546647412e-05, "loss": 0.7523, "step": 4432 }, { "epoch": 0.8528280107733743, "grad_norm": 0.19477946329765228, "learning_rate": 1.1146756317951224e-05, "loss": 0.7646, "step": 4433 }, { "epoch": 0.8530203924586379, "grad_norm": 0.2079391640751331, "learning_rate": 1.1118185585143537e-05, "loss": 0.7912, "step": 4434 }, { "epoch": 0.8532127741439015, "grad_norm": 0.1996332925000733, "learning_rate": 1.1089649359316501e-05, "loss": 0.7822, "step": 4435 }, { "epoch": 0.8534051558291651, "grad_norm": 0.17547768961629726, "learning_rate": 1.1061147651548853e-05, "loss": 0.7903, "step": 4436 }, { "epoch": 0.8535975375144286, "grad_norm": 0.21866452486783677, "learning_rate": 1.1032680472905932e-05, "loss": 0.7559, "step": 4437 }, { "epoch": 0.8537899191996922, "grad_norm": 0.22645548134665794, "learning_rate": 1.1004247834439696e-05, "loss": 0.7906, "step": 4438 }, { "epoch": 0.8539823008849557, "grad_norm": 0.19787704171158566, "learning_rate": 1.097584974718866e-05, "loss": 0.8018, "step": 4439 }, { "epoch": 0.8541746825702193, "grad_norm": 0.19964266914306444, "learning_rate": 1.0947486222177928e-05, "loss": 0.78, "step": 4440 }, { "epoch": 0.8543670642554829, "grad_norm": 0.2070547580267456, "learning_rate": 1.0919157270419256e-05, "loss": 0.7333, "step": 4441 }, { "epoch": 0.8545594459407464, "grad_norm": 0.18268207742183493, "learning_rate": 1.0890862902910847e-05, "loss": 0.7827, "step": 4442 }, { "epoch": 0.85475182762601, "grad_norm": 0.1965886895771076, "learning_rate": 1.0862603130637561e-05, "loss": 0.7705, "step": 4443 }, { "epoch": 0.8549442093112736, "grad_norm": 0.2141725881967245, "learning_rate": 1.0834377964570864e-05, "loss": 0.7373, "step": 4444 }, { "epoch": 0.8551365909965372, "grad_norm": 0.19635034550853533, "learning_rate": 1.0806187415668667e-05, "loss": 0.7693, "step": 4445 }, { "epoch": 0.8553289726818007, "grad_norm": 0.19477408126717013, "learning_rate": 1.0778031494875573e-05, "loss": 0.7572, "step": 4446 }, { "epoch": 0.8555213543670642, "grad_norm": 0.20344900990198067, "learning_rate": 1.0749910213122649e-05, "loss": 0.7351, "step": 4447 }, { "epoch": 0.8557137360523278, "grad_norm": 0.19111212864784818, "learning_rate": 1.072182358132755e-05, "loss": 0.719, "step": 4448 }, { "epoch": 0.8559061177375914, "grad_norm": 0.18390283580712868, "learning_rate": 1.0693771610394476e-05, "loss": 0.7555, "step": 4449 }, { "epoch": 0.856098499422855, "grad_norm": 0.21640397533977496, "learning_rate": 1.0665754311214172e-05, "loss": 0.8084, "step": 4450 }, { "epoch": 0.8562908811081185, "grad_norm": 0.18889035799907603, "learning_rate": 1.0637771694663901e-05, "loss": 0.7885, "step": 4451 }, { "epoch": 0.856483262793382, "grad_norm": 0.20085660260440702, "learning_rate": 1.0609823771607485e-05, "loss": 0.7778, "step": 4452 }, { "epoch": 0.8566756444786456, "grad_norm": 0.18022620301294034, "learning_rate": 1.0581910552895302e-05, "loss": 0.7971, "step": 4453 }, { "epoch": 0.8568680261639092, "grad_norm": 0.19294939304878175, "learning_rate": 1.055403204936416e-05, "loss": 0.7956, "step": 4454 }, { "epoch": 0.8570604078491728, "grad_norm": 0.2034145618837739, "learning_rate": 1.0526188271837511e-05, "loss": 0.7801, "step": 4455 }, { "epoch": 0.8572527895344363, "grad_norm": 0.1995788704436257, "learning_rate": 1.0498379231125278e-05, "loss": 0.8018, "step": 4456 }, { "epoch": 0.8574451712196999, "grad_norm": 0.1820264417877339, "learning_rate": 1.0470604938023808e-05, "loss": 0.7935, "step": 4457 }, { "epoch": 0.8576375529049635, "grad_norm": 0.19293056112071563, "learning_rate": 1.0442865403316116e-05, "loss": 0.7202, "step": 4458 }, { "epoch": 0.857829934590227, "grad_norm": 0.11296622055744705, "learning_rate": 1.0415160637771603e-05, "loss": 0.6788, "step": 4459 }, { "epoch": 0.8580223162754905, "grad_norm": 0.18863176240480672, "learning_rate": 1.0387490652146237e-05, "loss": 0.8516, "step": 4460 }, { "epoch": 0.8582146979607541, "grad_norm": 0.2012105470617281, "learning_rate": 1.0359855457182454e-05, "loss": 0.7469, "step": 4461 }, { "epoch": 0.8584070796460177, "grad_norm": 0.19887029559819336, "learning_rate": 1.0332255063609175e-05, "loss": 0.7916, "step": 4462 }, { "epoch": 0.8585994613312813, "grad_norm": 0.19213160763405485, "learning_rate": 1.0304689482141838e-05, "loss": 0.8095, "step": 4463 }, { "epoch": 0.8587918430165449, "grad_norm": 0.1854268277109583, "learning_rate": 1.0277158723482339e-05, "loss": 0.749, "step": 4464 }, { "epoch": 0.8589842247018084, "grad_norm": 0.2045339710692265, "learning_rate": 1.024966279831907e-05, "loss": 0.7734, "step": 4465 }, { "epoch": 0.8591766063870719, "grad_norm": 0.21770283313210287, "learning_rate": 1.0222201717326885e-05, "loss": 0.7558, "step": 4466 }, { "epoch": 0.8593689880723355, "grad_norm": 0.1938978382597789, "learning_rate": 1.0194775491167163e-05, "loss": 0.7712, "step": 4467 }, { "epoch": 0.8595613697575991, "grad_norm": 0.19905433573780268, "learning_rate": 1.0167384130487667e-05, "loss": 0.79, "step": 4468 }, { "epoch": 0.8597537514428626, "grad_norm": 0.2035658371687959, "learning_rate": 1.0140027645922657e-05, "loss": 0.7257, "step": 4469 }, { "epoch": 0.8599461331281262, "grad_norm": 0.21383273495910957, "learning_rate": 1.0112706048092923e-05, "loss": 0.7938, "step": 4470 }, { "epoch": 0.8601385148133898, "grad_norm": 0.19952322486004134, "learning_rate": 1.0085419347605574e-05, "loss": 0.7936, "step": 4471 }, { "epoch": 0.8603308964986534, "grad_norm": 0.180866576944775, "learning_rate": 1.0058167555054298e-05, "loss": 0.7677, "step": 4472 }, { "epoch": 0.8605232781839169, "grad_norm": 0.19719837924598319, "learning_rate": 1.003095068101917e-05, "loss": 0.7146, "step": 4473 }, { "epoch": 0.8607156598691804, "grad_norm": 0.20736788314950128, "learning_rate": 1.0003768736066722e-05, "loss": 0.7377, "step": 4474 }, { "epoch": 0.860908041554444, "grad_norm": 0.19724224497486995, "learning_rate": 9.97662173074989e-06, "loss": 0.7832, "step": 4475 }, { "epoch": 0.8611004232397076, "grad_norm": 0.20902615595712914, "learning_rate": 9.949509675608115e-06, "loss": 0.774, "step": 4476 }, { "epoch": 0.8612928049249712, "grad_norm": 0.18551209402984364, "learning_rate": 9.922432581167207e-06, "loss": 0.8029, "step": 4477 }, { "epoch": 0.8614851866102347, "grad_norm": 0.197645889883799, "learning_rate": 9.895390457939413e-06, "loss": 0.795, "step": 4478 }, { "epoch": 0.8616775682954982, "grad_norm": 0.18376020006983634, "learning_rate": 9.86838331642348e-06, "loss": 0.755, "step": 4479 }, { "epoch": 0.8618699499807618, "grad_norm": 0.20407595071342044, "learning_rate": 9.84141116710442e-06, "loss": 0.7815, "step": 4480 }, { "epoch": 0.8620623316660254, "grad_norm": 0.21507671845014145, "learning_rate": 9.814474020453822e-06, "loss": 0.7036, "step": 4481 }, { "epoch": 0.862254713351289, "grad_norm": 0.19112802984714997, "learning_rate": 9.787571886929603e-06, "loss": 0.8219, "step": 4482 }, { "epoch": 0.8624470950365525, "grad_norm": 0.18842139640874606, "learning_rate": 9.760704776976049e-06, "loss": 0.7293, "step": 4483 }, { "epoch": 0.8626394767218161, "grad_norm": 0.19461604528166135, "learning_rate": 9.733872701023939e-06, "loss": 0.7897, "step": 4484 }, { "epoch": 0.8628318584070797, "grad_norm": 0.20183707232048778, "learning_rate": 9.707075669490407e-06, "loss": 0.7846, "step": 4485 }, { "epoch": 0.8630242400923432, "grad_norm": 0.1893175112021845, "learning_rate": 9.680313692778974e-06, "loss": 0.7333, "step": 4486 }, { "epoch": 0.8632166217776067, "grad_norm": 0.2098912713247036, "learning_rate": 9.653586781279567e-06, "loss": 0.7399, "step": 4487 }, { "epoch": 0.8634090034628703, "grad_norm": 0.20125128949072188, "learning_rate": 9.626894945368491e-06, "loss": 0.7519, "step": 4488 }, { "epoch": 0.8636013851481339, "grad_norm": 0.19248194498503549, "learning_rate": 9.600238195408428e-06, "loss": 0.7929, "step": 4489 }, { "epoch": 0.8637937668333975, "grad_norm": 0.19405367652298422, "learning_rate": 9.573616541748464e-06, "loss": 0.7703, "step": 4490 }, { "epoch": 0.8639861485186611, "grad_norm": 0.188703684359287, "learning_rate": 9.547029994724022e-06, "loss": 0.7236, "step": 4491 }, { "epoch": 0.8641785302039245, "grad_norm": 0.20285865436243167, "learning_rate": 9.520478564656899e-06, "loss": 0.7339, "step": 4492 }, { "epoch": 0.8643709118891881, "grad_norm": 0.1952821673693212, "learning_rate": 9.493962261855349e-06, "loss": 0.7613, "step": 4493 }, { "epoch": 0.8645632935744517, "grad_norm": 0.20106424905637407, "learning_rate": 9.467481096613828e-06, "loss": 0.7725, "step": 4494 }, { "epoch": 0.8647556752597153, "grad_norm": 0.211044029579712, "learning_rate": 9.441035079213267e-06, "loss": 0.8185, "step": 4495 }, { "epoch": 0.8649480569449788, "grad_norm": 0.19510758399723638, "learning_rate": 9.414624219920953e-06, "loss": 0.7676, "step": 4496 }, { "epoch": 0.8651404386302424, "grad_norm": 0.19239473173228272, "learning_rate": 9.388248528990428e-06, "loss": 0.7389, "step": 4497 }, { "epoch": 0.865332820315506, "grad_norm": 0.2065013979365267, "learning_rate": 9.361908016661702e-06, "loss": 0.7727, "step": 4498 }, { "epoch": 0.8655252020007695, "grad_norm": 0.10560754397345348, "learning_rate": 9.335602693161038e-06, "loss": 0.6151, "step": 4499 }, { "epoch": 0.8657175836860331, "grad_norm": 0.18297134242969187, "learning_rate": 9.309332568701079e-06, "loss": 0.7828, "step": 4500 }, { "epoch": 0.8659099653712966, "grad_norm": 0.18904044167918727, "learning_rate": 9.283097653480789e-06, "loss": 0.7567, "step": 4501 }, { "epoch": 0.8661023470565602, "grad_norm": 0.18789997351907825, "learning_rate": 9.256897957685462e-06, "loss": 0.8339, "step": 4502 }, { "epoch": 0.8662947287418238, "grad_norm": 0.18921428360827067, "learning_rate": 9.23073349148672e-06, "loss": 0.7625, "step": 4503 }, { "epoch": 0.8664871104270874, "grad_norm": 0.17998975636287143, "learning_rate": 9.204604265042505e-06, "loss": 0.779, "step": 4504 }, { "epoch": 0.866679492112351, "grad_norm": 0.18755927000978329, "learning_rate": 9.178510288497122e-06, "loss": 0.7289, "step": 4505 }, { "epoch": 0.8668718737976144, "grad_norm": 0.23687559513302578, "learning_rate": 9.15245157198108e-06, "loss": 0.7874, "step": 4506 }, { "epoch": 0.867064255482878, "grad_norm": 0.21081522631347574, "learning_rate": 9.126428125611342e-06, "loss": 0.8083, "step": 4507 }, { "epoch": 0.8672566371681416, "grad_norm": 0.18621686829232237, "learning_rate": 9.10043995949108e-06, "loss": 0.7201, "step": 4508 }, { "epoch": 0.8674490188534052, "grad_norm": 0.19517393896167823, "learning_rate": 9.07448708370976e-06, "loss": 0.8035, "step": 4509 }, { "epoch": 0.8676414005386687, "grad_norm": 0.18732497337533258, "learning_rate": 9.04856950834323e-06, "loss": 0.8208, "step": 4510 }, { "epoch": 0.8678337822239323, "grad_norm": 0.20541681538587375, "learning_rate": 9.022687243453554e-06, "loss": 0.7689, "step": 4511 }, { "epoch": 0.8680261639091958, "grad_norm": 0.20975837404981398, "learning_rate": 8.996840299089149e-06, "loss": 0.7565, "step": 4512 }, { "epoch": 0.8682185455944594, "grad_norm": 0.18973459944299687, "learning_rate": 8.971028685284655e-06, "loss": 0.7685, "step": 4513 }, { "epoch": 0.868410927279723, "grad_norm": 0.19131555390703198, "learning_rate": 8.945252412061055e-06, "loss": 0.7651, "step": 4514 }, { "epoch": 0.8686033089649865, "grad_norm": 0.20911815669309386, "learning_rate": 8.91951148942557e-06, "loss": 0.8136, "step": 4515 }, { "epoch": 0.8687956906502501, "grad_norm": 0.2090803994254754, "learning_rate": 8.893805927371723e-06, "loss": 0.7973, "step": 4516 }, { "epoch": 0.8689880723355137, "grad_norm": 0.1848967671977474, "learning_rate": 8.86813573587929e-06, "loss": 0.8043, "step": 4517 }, { "epoch": 0.8691804540207773, "grad_norm": 0.19747436210363217, "learning_rate": 8.842500924914298e-06, "loss": 0.832, "step": 4518 }, { "epoch": 0.8693728357060407, "grad_norm": 0.19904788914249127, "learning_rate": 8.816901504429143e-06, "loss": 0.7572, "step": 4519 }, { "epoch": 0.8695652173913043, "grad_norm": 0.19291159934934496, "learning_rate": 8.791337484362306e-06, "loss": 0.7598, "step": 4520 }, { "epoch": 0.8697575990765679, "grad_norm": 0.1924002969461419, "learning_rate": 8.765808874638681e-06, "loss": 0.7393, "step": 4521 }, { "epoch": 0.8699499807618315, "grad_norm": 0.1765793170428404, "learning_rate": 8.740315685169365e-06, "loss": 0.7728, "step": 4522 }, { "epoch": 0.8701423624470951, "grad_norm": 0.1991678990430213, "learning_rate": 8.714857925851615e-06, "loss": 0.7632, "step": 4523 }, { "epoch": 0.8703347441323586, "grad_norm": 0.19518428593444248, "learning_rate": 8.689435606569085e-06, "loss": 0.8346, "step": 4524 }, { "epoch": 0.8705271258176221, "grad_norm": 0.19431802485189092, "learning_rate": 8.664048737191566e-06, "loss": 0.8019, "step": 4525 }, { "epoch": 0.8707195075028857, "grad_norm": 0.18537198863060894, "learning_rate": 8.638697327575108e-06, "loss": 0.8064, "step": 4526 }, { "epoch": 0.8709118891881493, "grad_norm": 0.1924404092730313, "learning_rate": 8.613381387562015e-06, "loss": 0.7994, "step": 4527 }, { "epoch": 0.8711042708734128, "grad_norm": 0.19302207255235962, "learning_rate": 8.588100926980802e-06, "loss": 0.7675, "step": 4528 }, { "epoch": 0.8712966525586764, "grad_norm": 0.21526523086741792, "learning_rate": 8.56285595564621e-06, "loss": 0.7823, "step": 4529 }, { "epoch": 0.87148903424394, "grad_norm": 0.2071391144558886, "learning_rate": 8.537646483359186e-06, "loss": 0.6944, "step": 4530 }, { "epoch": 0.8716814159292036, "grad_norm": 0.177114109960709, "learning_rate": 8.512472519906978e-06, "loss": 0.8103, "step": 4531 }, { "epoch": 0.8718737976144671, "grad_norm": 0.21748166761805462, "learning_rate": 8.487334075062913e-06, "loss": 0.8164, "step": 4532 }, { "epoch": 0.8720661792997306, "grad_norm": 0.18590716436856203, "learning_rate": 8.462231158586653e-06, "loss": 0.7449, "step": 4533 }, { "epoch": 0.8722585609849942, "grad_norm": 0.10860389054791555, "learning_rate": 8.437163780224011e-06, "loss": 0.6419, "step": 4534 }, { "epoch": 0.8724509426702578, "grad_norm": 0.1940609132096833, "learning_rate": 8.41213194970696e-06, "loss": 0.8251, "step": 4535 }, { "epoch": 0.8726433243555214, "grad_norm": 0.21996313258976719, "learning_rate": 8.387135676753755e-06, "loss": 0.7701, "step": 4536 }, { "epoch": 0.8728357060407849, "grad_norm": 0.20826830549179756, "learning_rate": 8.362174971068803e-06, "loss": 0.7994, "step": 4537 }, { "epoch": 0.8730280877260485, "grad_norm": 0.18494984178499607, "learning_rate": 8.33724984234272e-06, "loss": 0.7488, "step": 4538 }, { "epoch": 0.873220469411312, "grad_norm": 0.19806648763341436, "learning_rate": 8.312360300252286e-06, "loss": 0.7684, "step": 4539 }, { "epoch": 0.8734128510965756, "grad_norm": 0.19131844148787466, "learning_rate": 8.287506354460484e-06, "loss": 0.7996, "step": 4540 }, { "epoch": 0.8736052327818392, "grad_norm": 0.19326017983909902, "learning_rate": 8.26268801461646e-06, "loss": 0.7737, "step": 4541 }, { "epoch": 0.8737976144671027, "grad_norm": 0.1930338838623802, "learning_rate": 8.237905290355564e-06, "loss": 0.762, "step": 4542 }, { "epoch": 0.8739899961523663, "grad_norm": 0.18731373082348163, "learning_rate": 8.213158191299297e-06, "loss": 0.7659, "step": 4543 }, { "epoch": 0.8741823778376299, "grad_norm": 0.2108179328393187, "learning_rate": 8.18844672705531e-06, "loss": 0.738, "step": 4544 }, { "epoch": 0.8743747595228935, "grad_norm": 0.20567489873683856, "learning_rate": 8.163770907217505e-06, "loss": 0.792, "step": 4545 }, { "epoch": 0.8745671412081569, "grad_norm": 0.19860125674180668, "learning_rate": 8.13913074136582e-06, "loss": 0.7893, "step": 4546 }, { "epoch": 0.8747595228934205, "grad_norm": 0.18251160905214356, "learning_rate": 8.114526239066455e-06, "loss": 0.7871, "step": 4547 }, { "epoch": 0.8749519045786841, "grad_norm": 0.20023053235252114, "learning_rate": 8.08995740987173e-06, "loss": 0.7668, "step": 4548 }, { "epoch": 0.8751442862639477, "grad_norm": 0.17744432331273233, "learning_rate": 8.065424263320054e-06, "loss": 0.7553, "step": 4549 }, { "epoch": 0.8753366679492113, "grad_norm": 0.18431219473115476, "learning_rate": 8.040926808936111e-06, "loss": 0.7467, "step": 4550 }, { "epoch": 0.8755290496344748, "grad_norm": 0.1891241064536342, "learning_rate": 8.016465056230615e-06, "loss": 0.7864, "step": 4551 }, { "epoch": 0.8757214313197383, "grad_norm": 0.19923827998553587, "learning_rate": 7.99203901470047e-06, "loss": 0.7727, "step": 4552 }, { "epoch": 0.8759138130050019, "grad_norm": 0.2067887240310482, "learning_rate": 7.967648693828712e-06, "loss": 0.7432, "step": 4553 }, { "epoch": 0.8761061946902655, "grad_norm": 0.17542841848249932, "learning_rate": 7.943294103084487e-06, "loss": 0.7359, "step": 4554 }, { "epoch": 0.8762985763755291, "grad_norm": 0.18794246349518617, "learning_rate": 7.918975251923099e-06, "loss": 0.7093, "step": 4555 }, { "epoch": 0.8764909580607926, "grad_norm": 0.19275428303654965, "learning_rate": 7.894692149785954e-06, "loss": 0.7637, "step": 4556 }, { "epoch": 0.8766833397460562, "grad_norm": 0.18530200442409778, "learning_rate": 7.87044480610062e-06, "loss": 0.7554, "step": 4557 }, { "epoch": 0.8768757214313198, "grad_norm": 0.20050953923453818, "learning_rate": 7.846233230280697e-06, "loss": 0.7457, "step": 4558 }, { "epoch": 0.8770681031165833, "grad_norm": 0.20336574123677917, "learning_rate": 7.822057431725993e-06, "loss": 0.7812, "step": 4559 }, { "epoch": 0.8772604848018468, "grad_norm": 0.1925442747845435, "learning_rate": 7.797917419822375e-06, "loss": 0.7658, "step": 4560 }, { "epoch": 0.8774528664871104, "grad_norm": 0.22411731990137607, "learning_rate": 7.773813203941826e-06, "loss": 0.8031, "step": 4561 }, { "epoch": 0.877645248172374, "grad_norm": 0.18617265446981476, "learning_rate": 7.749744793442448e-06, "loss": 0.7777, "step": 4562 }, { "epoch": 0.8778376298576376, "grad_norm": 0.19995981058551837, "learning_rate": 7.725712197668378e-06, "loss": 0.7414, "step": 4563 }, { "epoch": 0.8780300115429012, "grad_norm": 0.19336477822175915, "learning_rate": 7.70171542594995e-06, "loss": 0.7607, "step": 4564 }, { "epoch": 0.8782223932281646, "grad_norm": 0.18791110960687726, "learning_rate": 7.677754487603517e-06, "loss": 0.7223, "step": 4565 }, { "epoch": 0.8784147749134282, "grad_norm": 0.19795878789781368, "learning_rate": 7.653829391931534e-06, "loss": 0.6697, "step": 4566 }, { "epoch": 0.8786071565986918, "grad_norm": 0.20720763267467132, "learning_rate": 7.629940148222559e-06, "loss": 0.7164, "step": 4567 }, { "epoch": 0.8787995382839554, "grad_norm": 0.18110632179175415, "learning_rate": 7.606086765751208e-06, "loss": 0.7336, "step": 4568 }, { "epoch": 0.8789919199692189, "grad_norm": 0.19021460458924103, "learning_rate": 7.582269253778185e-06, "loss": 0.8092, "step": 4569 }, { "epoch": 0.8791843016544825, "grad_norm": 0.20914366248193098, "learning_rate": 7.5584876215502695e-06, "loss": 0.7757, "step": 4570 }, { "epoch": 0.8793766833397461, "grad_norm": 0.18678327969946681, "learning_rate": 7.534741878300333e-06, "loss": 0.7372, "step": 4571 }, { "epoch": 0.8795690650250096, "grad_norm": 0.18379231322215284, "learning_rate": 7.511032033247256e-06, "loss": 0.7664, "step": 4572 }, { "epoch": 0.8797614467102732, "grad_norm": 0.21154364687638014, "learning_rate": 7.487358095596031e-06, "loss": 0.8516, "step": 4573 }, { "epoch": 0.8799538283955367, "grad_norm": 0.21516486762058154, "learning_rate": 7.463720074537728e-06, "loss": 0.8064, "step": 4574 }, { "epoch": 0.8801462100808003, "grad_norm": 0.18604382059147812, "learning_rate": 7.440117979249362e-06, "loss": 0.7915, "step": 4575 }, { "epoch": 0.8803385917660639, "grad_norm": 0.1965846361948933, "learning_rate": 7.416551818894157e-06, "loss": 0.8149, "step": 4576 }, { "epoch": 0.8805309734513275, "grad_norm": 0.18222726372857442, "learning_rate": 7.393021602621264e-06, "loss": 0.7746, "step": 4577 }, { "epoch": 0.880723355136591, "grad_norm": 0.20245304014607304, "learning_rate": 7.3695273395659515e-06, "loss": 0.7753, "step": 4578 }, { "epoch": 0.8809157368218545, "grad_norm": 0.1891857784114552, "learning_rate": 7.346069038849468e-06, "loss": 0.7487, "step": 4579 }, { "epoch": 0.8811081185071181, "grad_norm": 0.1862621822412771, "learning_rate": 7.322646709579173e-06, "loss": 0.7203, "step": 4580 }, { "epoch": 0.8813005001923817, "grad_norm": 0.19985519736945798, "learning_rate": 7.299260360848381e-06, "loss": 0.7783, "step": 4581 }, { "epoch": 0.8814928818776453, "grad_norm": 0.2041224797575403, "learning_rate": 7.275910001736497e-06, "loss": 0.7674, "step": 4582 }, { "epoch": 0.8816852635629088, "grad_norm": 0.18485359709359533, "learning_rate": 7.2525956413089565e-06, "loss": 0.8109, "step": 4583 }, { "epoch": 0.8818776452481724, "grad_norm": 0.20553142616321554, "learning_rate": 7.2293172886171435e-06, "loss": 0.7795, "step": 4584 }, { "epoch": 0.882070026933436, "grad_norm": 0.20125252820284945, "learning_rate": 7.206074952698561e-06, "loss": 0.7022, "step": 4585 }, { "epoch": 0.8822624086186995, "grad_norm": 0.18067019049643965, "learning_rate": 7.1828686425766785e-06, "loss": 0.7414, "step": 4586 }, { "epoch": 0.882454790303963, "grad_norm": 0.2093495436357559, "learning_rate": 7.15969836726097e-06, "loss": 0.7709, "step": 4587 }, { "epoch": 0.8826471719892266, "grad_norm": 0.19441992383900764, "learning_rate": 7.136564135746959e-06, "loss": 0.7398, "step": 4588 }, { "epoch": 0.8828395536744902, "grad_norm": 0.19076440471948902, "learning_rate": 7.113465957016097e-06, "loss": 0.8013, "step": 4589 }, { "epoch": 0.8830319353597538, "grad_norm": 0.18679902026113782, "learning_rate": 7.090403840035942e-06, "loss": 0.7634, "step": 4590 }, { "epoch": 0.8832243170450174, "grad_norm": 0.1933792556482483, "learning_rate": 7.067377793759999e-06, "loss": 0.7678, "step": 4591 }, { "epoch": 0.8834166987302808, "grad_norm": 0.2010873387854565, "learning_rate": 7.044387827127752e-06, "loss": 0.7899, "step": 4592 }, { "epoch": 0.8836090804155444, "grad_norm": 0.18224772349045687, "learning_rate": 7.0214339490647045e-06, "loss": 0.7689, "step": 4593 }, { "epoch": 0.883801462100808, "grad_norm": 0.2046499015356077, "learning_rate": 6.99851616848235e-06, "loss": 0.7672, "step": 4594 }, { "epoch": 0.8839938437860716, "grad_norm": 0.20877963788903328, "learning_rate": 6.975634494278149e-06, "loss": 0.7451, "step": 4595 }, { "epoch": 0.8841862254713351, "grad_norm": 0.19789471061564437, "learning_rate": 6.952788935335541e-06, "loss": 0.7605, "step": 4596 }, { "epoch": 0.8843786071565987, "grad_norm": 0.19441469189631955, "learning_rate": 6.92997950052402e-06, "loss": 0.7836, "step": 4597 }, { "epoch": 0.8845709888418622, "grad_norm": 0.18683019841981235, "learning_rate": 6.907206198698913e-06, "loss": 0.7723, "step": 4598 }, { "epoch": 0.8847633705271258, "grad_norm": 0.18738616030154015, "learning_rate": 6.884469038701646e-06, "loss": 0.7603, "step": 4599 }, { "epoch": 0.8849557522123894, "grad_norm": 0.19619546595044765, "learning_rate": 6.861768029359594e-06, "loss": 0.7889, "step": 4600 }, { "epoch": 0.8851481338976529, "grad_norm": 0.1813741711476252, "learning_rate": 6.839103179485995e-06, "loss": 0.7915, "step": 4601 }, { "epoch": 0.8853405155829165, "grad_norm": 0.19468223583876085, "learning_rate": 6.8164744978801765e-06, "loss": 0.7757, "step": 4602 }, { "epoch": 0.8855328972681801, "grad_norm": 0.1923416921535734, "learning_rate": 6.793881993327367e-06, "loss": 0.7459, "step": 4603 }, { "epoch": 0.8857252789534437, "grad_norm": 0.19628407952707827, "learning_rate": 6.771325674598749e-06, "loss": 0.7657, "step": 4604 }, { "epoch": 0.8859176606387072, "grad_norm": 0.19020516841480986, "learning_rate": 6.748805550451454e-06, "loss": 0.7787, "step": 4605 }, { "epoch": 0.8861100423239707, "grad_norm": 0.19937677732493206, "learning_rate": 6.726321629628585e-06, "loss": 0.8092, "step": 4606 }, { "epoch": 0.8863024240092343, "grad_norm": 0.19496248727358698, "learning_rate": 6.7038739208591605e-06, "loss": 0.7555, "step": 4607 }, { "epoch": 0.8864948056944979, "grad_norm": 0.17202209421588469, "learning_rate": 6.681462432858154e-06, "loss": 0.7809, "step": 4608 }, { "epoch": 0.8866871873797615, "grad_norm": 0.10483604307812568, "learning_rate": 6.659087174326506e-06, "loss": 0.6279, "step": 4609 }, { "epoch": 0.886879569065025, "grad_norm": 0.20175933083576061, "learning_rate": 6.636748153951e-06, "loss": 0.7768, "step": 4610 }, { "epoch": 0.8870719507502886, "grad_norm": 0.11010225011173323, "learning_rate": 6.614445380404477e-06, "loss": 0.635, "step": 4611 }, { "epoch": 0.8872643324355521, "grad_norm": 0.19208803051745857, "learning_rate": 6.5921788623456216e-06, "loss": 0.7258, "step": 4612 }, { "epoch": 0.8874567141208157, "grad_norm": 0.20058140375081535, "learning_rate": 6.569948608419041e-06, "loss": 0.7893, "step": 4613 }, { "epoch": 0.8876490958060793, "grad_norm": 0.19321390146565867, "learning_rate": 6.547754627255331e-06, "loss": 0.7632, "step": 4614 }, { "epoch": 0.8878414774913428, "grad_norm": 0.19953492919145915, "learning_rate": 6.5255969274708896e-06, "loss": 0.7995, "step": 4615 }, { "epoch": 0.8880338591766064, "grad_norm": 0.21993188120148144, "learning_rate": 6.503475517668167e-06, "loss": 0.7802, "step": 4616 }, { "epoch": 0.88822624086187, "grad_norm": 0.19534923176333724, "learning_rate": 6.481390406435417e-06, "loss": 0.7596, "step": 4617 }, { "epoch": 0.8884186225471336, "grad_norm": 0.21608964700264677, "learning_rate": 6.4593416023468575e-06, "loss": 0.8094, "step": 4618 }, { "epoch": 0.888611004232397, "grad_norm": 0.10080825433703194, "learning_rate": 6.437329113962576e-06, "loss": 0.6311, "step": 4619 }, { "epoch": 0.8888033859176606, "grad_norm": 0.18525244405965424, "learning_rate": 6.4153529498286014e-06, "loss": 0.7882, "step": 4620 }, { "epoch": 0.8889957676029242, "grad_norm": 0.19292661761475716, "learning_rate": 6.393413118476821e-06, "loss": 0.7944, "step": 4621 }, { "epoch": 0.8891881492881878, "grad_norm": 0.18750844482390328, "learning_rate": 6.3715096284250205e-06, "loss": 0.7707, "step": 4622 }, { "epoch": 0.8893805309734514, "grad_norm": 0.19571676489711146, "learning_rate": 6.349642488176944e-06, "loss": 0.7759, "step": 4623 }, { "epoch": 0.8895729126587149, "grad_norm": 0.1999374493424381, "learning_rate": 6.327811706222098e-06, "loss": 0.773, "step": 4624 }, { "epoch": 0.8897652943439784, "grad_norm": 0.18994904555485823, "learning_rate": 6.3060172910359815e-06, "loss": 0.7758, "step": 4625 }, { "epoch": 0.889957676029242, "grad_norm": 0.20194084379865584, "learning_rate": 6.284259251079938e-06, "loss": 0.7296, "step": 4626 }, { "epoch": 0.8901500577145056, "grad_norm": 0.20003183431006835, "learning_rate": 6.262537594801177e-06, "loss": 0.7613, "step": 4627 }, { "epoch": 0.8903424393997691, "grad_norm": 0.2002616895227134, "learning_rate": 6.240852330632796e-06, "loss": 0.7706, "step": 4628 }, { "epoch": 0.8905348210850327, "grad_norm": 0.18494771783457797, "learning_rate": 6.219203466993761e-06, "loss": 0.7788, "step": 4629 }, { "epoch": 0.8907272027702963, "grad_norm": 0.186778675848423, "learning_rate": 6.197591012288917e-06, "loss": 0.7564, "step": 4630 }, { "epoch": 0.8909195844555599, "grad_norm": 0.20151651508987037, "learning_rate": 6.17601497490895e-06, "loss": 0.767, "step": 4631 }, { "epoch": 0.8911119661408234, "grad_norm": 0.19045957931853313, "learning_rate": 6.1544753632304165e-06, "loss": 0.8084, "step": 4632 }, { "epoch": 0.8913043478260869, "grad_norm": 0.22473783238469086, "learning_rate": 6.132972185615749e-06, "loss": 0.7911, "step": 4633 }, { "epoch": 0.8914967295113505, "grad_norm": 0.18194978836961323, "learning_rate": 6.111505450413202e-06, "loss": 0.8128, "step": 4634 }, { "epoch": 0.8916891111966141, "grad_norm": 0.19665062036198275, "learning_rate": 6.090075165956943e-06, "loss": 0.7692, "step": 4635 }, { "epoch": 0.8918814928818777, "grad_norm": 0.1922861356216876, "learning_rate": 6.068681340566895e-06, "loss": 0.7297, "step": 4636 }, { "epoch": 0.8920738745671412, "grad_norm": 0.1875684489904389, "learning_rate": 6.0473239825489244e-06, "loss": 0.7703, "step": 4637 }, { "epoch": 0.8922662562524047, "grad_norm": 0.19197376913574055, "learning_rate": 6.026003100194632e-06, "loss": 0.8087, "step": 4638 }, { "epoch": 0.8924586379376683, "grad_norm": 0.19442852972011476, "learning_rate": 6.004718701781575e-06, "loss": 0.7657, "step": 4639 }, { "epoch": 0.8926510196229319, "grad_norm": 0.17975702964796988, "learning_rate": 5.983470795573088e-06, "loss": 0.7735, "step": 4640 }, { "epoch": 0.8928434013081955, "grad_norm": 0.20243694955561695, "learning_rate": 5.962259389818292e-06, "loss": 0.7719, "step": 4641 }, { "epoch": 0.893035782993459, "grad_norm": 0.1171192061653547, "learning_rate": 5.941084492752235e-06, "loss": 0.6383, "step": 4642 }, { "epoch": 0.8932281646787226, "grad_norm": 0.19174299174884218, "learning_rate": 5.91994611259572e-06, "loss": 0.7618, "step": 4643 }, { "epoch": 0.8934205463639862, "grad_norm": 0.1930218466621714, "learning_rate": 5.898844257555391e-06, "loss": 0.8061, "step": 4644 }, { "epoch": 0.8936129280492497, "grad_norm": 0.21770368699605006, "learning_rate": 5.877778935823719e-06, "loss": 0.7447, "step": 4645 }, { "epoch": 0.8938053097345132, "grad_norm": 0.1859471301284907, "learning_rate": 5.856750155578983e-06, "loss": 0.7751, "step": 4646 }, { "epoch": 0.8939976914197768, "grad_norm": 0.1966233249243151, "learning_rate": 5.835757924985286e-06, "loss": 0.7943, "step": 4647 }, { "epoch": 0.8941900731050404, "grad_norm": 0.1996643764881056, "learning_rate": 5.8148022521925195e-06, "loss": 0.7612, "step": 4648 }, { "epoch": 0.894382454790304, "grad_norm": 0.1907325560723981, "learning_rate": 5.793883145336443e-06, "loss": 0.7681, "step": 4649 }, { "epoch": 0.8945748364755676, "grad_norm": 0.19059796367068904, "learning_rate": 5.773000612538504e-06, "loss": 0.814, "step": 4650 }, { "epoch": 0.894767218160831, "grad_norm": 0.2003502065142198, "learning_rate": 5.752154661906084e-06, "loss": 0.7852, "step": 4651 }, { "epoch": 0.8949595998460946, "grad_norm": 0.2046664456661381, "learning_rate": 5.731345301532265e-06, "loss": 0.7886, "step": 4652 }, { "epoch": 0.8951519815313582, "grad_norm": 0.1886721005559603, "learning_rate": 5.710572539495962e-06, "loss": 0.7397, "step": 4653 }, { "epoch": 0.8953443632166218, "grad_norm": 0.20252396077618606, "learning_rate": 5.689836383861879e-06, "loss": 0.8196, "step": 4654 }, { "epoch": 0.8955367449018854, "grad_norm": 0.18661847731271602, "learning_rate": 5.6691368426805114e-06, "loss": 0.7891, "step": 4655 }, { "epoch": 0.8957291265871489, "grad_norm": 0.20515000217356816, "learning_rate": 5.648473923988129e-06, "loss": 0.7397, "step": 4656 }, { "epoch": 0.8959215082724125, "grad_norm": 0.18089800030007422, "learning_rate": 5.627847635806771e-06, "loss": 0.7142, "step": 4657 }, { "epoch": 0.896113889957676, "grad_norm": 0.20177325931086587, "learning_rate": 5.607257986144321e-06, "loss": 0.738, "step": 4658 }, { "epoch": 0.8963062716429396, "grad_norm": 0.17728154702739815, "learning_rate": 5.5867049829943395e-06, "loss": 0.786, "step": 4659 }, { "epoch": 0.8964986533282031, "grad_norm": 0.204262865935118, "learning_rate": 5.566188634336211e-06, "loss": 0.7688, "step": 4660 }, { "epoch": 0.8966910350134667, "grad_norm": 0.2221169117570336, "learning_rate": 5.545708948135142e-06, "loss": 0.7666, "step": 4661 }, { "epoch": 0.8968834166987303, "grad_norm": 0.20289591880549468, "learning_rate": 5.525265932341983e-06, "loss": 0.8059, "step": 4662 }, { "epoch": 0.8970757983839939, "grad_norm": 0.18419990660447502, "learning_rate": 5.504859594893474e-06, "loss": 0.803, "step": 4663 }, { "epoch": 0.8972681800692575, "grad_norm": 0.22517659715307373, "learning_rate": 5.484489943712012e-06, "loss": 0.7276, "step": 4664 }, { "epoch": 0.8974605617545209, "grad_norm": 0.18592153418865254, "learning_rate": 5.464156986705826e-06, "loss": 0.7922, "step": 4665 }, { "epoch": 0.8976529434397845, "grad_norm": 0.18657499499818073, "learning_rate": 5.443860731768868e-06, "loss": 0.8129, "step": 4666 }, { "epoch": 0.8978453251250481, "grad_norm": 0.182011297182935, "learning_rate": 5.4236011867808355e-06, "loss": 0.7851, "step": 4667 }, { "epoch": 0.8980377068103117, "grad_norm": 0.18314779269950743, "learning_rate": 5.403378359607181e-06, "loss": 0.7718, "step": 4668 }, { "epoch": 0.8982300884955752, "grad_norm": 0.20181334321452982, "learning_rate": 5.383192258099112e-06, "loss": 0.7821, "step": 4669 }, { "epoch": 0.8984224701808388, "grad_norm": 0.2065427270914259, "learning_rate": 5.36304289009355e-06, "loss": 0.7519, "step": 4670 }, { "epoch": 0.8986148518661023, "grad_norm": 0.18840619972827718, "learning_rate": 5.342930263413193e-06, "loss": 0.7225, "step": 4671 }, { "epoch": 0.8988072335513659, "grad_norm": 0.18464493835645981, "learning_rate": 5.322854385866438e-06, "loss": 0.7231, "step": 4672 }, { "epoch": 0.8989996152366295, "grad_norm": 0.20246192820838602, "learning_rate": 5.3028152652474514e-06, "loss": 0.817, "step": 4673 }, { "epoch": 0.899191996921893, "grad_norm": 0.21171783357267218, "learning_rate": 5.282812909336077e-06, "loss": 0.7243, "step": 4674 }, { "epoch": 0.8993843786071566, "grad_norm": 0.20013801848978793, "learning_rate": 5.262847325897968e-06, "loss": 0.7637, "step": 4675 }, { "epoch": 0.8995767602924202, "grad_norm": 0.19338503754017008, "learning_rate": 5.242918522684393e-06, "loss": 0.7967, "step": 4676 }, { "epoch": 0.8997691419776838, "grad_norm": 0.1837536360208968, "learning_rate": 5.223026507432449e-06, "loss": 0.7957, "step": 4677 }, { "epoch": 0.8999615236629472, "grad_norm": 0.1897811195177453, "learning_rate": 5.203171287864872e-06, "loss": 0.759, "step": 4678 }, { "epoch": 0.9001539053482108, "grad_norm": 0.11826548530772632, "learning_rate": 5.183352871690162e-06, "loss": 0.7177, "step": 4679 }, { "epoch": 0.9003462870334744, "grad_norm": 0.19834946817447044, "learning_rate": 5.163571266602485e-06, "loss": 0.7675, "step": 4680 }, { "epoch": 0.900538668718738, "grad_norm": 0.18726840118872926, "learning_rate": 5.143826480281777e-06, "loss": 0.7465, "step": 4681 }, { "epoch": 0.9007310504040016, "grad_norm": 0.22021925227898292, "learning_rate": 5.124118520393606e-06, "loss": 0.8108, "step": 4682 }, { "epoch": 0.9009234320892651, "grad_norm": 0.20451459471621225, "learning_rate": 5.104447394589296e-06, "loss": 0.7488, "step": 4683 }, { "epoch": 0.9011158137745287, "grad_norm": 0.19106161469644667, "learning_rate": 5.0848131105058704e-06, "loss": 0.7944, "step": 4684 }, { "epoch": 0.9013081954597922, "grad_norm": 0.19822988970898517, "learning_rate": 5.0652156757660215e-06, "loss": 0.7839, "step": 4685 }, { "epoch": 0.9015005771450558, "grad_norm": 0.1840096682911428, "learning_rate": 5.045655097978131e-06, "loss": 0.7535, "step": 4686 }, { "epoch": 0.9016929588303193, "grad_norm": 0.20169868780094893, "learning_rate": 5.026131384736321e-06, "loss": 0.7732, "step": 4687 }, { "epoch": 0.9018853405155829, "grad_norm": 0.1783708740679341, "learning_rate": 5.006644543620342e-06, "loss": 0.8209, "step": 4688 }, { "epoch": 0.9020777222008465, "grad_norm": 0.1978800681457316, "learning_rate": 4.987194582195687e-06, "loss": 0.7537, "step": 4689 }, { "epoch": 0.9022701038861101, "grad_norm": 0.17556400543441414, "learning_rate": 4.967781508013458e-06, "loss": 0.777, "step": 4690 }, { "epoch": 0.9024624855713737, "grad_norm": 0.18950609990492764, "learning_rate": 4.9484053286105055e-06, "loss": 0.7979, "step": 4691 }, { "epoch": 0.9026548672566371, "grad_norm": 0.19317179226019368, "learning_rate": 4.929066051509346e-06, "loss": 0.8211, "step": 4692 }, { "epoch": 0.9028472489419007, "grad_norm": 0.20167094819533588, "learning_rate": 4.909763684218116e-06, "loss": 0.7168, "step": 4693 }, { "epoch": 0.9030396306271643, "grad_norm": 0.19205294461458144, "learning_rate": 4.89049823423069e-06, "loss": 0.783, "step": 4694 }, { "epoch": 0.9032320123124279, "grad_norm": 0.18991780890729051, "learning_rate": 4.87126970902656e-06, "loss": 0.7857, "step": 4695 }, { "epoch": 0.9034243939976914, "grad_norm": 0.2153359317621122, "learning_rate": 4.852078116070902e-06, "loss": 0.7384, "step": 4696 }, { "epoch": 0.903616775682955, "grad_norm": 0.1884554067519739, "learning_rate": 4.832923462814565e-06, "loss": 0.7749, "step": 4697 }, { "epoch": 0.9038091573682185, "grad_norm": 0.1868331279595106, "learning_rate": 4.813805756694034e-06, "loss": 0.7575, "step": 4698 }, { "epoch": 0.9040015390534821, "grad_norm": 0.2150226005137855, "learning_rate": 4.794725005131462e-06, "loss": 0.7789, "step": 4699 }, { "epoch": 0.9041939207387457, "grad_norm": 0.20689426281194861, "learning_rate": 4.775681215534656e-06, "loss": 0.7528, "step": 4700 }, { "epoch": 0.9043863024240092, "grad_norm": 0.21884301807668008, "learning_rate": 4.756674395297089e-06, "loss": 0.7537, "step": 4701 }, { "epoch": 0.9045786841092728, "grad_norm": 0.10768389034557221, "learning_rate": 4.737704551797817e-06, "loss": 0.5839, "step": 4702 }, { "epoch": 0.9047710657945364, "grad_norm": 0.19606951518741877, "learning_rate": 4.718771692401636e-06, "loss": 0.7972, "step": 4703 }, { "epoch": 0.9049634474798, "grad_norm": 0.1940866242551663, "learning_rate": 4.699875824458899e-06, "loss": 0.7405, "step": 4704 }, { "epoch": 0.9051558291650635, "grad_norm": 0.20746921642527533, "learning_rate": 4.6810169553056615e-06, "loss": 0.7123, "step": 4705 }, { "epoch": 0.905348210850327, "grad_norm": 0.19465119441159293, "learning_rate": 4.662195092263566e-06, "loss": 0.7518, "step": 4706 }, { "epoch": 0.9055405925355906, "grad_norm": 0.18823664032949286, "learning_rate": 4.643410242639911e-06, "loss": 0.7657, "step": 4707 }, { "epoch": 0.9057329742208542, "grad_norm": 0.17374999916272726, "learning_rate": 4.624662413727621e-06, "loss": 0.8142, "step": 4708 }, { "epoch": 0.9059253559061178, "grad_norm": 0.19716023382998762, "learning_rate": 4.605951612805237e-06, "loss": 0.779, "step": 4709 }, { "epoch": 0.9061177375913813, "grad_norm": 0.21326397089146268, "learning_rate": 4.587277847136984e-06, "loss": 0.7477, "step": 4710 }, { "epoch": 0.9063101192766448, "grad_norm": 0.20257097027509155, "learning_rate": 4.5686411239726055e-06, "loss": 0.7752, "step": 4711 }, { "epoch": 0.9065025009619084, "grad_norm": 0.1918958417644558, "learning_rate": 4.550041450547548e-06, "loss": 0.7444, "step": 4712 }, { "epoch": 0.906694882647172, "grad_norm": 0.23099503228140347, "learning_rate": 4.531478834082836e-06, "loss": 0.7675, "step": 4713 }, { "epoch": 0.9068872643324356, "grad_norm": 0.19171279679140293, "learning_rate": 4.512953281785104e-06, "loss": 0.7965, "step": 4714 }, { "epoch": 0.9070796460176991, "grad_norm": 0.19224777395657328, "learning_rate": 4.494464800846654e-06, "loss": 0.7673, "step": 4715 }, { "epoch": 0.9072720277029627, "grad_norm": 0.19121749811550987, "learning_rate": 4.476013398445289e-06, "loss": 0.7106, "step": 4716 }, { "epoch": 0.9074644093882263, "grad_norm": 0.1845094650731676, "learning_rate": 4.457599081744523e-06, "loss": 0.8391, "step": 4717 }, { "epoch": 0.9076567910734898, "grad_norm": 0.20200066743720244, "learning_rate": 4.439221857893416e-06, "loss": 0.7873, "step": 4718 }, { "epoch": 0.9078491727587533, "grad_norm": 0.18086756091085124, "learning_rate": 4.420881734026639e-06, "loss": 0.7835, "step": 4719 }, { "epoch": 0.9080415544440169, "grad_norm": 0.18446014333745825, "learning_rate": 4.402578717264449e-06, "loss": 0.8157, "step": 4720 }, { "epoch": 0.9082339361292805, "grad_norm": 0.18742670354582078, "learning_rate": 4.384312814712721e-06, "loss": 0.7388, "step": 4721 }, { "epoch": 0.9084263178145441, "grad_norm": 0.20600587540628487, "learning_rate": 4.366084033462914e-06, "loss": 0.802, "step": 4722 }, { "epoch": 0.9086186994998077, "grad_norm": 0.1959142462664383, "learning_rate": 4.347892380592034e-06, "loss": 0.7641, "step": 4723 }, { "epoch": 0.9088110811850711, "grad_norm": 0.18474859806412222, "learning_rate": 4.329737863162752e-06, "loss": 0.7164, "step": 4724 }, { "epoch": 0.9090034628703347, "grad_norm": 0.1848348067009686, "learning_rate": 4.311620488223256e-06, "loss": 0.7319, "step": 4725 }, { "epoch": 0.9091958445555983, "grad_norm": 0.18985307500522294, "learning_rate": 4.293540262807317e-06, "loss": 0.7986, "step": 4726 }, { "epoch": 0.9093882262408619, "grad_norm": 0.19862933125317733, "learning_rate": 4.27549719393433e-06, "loss": 0.7824, "step": 4727 }, { "epoch": 0.9095806079261254, "grad_norm": 0.19925077440904937, "learning_rate": 4.257491288609217e-06, "loss": 0.7684, "step": 4728 }, { "epoch": 0.909772989611389, "grad_norm": 0.1915042774682735, "learning_rate": 4.239522553822495e-06, "loss": 0.7989, "step": 4729 }, { "epoch": 0.9099653712966526, "grad_norm": 0.19096513596462666, "learning_rate": 4.221590996550251e-06, "loss": 0.7484, "step": 4730 }, { "epoch": 0.9101577529819161, "grad_norm": 0.20799090210473403, "learning_rate": 4.203696623754138e-06, "loss": 0.7632, "step": 4731 }, { "epoch": 0.9103501346671797, "grad_norm": 0.20167350850744314, "learning_rate": 4.185839442381357e-06, "loss": 0.7804, "step": 4732 }, { "epoch": 0.9105425163524432, "grad_norm": 0.21028743903737962, "learning_rate": 4.168019459364669e-06, "loss": 0.7488, "step": 4733 }, { "epoch": 0.9107348980377068, "grad_norm": 0.18675213978527164, "learning_rate": 4.150236681622433e-06, "loss": 0.7996, "step": 4734 }, { "epoch": 0.9109272797229704, "grad_norm": 0.19391096764743346, "learning_rate": 4.132491116058501e-06, "loss": 0.8067, "step": 4735 }, { "epoch": 0.911119661408234, "grad_norm": 0.19702943335462267, "learning_rate": 4.114782769562364e-06, "loss": 0.7922, "step": 4736 }, { "epoch": 0.9113120430934974, "grad_norm": 0.19714578375166253, "learning_rate": 4.097111649008967e-06, "loss": 0.7945, "step": 4737 }, { "epoch": 0.911504424778761, "grad_norm": 0.1900819244215331, "learning_rate": 4.079477761258854e-06, "loss": 0.7769, "step": 4738 }, { "epoch": 0.9116968064640246, "grad_norm": 0.1861620424960659, "learning_rate": 4.061881113158117e-06, "loss": 0.7789, "step": 4739 }, { "epoch": 0.9118891881492882, "grad_norm": 0.1963653472156607, "learning_rate": 4.044321711538368e-06, "loss": 0.7832, "step": 4740 }, { "epoch": 0.9120815698345518, "grad_norm": 0.1956840918361176, "learning_rate": 4.0267995632168094e-06, "loss": 0.7586, "step": 4741 }, { "epoch": 0.9122739515198153, "grad_norm": 0.20031443900888837, "learning_rate": 4.00931467499609e-06, "loss": 0.8147, "step": 4742 }, { "epoch": 0.9124663332050789, "grad_norm": 0.21148685772533377, "learning_rate": 3.9918670536644775e-06, "loss": 0.8296, "step": 4743 }, { "epoch": 0.9126587148903424, "grad_norm": 0.19424622466545238, "learning_rate": 3.974456705995733e-06, "loss": 0.7752, "step": 4744 }, { "epoch": 0.912851096575606, "grad_norm": 0.20879675561266056, "learning_rate": 3.957083638749148e-06, "loss": 0.7546, "step": 4745 }, { "epoch": 0.9130434782608695, "grad_norm": 0.19969115128574083, "learning_rate": 3.939747858669551e-06, "loss": 0.7827, "step": 4746 }, { "epoch": 0.9132358599461331, "grad_norm": 0.1858608686732823, "learning_rate": 3.922449372487291e-06, "loss": 0.796, "step": 4747 }, { "epoch": 0.9134282416313967, "grad_norm": 0.20458982655880292, "learning_rate": 3.905188186918229e-06, "loss": 0.7656, "step": 4748 }, { "epoch": 0.9136206233166603, "grad_norm": 0.20778705158060243, "learning_rate": 3.887964308663739e-06, "loss": 0.7925, "step": 4749 }, { "epoch": 0.9138130050019239, "grad_norm": 0.1967838404696935, "learning_rate": 3.8707777444107695e-06, "loss": 0.7941, "step": 4750 }, { "epoch": 0.9140053866871873, "grad_norm": 0.18842483703543916, "learning_rate": 3.853628500831685e-06, "loss": 0.8633, "step": 4751 }, { "epoch": 0.9141977683724509, "grad_norm": 0.1920060515735726, "learning_rate": 3.836516584584426e-06, "loss": 0.7977, "step": 4752 }, { "epoch": 0.9143901500577145, "grad_norm": 0.19282411935003008, "learning_rate": 3.819442002312457e-06, "loss": 0.7184, "step": 4753 }, { "epoch": 0.9145825317429781, "grad_norm": 0.19724663277046794, "learning_rate": 3.8024047606446735e-06, "loss": 0.7642, "step": 4754 }, { "epoch": 0.9147749134282416, "grad_norm": 0.19039887890627913, "learning_rate": 3.7854048661955522e-06, "loss": 0.7497, "step": 4755 }, { "epoch": 0.9149672951135052, "grad_norm": 0.18862191888675656, "learning_rate": 3.7684423255650357e-06, "loss": 0.8167, "step": 4756 }, { "epoch": 0.9151596767987688, "grad_norm": 0.19258251688782502, "learning_rate": 3.7515171453385458e-06, "loss": 0.772, "step": 4757 }, { "epoch": 0.9153520584840323, "grad_norm": 0.18294411545870143, "learning_rate": 3.7346293320870363e-06, "loss": 0.781, "step": 4758 }, { "epoch": 0.9155444401692959, "grad_norm": 0.19340722189822188, "learning_rate": 3.717778892366941e-06, "loss": 0.7833, "step": 4759 }, { "epoch": 0.9157368218545594, "grad_norm": 0.11462931390627669, "learning_rate": 3.700965832720171e-06, "loss": 0.604, "step": 4760 }, { "epoch": 0.915929203539823, "grad_norm": 0.11994223054121753, "learning_rate": 3.684190159674117e-06, "loss": 0.6378, "step": 4761 }, { "epoch": 0.9161215852250866, "grad_norm": 0.19225796859259753, "learning_rate": 3.6674518797417235e-06, "loss": 0.7609, "step": 4762 }, { "epoch": 0.9163139669103502, "grad_norm": 0.20479998859868798, "learning_rate": 3.650750999421315e-06, "loss": 0.7664, "step": 4763 }, { "epoch": 0.9165063485956138, "grad_norm": 0.187052209015783, "learning_rate": 3.634087525196794e-06, "loss": 0.8016, "step": 4764 }, { "epoch": 0.9166987302808772, "grad_norm": 0.20184558174096046, "learning_rate": 3.6174614635374637e-06, "loss": 0.7835, "step": 4765 }, { "epoch": 0.9168911119661408, "grad_norm": 0.17420612966903068, "learning_rate": 3.6008728208981156e-06, "loss": 0.8, "step": 4766 }, { "epoch": 0.9170834936514044, "grad_norm": 0.1918929320945558, "learning_rate": 3.5843216037190876e-06, "loss": 0.8056, "step": 4767 }, { "epoch": 0.917275875336668, "grad_norm": 0.19762000028718582, "learning_rate": 3.567807818426083e-06, "loss": 0.769, "step": 4768 }, { "epoch": 0.9174682570219315, "grad_norm": 0.10465281084943366, "learning_rate": 3.5513314714303526e-06, "loss": 0.6386, "step": 4769 }, { "epoch": 0.917660638707195, "grad_norm": 0.18680539488062053, "learning_rate": 3.534892569128567e-06, "loss": 0.7907, "step": 4770 }, { "epoch": 0.9178530203924586, "grad_norm": 0.19248874324415916, "learning_rate": 3.518491117902878e-06, "loss": 0.7977, "step": 4771 }, { "epoch": 0.9180454020777222, "grad_norm": 0.19856124006749082, "learning_rate": 3.5021271241208907e-06, "loss": 0.7669, "step": 4772 }, { "epoch": 0.9182377837629858, "grad_norm": 0.1861552498177926, "learning_rate": 3.48580059413568e-06, "loss": 0.8356, "step": 4773 }, { "epoch": 0.9184301654482493, "grad_norm": 0.17682149861425725, "learning_rate": 3.469511534285752e-06, "loss": 0.7837, "step": 4774 }, { "epoch": 0.9186225471335129, "grad_norm": 0.17889635027348508, "learning_rate": 3.4532599508950824e-06, "loss": 0.7981, "step": 4775 }, { "epoch": 0.9188149288187765, "grad_norm": 0.1926570972104963, "learning_rate": 3.437045850273113e-06, "loss": 0.7508, "step": 4776 }, { "epoch": 0.91900731050404, "grad_norm": 0.18934660103126097, "learning_rate": 3.420869238714708e-06, "loss": 0.8064, "step": 4777 }, { "epoch": 0.9191996921893035, "grad_norm": 0.2150319174289337, "learning_rate": 3.4047301225001547e-06, "loss": 0.7497, "step": 4778 }, { "epoch": 0.9193920738745671, "grad_norm": 0.1881524948877378, "learning_rate": 3.388628507895275e-06, "loss": 0.7582, "step": 4779 }, { "epoch": 0.9195844555598307, "grad_norm": 0.11519188186038672, "learning_rate": 3.3725644011512124e-06, "loss": 0.6262, "step": 4780 }, { "epoch": 0.9197768372450943, "grad_norm": 0.20948151506581958, "learning_rate": 3.3565378085046337e-06, "loss": 0.7552, "step": 4781 }, { "epoch": 0.9199692189303579, "grad_norm": 0.19718659542603822, "learning_rate": 3.340548736177618e-06, "loss": 0.7995, "step": 4782 }, { "epoch": 0.9201616006156214, "grad_norm": 0.19875541758925855, "learning_rate": 3.324597190377665e-06, "loss": 0.7437, "step": 4783 }, { "epoch": 0.9203539823008849, "grad_norm": 0.19728971849814478, "learning_rate": 3.308683177297711e-06, "loss": 0.8042, "step": 4784 }, { "epoch": 0.9205463639861485, "grad_norm": 0.18881170363613492, "learning_rate": 3.2928067031161247e-06, "loss": 0.7826, "step": 4785 }, { "epoch": 0.9207387456714121, "grad_norm": 0.1887703087019458, "learning_rate": 3.2769677739966973e-06, "loss": 0.7448, "step": 4786 }, { "epoch": 0.9209311273566756, "grad_norm": 0.197975162882154, "learning_rate": 3.2611663960886664e-06, "loss": 0.7557, "step": 4787 }, { "epoch": 0.9211235090419392, "grad_norm": 0.1955893665131825, "learning_rate": 3.245402575526646e-06, "loss": 0.782, "step": 4788 }, { "epoch": 0.9213158907272028, "grad_norm": 0.1959253487569719, "learning_rate": 3.229676318430697e-06, "loss": 0.7952, "step": 4789 }, { "epoch": 0.9215082724124664, "grad_norm": 0.20014482527440003, "learning_rate": 3.2139876309063234e-06, "loss": 0.7218, "step": 4790 }, { "epoch": 0.9217006540977299, "grad_norm": 0.10568015525153242, "learning_rate": 3.198336519044376e-06, "loss": 0.614, "step": 4791 }, { "epoch": 0.9218930357829934, "grad_norm": 0.18355528715364375, "learning_rate": 3.1827229889211606e-06, "loss": 0.7326, "step": 4792 }, { "epoch": 0.922085417468257, "grad_norm": 0.19956508801972092, "learning_rate": 3.1671470465984177e-06, "loss": 0.7667, "step": 4793 }, { "epoch": 0.9222777991535206, "grad_norm": 0.20940458889843155, "learning_rate": 3.151608698123232e-06, "loss": 0.8077, "step": 4794 }, { "epoch": 0.9224701808387842, "grad_norm": 0.1909081675988135, "learning_rate": 3.1361079495281444e-06, "loss": 0.7532, "step": 4795 }, { "epoch": 0.9226625625240477, "grad_norm": 0.1859016267679606, "learning_rate": 3.1206448068310633e-06, "loss": 0.7194, "step": 4796 }, { "epoch": 0.9228549442093112, "grad_norm": 0.19752295222047572, "learning_rate": 3.1052192760353314e-06, "loss": 0.8253, "step": 4797 }, { "epoch": 0.9230473258945748, "grad_norm": 0.19064250990109716, "learning_rate": 3.0898313631296583e-06, "loss": 0.8051, "step": 4798 }, { "epoch": 0.9232397075798384, "grad_norm": 0.19237074917569408, "learning_rate": 3.074481074088165e-06, "loss": 0.7505, "step": 4799 }, { "epoch": 0.923432089265102, "grad_norm": 0.21145289552740695, "learning_rate": 3.0591684148703614e-06, "loss": 0.8034, "step": 4800 }, { "epoch": 0.9236244709503655, "grad_norm": 0.19529796671342445, "learning_rate": 3.0438933914211486e-06, "loss": 0.7317, "step": 4801 }, { "epoch": 0.9238168526356291, "grad_norm": 0.19319345668735713, "learning_rate": 3.028656009670827e-06, "loss": 0.7943, "step": 4802 }, { "epoch": 0.9240092343208927, "grad_norm": 0.19926248988552905, "learning_rate": 3.0134562755350537e-06, "loss": 0.7747, "step": 4803 }, { "epoch": 0.9242016160061562, "grad_norm": 0.19751753368308136, "learning_rate": 2.998294194914897e-06, "loss": 0.7098, "step": 4804 }, { "epoch": 0.9243939976914197, "grad_norm": 0.19319192402840685, "learning_rate": 2.9831697736968146e-06, "loss": 0.7621, "step": 4805 }, { "epoch": 0.9245863793766833, "grad_norm": 0.18672181638552762, "learning_rate": 2.9680830177525986e-06, "loss": 0.8621, "step": 4806 }, { "epoch": 0.9247787610619469, "grad_norm": 0.18227041105346412, "learning_rate": 2.953033932939464e-06, "loss": 0.7239, "step": 4807 }, { "epoch": 0.9249711427472105, "grad_norm": 0.17748903740080804, "learning_rate": 2.9380225250999816e-06, "loss": 0.8031, "step": 4808 }, { "epoch": 0.9251635244324741, "grad_norm": 0.2032666870259744, "learning_rate": 2.9230488000621004e-06, "loss": 0.8091, "step": 4809 }, { "epoch": 0.9253559061177375, "grad_norm": 0.20598368195654762, "learning_rate": 2.908112763639137e-06, "loss": 0.766, "step": 4810 }, { "epoch": 0.9255482878030011, "grad_norm": 0.11262419680713065, "learning_rate": 2.893214421629764e-06, "loss": 0.6096, "step": 4811 }, { "epoch": 0.9257406694882647, "grad_norm": 0.20791381519471333, "learning_rate": 2.8783537798180437e-06, "loss": 0.7741, "step": 4812 }, { "epoch": 0.9259330511735283, "grad_norm": 0.2059390205025462, "learning_rate": 2.863530843973372e-06, "loss": 0.7962, "step": 4813 }, { "epoch": 0.9261254328587919, "grad_norm": 0.1996034762302146, "learning_rate": 2.848745619850546e-06, "loss": 0.7523, "step": 4814 }, { "epoch": 0.9263178145440554, "grad_norm": 0.20225180450512809, "learning_rate": 2.833998113189662e-06, "loss": 0.8122, "step": 4815 }, { "epoch": 0.926510196229319, "grad_norm": 0.19397093813913494, "learning_rate": 2.819288329716263e-06, "loss": 0.7961, "step": 4816 }, { "epoch": 0.9267025779145825, "grad_norm": 0.20771910093003587, "learning_rate": 2.8046162751411475e-06, "loss": 0.7934, "step": 4817 }, { "epoch": 0.9268949595998461, "grad_norm": 0.19354749927886677, "learning_rate": 2.7899819551605257e-06, "loss": 0.7919, "step": 4818 }, { "epoch": 0.9270873412851096, "grad_norm": 0.19402914091225928, "learning_rate": 2.7753853754559635e-06, "loss": 0.7865, "step": 4819 }, { "epoch": 0.9272797229703732, "grad_norm": 0.20919967757502353, "learning_rate": 2.760826541694328e-06, "loss": 0.8222, "step": 4820 }, { "epoch": 0.9274721046556368, "grad_norm": 0.19997916970845292, "learning_rate": 2.746305459527876e-06, "loss": 0.7389, "step": 4821 }, { "epoch": 0.9276644863409004, "grad_norm": 0.10901687183001066, "learning_rate": 2.7318221345941863e-06, "loss": 0.6619, "step": 4822 }, { "epoch": 0.927856868026164, "grad_norm": 0.18617928141341913, "learning_rate": 2.7173765725161837e-06, "loss": 0.7646, "step": 4823 }, { "epoch": 0.9280492497114274, "grad_norm": 0.20266558594369888, "learning_rate": 2.7029687789021373e-06, "loss": 0.7001, "step": 4824 }, { "epoch": 0.928241631396691, "grad_norm": 0.19912700779581868, "learning_rate": 2.6885987593456507e-06, "loss": 0.7958, "step": 4825 }, { "epoch": 0.9284340130819546, "grad_norm": 0.17867112824358575, "learning_rate": 2.67426651942565e-06, "loss": 0.7622, "step": 4826 }, { "epoch": 0.9286263947672182, "grad_norm": 0.18443818946689194, "learning_rate": 2.6599720647064062e-06, "loss": 0.7994, "step": 4827 }, { "epoch": 0.9288187764524817, "grad_norm": 0.18431713207197442, "learning_rate": 2.645715400737536e-06, "loss": 0.7596, "step": 4828 }, { "epoch": 0.9290111581377453, "grad_norm": 0.1882230543760777, "learning_rate": 2.631496533053934e-06, "loss": 0.7973, "step": 4829 }, { "epoch": 0.9292035398230089, "grad_norm": 0.10341332321275434, "learning_rate": 2.617315467175885e-06, "loss": 0.5707, "step": 4830 }, { "epoch": 0.9293959215082724, "grad_norm": 0.10994443437820528, "learning_rate": 2.6031722086089615e-06, "loss": 0.6139, "step": 4831 }, { "epoch": 0.929588303193536, "grad_norm": 0.22303109147280273, "learning_rate": 2.589066762844039e-06, "loss": 0.81, "step": 4832 }, { "epoch": 0.9297806848787995, "grad_norm": 0.21613237101601113, "learning_rate": 2.57499913535737e-06, "loss": 0.7698, "step": 4833 }, { "epoch": 0.9299730665640631, "grad_norm": 0.18234502707450453, "learning_rate": 2.5609693316104744e-06, "loss": 0.7618, "step": 4834 }, { "epoch": 0.9301654482493267, "grad_norm": 0.18585451383001447, "learning_rate": 2.5469773570502065e-06, "loss": 0.7839, "step": 4835 }, { "epoch": 0.9303578299345903, "grad_norm": 0.10589831042333893, "learning_rate": 2.5330232171087433e-06, "loss": 0.6283, "step": 4836 }, { "epoch": 0.9305502116198537, "grad_norm": 0.19847893262471164, "learning_rate": 2.5191069172035396e-06, "loss": 0.7798, "step": 4837 }, { "epoch": 0.9307425933051173, "grad_norm": 0.21800122123778456, "learning_rate": 2.5052284627374077e-06, "loss": 0.7816, "step": 4838 }, { "epoch": 0.9309349749903809, "grad_norm": 0.19938886270551584, "learning_rate": 2.491387859098426e-06, "loss": 0.8146, "step": 4839 }, { "epoch": 0.9311273566756445, "grad_norm": 0.18278897489950557, "learning_rate": 2.4775851116599967e-06, "loss": 0.7348, "step": 4840 }, { "epoch": 0.9313197383609081, "grad_norm": 0.19855034471302424, "learning_rate": 2.4638202257808107e-06, "loss": 0.7875, "step": 4841 }, { "epoch": 0.9315121200461716, "grad_norm": 0.21920368909330573, "learning_rate": 2.4500932068049043e-06, "loss": 0.7691, "step": 4842 }, { "epoch": 0.9317045017314352, "grad_norm": 0.2142706253472343, "learning_rate": 2.436404060061548e-06, "loss": 0.7443, "step": 4843 }, { "epoch": 0.9318968834166987, "grad_norm": 0.1976902686035109, "learning_rate": 2.4227527908653458e-06, "loss": 0.7335, "step": 4844 }, { "epoch": 0.9320892651019623, "grad_norm": 0.18980604184089014, "learning_rate": 2.409139404516203e-06, "loss": 0.7727, "step": 4845 }, { "epoch": 0.9322816467872258, "grad_norm": 0.17855671176805554, "learning_rate": 2.3955639062992696e-06, "loss": 0.7539, "step": 4846 }, { "epoch": 0.9324740284724894, "grad_norm": 0.20517518420965083, "learning_rate": 2.382026301485074e-06, "loss": 0.7818, "step": 4847 }, { "epoch": 0.932666410157753, "grad_norm": 0.19732107769211174, "learning_rate": 2.3685265953293343e-06, "loss": 0.787, "step": 4848 }, { "epoch": 0.9328587918430166, "grad_norm": 0.1820851953063446, "learning_rate": 2.3550647930731364e-06, "loss": 0.7366, "step": 4849 }, { "epoch": 0.9330511735282802, "grad_norm": 0.2027528686254674, "learning_rate": 2.3416408999427875e-06, "loss": 0.7727, "step": 4850 }, { "epoch": 0.9332435552135436, "grad_norm": 0.18978540051986406, "learning_rate": 2.3282549211499304e-06, "loss": 0.8125, "step": 4851 }, { "epoch": 0.9334359368988072, "grad_norm": 0.19196430802514042, "learning_rate": 2.3149068618914415e-06, "loss": 0.7799, "step": 4852 }, { "epoch": 0.9336283185840708, "grad_norm": 0.18452875767656984, "learning_rate": 2.3015967273494864e-06, "loss": 0.7784, "step": 4853 }, { "epoch": 0.9338207002693344, "grad_norm": 0.12401213177927917, "learning_rate": 2.288324522691565e-06, "loss": 0.6549, "step": 4854 }, { "epoch": 0.9340130819545979, "grad_norm": 0.18110239409774787, "learning_rate": 2.2750902530703666e-06, "loss": 0.7687, "step": 4855 }, { "epoch": 0.9342054636398615, "grad_norm": 0.1937189198808449, "learning_rate": 2.2618939236238923e-06, "loss": 0.7248, "step": 4856 }, { "epoch": 0.934397845325125, "grad_norm": 0.17578260749676172, "learning_rate": 2.248735539475433e-06, "loss": 0.7947, "step": 4857 }, { "epoch": 0.9345902270103886, "grad_norm": 0.19697689682280442, "learning_rate": 2.2356151057334906e-06, "loss": 0.7646, "step": 4858 }, { "epoch": 0.9347826086956522, "grad_norm": 0.10666682050430694, "learning_rate": 2.2225326274919133e-06, "loss": 0.6104, "step": 4859 }, { "epoch": 0.9349749903809157, "grad_norm": 0.18947844884398357, "learning_rate": 2.209488109829727e-06, "loss": 0.7627, "step": 4860 }, { "epoch": 0.9351673720661793, "grad_norm": 0.2254509020765301, "learning_rate": 2.196481557811303e-06, "loss": 0.7434, "step": 4861 }, { "epoch": 0.9353597537514429, "grad_norm": 0.19411046214089667, "learning_rate": 2.1835129764861906e-06, "loss": 0.8105, "step": 4862 }, { "epoch": 0.9355521354367065, "grad_norm": 0.17615362448554903, "learning_rate": 2.1705823708892734e-06, "loss": 0.7896, "step": 4863 }, { "epoch": 0.93574451712197, "grad_norm": 0.20320615574688944, "learning_rate": 2.1576897460406474e-06, "loss": 0.7491, "step": 4864 }, { "epoch": 0.9359368988072335, "grad_norm": 0.21440202076843565, "learning_rate": 2.144835106945664e-06, "loss": 0.7361, "step": 4865 }, { "epoch": 0.9361292804924971, "grad_norm": 0.2005583070597083, "learning_rate": 2.1320184585949533e-06, "loss": 0.8093, "step": 4866 }, { "epoch": 0.9363216621777607, "grad_norm": 0.1881753601187003, "learning_rate": 2.119239805964357e-06, "loss": 0.7903, "step": 4867 }, { "epoch": 0.9365140438630243, "grad_norm": 0.12220584321953981, "learning_rate": 2.106499154015018e-06, "loss": 0.655, "step": 4868 }, { "epoch": 0.9367064255482878, "grad_norm": 0.1927760998458783, "learning_rate": 2.0937965076932574e-06, "loss": 0.7733, "step": 4869 }, { "epoch": 0.9368988072335513, "grad_norm": 0.21400972413728073, "learning_rate": 2.0811318719307193e-06, "loss": 0.783, "step": 4870 }, { "epoch": 0.9370911889188149, "grad_norm": 0.18798519069169084, "learning_rate": 2.068505251644237e-06, "loss": 0.7444, "step": 4871 }, { "epoch": 0.9372835706040785, "grad_norm": 0.1801440736666902, "learning_rate": 2.0559166517358785e-06, "loss": 0.7446, "step": 4872 }, { "epoch": 0.9374759522893421, "grad_norm": 0.18587309576974476, "learning_rate": 2.0433660770930007e-06, "loss": 0.7693, "step": 4873 }, { "epoch": 0.9376683339746056, "grad_norm": 0.19961480445612934, "learning_rate": 2.030853532588162e-06, "loss": 0.7875, "step": 4874 }, { "epoch": 0.9378607156598692, "grad_norm": 0.20576486223631482, "learning_rate": 2.0183790230791534e-06, "loss": 0.7696, "step": 4875 }, { "epoch": 0.9380530973451328, "grad_norm": 0.19850292463528263, "learning_rate": 2.0059425534090125e-06, "loss": 0.741, "step": 4876 }, { "epoch": 0.9382454790303963, "grad_norm": 0.19638993434148566, "learning_rate": 1.993544128406e-06, "loss": 0.8179, "step": 4877 }, { "epoch": 0.9384378607156598, "grad_norm": 0.20225885366319712, "learning_rate": 1.981183752883631e-06, "loss": 0.7853, "step": 4878 }, { "epoch": 0.9386302424009234, "grad_norm": 0.2193934949927163, "learning_rate": 1.9688614316406007e-06, "loss": 0.7741, "step": 4879 }, { "epoch": 0.938822624086187, "grad_norm": 0.20894235653628826, "learning_rate": 1.9565771694608936e-06, "loss": 0.7626, "step": 4880 }, { "epoch": 0.9390150057714506, "grad_norm": 0.19065674525124382, "learning_rate": 1.944330971113639e-06, "loss": 0.7695, "step": 4881 }, { "epoch": 0.9392073874567142, "grad_norm": 0.2114036024160911, "learning_rate": 1.9321228413532787e-06, "loss": 0.7322, "step": 4882 }, { "epoch": 0.9393997691419776, "grad_norm": 0.1987843233463899, "learning_rate": 1.91995278491941e-06, "loss": 0.7517, "step": 4883 }, { "epoch": 0.9395921508272412, "grad_norm": 0.11696084710148291, "learning_rate": 1.907820806536842e-06, "loss": 0.6329, "step": 4884 }, { "epoch": 0.9397845325125048, "grad_norm": 0.1903900872909262, "learning_rate": 1.8957269109156628e-06, "loss": 0.7402, "step": 4885 }, { "epoch": 0.9399769141977684, "grad_norm": 0.20027924952346546, "learning_rate": 1.8836711027511277e-06, "loss": 0.7894, "step": 4886 }, { "epoch": 0.9401692958830319, "grad_norm": 0.10715140804826326, "learning_rate": 1.8716533867237152e-06, "loss": 0.6211, "step": 4887 }, { "epoch": 0.9403616775682955, "grad_norm": 0.17205066744458097, "learning_rate": 1.859673767499115e-06, "loss": 0.8145, "step": 4888 }, { "epoch": 0.9405540592535591, "grad_norm": 0.10802170384788752, "learning_rate": 1.8477322497282178e-06, "loss": 0.5962, "step": 4889 }, { "epoch": 0.9407464409388226, "grad_norm": 0.20425828621161138, "learning_rate": 1.83582883804716e-06, "loss": 0.7966, "step": 4890 }, { "epoch": 0.9409388226240862, "grad_norm": 0.17705201809193238, "learning_rate": 1.8239635370772223e-06, "loss": 0.733, "step": 4891 }, { "epoch": 0.9411312043093497, "grad_norm": 0.12058122223139983, "learning_rate": 1.8121363514249535e-06, "loss": 0.6325, "step": 4892 }, { "epoch": 0.9413235859946133, "grad_norm": 0.20747470258917125, "learning_rate": 1.8003472856820469e-06, "loss": 0.7487, "step": 4893 }, { "epoch": 0.9415159676798769, "grad_norm": 0.20394721578843136, "learning_rate": 1.7885963444254528e-06, "loss": 0.7689, "step": 4894 }, { "epoch": 0.9417083493651405, "grad_norm": 0.17555055194525312, "learning_rate": 1.776883532217255e-06, "loss": 0.7623, "step": 4895 }, { "epoch": 0.941900731050404, "grad_norm": 0.19327367852579702, "learning_rate": 1.7652088536048051e-06, "loss": 0.8155, "step": 4896 }, { "epoch": 0.9420931127356675, "grad_norm": 0.19625057597689977, "learning_rate": 1.7535723131206106e-06, "loss": 0.7783, "step": 4897 }, { "epoch": 0.9422854944209311, "grad_norm": 0.20400440694252497, "learning_rate": 1.7419739152823467e-06, "loss": 0.828, "step": 4898 }, { "epoch": 0.9424778761061947, "grad_norm": 0.18393488664735191, "learning_rate": 1.7304136645929447e-06, "loss": 0.8034, "step": 4899 }, { "epoch": 0.9426702577914583, "grad_norm": 0.20094772619967696, "learning_rate": 1.7188915655404813e-06, "loss": 0.7625, "step": 4900 }, { "epoch": 0.9428626394767218, "grad_norm": 0.20889175951689815, "learning_rate": 1.7074076225982228e-06, "loss": 0.7387, "step": 4901 }, { "epoch": 0.9430550211619854, "grad_norm": 0.20877520388779924, "learning_rate": 1.6959618402246357e-06, "loss": 0.7826, "step": 4902 }, { "epoch": 0.943247402847249, "grad_norm": 0.20604824298646984, "learning_rate": 1.684554222863377e-06, "loss": 0.7536, "step": 4903 }, { "epoch": 0.9434397845325125, "grad_norm": 0.20833188174105638, "learning_rate": 1.6731847749432705e-06, "loss": 0.7726, "step": 4904 }, { "epoch": 0.943632166217776, "grad_norm": 0.1905939338611551, "learning_rate": 1.6618535008783075e-06, "loss": 0.7851, "step": 4905 }, { "epoch": 0.9438245479030396, "grad_norm": 0.17473339925975714, "learning_rate": 1.650560405067725e-06, "loss": 0.8075, "step": 4906 }, { "epoch": 0.9440169295883032, "grad_norm": 0.2098079024843686, "learning_rate": 1.6393054918958373e-06, "loss": 0.7877, "step": 4907 }, { "epoch": 0.9442093112735668, "grad_norm": 0.1765590351813623, "learning_rate": 1.6280887657322275e-06, "loss": 0.7026, "step": 4908 }, { "epoch": 0.9444016929588304, "grad_norm": 0.20522060038809012, "learning_rate": 1.616910230931612e-06, "loss": 0.7803, "step": 4909 }, { "epoch": 0.9445940746440938, "grad_norm": 0.1895634858240166, "learning_rate": 1.6057698918338526e-06, "loss": 0.803, "step": 4910 }, { "epoch": 0.9447864563293574, "grad_norm": 0.20887823696682695, "learning_rate": 1.5946677527640564e-06, "loss": 0.768, "step": 4911 }, { "epoch": 0.944978838014621, "grad_norm": 0.2011338951964416, "learning_rate": 1.5836038180324198e-06, "loss": 0.7596, "step": 4912 }, { "epoch": 0.9451712196998846, "grad_norm": 0.09857462823425739, "learning_rate": 1.5725780919343625e-06, "loss": 0.6149, "step": 4913 }, { "epoch": 0.9453636013851482, "grad_norm": 0.19110070445995705, "learning_rate": 1.561590578750438e-06, "loss": 0.8431, "step": 4914 }, { "epoch": 0.9455559830704117, "grad_norm": 0.10705183063129968, "learning_rate": 1.55064128274639e-06, "loss": 0.6627, "step": 4915 }, { "epoch": 0.9457483647556753, "grad_norm": 0.18767715002377552, "learning_rate": 1.5397302081731068e-06, "loss": 0.7732, "step": 4916 }, { "epoch": 0.9459407464409388, "grad_norm": 0.18658049017053494, "learning_rate": 1.5288573592666444e-06, "loss": 0.7792, "step": 4917 }, { "epoch": 0.9461331281262024, "grad_norm": 0.19410935810776853, "learning_rate": 1.518022740248215e-06, "loss": 0.7653, "step": 4918 }, { "epoch": 0.9463255098114659, "grad_norm": 0.1969480834315606, "learning_rate": 1.5072263553241872e-06, "loss": 0.7391, "step": 4919 }, { "epoch": 0.9465178914967295, "grad_norm": 0.22042562906352212, "learning_rate": 1.4964682086861082e-06, "loss": 0.8014, "step": 4920 }, { "epoch": 0.9467102731819931, "grad_norm": 0.21434565530116545, "learning_rate": 1.4857483045106257e-06, "loss": 0.8255, "step": 4921 }, { "epoch": 0.9469026548672567, "grad_norm": 0.1880478114098402, "learning_rate": 1.4750666469596108e-06, "loss": 0.8526, "step": 4922 }, { "epoch": 0.9470950365525203, "grad_norm": 0.19301804035262432, "learning_rate": 1.464423240180035e-06, "loss": 0.7552, "step": 4923 }, { "epoch": 0.9472874182377837, "grad_norm": 0.18277539605426668, "learning_rate": 1.4538180883040265e-06, "loss": 0.7738, "step": 4924 }, { "epoch": 0.9474797999230473, "grad_norm": 0.19255964404180093, "learning_rate": 1.4432511954488914e-06, "loss": 0.7679, "step": 4925 }, { "epoch": 0.9476721816083109, "grad_norm": 0.2003109690389381, "learning_rate": 1.4327225657170484e-06, "loss": 0.8274, "step": 4926 }, { "epoch": 0.9478645632935745, "grad_norm": 0.2529256838230739, "learning_rate": 1.4222322031960722e-06, "loss": 0.7622, "step": 4927 }, { "epoch": 0.948056944978838, "grad_norm": 0.2007798128792344, "learning_rate": 1.4117801119586938e-06, "loss": 0.7295, "step": 4928 }, { "epoch": 0.9482493266641016, "grad_norm": 0.2064146612887807, "learning_rate": 1.401366296062756e-06, "loss": 0.7988, "step": 4929 }, { "epoch": 0.9484417083493651, "grad_norm": 0.21253122488128606, "learning_rate": 1.3909907595512805e-06, "loss": 0.7563, "step": 4930 }, { "epoch": 0.9486340900346287, "grad_norm": 0.19594401098696818, "learning_rate": 1.3806535064524006e-06, "loss": 0.7622, "step": 4931 }, { "epoch": 0.9488264717198923, "grad_norm": 0.20006590974740868, "learning_rate": 1.370354540779395e-06, "loss": 0.8225, "step": 4932 }, { "epoch": 0.9490188534051558, "grad_norm": 0.19961666934782463, "learning_rate": 1.360093866530665e-06, "loss": 0.8489, "step": 4933 }, { "epoch": 0.9492112350904194, "grad_norm": 0.1787682012740133, "learning_rate": 1.34987148768978e-06, "loss": 0.7803, "step": 4934 }, { "epoch": 0.949403616775683, "grad_norm": 0.19332358414261142, "learning_rate": 1.3396874082253986e-06, "loss": 0.794, "step": 4935 }, { "epoch": 0.9495959984609466, "grad_norm": 0.1812641512625367, "learning_rate": 1.3295416320913357e-06, "loss": 0.7856, "step": 4936 }, { "epoch": 0.94978838014621, "grad_norm": 0.1989031935941376, "learning_rate": 1.3194341632265516e-06, "loss": 0.791, "step": 4937 }, { "epoch": 0.9499807618314736, "grad_norm": 0.18811243738438546, "learning_rate": 1.3093650055550854e-06, "loss": 0.8269, "step": 4938 }, { "epoch": 0.9501731435167372, "grad_norm": 0.18257886050064767, "learning_rate": 1.2993341629861433e-06, "loss": 0.7789, "step": 4939 }, { "epoch": 0.9503655252020008, "grad_norm": 0.2133245520158746, "learning_rate": 1.2893416394140323e-06, "loss": 0.7619, "step": 4940 }, { "epoch": 0.9505579068872644, "grad_norm": 0.18975396720354148, "learning_rate": 1.279387438718216e-06, "loss": 0.7991, "step": 4941 }, { "epoch": 0.9507502885725279, "grad_norm": 0.19857064323121174, "learning_rate": 1.269471564763247e-06, "loss": 0.765, "step": 4942 }, { "epoch": 0.9509426702577914, "grad_norm": 0.18662828700123513, "learning_rate": 1.2595940213988022e-06, "loss": 0.7541, "step": 4943 }, { "epoch": 0.951135051943055, "grad_norm": 0.18835398230614397, "learning_rate": 1.2497548124597024e-06, "loss": 0.7793, "step": 4944 }, { "epoch": 0.9513274336283186, "grad_norm": 0.17478387023669567, "learning_rate": 1.2399539417658369e-06, "loss": 0.7568, "step": 4945 }, { "epoch": 0.9515198153135821, "grad_norm": 0.202205695145859, "learning_rate": 1.2301914131222726e-06, "loss": 0.7676, "step": 4946 }, { "epoch": 0.9517121969988457, "grad_norm": 0.17557825035054928, "learning_rate": 1.2204672303191334e-06, "loss": 0.7156, "step": 4947 }, { "epoch": 0.9519045786841093, "grad_norm": 0.2104179523704845, "learning_rate": 1.2107813971317105e-06, "loss": 0.7211, "step": 4948 }, { "epoch": 0.9520969603693729, "grad_norm": 0.1824421876199497, "learning_rate": 1.2011339173203628e-06, "loss": 0.764, "step": 4949 }, { "epoch": 0.9522893420546364, "grad_norm": 0.18573935390346746, "learning_rate": 1.1915247946305496e-06, "loss": 0.8043, "step": 4950 }, { "epoch": 0.9524817237398999, "grad_norm": 0.19785375132983224, "learning_rate": 1.181954032792909e-06, "loss": 0.7648, "step": 4951 }, { "epoch": 0.9526741054251635, "grad_norm": 0.20543860631318658, "learning_rate": 1.1724216355231022e-06, "loss": 0.7852, "step": 4952 }, { "epoch": 0.9528664871104271, "grad_norm": 0.2028340259818645, "learning_rate": 1.1629276065219575e-06, "loss": 0.7563, "step": 4953 }, { "epoch": 0.9530588687956907, "grad_norm": 0.19028851020885937, "learning_rate": 1.153471949475382e-06, "loss": 0.8187, "step": 4954 }, { "epoch": 0.9532512504809542, "grad_norm": 0.19770772477517148, "learning_rate": 1.1440546680543728e-06, "loss": 0.7667, "step": 4955 }, { "epoch": 0.9534436321662177, "grad_norm": 0.1881578344888228, "learning_rate": 1.1346757659150497e-06, "loss": 0.7922, "step": 4956 }, { "epoch": 0.9536360138514813, "grad_norm": 0.19105553839183334, "learning_rate": 1.1253352466986334e-06, "loss": 0.7829, "step": 4957 }, { "epoch": 0.9538283955367449, "grad_norm": 0.18596201942696644, "learning_rate": 1.1160331140314339e-06, "loss": 0.7951, "step": 4958 }, { "epoch": 0.9540207772220085, "grad_norm": 0.17964246648224086, "learning_rate": 1.1067693715248406e-06, "loss": 0.7522, "step": 4959 }, { "epoch": 0.954213158907272, "grad_norm": 0.18554159631262146, "learning_rate": 1.0975440227753764e-06, "loss": 0.7812, "step": 4960 }, { "epoch": 0.9544055405925356, "grad_norm": 0.19747483620391637, "learning_rate": 1.0883570713646319e-06, "loss": 0.7837, "step": 4961 }, { "epoch": 0.9545979222777992, "grad_norm": 0.1912426147306408, "learning_rate": 1.0792085208593094e-06, "loss": 0.7618, "step": 4962 }, { "epoch": 0.9547903039630627, "grad_norm": 0.20649336454128886, "learning_rate": 1.0700983748111793e-06, "loss": 0.7405, "step": 4963 }, { "epoch": 0.9549826856483262, "grad_norm": 0.20015338461119767, "learning_rate": 1.0610266367571009e-06, "loss": 0.7931, "step": 4964 }, { "epoch": 0.9551750673335898, "grad_norm": 0.1820167678985934, "learning_rate": 1.0519933102190683e-06, "loss": 0.7805, "step": 4965 }, { "epoch": 0.9553674490188534, "grad_norm": 0.18283534087693584, "learning_rate": 1.0429983987041092e-06, "loss": 0.7855, "step": 4966 }, { "epoch": 0.955559830704117, "grad_norm": 0.20364138715242652, "learning_rate": 1.0340419057043526e-06, "loss": 0.7765, "step": 4967 }, { "epoch": 0.9557522123893806, "grad_norm": 0.19497143695534494, "learning_rate": 1.0251238346970393e-06, "loss": 0.7939, "step": 4968 }, { "epoch": 0.955944594074644, "grad_norm": 0.17676365113299417, "learning_rate": 1.0162441891444441e-06, "loss": 0.7811, "step": 4969 }, { "epoch": 0.9561369757599076, "grad_norm": 0.198667291956212, "learning_rate": 1.007402972493976e-06, "loss": 0.7578, "step": 4970 }, { "epoch": 0.9563293574451712, "grad_norm": 0.19772179179023933, "learning_rate": 9.986001881780782e-07, "loss": 0.7611, "step": 4971 }, { "epoch": 0.9565217391304348, "grad_norm": 0.11266567872630005, "learning_rate": 9.898358396143171e-07, "loss": 0.6401, "step": 4972 }, { "epoch": 0.9567141208156984, "grad_norm": 0.18102497242599921, "learning_rate": 9.811099302052929e-07, "loss": 0.7433, "step": 4973 }, { "epoch": 0.9569065025009619, "grad_norm": 0.2097629361828104, "learning_rate": 9.72422463338718e-07, "loss": 0.7575, "step": 4974 }, { "epoch": 0.9570988841862255, "grad_norm": 0.1866565549942456, "learning_rate": 9.637734423873613e-07, "loss": 0.7333, "step": 4975 }, { "epoch": 0.957291265871489, "grad_norm": 0.22394263617154492, "learning_rate": 9.551628707090809e-07, "loss": 0.7595, "step": 4976 }, { "epoch": 0.9574836475567526, "grad_norm": 0.2137917292551065, "learning_rate": 9.465907516467698e-07, "loss": 0.8019, "step": 4977 }, { "epoch": 0.9576760292420161, "grad_norm": 0.11622447073325953, "learning_rate": 9.380570885284545e-07, "loss": 0.6613, "step": 4978 }, { "epoch": 0.9578684109272797, "grad_norm": 0.17949066697305255, "learning_rate": 9.295618846671738e-07, "loss": 0.8253, "step": 4979 }, { "epoch": 0.9580607926125433, "grad_norm": 0.21625228337431615, "learning_rate": 9.211051433610674e-07, "loss": 0.8156, "step": 4980 }, { "epoch": 0.9582531742978069, "grad_norm": 0.18800668216578673, "learning_rate": 9.126868678933198e-07, "loss": 0.7598, "step": 4981 }, { "epoch": 0.9584455559830705, "grad_norm": 0.17700795264665464, "learning_rate": 9.043070615322169e-07, "loss": 0.7401, "step": 4982 }, { "epoch": 0.9586379376683339, "grad_norm": 0.1856753152615524, "learning_rate": 8.959657275310674e-07, "loss": 0.774, "step": 4983 }, { "epoch": 0.9588303193535975, "grad_norm": 0.20692281116635197, "learning_rate": 8.876628691282918e-07, "loss": 0.8069, "step": 4984 }, { "epoch": 0.9590227010388611, "grad_norm": 0.18833273792377028, "learning_rate": 8.793984895473117e-07, "loss": 0.737, "step": 4985 }, { "epoch": 0.9592150827241247, "grad_norm": 0.20290412335696092, "learning_rate": 8.711725919966718e-07, "loss": 0.7788, "step": 4986 }, { "epoch": 0.9594074644093882, "grad_norm": 0.181739526294889, "learning_rate": 8.629851796699284e-07, "loss": 0.6833, "step": 4987 }, { "epoch": 0.9595998460946518, "grad_norm": 0.19646223883085834, "learning_rate": 8.548362557457279e-07, "loss": 0.7525, "step": 4988 }, { "epoch": 0.9597922277799154, "grad_norm": 0.2132240045041183, "learning_rate": 8.467258233877728e-07, "loss": 0.7485, "step": 4989 }, { "epoch": 0.9599846094651789, "grad_norm": 0.20862128295916393, "learning_rate": 8.386538857447779e-07, "loss": 0.8133, "step": 4990 }, { "epoch": 0.9601769911504425, "grad_norm": 0.1975159854136153, "learning_rate": 8.306204459505806e-07, "loss": 0.745, "step": 4991 }, { "epoch": 0.960369372835706, "grad_norm": 0.1840157196169249, "learning_rate": 8.226255071240308e-07, "loss": 0.7569, "step": 4992 }, { "epoch": 0.9605617545209696, "grad_norm": 0.18406179580112772, "learning_rate": 8.146690723690342e-07, "loss": 0.7831, "step": 4993 }, { "epoch": 0.9607541362062332, "grad_norm": 0.17326520666545028, "learning_rate": 8.067511447745535e-07, "loss": 0.8103, "step": 4994 }, { "epoch": 0.9609465178914968, "grad_norm": 0.19250424054242307, "learning_rate": 7.988717274146074e-07, "loss": 0.7143, "step": 4995 }, { "epoch": 0.9611388995767602, "grad_norm": 0.20070078728035115, "learning_rate": 7.910308233482488e-07, "loss": 0.8043, "step": 4996 }, { "epoch": 0.9613312812620238, "grad_norm": 0.20411612167407203, "learning_rate": 7.832284356195762e-07, "loss": 0.7799, "step": 4997 }, { "epoch": 0.9615236629472874, "grad_norm": 0.18516157612719727, "learning_rate": 7.754645672577776e-07, "loss": 0.7465, "step": 4998 }, { "epoch": 0.961716044632551, "grad_norm": 0.1941259521495904, "learning_rate": 7.677392212770196e-07, "loss": 0.7096, "step": 4999 }, { "epoch": 0.9619084263178146, "grad_norm": 0.20819838693777953, "learning_rate": 7.600524006765808e-07, "loss": 0.7652, "step": 5000 }, { "epoch": 0.9621008080030781, "grad_norm": 0.19588782942505667, "learning_rate": 7.524041084407185e-07, "loss": 0.7071, "step": 5001 }, { "epoch": 0.9622931896883417, "grad_norm": 0.21246260892972404, "learning_rate": 7.447943475387797e-07, "loss": 0.7874, "step": 5002 }, { "epoch": 0.9624855713736052, "grad_norm": 0.1949979119927462, "learning_rate": 7.372231209251346e-07, "loss": 0.775, "step": 5003 }, { "epoch": 0.9626779530588688, "grad_norm": 0.21354912175172866, "learning_rate": 7.296904315391873e-07, "loss": 0.8135, "step": 5004 }, { "epoch": 0.9628703347441323, "grad_norm": 0.19484645499157213, "learning_rate": 7.221962823053874e-07, "loss": 0.7851, "step": 5005 }, { "epoch": 0.9630627164293959, "grad_norm": 0.18998372335941155, "learning_rate": 7.147406761332298e-07, "loss": 0.7362, "step": 5006 }, { "epoch": 0.9632550981146595, "grad_norm": 0.17977424263632477, "learning_rate": 7.073236159172325e-07, "loss": 0.7523, "step": 5007 }, { "epoch": 0.9634474797999231, "grad_norm": 0.1959998299391742, "learning_rate": 6.999451045369587e-07, "loss": 0.7704, "step": 5008 }, { "epoch": 0.9636398614851867, "grad_norm": 0.21347735364247225, "learning_rate": 6.926051448569948e-07, "loss": 0.8162, "step": 5009 }, { "epoch": 0.9638322431704501, "grad_norm": 0.19692767990124835, "learning_rate": 6.853037397269724e-07, "loss": 0.7214, "step": 5010 }, { "epoch": 0.9640246248557137, "grad_norm": 0.19993228508115812, "learning_rate": 6.78040891981524e-07, "loss": 0.8017, "step": 5011 }, { "epoch": 0.9642170065409773, "grad_norm": 0.20248159686701922, "learning_rate": 6.70816604440383e-07, "loss": 0.8383, "step": 5012 }, { "epoch": 0.9644093882262409, "grad_norm": 0.18702463029986452, "learning_rate": 6.63630879908217e-07, "loss": 0.8016, "step": 5013 }, { "epoch": 0.9646017699115044, "grad_norm": 0.19820315435871988, "learning_rate": 6.564837211748054e-07, "loss": 0.7541, "step": 5014 }, { "epoch": 0.964794151596768, "grad_norm": 0.19883964462108578, "learning_rate": 6.493751310149176e-07, "loss": 0.7952, "step": 5015 }, { "epoch": 0.9649865332820315, "grad_norm": 0.19377594712113747, "learning_rate": 6.42305112188335e-07, "loss": 0.8193, "step": 5016 }, { "epoch": 0.9651789149672951, "grad_norm": 0.1921864125979164, "learning_rate": 6.352736674398951e-07, "loss": 0.78, "step": 5017 }, { "epoch": 0.9653712966525587, "grad_norm": 0.194551905283146, "learning_rate": 6.282807994994478e-07, "loss": 0.7631, "step": 5018 }, { "epoch": 0.9655636783378222, "grad_norm": 0.17570619944084792, "learning_rate": 6.213265110818655e-07, "loss": 0.782, "step": 5019 }, { "epoch": 0.9657560600230858, "grad_norm": 0.18169129138035253, "learning_rate": 6.144108048870333e-07, "loss": 0.7703, "step": 5020 }, { "epoch": 0.9659484417083494, "grad_norm": 0.21016282548858248, "learning_rate": 6.075336835998812e-07, "loss": 0.7947, "step": 5021 }, { "epoch": 0.966140823393613, "grad_norm": 0.11602605183897413, "learning_rate": 6.00695149890329e-07, "loss": 0.6395, "step": 5022 }, { "epoch": 0.9663332050788765, "grad_norm": 0.2142042767474832, "learning_rate": 5.938952064133418e-07, "loss": 0.7693, "step": 5023 }, { "epoch": 0.96652558676414, "grad_norm": 0.19911997542886153, "learning_rate": 5.871338558088857e-07, "loss": 0.8087, "step": 5024 }, { "epoch": 0.9667179684494036, "grad_norm": 0.2167861865835512, "learning_rate": 5.804111007019497e-07, "loss": 0.79, "step": 5025 }, { "epoch": 0.9669103501346672, "grad_norm": 0.18195584788577632, "learning_rate": 5.737269437025461e-07, "loss": 0.8169, "step": 5026 }, { "epoch": 0.9671027318199308, "grad_norm": 0.23344236786655384, "learning_rate": 5.67081387405688e-07, "loss": 0.7263, "step": 5027 }, { "epoch": 0.9672951135051943, "grad_norm": 0.20189353040548708, "learning_rate": 5.604744343914115e-07, "loss": 0.7702, "step": 5028 }, { "epoch": 0.9674874951904578, "grad_norm": 0.20175084562996315, "learning_rate": 5.539060872247537e-07, "loss": 0.7657, "step": 5029 }, { "epoch": 0.9676798768757214, "grad_norm": 0.17002069222732422, "learning_rate": 5.47376348455797e-07, "loss": 0.766, "step": 5030 }, { "epoch": 0.967872258560985, "grad_norm": 0.10332494014630528, "learning_rate": 5.408852206195913e-07, "loss": 0.5785, "step": 5031 }, { "epoch": 0.9680646402462486, "grad_norm": 0.21591147916715422, "learning_rate": 5.344327062362098e-07, "loss": 0.7479, "step": 5032 }, { "epoch": 0.9682570219315121, "grad_norm": 0.11089068070734753, "learning_rate": 5.280188078107595e-07, "loss": 0.6193, "step": 5033 }, { "epoch": 0.9684494036167757, "grad_norm": 0.20050647352871828, "learning_rate": 5.216435278333376e-07, "loss": 0.7586, "step": 5034 }, { "epoch": 0.9686417853020393, "grad_norm": 0.1875644905360567, "learning_rate": 5.153068687790197e-07, "loss": 0.7914, "step": 5035 }, { "epoch": 0.9688341669873028, "grad_norm": 0.19968339147060682, "learning_rate": 5.09008833107949e-07, "loss": 0.794, "step": 5036 }, { "epoch": 0.9690265486725663, "grad_norm": 0.20628600864694732, "learning_rate": 5.027494232652142e-07, "loss": 0.7548, "step": 5037 }, { "epoch": 0.9692189303578299, "grad_norm": 0.2157308611527685, "learning_rate": 4.9652864168096e-07, "loss": 0.732, "step": 5038 }, { "epoch": 0.9694113120430935, "grad_norm": 0.21345933822415325, "learning_rate": 4.90346490770277e-07, "loss": 0.7591, "step": 5039 }, { "epoch": 0.9696036937283571, "grad_norm": 0.19128477680224762, "learning_rate": 4.84202972933312e-07, "loss": 0.7812, "step": 5040 }, { "epoch": 0.9697960754136207, "grad_norm": 0.21339914228217216, "learning_rate": 4.78098090555179e-07, "loss": 0.8259, "step": 5041 }, { "epoch": 0.9699884570988841, "grad_norm": 0.20383245000477535, "learning_rate": 4.720318460060047e-07, "loss": 0.7982, "step": 5042 }, { "epoch": 0.9701808387841477, "grad_norm": 0.19140913937608145, "learning_rate": 4.6600424164091605e-07, "loss": 0.765, "step": 5043 }, { "epoch": 0.9703732204694113, "grad_norm": 0.20633388408216735, "learning_rate": 4.6001527980004124e-07, "loss": 0.7242, "step": 5044 }, { "epoch": 0.9705656021546749, "grad_norm": 0.18694897486970127, "learning_rate": 4.54064962808487e-07, "loss": 0.7693, "step": 5045 }, { "epoch": 0.9707579838399384, "grad_norm": 0.1888035066781795, "learning_rate": 4.481532929763943e-07, "loss": 0.7847, "step": 5046 }, { "epoch": 0.970950365525202, "grad_norm": 0.19893547526291433, "learning_rate": 4.422802725988606e-07, "loss": 0.7527, "step": 5047 }, { "epoch": 0.9711427472104656, "grad_norm": 0.2163130427919572, "learning_rate": 4.364459039559843e-07, "loss": 0.764, "step": 5048 }, { "epoch": 0.9713351288957291, "grad_norm": 0.18804250641799017, "learning_rate": 4.306501893128978e-07, "loss": 0.8042, "step": 5049 }, { "epoch": 0.9715275105809927, "grad_norm": 0.1909945712697846, "learning_rate": 4.2489313091967906e-07, "loss": 0.7691, "step": 5050 }, { "epoch": 0.9717198922662562, "grad_norm": 0.19908667724496418, "learning_rate": 4.191747310114069e-07, "loss": 0.743, "step": 5051 }, { "epoch": 0.9719122739515198, "grad_norm": 0.1827943018442864, "learning_rate": 4.134949918081832e-07, "loss": 0.747, "step": 5052 }, { "epoch": 0.9721046556367834, "grad_norm": 0.18919658781578505, "learning_rate": 4.0785391551506623e-07, "loss": 0.7517, "step": 5053 }, { "epoch": 0.972297037322047, "grad_norm": 0.1998893277816455, "learning_rate": 4.0225150432211535e-07, "loss": 0.7503, "step": 5054 }, { "epoch": 0.9724894190073105, "grad_norm": 0.20985290035192655, "learning_rate": 3.9668776040437947e-07, "loss": 0.7642, "step": 5055 }, { "epoch": 0.972681800692574, "grad_norm": 0.21880896773764547, "learning_rate": 3.9116268592189755e-07, "loss": 0.7764, "step": 5056 }, { "epoch": 0.9728741823778376, "grad_norm": 0.18600435190874517, "learning_rate": 3.8567628301969803e-07, "loss": 0.7978, "step": 5057 }, { "epoch": 0.9730665640631012, "grad_norm": 0.18519041434863548, "learning_rate": 3.8022855382777724e-07, "loss": 0.8002, "step": 5058 }, { "epoch": 0.9732589457483648, "grad_norm": 0.1997045744254098, "learning_rate": 3.7481950046115433e-07, "loss": 0.7327, "step": 5059 }, { "epoch": 0.9734513274336283, "grad_norm": 0.1975434282203966, "learning_rate": 3.69449125019794e-07, "loss": 0.7813, "step": 5060 }, { "epoch": 0.9736437091188919, "grad_norm": 0.2102395205307336, "learning_rate": 3.641174295886618e-07, "loss": 0.7418, "step": 5061 }, { "epoch": 0.9738360908041555, "grad_norm": 0.21084302766332016, "learning_rate": 3.588244162377019e-07, "loss": 0.7972, "step": 5062 }, { "epoch": 0.974028472489419, "grad_norm": 0.10680780813282693, "learning_rate": 3.5357008702185944e-07, "loss": 0.6051, "step": 5063 }, { "epoch": 0.9742208541746825, "grad_norm": 0.10767965415731404, "learning_rate": 3.483544439810249e-07, "loss": 0.6655, "step": 5064 }, { "epoch": 0.9744132358599461, "grad_norm": 0.1818032987813939, "learning_rate": 3.431774891401118e-07, "loss": 0.7284, "step": 5065 }, { "epoch": 0.9746056175452097, "grad_norm": 0.1822247195244641, "learning_rate": 3.380392245089792e-07, "loss": 0.7306, "step": 5066 }, { "epoch": 0.9747979992304733, "grad_norm": 0.18866309488433697, "learning_rate": 3.329396520824757e-07, "loss": 0.7773, "step": 5067 }, { "epoch": 0.9749903809157369, "grad_norm": 0.2179190505938272, "learning_rate": 3.2787877384045096e-07, "loss": 0.7313, "step": 5068 }, { "epoch": 0.9751827626010003, "grad_norm": 0.1929898169420656, "learning_rate": 3.228565917476889e-07, "loss": 0.7985, "step": 5069 }, { "epoch": 0.9753751442862639, "grad_norm": 0.21962761836725742, "learning_rate": 3.178731077539743e-07, "loss": 0.8124, "step": 5070 }, { "epoch": 0.9755675259715275, "grad_norm": 0.17868743942757817, "learning_rate": 3.1292832379409274e-07, "loss": 0.7789, "step": 5071 }, { "epoch": 0.9757599076567911, "grad_norm": 0.2077544343128289, "learning_rate": 3.080222417877421e-07, "loss": 0.7516, "step": 5072 }, { "epoch": 0.9759522893420547, "grad_norm": 0.20627621441209268, "learning_rate": 3.031548636396764e-07, "loss": 0.7689, "step": 5073 }, { "epoch": 0.9761446710273182, "grad_norm": 0.21086613789835984, "learning_rate": 2.9832619123953965e-07, "loss": 0.785, "step": 5074 }, { "epoch": 0.9763370527125818, "grad_norm": 0.20857101808306772, "learning_rate": 2.9353622646199897e-07, "loss": 0.7745, "step": 5075 }, { "epoch": 0.9765294343978453, "grad_norm": 0.19009498856138834, "learning_rate": 2.8878497116671123e-07, "loss": 0.738, "step": 5076 }, { "epoch": 0.9767218160831089, "grad_norm": 0.20730684470774755, "learning_rate": 2.8407242719823424e-07, "loss": 0.8139, "step": 5077 }, { "epoch": 0.9769141977683724, "grad_norm": 0.18686591980159883, "learning_rate": 2.793985963861712e-07, "loss": 0.7944, "step": 5078 }, { "epoch": 0.977106579453636, "grad_norm": 0.1940066358235648, "learning_rate": 2.747634805450483e-07, "loss": 0.7608, "step": 5079 }, { "epoch": 0.9772989611388996, "grad_norm": 0.19305045317871433, "learning_rate": 2.7016708147439286e-07, "loss": 0.7786, "step": 5080 }, { "epoch": 0.9774913428241632, "grad_norm": 0.19943525913439164, "learning_rate": 2.656094009586663e-07, "loss": 0.8118, "step": 5081 }, { "epoch": 0.9776837245094268, "grad_norm": 0.09963984998993175, "learning_rate": 2.6109044076733093e-07, "loss": 0.6275, "step": 5082 }, { "epoch": 0.9778761061946902, "grad_norm": 0.22755440638601876, "learning_rate": 2.5661020265479454e-07, "loss": 0.8011, "step": 5083 }, { "epoch": 0.9780684878799538, "grad_norm": 0.19079404727582064, "learning_rate": 2.521686883604324e-07, "loss": 0.776, "step": 5084 }, { "epoch": 0.9782608695652174, "grad_norm": 0.19379450940529283, "learning_rate": 2.4776589960862075e-07, "loss": 0.7955, "step": 5085 }, { "epoch": 0.978453251250481, "grad_norm": 0.19595624547578414, "learning_rate": 2.434018381086589e-07, "loss": 0.7852, "step": 5086 }, { "epoch": 0.9786456329357445, "grad_norm": 0.1835692366071131, "learning_rate": 2.3907650555481385e-07, "loss": 0.7392, "step": 5087 }, { "epoch": 0.9788380146210081, "grad_norm": 0.20008555018911575, "learning_rate": 2.3478990362634234e-07, "loss": 0.7735, "step": 5088 }, { "epoch": 0.9790303963062716, "grad_norm": 0.1983932708883502, "learning_rate": 2.3054203398743534e-07, "loss": 0.7946, "step": 5089 }, { "epoch": 0.9792227779915352, "grad_norm": 0.22099009311343099, "learning_rate": 2.2633289828729586e-07, "loss": 0.8304, "step": 5090 }, { "epoch": 0.9794151596767988, "grad_norm": 0.19216193296672, "learning_rate": 2.221624981600168e-07, "loss": 0.7802, "step": 5091 }, { "epoch": 0.9796075413620623, "grad_norm": 0.21007357508674124, "learning_rate": 2.1803083522471402e-07, "loss": 0.8247, "step": 5092 }, { "epoch": 0.9797999230473259, "grad_norm": 0.19753752515211856, "learning_rate": 2.139379110854267e-07, "loss": 0.8051, "step": 5093 }, { "epoch": 0.9799923047325895, "grad_norm": 0.1827165435165928, "learning_rate": 2.0988372733118379e-07, "loss": 0.7802, "step": 5094 }, { "epoch": 0.9801846864178531, "grad_norm": 0.18622777495173687, "learning_rate": 2.058682855359595e-07, "loss": 0.7934, "step": 5095 }, { "epoch": 0.9803770681031165, "grad_norm": 0.183706820747639, "learning_rate": 2.0189158725867352e-07, "loss": 0.7422, "step": 5096 }, { "epoch": 0.9805694497883801, "grad_norm": 0.18746874640578032, "learning_rate": 1.979536340432131e-07, "loss": 0.7467, "step": 5097 }, { "epoch": 0.9807618314736437, "grad_norm": 0.1912590247124948, "learning_rate": 1.9405442741844416e-07, "loss": 0.6998, "step": 5098 }, { "epoch": 0.9809542131589073, "grad_norm": 0.21484572570495, "learning_rate": 1.9019396889816688e-07, "loss": 0.7565, "step": 5099 }, { "epoch": 0.9811465948441709, "grad_norm": 0.21122522679005448, "learning_rate": 1.8637225998114906e-07, "loss": 0.7631, "step": 5100 }, { "epoch": 0.9813389765294344, "grad_norm": 0.19795280768634993, "learning_rate": 1.825893021510927e-07, "loss": 0.7658, "step": 5101 }, { "epoch": 0.981531358214698, "grad_norm": 0.19367372255127005, "learning_rate": 1.788450968766897e-07, "loss": 0.805, "step": 5102 }, { "epoch": 0.9817237398999615, "grad_norm": 0.1871665897390656, "learning_rate": 1.7513964561156615e-07, "loss": 0.766, "step": 5103 }, { "epoch": 0.9819161215852251, "grad_norm": 0.19061395097653064, "learning_rate": 1.7147294979429352e-07, "loss": 0.7841, "step": 5104 }, { "epoch": 0.9821085032704886, "grad_norm": 0.1880005313120643, "learning_rate": 1.6784501084843308e-07, "loss": 0.7995, "step": 5105 }, { "epoch": 0.9823008849557522, "grad_norm": 0.20199317780318157, "learning_rate": 1.6425583018244706e-07, "loss": 0.7775, "step": 5106 }, { "epoch": 0.9824932666410158, "grad_norm": 0.21937532921902397, "learning_rate": 1.607054091897986e-07, "loss": 0.6975, "step": 5107 }, { "epoch": 0.9826856483262794, "grad_norm": 0.1861072491722188, "learning_rate": 1.57193749248874e-07, "loss": 0.7769, "step": 5108 }, { "epoch": 0.982878030011543, "grad_norm": 0.18030684090575236, "learning_rate": 1.5372085172302708e-07, "loss": 0.7932, "step": 5109 }, { "epoch": 0.9830704116968064, "grad_norm": 0.1844455299078495, "learning_rate": 1.5028671796055715e-07, "loss": 0.7833, "step": 5110 }, { "epoch": 0.98326279338207, "grad_norm": 0.19913124324013712, "learning_rate": 1.4689134929470882e-07, "loss": 0.8023, "step": 5111 }, { "epoch": 0.9834551750673336, "grad_norm": 0.1892517285022903, "learning_rate": 1.4353474704368319e-07, "loss": 0.7633, "step": 5112 }, { "epoch": 0.9836475567525972, "grad_norm": 0.19826541196280156, "learning_rate": 1.4021691251062673e-07, "loss": 0.7567, "step": 5113 }, { "epoch": 0.9838399384378607, "grad_norm": 0.23297162537218938, "learning_rate": 1.3693784698363133e-07, "loss": 0.7364, "step": 5114 }, { "epoch": 0.9840323201231242, "grad_norm": 0.19029348116699712, "learning_rate": 1.3369755173575637e-07, "loss": 0.7517, "step": 5115 }, { "epoch": 0.9842247018083878, "grad_norm": 0.19737565720656342, "learning_rate": 1.304960280249845e-07, "loss": 0.8146, "step": 5116 }, { "epoch": 0.9844170834936514, "grad_norm": 0.19269211363464667, "learning_rate": 1.273332770942659e-07, "loss": 0.7901, "step": 5117 }, { "epoch": 0.984609465178915, "grad_norm": 0.11070087975366716, "learning_rate": 1.2420930017148502e-07, "loss": 0.6399, "step": 5118 }, { "epoch": 0.9848018468641785, "grad_norm": 0.22423369184213157, "learning_rate": 1.211240984694717e-07, "loss": 0.7771, "step": 5119 }, { "epoch": 0.9849942285494421, "grad_norm": 0.19576126814547853, "learning_rate": 1.1807767318602336e-07, "loss": 0.7873, "step": 5120 }, { "epoch": 0.9851866102347057, "grad_norm": 0.20547520569183636, "learning_rate": 1.1507002550386059e-07, "loss": 0.7683, "step": 5121 }, { "epoch": 0.9853789919199692, "grad_norm": 0.2104073511347717, "learning_rate": 1.1210115659063825e-07, "loss": 0.7497, "step": 5122 }, { "epoch": 0.9855713736052328, "grad_norm": 0.18731916782556696, "learning_rate": 1.0917106759900097e-07, "loss": 0.725, "step": 5123 }, { "epoch": 0.9857637552904963, "grad_norm": 0.1931963709748766, "learning_rate": 1.062797596664944e-07, "loss": 0.7838, "step": 5124 }, { "epoch": 0.9859561369757599, "grad_norm": 0.19281810610017075, "learning_rate": 1.034272339156206e-07, "loss": 0.7959, "step": 5125 }, { "epoch": 0.9861485186610235, "grad_norm": 0.19199200242841463, "learning_rate": 1.0061349145383813e-07, "loss": 0.7536, "step": 5126 }, { "epoch": 0.9863409003462871, "grad_norm": 0.18017234032728396, "learning_rate": 9.783853337353988e-08, "loss": 0.7973, "step": 5127 }, { "epoch": 0.9865332820315506, "grad_norm": 0.10311522873340201, "learning_rate": 9.510236075205292e-08, "loss": 0.6182, "step": 5128 }, { "epoch": 0.9867256637168141, "grad_norm": 0.18794607505738686, "learning_rate": 9.240497465164976e-08, "loss": 0.7462, "step": 5129 }, { "epoch": 0.9869180454020777, "grad_norm": 0.19636649771026743, "learning_rate": 8.974637611955938e-08, "loss": 0.8229, "step": 5130 }, { "epoch": 0.9871104270873413, "grad_norm": 0.18813317095840965, "learning_rate": 8.712656618793391e-08, "loss": 0.7673, "step": 5131 }, { "epoch": 0.9873028087726049, "grad_norm": 0.21718942466708446, "learning_rate": 8.454554587388197e-08, "loss": 0.7341, "step": 5132 }, { "epoch": 0.9874951904578684, "grad_norm": 0.19844994171633462, "learning_rate": 8.200331617943536e-08, "loss": 0.823, "step": 5133 }, { "epoch": 0.987687572143132, "grad_norm": 0.19222882749808326, "learning_rate": 7.949987809158233e-08, "loss": 0.7867, "step": 5134 }, { "epoch": 0.9878799538283956, "grad_norm": 0.10469913908107316, "learning_rate": 7.703523258223432e-08, "loss": 0.6253, "step": 5135 }, { "epoch": 0.9880723355136591, "grad_norm": 0.18804233936091394, "learning_rate": 7.460938060825929e-08, "loss": 0.7794, "step": 5136 }, { "epoch": 0.9882647171989226, "grad_norm": 0.18086582585192654, "learning_rate": 7.222232311145937e-08, "loss": 0.8065, "step": 5137 }, { "epoch": 0.9884570988841862, "grad_norm": 0.2007293588895569, "learning_rate": 6.987406101855998e-08, "loss": 0.7645, "step": 5138 }, { "epoch": 0.9886494805694498, "grad_norm": 0.2119409422284636, "learning_rate": 6.756459524125402e-08, "loss": 0.8233, "step": 5139 }, { "epoch": 0.9888418622547134, "grad_norm": 0.1938846620026217, "learning_rate": 6.529392667613543e-08, "loss": 0.8551, "step": 5140 }, { "epoch": 0.989034243939977, "grad_norm": 0.1914656596139556, "learning_rate": 6.306205620477678e-08, "loss": 0.7643, "step": 5141 }, { "epoch": 0.9892266256252404, "grad_norm": 0.19143707214913225, "learning_rate": 6.086898469365165e-08, "loss": 0.7807, "step": 5142 }, { "epoch": 0.989419007310504, "grad_norm": 0.18927910018402747, "learning_rate": 5.8714712994190067e-08, "loss": 0.796, "step": 5143 }, { "epoch": 0.9896113889957676, "grad_norm": 0.19474323003571536, "learning_rate": 5.659924194276745e-08, "loss": 0.7755, "step": 5144 }, { "epoch": 0.9898037706810312, "grad_norm": 0.20410927344533158, "learning_rate": 5.4522572360660164e-08, "loss": 0.7691, "step": 5145 }, { "epoch": 0.9899961523662947, "grad_norm": 0.11393783954489818, "learning_rate": 5.2484705054123286e-08, "loss": 0.6302, "step": 5146 }, { "epoch": 0.9901885340515583, "grad_norm": 0.19861871129457176, "learning_rate": 5.048564081431284e-08, "loss": 0.8242, "step": 5147 }, { "epoch": 0.9903809157368219, "grad_norm": 0.20111461259208546, "learning_rate": 4.8525380417330236e-08, "loss": 0.8026, "step": 5148 }, { "epoch": 0.9905732974220854, "grad_norm": 0.18583408049516115, "learning_rate": 4.6603924624244456e-08, "loss": 0.772, "step": 5149 }, { "epoch": 0.990765679107349, "grad_norm": 0.21040636528418216, "learning_rate": 4.472127418099214e-08, "loss": 0.7978, "step": 5150 }, { "epoch": 0.9909580607926125, "grad_norm": 0.220046714913351, "learning_rate": 4.287742981851084e-08, "loss": 0.7765, "step": 5151 }, { "epoch": 0.9911504424778761, "grad_norm": 0.1853226921363562, "learning_rate": 4.1072392252639036e-08, "loss": 0.7538, "step": 5152 }, { "epoch": 0.9913428241631397, "grad_norm": 0.19376895798096017, "learning_rate": 3.930616218414951e-08, "loss": 0.7859, "step": 5153 }, { "epoch": 0.9915352058484033, "grad_norm": 0.17989595673241532, "learning_rate": 3.7578740298749306e-08, "loss": 0.8312, "step": 5154 }, { "epoch": 0.9917275875336667, "grad_norm": 0.19669281826875135, "learning_rate": 3.5890127267090846e-08, "loss": 0.7906, "step": 5155 }, { "epoch": 0.9919199692189303, "grad_norm": 0.19095446253352055, "learning_rate": 3.424032374476083e-08, "loss": 0.7444, "step": 5156 }, { "epoch": 0.9921123509041939, "grad_norm": 0.19292529020090549, "learning_rate": 3.262933037224691e-08, "loss": 0.8027, "step": 5157 }, { "epoch": 0.9923047325894575, "grad_norm": 0.1982167804291955, "learning_rate": 3.105714777501545e-08, "loss": 0.7379, "step": 5158 }, { "epoch": 0.9924971142747211, "grad_norm": 0.19112483656225485, "learning_rate": 2.9523776563422644e-08, "loss": 0.7628, "step": 5159 }, { "epoch": 0.9926894959599846, "grad_norm": 0.19147220030652448, "learning_rate": 2.8029217332781188e-08, "loss": 0.7702, "step": 5160 }, { "epoch": 0.9928818776452482, "grad_norm": 0.21246415403730198, "learning_rate": 2.657347066333804e-08, "loss": 0.7891, "step": 5161 }, { "epoch": 0.9930742593305117, "grad_norm": 0.22470792796395014, "learning_rate": 2.5156537120263335e-08, "loss": 0.804, "step": 5162 }, { "epoch": 0.9932666410157753, "grad_norm": 0.19166106887478293, "learning_rate": 2.3778417253650376e-08, "loss": 0.8184, "step": 5163 }, { "epoch": 0.9934590227010388, "grad_norm": 0.18612943208601795, "learning_rate": 2.243911159853784e-08, "loss": 0.7933, "step": 5164 }, { "epoch": 0.9936514043863024, "grad_norm": 0.20016285568635114, "learning_rate": 2.113862067488759e-08, "loss": 0.8284, "step": 5165 }, { "epoch": 0.993843786071566, "grad_norm": 0.1766278704534028, "learning_rate": 1.9876944987606838e-08, "loss": 0.7625, "step": 5166 }, { "epoch": 0.9940361677568296, "grad_norm": 0.21617060197493493, "learning_rate": 1.8654085026503787e-08, "loss": 0.7768, "step": 5167 }, { "epoch": 0.9942285494420932, "grad_norm": 0.2029689281156863, "learning_rate": 1.747004126635421e-08, "loss": 0.7614, "step": 5168 }, { "epoch": 0.9944209311273566, "grad_norm": 0.18208797593959905, "learning_rate": 1.6324814166823744e-08, "loss": 0.7792, "step": 5169 }, { "epoch": 0.9946133128126202, "grad_norm": 0.18501437065654233, "learning_rate": 1.5218404172545608e-08, "loss": 0.7788, "step": 5170 }, { "epoch": 0.9948056944978838, "grad_norm": 0.20813310847055178, "learning_rate": 1.415081171305399e-08, "loss": 0.7755, "step": 5171 }, { "epoch": 0.9949980761831474, "grad_norm": 0.1979265396554005, "learning_rate": 1.3122037202828452e-08, "loss": 0.8228, "step": 5172 }, { "epoch": 0.995190457868411, "grad_norm": 0.18640071594574373, "learning_rate": 1.2132081041282827e-08, "loss": 0.8138, "step": 5173 }, { "epoch": 0.9953828395536745, "grad_norm": 0.2083895863958879, "learning_rate": 1.1180943612754124e-08, "loss": 0.7739, "step": 5174 }, { "epoch": 0.995575221238938, "grad_norm": 0.20696452130605664, "learning_rate": 1.0268625286491418e-08, "loss": 0.7921, "step": 5175 }, { "epoch": 0.9957676029242016, "grad_norm": 0.2035773529328211, "learning_rate": 9.39512641668916e-09, "loss": 0.7348, "step": 5176 }, { "epoch": 0.9959599846094652, "grad_norm": 0.21153165157228784, "learning_rate": 8.560447342487176e-09, "loss": 0.7698, "step": 5177 }, { "epoch": 0.9961523662947287, "grad_norm": 0.17992222065602068, "learning_rate": 7.76458838791516e-09, "loss": 0.7607, "step": 5178 }, { "epoch": 0.9963447479799923, "grad_norm": 0.1848756510589712, "learning_rate": 7.0075498619703865e-09, "loss": 0.6971, "step": 5179 }, { "epoch": 0.9965371296652559, "grad_norm": 0.20992541158443592, "learning_rate": 6.28933205855109e-09, "loss": 0.8376, "step": 5180 }, { "epoch": 0.9967295113505195, "grad_norm": 0.18132999096615668, "learning_rate": 5.609935256500886e-09, "loss": 0.776, "step": 5181 }, { "epoch": 0.996921893035783, "grad_norm": 0.21152310471483235, "learning_rate": 4.9693597195865634e-09, "loss": 0.739, "step": 5182 }, { "epoch": 0.9971142747210465, "grad_norm": 0.1955628202465568, "learning_rate": 4.367605696486976e-09, "loss": 0.7825, "step": 5183 }, { "epoch": 0.9973066564063101, "grad_norm": 0.18520212965858807, "learning_rate": 3.8046734208374565e-09, "loss": 0.7781, "step": 5184 }, { "epoch": 0.9974990380915737, "grad_norm": 0.19116497720535788, "learning_rate": 3.2805631111743062e-09, "loss": 0.7965, "step": 5185 }, { "epoch": 0.9976914197768373, "grad_norm": 0.18602327312483238, "learning_rate": 2.7952749710014047e-09, "loss": 0.7387, "step": 5186 }, { "epoch": 0.9978838014621008, "grad_norm": 0.19509988037983395, "learning_rate": 2.348809188690293e-09, "loss": 0.7671, "step": 5187 }, { "epoch": 0.9980761831473643, "grad_norm": 0.18617150853959633, "learning_rate": 1.941165937602296e-09, "loss": 0.8128, "step": 5188 }, { "epoch": 0.9982685648326279, "grad_norm": 0.1917150250502024, "learning_rate": 1.5723453759886041e-09, "loss": 0.8479, "step": 5189 }, { "epoch": 0.9984609465178915, "grad_norm": 0.20174714564095567, "learning_rate": 1.2423476470346806e-09, "loss": 0.7517, "step": 5190 }, { "epoch": 0.9986533282031551, "grad_norm": 0.1923638037639128, "learning_rate": 9.511728788602625e-10, "loss": 0.7979, "step": 5191 }, { "epoch": 0.9988457098884186, "grad_norm": 0.19866311819006563, "learning_rate": 6.988211845082582e-10, "loss": 0.8262, "step": 5192 }, { "epoch": 0.9990380915736822, "grad_norm": 0.21096707279446755, "learning_rate": 4.852926619447473e-10, "loss": 0.8372, "step": 5193 }, { "epoch": 0.9992304732589458, "grad_norm": 0.19295171346703938, "learning_rate": 3.1058739408118544e-10, "loss": 0.8045, "step": 5194 }, { "epoch": 0.9994228549442093, "grad_norm": 0.21253871127068089, "learning_rate": 1.7470544874109706e-10, "loss": 0.734, "step": 5195 }, { "epoch": 0.9996152366294728, "grad_norm": 0.1935330653215911, "learning_rate": 7.764687866007592e-11, "loss": 0.7713, "step": 5196 }, { "epoch": 0.9998076183147364, "grad_norm": 0.31249226226528926, "learning_rate": 1.941172155239812e-11, "loss": 0.6309, "step": 5197 }, { "epoch": 1.0, "grad_norm": 0.16393242576311134, "learning_rate": 0.0, "loss": 0.7415, "step": 5198 }, { "epoch": 1.0, "step": 5198, "total_flos": 1375951590064128.0, "train_loss": 0.805585906110208, "train_runtime": 22417.1982, "train_samples_per_second": 29.678, "train_steps_per_second": 0.232 } ], "logging_steps": 1.0, "max_steps": 5198, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 50000, "total_flos": 1375951590064128.0, "train_batch_size": 16, "trial_name": null, "trial_params": null }