diff --git "a/checkpoint-2830/trainer_state.json" "b/checkpoint-2830/trainer_state.json" new file mode 100644--- /dev/null +++ "b/checkpoint-2830/trainer_state.json" @@ -0,0 +1,19843 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 9.99117387466902, + "eval_steps": 500, + "global_step": 2830, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00353045013239188, + "grad_norm": 20.178845075428434, + "learning_rate": 1.1764705882352942e-07, + "loss": 0.7173, + "step": 1 + }, + { + "epoch": 0.00706090026478376, + "grad_norm": 21.408372160532007, + "learning_rate": 2.3529411764705883e-07, + "loss": 0.8369, + "step": 2 + }, + { + "epoch": 0.01059135039717564, + "grad_norm": 22.339887896461786, + "learning_rate": 3.529411764705883e-07, + "loss": 0.7742, + "step": 3 + }, + { + "epoch": 0.01412180052956752, + "grad_norm": 20.114111492682788, + "learning_rate": 4.7058823529411767e-07, + "loss": 0.6685, + "step": 4 + }, + { + "epoch": 0.0176522506619594, + "grad_norm": 19.395831354668825, + "learning_rate": 5.882352941176471e-07, + "loss": 0.724, + "step": 5 + }, + { + "epoch": 0.02118270079435128, + "grad_norm": 20.441308282025314, + "learning_rate": 7.058823529411766e-07, + "loss": 0.8051, + "step": 6 + }, + { + "epoch": 0.02471315092674316, + "grad_norm": 21.439520175660693, + "learning_rate": 8.235294117647059e-07, + "loss": 0.7853, + "step": 7 + }, + { + "epoch": 0.02824360105913504, + "grad_norm": 16.361184069343224, + "learning_rate": 9.411764705882353e-07, + "loss": 0.6029, + "step": 8 + }, + { + "epoch": 0.03177405119152692, + "grad_norm": 16.50966475117204, + "learning_rate": 1.0588235294117648e-06, + "loss": 0.607, + "step": 9 + }, + { + "epoch": 0.0353045013239188, + "grad_norm": 10.733426578588741, + "learning_rate": 1.1764705882352942e-06, + "loss": 0.5416, + "step": 10 + }, + { + "epoch": 0.038834951456310676, + "grad_norm": 10.395447257616585, + "learning_rate": 1.2941176470588237e-06, + "loss": 0.4926, + "step": 11 + }, + { + "epoch": 0.04236540158870256, + "grad_norm": 12.789389493634516, + "learning_rate": 1.4117647058823531e-06, + "loss": 0.5638, + "step": 12 + }, + { + "epoch": 0.04589585172109444, + "grad_norm": 8.94856882496586, + "learning_rate": 1.5294117647058826e-06, + "loss": 0.459, + "step": 13 + }, + { + "epoch": 0.04942630185348632, + "grad_norm": 4.421740571415892, + "learning_rate": 1.6470588235294118e-06, + "loss": 0.322, + "step": 14 + }, + { + "epoch": 0.0529567519858782, + "grad_norm": 4.035083034631525, + "learning_rate": 1.7647058823529414e-06, + "loss": 0.3224, + "step": 15 + }, + { + "epoch": 0.05648720211827008, + "grad_norm": 4.319827920176901, + "learning_rate": 1.8823529411764707e-06, + "loss": 0.3079, + "step": 16 + }, + { + "epoch": 0.06001765225066196, + "grad_norm": 4.9583914837299, + "learning_rate": 2.0000000000000003e-06, + "loss": 0.3102, + "step": 17 + }, + { + "epoch": 0.06354810238305383, + "grad_norm": 4.011276537282153, + "learning_rate": 2.1176470588235296e-06, + "loss": 0.2933, + "step": 18 + }, + { + "epoch": 0.06707855251544571, + "grad_norm": 3.1423743425131607, + "learning_rate": 2.2352941176470592e-06, + "loss": 0.2419, + "step": 19 + }, + { + "epoch": 0.0706090026478376, + "grad_norm": 3.526182340817241, + "learning_rate": 2.3529411764705885e-06, + "loss": 0.2649, + "step": 20 + }, + { + "epoch": 0.07413945278022947, + "grad_norm": 2.1822830906532205, + "learning_rate": 2.470588235294118e-06, + "loss": 0.241, + "step": 21 + }, + { + "epoch": 0.07766990291262135, + "grad_norm": 3.01235390581002, + "learning_rate": 2.5882352941176473e-06, + "loss": 0.2352, + "step": 22 + }, + { + "epoch": 0.08120035304501325, + "grad_norm": 3.34743138297086, + "learning_rate": 2.7058823529411766e-06, + "loss": 0.2196, + "step": 23 + }, + { + "epoch": 0.08473080317740513, + "grad_norm": 2.453669230978404, + "learning_rate": 2.8235294117647062e-06, + "loss": 0.2577, + "step": 24 + }, + { + "epoch": 0.088261253309797, + "grad_norm": 2.0737233723937765, + "learning_rate": 2.9411764705882355e-06, + "loss": 0.2099, + "step": 25 + }, + { + "epoch": 0.09179170344218888, + "grad_norm": 2.3481110465941653, + "learning_rate": 3.058823529411765e-06, + "loss": 0.2302, + "step": 26 + }, + { + "epoch": 0.09532215357458076, + "grad_norm": 2.4508501080365086, + "learning_rate": 3.1764705882352943e-06, + "loss": 0.2183, + "step": 27 + }, + { + "epoch": 0.09885260370697264, + "grad_norm": 1.9310659307898554, + "learning_rate": 3.2941176470588236e-06, + "loss": 0.22, + "step": 28 + }, + { + "epoch": 0.10238305383936452, + "grad_norm": 1.731861214941761, + "learning_rate": 3.4117647058823532e-06, + "loss": 0.1702, + "step": 29 + }, + { + "epoch": 0.1059135039717564, + "grad_norm": 1.7963688467168324, + "learning_rate": 3.529411764705883e-06, + "loss": 0.2141, + "step": 30 + }, + { + "epoch": 0.10944395410414828, + "grad_norm": 1.8118405787472915, + "learning_rate": 3.6470588235294117e-06, + "loss": 0.2155, + "step": 31 + }, + { + "epoch": 0.11297440423654016, + "grad_norm": 2.1536428426895764, + "learning_rate": 3.7647058823529414e-06, + "loss": 0.2348, + "step": 32 + }, + { + "epoch": 0.11650485436893204, + "grad_norm": 1.3361162643597526, + "learning_rate": 3.882352941176471e-06, + "loss": 0.211, + "step": 33 + }, + { + "epoch": 0.12003530450132392, + "grad_norm": 1.3679923730470394, + "learning_rate": 4.000000000000001e-06, + "loss": 0.2006, + "step": 34 + }, + { + "epoch": 0.1235657546337158, + "grad_norm": 1.6951013269663189, + "learning_rate": 4.11764705882353e-06, + "loss": 0.2173, + "step": 35 + }, + { + "epoch": 0.12709620476610767, + "grad_norm": 1.1243675718278554, + "learning_rate": 4.235294117647059e-06, + "loss": 0.21, + "step": 36 + }, + { + "epoch": 0.13062665489849956, + "grad_norm": 1.798056949309266, + "learning_rate": 4.352941176470588e-06, + "loss": 0.2151, + "step": 37 + }, + { + "epoch": 0.13415710503089143, + "grad_norm": 1.5535634835962755, + "learning_rate": 4.4705882352941184e-06, + "loss": 0.2136, + "step": 38 + }, + { + "epoch": 0.13768755516328332, + "grad_norm": 1.3322693781225434, + "learning_rate": 4.588235294117647e-06, + "loss": 0.2016, + "step": 39 + }, + { + "epoch": 0.1412180052956752, + "grad_norm": 1.2353366230952958, + "learning_rate": 4.705882352941177e-06, + "loss": 0.2158, + "step": 40 + }, + { + "epoch": 0.14474845542806708, + "grad_norm": 1.649116854499638, + "learning_rate": 4.823529411764706e-06, + "loss": 0.2085, + "step": 41 + }, + { + "epoch": 0.14827890556045895, + "grad_norm": 1.5327812092407431, + "learning_rate": 4.941176470588236e-06, + "loss": 0.223, + "step": 42 + }, + { + "epoch": 0.15180935569285084, + "grad_norm": 1.3654123529518498, + "learning_rate": 5.058823529411765e-06, + "loss": 0.1897, + "step": 43 + }, + { + "epoch": 0.1553398058252427, + "grad_norm": 1.2874875896720268, + "learning_rate": 5.176470588235295e-06, + "loss": 0.208, + "step": 44 + }, + { + "epoch": 0.1588702559576346, + "grad_norm": 1.5894455663917926, + "learning_rate": 5.294117647058824e-06, + "loss": 0.2114, + "step": 45 + }, + { + "epoch": 0.1624007060900265, + "grad_norm": 1.4199861690645073, + "learning_rate": 5.411764705882353e-06, + "loss": 0.2204, + "step": 46 + }, + { + "epoch": 0.16593115622241836, + "grad_norm": 1.6111716054466962, + "learning_rate": 5.529411764705883e-06, + "loss": 0.2296, + "step": 47 + }, + { + "epoch": 0.16946160635481025, + "grad_norm": 1.4732964348658326, + "learning_rate": 5.6470588235294125e-06, + "loss": 0.2236, + "step": 48 + }, + { + "epoch": 0.17299205648720212, + "grad_norm": 1.7125830950347147, + "learning_rate": 5.764705882352941e-06, + "loss": 0.1889, + "step": 49 + }, + { + "epoch": 0.176522506619594, + "grad_norm": 1.4473702644438498, + "learning_rate": 5.882352941176471e-06, + "loss": 0.2002, + "step": 50 + }, + { + "epoch": 0.18005295675198588, + "grad_norm": 1.5337680099050695, + "learning_rate": 6e-06, + "loss": 0.2162, + "step": 51 + }, + { + "epoch": 0.18358340688437777, + "grad_norm": 1.6166188689363534, + "learning_rate": 6.11764705882353e-06, + "loss": 0.1984, + "step": 52 + }, + { + "epoch": 0.18711385701676964, + "grad_norm": 1.193587355481407, + "learning_rate": 6.2352941176470595e-06, + "loss": 0.1952, + "step": 53 + }, + { + "epoch": 0.19064430714916153, + "grad_norm": 1.150057201671324, + "learning_rate": 6.352941176470589e-06, + "loss": 0.2181, + "step": 54 + }, + { + "epoch": 0.1941747572815534, + "grad_norm": 1.6229574548509673, + "learning_rate": 6.470588235294119e-06, + "loss": 0.2064, + "step": 55 + }, + { + "epoch": 0.1977052074139453, + "grad_norm": 1.3194440409324901, + "learning_rate": 6.588235294117647e-06, + "loss": 0.2024, + "step": 56 + }, + { + "epoch": 0.20123565754633715, + "grad_norm": 1.5072216772819462, + "learning_rate": 6.705882352941176e-06, + "loss": 0.2041, + "step": 57 + }, + { + "epoch": 0.20476610767872905, + "grad_norm": 1.2184767982361704, + "learning_rate": 6.8235294117647065e-06, + "loss": 0.2165, + "step": 58 + }, + { + "epoch": 0.2082965578111209, + "grad_norm": 1.4524431095544894, + "learning_rate": 6.941176470588236e-06, + "loss": 0.2108, + "step": 59 + }, + { + "epoch": 0.2118270079435128, + "grad_norm": 1.132320106750761, + "learning_rate": 7.058823529411766e-06, + "loss": 0.1904, + "step": 60 + }, + { + "epoch": 0.21535745807590467, + "grad_norm": 1.215224093914693, + "learning_rate": 7.176470588235295e-06, + "loss": 0.1994, + "step": 61 + }, + { + "epoch": 0.21888790820829657, + "grad_norm": 1.0022175847945094, + "learning_rate": 7.294117647058823e-06, + "loss": 0.186, + "step": 62 + }, + { + "epoch": 0.22241835834068843, + "grad_norm": 1.3628751075073207, + "learning_rate": 7.4117647058823535e-06, + "loss": 0.1911, + "step": 63 + }, + { + "epoch": 0.22594880847308033, + "grad_norm": 1.2505219061147377, + "learning_rate": 7.529411764705883e-06, + "loss": 0.2018, + "step": 64 + }, + { + "epoch": 0.2294792586054722, + "grad_norm": 1.3707206838948995, + "learning_rate": 7.647058823529411e-06, + "loss": 0.1997, + "step": 65 + }, + { + "epoch": 0.23300970873786409, + "grad_norm": 1.3412107722466498, + "learning_rate": 7.764705882352941e-06, + "loss": 0.2174, + "step": 66 + }, + { + "epoch": 0.23654015887025595, + "grad_norm": 1.3129029846369458, + "learning_rate": 7.882352941176471e-06, + "loss": 0.1947, + "step": 67 + }, + { + "epoch": 0.24007060900264784, + "grad_norm": 1.6510384082728948, + "learning_rate": 8.000000000000001e-06, + "loss": 0.1724, + "step": 68 + }, + { + "epoch": 0.2436010591350397, + "grad_norm": 1.2774008213531902, + "learning_rate": 8.11764705882353e-06, + "loss": 0.2002, + "step": 69 + }, + { + "epoch": 0.2471315092674316, + "grad_norm": 1.3287314560957624, + "learning_rate": 8.23529411764706e-06, + "loss": 0.1928, + "step": 70 + }, + { + "epoch": 0.2506619593998235, + "grad_norm": 1.0681415016582574, + "learning_rate": 8.35294117647059e-06, + "loss": 0.1721, + "step": 71 + }, + { + "epoch": 0.25419240953221534, + "grad_norm": 1.21518473330371, + "learning_rate": 8.470588235294118e-06, + "loss": 0.1985, + "step": 72 + }, + { + "epoch": 0.25772285966460723, + "grad_norm": 1.4756769784911512, + "learning_rate": 8.588235294117647e-06, + "loss": 0.2281, + "step": 73 + }, + { + "epoch": 0.2612533097969991, + "grad_norm": 0.8760360831121267, + "learning_rate": 8.705882352941177e-06, + "loss": 0.1737, + "step": 74 + }, + { + "epoch": 0.264783759929391, + "grad_norm": 1.336571680306999, + "learning_rate": 8.823529411764707e-06, + "loss": 0.1901, + "step": 75 + }, + { + "epoch": 0.26831421006178285, + "grad_norm": 1.4034937627691704, + "learning_rate": 8.941176470588237e-06, + "loss": 0.2238, + "step": 76 + }, + { + "epoch": 0.27184466019417475, + "grad_norm": 1.1091052692159387, + "learning_rate": 9.058823529411765e-06, + "loss": 0.2, + "step": 77 + }, + { + "epoch": 0.27537511032656664, + "grad_norm": 1.5465224562234414, + "learning_rate": 9.176470588235294e-06, + "loss": 0.192, + "step": 78 + }, + { + "epoch": 0.27890556045895853, + "grad_norm": 1.4355050182008238, + "learning_rate": 9.294117647058824e-06, + "loss": 0.1853, + "step": 79 + }, + { + "epoch": 0.2824360105913504, + "grad_norm": 1.6148908553292696, + "learning_rate": 9.411764705882354e-06, + "loss": 0.1944, + "step": 80 + }, + { + "epoch": 0.28596646072374227, + "grad_norm": 1.006979010091179, + "learning_rate": 9.529411764705882e-06, + "loss": 0.1842, + "step": 81 + }, + { + "epoch": 0.28949691085613416, + "grad_norm": 1.2631451183457176, + "learning_rate": 9.647058823529412e-06, + "loss": 0.1889, + "step": 82 + }, + { + "epoch": 0.29302736098852605, + "grad_norm": 1.0554142641488224, + "learning_rate": 9.764705882352942e-06, + "loss": 0.1954, + "step": 83 + }, + { + "epoch": 0.2965578111209179, + "grad_norm": 1.2871157434331197, + "learning_rate": 9.882352941176472e-06, + "loss": 0.1783, + "step": 84 + }, + { + "epoch": 0.3000882612533098, + "grad_norm": 1.8307918651683295, + "learning_rate": 1e-05, + "loss": 0.2095, + "step": 85 + }, + { + "epoch": 0.3036187113857017, + "grad_norm": 1.27120661506548, + "learning_rate": 9.999996725424495e-06, + "loss": 0.1818, + "step": 86 + }, + { + "epoch": 0.30714916151809357, + "grad_norm": 1.3010925765539272, + "learning_rate": 9.999986901702262e-06, + "loss": 0.1967, + "step": 87 + }, + { + "epoch": 0.3106796116504854, + "grad_norm": 1.012603025593678, + "learning_rate": 9.999970528846173e-06, + "loss": 0.2084, + "step": 88 + }, + { + "epoch": 0.3142100617828773, + "grad_norm": 1.042462350269043, + "learning_rate": 9.99994760687767e-06, + "loss": 0.2113, + "step": 89 + }, + { + "epoch": 0.3177405119152692, + "grad_norm": 1.479637631853533, + "learning_rate": 9.999918135826783e-06, + "loss": 0.1923, + "step": 90 + }, + { + "epoch": 0.3212709620476611, + "grad_norm": 1.6536152705406355, + "learning_rate": 9.999882115732109e-06, + "loss": 0.1757, + "step": 91 + }, + { + "epoch": 0.324801412180053, + "grad_norm": 1.0022195112884287, + "learning_rate": 9.999839546640827e-06, + "loss": 0.2014, + "step": 92 + }, + { + "epoch": 0.3283318623124448, + "grad_norm": 1.0752815706731398, + "learning_rate": 9.9997904286087e-06, + "loss": 0.1947, + "step": 93 + }, + { + "epoch": 0.3318623124448367, + "grad_norm": 0.9640451554876094, + "learning_rate": 9.999734761700061e-06, + "loss": 0.1947, + "step": 94 + }, + { + "epoch": 0.3353927625772286, + "grad_norm": 0.9472529792678961, + "learning_rate": 9.999672545987826e-06, + "loss": 0.1933, + "step": 95 + }, + { + "epoch": 0.3389232127096205, + "grad_norm": 0.8262321839814756, + "learning_rate": 9.999603781553487e-06, + "loss": 0.1824, + "step": 96 + }, + { + "epoch": 0.34245366284201234, + "grad_norm": 1.302452468053817, + "learning_rate": 9.999528468487113e-06, + "loss": 0.2068, + "step": 97 + }, + { + "epoch": 0.34598411297440423, + "grad_norm": 0.903902282892577, + "learning_rate": 9.999446606887349e-06, + "loss": 0.1978, + "step": 98 + }, + { + "epoch": 0.34951456310679613, + "grad_norm": 1.0828870366995145, + "learning_rate": 9.999358196861422e-06, + "loss": 0.218, + "step": 99 + }, + { + "epoch": 0.353045013239188, + "grad_norm": 1.1262059857251303, + "learning_rate": 9.999263238525135e-06, + "loss": 0.2054, + "step": 100 + }, + { + "epoch": 0.35657546337157986, + "grad_norm": 0.9433649602156967, + "learning_rate": 9.999161732002867e-06, + "loss": 0.1811, + "step": 101 + }, + { + "epoch": 0.36010591350397175, + "grad_norm": 1.2486258616646937, + "learning_rate": 9.999053677427573e-06, + "loss": 0.2028, + "step": 102 + }, + { + "epoch": 0.36363636363636365, + "grad_norm": 1.0344680877191332, + "learning_rate": 9.998939074940788e-06, + "loss": 0.2063, + "step": 103 + }, + { + "epoch": 0.36716681376875554, + "grad_norm": 1.3236038216383461, + "learning_rate": 9.99881792469262e-06, + "loss": 0.2014, + "step": 104 + }, + { + "epoch": 0.3706972639011474, + "grad_norm": 1.0486814111295473, + "learning_rate": 9.998690226841756e-06, + "loss": 0.2005, + "step": 105 + }, + { + "epoch": 0.37422771403353927, + "grad_norm": 0.9052783364228919, + "learning_rate": 9.998555981555459e-06, + "loss": 0.1806, + "step": 106 + }, + { + "epoch": 0.37775816416593117, + "grad_norm": 0.8762864385995432, + "learning_rate": 9.998415189009566e-06, + "loss": 0.1963, + "step": 107 + }, + { + "epoch": 0.38128861429832306, + "grad_norm": 1.0133330127905669, + "learning_rate": 9.998267849388494e-06, + "loss": 0.2063, + "step": 108 + }, + { + "epoch": 0.3848190644307149, + "grad_norm": 1.1032798889585722, + "learning_rate": 9.99811396288523e-06, + "loss": 0.2136, + "step": 109 + }, + { + "epoch": 0.3883495145631068, + "grad_norm": 0.9526777003854597, + "learning_rate": 9.99795352970134e-06, + "loss": 0.1997, + "step": 110 + }, + { + "epoch": 0.3918799646954987, + "grad_norm": 1.3675080989097932, + "learning_rate": 9.997786550046965e-06, + "loss": 0.1866, + "step": 111 + }, + { + "epoch": 0.3954104148278906, + "grad_norm": 1.0304325204839548, + "learning_rate": 9.99761302414082e-06, + "loss": 0.2001, + "step": 112 + }, + { + "epoch": 0.3989408649602824, + "grad_norm": 0.8029445134568592, + "learning_rate": 9.997432952210193e-06, + "loss": 0.1806, + "step": 113 + }, + { + "epoch": 0.4024713150926743, + "grad_norm": 0.9662432601060303, + "learning_rate": 9.997246334490952e-06, + "loss": 0.1983, + "step": 114 + }, + { + "epoch": 0.4060017652250662, + "grad_norm": 0.6650992554809413, + "learning_rate": 9.997053171227527e-06, + "loss": 0.1904, + "step": 115 + }, + { + "epoch": 0.4095322153574581, + "grad_norm": 1.2497143912397224, + "learning_rate": 9.996853462672935e-06, + "loss": 0.1921, + "step": 116 + }, + { + "epoch": 0.41306266548984993, + "grad_norm": 1.2105700846879488, + "learning_rate": 9.996647209088759e-06, + "loss": 0.1913, + "step": 117 + }, + { + "epoch": 0.4165931156222418, + "grad_norm": 0.8745391363159488, + "learning_rate": 9.996434410745158e-06, + "loss": 0.1877, + "step": 118 + }, + { + "epoch": 0.4201235657546337, + "grad_norm": 1.155984605743064, + "learning_rate": 9.996215067920855e-06, + "loss": 0.1931, + "step": 119 + }, + { + "epoch": 0.4236540158870256, + "grad_norm": 0.9959150019854762, + "learning_rate": 9.99598918090316e-06, + "loss": 0.1901, + "step": 120 + }, + { + "epoch": 0.42718446601941745, + "grad_norm": 1.051448796745704, + "learning_rate": 9.995756749987942e-06, + "loss": 0.1642, + "step": 121 + }, + { + "epoch": 0.43071491615180935, + "grad_norm": 0.935738059968614, + "learning_rate": 9.995517775479647e-06, + "loss": 0.1702, + "step": 122 + }, + { + "epoch": 0.43424536628420124, + "grad_norm": 1.085401876613817, + "learning_rate": 9.995272257691291e-06, + "loss": 0.1992, + "step": 123 + }, + { + "epoch": 0.43777581641659313, + "grad_norm": 0.9924411904534596, + "learning_rate": 9.99502019694446e-06, + "loss": 0.1922, + "step": 124 + }, + { + "epoch": 0.44130626654898497, + "grad_norm": 1.3346252020931497, + "learning_rate": 9.994761593569312e-06, + "loss": 0.1894, + "step": 125 + }, + { + "epoch": 0.44483671668137686, + "grad_norm": 0.996928968088298, + "learning_rate": 9.994496447904573e-06, + "loss": 0.18, + "step": 126 + }, + { + "epoch": 0.44836716681376876, + "grad_norm": 1.185034118977359, + "learning_rate": 9.99422476029754e-06, + "loss": 0.2015, + "step": 127 + }, + { + "epoch": 0.45189761694616065, + "grad_norm": 1.079465604221569, + "learning_rate": 9.993946531104075e-06, + "loss": 0.1795, + "step": 128 + }, + { + "epoch": 0.4554280670785525, + "grad_norm": 1.0665272432739623, + "learning_rate": 9.993661760688611e-06, + "loss": 0.199, + "step": 129 + }, + { + "epoch": 0.4589585172109444, + "grad_norm": 1.2364225145584533, + "learning_rate": 9.993370449424153e-06, + "loss": 0.2198, + "step": 130 + }, + { + "epoch": 0.4624889673433363, + "grad_norm": 0.755126320981263, + "learning_rate": 9.993072597692266e-06, + "loss": 0.2006, + "step": 131 + }, + { + "epoch": 0.46601941747572817, + "grad_norm": 1.3871006955467502, + "learning_rate": 9.992768205883085e-06, + "loss": 0.2037, + "step": 132 + }, + { + "epoch": 0.46954986760812, + "grad_norm": 0.7426689064084516, + "learning_rate": 9.992457274395312e-06, + "loss": 0.1774, + "step": 133 + }, + { + "epoch": 0.4730803177405119, + "grad_norm": 0.9196839690873588, + "learning_rate": 9.992139803636218e-06, + "loss": 0.1973, + "step": 134 + }, + { + "epoch": 0.4766107678729038, + "grad_norm": 0.7786869026927636, + "learning_rate": 9.99181579402163e-06, + "loss": 0.1842, + "step": 135 + }, + { + "epoch": 0.4801412180052957, + "grad_norm": 1.03244467425382, + "learning_rate": 9.991485245975949e-06, + "loss": 0.1663, + "step": 136 + }, + { + "epoch": 0.4836716681376876, + "grad_norm": 1.0713507973139096, + "learning_rate": 9.991148159932134e-06, + "loss": 0.1941, + "step": 137 + }, + { + "epoch": 0.4872021182700794, + "grad_norm": 0.9453234857914946, + "learning_rate": 9.990804536331715e-06, + "loss": 0.2007, + "step": 138 + }, + { + "epoch": 0.4907325684024713, + "grad_norm": 0.7607319035095186, + "learning_rate": 9.990454375624778e-06, + "loss": 0.18, + "step": 139 + }, + { + "epoch": 0.4942630185348632, + "grad_norm": 0.9915191845505621, + "learning_rate": 9.990097678269972e-06, + "loss": 0.1832, + "step": 140 + }, + { + "epoch": 0.4977934686672551, + "grad_norm": 0.6818554079855815, + "learning_rate": 9.989734444734513e-06, + "loss": 0.1779, + "step": 141 + }, + { + "epoch": 0.501323918799647, + "grad_norm": 0.8679707616739033, + "learning_rate": 9.989364675494175e-06, + "loss": 0.1585, + "step": 142 + }, + { + "epoch": 0.5048543689320388, + "grad_norm": 0.834867345483289, + "learning_rate": 9.988988371033292e-06, + "loss": 0.1684, + "step": 143 + }, + { + "epoch": 0.5083848190644307, + "grad_norm": 0.7262565816433312, + "learning_rate": 9.98860553184476e-06, + "loss": 0.1688, + "step": 144 + }, + { + "epoch": 0.5119152691968226, + "grad_norm": 0.9922761294466914, + "learning_rate": 9.988216158430033e-06, + "loss": 0.1846, + "step": 145 + }, + { + "epoch": 0.5154457193292145, + "grad_norm": 1.0610599926539903, + "learning_rate": 9.987820251299121e-06, + "loss": 0.1965, + "step": 146 + }, + { + "epoch": 0.5189761694616064, + "grad_norm": 0.8658381475355275, + "learning_rate": 9.987417810970601e-06, + "loss": 0.1923, + "step": 147 + }, + { + "epoch": 0.5225066195939982, + "grad_norm": 1.0687893341105537, + "learning_rate": 9.987008837971595e-06, + "loss": 0.2084, + "step": 148 + }, + { + "epoch": 0.5260370697263901, + "grad_norm": 1.0221589345087516, + "learning_rate": 9.986593332837795e-06, + "loss": 0.1825, + "step": 149 + }, + { + "epoch": 0.529567519858782, + "grad_norm": 0.9977245435165767, + "learning_rate": 9.986171296113437e-06, + "loss": 0.2014, + "step": 150 + }, + { + "epoch": 0.5330979699911739, + "grad_norm": 1.0012549684835264, + "learning_rate": 9.985742728351317e-06, + "loss": 0.1885, + "step": 151 + }, + { + "epoch": 0.5366284201235657, + "grad_norm": 1.2108459546620967, + "learning_rate": 9.98530763011279e-06, + "loss": 0.1888, + "step": 152 + }, + { + "epoch": 0.5401588702559577, + "grad_norm": 0.849586927833, + "learning_rate": 9.98486600196776e-06, + "loss": 0.1957, + "step": 153 + }, + { + "epoch": 0.5436893203883495, + "grad_norm": 0.929219836766702, + "learning_rate": 9.984417844494681e-06, + "loss": 0.2051, + "step": 154 + }, + { + "epoch": 0.5472197705207414, + "grad_norm": 0.7551838277587243, + "learning_rate": 9.983963158280568e-06, + "loss": 0.1687, + "step": 155 + }, + { + "epoch": 0.5507502206531333, + "grad_norm": 0.8965218043033845, + "learning_rate": 9.983501943920978e-06, + "loss": 0.2092, + "step": 156 + }, + { + "epoch": 0.5542806707855251, + "grad_norm": 0.9118641687914303, + "learning_rate": 9.98303420202003e-06, + "loss": 0.1857, + "step": 157 + }, + { + "epoch": 0.5578111209179171, + "grad_norm": 1.0557315457012684, + "learning_rate": 9.98255993319038e-06, + "loss": 0.1796, + "step": 158 + }, + { + "epoch": 0.5613415710503089, + "grad_norm": 0.9917786126115438, + "learning_rate": 9.982079138053243e-06, + "loss": 0.1797, + "step": 159 + }, + { + "epoch": 0.5648720211827007, + "grad_norm": 0.9264989403126399, + "learning_rate": 9.981591817238379e-06, + "loss": 0.1814, + "step": 160 + }, + { + "epoch": 0.5684024713150927, + "grad_norm": 1.1038265073069693, + "learning_rate": 9.981097971384094e-06, + "loss": 0.2039, + "step": 161 + }, + { + "epoch": 0.5719329214474845, + "grad_norm": 1.1450789849461764, + "learning_rate": 9.980597601137244e-06, + "loss": 0.2262, + "step": 162 + }, + { + "epoch": 0.5754633715798765, + "grad_norm": 1.0093808973803213, + "learning_rate": 9.980090707153226e-06, + "loss": 0.1767, + "step": 163 + }, + { + "epoch": 0.5789938217122683, + "grad_norm": 0.8305835473538516, + "learning_rate": 9.979577290095988e-06, + "loss": 0.1854, + "step": 164 + }, + { + "epoch": 0.5825242718446602, + "grad_norm": 0.8814645905158526, + "learning_rate": 9.97905735063802e-06, + "loss": 0.207, + "step": 165 + }, + { + "epoch": 0.5860547219770521, + "grad_norm": 1.1283393561929782, + "learning_rate": 9.978530889460351e-06, + "loss": 0.1482, + "step": 166 + }, + { + "epoch": 0.589585172109444, + "grad_norm": 0.8882147790176694, + "learning_rate": 9.977997907252557e-06, + "loss": 0.1778, + "step": 167 + }, + { + "epoch": 0.5931156222418358, + "grad_norm": 0.8063636359843489, + "learning_rate": 9.977458404712755e-06, + "loss": 0.1922, + "step": 168 + }, + { + "epoch": 0.5966460723742277, + "grad_norm": 0.8023695735241588, + "learning_rate": 9.976912382547603e-06, + "loss": 0.1979, + "step": 169 + }, + { + "epoch": 0.6001765225066196, + "grad_norm": 0.8128057874275411, + "learning_rate": 9.976359841472294e-06, + "loss": 0.196, + "step": 170 + }, + { + "epoch": 0.6037069726390115, + "grad_norm": 0.8202708710729476, + "learning_rate": 9.975800782210565e-06, + "loss": 0.1829, + "step": 171 + }, + { + "epoch": 0.6072374227714034, + "grad_norm": 0.9514477549047732, + "learning_rate": 9.975235205494689e-06, + "loss": 0.218, + "step": 172 + }, + { + "epoch": 0.6107678729037952, + "grad_norm": 1.0222112927480378, + "learning_rate": 9.974663112065474e-06, + "loss": 0.1783, + "step": 173 + }, + { + "epoch": 0.6142983230361871, + "grad_norm": 0.891305185008567, + "learning_rate": 9.974084502672266e-06, + "loss": 0.1588, + "step": 174 + }, + { + "epoch": 0.617828773168579, + "grad_norm": 0.8521407342632297, + "learning_rate": 9.973499378072947e-06, + "loss": 0.1997, + "step": 175 + }, + { + "epoch": 0.6213592233009708, + "grad_norm": 0.8261272762704142, + "learning_rate": 9.972907739033927e-06, + "loss": 0.2064, + "step": 176 + }, + { + "epoch": 0.6248896734333628, + "grad_norm": 0.9844072163334512, + "learning_rate": 9.972309586330156e-06, + "loss": 0.1934, + "step": 177 + }, + { + "epoch": 0.6284201235657546, + "grad_norm": 0.9760362259758513, + "learning_rate": 9.97170492074511e-06, + "loss": 0.1711, + "step": 178 + }, + { + "epoch": 0.6319505736981466, + "grad_norm": 0.8171800425370801, + "learning_rate": 9.9710937430708e-06, + "loss": 0.1814, + "step": 179 + }, + { + "epoch": 0.6354810238305384, + "grad_norm": 0.7848321856443494, + "learning_rate": 9.970476054107763e-06, + "loss": 0.1688, + "step": 180 + }, + { + "epoch": 0.6390114739629302, + "grad_norm": 0.7304874742153669, + "learning_rate": 9.969851854665072e-06, + "loss": 0.1975, + "step": 181 + }, + { + "epoch": 0.6425419240953222, + "grad_norm": 0.8122652782447508, + "learning_rate": 9.969221145560315e-06, + "loss": 0.1856, + "step": 182 + }, + { + "epoch": 0.646072374227714, + "grad_norm": 0.5663405788505017, + "learning_rate": 9.968583927619618e-06, + "loss": 0.1823, + "step": 183 + }, + { + "epoch": 0.649602824360106, + "grad_norm": 0.8631102591025225, + "learning_rate": 9.967940201677628e-06, + "loss": 0.191, + "step": 184 + }, + { + "epoch": 0.6531332744924978, + "grad_norm": 0.9219061678650854, + "learning_rate": 9.967289968577515e-06, + "loss": 0.1867, + "step": 185 + }, + { + "epoch": 0.6566637246248896, + "grad_norm": 0.8104447482059561, + "learning_rate": 9.966633229170975e-06, + "loss": 0.2085, + "step": 186 + }, + { + "epoch": 0.6601941747572816, + "grad_norm": 0.696325768189324, + "learning_rate": 9.965969984318227e-06, + "loss": 0.1955, + "step": 187 + }, + { + "epoch": 0.6637246248896734, + "grad_norm": 0.957747594144563, + "learning_rate": 9.965300234888007e-06, + "loss": 0.1973, + "step": 188 + }, + { + "epoch": 0.6672550750220653, + "grad_norm": 0.8838107190805117, + "learning_rate": 9.964623981757572e-06, + "loss": 0.1862, + "step": 189 + }, + { + "epoch": 0.6707855251544572, + "grad_norm": 1.0086692709079665, + "learning_rate": 9.9639412258127e-06, + "loss": 0.1876, + "step": 190 + }, + { + "epoch": 0.6743159752868491, + "grad_norm": 0.7851582403291492, + "learning_rate": 9.963251967947687e-06, + "loss": 0.1925, + "step": 191 + }, + { + "epoch": 0.677846425419241, + "grad_norm": 0.9006177750516297, + "learning_rate": 9.962556209065343e-06, + "loss": 0.1651, + "step": 192 + }, + { + "epoch": 0.6813768755516328, + "grad_norm": 0.8917643734840932, + "learning_rate": 9.961853950076992e-06, + "loss": 0.164, + "step": 193 + }, + { + "epoch": 0.6849073256840247, + "grad_norm": 0.8833175487061256, + "learning_rate": 9.961145191902475e-06, + "loss": 0.1858, + "step": 194 + }, + { + "epoch": 0.6884377758164166, + "grad_norm": 0.7703074827448632, + "learning_rate": 9.960429935470146e-06, + "loss": 0.1719, + "step": 195 + }, + { + "epoch": 0.6919682259488085, + "grad_norm": 0.978543268699946, + "learning_rate": 9.959708181716869e-06, + "loss": 0.1843, + "step": 196 + }, + { + "epoch": 0.6954986760812003, + "grad_norm": 1.0182659329814947, + "learning_rate": 9.958979931588018e-06, + "loss": 0.1698, + "step": 197 + }, + { + "epoch": 0.6990291262135923, + "grad_norm": 1.0948049910788071, + "learning_rate": 9.958245186037477e-06, + "loss": 0.1892, + "step": 198 + }, + { + "epoch": 0.7025595763459841, + "grad_norm": 0.6933804652704536, + "learning_rate": 9.957503946027638e-06, + "loss": 0.1672, + "step": 199 + }, + { + "epoch": 0.706090026478376, + "grad_norm": 0.8077103768832906, + "learning_rate": 9.956756212529402e-06, + "loss": 0.1585, + "step": 200 + }, + { + "epoch": 0.7096204766107679, + "grad_norm": 0.822023166461492, + "learning_rate": 9.95600198652217e-06, + "loss": 0.1801, + "step": 201 + }, + { + "epoch": 0.7131509267431597, + "grad_norm": 1.1206342687945703, + "learning_rate": 9.955241268993852e-06, + "loss": 0.201, + "step": 202 + }, + { + "epoch": 0.7166813768755517, + "grad_norm": 0.974499750770202, + "learning_rate": 9.954474060940856e-06, + "loss": 0.2085, + "step": 203 + }, + { + "epoch": 0.7202118270079435, + "grad_norm": 0.6598260944598164, + "learning_rate": 9.953700363368097e-06, + "loss": 0.1859, + "step": 204 + }, + { + "epoch": 0.7237422771403353, + "grad_norm": 0.8633415747390508, + "learning_rate": 9.952920177288985e-06, + "loss": 0.1963, + "step": 205 + }, + { + "epoch": 0.7272727272727273, + "grad_norm": 0.5898944244131702, + "learning_rate": 9.952133503725435e-06, + "loss": 0.2049, + "step": 206 + }, + { + "epoch": 0.7308031774051191, + "grad_norm": 1.2714984543033625, + "learning_rate": 9.951340343707852e-06, + "loss": 0.1837, + "step": 207 + }, + { + "epoch": 0.7343336275375111, + "grad_norm": 0.9097980008759828, + "learning_rate": 9.950540698275144e-06, + "loss": 0.1849, + "step": 208 + }, + { + "epoch": 0.7378640776699029, + "grad_norm": 0.6941779662092539, + "learning_rate": 9.949734568474708e-06, + "loss": 0.176, + "step": 209 + }, + { + "epoch": 0.7413945278022948, + "grad_norm": 1.0789428661342089, + "learning_rate": 9.948921955362438e-06, + "loss": 0.1845, + "step": 210 + }, + { + "epoch": 0.7449249779346867, + "grad_norm": 1.002969780293264, + "learning_rate": 9.94810286000272e-06, + "loss": 0.1932, + "step": 211 + }, + { + "epoch": 0.7484554280670785, + "grad_norm": 0.7346857479212374, + "learning_rate": 9.94727728346843e-06, + "loss": 0.1895, + "step": 212 + }, + { + "epoch": 0.7519858781994704, + "grad_norm": 1.1214956368968594, + "learning_rate": 9.946445226840932e-06, + "loss": 0.1872, + "step": 213 + }, + { + "epoch": 0.7555163283318623, + "grad_norm": 0.8098364912886733, + "learning_rate": 9.945606691210081e-06, + "loss": 0.182, + "step": 214 + }, + { + "epoch": 0.7590467784642542, + "grad_norm": 0.9351430688337309, + "learning_rate": 9.944761677674214e-06, + "loss": 0.194, + "step": 215 + }, + { + "epoch": 0.7625772285966461, + "grad_norm": 1.1657299434860284, + "learning_rate": 9.943910187340154e-06, + "loss": 0.1875, + "step": 216 + }, + { + "epoch": 0.766107678729038, + "grad_norm": 1.0056936253752848, + "learning_rate": 9.943052221323212e-06, + "loss": 0.1993, + "step": 217 + }, + { + "epoch": 0.7696381288614298, + "grad_norm": 1.018936012562021, + "learning_rate": 9.942187780747177e-06, + "loss": 0.205, + "step": 218 + }, + { + "epoch": 0.7731685789938217, + "grad_norm": 0.9704710377674022, + "learning_rate": 9.94131686674432e-06, + "loss": 0.1803, + "step": 219 + }, + { + "epoch": 0.7766990291262136, + "grad_norm": 0.844440847315935, + "learning_rate": 9.940439480455386e-06, + "loss": 0.1903, + "step": 220 + }, + { + "epoch": 0.7802294792586054, + "grad_norm": 0.6911229516194293, + "learning_rate": 9.939555623029608e-06, + "loss": 0.1648, + "step": 221 + }, + { + "epoch": 0.7837599293909974, + "grad_norm": 0.977598516917381, + "learning_rate": 9.938665295624686e-06, + "loss": 0.1775, + "step": 222 + }, + { + "epoch": 0.7872903795233892, + "grad_norm": 0.7894216874582392, + "learning_rate": 9.937768499406798e-06, + "loss": 0.1734, + "step": 223 + }, + { + "epoch": 0.7908208296557812, + "grad_norm": 0.8720889047054564, + "learning_rate": 9.936865235550595e-06, + "loss": 0.1818, + "step": 224 + }, + { + "epoch": 0.794351279788173, + "grad_norm": 0.7210753676381008, + "learning_rate": 9.9359555052392e-06, + "loss": 0.2001, + "step": 225 + }, + { + "epoch": 0.7978817299205648, + "grad_norm": 0.6358344164302707, + "learning_rate": 9.935039309664204e-06, + "loss": 0.1703, + "step": 226 + }, + { + "epoch": 0.8014121800529568, + "grad_norm": 0.8138407178306513, + "learning_rate": 9.93411665002567e-06, + "loss": 0.172, + "step": 227 + }, + { + "epoch": 0.8049426301853486, + "grad_norm": 0.974591399579874, + "learning_rate": 9.93318752753212e-06, + "loss": 0.2067, + "step": 228 + }, + { + "epoch": 0.8084730803177406, + "grad_norm": 0.8238739378175247, + "learning_rate": 9.932251943400554e-06, + "loss": 0.1607, + "step": 229 + }, + { + "epoch": 0.8120035304501324, + "grad_norm": 1.0422244856011833, + "learning_rate": 9.931309898856423e-06, + "loss": 0.1693, + "step": 230 + }, + { + "epoch": 0.8155339805825242, + "grad_norm": 0.812884439844158, + "learning_rate": 9.930361395133647e-06, + "loss": 0.1683, + "step": 231 + }, + { + "epoch": 0.8190644307149162, + "grad_norm": 0.6222138282668747, + "learning_rate": 9.929406433474606e-06, + "loss": 0.1636, + "step": 232 + }, + { + "epoch": 0.822594880847308, + "grad_norm": 0.8762443957945536, + "learning_rate": 9.928445015130136e-06, + "loss": 0.191, + "step": 233 + }, + { + "epoch": 0.8261253309796999, + "grad_norm": 0.9859315198495622, + "learning_rate": 9.927477141359533e-06, + "loss": 0.1895, + "step": 234 + }, + { + "epoch": 0.8296557811120918, + "grad_norm": 0.9589349463946001, + "learning_rate": 9.926502813430545e-06, + "loss": 0.197, + "step": 235 + }, + { + "epoch": 0.8331862312444837, + "grad_norm": 0.8718214863000249, + "learning_rate": 9.92552203261938e-06, + "loss": 0.1799, + "step": 236 + }, + { + "epoch": 0.8367166813768756, + "grad_norm": 1.1335307020440895, + "learning_rate": 9.92453480021069e-06, + "loss": 0.198, + "step": 237 + }, + { + "epoch": 0.8402471315092674, + "grad_norm": 0.8430408249858925, + "learning_rate": 9.923541117497586e-06, + "loss": 0.1478, + "step": 238 + }, + { + "epoch": 0.8437775816416593, + "grad_norm": 0.6769109520694074, + "learning_rate": 9.922540985781621e-06, + "loss": 0.1923, + "step": 239 + }, + { + "epoch": 0.8473080317740512, + "grad_norm": 0.6377271343571659, + "learning_rate": 9.921534406372797e-06, + "loss": 0.1879, + "step": 240 + }, + { + "epoch": 0.8508384819064431, + "grad_norm": 0.7271721148140101, + "learning_rate": 9.920521380589566e-06, + "loss": 0.1938, + "step": 241 + }, + { + "epoch": 0.8543689320388349, + "grad_norm": 0.9160321910693624, + "learning_rate": 9.919501909758815e-06, + "loss": 0.1689, + "step": 242 + }, + { + "epoch": 0.8578993821712269, + "grad_norm": 0.8451022682499405, + "learning_rate": 9.918475995215881e-06, + "loss": 0.1567, + "step": 243 + }, + { + "epoch": 0.8614298323036187, + "grad_norm": 0.7383618763118188, + "learning_rate": 9.917443638304536e-06, + "loss": 0.1998, + "step": 244 + }, + { + "epoch": 0.8649602824360106, + "grad_norm": 0.9977901978127347, + "learning_rate": 9.916404840376993e-06, + "loss": 0.2184, + "step": 245 + }, + { + "epoch": 0.8684907325684025, + "grad_norm": 0.905766448884252, + "learning_rate": 9.915359602793902e-06, + "loss": 0.1822, + "step": 246 + }, + { + "epoch": 0.8720211827007943, + "grad_norm": 0.668257498288602, + "learning_rate": 9.914307926924344e-06, + "loss": 0.1755, + "step": 247 + }, + { + "epoch": 0.8755516328331863, + "grad_norm": 1.2898306231343724, + "learning_rate": 9.913249814145839e-06, + "loss": 0.1943, + "step": 248 + }, + { + "epoch": 0.8790820829655781, + "grad_norm": 0.7801616088781465, + "learning_rate": 9.912185265844333e-06, + "loss": 0.1961, + "step": 249 + }, + { + "epoch": 0.8826125330979699, + "grad_norm": 0.8499032108492505, + "learning_rate": 9.911114283414204e-06, + "loss": 0.2022, + "step": 250 + }, + { + "epoch": 0.8861429832303619, + "grad_norm": 0.9986716197227599, + "learning_rate": 9.910036868258256e-06, + "loss": 0.1935, + "step": 251 + }, + { + "epoch": 0.8896734333627537, + "grad_norm": 0.8164649483937467, + "learning_rate": 9.908953021787725e-06, + "loss": 0.1772, + "step": 252 + }, + { + "epoch": 0.8932038834951457, + "grad_norm": 0.6585245832704895, + "learning_rate": 9.907862745422258e-06, + "loss": 0.1762, + "step": 253 + }, + { + "epoch": 0.8967343336275375, + "grad_norm": 0.7669263548180228, + "learning_rate": 9.906766040589937e-06, + "loss": 0.1673, + "step": 254 + }, + { + "epoch": 0.9002647837599294, + "grad_norm": 0.8311965531630117, + "learning_rate": 9.905662908727258e-06, + "loss": 0.1657, + "step": 255 + }, + { + "epoch": 0.9037952338923213, + "grad_norm": 0.6844670189568429, + "learning_rate": 9.904553351279139e-06, + "loss": 0.1594, + "step": 256 + }, + { + "epoch": 0.9073256840247131, + "grad_norm": 0.8500217928271374, + "learning_rate": 9.903437369698906e-06, + "loss": 0.2076, + "step": 257 + }, + { + "epoch": 0.910856134157105, + "grad_norm": 0.8553603507750409, + "learning_rate": 9.902314965448309e-06, + "loss": 0.1838, + "step": 258 + }, + { + "epoch": 0.9143865842894969, + "grad_norm": 0.8770431320611423, + "learning_rate": 9.901186139997507e-06, + "loss": 0.1935, + "step": 259 + }, + { + "epoch": 0.9179170344218888, + "grad_norm": 0.7281161731708585, + "learning_rate": 9.90005089482507e-06, + "loss": 0.1888, + "step": 260 + }, + { + "epoch": 0.9214474845542807, + "grad_norm": 0.6456235823671523, + "learning_rate": 9.898909231417976e-06, + "loss": 0.1764, + "step": 261 + }, + { + "epoch": 0.9249779346866726, + "grad_norm": 0.6768287107038422, + "learning_rate": 9.89776115127161e-06, + "loss": 0.163, + "step": 262 + }, + { + "epoch": 0.9285083848190644, + "grad_norm": 0.834905164466666, + "learning_rate": 9.89660665588976e-06, + "loss": 0.1842, + "step": 263 + }, + { + "epoch": 0.9320388349514563, + "grad_norm": 0.6460482257383012, + "learning_rate": 9.895445746784623e-06, + "loss": 0.165, + "step": 264 + }, + { + "epoch": 0.9355692850838482, + "grad_norm": 0.9772687205581816, + "learning_rate": 9.89427842547679e-06, + "loss": 0.2003, + "step": 265 + }, + { + "epoch": 0.93909973521624, + "grad_norm": 0.7686896072265816, + "learning_rate": 9.893104693495255e-06, + "loss": 0.1753, + "step": 266 + }, + { + "epoch": 0.942630185348632, + "grad_norm": 0.7667904026391998, + "learning_rate": 9.891924552377405e-06, + "loss": 0.1802, + "step": 267 + }, + { + "epoch": 0.9461606354810238, + "grad_norm": 0.6993230964449527, + "learning_rate": 9.890738003669029e-06, + "loss": 0.1701, + "step": 268 + }, + { + "epoch": 0.9496910856134158, + "grad_norm": 0.7632215639559887, + "learning_rate": 9.8895450489243e-06, + "loss": 0.1607, + "step": 269 + }, + { + "epoch": 0.9532215357458076, + "grad_norm": 0.8307898449698382, + "learning_rate": 9.88834568970579e-06, + "loss": 0.1631, + "step": 270 + }, + { + "epoch": 0.9567519858781994, + "grad_norm": 0.7176869974666727, + "learning_rate": 9.887139927584453e-06, + "loss": 0.1665, + "step": 271 + }, + { + "epoch": 0.9602824360105914, + "grad_norm": 0.924669948462229, + "learning_rate": 9.885927764139633e-06, + "loss": 0.2065, + "step": 272 + }, + { + "epoch": 0.9638128861429832, + "grad_norm": 0.7237412872405012, + "learning_rate": 9.88470920095906e-06, + "loss": 0.1802, + "step": 273 + }, + { + "epoch": 0.9673433362753752, + "grad_norm": 0.6998624857232224, + "learning_rate": 9.883484239638842e-06, + "loss": 0.1642, + "step": 274 + }, + { + "epoch": 0.970873786407767, + "grad_norm": 0.8569858798701728, + "learning_rate": 9.882252881783472e-06, + "loss": 0.1956, + "step": 275 + }, + { + "epoch": 0.9744042365401588, + "grad_norm": 0.7621961932465209, + "learning_rate": 9.881015129005822e-06, + "loss": 0.1841, + "step": 276 + }, + { + "epoch": 0.9779346866725508, + "grad_norm": 0.7592469083950414, + "learning_rate": 9.879770982927135e-06, + "loss": 0.1937, + "step": 277 + }, + { + "epoch": 0.9814651368049426, + "grad_norm": 0.6865266207305956, + "learning_rate": 9.87852044517703e-06, + "loss": 0.1957, + "step": 278 + }, + { + "epoch": 0.9849955869373345, + "grad_norm": 0.7228023197310592, + "learning_rate": 9.877263517393503e-06, + "loss": 0.1786, + "step": 279 + }, + { + "epoch": 0.9885260370697264, + "grad_norm": 0.9465468675591545, + "learning_rate": 9.876000201222912e-06, + "loss": 0.1661, + "step": 280 + }, + { + "epoch": 0.9920564872021183, + "grad_norm": 0.8521304202598817, + "learning_rate": 9.87473049831999e-06, + "loss": 0.154, + "step": 281 + }, + { + "epoch": 0.9955869373345102, + "grad_norm": 0.669879136764297, + "learning_rate": 9.87345441034783e-06, + "loss": 0.1624, + "step": 282 + }, + { + "epoch": 0.999117387466902, + "grad_norm": 0.9684619377618936, + "learning_rate": 9.872171938977895e-06, + "loss": 0.1881, + "step": 283 + }, + { + "epoch": 1.002647837599294, + "grad_norm": 0.85439046023442, + "learning_rate": 9.870883085889997e-06, + "loss": 0.1865, + "step": 284 + }, + { + "epoch": 1.0061782877316858, + "grad_norm": 0.8191806695595788, + "learning_rate": 9.86958785277232e-06, + "loss": 0.163, + "step": 285 + }, + { + "epoch": 1.0097087378640777, + "grad_norm": 0.9382251565209051, + "learning_rate": 9.8682862413214e-06, + "loss": 0.1758, + "step": 286 + }, + { + "epoch": 1.0132391879964695, + "grad_norm": 0.5193721161505495, + "learning_rate": 9.866978253242123e-06, + "loss": 0.1751, + "step": 287 + }, + { + "epoch": 1.0167696381288613, + "grad_norm": 0.8198070617026992, + "learning_rate": 9.865663890247735e-06, + "loss": 0.1535, + "step": 288 + }, + { + "epoch": 1.0203000882612534, + "grad_norm": 0.9683137087655609, + "learning_rate": 9.864343154059825e-06, + "loss": 0.193, + "step": 289 + }, + { + "epoch": 1.0238305383936452, + "grad_norm": 1.0396575123509755, + "learning_rate": 9.863016046408336e-06, + "loss": 0.1584, + "step": 290 + }, + { + "epoch": 1.027360988526037, + "grad_norm": 0.8171505980553876, + "learning_rate": 9.861682569031553e-06, + "loss": 0.1681, + "step": 291 + }, + { + "epoch": 1.030891438658429, + "grad_norm": 0.6927072039886698, + "learning_rate": 9.860342723676105e-06, + "loss": 0.1841, + "step": 292 + }, + { + "epoch": 1.0344218887908208, + "grad_norm": 0.9193349315868659, + "learning_rate": 9.85899651209696e-06, + "loss": 0.1786, + "step": 293 + }, + { + "epoch": 1.0379523389232128, + "grad_norm": 0.6395800368416922, + "learning_rate": 9.857643936057431e-06, + "loss": 0.1714, + "step": 294 + }, + { + "epoch": 1.0414827890556047, + "grad_norm": 0.7076027666270328, + "learning_rate": 9.856284997329158e-06, + "loss": 0.1628, + "step": 295 + }, + { + "epoch": 1.0450132391879965, + "grad_norm": 0.5571080103911318, + "learning_rate": 9.854919697692123e-06, + "loss": 0.1765, + "step": 296 + }, + { + "epoch": 1.0485436893203883, + "grad_norm": 0.8245809567555668, + "learning_rate": 9.853548038934635e-06, + "loss": 0.178, + "step": 297 + }, + { + "epoch": 1.0520741394527802, + "grad_norm": 1.3356935667806615, + "learning_rate": 9.852170022853335e-06, + "loss": 0.1629, + "step": 298 + }, + { + "epoch": 1.055604589585172, + "grad_norm": 0.5648163532404421, + "learning_rate": 9.85078565125319e-06, + "loss": 0.1638, + "step": 299 + }, + { + "epoch": 1.059135039717564, + "grad_norm": 0.8602300175351031, + "learning_rate": 9.849394925947493e-06, + "loss": 0.1801, + "step": 300 + }, + { + "epoch": 1.062665489849956, + "grad_norm": 0.8714458340169661, + "learning_rate": 9.847997848757855e-06, + "loss": 0.1494, + "step": 301 + }, + { + "epoch": 1.0661959399823477, + "grad_norm": 0.7793786634487098, + "learning_rate": 9.846594421514213e-06, + "loss": 0.1627, + "step": 302 + }, + { + "epoch": 1.0697263901147396, + "grad_norm": 0.7080125846941803, + "learning_rate": 9.845184646054817e-06, + "loss": 0.1776, + "step": 303 + }, + { + "epoch": 1.0732568402471314, + "grad_norm": 0.6880805275018903, + "learning_rate": 9.843768524226233e-06, + "loss": 0.1597, + "step": 304 + }, + { + "epoch": 1.0767872903795235, + "grad_norm": 0.7703863699473644, + "learning_rate": 9.842346057883341e-06, + "loss": 0.1603, + "step": 305 + }, + { + "epoch": 1.0803177405119153, + "grad_norm": 0.833743875181872, + "learning_rate": 9.84091724888933e-06, + "loss": 0.1613, + "step": 306 + }, + { + "epoch": 1.0838481906443072, + "grad_norm": 0.7125550254627817, + "learning_rate": 9.839482099115696e-06, + "loss": 0.1511, + "step": 307 + }, + { + "epoch": 1.087378640776699, + "grad_norm": 0.6773769437101177, + "learning_rate": 9.838040610442243e-06, + "loss": 0.1578, + "step": 308 + }, + { + "epoch": 1.0909090909090908, + "grad_norm": 0.7853675801397739, + "learning_rate": 9.836592784757077e-06, + "loss": 0.1548, + "step": 309 + }, + { + "epoch": 1.0944395410414829, + "grad_norm": 0.9629392348111224, + "learning_rate": 9.835138623956603e-06, + "loss": 0.1787, + "step": 310 + }, + { + "epoch": 1.0979699911738747, + "grad_norm": 1.0276282328498774, + "learning_rate": 9.833678129945524e-06, + "loss": 0.1798, + "step": 311 + }, + { + "epoch": 1.1015004413062666, + "grad_norm": 0.9134089726849168, + "learning_rate": 9.83221130463684e-06, + "loss": 0.1888, + "step": 312 + }, + { + "epoch": 1.1050308914386584, + "grad_norm": 0.9917300375319624, + "learning_rate": 9.830738149951843e-06, + "loss": 0.1884, + "step": 313 + }, + { + "epoch": 1.1085613415710502, + "grad_norm": 0.7089950679554146, + "learning_rate": 9.829258667820114e-06, + "loss": 0.1686, + "step": 314 + }, + { + "epoch": 1.1120917917034423, + "grad_norm": 0.578692441690528, + "learning_rate": 9.827772860179528e-06, + "loss": 0.1804, + "step": 315 + }, + { + "epoch": 1.1156222418358341, + "grad_norm": 0.739188688334005, + "learning_rate": 9.826280728976234e-06, + "loss": 0.1699, + "step": 316 + }, + { + "epoch": 1.119152691968226, + "grad_norm": 1.0203778463290891, + "learning_rate": 9.824782276164678e-06, + "loss": 0.1679, + "step": 317 + }, + { + "epoch": 1.1226831421006178, + "grad_norm": 0.9103291412358969, + "learning_rate": 9.823277503707571e-06, + "loss": 0.1885, + "step": 318 + }, + { + "epoch": 1.1262135922330097, + "grad_norm": 1.1240057634668683, + "learning_rate": 9.821766413575915e-06, + "loss": 0.1467, + "step": 319 + }, + { + "epoch": 1.1297440423654015, + "grad_norm": 0.7591033795680634, + "learning_rate": 9.820249007748978e-06, + "loss": 0.1708, + "step": 320 + }, + { + "epoch": 1.1332744924977936, + "grad_norm": 0.6461323484499221, + "learning_rate": 9.818725288214309e-06, + "loss": 0.1576, + "step": 321 + }, + { + "epoch": 1.1368049426301854, + "grad_norm": 0.8717085794666202, + "learning_rate": 9.817195256967715e-06, + "loss": 0.1846, + "step": 322 + }, + { + "epoch": 1.1403353927625772, + "grad_norm": 0.5766027007180119, + "learning_rate": 9.815658916013283e-06, + "loss": 0.1502, + "step": 323 + }, + { + "epoch": 1.143865842894969, + "grad_norm": 0.8409708523916547, + "learning_rate": 9.814116267363355e-06, + "loss": 0.1522, + "step": 324 + }, + { + "epoch": 1.147396293027361, + "grad_norm": 0.9388226084700922, + "learning_rate": 9.812567313038542e-06, + "loss": 0.1681, + "step": 325 + }, + { + "epoch": 1.150926743159753, + "grad_norm": 0.9369552573302047, + "learning_rate": 9.811012055067708e-06, + "loss": 0.1905, + "step": 326 + }, + { + "epoch": 1.1544571932921448, + "grad_norm": 1.0468936130746385, + "learning_rate": 9.809450495487977e-06, + "loss": 0.1838, + "step": 327 + }, + { + "epoch": 1.1579876434245366, + "grad_norm": 0.8466637747576687, + "learning_rate": 9.80788263634473e-06, + "loss": 0.1536, + "step": 328 + }, + { + "epoch": 1.1615180935569285, + "grad_norm": 0.8133260085099403, + "learning_rate": 9.806308479691595e-06, + "loss": 0.185, + "step": 329 + }, + { + "epoch": 1.1650485436893203, + "grad_norm": 0.634815173007273, + "learning_rate": 9.80472802759045e-06, + "loss": 0.1763, + "step": 330 + }, + { + "epoch": 1.1685789938217122, + "grad_norm": 0.7155087062007245, + "learning_rate": 9.803141282111414e-06, + "loss": 0.1689, + "step": 331 + }, + { + "epoch": 1.1721094439541042, + "grad_norm": 0.918184676277347, + "learning_rate": 9.801548245332863e-06, + "loss": 0.1771, + "step": 332 + }, + { + "epoch": 1.175639894086496, + "grad_norm": 0.847842933527483, + "learning_rate": 9.799948919341398e-06, + "loss": 0.1982, + "step": 333 + }, + { + "epoch": 1.179170344218888, + "grad_norm": 0.950371294714106, + "learning_rate": 9.798343306231869e-06, + "loss": 0.1715, + "step": 334 + }, + { + "epoch": 1.1827007943512797, + "grad_norm": 0.8452751079657315, + "learning_rate": 9.796731408107349e-06, + "loss": 0.1898, + "step": 335 + }, + { + "epoch": 1.1862312444836718, + "grad_norm": 1.0132828873057207, + "learning_rate": 9.79511322707916e-06, + "loss": 0.1832, + "step": 336 + }, + { + "epoch": 1.1897616946160636, + "grad_norm": 0.6796613513123204, + "learning_rate": 9.793488765266838e-06, + "loss": 0.1795, + "step": 337 + }, + { + "epoch": 1.1932921447484555, + "grad_norm": 0.7516255039006127, + "learning_rate": 9.791858024798157e-06, + "loss": 0.1682, + "step": 338 + }, + { + "epoch": 1.1968225948808473, + "grad_norm": 0.6882391031969943, + "learning_rate": 9.790221007809106e-06, + "loss": 0.158, + "step": 339 + }, + { + "epoch": 1.2003530450132391, + "grad_norm": 0.7235480772368882, + "learning_rate": 9.788577716443903e-06, + "loss": 0.1685, + "step": 340 + }, + { + "epoch": 1.203883495145631, + "grad_norm": 0.8054680221859308, + "learning_rate": 9.786928152854977e-06, + "loss": 0.1679, + "step": 341 + }, + { + "epoch": 1.207413945278023, + "grad_norm": 0.6537575157321099, + "learning_rate": 9.785272319202978e-06, + "loss": 0.1535, + "step": 342 + }, + { + "epoch": 1.2109443954104149, + "grad_norm": 0.8631226400216473, + "learning_rate": 9.783610217656766e-06, + "loss": 0.1686, + "step": 343 + }, + { + "epoch": 1.2144748455428067, + "grad_norm": 0.7074978790315094, + "learning_rate": 9.781941850393412e-06, + "loss": 0.1557, + "step": 344 + }, + { + "epoch": 1.2180052956751986, + "grad_norm": 0.7508923027783329, + "learning_rate": 9.780267219598196e-06, + "loss": 0.1734, + "step": 345 + }, + { + "epoch": 1.2215357458075904, + "grad_norm": 0.6041756898423745, + "learning_rate": 9.7785863274646e-06, + "loss": 0.1676, + "step": 346 + }, + { + "epoch": 1.2250661959399824, + "grad_norm": 0.8916407715059216, + "learning_rate": 9.776899176194302e-06, + "loss": 0.1612, + "step": 347 + }, + { + "epoch": 1.2285966460723743, + "grad_norm": 0.9080331160273555, + "learning_rate": 9.775205767997188e-06, + "loss": 0.1718, + "step": 348 + }, + { + "epoch": 1.2321270962047661, + "grad_norm": 0.844342085425393, + "learning_rate": 9.773506105091336e-06, + "loss": 0.1673, + "step": 349 + }, + { + "epoch": 1.235657546337158, + "grad_norm": 0.7564502855821623, + "learning_rate": 9.771800189703014e-06, + "loss": 0.1535, + "step": 350 + }, + { + "epoch": 1.2391879964695498, + "grad_norm": 0.6268844715263006, + "learning_rate": 9.770088024066681e-06, + "loss": 0.147, + "step": 351 + }, + { + "epoch": 1.2427184466019416, + "grad_norm": 0.7824396074378344, + "learning_rate": 9.768369610424984e-06, + "loss": 0.1865, + "step": 352 + }, + { + "epoch": 1.2462488967343337, + "grad_norm": 0.7832814173090706, + "learning_rate": 9.766644951028755e-06, + "loss": 0.1792, + "step": 353 + }, + { + "epoch": 1.2497793468667255, + "grad_norm": 0.9379854009237389, + "learning_rate": 9.764914048137001e-06, + "loss": 0.1657, + "step": 354 + }, + { + "epoch": 1.2533097969991174, + "grad_norm": 0.8058449874416329, + "learning_rate": 9.763176904016914e-06, + "loss": 0.1718, + "step": 355 + }, + { + "epoch": 1.2568402471315092, + "grad_norm": 0.8085192046750732, + "learning_rate": 9.761433520943856e-06, + "loss": 0.1776, + "step": 356 + }, + { + "epoch": 1.2603706972639013, + "grad_norm": 0.9591752385660486, + "learning_rate": 9.759683901201364e-06, + "loss": 0.1979, + "step": 357 + }, + { + "epoch": 1.2639011473962931, + "grad_norm": 1.0461750891605788, + "learning_rate": 9.757928047081144e-06, + "loss": 0.1926, + "step": 358 + }, + { + "epoch": 1.267431597528685, + "grad_norm": 0.7954582531988508, + "learning_rate": 9.756165960883063e-06, + "loss": 0.1648, + "step": 359 + }, + { + "epoch": 1.2709620476610768, + "grad_norm": 0.8863013446887811, + "learning_rate": 9.75439764491516e-06, + "loss": 0.1807, + "step": 360 + }, + { + "epoch": 1.2744924977934686, + "grad_norm": 0.8321220073097337, + "learning_rate": 9.752623101493624e-06, + "loss": 0.1771, + "step": 361 + }, + { + "epoch": 1.2780229479258605, + "grad_norm": 0.8204672244613037, + "learning_rate": 9.750842332942805e-06, + "loss": 0.1669, + "step": 362 + }, + { + "epoch": 1.2815533980582523, + "grad_norm": 0.8746356076847612, + "learning_rate": 9.749055341595213e-06, + "loss": 0.1511, + "step": 363 + }, + { + "epoch": 1.2850838481906444, + "grad_norm": 0.7387304926466468, + "learning_rate": 9.747262129791497e-06, + "loss": 0.1696, + "step": 364 + }, + { + "epoch": 1.2886142983230362, + "grad_norm": 0.6434577859365603, + "learning_rate": 9.745462699880464e-06, + "loss": 0.1617, + "step": 365 + }, + { + "epoch": 1.292144748455428, + "grad_norm": 0.8994503573558672, + "learning_rate": 9.74365705421906e-06, + "loss": 0.1771, + "step": 366 + }, + { + "epoch": 1.2956751985878199, + "grad_norm": 0.9104255097758082, + "learning_rate": 9.741845195172376e-06, + "loss": 0.1507, + "step": 367 + }, + { + "epoch": 1.299205648720212, + "grad_norm": 0.8237793644377707, + "learning_rate": 9.740027125113638e-06, + "loss": 0.1619, + "step": 368 + }, + { + "epoch": 1.3027360988526038, + "grad_norm": 0.7759651867155161, + "learning_rate": 9.73820284642421e-06, + "loss": 0.1671, + "step": 369 + }, + { + "epoch": 1.3062665489849956, + "grad_norm": 0.8529971231768904, + "learning_rate": 9.736372361493584e-06, + "loss": 0.1601, + "step": 370 + }, + { + "epoch": 1.3097969991173875, + "grad_norm": 0.6232528534187706, + "learning_rate": 9.734535672719391e-06, + "loss": 0.1662, + "step": 371 + }, + { + "epoch": 1.3133274492497793, + "grad_norm": 0.9946599681832587, + "learning_rate": 9.732692782507376e-06, + "loss": 0.1687, + "step": 372 + }, + { + "epoch": 1.3168578993821711, + "grad_norm": 0.7623188795800929, + "learning_rate": 9.730843693271413e-06, + "loss": 0.1683, + "step": 373 + }, + { + "epoch": 1.3203883495145632, + "grad_norm": 0.5696437347911626, + "learning_rate": 9.728988407433497e-06, + "loss": 0.1555, + "step": 374 + }, + { + "epoch": 1.323918799646955, + "grad_norm": 0.7695474373449546, + "learning_rate": 9.727126927423736e-06, + "loss": 0.1891, + "step": 375 + }, + { + "epoch": 1.3274492497793469, + "grad_norm": 0.7640349287167723, + "learning_rate": 9.725259255680352e-06, + "loss": 0.1822, + "step": 376 + }, + { + "epoch": 1.3309796999117387, + "grad_norm": 0.7909928993460938, + "learning_rate": 9.72338539464968e-06, + "loss": 0.1742, + "step": 377 + }, + { + "epoch": 1.3345101500441308, + "grad_norm": 1.1113868103845201, + "learning_rate": 9.72150534678616e-06, + "loss": 0.1647, + "step": 378 + }, + { + "epoch": 1.3380406001765226, + "grad_norm": 1.0428448743119225, + "learning_rate": 9.719619114552332e-06, + "loss": 0.1354, + "step": 379 + }, + { + "epoch": 1.3415710503089144, + "grad_norm": 0.8715636876334808, + "learning_rate": 9.717726700418842e-06, + "loss": 0.1772, + "step": 380 + }, + { + "epoch": 1.3451015004413063, + "grad_norm": 0.878842141186717, + "learning_rate": 9.715828106864434e-06, + "loss": 0.1657, + "step": 381 + }, + { + "epoch": 1.3486319505736981, + "grad_norm": 0.6697201435871655, + "learning_rate": 9.713923336375936e-06, + "loss": 0.1585, + "step": 382 + }, + { + "epoch": 1.35216240070609, + "grad_norm": 0.9696617071339634, + "learning_rate": 9.712012391448283e-06, + "loss": 0.1939, + "step": 383 + }, + { + "epoch": 1.3556928508384818, + "grad_norm": 0.8187281006125846, + "learning_rate": 9.710095274584482e-06, + "loss": 0.1653, + "step": 384 + }, + { + "epoch": 1.3592233009708738, + "grad_norm": 0.8974963672760499, + "learning_rate": 9.70817198829563e-06, + "loss": 0.1737, + "step": 385 + }, + { + "epoch": 1.3627537511032657, + "grad_norm": 0.7456778356448868, + "learning_rate": 9.70624253510091e-06, + "loss": 0.1635, + "step": 386 + }, + { + "epoch": 1.3662842012356575, + "grad_norm": 1.1629774820732057, + "learning_rate": 9.704306917527575e-06, + "loss": 0.2045, + "step": 387 + }, + { + "epoch": 1.3698146513680494, + "grad_norm": 0.8069603325247644, + "learning_rate": 9.702365138110958e-06, + "loss": 0.1603, + "step": 388 + }, + { + "epoch": 1.3733451015004414, + "grad_norm": 0.9350513325401503, + "learning_rate": 9.700417199394457e-06, + "loss": 0.1764, + "step": 389 + }, + { + "epoch": 1.3768755516328333, + "grad_norm": 0.5965930357229629, + "learning_rate": 9.698463103929542e-06, + "loss": 0.1688, + "step": 390 + }, + { + "epoch": 1.380406001765225, + "grad_norm": 0.9881802502016954, + "learning_rate": 9.69650285427575e-06, + "loss": 0.1996, + "step": 391 + }, + { + "epoch": 1.383936451897617, + "grad_norm": 0.6943128810137366, + "learning_rate": 9.694536453000668e-06, + "loss": 0.1685, + "step": 392 + }, + { + "epoch": 1.3874669020300088, + "grad_norm": 0.6982977323410529, + "learning_rate": 9.692563902679954e-06, + "loss": 0.1563, + "step": 393 + }, + { + "epoch": 1.3909973521624006, + "grad_norm": 0.7295520907046743, + "learning_rate": 9.69058520589731e-06, + "loss": 0.1692, + "step": 394 + }, + { + "epoch": 1.3945278022947925, + "grad_norm": 0.80781982226485, + "learning_rate": 9.688600365244495e-06, + "loss": 0.1762, + "step": 395 + }, + { + "epoch": 1.3980582524271845, + "grad_norm": 0.6182663519447118, + "learning_rate": 9.686609383321316e-06, + "loss": 0.1686, + "step": 396 + }, + { + "epoch": 1.4015887025595763, + "grad_norm": 0.5145110224464513, + "learning_rate": 9.684612262735614e-06, + "loss": 0.1626, + "step": 397 + }, + { + "epoch": 1.4051191526919682, + "grad_norm": 0.8760399209586848, + "learning_rate": 9.682609006103283e-06, + "loss": 0.1754, + "step": 398 + }, + { + "epoch": 1.40864960282436, + "grad_norm": 0.8161048235153104, + "learning_rate": 9.680599616048247e-06, + "loss": 0.164, + "step": 399 + }, + { + "epoch": 1.412180052956752, + "grad_norm": 0.7009270806175059, + "learning_rate": 9.678584095202468e-06, + "loss": 0.1613, + "step": 400 + }, + { + "epoch": 1.415710503089144, + "grad_norm": 0.6567037684930597, + "learning_rate": 9.676562446205936e-06, + "loss": 0.1498, + "step": 401 + }, + { + "epoch": 1.4192409532215358, + "grad_norm": 0.7885493080145554, + "learning_rate": 9.674534671706666e-06, + "loss": 0.1834, + "step": 402 + }, + { + "epoch": 1.4227714033539276, + "grad_norm": 0.561383249284057, + "learning_rate": 9.672500774360698e-06, + "loss": 0.1664, + "step": 403 + }, + { + "epoch": 1.4263018534863194, + "grad_norm": 0.8130649369686304, + "learning_rate": 9.670460756832094e-06, + "loss": 0.1375, + "step": 404 + }, + { + "epoch": 1.4298323036187113, + "grad_norm": 0.730228217232664, + "learning_rate": 9.668414621792928e-06, + "loss": 0.1602, + "step": 405 + }, + { + "epoch": 1.4333627537511033, + "grad_norm": 0.701767347722211, + "learning_rate": 9.666362371923294e-06, + "loss": 0.1619, + "step": 406 + }, + { + "epoch": 1.4368932038834952, + "grad_norm": 0.6142164548390184, + "learning_rate": 9.664304009911286e-06, + "loss": 0.1578, + "step": 407 + }, + { + "epoch": 1.440423654015887, + "grad_norm": 1.1635404922505252, + "learning_rate": 9.66223953845301e-06, + "loss": 0.1835, + "step": 408 + }, + { + "epoch": 1.4439541041482788, + "grad_norm": 0.8373651861820354, + "learning_rate": 9.660168960252575e-06, + "loss": 0.17, + "step": 409 + }, + { + "epoch": 1.447484554280671, + "grad_norm": 0.5626164848515763, + "learning_rate": 9.658092278022088e-06, + "loss": 0.1638, + "step": 410 + }, + { + "epoch": 1.4510150044130627, + "grad_norm": 0.9647932895601038, + "learning_rate": 9.656009494481646e-06, + "loss": 0.1737, + "step": 411 + }, + { + "epoch": 1.4545454545454546, + "grad_norm": 0.6769901378936639, + "learning_rate": 9.653920612359344e-06, + "loss": 0.1719, + "step": 412 + }, + { + "epoch": 1.4580759046778464, + "grad_norm": 0.5912616414391476, + "learning_rate": 9.651825634391262e-06, + "loss": 0.1547, + "step": 413 + }, + { + "epoch": 1.4616063548102383, + "grad_norm": 0.6130890022296159, + "learning_rate": 9.649724563321465e-06, + "loss": 0.1509, + "step": 414 + }, + { + "epoch": 1.46513680494263, + "grad_norm": 0.6981352025940074, + "learning_rate": 9.647617401902003e-06, + "loss": 0.1612, + "step": 415 + }, + { + "epoch": 1.468667255075022, + "grad_norm": 0.7832373213646434, + "learning_rate": 9.645504152892895e-06, + "loss": 0.1867, + "step": 416 + }, + { + "epoch": 1.472197705207414, + "grad_norm": 0.8544155478419707, + "learning_rate": 9.64338481906214e-06, + "loss": 0.1721, + "step": 417 + }, + { + "epoch": 1.4757281553398058, + "grad_norm": 0.9502151610480191, + "learning_rate": 9.641259403185706e-06, + "loss": 0.1743, + "step": 418 + }, + { + "epoch": 1.4792586054721977, + "grad_norm": 0.7733505393875248, + "learning_rate": 9.639127908047527e-06, + "loss": 0.1734, + "step": 419 + }, + { + "epoch": 1.4827890556045895, + "grad_norm": 0.8264946927060004, + "learning_rate": 9.6369903364395e-06, + "loss": 0.1911, + "step": 420 + }, + { + "epoch": 1.4863195057369816, + "grad_norm": 0.8237623458796101, + "learning_rate": 9.63484669116148e-06, + "loss": 0.169, + "step": 421 + }, + { + "epoch": 1.4898499558693734, + "grad_norm": 1.0384652928610334, + "learning_rate": 9.632696975021278e-06, + "loss": 0.1761, + "step": 422 + }, + { + "epoch": 1.4933804060017652, + "grad_norm": 0.9824278139544069, + "learning_rate": 9.630541190834659e-06, + "loss": 0.1854, + "step": 423 + }, + { + "epoch": 1.496910856134157, + "grad_norm": 0.7655842553137074, + "learning_rate": 9.628379341425332e-06, + "loss": 0.1762, + "step": 424 + }, + { + "epoch": 1.500441306266549, + "grad_norm": 0.7573467640480261, + "learning_rate": 9.626211429624955e-06, + "loss": 0.1719, + "step": 425 + }, + { + "epoch": 1.5039717563989408, + "grad_norm": 0.6342899695339446, + "learning_rate": 9.624037458273121e-06, + "loss": 0.1663, + "step": 426 + }, + { + "epoch": 1.5075022065313326, + "grad_norm": 0.7241886079251669, + "learning_rate": 9.621857430217366e-06, + "loss": 0.1612, + "step": 427 + }, + { + "epoch": 1.5110326566637247, + "grad_norm": 0.7713579132775925, + "learning_rate": 9.619671348313156e-06, + "loss": 0.1414, + "step": 428 + }, + { + "epoch": 1.5145631067961165, + "grad_norm": 0.6070241168065806, + "learning_rate": 9.617479215423887e-06, + "loss": 0.1752, + "step": 429 + }, + { + "epoch": 1.5180935569285083, + "grad_norm": 0.8462668684183862, + "learning_rate": 9.615281034420882e-06, + "loss": 0.1769, + "step": 430 + }, + { + "epoch": 1.5216240070609004, + "grad_norm": 0.7218104922456425, + "learning_rate": 9.613076808183382e-06, + "loss": 0.1531, + "step": 431 + }, + { + "epoch": 1.5251544571932922, + "grad_norm": 0.7261686662514826, + "learning_rate": 9.610866539598551e-06, + "loss": 0.1638, + "step": 432 + }, + { + "epoch": 1.528684907325684, + "grad_norm": 1.2087576640937627, + "learning_rate": 9.608650231561466e-06, + "loss": 0.1848, + "step": 433 + }, + { + "epoch": 1.532215357458076, + "grad_norm": 0.6381008529495783, + "learning_rate": 9.606427886975113e-06, + "loss": 0.1719, + "step": 434 + }, + { + "epoch": 1.5357458075904677, + "grad_norm": 0.9801974645943355, + "learning_rate": 9.604199508750386e-06, + "loss": 0.189, + "step": 435 + }, + { + "epoch": 1.5392762577228596, + "grad_norm": 0.7249244172345439, + "learning_rate": 9.601965099806085e-06, + "loss": 0.1729, + "step": 436 + }, + { + "epoch": 1.5428067078552514, + "grad_norm": 0.852227223208167, + "learning_rate": 9.599724663068904e-06, + "loss": 0.1522, + "step": 437 + }, + { + "epoch": 1.5463371579876433, + "grad_norm": 0.70519515152008, + "learning_rate": 9.597478201473433e-06, + "loss": 0.1682, + "step": 438 + }, + { + "epoch": 1.5498676081200353, + "grad_norm": 0.7777614025947788, + "learning_rate": 9.595225717962157e-06, + "loss": 0.16, + "step": 439 + }, + { + "epoch": 1.5533980582524272, + "grad_norm": 0.6144264510883923, + "learning_rate": 9.592967215485449e-06, + "loss": 0.1737, + "step": 440 + }, + { + "epoch": 1.5569285083848192, + "grad_norm": 0.6723553147570778, + "learning_rate": 9.590702697001559e-06, + "loss": 0.1864, + "step": 441 + }, + { + "epoch": 1.560458958517211, + "grad_norm": 0.786872255669317, + "learning_rate": 9.588432165476626e-06, + "loss": 0.1678, + "step": 442 + }, + { + "epoch": 1.563989408649603, + "grad_norm": 1.060953562194783, + "learning_rate": 9.58615562388466e-06, + "loss": 0.1743, + "step": 443 + }, + { + "epoch": 1.5675198587819947, + "grad_norm": 0.8590680802220704, + "learning_rate": 9.583873075207541e-06, + "loss": 0.1449, + "step": 444 + }, + { + "epoch": 1.5710503089143866, + "grad_norm": 0.8592072746191091, + "learning_rate": 9.581584522435025e-06, + "loss": 0.1689, + "step": 445 + }, + { + "epoch": 1.5745807590467784, + "grad_norm": 0.6431750629028619, + "learning_rate": 9.579289968564723e-06, + "loss": 0.1549, + "step": 446 + }, + { + "epoch": 1.5781112091791702, + "grad_norm": 0.7880523755608316, + "learning_rate": 9.576989416602113e-06, + "loss": 0.1569, + "step": 447 + }, + { + "epoch": 1.581641659311562, + "grad_norm": 0.7826920782979999, + "learning_rate": 9.574682869560528e-06, + "loss": 0.1655, + "step": 448 + }, + { + "epoch": 1.5851721094439541, + "grad_norm": 0.6507531502926956, + "learning_rate": 9.57237033046115e-06, + "loss": 0.1704, + "step": 449 + }, + { + "epoch": 1.588702559576346, + "grad_norm": 0.6131816635944837, + "learning_rate": 9.570051802333017e-06, + "loss": 0.1722, + "step": 450 + }, + { + "epoch": 1.5922330097087378, + "grad_norm": 0.8615528127088715, + "learning_rate": 9.567727288213005e-06, + "loss": 0.1671, + "step": 451 + }, + { + "epoch": 1.5957634598411299, + "grad_norm": 0.7175178771131497, + "learning_rate": 9.565396791145833e-06, + "loss": 0.1646, + "step": 452 + }, + { + "epoch": 1.5992939099735217, + "grad_norm": 0.6985637923301037, + "learning_rate": 9.563060314184055e-06, + "loss": 0.1919, + "step": 453 + }, + { + "epoch": 1.6028243601059136, + "grad_norm": 0.854067298117589, + "learning_rate": 9.560717860388061e-06, + "loss": 0.1819, + "step": 454 + }, + { + "epoch": 1.6063548102383054, + "grad_norm": 0.884963758284598, + "learning_rate": 9.558369432826067e-06, + "loss": 0.169, + "step": 455 + }, + { + "epoch": 1.6098852603706972, + "grad_norm": 0.7696374374267597, + "learning_rate": 9.556015034574114e-06, + "loss": 0.1538, + "step": 456 + }, + { + "epoch": 1.613415710503089, + "grad_norm": 0.6854959076334332, + "learning_rate": 9.553654668716065e-06, + "loss": 0.1743, + "step": 457 + }, + { + "epoch": 1.616946160635481, + "grad_norm": 0.7796697126004593, + "learning_rate": 9.551288338343597e-06, + "loss": 0.1841, + "step": 458 + }, + { + "epoch": 1.6204766107678727, + "grad_norm": 0.8659991614483475, + "learning_rate": 9.548916046556202e-06, + "loss": 0.1624, + "step": 459 + }, + { + "epoch": 1.6240070609002648, + "grad_norm": 0.9487399709116949, + "learning_rate": 9.54653779646118e-06, + "loss": 0.1855, + "step": 460 + }, + { + "epoch": 1.6275375110326566, + "grad_norm": 0.6655499212908461, + "learning_rate": 9.544153591173632e-06, + "loss": 0.1582, + "step": 461 + }, + { + "epoch": 1.6310679611650487, + "grad_norm": 0.6850321878653797, + "learning_rate": 9.541763433816467e-06, + "loss": 0.161, + "step": 462 + }, + { + "epoch": 1.6345984112974405, + "grad_norm": 0.6344259176675145, + "learning_rate": 9.539367327520382e-06, + "loss": 0.1899, + "step": 463 + }, + { + "epoch": 1.6381288614298324, + "grad_norm": 0.8057880340454765, + "learning_rate": 9.536965275423869e-06, + "loss": 0.1849, + "step": 464 + }, + { + "epoch": 1.6416593115622242, + "grad_norm": 0.6524278173237419, + "learning_rate": 9.53455728067321e-06, + "loss": 0.1616, + "step": 465 + }, + { + "epoch": 1.645189761694616, + "grad_norm": 0.721136853352441, + "learning_rate": 9.532143346422466e-06, + "loss": 0.1561, + "step": 466 + }, + { + "epoch": 1.648720211827008, + "grad_norm": 0.9490769657599735, + "learning_rate": 9.529723475833485e-06, + "loss": 0.1754, + "step": 467 + }, + { + "epoch": 1.6522506619593997, + "grad_norm": 0.5644282783067912, + "learning_rate": 9.527297672075886e-06, + "loss": 0.1627, + "step": 468 + }, + { + "epoch": 1.6557811120917916, + "grad_norm": 0.7598201706804275, + "learning_rate": 9.524865938327058e-06, + "loss": 0.1716, + "step": 469 + }, + { + "epoch": 1.6593115622241836, + "grad_norm": 0.7591422889029776, + "learning_rate": 9.522428277772162e-06, + "loss": 0.1451, + "step": 470 + }, + { + "epoch": 1.6628420123565755, + "grad_norm": 0.6982093938144709, + "learning_rate": 9.519984693604116e-06, + "loss": 0.1645, + "step": 471 + }, + { + "epoch": 1.6663724624889673, + "grad_norm": 0.5228701865370198, + "learning_rate": 9.517535189023602e-06, + "loss": 0.1563, + "step": 472 + }, + { + "epoch": 1.6699029126213594, + "grad_norm": 0.8204330233451095, + "learning_rate": 9.515079767239056e-06, + "loss": 0.1678, + "step": 473 + }, + { + "epoch": 1.6734333627537512, + "grad_norm": 0.9523128967867128, + "learning_rate": 9.512618431466662e-06, + "loss": 0.1664, + "step": 474 + }, + { + "epoch": 1.676963812886143, + "grad_norm": 0.7078287409806696, + "learning_rate": 9.510151184930354e-06, + "loss": 0.1658, + "step": 475 + }, + { + "epoch": 1.6804942630185349, + "grad_norm": 0.7520543206855101, + "learning_rate": 9.507678030861805e-06, + "loss": 0.1646, + "step": 476 + }, + { + "epoch": 1.6840247131509267, + "grad_norm": 0.9284654185473334, + "learning_rate": 9.505198972500426e-06, + "loss": 0.1684, + "step": 477 + }, + { + "epoch": 1.6875551632833186, + "grad_norm": 0.6224185496516447, + "learning_rate": 9.502714013093364e-06, + "loss": 0.1475, + "step": 478 + }, + { + "epoch": 1.6910856134157104, + "grad_norm": 0.7200200167776666, + "learning_rate": 9.500223155895491e-06, + "loss": 0.1739, + "step": 479 + }, + { + "epoch": 1.6946160635481022, + "grad_norm": 0.7893065612440167, + "learning_rate": 9.497726404169411e-06, + "loss": 0.1577, + "step": 480 + }, + { + "epoch": 1.6981465136804943, + "grad_norm": 0.806998234582965, + "learning_rate": 9.495223761185443e-06, + "loss": 0.1692, + "step": 481 + }, + { + "epoch": 1.7016769638128861, + "grad_norm": 0.8896910612567586, + "learning_rate": 9.492715230221623e-06, + "loss": 0.155, + "step": 482 + }, + { + "epoch": 1.705207413945278, + "grad_norm": 0.7929727685131104, + "learning_rate": 9.490200814563704e-06, + "loss": 0.1693, + "step": 483 + }, + { + "epoch": 1.70873786407767, + "grad_norm": 0.8281002779014005, + "learning_rate": 9.48768051750514e-06, + "loss": 0.162, + "step": 484 + }, + { + "epoch": 1.7122683142100619, + "grad_norm": 0.7574881698882824, + "learning_rate": 9.485154342347093e-06, + "loss": 0.1822, + "step": 485 + }, + { + "epoch": 1.7157987643424537, + "grad_norm": 0.7603690459222415, + "learning_rate": 9.482622292398424e-06, + "loss": 0.1705, + "step": 486 + }, + { + "epoch": 1.7193292144748455, + "grad_norm": 0.6922874253842599, + "learning_rate": 9.480084370975691e-06, + "loss": 0.1687, + "step": 487 + }, + { + "epoch": 1.7228596646072374, + "grad_norm": 0.80394336074322, + "learning_rate": 9.477540581403137e-06, + "loss": 0.1688, + "step": 488 + }, + { + "epoch": 1.7263901147396292, + "grad_norm": 0.8802309764956326, + "learning_rate": 9.474990927012694e-06, + "loss": 0.1304, + "step": 489 + }, + { + "epoch": 1.729920564872021, + "grad_norm": 0.6137981646644614, + "learning_rate": 9.472435411143979e-06, + "loss": 0.1664, + "step": 490 + }, + { + "epoch": 1.733451015004413, + "grad_norm": 0.9282185888184011, + "learning_rate": 9.469874037144281e-06, + "loss": 0.2012, + "step": 491 + }, + { + "epoch": 1.736981465136805, + "grad_norm": 0.6630691029948931, + "learning_rate": 9.467306808368568e-06, + "loss": 0.1502, + "step": 492 + }, + { + "epoch": 1.7405119152691968, + "grad_norm": 0.7577885818393465, + "learning_rate": 9.46473372817947e-06, + "loss": 0.1702, + "step": 493 + }, + { + "epoch": 1.7440423654015889, + "grad_norm": 0.5899998168235006, + "learning_rate": 9.46215479994729e-06, + "loss": 0.1602, + "step": 494 + }, + { + "epoch": 1.7475728155339807, + "grad_norm": 1.0306863382019156, + "learning_rate": 9.459570027049983e-06, + "loss": 0.1334, + "step": 495 + }, + { + "epoch": 1.7511032656663725, + "grad_norm": 0.865412362993785, + "learning_rate": 9.456979412873163e-06, + "loss": 0.1852, + "step": 496 + }, + { + "epoch": 1.7546337157987644, + "grad_norm": 0.8407963911360593, + "learning_rate": 9.454382960810093e-06, + "loss": 0.1691, + "step": 497 + }, + { + "epoch": 1.7581641659311562, + "grad_norm": 0.6988074166274713, + "learning_rate": 9.451780674261687e-06, + "loss": 0.16, + "step": 498 + }, + { + "epoch": 1.761694616063548, + "grad_norm": 0.6567772580817499, + "learning_rate": 9.4491725566365e-06, + "loss": 0.1678, + "step": 499 + }, + { + "epoch": 1.7652250661959399, + "grad_norm": 0.8826194122518776, + "learning_rate": 9.446558611350718e-06, + "loss": 0.177, + "step": 500 + }, + { + "epoch": 1.7687555163283317, + "grad_norm": 0.6900513296683881, + "learning_rate": 9.44393884182817e-06, + "loss": 0.1852, + "step": 501 + }, + { + "epoch": 1.7722859664607238, + "grad_norm": 0.6605181197941491, + "learning_rate": 9.441313251500307e-06, + "loss": 0.1557, + "step": 502 + }, + { + "epoch": 1.7758164165931156, + "grad_norm": 0.6146198835230601, + "learning_rate": 9.438681843806207e-06, + "loss": 0.1594, + "step": 503 + }, + { + "epoch": 1.7793468667255075, + "grad_norm": 0.5149059845127192, + "learning_rate": 9.436044622192569e-06, + "loss": 0.1665, + "step": 504 + }, + { + "epoch": 1.7828773168578995, + "grad_norm": 0.9135519362694595, + "learning_rate": 9.433401590113702e-06, + "loss": 0.1797, + "step": 505 + }, + { + "epoch": 1.7864077669902914, + "grad_norm": 0.8916844443544416, + "learning_rate": 9.43075275103153e-06, + "loss": 0.17, + "step": 506 + }, + { + "epoch": 1.7899382171226832, + "grad_norm": 0.8009121594911396, + "learning_rate": 9.428098108415587e-06, + "loss": 0.1599, + "step": 507 + }, + { + "epoch": 1.793468667255075, + "grad_norm": 0.8269496601871587, + "learning_rate": 9.425437665742998e-06, + "loss": 0.1633, + "step": 508 + }, + { + "epoch": 1.7969991173874669, + "grad_norm": 0.6798890374153103, + "learning_rate": 9.422771426498493e-06, + "loss": 0.1605, + "step": 509 + }, + { + "epoch": 1.8005295675198587, + "grad_norm": 0.6315062922758498, + "learning_rate": 9.420099394174397e-06, + "loss": 0.1795, + "step": 510 + }, + { + "epoch": 1.8040600176522505, + "grad_norm": 0.5746772810241558, + "learning_rate": 9.417421572270612e-06, + "loss": 0.1567, + "step": 511 + }, + { + "epoch": 1.8075904677846424, + "grad_norm": 0.8050650153052257, + "learning_rate": 9.414737964294636e-06, + "loss": 0.1729, + "step": 512 + }, + { + "epoch": 1.8111209179170344, + "grad_norm": 0.480452501514603, + "learning_rate": 9.412048573761535e-06, + "loss": 0.1766, + "step": 513 + }, + { + "epoch": 1.8146513680494263, + "grad_norm": 0.9104720825841831, + "learning_rate": 9.409353404193958e-06, + "loss": 0.1745, + "step": 514 + }, + { + "epoch": 1.8181818181818183, + "grad_norm": 0.8780636579991754, + "learning_rate": 9.406652459122115e-06, + "loss": 0.1762, + "step": 515 + }, + { + "epoch": 1.8217122683142102, + "grad_norm": 0.6984924204172243, + "learning_rate": 9.403945742083788e-06, + "loss": 0.1793, + "step": 516 + }, + { + "epoch": 1.825242718446602, + "grad_norm": 0.8238701991708997, + "learning_rate": 9.401233256624318e-06, + "loss": 0.1485, + "step": 517 + }, + { + "epoch": 1.8287731685789939, + "grad_norm": 0.8696582194623638, + "learning_rate": 9.398515006296598e-06, + "loss": 0.1891, + "step": 518 + }, + { + "epoch": 1.8323036187113857, + "grad_norm": 0.699771115532138, + "learning_rate": 9.395790994661075e-06, + "loss": 0.1553, + "step": 519 + }, + { + "epoch": 1.8358340688437775, + "grad_norm": 0.9125794295579036, + "learning_rate": 9.393061225285743e-06, + "loss": 0.163, + "step": 520 + }, + { + "epoch": 1.8393645189761694, + "grad_norm": 0.8950419906565266, + "learning_rate": 9.390325701746135e-06, + "loss": 0.167, + "step": 521 + }, + { + "epoch": 1.8428949691085612, + "grad_norm": 0.8710778282196566, + "learning_rate": 9.387584427625323e-06, + "loss": 0.1585, + "step": 522 + }, + { + "epoch": 1.8464254192409533, + "grad_norm": 0.8901232516687565, + "learning_rate": 9.38483740651391e-06, + "loss": 0.171, + "step": 523 + }, + { + "epoch": 1.849955869373345, + "grad_norm": 0.7244577661750393, + "learning_rate": 9.382084642010026e-06, + "loss": 0.1718, + "step": 524 + }, + { + "epoch": 1.853486319505737, + "grad_norm": 0.9044306478777916, + "learning_rate": 9.379326137719329e-06, + "loss": 0.1762, + "step": 525 + }, + { + "epoch": 1.857016769638129, + "grad_norm": 0.6730866435868901, + "learning_rate": 9.376561897254987e-06, + "loss": 0.1567, + "step": 526 + }, + { + "epoch": 1.8605472197705208, + "grad_norm": 0.816989704732525, + "learning_rate": 9.37379192423769e-06, + "loss": 0.1919, + "step": 527 + }, + { + "epoch": 1.8640776699029127, + "grad_norm": 0.7655388484499017, + "learning_rate": 9.371016222295628e-06, + "loss": 0.1736, + "step": 528 + }, + { + "epoch": 1.8676081200353045, + "grad_norm": 0.8699212931081676, + "learning_rate": 9.368234795064502e-06, + "loss": 0.1625, + "step": 529 + }, + { + "epoch": 1.8711385701676964, + "grad_norm": 1.0534597898924092, + "learning_rate": 9.365447646187509e-06, + "loss": 0.1381, + "step": 530 + }, + { + "epoch": 1.8746690203000882, + "grad_norm": 0.6614479535387474, + "learning_rate": 9.36265477931534e-06, + "loss": 0.1467, + "step": 531 + }, + { + "epoch": 1.87819947043248, + "grad_norm": 0.8393186062366995, + "learning_rate": 9.359856198106176e-06, + "loss": 0.1913, + "step": 532 + }, + { + "epoch": 1.8817299205648719, + "grad_norm": 0.5866280093193419, + "learning_rate": 9.357051906225685e-06, + "loss": 0.1566, + "step": 533 + }, + { + "epoch": 1.885260370697264, + "grad_norm": 0.9996266820247056, + "learning_rate": 9.35424190734701e-06, + "loss": 0.1919, + "step": 534 + }, + { + "epoch": 1.8887908208296558, + "grad_norm": 0.8127958941754003, + "learning_rate": 9.351426205150778e-06, + "loss": 0.1695, + "step": 535 + }, + { + "epoch": 1.8923212709620476, + "grad_norm": 0.6634743250735253, + "learning_rate": 9.348604803325075e-06, + "loss": 0.1555, + "step": 536 + }, + { + "epoch": 1.8958517210944397, + "grad_norm": 0.812340369225641, + "learning_rate": 9.34577770556546e-06, + "loss": 0.1645, + "step": 537 + }, + { + "epoch": 1.8993821712268315, + "grad_norm": 0.7497218737066156, + "learning_rate": 9.342944915574952e-06, + "loss": 0.1668, + "step": 538 + }, + { + "epoch": 1.9029126213592233, + "grad_norm": 0.5695613946823557, + "learning_rate": 9.340106437064026e-06, + "loss": 0.1625, + "step": 539 + }, + { + "epoch": 1.9064430714916152, + "grad_norm": 0.8336306538843815, + "learning_rate": 9.337262273750604e-06, + "loss": 0.1679, + "step": 540 + }, + { + "epoch": 1.909973521624007, + "grad_norm": 0.7985645323717766, + "learning_rate": 9.334412429360058e-06, + "loss": 0.1717, + "step": 541 + }, + { + "epoch": 1.9135039717563989, + "grad_norm": 0.803743435053368, + "learning_rate": 9.3315569076252e-06, + "loss": 0.1811, + "step": 542 + }, + { + "epoch": 1.9170344218887907, + "grad_norm": 0.7792918327227041, + "learning_rate": 9.328695712286283e-06, + "loss": 0.178, + "step": 543 + }, + { + "epoch": 1.9205648720211828, + "grad_norm": 1.0216399627321537, + "learning_rate": 9.32582884709098e-06, + "loss": 0.1702, + "step": 544 + }, + { + "epoch": 1.9240953221535746, + "grad_norm": 0.8524211551346756, + "learning_rate": 9.322956315794403e-06, + "loss": 0.1769, + "step": 545 + }, + { + "epoch": 1.9276257722859664, + "grad_norm": 1.0061876178961442, + "learning_rate": 9.320078122159077e-06, + "loss": 0.1725, + "step": 546 + }, + { + "epoch": 1.9311562224183585, + "grad_norm": 0.6449173961835437, + "learning_rate": 9.317194269954948e-06, + "loss": 0.1669, + "step": 547 + }, + { + "epoch": 1.9346866725507503, + "grad_norm": 0.6716320314286484, + "learning_rate": 9.314304762959373e-06, + "loss": 0.1837, + "step": 548 + }, + { + "epoch": 1.9382171226831422, + "grad_norm": 0.7847559077293133, + "learning_rate": 9.311409604957116e-06, + "loss": 0.1697, + "step": 549 + }, + { + "epoch": 1.941747572815534, + "grad_norm": 0.6479063264651344, + "learning_rate": 9.30850879974034e-06, + "loss": 0.1887, + "step": 550 + }, + { + "epoch": 1.9452780229479258, + "grad_norm": 0.5778944362755724, + "learning_rate": 9.305602351108613e-06, + "loss": 0.1732, + "step": 551 + }, + { + "epoch": 1.9488084730803177, + "grad_norm": 0.5993363081022134, + "learning_rate": 9.302690262868882e-06, + "loss": 0.1524, + "step": 552 + }, + { + "epoch": 1.9523389232127095, + "grad_norm": 0.7186259440033298, + "learning_rate": 9.299772538835492e-06, + "loss": 0.1562, + "step": 553 + }, + { + "epoch": 1.9558693733451014, + "grad_norm": 0.7547445542865213, + "learning_rate": 9.296849182830167e-06, + "loss": 0.159, + "step": 554 + }, + { + "epoch": 1.9593998234774934, + "grad_norm": 0.742454944410959, + "learning_rate": 9.293920198682004e-06, + "loss": 0.178, + "step": 555 + }, + { + "epoch": 1.9629302736098853, + "grad_norm": 0.731704919378445, + "learning_rate": 9.290985590227478e-06, + "loss": 0.178, + "step": 556 + }, + { + "epoch": 1.966460723742277, + "grad_norm": 0.6326777732721525, + "learning_rate": 9.288045361310423e-06, + "loss": 0.1572, + "step": 557 + }, + { + "epoch": 1.9699911738746692, + "grad_norm": 0.6096205323065964, + "learning_rate": 9.285099515782045e-06, + "loss": 0.1609, + "step": 558 + }, + { + "epoch": 1.973521624007061, + "grad_norm": 0.968981563236468, + "learning_rate": 9.2821480575009e-06, + "loss": 0.1554, + "step": 559 + }, + { + "epoch": 1.9770520741394528, + "grad_norm": 0.6731379335046332, + "learning_rate": 9.279190990332896e-06, + "loss": 0.1786, + "step": 560 + }, + { + "epoch": 1.9805825242718447, + "grad_norm": 0.5474632434518878, + "learning_rate": 9.276228318151288e-06, + "loss": 0.1825, + "step": 561 + }, + { + "epoch": 1.9841129744042365, + "grad_norm": 0.5300134793710058, + "learning_rate": 9.273260044836675e-06, + "loss": 0.1507, + "step": 562 + }, + { + "epoch": 1.9876434245366283, + "grad_norm": 0.6263089144886929, + "learning_rate": 9.270286174276992e-06, + "loss": 0.1753, + "step": 563 + }, + { + "epoch": 1.9911738746690202, + "grad_norm": 0.7895430535402537, + "learning_rate": 9.267306710367503e-06, + "loss": 0.1772, + "step": 564 + }, + { + "epoch": 1.994704324801412, + "grad_norm": 0.7740002132732854, + "learning_rate": 9.2643216570108e-06, + "loss": 0.1874, + "step": 565 + }, + { + "epoch": 1.998234774933804, + "grad_norm": 0.9874775418494401, + "learning_rate": 9.261331018116795e-06, + "loss": 0.173, + "step": 566 + }, + { + "epoch": 2.001765225066196, + "grad_norm": 0.6888142224041766, + "learning_rate": 9.25833479760272e-06, + "loss": 0.1811, + "step": 567 + }, + { + "epoch": 2.005295675198588, + "grad_norm": 0.6459025197429961, + "learning_rate": 9.255332999393114e-06, + "loss": 0.1507, + "step": 568 + }, + { + "epoch": 2.00882612533098, + "grad_norm": 0.6618092460843229, + "learning_rate": 9.252325627419821e-06, + "loss": 0.1302, + "step": 569 + }, + { + "epoch": 2.0123565754633717, + "grad_norm": 0.8386091353618218, + "learning_rate": 9.249312685621989e-06, + "loss": 0.1531, + "step": 570 + }, + { + "epoch": 2.0158870255957635, + "grad_norm": 0.7735686550828623, + "learning_rate": 9.246294177946062e-06, + "loss": 0.1554, + "step": 571 + }, + { + "epoch": 2.0194174757281553, + "grad_norm": 0.6350949023955103, + "learning_rate": 9.24327010834577e-06, + "loss": 0.1161, + "step": 572 + }, + { + "epoch": 2.022947925860547, + "grad_norm": 0.7901812966196844, + "learning_rate": 9.24024048078213e-06, + "loss": 0.1498, + "step": 573 + }, + { + "epoch": 2.026478375992939, + "grad_norm": 0.7745266020928578, + "learning_rate": 9.237205299223444e-06, + "loss": 0.1665, + "step": 574 + }, + { + "epoch": 2.030008826125331, + "grad_norm": 0.6972058102385509, + "learning_rate": 9.234164567645278e-06, + "loss": 0.1257, + "step": 575 + }, + { + "epoch": 2.0335392762577227, + "grad_norm": 0.6137168822739238, + "learning_rate": 9.231118290030481e-06, + "loss": 0.121, + "step": 576 + }, + { + "epoch": 2.0370697263901145, + "grad_norm": 0.7533855618361971, + "learning_rate": 9.228066470369155e-06, + "loss": 0.1352, + "step": 577 + }, + { + "epoch": 2.040600176522507, + "grad_norm": 0.792710448056873, + "learning_rate": 9.225009112658667e-06, + "loss": 0.144, + "step": 578 + }, + { + "epoch": 2.0441306266548986, + "grad_norm": 0.865387452735705, + "learning_rate": 9.221946220903634e-06, + "loss": 0.1521, + "step": 579 + }, + { + "epoch": 2.0476610767872905, + "grad_norm": 1.0235916124259214, + "learning_rate": 9.218877799115929e-06, + "loss": 0.1441, + "step": 580 + }, + { + "epoch": 2.0511915269196823, + "grad_norm": 1.0508276358903932, + "learning_rate": 9.21580385131466e-06, + "loss": 0.1685, + "step": 581 + }, + { + "epoch": 2.054721977052074, + "grad_norm": 0.8910905064383835, + "learning_rate": 9.212724381526177e-06, + "loss": 0.1736, + "step": 582 + }, + { + "epoch": 2.058252427184466, + "grad_norm": 1.0571601952147707, + "learning_rate": 9.209639393784064e-06, + "loss": 0.1565, + "step": 583 + }, + { + "epoch": 2.061782877316858, + "grad_norm": 0.8815436783906608, + "learning_rate": 9.206548892129129e-06, + "loss": 0.1371, + "step": 584 + }, + { + "epoch": 2.0653133274492497, + "grad_norm": 1.0517504956260713, + "learning_rate": 9.203452880609407e-06, + "loss": 0.1464, + "step": 585 + }, + { + "epoch": 2.0688437775816415, + "grad_norm": 0.8529237538392984, + "learning_rate": 9.200351363280144e-06, + "loss": 0.1266, + "step": 586 + }, + { + "epoch": 2.0723742277140333, + "grad_norm": 0.874098127040573, + "learning_rate": 9.197244344203804e-06, + "loss": 0.1349, + "step": 587 + }, + { + "epoch": 2.0759046778464256, + "grad_norm": 0.8109927532199982, + "learning_rate": 9.194131827450054e-06, + "loss": 0.1445, + "step": 588 + }, + { + "epoch": 2.0794351279788175, + "grad_norm": 0.6489560116761195, + "learning_rate": 9.191013817095762e-06, + "loss": 0.1294, + "step": 589 + }, + { + "epoch": 2.0829655781112093, + "grad_norm": 0.7508144606670216, + "learning_rate": 9.18789031722499e-06, + "loss": 0.1135, + "step": 590 + }, + { + "epoch": 2.086496028243601, + "grad_norm": 0.7295793522432309, + "learning_rate": 9.184761331928997e-06, + "loss": 0.1363, + "step": 591 + }, + { + "epoch": 2.090026478375993, + "grad_norm": 0.861095430673198, + "learning_rate": 9.18162686530622e-06, + "loss": 0.1279, + "step": 592 + }, + { + "epoch": 2.093556928508385, + "grad_norm": 0.6861216733576722, + "learning_rate": 9.178486921462276e-06, + "loss": 0.1224, + "step": 593 + }, + { + "epoch": 2.0970873786407767, + "grad_norm": 0.8019378401083357, + "learning_rate": 9.175341504509961e-06, + "loss": 0.1545, + "step": 594 + }, + { + "epoch": 2.1006178287731685, + "grad_norm": 1.526043856422149, + "learning_rate": 9.172190618569236e-06, + "loss": 0.1306, + "step": 595 + }, + { + "epoch": 2.1041482789055603, + "grad_norm": 0.8476895681346993, + "learning_rate": 9.16903426776723e-06, + "loss": 0.149, + "step": 596 + }, + { + "epoch": 2.107678729037952, + "grad_norm": 0.8161608102533534, + "learning_rate": 9.16587245623822e-06, + "loss": 0.1364, + "step": 597 + }, + { + "epoch": 2.111209179170344, + "grad_norm": 0.7223711233696612, + "learning_rate": 9.162705188123647e-06, + "loss": 0.142, + "step": 598 + }, + { + "epoch": 2.1147396293027363, + "grad_norm": 0.8507390490828752, + "learning_rate": 9.159532467572094e-06, + "loss": 0.1547, + "step": 599 + }, + { + "epoch": 2.118270079435128, + "grad_norm": 0.8316619115260057, + "learning_rate": 9.156354298739285e-06, + "loss": 0.1563, + "step": 600 + }, + { + "epoch": 2.12180052956752, + "grad_norm": 0.7544897788791859, + "learning_rate": 9.153170685788081e-06, + "loss": 0.1421, + "step": 601 + }, + { + "epoch": 2.125330979699912, + "grad_norm": 0.800819425552279, + "learning_rate": 9.149981632888476e-06, + "loss": 0.1579, + "step": 602 + }, + { + "epoch": 2.1288614298323036, + "grad_norm": 0.7838909562518437, + "learning_rate": 9.146787144217588e-06, + "loss": 0.1681, + "step": 603 + }, + { + "epoch": 2.1323918799646955, + "grad_norm": 0.9303054432448055, + "learning_rate": 9.143587223959653e-06, + "loss": 0.1397, + "step": 604 + }, + { + "epoch": 2.1359223300970873, + "grad_norm": 0.7625989738070533, + "learning_rate": 9.140381876306026e-06, + "loss": 0.1581, + "step": 605 + }, + { + "epoch": 2.139452780229479, + "grad_norm": 0.7857755303539646, + "learning_rate": 9.137171105455165e-06, + "loss": 0.1258, + "step": 606 + }, + { + "epoch": 2.142983230361871, + "grad_norm": 0.6441805120671155, + "learning_rate": 9.133954915612635e-06, + "loss": 0.1305, + "step": 607 + }, + { + "epoch": 2.146513680494263, + "grad_norm": 0.8467018145382312, + "learning_rate": 9.130733310991102e-06, + "loss": 0.1432, + "step": 608 + }, + { + "epoch": 2.1500441306266547, + "grad_norm": 0.7571331987070536, + "learning_rate": 9.127506295810316e-06, + "loss": 0.1536, + "step": 609 + }, + { + "epoch": 2.153574580759047, + "grad_norm": 1.0166648071075606, + "learning_rate": 9.124273874297123e-06, + "loss": 0.1641, + "step": 610 + }, + { + "epoch": 2.157105030891439, + "grad_norm": 1.0624683994204978, + "learning_rate": 9.121036050685443e-06, + "loss": 0.1568, + "step": 611 + }, + { + "epoch": 2.1606354810238306, + "grad_norm": 0.754254288642853, + "learning_rate": 9.11779282921628e-06, + "loss": 0.1456, + "step": 612 + }, + { + "epoch": 2.1641659311562225, + "grad_norm": 0.8711250730686613, + "learning_rate": 9.114544214137699e-06, + "loss": 0.1664, + "step": 613 + }, + { + "epoch": 2.1676963812886143, + "grad_norm": 0.8545203984797597, + "learning_rate": 9.111290209704835e-06, + "loss": 0.1494, + "step": 614 + }, + { + "epoch": 2.171226831421006, + "grad_norm": 0.8374462073082292, + "learning_rate": 9.108030820179882e-06, + "loss": 0.1541, + "step": 615 + }, + { + "epoch": 2.174757281553398, + "grad_norm": 0.7357997781637011, + "learning_rate": 9.104766049832088e-06, + "loss": 0.1491, + "step": 616 + }, + { + "epoch": 2.17828773168579, + "grad_norm": 0.7606936499199851, + "learning_rate": 9.101495902937744e-06, + "loss": 0.1577, + "step": 617 + }, + { + "epoch": 2.1818181818181817, + "grad_norm": 0.8031929036786638, + "learning_rate": 9.098220383780191e-06, + "loss": 0.1649, + "step": 618 + }, + { + "epoch": 2.1853486319505735, + "grad_norm": 1.064933842253509, + "learning_rate": 9.094939496649803e-06, + "loss": 0.1218, + "step": 619 + }, + { + "epoch": 2.1888790820829658, + "grad_norm": 0.7301032836107995, + "learning_rate": 9.091653245843984e-06, + "loss": 0.1376, + "step": 620 + }, + { + "epoch": 2.1924095322153576, + "grad_norm": 0.8431618205640518, + "learning_rate": 9.088361635667162e-06, + "loss": 0.148, + "step": 621 + }, + { + "epoch": 2.1959399823477495, + "grad_norm": 1.340479354749448, + "learning_rate": 9.08506467043079e-06, + "loss": 0.1616, + "step": 622 + }, + { + "epoch": 2.1994704324801413, + "grad_norm": 0.8064011593740745, + "learning_rate": 9.081762354453333e-06, + "loss": 0.1488, + "step": 623 + }, + { + "epoch": 2.203000882612533, + "grad_norm": 0.8004191241563091, + "learning_rate": 9.078454692060264e-06, + "loss": 0.1403, + "step": 624 + }, + { + "epoch": 2.206531332744925, + "grad_norm": 0.8885936871876676, + "learning_rate": 9.075141687584056e-06, + "loss": 0.1298, + "step": 625 + }, + { + "epoch": 2.210061782877317, + "grad_norm": 1.013124408250628, + "learning_rate": 9.071823345364187e-06, + "loss": 0.1461, + "step": 626 + }, + { + "epoch": 2.2135922330097086, + "grad_norm": 0.8812285753845427, + "learning_rate": 9.068499669747122e-06, + "loss": 0.1475, + "step": 627 + }, + { + "epoch": 2.2171226831421005, + "grad_norm": 0.8390832015738438, + "learning_rate": 9.065170665086307e-06, + "loss": 0.152, + "step": 628 + }, + { + "epoch": 2.2206531332744923, + "grad_norm": 0.6214987457344594, + "learning_rate": 9.061836335742175e-06, + "loss": 0.1418, + "step": 629 + }, + { + "epoch": 2.2241835834068846, + "grad_norm": 0.8210501747316103, + "learning_rate": 9.058496686082132e-06, + "loss": 0.1385, + "step": 630 + }, + { + "epoch": 2.2277140335392764, + "grad_norm": 0.8877640524225171, + "learning_rate": 9.055151720480553e-06, + "loss": 0.1636, + "step": 631 + }, + { + "epoch": 2.2312444836716683, + "grad_norm": 0.6993342223349599, + "learning_rate": 9.051801443318772e-06, + "loss": 0.1267, + "step": 632 + }, + { + "epoch": 2.23477493380406, + "grad_norm": 0.915383376214395, + "learning_rate": 9.048445858985085e-06, + "loss": 0.1346, + "step": 633 + }, + { + "epoch": 2.238305383936452, + "grad_norm": 0.8226201621022076, + "learning_rate": 9.045084971874738e-06, + "loss": 0.145, + "step": 634 + }, + { + "epoch": 2.241835834068844, + "grad_norm": 0.949771207964201, + "learning_rate": 9.041718786389921e-06, + "loss": 0.1729, + "step": 635 + }, + { + "epoch": 2.2453662842012356, + "grad_norm": 0.7669826168302643, + "learning_rate": 9.038347306939768e-06, + "loss": 0.1491, + "step": 636 + }, + { + "epoch": 2.2488967343336275, + "grad_norm": 0.8401151453900115, + "learning_rate": 9.034970537940342e-06, + "loss": 0.1587, + "step": 637 + }, + { + "epoch": 2.2524271844660193, + "grad_norm": 0.8362745634336628, + "learning_rate": 9.031588483814639e-06, + "loss": 0.152, + "step": 638 + }, + { + "epoch": 2.255957634598411, + "grad_norm": 0.828931692995688, + "learning_rate": 9.028201148992572e-06, + "loss": 0.127, + "step": 639 + }, + { + "epoch": 2.259488084730803, + "grad_norm": 0.8901277435562056, + "learning_rate": 9.024808537910981e-06, + "loss": 0.1472, + "step": 640 + }, + { + "epoch": 2.2630185348631953, + "grad_norm": 0.8458917100010721, + "learning_rate": 9.021410655013604e-06, + "loss": 0.1549, + "step": 641 + }, + { + "epoch": 2.266548984995587, + "grad_norm": 0.7511334646906924, + "learning_rate": 9.018007504751094e-06, + "loss": 0.1401, + "step": 642 + }, + { + "epoch": 2.270079435127979, + "grad_norm": 0.8998427764695459, + "learning_rate": 9.014599091581e-06, + "loss": 0.1665, + "step": 643 + }, + { + "epoch": 2.2736098852603708, + "grad_norm": 0.8241701592002284, + "learning_rate": 9.011185419967761e-06, + "loss": 0.136, + "step": 644 + }, + { + "epoch": 2.2771403353927626, + "grad_norm": 0.7830126296047867, + "learning_rate": 9.007766494382713e-06, + "loss": 0.151, + "step": 645 + }, + { + "epoch": 2.2806707855251545, + "grad_norm": 1.0742703925847628, + "learning_rate": 9.004342319304063e-06, + "loss": 0.1606, + "step": 646 + }, + { + "epoch": 2.2842012356575463, + "grad_norm": 0.9139889110017969, + "learning_rate": 9.000912899216902e-06, + "loss": 0.1556, + "step": 647 + }, + { + "epoch": 2.287731685789938, + "grad_norm": 0.9228198414722247, + "learning_rate": 8.997478238613186e-06, + "loss": 0.1635, + "step": 648 + }, + { + "epoch": 2.29126213592233, + "grad_norm": 0.6979642688439006, + "learning_rate": 8.994038341991739e-06, + "loss": 0.1443, + "step": 649 + }, + { + "epoch": 2.294792586054722, + "grad_norm": 0.7432603342893744, + "learning_rate": 8.990593213858239e-06, + "loss": 0.1497, + "step": 650 + }, + { + "epoch": 2.2983230361871136, + "grad_norm": 0.9788738069579437, + "learning_rate": 8.987142858725221e-06, + "loss": 0.1388, + "step": 651 + }, + { + "epoch": 2.301853486319506, + "grad_norm": 0.9451424967898011, + "learning_rate": 8.983687281112066e-06, + "loss": 0.1558, + "step": 652 + }, + { + "epoch": 2.3053839364518978, + "grad_norm": 0.796516553881763, + "learning_rate": 8.98022648554499e-06, + "loss": 0.1395, + "step": 653 + }, + { + "epoch": 2.3089143865842896, + "grad_norm": 0.6535029248700488, + "learning_rate": 8.976760476557049e-06, + "loss": 0.1466, + "step": 654 + }, + { + "epoch": 2.3124448367166814, + "grad_norm": 1.0145090677139883, + "learning_rate": 8.973289258688125e-06, + "loss": 0.1601, + "step": 655 + }, + { + "epoch": 2.3159752868490733, + "grad_norm": 0.9360528769068794, + "learning_rate": 8.969812836484929e-06, + "loss": 0.164, + "step": 656 + }, + { + "epoch": 2.319505736981465, + "grad_norm": 0.8047358252615086, + "learning_rate": 8.966331214500978e-06, + "loss": 0.1756, + "step": 657 + }, + { + "epoch": 2.323036187113857, + "grad_norm": 0.7975682754458384, + "learning_rate": 8.962844397296608e-06, + "loss": 0.1435, + "step": 658 + }, + { + "epoch": 2.326566637246249, + "grad_norm": 0.7277448447990951, + "learning_rate": 8.959352389438959e-06, + "loss": 0.1352, + "step": 659 + }, + { + "epoch": 2.3300970873786406, + "grad_norm": 0.8639595145485759, + "learning_rate": 8.955855195501966e-06, + "loss": 0.1497, + "step": 660 + }, + { + "epoch": 2.3336275375110325, + "grad_norm": 0.9741801125022028, + "learning_rate": 8.952352820066359e-06, + "loss": 0.1743, + "step": 661 + }, + { + "epoch": 2.3371579876434243, + "grad_norm": 0.7011036253403068, + "learning_rate": 8.948845267719658e-06, + "loss": 0.1483, + "step": 662 + }, + { + "epoch": 2.3406884377758166, + "grad_norm": 0.7915113319296646, + "learning_rate": 8.945332543056158e-06, + "loss": 0.1674, + "step": 663 + }, + { + "epoch": 2.3442188879082084, + "grad_norm": 0.7246375584288023, + "learning_rate": 8.941814650676934e-06, + "loss": 0.1429, + "step": 664 + }, + { + "epoch": 2.3477493380406003, + "grad_norm": 0.6572418012169463, + "learning_rate": 8.938291595189826e-06, + "loss": 0.1464, + "step": 665 + }, + { + "epoch": 2.351279788172992, + "grad_norm": 1.0088944178687222, + "learning_rate": 8.93476338120944e-06, + "loss": 0.1723, + "step": 666 + }, + { + "epoch": 2.354810238305384, + "grad_norm": 0.8873624823759165, + "learning_rate": 8.931230013357136e-06, + "loss": 0.1478, + "step": 667 + }, + { + "epoch": 2.358340688437776, + "grad_norm": 0.8232927235637216, + "learning_rate": 8.927691496261025e-06, + "loss": 0.1509, + "step": 668 + }, + { + "epoch": 2.3618711385701676, + "grad_norm": 0.7448944968086746, + "learning_rate": 8.924147834555966e-06, + "loss": 0.144, + "step": 669 + }, + { + "epoch": 2.3654015887025595, + "grad_norm": 0.6545431722863853, + "learning_rate": 8.920599032883553e-06, + "loss": 0.1442, + "step": 670 + }, + { + "epoch": 2.3689320388349513, + "grad_norm": 0.9857690033050072, + "learning_rate": 8.917045095892115e-06, + "loss": 0.1534, + "step": 671 + }, + { + "epoch": 2.3724624889673436, + "grad_norm": 0.8890672612580922, + "learning_rate": 8.913486028236703e-06, + "loss": 0.1414, + "step": 672 + }, + { + "epoch": 2.375992939099735, + "grad_norm": 0.7720214368079358, + "learning_rate": 8.909921834579093e-06, + "loss": 0.1379, + "step": 673 + }, + { + "epoch": 2.3795233892321273, + "grad_norm": 0.7968382535551254, + "learning_rate": 8.906352519587774e-06, + "loss": 0.1665, + "step": 674 + }, + { + "epoch": 2.383053839364519, + "grad_norm": 1.012361224051788, + "learning_rate": 8.902778087937943e-06, + "loss": 0.1506, + "step": 675 + }, + { + "epoch": 2.386584289496911, + "grad_norm": 1.6922659485205467, + "learning_rate": 8.899198544311496e-06, + "loss": 0.1621, + "step": 676 + }, + { + "epoch": 2.3901147396293028, + "grad_norm": 0.8810742656212996, + "learning_rate": 8.89561389339703e-06, + "loss": 0.1342, + "step": 677 + }, + { + "epoch": 2.3936451897616946, + "grad_norm": 0.8922813495137972, + "learning_rate": 8.892024139889827e-06, + "loss": 0.1348, + "step": 678 + }, + { + "epoch": 2.3971756398940864, + "grad_norm": 0.6428583850540706, + "learning_rate": 8.888429288491857e-06, + "loss": 0.1365, + "step": 679 + }, + { + "epoch": 2.4007060900264783, + "grad_norm": 0.959500169941628, + "learning_rate": 8.884829343911763e-06, + "loss": 0.1579, + "step": 680 + }, + { + "epoch": 2.40423654015887, + "grad_norm": 0.8928078544095518, + "learning_rate": 8.88122431086486e-06, + "loss": 0.158, + "step": 681 + }, + { + "epoch": 2.407766990291262, + "grad_norm": 0.7299606081515114, + "learning_rate": 8.877614194073134e-06, + "loss": 0.1332, + "step": 682 + }, + { + "epoch": 2.4112974404236542, + "grad_norm": 0.867643203493481, + "learning_rate": 8.87399899826522e-06, + "loss": 0.1567, + "step": 683 + }, + { + "epoch": 2.414827890556046, + "grad_norm": 0.8544098798566452, + "learning_rate": 8.870378728176412e-06, + "loss": 0.1479, + "step": 684 + }, + { + "epoch": 2.418358340688438, + "grad_norm": 1.0065430757516332, + "learning_rate": 8.86675338854865e-06, + "loss": 0.1826, + "step": 685 + }, + { + "epoch": 2.4218887908208298, + "grad_norm": 0.9392726338359665, + "learning_rate": 8.863122984130514e-06, + "loss": 0.1573, + "step": 686 + }, + { + "epoch": 2.4254192409532216, + "grad_norm": 0.815039972544626, + "learning_rate": 8.859487519677215e-06, + "loss": 0.1296, + "step": 687 + }, + { + "epoch": 2.4289496910856134, + "grad_norm": 0.8603703909617912, + "learning_rate": 8.855846999950595e-06, + "loss": 0.1547, + "step": 688 + }, + { + "epoch": 2.4324801412180053, + "grad_norm": 1.0058587648363357, + "learning_rate": 8.852201429719119e-06, + "loss": 0.1394, + "step": 689 + }, + { + "epoch": 2.436010591350397, + "grad_norm": 0.895199851785628, + "learning_rate": 8.848550813757861e-06, + "loss": 0.1379, + "step": 690 + }, + { + "epoch": 2.439541041482789, + "grad_norm": 0.7927306345545818, + "learning_rate": 8.844895156848511e-06, + "loss": 0.1347, + "step": 691 + }, + { + "epoch": 2.443071491615181, + "grad_norm": 0.7501822257778045, + "learning_rate": 8.841234463779359e-06, + "loss": 0.1472, + "step": 692 + }, + { + "epoch": 2.4466019417475726, + "grad_norm": 0.8155634914790267, + "learning_rate": 8.83756873934529e-06, + "loss": 0.1637, + "step": 693 + }, + { + "epoch": 2.450132391879965, + "grad_norm": 0.9416808234395326, + "learning_rate": 8.83389798834778e-06, + "loss": 0.1619, + "step": 694 + }, + { + "epoch": 2.4536628420123567, + "grad_norm": 0.7762931486236544, + "learning_rate": 8.83022221559489e-06, + "loss": 0.1607, + "step": 695 + }, + { + "epoch": 2.4571932921447486, + "grad_norm": 0.6647840629115083, + "learning_rate": 8.82654142590126e-06, + "loss": 0.1474, + "step": 696 + }, + { + "epoch": 2.4607237422771404, + "grad_norm": 0.809281030127039, + "learning_rate": 8.822855624088099e-06, + "loss": 0.1329, + "step": 697 + }, + { + "epoch": 2.4642541924095323, + "grad_norm": 0.7105112267949759, + "learning_rate": 8.819164814983179e-06, + "loss": 0.1513, + "step": 698 + }, + { + "epoch": 2.467784642541924, + "grad_norm": 0.6643065745395029, + "learning_rate": 8.815469003420835e-06, + "loss": 0.1355, + "step": 699 + }, + { + "epoch": 2.471315092674316, + "grad_norm": 0.6786894914054595, + "learning_rate": 8.811768194241951e-06, + "loss": 0.1491, + "step": 700 + }, + { + "epoch": 2.4748455428067078, + "grad_norm": 0.8985899448983452, + "learning_rate": 8.808062392293964e-06, + "loss": 0.1683, + "step": 701 + }, + { + "epoch": 2.4783759929390996, + "grad_norm": 0.8051939382089317, + "learning_rate": 8.804351602430838e-06, + "loss": 0.1316, + "step": 702 + }, + { + "epoch": 2.4819064430714914, + "grad_norm": 0.9657848784459468, + "learning_rate": 8.800635829513083e-06, + "loss": 0.1695, + "step": 703 + }, + { + "epoch": 2.4854368932038833, + "grad_norm": 0.8063728250819603, + "learning_rate": 8.796915078407729e-06, + "loss": 0.1524, + "step": 704 + }, + { + "epoch": 2.4889673433362756, + "grad_norm": 0.6461209785079439, + "learning_rate": 8.793189353988325e-06, + "loss": 0.1487, + "step": 705 + }, + { + "epoch": 2.4924977934686674, + "grad_norm": 0.6869278039173332, + "learning_rate": 8.789458661134943e-06, + "loss": 0.1442, + "step": 706 + }, + { + "epoch": 2.4960282436010592, + "grad_norm": 1.1034616460480864, + "learning_rate": 8.785723004734153e-06, + "loss": 0.1714, + "step": 707 + }, + { + "epoch": 2.499558693733451, + "grad_norm": 0.6104590542245811, + "learning_rate": 8.781982389679033e-06, + "loss": 0.1339, + "step": 708 + }, + { + "epoch": 2.503089143865843, + "grad_norm": 0.8481696722493157, + "learning_rate": 8.778236820869152e-06, + "loss": 0.1536, + "step": 709 + }, + { + "epoch": 2.5066195939982348, + "grad_norm": 0.8111930888716574, + "learning_rate": 8.77448630321057e-06, + "loss": 0.1422, + "step": 710 + }, + { + "epoch": 2.5101500441306266, + "grad_norm": 0.7036188732668607, + "learning_rate": 8.770730841615827e-06, + "loss": 0.1445, + "step": 711 + }, + { + "epoch": 2.5136804942630184, + "grad_norm": 0.9697606827791236, + "learning_rate": 8.766970441003942e-06, + "loss": 0.167, + "step": 712 + }, + { + "epoch": 2.5172109443954103, + "grad_norm": 0.8387174374399691, + "learning_rate": 8.7632051063004e-06, + "loss": 0.1585, + "step": 713 + }, + { + "epoch": 2.5207413945278025, + "grad_norm": 0.6631550381542344, + "learning_rate": 8.75943484243715e-06, + "loss": 0.1428, + "step": 714 + }, + { + "epoch": 2.524271844660194, + "grad_norm": 0.7695360425616815, + "learning_rate": 8.755659654352599e-06, + "loss": 0.1463, + "step": 715 + }, + { + "epoch": 2.5278022947925862, + "grad_norm": 0.7011902192003712, + "learning_rate": 8.751879546991601e-06, + "loss": 0.1406, + "step": 716 + }, + { + "epoch": 2.531332744924978, + "grad_norm": 0.7991566673347344, + "learning_rate": 8.748094525305457e-06, + "loss": 0.1545, + "step": 717 + }, + { + "epoch": 2.53486319505737, + "grad_norm": 0.8486318435641594, + "learning_rate": 8.744304594251898e-06, + "loss": 0.1565, + "step": 718 + }, + { + "epoch": 2.5383936451897617, + "grad_norm": 0.6882492324618286, + "learning_rate": 8.740509758795094e-06, + "loss": 0.141, + "step": 719 + }, + { + "epoch": 2.5419240953221536, + "grad_norm": 0.7803012919767831, + "learning_rate": 8.736710023905634e-06, + "loss": 0.1671, + "step": 720 + }, + { + "epoch": 2.5454545454545454, + "grad_norm": 1.0460659891563937, + "learning_rate": 8.732905394560526e-06, + "loss": 0.1662, + "step": 721 + }, + { + "epoch": 2.5489849955869373, + "grad_norm": 0.8104918877045203, + "learning_rate": 8.729095875743187e-06, + "loss": 0.1707, + "step": 722 + }, + { + "epoch": 2.552515445719329, + "grad_norm": 0.822076128070119, + "learning_rate": 8.725281472443442e-06, + "loss": 0.1434, + "step": 723 + }, + { + "epoch": 2.556045895851721, + "grad_norm": 0.8428018638045196, + "learning_rate": 8.72146218965751e-06, + "loss": 0.1452, + "step": 724 + }, + { + "epoch": 2.559576345984113, + "grad_norm": 0.6860938138056261, + "learning_rate": 8.717638032388002e-06, + "loss": 0.1452, + "step": 725 + }, + { + "epoch": 2.5631067961165046, + "grad_norm": 0.6679034911744651, + "learning_rate": 8.713809005643919e-06, + "loss": 0.1415, + "step": 726 + }, + { + "epoch": 2.566637246248897, + "grad_norm": 0.8156715771369578, + "learning_rate": 8.709975114440632e-06, + "loss": 0.1664, + "step": 727 + }, + { + "epoch": 2.5701676963812887, + "grad_norm": 0.950713499966215, + "learning_rate": 8.706136363799888e-06, + "loss": 0.1579, + "step": 728 + }, + { + "epoch": 2.5736981465136806, + "grad_norm": 0.6736037466099428, + "learning_rate": 8.7022927587498e-06, + "loss": 0.1376, + "step": 729 + }, + { + "epoch": 2.5772285966460724, + "grad_norm": 0.8007987125298717, + "learning_rate": 8.698444304324837e-06, + "loss": 0.1604, + "step": 730 + }, + { + "epoch": 2.5807590467784642, + "grad_norm": 0.7909053805653098, + "learning_rate": 8.694591005565818e-06, + "loss": 0.1525, + "step": 731 + }, + { + "epoch": 2.584289496910856, + "grad_norm": 0.7510519443516507, + "learning_rate": 8.690732867519915e-06, + "loss": 0.152, + "step": 732 + }, + { + "epoch": 2.587819947043248, + "grad_norm": 0.7070326571398531, + "learning_rate": 8.686869895240631e-06, + "loss": 0.1478, + "step": 733 + }, + { + "epoch": 2.5913503971756398, + "grad_norm": 0.7843881178771545, + "learning_rate": 8.683002093787805e-06, + "loss": 0.155, + "step": 734 + }, + { + "epoch": 2.5948808473080316, + "grad_norm": 0.74442690372609, + "learning_rate": 8.679129468227601e-06, + "loss": 0.1742, + "step": 735 + }, + { + "epoch": 2.598411297440424, + "grad_norm": 0.8873243162061539, + "learning_rate": 8.675252023632497e-06, + "loss": 0.1433, + "step": 736 + }, + { + "epoch": 2.6019417475728153, + "grad_norm": 1.0558814263186358, + "learning_rate": 8.671369765081291e-06, + "loss": 0.1383, + "step": 737 + }, + { + "epoch": 2.6054721977052075, + "grad_norm": 0.9661052979159337, + "learning_rate": 8.66748269765908e-06, + "loss": 0.1519, + "step": 738 + }, + { + "epoch": 2.6090026478375994, + "grad_norm": 0.7940230362582856, + "learning_rate": 8.663590826457264e-06, + "loss": 0.1449, + "step": 739 + }, + { + "epoch": 2.6125330979699912, + "grad_norm": 0.7712249737573764, + "learning_rate": 8.659694156573533e-06, + "loss": 0.1567, + "step": 740 + }, + { + "epoch": 2.616063548102383, + "grad_norm": 0.7354585280144065, + "learning_rate": 8.655792693111864e-06, + "loss": 0.147, + "step": 741 + }, + { + "epoch": 2.619593998234775, + "grad_norm": 0.7686478452231044, + "learning_rate": 8.651886441182509e-06, + "loss": 0.1462, + "step": 742 + }, + { + "epoch": 2.6231244483671667, + "grad_norm": 0.9615765356144627, + "learning_rate": 8.647975405901996e-06, + "loss": 0.1598, + "step": 743 + }, + { + "epoch": 2.6266548984995586, + "grad_norm": 0.7899472619604829, + "learning_rate": 8.644059592393117e-06, + "loss": 0.1473, + "step": 744 + }, + { + "epoch": 2.6301853486319504, + "grad_norm": 0.8716879941644811, + "learning_rate": 8.640139005784924e-06, + "loss": 0.1512, + "step": 745 + }, + { + "epoch": 2.6337157987643423, + "grad_norm": 0.9497477417594384, + "learning_rate": 8.636213651212719e-06, + "loss": 0.1501, + "step": 746 + }, + { + "epoch": 2.6372462488967345, + "grad_norm": 0.8255927742841409, + "learning_rate": 8.632283533818048e-06, + "loss": 0.147, + "step": 747 + }, + { + "epoch": 2.6407766990291264, + "grad_norm": 0.6589014957846149, + "learning_rate": 8.628348658748702e-06, + "loss": 0.146, + "step": 748 + }, + { + "epoch": 2.644307149161518, + "grad_norm": 0.7383869029734058, + "learning_rate": 8.624409031158696e-06, + "loss": 0.1444, + "step": 749 + }, + { + "epoch": 2.64783759929391, + "grad_norm": 0.8754969443335403, + "learning_rate": 8.620464656208272e-06, + "loss": 0.1533, + "step": 750 + }, + { + "epoch": 2.651368049426302, + "grad_norm": 0.8093544382322876, + "learning_rate": 8.616515539063894e-06, + "loss": 0.1466, + "step": 751 + }, + { + "epoch": 2.6548984995586937, + "grad_norm": 0.8094232555314647, + "learning_rate": 8.612561684898235e-06, + "loss": 0.1414, + "step": 752 + }, + { + "epoch": 2.6584289496910856, + "grad_norm": 0.9276446354153174, + "learning_rate": 8.60860309889017e-06, + "loss": 0.1608, + "step": 753 + }, + { + "epoch": 2.6619593998234774, + "grad_norm": 0.8670503330866877, + "learning_rate": 8.604639786224778e-06, + "loss": 0.148, + "step": 754 + }, + { + "epoch": 2.6654898499558692, + "grad_norm": 0.9236407101629923, + "learning_rate": 8.600671752093323e-06, + "loss": 0.1465, + "step": 755 + }, + { + "epoch": 2.6690203000882615, + "grad_norm": 0.7947533360687092, + "learning_rate": 8.596699001693257e-06, + "loss": 0.1382, + "step": 756 + }, + { + "epoch": 2.672550750220653, + "grad_norm": 0.9401531663981636, + "learning_rate": 8.592721540228207e-06, + "loss": 0.1545, + "step": 757 + }, + { + "epoch": 2.676081200353045, + "grad_norm": 0.6849210881232853, + "learning_rate": 8.588739372907974e-06, + "loss": 0.1422, + "step": 758 + }, + { + "epoch": 2.679611650485437, + "grad_norm": 0.9020222283025484, + "learning_rate": 8.584752504948522e-06, + "loss": 0.148, + "step": 759 + }, + { + "epoch": 2.683142100617829, + "grad_norm": 0.763357957155723, + "learning_rate": 8.580760941571968e-06, + "loss": 0.1387, + "step": 760 + }, + { + "epoch": 2.6866725507502207, + "grad_norm": 0.8677207545705664, + "learning_rate": 8.576764688006583e-06, + "loss": 0.1694, + "step": 761 + }, + { + "epoch": 2.6902030008826125, + "grad_norm": 0.7387471803373706, + "learning_rate": 8.572763749486784e-06, + "loss": 0.1507, + "step": 762 + }, + { + "epoch": 2.6937334510150044, + "grad_norm": 0.6986240692420573, + "learning_rate": 8.568758131253115e-06, + "loss": 0.1632, + "step": 763 + }, + { + "epoch": 2.6972639011473962, + "grad_norm": 0.9057220759646399, + "learning_rate": 8.564747838552262e-06, + "loss": 0.1607, + "step": 764 + }, + { + "epoch": 2.700794351279788, + "grad_norm": 0.8635819293383579, + "learning_rate": 8.560732876637023e-06, + "loss": 0.1446, + "step": 765 + }, + { + "epoch": 2.70432480141218, + "grad_norm": 0.773759396974051, + "learning_rate": 8.55671325076632e-06, + "loss": 0.152, + "step": 766 + }, + { + "epoch": 2.707855251544572, + "grad_norm": 0.7726341748367627, + "learning_rate": 8.552688966205176e-06, + "loss": 0.1507, + "step": 767 + }, + { + "epoch": 2.7113857016769636, + "grad_norm": 0.8172707078015327, + "learning_rate": 8.548660028224724e-06, + "loss": 0.1624, + "step": 768 + }, + { + "epoch": 2.714916151809356, + "grad_norm": 0.7096834505741385, + "learning_rate": 8.544626442102188e-06, + "loss": 0.1451, + "step": 769 + }, + { + "epoch": 2.7184466019417477, + "grad_norm": 0.7394117544451063, + "learning_rate": 8.54058821312088e-06, + "loss": 0.163, + "step": 770 + }, + { + "epoch": 2.7219770520741395, + "grad_norm": 0.9101784701617734, + "learning_rate": 8.536545346570195e-06, + "loss": 0.1597, + "step": 771 + }, + { + "epoch": 2.7255075022065314, + "grad_norm": 0.8528637952954311, + "learning_rate": 8.532497847745602e-06, + "loss": 0.1455, + "step": 772 + }, + { + "epoch": 2.729037952338923, + "grad_norm": 0.7128839210498725, + "learning_rate": 8.528445721948637e-06, + "loss": 0.1562, + "step": 773 + }, + { + "epoch": 2.732568402471315, + "grad_norm": 0.7514251001247261, + "learning_rate": 8.524388974486896e-06, + "loss": 0.1787, + "step": 774 + }, + { + "epoch": 2.736098852603707, + "grad_norm": 0.8717097957173104, + "learning_rate": 8.520327610674029e-06, + "loss": 0.1686, + "step": 775 + }, + { + "epoch": 2.7396293027360987, + "grad_norm": 0.8134061335902161, + "learning_rate": 8.516261635829734e-06, + "loss": 0.1444, + "step": 776 + }, + { + "epoch": 2.7431597528684906, + "grad_norm": 0.8130838438482634, + "learning_rate": 8.512191055279749e-06, + "loss": 0.1621, + "step": 777 + }, + { + "epoch": 2.746690203000883, + "grad_norm": 0.7609974982707016, + "learning_rate": 8.50811587435584e-06, + "loss": 0.1639, + "step": 778 + }, + { + "epoch": 2.7502206531332742, + "grad_norm": 0.8656038452478626, + "learning_rate": 8.504036098395805e-06, + "loss": 0.1584, + "step": 779 + }, + { + "epoch": 2.7537511032656665, + "grad_norm": 0.7488569534666748, + "learning_rate": 8.499951732743457e-06, + "loss": 0.148, + "step": 780 + }, + { + "epoch": 2.7572815533980584, + "grad_norm": 0.77448853187229, + "learning_rate": 8.495862782748619e-06, + "loss": 0.1406, + "step": 781 + }, + { + "epoch": 2.76081200353045, + "grad_norm": 0.6757102359665434, + "learning_rate": 8.491769253767126e-06, + "loss": 0.1301, + "step": 782 + }, + { + "epoch": 2.764342453662842, + "grad_norm": 0.7775212534985085, + "learning_rate": 8.487671151160801e-06, + "loss": 0.1595, + "step": 783 + }, + { + "epoch": 2.767872903795234, + "grad_norm": 0.7117920982524355, + "learning_rate": 8.483568480297467e-06, + "loss": 0.1378, + "step": 784 + }, + { + "epoch": 2.7714033539276257, + "grad_norm": 0.7937162920192259, + "learning_rate": 8.479461246550922e-06, + "loss": 0.152, + "step": 785 + }, + { + "epoch": 2.7749338040600176, + "grad_norm": 0.7164290770402717, + "learning_rate": 8.475349455300947e-06, + "loss": 0.1249, + "step": 786 + }, + { + "epoch": 2.7784642541924094, + "grad_norm": 0.7347153157580293, + "learning_rate": 8.471233111933291e-06, + "loss": 0.169, + "step": 787 + }, + { + "epoch": 2.7819947043248012, + "grad_norm": 0.6345954182380398, + "learning_rate": 8.467112221839664e-06, + "loss": 0.1476, + "step": 788 + }, + { + "epoch": 2.7855251544571935, + "grad_norm": 0.7710274326642901, + "learning_rate": 8.462986790417733e-06, + "loss": 0.1266, + "step": 789 + }, + { + "epoch": 2.789055604589585, + "grad_norm": 0.7730009077854115, + "learning_rate": 8.458856823071111e-06, + "loss": 0.1347, + "step": 790 + }, + { + "epoch": 2.792586054721977, + "grad_norm": 0.9320951665857266, + "learning_rate": 8.454722325209354e-06, + "loss": 0.1544, + "step": 791 + }, + { + "epoch": 2.796116504854369, + "grad_norm": 0.854024287144252, + "learning_rate": 8.450583302247954e-06, + "loss": 0.1776, + "step": 792 + }, + { + "epoch": 2.799646954986761, + "grad_norm": 0.8267045018857162, + "learning_rate": 8.44643975960833e-06, + "loss": 0.1805, + "step": 793 + }, + { + "epoch": 2.8031774051191527, + "grad_norm": 0.8689865555638648, + "learning_rate": 8.442291702717814e-06, + "loss": 0.1643, + "step": 794 + }, + { + "epoch": 2.8067078552515445, + "grad_norm": 0.78855645463223, + "learning_rate": 8.43813913700966e-06, + "loss": 0.1663, + "step": 795 + }, + { + "epoch": 2.8102383053839364, + "grad_norm": 0.8423324894257369, + "learning_rate": 8.433982067923021e-06, + "loss": 0.1539, + "step": 796 + }, + { + "epoch": 2.813768755516328, + "grad_norm": 0.6841465341626565, + "learning_rate": 8.429820500902957e-06, + "loss": 0.1707, + "step": 797 + }, + { + "epoch": 2.81729920564872, + "grad_norm": 0.637410884505517, + "learning_rate": 8.425654441400408e-06, + "loss": 0.1631, + "step": 798 + }, + { + "epoch": 2.820829655781112, + "grad_norm": 0.7435035501128142, + "learning_rate": 8.421483894872208e-06, + "loss": 0.1518, + "step": 799 + }, + { + "epoch": 2.824360105913504, + "grad_norm": 0.7390502021511453, + "learning_rate": 8.417308866781065e-06, + "loss": 0.1513, + "step": 800 + }, + { + "epoch": 2.827890556045896, + "grad_norm": 0.7074112942115228, + "learning_rate": 8.413129362595555e-06, + "loss": 0.1626, + "step": 801 + }, + { + "epoch": 2.831421006178288, + "grad_norm": 0.8297265070014697, + "learning_rate": 8.408945387790121e-06, + "loss": 0.1447, + "step": 802 + }, + { + "epoch": 2.8349514563106797, + "grad_norm": 0.6691649963937963, + "learning_rate": 8.404756947845056e-06, + "loss": 0.1453, + "step": 803 + }, + { + "epoch": 2.8384819064430715, + "grad_norm": 0.7275626975495146, + "learning_rate": 8.40056404824651e-06, + "loss": 0.162, + "step": 804 + }, + { + "epoch": 2.8420123565754634, + "grad_norm": 0.9804621444439445, + "learning_rate": 8.396366694486466e-06, + "loss": 0.1711, + "step": 805 + }, + { + "epoch": 2.845542806707855, + "grad_norm": 0.7841205320098567, + "learning_rate": 8.392164892062746e-06, + "loss": 0.1682, + "step": 806 + }, + { + "epoch": 2.849073256840247, + "grad_norm": 0.7259305687219134, + "learning_rate": 8.387958646478998e-06, + "loss": 0.1349, + "step": 807 + }, + { + "epoch": 2.852603706972639, + "grad_norm": 0.6285666615213146, + "learning_rate": 8.383747963244688e-06, + "loss": 0.1622, + "step": 808 + }, + { + "epoch": 2.856134157105031, + "grad_norm": 0.8191314538526839, + "learning_rate": 8.379532847875099e-06, + "loss": 0.1765, + "step": 809 + }, + { + "epoch": 2.8596646072374226, + "grad_norm": 0.7451704716031862, + "learning_rate": 8.375313305891312e-06, + "loss": 0.1451, + "step": 810 + }, + { + "epoch": 2.863195057369815, + "grad_norm": 0.6744145077220661, + "learning_rate": 8.371089342820216e-06, + "loss": 0.1643, + "step": 811 + }, + { + "epoch": 2.8667255075022067, + "grad_norm": 0.6887811045404645, + "learning_rate": 8.366860964194479e-06, + "loss": 0.1522, + "step": 812 + }, + { + "epoch": 2.8702559576345985, + "grad_norm": 0.63573464613807, + "learning_rate": 8.362628175552566e-06, + "loss": 0.1511, + "step": 813 + }, + { + "epoch": 2.8737864077669903, + "grad_norm": 0.6897759046360326, + "learning_rate": 8.358390982438706e-06, + "loss": 0.1434, + "step": 814 + }, + { + "epoch": 2.877316857899382, + "grad_norm": 0.7279048382426626, + "learning_rate": 8.354149390402904e-06, + "loss": 0.1351, + "step": 815 + }, + { + "epoch": 2.880847308031774, + "grad_norm": 0.7458563780785276, + "learning_rate": 8.349903405000926e-06, + "loss": 0.1517, + "step": 816 + }, + { + "epoch": 2.884377758164166, + "grad_norm": 0.7101890658117982, + "learning_rate": 8.345653031794292e-06, + "loss": 0.1462, + "step": 817 + }, + { + "epoch": 2.8879082082965577, + "grad_norm": 0.7742297467061107, + "learning_rate": 8.341398276350268e-06, + "loss": 0.1555, + "step": 818 + }, + { + "epoch": 2.8914386584289495, + "grad_norm": 0.7438278623504806, + "learning_rate": 8.337139144241864e-06, + "loss": 0.1396, + "step": 819 + }, + { + "epoch": 2.894969108561342, + "grad_norm": 0.724140366387057, + "learning_rate": 8.332875641047817e-06, + "loss": 0.1361, + "step": 820 + }, + { + "epoch": 2.898499558693733, + "grad_norm": 0.7091482359305543, + "learning_rate": 8.328607772352593e-06, + "loss": 0.1477, + "step": 821 + }, + { + "epoch": 2.9020300088261255, + "grad_norm": 0.714577592926291, + "learning_rate": 8.324335543746377e-06, + "loss": 0.155, + "step": 822 + }, + { + "epoch": 2.9055604589585173, + "grad_norm": 0.6821542598622823, + "learning_rate": 8.32005896082506e-06, + "loss": 0.1449, + "step": 823 + }, + { + "epoch": 2.909090909090909, + "grad_norm": 0.7674294440664721, + "learning_rate": 8.31577802919024e-06, + "loss": 0.162, + "step": 824 + }, + { + "epoch": 2.912621359223301, + "grad_norm": 0.6635918469316766, + "learning_rate": 8.311492754449213e-06, + "loss": 0.1445, + "step": 825 + }, + { + "epoch": 2.916151809355693, + "grad_norm": 0.7039005991376192, + "learning_rate": 8.307203142214959e-06, + "loss": 0.1473, + "step": 826 + }, + { + "epoch": 2.9196822594880847, + "grad_norm": 0.7210854795121673, + "learning_rate": 8.302909198106144e-06, + "loss": 0.1431, + "step": 827 + }, + { + "epoch": 2.9232127096204765, + "grad_norm": 0.8804727621267001, + "learning_rate": 8.298610927747104e-06, + "loss": 0.1514, + "step": 828 + }, + { + "epoch": 2.9267431597528684, + "grad_norm": 0.6567640224299616, + "learning_rate": 8.294308336767843e-06, + "loss": 0.1356, + "step": 829 + }, + { + "epoch": 2.93027360988526, + "grad_norm": 0.7266522819100054, + "learning_rate": 8.290001430804026e-06, + "loss": 0.1515, + "step": 830 + }, + { + "epoch": 2.9338040600176525, + "grad_norm": 0.7006333424868092, + "learning_rate": 8.285690215496966e-06, + "loss": 0.1518, + "step": 831 + }, + { + "epoch": 2.937334510150044, + "grad_norm": 1.0457753137928258, + "learning_rate": 8.281374696493628e-06, + "loss": 0.1623, + "step": 832 + }, + { + "epoch": 2.940864960282436, + "grad_norm": 0.7835519475042352, + "learning_rate": 8.277054879446603e-06, + "loss": 0.1512, + "step": 833 + }, + { + "epoch": 2.944395410414828, + "grad_norm": 0.694852963337788, + "learning_rate": 8.27273077001412e-06, + "loss": 0.1408, + "step": 834 + }, + { + "epoch": 2.94792586054722, + "grad_norm": 0.6914754193933184, + "learning_rate": 8.26840237386003e-06, + "loss": 0.1553, + "step": 835 + }, + { + "epoch": 2.9514563106796117, + "grad_norm": 0.7640107634127185, + "learning_rate": 8.264069696653795e-06, + "loss": 0.1563, + "step": 836 + }, + { + "epoch": 2.9549867608120035, + "grad_norm": 0.7031909948366248, + "learning_rate": 8.25973274407049e-06, + "loss": 0.1513, + "step": 837 + }, + { + "epoch": 2.9585172109443953, + "grad_norm": 0.7212903191170139, + "learning_rate": 8.255391521790782e-06, + "loss": 0.1452, + "step": 838 + }, + { + "epoch": 2.962047661076787, + "grad_norm": 0.9288624940870822, + "learning_rate": 8.251046035500937e-06, + "loss": 0.1611, + "step": 839 + }, + { + "epoch": 2.965578111209179, + "grad_norm": 0.8307453353673667, + "learning_rate": 8.246696290892804e-06, + "loss": 0.1511, + "step": 840 + }, + { + "epoch": 2.969108561341571, + "grad_norm": 0.7346059651517513, + "learning_rate": 8.24234229366381e-06, + "loss": 0.1645, + "step": 841 + }, + { + "epoch": 2.972639011473963, + "grad_norm": 0.8211385940924918, + "learning_rate": 8.237984049516952e-06, + "loss": 0.1576, + "step": 842 + }, + { + "epoch": 2.9761694616063545, + "grad_norm": 0.6365132434025104, + "learning_rate": 8.23362156416079e-06, + "loss": 0.1228, + "step": 843 + }, + { + "epoch": 2.979699911738747, + "grad_norm": 0.8824809696079167, + "learning_rate": 8.229254843309438e-06, + "loss": 0.1509, + "step": 844 + }, + { + "epoch": 2.9832303618711387, + "grad_norm": 0.8601855800826128, + "learning_rate": 8.22488389268256e-06, + "loss": 0.1615, + "step": 845 + }, + { + "epoch": 2.9867608120035305, + "grad_norm": 0.731123229093688, + "learning_rate": 8.220508718005359e-06, + "loss": 0.1378, + "step": 846 + }, + { + "epoch": 2.9902912621359223, + "grad_norm": 0.6227120430460796, + "learning_rate": 8.21612932500857e-06, + "loss": 0.1399, + "step": 847 + }, + { + "epoch": 2.993821712268314, + "grad_norm": 0.7545771750137871, + "learning_rate": 8.211745719428456e-06, + "loss": 0.1701, + "step": 848 + }, + { + "epoch": 2.997352162400706, + "grad_norm": 0.8046466791224096, + "learning_rate": 8.207357907006794e-06, + "loss": 0.1549, + "step": 849 + }, + { + "epoch": 3.000882612533098, + "grad_norm": 0.7642023797697144, + "learning_rate": 8.202965893490877e-06, + "loss": 0.1368, + "step": 850 + }, + { + "epoch": 3.0044130626654897, + "grad_norm": 0.6716017754172908, + "learning_rate": 8.198569684633492e-06, + "loss": 0.1162, + "step": 851 + }, + { + "epoch": 3.0079435127978815, + "grad_norm": 0.6431034665161904, + "learning_rate": 8.194169286192928e-06, + "loss": 0.1195, + "step": 852 + }, + { + "epoch": 3.011473962930274, + "grad_norm": 0.7143235899902817, + "learning_rate": 8.18976470393296e-06, + "loss": 0.112, + "step": 853 + }, + { + "epoch": 3.0150044130626656, + "grad_norm": 0.698231889557549, + "learning_rate": 8.185355943622845e-06, + "loss": 0.1234, + "step": 854 + }, + { + "epoch": 3.0185348631950575, + "grad_norm": 0.9000098025363069, + "learning_rate": 8.180943011037306e-06, + "loss": 0.1249, + "step": 855 + }, + { + "epoch": 3.0220653133274493, + "grad_norm": 0.8015546761444414, + "learning_rate": 8.176525911956539e-06, + "loss": 0.1245, + "step": 856 + }, + { + "epoch": 3.025595763459841, + "grad_norm": 0.8094451013988107, + "learning_rate": 8.172104652166194e-06, + "loss": 0.1018, + "step": 857 + }, + { + "epoch": 3.029126213592233, + "grad_norm": 0.720799250577229, + "learning_rate": 8.167679237457368e-06, + "loss": 0.0992, + "step": 858 + }, + { + "epoch": 3.032656663724625, + "grad_norm": 0.8000553496291071, + "learning_rate": 8.163249673626603e-06, + "loss": 0.1194, + "step": 859 + }, + { + "epoch": 3.0361871138570167, + "grad_norm": 0.8669635173424244, + "learning_rate": 8.158815966475877e-06, + "loss": 0.0816, + "step": 860 + }, + { + "epoch": 3.0397175639894085, + "grad_norm": 0.9025742229651821, + "learning_rate": 8.154378121812593e-06, + "loss": 0.1018, + "step": 861 + }, + { + "epoch": 3.0432480141218003, + "grad_norm": 0.9845870522804518, + "learning_rate": 8.149936145449575e-06, + "loss": 0.1251, + "step": 862 + }, + { + "epoch": 3.046778464254192, + "grad_norm": 1.0256068011823871, + "learning_rate": 8.145490043205056e-06, + "loss": 0.1239, + "step": 863 + }, + { + "epoch": 3.0503089143865845, + "grad_norm": 1.148156571965961, + "learning_rate": 8.141039820902676e-06, + "loss": 0.1104, + "step": 864 + }, + { + "epoch": 3.0538393645189763, + "grad_norm": 1.2193696303353918, + "learning_rate": 8.136585484371471e-06, + "loss": 0.1152, + "step": 865 + }, + { + "epoch": 3.057369814651368, + "grad_norm": 1.2915877393106403, + "learning_rate": 8.132127039445864e-06, + "loss": 0.118, + "step": 866 + }, + { + "epoch": 3.06090026478376, + "grad_norm": 1.2149067405554672, + "learning_rate": 8.127664491965662e-06, + "loss": 0.1142, + "step": 867 + }, + { + "epoch": 3.064430714916152, + "grad_norm": 0.990566681159559, + "learning_rate": 8.123197847776043e-06, + "loss": 0.1078, + "step": 868 + }, + { + "epoch": 3.0679611650485437, + "grad_norm": 0.9565804354564945, + "learning_rate": 8.118727112727556e-06, + "loss": 0.1135, + "step": 869 + }, + { + "epoch": 3.0714916151809355, + "grad_norm": 1.0200728098865612, + "learning_rate": 8.1142522926761e-06, + "loss": 0.1195, + "step": 870 + }, + { + "epoch": 3.0750220653133273, + "grad_norm": 1.0243439506618488, + "learning_rate": 8.109773393482934e-06, + "loss": 0.1024, + "step": 871 + }, + { + "epoch": 3.078552515445719, + "grad_norm": 1.0049145977279004, + "learning_rate": 8.105290421014653e-06, + "loss": 0.1076, + "step": 872 + }, + { + "epoch": 3.082082965578111, + "grad_norm": 0.8112529372591591, + "learning_rate": 8.10080338114319e-06, + "loss": 0.0982, + "step": 873 + }, + { + "epoch": 3.0856134157105033, + "grad_norm": 1.0002477688608336, + "learning_rate": 8.096312279745804e-06, + "loss": 0.13, + "step": 874 + }, + { + "epoch": 3.089143865842895, + "grad_norm": 1.0623544903602675, + "learning_rate": 8.09181712270508e-06, + "loss": 0.1025, + "step": 875 + }, + { + "epoch": 3.092674315975287, + "grad_norm": 1.0410125624967994, + "learning_rate": 8.087317915908904e-06, + "loss": 0.1063, + "step": 876 + }, + { + "epoch": 3.096204766107679, + "grad_norm": 1.0118590449093119, + "learning_rate": 8.082814665250476e-06, + "loss": 0.0954, + "step": 877 + }, + { + "epoch": 3.0997352162400706, + "grad_norm": 0.9546611695399589, + "learning_rate": 8.078307376628292e-06, + "loss": 0.1245, + "step": 878 + }, + { + "epoch": 3.1032656663724625, + "grad_norm": 0.9269391315446083, + "learning_rate": 8.073796055946133e-06, + "loss": 0.1174, + "step": 879 + }, + { + "epoch": 3.1067961165048543, + "grad_norm": 1.0305117514176123, + "learning_rate": 8.06928070911306e-06, + "loss": 0.1026, + "step": 880 + }, + { + "epoch": 3.110326566637246, + "grad_norm": 1.208839666085655, + "learning_rate": 8.064761342043415e-06, + "loss": 0.1248, + "step": 881 + }, + { + "epoch": 3.113857016769638, + "grad_norm": 0.952513604276628, + "learning_rate": 8.060237960656799e-06, + "loss": 0.1085, + "step": 882 + }, + { + "epoch": 3.11738746690203, + "grad_norm": 0.8730540650291831, + "learning_rate": 8.055710570878076e-06, + "loss": 0.1134, + "step": 883 + }, + { + "epoch": 3.120917917034422, + "grad_norm": 1.006482604780559, + "learning_rate": 8.051179178637356e-06, + "loss": 0.1049, + "step": 884 + }, + { + "epoch": 3.124448367166814, + "grad_norm": 0.9112291303575337, + "learning_rate": 8.046643789869991e-06, + "loss": 0.108, + "step": 885 + }, + { + "epoch": 3.127978817299206, + "grad_norm": 1.0693444544873563, + "learning_rate": 8.042104410516576e-06, + "loss": 0.1313, + "step": 886 + }, + { + "epoch": 3.1315092674315976, + "grad_norm": 1.0053001607105967, + "learning_rate": 8.037561046522923e-06, + "loss": 0.1142, + "step": 887 + }, + { + "epoch": 3.1350397175639895, + "grad_norm": 0.9933955602184302, + "learning_rate": 8.033013703840067e-06, + "loss": 0.1278, + "step": 888 + }, + { + "epoch": 3.1385701676963813, + "grad_norm": 1.0821161416075462, + "learning_rate": 8.028462388424258e-06, + "loss": 0.1213, + "step": 889 + }, + { + "epoch": 3.142100617828773, + "grad_norm": 1.0448661581558534, + "learning_rate": 8.023907106236944e-06, + "loss": 0.1016, + "step": 890 + }, + { + "epoch": 3.145631067961165, + "grad_norm": 1.1354463844063831, + "learning_rate": 8.019347863244773e-06, + "loss": 0.1198, + "step": 891 + }, + { + "epoch": 3.149161518093557, + "grad_norm": 1.008070141194459, + "learning_rate": 8.014784665419577e-06, + "loss": 0.1252, + "step": 892 + }, + { + "epoch": 3.1526919682259487, + "grad_norm": 1.06009230334142, + "learning_rate": 8.01021751873837e-06, + "loss": 0.1286, + "step": 893 + }, + { + "epoch": 3.1562224183583405, + "grad_norm": 1.138468713844925, + "learning_rate": 8.005646429183342e-06, + "loss": 0.1129, + "step": 894 + }, + { + "epoch": 3.159752868490733, + "grad_norm": 1.2328861278363776, + "learning_rate": 8.001071402741843e-06, + "loss": 0.1235, + "step": 895 + }, + { + "epoch": 3.1632833186231246, + "grad_norm": 1.0853132184834922, + "learning_rate": 7.996492445406377e-06, + "loss": 0.1282, + "step": 896 + }, + { + "epoch": 3.1668137687555165, + "grad_norm": 0.933433925231955, + "learning_rate": 7.991909563174605e-06, + "loss": 0.1325, + "step": 897 + }, + { + "epoch": 3.1703442188879083, + "grad_norm": 0.8319438313846836, + "learning_rate": 7.987322762049323e-06, + "loss": 0.0987, + "step": 898 + }, + { + "epoch": 3.1738746690203, + "grad_norm": 0.9872815349839481, + "learning_rate": 7.982732048038462e-06, + "loss": 0.1195, + "step": 899 + }, + { + "epoch": 3.177405119152692, + "grad_norm": 0.9612402010886701, + "learning_rate": 7.978137427155078e-06, + "loss": 0.1258, + "step": 900 + }, + { + "epoch": 3.180935569285084, + "grad_norm": 0.8927263036280142, + "learning_rate": 7.973538905417344e-06, + "loss": 0.1072, + "step": 901 + }, + { + "epoch": 3.1844660194174756, + "grad_norm": 0.8841286752626785, + "learning_rate": 7.968936488848541e-06, + "loss": 0.1047, + "step": 902 + }, + { + "epoch": 3.1879964695498675, + "grad_norm": 0.912445757407591, + "learning_rate": 7.964330183477057e-06, + "loss": 0.1276, + "step": 903 + }, + { + "epoch": 3.1915269196822593, + "grad_norm": 1.059803669034267, + "learning_rate": 7.959719995336364e-06, + "loss": 0.1032, + "step": 904 + }, + { + "epoch": 3.195057369814651, + "grad_norm": 0.9513868377440023, + "learning_rate": 7.955105930465032e-06, + "loss": 0.0911, + "step": 905 + }, + { + "epoch": 3.1985878199470434, + "grad_norm": 0.8797745669418451, + "learning_rate": 7.9504879949067e-06, + "loss": 0.1143, + "step": 906 + }, + { + "epoch": 3.2021182700794353, + "grad_norm": 1.0029598114248315, + "learning_rate": 7.945866194710079e-06, + "loss": 0.1161, + "step": 907 + }, + { + "epoch": 3.205648720211827, + "grad_norm": 1.1319418617063524, + "learning_rate": 7.941240535928942e-06, + "loss": 0.1268, + "step": 908 + }, + { + "epoch": 3.209179170344219, + "grad_norm": 0.9958247294998045, + "learning_rate": 7.936611024622117e-06, + "loss": 0.1074, + "step": 909 + }, + { + "epoch": 3.212709620476611, + "grad_norm": 1.0922819941075517, + "learning_rate": 7.93197766685348e-06, + "loss": 0.1179, + "step": 910 + }, + { + "epoch": 3.2162400706090026, + "grad_norm": 1.136941227912527, + "learning_rate": 7.92734046869194e-06, + "loss": 0.1331, + "step": 911 + }, + { + "epoch": 3.2197705207413945, + "grad_norm": 1.0047937727286098, + "learning_rate": 7.922699436211438e-06, + "loss": 0.1104, + "step": 912 + }, + { + "epoch": 3.2233009708737863, + "grad_norm": 0.9964788630911344, + "learning_rate": 7.918054575490943e-06, + "loss": 0.1252, + "step": 913 + }, + { + "epoch": 3.226831421006178, + "grad_norm": 1.0926246384683764, + "learning_rate": 7.913405892614433e-06, + "loss": 0.1044, + "step": 914 + }, + { + "epoch": 3.23036187113857, + "grad_norm": 0.9688331888143029, + "learning_rate": 7.908753393670891e-06, + "loss": 0.0945, + "step": 915 + }, + { + "epoch": 3.233892321270962, + "grad_norm": 1.111353673933071, + "learning_rate": 7.904097084754301e-06, + "loss": 0.1148, + "step": 916 + }, + { + "epoch": 3.237422771403354, + "grad_norm": 1.092326321596579, + "learning_rate": 7.899436971963637e-06, + "loss": 0.1276, + "step": 917 + }, + { + "epoch": 3.240953221535746, + "grad_norm": 0.7874168063997805, + "learning_rate": 7.894773061402855e-06, + "loss": 0.0932, + "step": 918 + }, + { + "epoch": 3.244483671668138, + "grad_norm": 0.9597684028837579, + "learning_rate": 7.89010535918089e-06, + "loss": 0.1158, + "step": 919 + }, + { + "epoch": 3.2480141218005296, + "grad_norm": 1.0816606744001607, + "learning_rate": 7.885433871411634e-06, + "loss": 0.1204, + "step": 920 + }, + { + "epoch": 3.2515445719329215, + "grad_norm": 1.035099273475842, + "learning_rate": 7.880758604213945e-06, + "loss": 0.125, + "step": 921 + }, + { + "epoch": 3.2550750220653133, + "grad_norm": 0.8982660262483314, + "learning_rate": 7.876079563711631e-06, + "loss": 0.104, + "step": 922 + }, + { + "epoch": 3.258605472197705, + "grad_norm": 1.0276050394216014, + "learning_rate": 7.871396756033438e-06, + "loss": 0.119, + "step": 923 + }, + { + "epoch": 3.262135922330097, + "grad_norm": 1.0128946832960732, + "learning_rate": 7.866710187313048e-06, + "loss": 0.118, + "step": 924 + }, + { + "epoch": 3.265666372462489, + "grad_norm": 0.8704316103506846, + "learning_rate": 7.862019863689075e-06, + "loss": 0.1139, + "step": 925 + }, + { + "epoch": 3.269196822594881, + "grad_norm": 0.9671125371629974, + "learning_rate": 7.857325791305042e-06, + "loss": 0.137, + "step": 926 + }, + { + "epoch": 3.2727272727272725, + "grad_norm": 0.868940825788469, + "learning_rate": 7.852627976309388e-06, + "loss": 0.1124, + "step": 927 + }, + { + "epoch": 3.2762577228596648, + "grad_norm": 0.9630053987036548, + "learning_rate": 7.847926424855454e-06, + "loss": 0.1253, + "step": 928 + }, + { + "epoch": 3.2797881729920566, + "grad_norm": 1.0029586796082006, + "learning_rate": 7.843221143101474e-06, + "loss": 0.1137, + "step": 929 + }, + { + "epoch": 3.2833186231244484, + "grad_norm": 0.8820085770728214, + "learning_rate": 7.838512137210565e-06, + "loss": 0.1255, + "step": 930 + }, + { + "epoch": 3.2868490732568403, + "grad_norm": 0.9828634895838252, + "learning_rate": 7.833799413350732e-06, + "loss": 0.1035, + "step": 931 + }, + { + "epoch": 3.290379523389232, + "grad_norm": 1.319511988574606, + "learning_rate": 7.829082977694835e-06, + "loss": 0.1325, + "step": 932 + }, + { + "epoch": 3.293909973521624, + "grad_norm": 1.2144177327586667, + "learning_rate": 7.824362836420609e-06, + "loss": 0.1403, + "step": 933 + }, + { + "epoch": 3.297440423654016, + "grad_norm": 1.0719443808897844, + "learning_rate": 7.819638995710636e-06, + "loss": 0.1185, + "step": 934 + }, + { + "epoch": 3.3009708737864076, + "grad_norm": 1.0518286795857918, + "learning_rate": 7.814911461752346e-06, + "loss": 0.1065, + "step": 935 + }, + { + "epoch": 3.3045013239187995, + "grad_norm": 1.024493911176814, + "learning_rate": 7.810180240738004e-06, + "loss": 0.1151, + "step": 936 + }, + { + "epoch": 3.3080317740511918, + "grad_norm": 1.127575182515073, + "learning_rate": 7.805445338864709e-06, + "loss": 0.1308, + "step": 937 + }, + { + "epoch": 3.3115622241835836, + "grad_norm": 0.9456157508829596, + "learning_rate": 7.800706762334375e-06, + "loss": 0.107, + "step": 938 + }, + { + "epoch": 3.3150926743159754, + "grad_norm": 0.9074545977721913, + "learning_rate": 7.795964517353734e-06, + "loss": 0.1138, + "step": 939 + }, + { + "epoch": 3.3186231244483673, + "grad_norm": 1.094922914297947, + "learning_rate": 7.791218610134324e-06, + "loss": 0.1349, + "step": 940 + }, + { + "epoch": 3.322153574580759, + "grad_norm": 1.133601456805787, + "learning_rate": 7.786469046892475e-06, + "loss": 0.1069, + "step": 941 + }, + { + "epoch": 3.325684024713151, + "grad_norm": 1.0182500924518352, + "learning_rate": 7.781715833849311e-06, + "loss": 0.1154, + "step": 942 + }, + { + "epoch": 3.329214474845543, + "grad_norm": 1.046201515291234, + "learning_rate": 7.776958977230731e-06, + "loss": 0.1251, + "step": 943 + }, + { + "epoch": 3.3327449249779346, + "grad_norm": 0.957386976461524, + "learning_rate": 7.77219848326741e-06, + "loss": 0.1101, + "step": 944 + }, + { + "epoch": 3.3362753751103265, + "grad_norm": 1.1740948292667477, + "learning_rate": 7.76743435819479e-06, + "loss": 0.1443, + "step": 945 + }, + { + "epoch": 3.3398058252427183, + "grad_norm": 1.850891607235183, + "learning_rate": 7.76266660825306e-06, + "loss": 0.1018, + "step": 946 + }, + { + "epoch": 3.34333627537511, + "grad_norm": 0.9097207671724479, + "learning_rate": 7.757895239687168e-06, + "loss": 0.1041, + "step": 947 + }, + { + "epoch": 3.3468667255075024, + "grad_norm": 0.9016475882306597, + "learning_rate": 7.753120258746793e-06, + "loss": 0.1151, + "step": 948 + }, + { + "epoch": 3.3503971756398943, + "grad_norm": 1.399326889021019, + "learning_rate": 7.748341671686355e-06, + "loss": 0.1361, + "step": 949 + }, + { + "epoch": 3.353927625772286, + "grad_norm": 0.8958929274360947, + "learning_rate": 7.743559484764987e-06, + "loss": 0.109, + "step": 950 + }, + { + "epoch": 3.357458075904678, + "grad_norm": 0.9661628430088748, + "learning_rate": 7.73877370424654e-06, + "loss": 0.1184, + "step": 951 + }, + { + "epoch": 3.3609885260370698, + "grad_norm": 1.0856106807644708, + "learning_rate": 7.733984336399578e-06, + "loss": 0.1258, + "step": 952 + }, + { + "epoch": 3.3645189761694616, + "grad_norm": 1.1339299007520192, + "learning_rate": 7.729191387497357e-06, + "loss": 0.1313, + "step": 953 + }, + { + "epoch": 3.3680494263018534, + "grad_norm": 0.9462346995224554, + "learning_rate": 7.72439486381783e-06, + "loss": 0.1299, + "step": 954 + }, + { + "epoch": 3.3715798764342453, + "grad_norm": 0.9773900040503245, + "learning_rate": 7.719594771643623e-06, + "loss": 0.115, + "step": 955 + }, + { + "epoch": 3.375110326566637, + "grad_norm": 0.8812615869761742, + "learning_rate": 7.714791117262047e-06, + "loss": 0.1178, + "step": 956 + }, + { + "epoch": 3.378640776699029, + "grad_norm": 0.872321741116879, + "learning_rate": 7.709983906965071e-06, + "loss": 0.1177, + "step": 957 + }, + { + "epoch": 3.382171226831421, + "grad_norm": 0.9590632048655139, + "learning_rate": 7.705173147049326e-06, + "loss": 0.1209, + "step": 958 + }, + { + "epoch": 3.385701676963813, + "grad_norm": 0.9774010811015693, + "learning_rate": 7.700358843816087e-06, + "loss": 0.0988, + "step": 959 + }, + { + "epoch": 3.389232127096205, + "grad_norm": 1.05084415330861, + "learning_rate": 7.695541003571275e-06, + "loss": 0.1281, + "step": 960 + }, + { + "epoch": 3.3927625772285968, + "grad_norm": 0.9815810438992207, + "learning_rate": 7.690719632625447e-06, + "loss": 0.1127, + "step": 961 + }, + { + "epoch": 3.3962930273609886, + "grad_norm": 0.9979507342808466, + "learning_rate": 7.685894737293773e-06, + "loss": 0.1224, + "step": 962 + }, + { + "epoch": 3.3998234774933804, + "grad_norm": 1.0098708954426872, + "learning_rate": 7.681066323896053e-06, + "loss": 0.1418, + "step": 963 + }, + { + "epoch": 3.4033539276257723, + "grad_norm": 0.8683560303898923, + "learning_rate": 7.676234398756688e-06, + "loss": 0.11, + "step": 964 + }, + { + "epoch": 3.406884377758164, + "grad_norm": 0.9354248879782321, + "learning_rate": 7.671398968204675e-06, + "loss": 0.1125, + "step": 965 + }, + { + "epoch": 3.410414827890556, + "grad_norm": 0.9688499014186751, + "learning_rate": 7.66656003857361e-06, + "loss": 0.1233, + "step": 966 + }, + { + "epoch": 3.413945278022948, + "grad_norm": 1.0624016948474533, + "learning_rate": 7.66171761620167e-06, + "loss": 0.1051, + "step": 967 + }, + { + "epoch": 3.4174757281553396, + "grad_norm": 0.9463189432354194, + "learning_rate": 7.656871707431604e-06, + "loss": 0.1171, + "step": 968 + }, + { + "epoch": 3.4210061782877315, + "grad_norm": 0.9894483338890546, + "learning_rate": 7.65202231861073e-06, + "loss": 0.1108, + "step": 969 + }, + { + "epoch": 3.4245366284201237, + "grad_norm": 0.9687362431205614, + "learning_rate": 7.647169456090925e-06, + "loss": 0.1116, + "step": 970 + }, + { + "epoch": 3.4280670785525156, + "grad_norm": 0.9557897471038936, + "learning_rate": 7.642313126228618e-06, + "loss": 0.1153, + "step": 971 + }, + { + "epoch": 3.4315975286849074, + "grad_norm": 1.0869572132117586, + "learning_rate": 7.63745333538477e-06, + "loss": 0.1236, + "step": 972 + }, + { + "epoch": 3.4351279788172993, + "grad_norm": 0.9605589326914042, + "learning_rate": 7.632590089924885e-06, + "loss": 0.1065, + "step": 973 + }, + { + "epoch": 3.438658428949691, + "grad_norm": 1.0323514330728734, + "learning_rate": 7.627723396218988e-06, + "loss": 0.1175, + "step": 974 + }, + { + "epoch": 3.442188879082083, + "grad_norm": 1.0118187975842183, + "learning_rate": 7.622853260641623e-06, + "loss": 0.1186, + "step": 975 + }, + { + "epoch": 3.4457193292144748, + "grad_norm": 1.024656799596295, + "learning_rate": 7.61797968957184e-06, + "loss": 0.1271, + "step": 976 + }, + { + "epoch": 3.4492497793468666, + "grad_norm": 0.8727047804672294, + "learning_rate": 7.613102689393188e-06, + "loss": 0.1144, + "step": 977 + }, + { + "epoch": 3.4527802294792584, + "grad_norm": 0.8394841490325925, + "learning_rate": 7.608222266493712e-06, + "loss": 0.1123, + "step": 978 + }, + { + "epoch": 3.4563106796116507, + "grad_norm": 1.1351511857664562, + "learning_rate": 7.603338427265935e-06, + "loss": 0.1012, + "step": 979 + }, + { + "epoch": 3.459841129744042, + "grad_norm": 1.128335444374776, + "learning_rate": 7.598451178106857e-06, + "loss": 0.1206, + "step": 980 + }, + { + "epoch": 3.4633715798764344, + "grad_norm": 0.8576208704238955, + "learning_rate": 7.593560525417947e-06, + "loss": 0.103, + "step": 981 + }, + { + "epoch": 3.4669020300088262, + "grad_norm": 1.1074290695297295, + "learning_rate": 7.588666475605127e-06, + "loss": 0.1254, + "step": 982 + }, + { + "epoch": 3.470432480141218, + "grad_norm": 1.1040068892498724, + "learning_rate": 7.583769035078772e-06, + "loss": 0.1143, + "step": 983 + }, + { + "epoch": 3.47396293027361, + "grad_norm": 1.0664311386837333, + "learning_rate": 7.578868210253698e-06, + "loss": 0.1353, + "step": 984 + }, + { + "epoch": 3.4774933804060018, + "grad_norm": 1.059615015944644, + "learning_rate": 7.5739640075491546e-06, + "loss": 0.1386, + "step": 985 + }, + { + "epoch": 3.4810238305383936, + "grad_norm": 0.9899439186497315, + "learning_rate": 7.569056433388813e-06, + "loss": 0.1258, + "step": 986 + }, + { + "epoch": 3.4845542806707854, + "grad_norm": 0.9224864093833172, + "learning_rate": 7.5641454942007606e-06, + "loss": 0.1225, + "step": 987 + }, + { + "epoch": 3.4880847308031773, + "grad_norm": 0.887546649860455, + "learning_rate": 7.559231196417495e-06, + "loss": 0.1249, + "step": 988 + }, + { + "epoch": 3.491615180935569, + "grad_norm": 0.9660667346820182, + "learning_rate": 7.554313546475915e-06, + "loss": 0.1163, + "step": 989 + }, + { + "epoch": 3.4951456310679614, + "grad_norm": 0.9707962317309367, + "learning_rate": 7.5493925508173025e-06, + "loss": 0.1321, + "step": 990 + }, + { + "epoch": 3.4986760812003532, + "grad_norm": 1.1180387890566459, + "learning_rate": 7.5444682158873305e-06, + "loss": 0.1221, + "step": 991 + }, + { + "epoch": 3.502206531332745, + "grad_norm": 0.8799016210637428, + "learning_rate": 7.539540548136037e-06, + "loss": 0.1371, + "step": 992 + }, + { + "epoch": 3.505736981465137, + "grad_norm": 1.0482875409664072, + "learning_rate": 7.534609554017833e-06, + "loss": 0.1175, + "step": 993 + }, + { + "epoch": 3.5092674315975287, + "grad_norm": 0.8991365777316297, + "learning_rate": 7.529675239991483e-06, + "loss": 0.1293, + "step": 994 + }, + { + "epoch": 3.5127978817299206, + "grad_norm": 1.0386653386530387, + "learning_rate": 7.524737612520101e-06, + "loss": 0.1229, + "step": 995 + }, + { + "epoch": 3.5163283318623124, + "grad_norm": 1.0711541879063016, + "learning_rate": 7.5197966780711385e-06, + "loss": 0.1201, + "step": 996 + }, + { + "epoch": 3.5198587819947043, + "grad_norm": 1.0318173451390964, + "learning_rate": 7.5148524431163825e-06, + "loss": 0.1316, + "step": 997 + }, + { + "epoch": 3.523389232127096, + "grad_norm": 0.984403787758931, + "learning_rate": 7.509904914131941e-06, + "loss": 0.1197, + "step": 998 + }, + { + "epoch": 3.526919682259488, + "grad_norm": 0.8522121035569924, + "learning_rate": 7.504954097598238e-06, + "loss": 0.1061, + "step": 999 + }, + { + "epoch": 3.5304501323918798, + "grad_norm": 0.849756117125948, + "learning_rate": 7.500000000000001e-06, + "loss": 0.1067, + "step": 1000 + }, + { + "epoch": 3.533980582524272, + "grad_norm": 0.7751208269103894, + "learning_rate": 7.495042627826257e-06, + "loss": 0.1091, + "step": 1001 + }, + { + "epoch": 3.537511032656664, + "grad_norm": 0.9495031542830646, + "learning_rate": 7.490081987570321e-06, + "loss": 0.121, + "step": 1002 + }, + { + "epoch": 3.5410414827890557, + "grad_norm": 1.0312096390159498, + "learning_rate": 7.48511808572979e-06, + "loss": 0.1304, + "step": 1003 + }, + { + "epoch": 3.5445719329214476, + "grad_norm": 1.054899906830368, + "learning_rate": 7.480150928806534e-06, + "loss": 0.1361, + "step": 1004 + }, + { + "epoch": 3.5481023830538394, + "grad_norm": 0.979527615798883, + "learning_rate": 7.475180523306684e-06, + "loss": 0.1277, + "step": 1005 + }, + { + "epoch": 3.5516328331862312, + "grad_norm": 1.0015458182337016, + "learning_rate": 7.470206875740627e-06, + "loss": 0.1149, + "step": 1006 + }, + { + "epoch": 3.555163283318623, + "grad_norm": 1.0755807694863864, + "learning_rate": 7.465229992622996e-06, + "loss": 0.1178, + "step": 1007 + }, + { + "epoch": 3.558693733451015, + "grad_norm": 1.0077743388921196, + "learning_rate": 7.4602498804726646e-06, + "loss": 0.1267, + "step": 1008 + }, + { + "epoch": 3.5622241835834068, + "grad_norm": 0.9760852883342952, + "learning_rate": 7.455266545812733e-06, + "loss": 0.1144, + "step": 1009 + }, + { + "epoch": 3.565754633715799, + "grad_norm": 0.9599179886281429, + "learning_rate": 7.450279995170524e-06, + "loss": 0.1233, + "step": 1010 + }, + { + "epoch": 3.5692850838481904, + "grad_norm": 1.0506465130972065, + "learning_rate": 7.445290235077571e-06, + "loss": 0.1223, + "step": 1011 + }, + { + "epoch": 3.5728155339805827, + "grad_norm": 0.9904879926238966, + "learning_rate": 7.440297272069615e-06, + "loss": 0.1103, + "step": 1012 + }, + { + "epoch": 3.5763459841129746, + "grad_norm": 0.9373307300830205, + "learning_rate": 7.435301112686588e-06, + "loss": 0.1089, + "step": 1013 + }, + { + "epoch": 3.5798764342453664, + "grad_norm": 0.9642807996392302, + "learning_rate": 7.430301763472612e-06, + "loss": 0.1397, + "step": 1014 + }, + { + "epoch": 3.5834068843777582, + "grad_norm": 1.0798478829987526, + "learning_rate": 7.425299230975982e-06, + "loss": 0.1271, + "step": 1015 + }, + { + "epoch": 3.58693733451015, + "grad_norm": 0.9757898668217716, + "learning_rate": 7.420293521749169e-06, + "loss": 0.1254, + "step": 1016 + }, + { + "epoch": 3.590467784642542, + "grad_norm": 0.7957246885082836, + "learning_rate": 7.415284642348801e-06, + "loss": 0.1024, + "step": 1017 + }, + { + "epoch": 3.5939982347749337, + "grad_norm": 1.0499954601099537, + "learning_rate": 7.410272599335662e-06, + "loss": 0.1104, + "step": 1018 + }, + { + "epoch": 3.5975286849073256, + "grad_norm": 1.017468888424547, + "learning_rate": 7.4052573992746735e-06, + "loss": 0.1329, + "step": 1019 + }, + { + "epoch": 3.6010591350397174, + "grad_norm": 0.9050306794358908, + "learning_rate": 7.400239048734899e-06, + "loss": 0.1001, + "step": 1020 + }, + { + "epoch": 3.6045895851721097, + "grad_norm": 0.9892172267357415, + "learning_rate": 7.395217554289524e-06, + "loss": 0.1116, + "step": 1021 + }, + { + "epoch": 3.608120035304501, + "grad_norm": 0.825558862940724, + "learning_rate": 7.390192922515855e-06, + "loss": 0.1178, + "step": 1022 + }, + { + "epoch": 3.6116504854368934, + "grad_norm": 1.0870666189939848, + "learning_rate": 7.385165159995306e-06, + "loss": 0.1188, + "step": 1023 + }, + { + "epoch": 3.615180935569285, + "grad_norm": 0.9461150823847166, + "learning_rate": 7.38013427331339e-06, + "loss": 0.1387, + "step": 1024 + }, + { + "epoch": 3.618711385701677, + "grad_norm": 0.8931705317958296, + "learning_rate": 7.3751002690597175e-06, + "loss": 0.1145, + "step": 1025 + }, + { + "epoch": 3.622241835834069, + "grad_norm": 1.0044315340520118, + "learning_rate": 7.370063153827978e-06, + "loss": 0.1311, + "step": 1026 + }, + { + "epoch": 3.6257722859664607, + "grad_norm": 0.9366772554201618, + "learning_rate": 7.365022934215935e-06, + "loss": 0.1347, + "step": 1027 + }, + { + "epoch": 3.6293027360988526, + "grad_norm": 1.0287287282798234, + "learning_rate": 7.3599796168254256e-06, + "loss": 0.12, + "step": 1028 + }, + { + "epoch": 3.6328331862312444, + "grad_norm": 1.2655633592932094, + "learning_rate": 7.3549332082623336e-06, + "loss": 0.1225, + "step": 1029 + }, + { + "epoch": 3.6363636363636362, + "grad_norm": 0.962932765361898, + "learning_rate": 7.349883715136601e-06, + "loss": 0.1163, + "step": 1030 + }, + { + "epoch": 3.639894086496028, + "grad_norm": 1.1898739705783477, + "learning_rate": 7.344831144062204e-06, + "loss": 0.1342, + "step": 1031 + }, + { + "epoch": 3.6434245366284204, + "grad_norm": 0.9775960634954581, + "learning_rate": 7.339775501657156e-06, + "loss": 0.1159, + "step": 1032 + }, + { + "epoch": 3.6469549867608118, + "grad_norm": 0.9985532411205827, + "learning_rate": 7.334716794543487e-06, + "loss": 0.1215, + "step": 1033 + }, + { + "epoch": 3.650485436893204, + "grad_norm": 0.8371099521051029, + "learning_rate": 7.329655029347246e-06, + "loss": 0.133, + "step": 1034 + }, + { + "epoch": 3.654015887025596, + "grad_norm": 0.8282699034669408, + "learning_rate": 7.324590212698486e-06, + "loss": 0.1128, + "step": 1035 + }, + { + "epoch": 3.6575463371579877, + "grad_norm": 0.8892127258122363, + "learning_rate": 7.319522351231256e-06, + "loss": 0.1193, + "step": 1036 + }, + { + "epoch": 3.6610767872903796, + "grad_norm": 1.0611463535886227, + "learning_rate": 7.314451451583593e-06, + "loss": 0.1371, + "step": 1037 + }, + { + "epoch": 3.6646072374227714, + "grad_norm": 0.834905111221197, + "learning_rate": 7.309377520397517e-06, + "loss": 0.1198, + "step": 1038 + }, + { + "epoch": 3.6681376875551632, + "grad_norm": 0.8795916947723678, + "learning_rate": 7.304300564319013e-06, + "loss": 0.1154, + "step": 1039 + }, + { + "epoch": 3.671668137687555, + "grad_norm": 0.932134034033176, + "learning_rate": 7.299220589998037e-06, + "loss": 0.1172, + "step": 1040 + }, + { + "epoch": 3.675198587819947, + "grad_norm": 1.060320989726038, + "learning_rate": 7.294137604088487e-06, + "loss": 0.1272, + "step": 1041 + }, + { + "epoch": 3.6787290379523387, + "grad_norm": 1.0865767224751157, + "learning_rate": 7.2890516132482145e-06, + "loss": 0.1174, + "step": 1042 + }, + { + "epoch": 3.682259488084731, + "grad_norm": 0.9524420011955063, + "learning_rate": 7.283962624139004e-06, + "loss": 0.1025, + "step": 1043 + }, + { + "epoch": 3.6857899382171224, + "grad_norm": 1.0723266113080743, + "learning_rate": 7.278870643426565e-06, + "loss": 0.1188, + "step": 1044 + }, + { + "epoch": 3.6893203883495147, + "grad_norm": 0.9889202356506911, + "learning_rate": 7.27377567778053e-06, + "loss": 0.1171, + "step": 1045 + }, + { + "epoch": 3.6928508384819065, + "grad_norm": 1.1387504413356822, + "learning_rate": 7.268677733874438e-06, + "loss": 0.1347, + "step": 1046 + }, + { + "epoch": 3.6963812886142984, + "grad_norm": 0.9054547213610565, + "learning_rate": 7.263576818385728e-06, + "loss": 0.1192, + "step": 1047 + }, + { + "epoch": 3.69991173874669, + "grad_norm": 0.9389240088652642, + "learning_rate": 7.258472937995736e-06, + "loss": 0.105, + "step": 1048 + }, + { + "epoch": 3.703442188879082, + "grad_norm": 1.0097808039069447, + "learning_rate": 7.253366099389678e-06, + "loss": 0.1283, + "step": 1049 + }, + { + "epoch": 3.706972639011474, + "grad_norm": 0.9312349763358698, + "learning_rate": 7.2482563092566446e-06, + "loss": 0.1177, + "step": 1050 + }, + { + "epoch": 3.7105030891438657, + "grad_norm": 0.8462385685589844, + "learning_rate": 7.243143574289595e-06, + "loss": 0.1176, + "step": 1051 + }, + { + "epoch": 3.7140335392762576, + "grad_norm": 0.9999611373105524, + "learning_rate": 7.23802790118534e-06, + "loss": 0.12, + "step": 1052 + }, + { + "epoch": 3.7175639894086494, + "grad_norm": 1.029486778996518, + "learning_rate": 7.232909296644548e-06, + "loss": 0.1313, + "step": 1053 + }, + { + "epoch": 3.7210944395410417, + "grad_norm": 0.9598825118610063, + "learning_rate": 7.227787767371717e-06, + "loss": 0.1057, + "step": 1054 + }, + { + "epoch": 3.7246248896734335, + "grad_norm": 0.8914389737346524, + "learning_rate": 7.222663320075185e-06, + "loss": 0.1196, + "step": 1055 + }, + { + "epoch": 3.7281553398058254, + "grad_norm": 0.9459232821476742, + "learning_rate": 7.217535961467105e-06, + "loss": 0.1254, + "step": 1056 + }, + { + "epoch": 3.731685789938217, + "grad_norm": 1.121524174912343, + "learning_rate": 7.212405698263446e-06, + "loss": 0.1296, + "step": 1057 + }, + { + "epoch": 3.735216240070609, + "grad_norm": 1.0305189372866497, + "learning_rate": 7.2072725371839826e-06, + "loss": 0.1123, + "step": 1058 + }, + { + "epoch": 3.738746690203001, + "grad_norm": 1.0095800993562205, + "learning_rate": 7.202136484952286e-06, + "loss": 0.1109, + "step": 1059 + }, + { + "epoch": 3.7422771403353927, + "grad_norm": 1.091036374867166, + "learning_rate": 7.1969975482957075e-06, + "loss": 0.125, + "step": 1060 + }, + { + "epoch": 3.7458075904677846, + "grad_norm": 1.1262861615878523, + "learning_rate": 7.191855733945388e-06, + "loss": 0.1304, + "step": 1061 + }, + { + "epoch": 3.7493380406001764, + "grad_norm": 1.0644115251365422, + "learning_rate": 7.1867110486362255e-06, + "loss": 0.1221, + "step": 1062 + }, + { + "epoch": 3.7528684907325687, + "grad_norm": 1.0506019899313102, + "learning_rate": 7.1815634991068895e-06, + "loss": 0.1205, + "step": 1063 + }, + { + "epoch": 3.75639894086496, + "grad_norm": 1.0044867932848505, + "learning_rate": 7.176413092099791e-06, + "loss": 0.1234, + "step": 1064 + }, + { + "epoch": 3.7599293909973523, + "grad_norm": 0.9095279203965683, + "learning_rate": 7.171259834361092e-06, + "loss": 0.1205, + "step": 1065 + }, + { + "epoch": 3.763459841129744, + "grad_norm": 1.1445775552595612, + "learning_rate": 7.1661037326406825e-06, + "loss": 0.1275, + "step": 1066 + }, + { + "epoch": 3.766990291262136, + "grad_norm": 0.946883327804174, + "learning_rate": 7.1609447936921814e-06, + "loss": 0.1398, + "step": 1067 + }, + { + "epoch": 3.770520741394528, + "grad_norm": 0.9041398561076188, + "learning_rate": 7.155783024272925e-06, + "loss": 0.13, + "step": 1068 + }, + { + "epoch": 3.7740511915269197, + "grad_norm": 0.8457469332005921, + "learning_rate": 7.150618431143951e-06, + "loss": 0.1143, + "step": 1069 + }, + { + "epoch": 3.7775816416593115, + "grad_norm": 0.8578352898530813, + "learning_rate": 7.145451021070003e-06, + "loss": 0.1247, + "step": 1070 + }, + { + "epoch": 3.7811120917917034, + "grad_norm": 0.9352514012935873, + "learning_rate": 7.140280800819507e-06, + "loss": 0.1344, + "step": 1071 + }, + { + "epoch": 3.784642541924095, + "grad_norm": 0.9799389793882611, + "learning_rate": 7.135107777164576e-06, + "loss": 0.1223, + "step": 1072 + }, + { + "epoch": 3.788172992056487, + "grad_norm": 1.065041516403633, + "learning_rate": 7.129931956880992e-06, + "loss": 0.1375, + "step": 1073 + }, + { + "epoch": 3.7917034421888793, + "grad_norm": 1.0449759566953258, + "learning_rate": 7.124753346748201e-06, + "loss": 0.1171, + "step": 1074 + }, + { + "epoch": 3.7952338923212707, + "grad_norm": 1.1549382686805798, + "learning_rate": 7.119571953549305e-06, + "loss": 0.1281, + "step": 1075 + }, + { + "epoch": 3.798764342453663, + "grad_norm": 1.007228257856292, + "learning_rate": 7.114387784071044e-06, + "loss": 0.1269, + "step": 1076 + }, + { + "epoch": 3.802294792586055, + "grad_norm": 1.0225753082093523, + "learning_rate": 7.109200845103806e-06, + "loss": 0.1199, + "step": 1077 + }, + { + "epoch": 3.8058252427184467, + "grad_norm": 0.9355931698169853, + "learning_rate": 7.104011143441596e-06, + "loss": 0.1169, + "step": 1078 + }, + { + "epoch": 3.8093556928508385, + "grad_norm": 0.963698358188344, + "learning_rate": 7.098818685882044e-06, + "loss": 0.1179, + "step": 1079 + }, + { + "epoch": 3.8128861429832304, + "grad_norm": 0.7777902932062336, + "learning_rate": 7.093623479226386e-06, + "loss": 0.1186, + "step": 1080 + }, + { + "epoch": 3.816416593115622, + "grad_norm": 0.9518180229607743, + "learning_rate": 7.088425530279461e-06, + "loss": 0.1062, + "step": 1081 + }, + { + "epoch": 3.819947043248014, + "grad_norm": 1.034322676608895, + "learning_rate": 7.083224845849701e-06, + "loss": 0.1144, + "step": 1082 + }, + { + "epoch": 3.823477493380406, + "grad_norm": 1.0557347999282005, + "learning_rate": 7.078021432749121e-06, + "loss": 0.1243, + "step": 1083 + }, + { + "epoch": 3.8270079435127977, + "grad_norm": 0.8967864020526236, + "learning_rate": 7.072815297793303e-06, + "loss": 0.1189, + "step": 1084 + }, + { + "epoch": 3.83053839364519, + "grad_norm": 0.9394124246894394, + "learning_rate": 7.067606447801405e-06, + "loss": 0.1116, + "step": 1085 + }, + { + "epoch": 3.8340688437775814, + "grad_norm": 0.9390953780920334, + "learning_rate": 7.062394889596133e-06, + "loss": 0.1258, + "step": 1086 + }, + { + "epoch": 3.8375992939099737, + "grad_norm": 0.9777220254987733, + "learning_rate": 7.057180630003746e-06, + "loss": 0.1241, + "step": 1087 + }, + { + "epoch": 3.8411297440423655, + "grad_norm": 0.8779726966694773, + "learning_rate": 7.051963675854035e-06, + "loss": 0.1267, + "step": 1088 + }, + { + "epoch": 3.8446601941747574, + "grad_norm": 1.0359165893132334, + "learning_rate": 7.046744033980329e-06, + "loss": 0.1168, + "step": 1089 + }, + { + "epoch": 3.848190644307149, + "grad_norm": 0.9976911554856348, + "learning_rate": 7.041521711219468e-06, + "loss": 0.1317, + "step": 1090 + }, + { + "epoch": 3.851721094439541, + "grad_norm": 1.0169743441045915, + "learning_rate": 7.03629671441181e-06, + "loss": 0.1398, + "step": 1091 + }, + { + "epoch": 3.855251544571933, + "grad_norm": 0.9568634693122333, + "learning_rate": 7.031069050401215e-06, + "loss": 0.0919, + "step": 1092 + }, + { + "epoch": 3.8587819947043247, + "grad_norm": 0.9083188838175256, + "learning_rate": 7.025838726035032e-06, + "loss": 0.1098, + "step": 1093 + }, + { + "epoch": 3.8623124448367165, + "grad_norm": 0.9312177288214185, + "learning_rate": 7.020605748164101e-06, + "loss": 0.1275, + "step": 1094 + }, + { + "epoch": 3.8658428949691084, + "grad_norm": 0.8178614978701555, + "learning_rate": 7.015370123642733e-06, + "loss": 0.11, + "step": 1095 + }, + { + "epoch": 3.8693733451015007, + "grad_norm": 0.9943777665520181, + "learning_rate": 7.010131859328706e-06, + "loss": 0.1141, + "step": 1096 + }, + { + "epoch": 3.872903795233892, + "grad_norm": 0.7533041541573118, + "learning_rate": 7.00489096208326e-06, + "loss": 0.0988, + "step": 1097 + }, + { + "epoch": 3.8764342453662843, + "grad_norm": 0.9539898985521694, + "learning_rate": 6.999647438771076e-06, + "loss": 0.1158, + "step": 1098 + }, + { + "epoch": 3.879964695498676, + "grad_norm": 0.9694635032226555, + "learning_rate": 6.994401296260285e-06, + "loss": 0.1476, + "step": 1099 + }, + { + "epoch": 3.883495145631068, + "grad_norm": 0.9466515792078597, + "learning_rate": 6.989152541422438e-06, + "loss": 0.1207, + "step": 1100 + }, + { + "epoch": 3.88702559576346, + "grad_norm": 0.9232735267824136, + "learning_rate": 6.983901181132514e-06, + "loss": 0.1219, + "step": 1101 + }, + { + "epoch": 3.8905560458958517, + "grad_norm": 0.9035726923781363, + "learning_rate": 6.978647222268904e-06, + "loss": 0.1139, + "step": 1102 + }, + { + "epoch": 3.8940864960282435, + "grad_norm": 0.9447988957374035, + "learning_rate": 6.973390671713401e-06, + "loss": 0.1106, + "step": 1103 + }, + { + "epoch": 3.8976169461606354, + "grad_norm": 0.9993165292258056, + "learning_rate": 6.9681315363511946e-06, + "loss": 0.1126, + "step": 1104 + }, + { + "epoch": 3.901147396293027, + "grad_norm": 0.9567127281227304, + "learning_rate": 6.9628698230708604e-06, + "loss": 0.126, + "step": 1105 + }, + { + "epoch": 3.904677846425419, + "grad_norm": 0.9704513493237618, + "learning_rate": 6.957605538764347e-06, + "loss": 0.1135, + "step": 1106 + }, + { + "epoch": 3.9082082965578113, + "grad_norm": 0.9049083128378328, + "learning_rate": 6.9523386903269715e-06, + "loss": 0.118, + "step": 1107 + }, + { + "epoch": 3.911738746690203, + "grad_norm": 1.1410548052057965, + "learning_rate": 6.947069284657416e-06, + "loss": 0.1232, + "step": 1108 + }, + { + "epoch": 3.915269196822595, + "grad_norm": 0.9628185789378251, + "learning_rate": 6.941797328657702e-06, + "loss": 0.1012, + "step": 1109 + }, + { + "epoch": 3.918799646954987, + "grad_norm": 0.986500497408887, + "learning_rate": 6.936522829233202e-06, + "loss": 0.1267, + "step": 1110 + }, + { + "epoch": 3.9223300970873787, + "grad_norm": 0.9404574261439161, + "learning_rate": 6.93124579329261e-06, + "loss": 0.1177, + "step": 1111 + }, + { + "epoch": 3.9258605472197705, + "grad_norm": 0.9652812442302577, + "learning_rate": 6.92596622774795e-06, + "loss": 0.1099, + "step": 1112 + }, + { + "epoch": 3.9293909973521624, + "grad_norm": 0.9255614343816538, + "learning_rate": 6.920684139514555e-06, + "loss": 0.1157, + "step": 1113 + }, + { + "epoch": 3.932921447484554, + "grad_norm": 1.1133925624513497, + "learning_rate": 6.9153995355110645e-06, + "loss": 0.13, + "step": 1114 + }, + { + "epoch": 3.936451897616946, + "grad_norm": 1.0496747828703927, + "learning_rate": 6.9101124226594106e-06, + "loss": 0.1123, + "step": 1115 + }, + { + "epoch": 3.9399823477493383, + "grad_norm": 1.0246048820560874, + "learning_rate": 6.904822807884816e-06, + "loss": 0.1158, + "step": 1116 + }, + { + "epoch": 3.9435127978817297, + "grad_norm": 1.0333168676182993, + "learning_rate": 6.899530698115776e-06, + "loss": 0.1315, + "step": 1117 + }, + { + "epoch": 3.947043248014122, + "grad_norm": 1.0064120983759648, + "learning_rate": 6.894236100284058e-06, + "loss": 0.1282, + "step": 1118 + }, + { + "epoch": 3.950573698146514, + "grad_norm": 0.963684336758527, + "learning_rate": 6.888939021324681e-06, + "loss": 0.1217, + "step": 1119 + }, + { + "epoch": 3.9541041482789057, + "grad_norm": 0.9892019560831151, + "learning_rate": 6.883639468175926e-06, + "loss": 0.1239, + "step": 1120 + }, + { + "epoch": 3.9576345984112975, + "grad_norm": 1.0120690523201414, + "learning_rate": 6.878337447779304e-06, + "loss": 0.1335, + "step": 1121 + }, + { + "epoch": 3.9611650485436893, + "grad_norm": 0.8218662981254, + "learning_rate": 6.873032967079562e-06, + "loss": 0.1185, + "step": 1122 + }, + { + "epoch": 3.964695498676081, + "grad_norm": 0.9915669809027783, + "learning_rate": 6.867726033024667e-06, + "loss": 0.1181, + "step": 1123 + }, + { + "epoch": 3.968225948808473, + "grad_norm": 0.7674002943828144, + "learning_rate": 6.862416652565804e-06, + "loss": 0.11, + "step": 1124 + }, + { + "epoch": 3.971756398940865, + "grad_norm": 0.9775474573427773, + "learning_rate": 6.85710483265736e-06, + "loss": 0.1357, + "step": 1125 + }, + { + "epoch": 3.9752868490732567, + "grad_norm": 0.9046890209780649, + "learning_rate": 6.8517905802569164e-06, + "loss": 0.1145, + "step": 1126 + }, + { + "epoch": 3.978817299205649, + "grad_norm": 0.921557821840316, + "learning_rate": 6.846473902325243e-06, + "loss": 0.1392, + "step": 1127 + }, + { + "epoch": 3.9823477493380404, + "grad_norm": 0.9794967931322031, + "learning_rate": 6.8411548058262824e-06, + "loss": 0.1088, + "step": 1128 + }, + { + "epoch": 3.9858781994704326, + "grad_norm": 0.8740159825381886, + "learning_rate": 6.835833297727148e-06, + "loss": 0.1139, + "step": 1129 + }, + { + "epoch": 3.9894086496028245, + "grad_norm": 0.8628586385174893, + "learning_rate": 6.830509384998114e-06, + "loss": 0.1082, + "step": 1130 + }, + { + "epoch": 3.9929390997352163, + "grad_norm": 0.9488799231960956, + "learning_rate": 6.8251830746126e-06, + "loss": 0.1316, + "step": 1131 + }, + { + "epoch": 3.996469549867608, + "grad_norm": 1.0703456877526685, + "learning_rate": 6.81985437354717e-06, + "loss": 0.1279, + "step": 1132 + }, + { + "epoch": 4.0, + "grad_norm": 0.967853650509713, + "learning_rate": 6.814523288781516e-06, + "loss": 0.1305, + "step": 1133 + }, + { + "epoch": 4.003530450132392, + "grad_norm": 0.7981320722327793, + "learning_rate": 6.809189827298457e-06, + "loss": 0.0885, + "step": 1134 + }, + { + "epoch": 4.007060900264784, + "grad_norm": 0.5854902750468941, + "learning_rate": 6.803853996083918e-06, + "loss": 0.0793, + "step": 1135 + }, + { + "epoch": 4.010591350397176, + "grad_norm": 0.7049273298585778, + "learning_rate": 6.7985158021269335e-06, + "loss": 0.0758, + "step": 1136 + }, + { + "epoch": 4.014121800529567, + "grad_norm": 0.6869340400131053, + "learning_rate": 6.79317525241963e-06, + "loss": 0.0933, + "step": 1137 + }, + { + "epoch": 4.01765225066196, + "grad_norm": 0.5900057189092714, + "learning_rate": 6.787832353957225e-06, + "loss": 0.0706, + "step": 1138 + }, + { + "epoch": 4.021182700794351, + "grad_norm": 0.7601438078339049, + "learning_rate": 6.782487113738002e-06, + "loss": 0.0769, + "step": 1139 + }, + { + "epoch": 4.024713150926743, + "grad_norm": 0.851783451355363, + "learning_rate": 6.777139538763324e-06, + "loss": 0.0711, + "step": 1140 + }, + { + "epoch": 4.028243601059135, + "grad_norm": 0.82880112354727, + "learning_rate": 6.7717896360376025e-06, + "loss": 0.0717, + "step": 1141 + }, + { + "epoch": 4.031774051191527, + "grad_norm": 1.1143799167216393, + "learning_rate": 6.766437412568302e-06, + "loss": 0.0784, + "step": 1142 + }, + { + "epoch": 4.035304501323918, + "grad_norm": 0.8210313928830595, + "learning_rate": 6.761082875365929e-06, + "loss": 0.0626, + "step": 1143 + }, + { + "epoch": 4.038834951456311, + "grad_norm": 0.7779155627289819, + "learning_rate": 6.755726031444014e-06, + "loss": 0.0649, + "step": 1144 + }, + { + "epoch": 4.042365401588703, + "grad_norm": 1.1156761361527734, + "learning_rate": 6.750366887819116e-06, + "loss": 0.0787, + "step": 1145 + }, + { + "epoch": 4.045895851721094, + "grad_norm": 1.11782008947278, + "learning_rate": 6.745005451510804e-06, + "loss": 0.0742, + "step": 1146 + }, + { + "epoch": 4.049426301853487, + "grad_norm": 1.0344771965558572, + "learning_rate": 6.739641729541645e-06, + "loss": 0.0838, + "step": 1147 + }, + { + "epoch": 4.052956751985878, + "grad_norm": 2.0154736605675763, + "learning_rate": 6.734275728937208e-06, + "loss": 0.1034, + "step": 1148 + }, + { + "epoch": 4.05648720211827, + "grad_norm": 1.2294533756515769, + "learning_rate": 6.728907456726042e-06, + "loss": 0.0619, + "step": 1149 + }, + { + "epoch": 4.060017652250662, + "grad_norm": 1.4290104269014912, + "learning_rate": 6.723536919939669e-06, + "loss": 0.0686, + "step": 1150 + }, + { + "epoch": 4.063548102383054, + "grad_norm": 1.5020515871139606, + "learning_rate": 6.718164125612585e-06, + "loss": 0.0761, + "step": 1151 + }, + { + "epoch": 4.067078552515445, + "grad_norm": 1.4908120498463364, + "learning_rate": 6.712789080782235e-06, + "loss": 0.0828, + "step": 1152 + }, + { + "epoch": 4.070609002647838, + "grad_norm": 1.2267050202060312, + "learning_rate": 6.707411792489016e-06, + "loss": 0.0643, + "step": 1153 + }, + { + "epoch": 4.074139452780229, + "grad_norm": 1.3569581964127149, + "learning_rate": 6.7020322677762614e-06, + "loss": 0.082, + "step": 1154 + }, + { + "epoch": 4.077669902912621, + "grad_norm": 1.1848448933110132, + "learning_rate": 6.696650513690238e-06, + "loss": 0.0639, + "step": 1155 + }, + { + "epoch": 4.081200353045014, + "grad_norm": 1.181765635994709, + "learning_rate": 6.691266537280128e-06, + "loss": 0.057, + "step": 1156 + }, + { + "epoch": 4.084730803177405, + "grad_norm": 1.016848718981542, + "learning_rate": 6.685880345598027e-06, + "loss": 0.0537, + "step": 1157 + }, + { + "epoch": 4.088261253309797, + "grad_norm": 1.3224355302362136, + "learning_rate": 6.680491945698929e-06, + "loss": 0.064, + "step": 1158 + }, + { + "epoch": 4.091791703442189, + "grad_norm": 1.4746292032964976, + "learning_rate": 6.675101344640727e-06, + "loss": 0.0892, + "step": 1159 + }, + { + "epoch": 4.095322153574581, + "grad_norm": 1.5383513145198393, + "learning_rate": 6.66970854948419e-06, + "loss": 0.0766, + "step": 1160 + }, + { + "epoch": 4.098852603706972, + "grad_norm": 1.207790515633538, + "learning_rate": 6.664313567292966e-06, + "loss": 0.0693, + "step": 1161 + }, + { + "epoch": 4.102383053839365, + "grad_norm": 1.288949461541623, + "learning_rate": 6.658916405133562e-06, + "loss": 0.0826, + "step": 1162 + }, + { + "epoch": 4.105913503971756, + "grad_norm": 1.0936618806184264, + "learning_rate": 6.6535170700753506e-06, + "loss": 0.0789, + "step": 1163 + }, + { + "epoch": 4.109443954104148, + "grad_norm": 1.087774369178708, + "learning_rate": 6.6481155691905375e-06, + "loss": 0.0571, + "step": 1164 + }, + { + "epoch": 4.112974404236541, + "grad_norm": 1.3919202257673968, + "learning_rate": 6.6427119095541745e-06, + "loss": 0.0655, + "step": 1165 + }, + { + "epoch": 4.116504854368932, + "grad_norm": 1.0692129775554304, + "learning_rate": 6.637306098244138e-06, + "loss": 0.0758, + "step": 1166 + }, + { + "epoch": 4.120035304501324, + "grad_norm": 1.1487653554819723, + "learning_rate": 6.631898142341122e-06, + "loss": 0.0619, + "step": 1167 + }, + { + "epoch": 4.123565754633716, + "grad_norm": 1.036223404954323, + "learning_rate": 6.626488048928632e-06, + "loss": 0.0593, + "step": 1168 + }, + { + "epoch": 4.127096204766108, + "grad_norm": 1.0121114014928942, + "learning_rate": 6.621075825092972e-06, + "loss": 0.0667, + "step": 1169 + }, + { + "epoch": 4.130626654898499, + "grad_norm": 1.1972992188904066, + "learning_rate": 6.6156614779232345e-06, + "loss": 0.0733, + "step": 1170 + }, + { + "epoch": 4.134157105030892, + "grad_norm": 1.197590502375495, + "learning_rate": 6.610245014511296e-06, + "loss": 0.0611, + "step": 1171 + }, + { + "epoch": 4.137687555163283, + "grad_norm": 1.0881571550345615, + "learning_rate": 6.604826441951803e-06, + "loss": 0.061, + "step": 1172 + }, + { + "epoch": 4.141218005295675, + "grad_norm": 0.9686194494515865, + "learning_rate": 6.599405767342167e-06, + "loss": 0.0559, + "step": 1173 + }, + { + "epoch": 4.144748455428067, + "grad_norm": 1.3831994016631128, + "learning_rate": 6.593982997782549e-06, + "loss": 0.0642, + "step": 1174 + }, + { + "epoch": 4.148278905560459, + "grad_norm": 1.241798068520593, + "learning_rate": 6.588558140375858e-06, + "loss": 0.0707, + "step": 1175 + }, + { + "epoch": 4.151809355692851, + "grad_norm": 1.2592043421219463, + "learning_rate": 6.583131202227736e-06, + "loss": 0.0851, + "step": 1176 + }, + { + "epoch": 4.155339805825243, + "grad_norm": 1.1890876424522738, + "learning_rate": 6.577702190446552e-06, + "loss": 0.0764, + "step": 1177 + }, + { + "epoch": 4.158870255957635, + "grad_norm": 1.1892762885244956, + "learning_rate": 6.572271112143386e-06, + "loss": 0.0823, + "step": 1178 + }, + { + "epoch": 4.162400706090026, + "grad_norm": 1.098687094404278, + "learning_rate": 6.566837974432032e-06, + "loss": 0.0582, + "step": 1179 + }, + { + "epoch": 4.165931156222419, + "grad_norm": 1.1795811578347426, + "learning_rate": 6.561402784428974e-06, + "loss": 0.0787, + "step": 1180 + }, + { + "epoch": 4.16946160635481, + "grad_norm": 1.0850929765127277, + "learning_rate": 6.5559655492533915e-06, + "loss": 0.0576, + "step": 1181 + }, + { + "epoch": 4.172992056487202, + "grad_norm": 1.2039461760737675, + "learning_rate": 6.550526276027138e-06, + "loss": 0.075, + "step": 1182 + }, + { + "epoch": 4.176522506619594, + "grad_norm": 1.1924260467355354, + "learning_rate": 6.545084971874738e-06, + "loss": 0.0759, + "step": 1183 + }, + { + "epoch": 4.180052956751986, + "grad_norm": 1.349601668880594, + "learning_rate": 6.539641643923376e-06, + "loss": 0.0762, + "step": 1184 + }, + { + "epoch": 4.183583406884377, + "grad_norm": 1.5302048823422634, + "learning_rate": 6.534196299302887e-06, + "loss": 0.0721, + "step": 1185 + }, + { + "epoch": 4.18711385701677, + "grad_norm": 1.1161238197658445, + "learning_rate": 6.528748945145749e-06, + "loss": 0.0752, + "step": 1186 + }, + { + "epoch": 4.190644307149162, + "grad_norm": 1.0876160222524982, + "learning_rate": 6.523299588587069e-06, + "loss": 0.0672, + "step": 1187 + }, + { + "epoch": 4.194174757281553, + "grad_norm": 1.1607986364903538, + "learning_rate": 6.51784823676458e-06, + "loss": 0.0743, + "step": 1188 + }, + { + "epoch": 4.197705207413946, + "grad_norm": 1.0772412583188433, + "learning_rate": 6.5123948968186275e-06, + "loss": 0.0673, + "step": 1189 + }, + { + "epoch": 4.201235657546337, + "grad_norm": 1.2617922130605492, + "learning_rate": 6.506939575892161e-06, + "loss": 0.0619, + "step": 1190 + }, + { + "epoch": 4.204766107678729, + "grad_norm": 1.2483711396156587, + "learning_rate": 6.501482281130724e-06, + "loss": 0.0871, + "step": 1191 + }, + { + "epoch": 4.208296557811121, + "grad_norm": 1.1989951304725375, + "learning_rate": 6.496023019682447e-06, + "loss": 0.068, + "step": 1192 + }, + { + "epoch": 4.211827007943513, + "grad_norm": 1.038016185342428, + "learning_rate": 6.490561798698033e-06, + "loss": 0.0672, + "step": 1193 + }, + { + "epoch": 4.215357458075904, + "grad_norm": 1.2879497667669522, + "learning_rate": 6.4850986253307566e-06, + "loss": 0.0813, + "step": 1194 + }, + { + "epoch": 4.218887908208297, + "grad_norm": 1.138562370359682, + "learning_rate": 6.479633506736447e-06, + "loss": 0.0798, + "step": 1195 + }, + { + "epoch": 4.222418358340688, + "grad_norm": 1.2141342731122955, + "learning_rate": 6.474166450073482e-06, + "loss": 0.0753, + "step": 1196 + }, + { + "epoch": 4.22594880847308, + "grad_norm": 1.1524180729165578, + "learning_rate": 6.468697462502776e-06, + "loss": 0.0669, + "step": 1197 + }, + { + "epoch": 4.229479258605473, + "grad_norm": 1.2503618917005421, + "learning_rate": 6.463226551187776e-06, + "loss": 0.0802, + "step": 1198 + }, + { + "epoch": 4.233009708737864, + "grad_norm": 1.0897556077719, + "learning_rate": 6.457753723294443e-06, + "loss": 0.0687, + "step": 1199 + }, + { + "epoch": 4.236540158870256, + "grad_norm": 1.2215195146222155, + "learning_rate": 6.452278985991257e-06, + "loss": 0.0859, + "step": 1200 + }, + { + "epoch": 4.240070609002648, + "grad_norm": 1.123982384909271, + "learning_rate": 6.4468023464491906e-06, + "loss": 0.0693, + "step": 1201 + }, + { + "epoch": 4.24360105913504, + "grad_norm": 1.161677655741415, + "learning_rate": 6.441323811841715e-06, + "loss": 0.0749, + "step": 1202 + }, + { + "epoch": 4.247131509267431, + "grad_norm": 1.1029803998203793, + "learning_rate": 6.435843389344778e-06, + "loss": 0.0666, + "step": 1203 + }, + { + "epoch": 4.250661959399824, + "grad_norm": 1.3744931507694842, + "learning_rate": 6.4303610861368026e-06, + "loss": 0.0844, + "step": 1204 + }, + { + "epoch": 4.254192409532215, + "grad_norm": 1.0659314315032296, + "learning_rate": 6.4248769093986765e-06, + "loss": 0.0723, + "step": 1205 + }, + { + "epoch": 4.257722859664607, + "grad_norm": 1.2659662434004557, + "learning_rate": 6.419390866313741e-06, + "loss": 0.0786, + "step": 1206 + }, + { + "epoch": 4.261253309796999, + "grad_norm": 1.2058156571448462, + "learning_rate": 6.413902964067776e-06, + "loss": 0.0708, + "step": 1207 + }, + { + "epoch": 4.264783759929391, + "grad_norm": 1.4552116581087131, + "learning_rate": 6.408413209849007e-06, + "loss": 0.0802, + "step": 1208 + }, + { + "epoch": 4.268314210061783, + "grad_norm": 1.3140338475558748, + "learning_rate": 6.402921610848076e-06, + "loss": 0.0701, + "step": 1209 + }, + { + "epoch": 4.271844660194175, + "grad_norm": 1.4948147235906752, + "learning_rate": 6.397428174258048e-06, + "loss": 0.0765, + "step": 1210 + }, + { + "epoch": 4.275375110326567, + "grad_norm": 1.4432674805143244, + "learning_rate": 6.391932907274389e-06, + "loss": 0.0736, + "step": 1211 + }, + { + "epoch": 4.278905560458958, + "grad_norm": 1.1269206127304876, + "learning_rate": 6.386435817094971e-06, + "loss": 0.0762, + "step": 1212 + }, + { + "epoch": 4.282436010591351, + "grad_norm": 1.4426602486276199, + "learning_rate": 6.380936910920046e-06, + "loss": 0.0834, + "step": 1213 + }, + { + "epoch": 4.285966460723742, + "grad_norm": 0.9454852560163308, + "learning_rate": 6.375436195952245e-06, + "loss": 0.0677, + "step": 1214 + }, + { + "epoch": 4.289496910856134, + "grad_norm": 1.1295627967839161, + "learning_rate": 6.369933679396571e-06, + "loss": 0.057, + "step": 1215 + }, + { + "epoch": 4.293027360988526, + "grad_norm": 1.19896192193708, + "learning_rate": 6.364429368460391e-06, + "loss": 0.0786, + "step": 1216 + }, + { + "epoch": 4.296557811120918, + "grad_norm": 1.2164688961559822, + "learning_rate": 6.358923270353412e-06, + "loss": 0.0743, + "step": 1217 + }, + { + "epoch": 4.300088261253309, + "grad_norm": 1.4316527320361974, + "learning_rate": 6.353415392287692e-06, + "loss": 0.0735, + "step": 1218 + }, + { + "epoch": 4.303618711385702, + "grad_norm": 1.1280198948959894, + "learning_rate": 6.347905741477613e-06, + "loss": 0.0638, + "step": 1219 + }, + { + "epoch": 4.307149161518094, + "grad_norm": 1.3537055389950547, + "learning_rate": 6.342394325139886e-06, + "loss": 0.0844, + "step": 1220 + }, + { + "epoch": 4.310679611650485, + "grad_norm": 1.175521881817613, + "learning_rate": 6.336881150493525e-06, + "loss": 0.0697, + "step": 1221 + }, + { + "epoch": 4.314210061782878, + "grad_norm": 1.3032526032205891, + "learning_rate": 6.3313662247598575e-06, + "loss": 0.0681, + "step": 1222 + }, + { + "epoch": 4.317740511915269, + "grad_norm": 1.2291592653948193, + "learning_rate": 6.325849555162496e-06, + "loss": 0.078, + "step": 1223 + }, + { + "epoch": 4.321270962047661, + "grad_norm": 1.2292408189938648, + "learning_rate": 6.320331148927344e-06, + "loss": 0.0605, + "step": 1224 + }, + { + "epoch": 4.324801412180053, + "grad_norm": 1.317695842582305, + "learning_rate": 6.314811013282574e-06, + "loss": 0.074, + "step": 1225 + }, + { + "epoch": 4.328331862312445, + "grad_norm": 1.3300537195408657, + "learning_rate": 6.309289155458629e-06, + "loss": 0.0781, + "step": 1226 + }, + { + "epoch": 4.331862312444836, + "grad_norm": 1.2795961802893723, + "learning_rate": 6.303765582688204e-06, + "loss": 0.0758, + "step": 1227 + }, + { + "epoch": 4.335392762577229, + "grad_norm": 1.089546205227726, + "learning_rate": 6.298240302206242e-06, + "loss": 0.0716, + "step": 1228 + }, + { + "epoch": 4.338923212709621, + "grad_norm": 1.4436985362559518, + "learning_rate": 6.29271332124992e-06, + "loss": 0.0733, + "step": 1229 + }, + { + "epoch": 4.342453662842012, + "grad_norm": 1.2123854120814914, + "learning_rate": 6.287184647058648e-06, + "loss": 0.063, + "step": 1230 + }, + { + "epoch": 4.345984112974405, + "grad_norm": 1.1599856785429075, + "learning_rate": 6.281654286874047e-06, + "loss": 0.0703, + "step": 1231 + }, + { + "epoch": 4.349514563106796, + "grad_norm": 1.2863269775769035, + "learning_rate": 6.276122247939953e-06, + "loss": 0.08, + "step": 1232 + }, + { + "epoch": 4.353045013239188, + "grad_norm": 0.9802224693174152, + "learning_rate": 6.2705885375023955e-06, + "loss": 0.0669, + "step": 1233 + }, + { + "epoch": 4.35657546337158, + "grad_norm": 1.3003686599476758, + "learning_rate": 6.265053162809597e-06, + "loss": 0.0758, + "step": 1234 + }, + { + "epoch": 4.360105913503972, + "grad_norm": 1.2476813765188703, + "learning_rate": 6.259516131111957e-06, + "loss": 0.0719, + "step": 1235 + }, + { + "epoch": 4.363636363636363, + "grad_norm": 1.0504270751123217, + "learning_rate": 6.253977449662047e-06, + "loss": 0.0614, + "step": 1236 + }, + { + "epoch": 4.367166813768756, + "grad_norm": 1.1435259908690352, + "learning_rate": 6.2484371257146e-06, + "loss": 0.0768, + "step": 1237 + }, + { + "epoch": 4.370697263901147, + "grad_norm": 1.3840251434761308, + "learning_rate": 6.242895166526502e-06, + "loss": 0.0742, + "step": 1238 + }, + { + "epoch": 4.374227714033539, + "grad_norm": 1.2422183469141699, + "learning_rate": 6.2373515793567715e-06, + "loss": 0.08, + "step": 1239 + }, + { + "epoch": 4.3777581641659316, + "grad_norm": 1.1315889355948883, + "learning_rate": 6.231806371466574e-06, + "loss": 0.0697, + "step": 1240 + }, + { + "epoch": 4.381288614298323, + "grad_norm": 1.081317701161799, + "learning_rate": 6.226259550119187e-06, + "loss": 0.0672, + "step": 1241 + }, + { + "epoch": 4.384819064430715, + "grad_norm": 1.139178949237585, + "learning_rate": 6.220711122580004e-06, + "loss": 0.0762, + "step": 1242 + }, + { + "epoch": 4.388349514563107, + "grad_norm": 1.2934745666925676, + "learning_rate": 6.215161096116524e-06, + "loss": 0.081, + "step": 1243 + }, + { + "epoch": 4.391879964695499, + "grad_norm": 1.4026385643923822, + "learning_rate": 6.209609477998339e-06, + "loss": 0.0759, + "step": 1244 + }, + { + "epoch": 4.39541041482789, + "grad_norm": 1.2849606159130602, + "learning_rate": 6.2040562754971265e-06, + "loss": 0.0713, + "step": 1245 + }, + { + "epoch": 4.398940864960283, + "grad_norm": 1.0919624470047933, + "learning_rate": 6.1985014958866386e-06, + "loss": 0.0731, + "step": 1246 + }, + { + "epoch": 4.402471315092674, + "grad_norm": 1.2552988710153292, + "learning_rate": 6.192945146442693e-06, + "loss": 0.0802, + "step": 1247 + }, + { + "epoch": 4.406001765225066, + "grad_norm": 1.2944459168853082, + "learning_rate": 6.187387234443164e-06, + "loss": 0.084, + "step": 1248 + }, + { + "epoch": 4.4095322153574585, + "grad_norm": 1.1228808317446295, + "learning_rate": 6.181827767167974e-06, + "loss": 0.0716, + "step": 1249 + }, + { + "epoch": 4.41306266548985, + "grad_norm": 1.1881280477511678, + "learning_rate": 6.176266751899078e-06, + "loss": 0.067, + "step": 1250 + }, + { + "epoch": 4.416593115622242, + "grad_norm": 1.1569504030576783, + "learning_rate": 6.170704195920465e-06, + "loss": 0.0702, + "step": 1251 + }, + { + "epoch": 4.420123565754634, + "grad_norm": 1.189002472009121, + "learning_rate": 6.165140106518137e-06, + "loss": 0.0694, + "step": 1252 + }, + { + "epoch": 4.423654015887026, + "grad_norm": 1.232505313641185, + "learning_rate": 6.1595744909801085e-06, + "loss": 0.0754, + "step": 1253 + }, + { + "epoch": 4.427184466019417, + "grad_norm": 1.2147460452464223, + "learning_rate": 6.154007356596387e-06, + "loss": 0.0652, + "step": 1254 + }, + { + "epoch": 4.43071491615181, + "grad_norm": 1.1480977785419901, + "learning_rate": 6.148438710658979e-06, + "loss": 0.0625, + "step": 1255 + }, + { + "epoch": 4.434245366284201, + "grad_norm": 1.218331723091377, + "learning_rate": 6.142868560461859e-06, + "loss": 0.0821, + "step": 1256 + }, + { + "epoch": 4.437775816416593, + "grad_norm": 1.1738427410555219, + "learning_rate": 6.137296913300982e-06, + "loss": 0.074, + "step": 1257 + }, + { + "epoch": 4.441306266548985, + "grad_norm": 1.19855168208864, + "learning_rate": 6.131723776474258e-06, + "loss": 0.0738, + "step": 1258 + }, + { + "epoch": 4.444836716681377, + "grad_norm": 1.3103544915169332, + "learning_rate": 6.12614915728155e-06, + "loss": 0.0781, + "step": 1259 + }, + { + "epoch": 4.448367166813769, + "grad_norm": 1.0115359906207875, + "learning_rate": 6.120573063024663e-06, + "loss": 0.0715, + "step": 1260 + }, + { + "epoch": 4.451897616946161, + "grad_norm": 0.9684920252271274, + "learning_rate": 6.114995501007334e-06, + "loss": 0.0626, + "step": 1261 + }, + { + "epoch": 4.455428067078553, + "grad_norm": 1.0836045218047463, + "learning_rate": 6.109416478535222e-06, + "loss": 0.0707, + "step": 1262 + }, + { + "epoch": 4.458958517210944, + "grad_norm": 1.1090014070135534, + "learning_rate": 6.103836002915898e-06, + "loss": 0.0638, + "step": 1263 + }, + { + "epoch": 4.4624889673433366, + "grad_norm": 1.300961015257033, + "learning_rate": 6.098254081458839e-06, + "loss": 0.0727, + "step": 1264 + }, + { + "epoch": 4.466019417475728, + "grad_norm": 1.2337280261124597, + "learning_rate": 6.092670721475414e-06, + "loss": 0.0748, + "step": 1265 + }, + { + "epoch": 4.46954986760812, + "grad_norm": 1.1278101885787395, + "learning_rate": 6.0870859302788756e-06, + "loss": 0.0703, + "step": 1266 + }, + { + "epoch": 4.473080317740512, + "grad_norm": 1.0631423710066719, + "learning_rate": 6.081499715184353e-06, + "loss": 0.0706, + "step": 1267 + }, + { + "epoch": 4.476610767872904, + "grad_norm": 1.143193764341101, + "learning_rate": 6.075912083508839e-06, + "loss": 0.0694, + "step": 1268 + }, + { + "epoch": 4.480141218005295, + "grad_norm": 1.1879992175907763, + "learning_rate": 6.070323042571184e-06, + "loss": 0.0683, + "step": 1269 + }, + { + "epoch": 4.483671668137688, + "grad_norm": 1.4154753236711644, + "learning_rate": 6.064732599692079e-06, + "loss": 0.0841, + "step": 1270 + }, + { + "epoch": 4.48720211827008, + "grad_norm": 1.0845224887804554, + "learning_rate": 6.059140762194059e-06, + "loss": 0.0753, + "step": 1271 + }, + { + "epoch": 4.490732568402471, + "grad_norm": 1.1903474371803968, + "learning_rate": 6.053547537401477e-06, + "loss": 0.0706, + "step": 1272 + }, + { + "epoch": 4.4942630185348635, + "grad_norm": 1.616858067887646, + "learning_rate": 6.047952932640513e-06, + "loss": 0.0912, + "step": 1273 + }, + { + "epoch": 4.497793468667255, + "grad_norm": 1.0913358558919948, + "learning_rate": 6.042356955239145e-06, + "loss": 0.0604, + "step": 1274 + }, + { + "epoch": 4.501323918799647, + "grad_norm": 1.1880905119942522, + "learning_rate": 6.036759612527157e-06, + "loss": 0.0759, + "step": 1275 + }, + { + "epoch": 4.504854368932039, + "grad_norm": 1.1747169454730277, + "learning_rate": 6.031160911836112e-06, + "loss": 0.0679, + "step": 1276 + }, + { + "epoch": 4.508384819064431, + "grad_norm": 1.211427366915763, + "learning_rate": 6.025560860499365e-06, + "loss": 0.0689, + "step": 1277 + }, + { + "epoch": 4.511915269196822, + "grad_norm": 1.417931942561815, + "learning_rate": 6.019959465852025e-06, + "loss": 0.0868, + "step": 1278 + }, + { + "epoch": 4.515445719329215, + "grad_norm": 1.2209993004826432, + "learning_rate": 6.014356735230973e-06, + "loss": 0.0646, + "step": 1279 + }, + { + "epoch": 4.518976169461606, + "grad_norm": 1.1649085744269059, + "learning_rate": 6.00875267597483e-06, + "loss": 0.0639, + "step": 1280 + }, + { + "epoch": 4.522506619593998, + "grad_norm": 1.011949914685396, + "learning_rate": 6.003147295423967e-06, + "loss": 0.0597, + "step": 1281 + }, + { + "epoch": 4.5260370697263905, + "grad_norm": 1.1309980576687069, + "learning_rate": 5.997540600920479e-06, + "loss": 0.0782, + "step": 1282 + }, + { + "epoch": 4.529567519858782, + "grad_norm": 1.0299510613826361, + "learning_rate": 5.9919325998081825e-06, + "loss": 0.066, + "step": 1283 + }, + { + "epoch": 4.533097969991174, + "grad_norm": 1.0789969131073938, + "learning_rate": 5.986323299432608e-06, + "loss": 0.0719, + "step": 1284 + }, + { + "epoch": 4.536628420123566, + "grad_norm": 1.0170912861541357, + "learning_rate": 5.980712707140985e-06, + "loss": 0.064, + "step": 1285 + }, + { + "epoch": 4.540158870255958, + "grad_norm": 1.1901299269101344, + "learning_rate": 5.97510083028224e-06, + "loss": 0.0707, + "step": 1286 + }, + { + "epoch": 4.543689320388349, + "grad_norm": 1.236036942535051, + "learning_rate": 5.969487676206975e-06, + "loss": 0.0695, + "step": 1287 + }, + { + "epoch": 4.5472197705207416, + "grad_norm": 1.0739729905219972, + "learning_rate": 5.9638732522674716e-06, + "loss": 0.0683, + "step": 1288 + }, + { + "epoch": 4.550750220653133, + "grad_norm": 1.1896504864727457, + "learning_rate": 5.95825756581767e-06, + "loss": 0.0692, + "step": 1289 + }, + { + "epoch": 4.554280670785525, + "grad_norm": 1.0099956859061028, + "learning_rate": 5.952640624213166e-06, + "loss": 0.0717, + "step": 1290 + }, + { + "epoch": 4.557811120917917, + "grad_norm": 1.1995738747246374, + "learning_rate": 5.947022434811202e-06, + "loss": 0.0611, + "step": 1291 + }, + { + "epoch": 4.561341571050309, + "grad_norm": 1.6264487607523304, + "learning_rate": 5.94140300497065e-06, + "loss": 0.0817, + "step": 1292 + }, + { + "epoch": 4.564872021182701, + "grad_norm": 1.2030247661685802, + "learning_rate": 5.935782342052007e-06, + "loss": 0.0605, + "step": 1293 + }, + { + "epoch": 4.568402471315093, + "grad_norm": 1.117896941640317, + "learning_rate": 5.93016045341739e-06, + "loss": 0.0521, + "step": 1294 + }, + { + "epoch": 4.571932921447485, + "grad_norm": 1.282540263365002, + "learning_rate": 5.924537346430517e-06, + "loss": 0.0719, + "step": 1295 + }, + { + "epoch": 4.575463371579876, + "grad_norm": 1.4173574483743017, + "learning_rate": 5.918913028456705e-06, + "loss": 0.0665, + "step": 1296 + }, + { + "epoch": 4.5789938217122685, + "grad_norm": 1.7378141839355061, + "learning_rate": 5.9132875068628555e-06, + "loss": 0.082, + "step": 1297 + }, + { + "epoch": 4.58252427184466, + "grad_norm": 1.5474239856959229, + "learning_rate": 5.907660789017446e-06, + "loss": 0.0747, + "step": 1298 + }, + { + "epoch": 4.586054721977052, + "grad_norm": 1.2885896943234003, + "learning_rate": 5.902032882290518e-06, + "loss": 0.0744, + "step": 1299 + }, + { + "epoch": 4.589585172109444, + "grad_norm": 1.3460860331275364, + "learning_rate": 5.896403794053679e-06, + "loss": 0.0767, + "step": 1300 + }, + { + "epoch": 4.593115622241836, + "grad_norm": 1.1096734280955396, + "learning_rate": 5.890773531680076e-06, + "loss": 0.0777, + "step": 1301 + }, + { + "epoch": 4.596646072374227, + "grad_norm": 1.1779234981799707, + "learning_rate": 5.885142102544398e-06, + "loss": 0.0636, + "step": 1302 + }, + { + "epoch": 4.60017652250662, + "grad_norm": 1.3624220771981437, + "learning_rate": 5.8795095140228595e-06, + "loss": 0.0894, + "step": 1303 + }, + { + "epoch": 4.603706972639012, + "grad_norm": 1.343188625919169, + "learning_rate": 5.8738757734931975e-06, + "loss": 0.0809, + "step": 1304 + }, + { + "epoch": 4.607237422771403, + "grad_norm": 1.4446432282791357, + "learning_rate": 5.8682408883346535e-06, + "loss": 0.0766, + "step": 1305 + }, + { + "epoch": 4.6107678729037955, + "grad_norm": 1.3556894299906552, + "learning_rate": 5.86260486592797e-06, + "loss": 0.0797, + "step": 1306 + }, + { + "epoch": 4.614298323036187, + "grad_norm": 1.278747621381562, + "learning_rate": 5.856967713655379e-06, + "loss": 0.0833, + "step": 1307 + }, + { + "epoch": 4.617828773168579, + "grad_norm": 1.1918217353301246, + "learning_rate": 5.851329438900595e-06, + "loss": 0.0799, + "step": 1308 + }, + { + "epoch": 4.621359223300971, + "grad_norm": 1.1278725314966394, + "learning_rate": 5.845690049048799e-06, + "loss": 0.0643, + "step": 1309 + }, + { + "epoch": 4.624889673433363, + "grad_norm": 1.2557557372315369, + "learning_rate": 5.8400495514866355e-06, + "loss": 0.0879, + "step": 1310 + }, + { + "epoch": 4.628420123565754, + "grad_norm": 1.2461622162231867, + "learning_rate": 5.8344079536021955e-06, + "loss": 0.0809, + "step": 1311 + }, + { + "epoch": 4.631950573698147, + "grad_norm": 1.2846191093338162, + "learning_rate": 5.82876526278502e-06, + "loss": 0.0795, + "step": 1312 + }, + { + "epoch": 4.635481023830538, + "grad_norm": 1.1240893202153388, + "learning_rate": 5.823121486426069e-06, + "loss": 0.073, + "step": 1313 + }, + { + "epoch": 4.63901147396293, + "grad_norm": 1.2575691959822162, + "learning_rate": 5.817476631917736e-06, + "loss": 0.0843, + "step": 1314 + }, + { + "epoch": 4.6425419240953225, + "grad_norm": 1.017222394290193, + "learning_rate": 5.811830706653819e-06, + "loss": 0.0673, + "step": 1315 + }, + { + "epoch": 4.646072374227714, + "grad_norm": 1.0729116322930856, + "learning_rate": 5.806183718029524e-06, + "loss": 0.0732, + "step": 1316 + }, + { + "epoch": 4.649602824360106, + "grad_norm": 1.1579892771980353, + "learning_rate": 5.800535673441446e-06, + "loss": 0.0775, + "step": 1317 + }, + { + "epoch": 4.653133274492498, + "grad_norm": 1.0293530970686642, + "learning_rate": 5.794886580287565e-06, + "loss": 0.0636, + "step": 1318 + }, + { + "epoch": 4.65666372462489, + "grad_norm": 0.9376266948248516, + "learning_rate": 5.789236445967233e-06, + "loss": 0.0603, + "step": 1319 + }, + { + "epoch": 4.660194174757281, + "grad_norm": 0.9987883216154062, + "learning_rate": 5.783585277881167e-06, + "loss": 0.0762, + "step": 1320 + }, + { + "epoch": 4.6637246248896735, + "grad_norm": 1.2329836983396925, + "learning_rate": 5.777933083431437e-06, + "loss": 0.0831, + "step": 1321 + }, + { + "epoch": 4.667255075022065, + "grad_norm": 1.1400090314981657, + "learning_rate": 5.77227987002146e-06, + "loss": 0.0715, + "step": 1322 + }, + { + "epoch": 4.670785525154457, + "grad_norm": 1.108926279713076, + "learning_rate": 5.766625645055982e-06, + "loss": 0.0727, + "step": 1323 + }, + { + "epoch": 4.674315975286849, + "grad_norm": 1.036285195186968, + "learning_rate": 5.760970415941082e-06, + "loss": 0.0619, + "step": 1324 + }, + { + "epoch": 4.677846425419241, + "grad_norm": 1.2377457117763253, + "learning_rate": 5.755314190084146e-06, + "loss": 0.0679, + "step": 1325 + }, + { + "epoch": 4.681376875551633, + "grad_norm": 1.2816857024923805, + "learning_rate": 5.749656974893873e-06, + "loss": 0.0761, + "step": 1326 + }, + { + "epoch": 4.684907325684025, + "grad_norm": 1.4320175248552778, + "learning_rate": 5.743998777780252e-06, + "loss": 0.0735, + "step": 1327 + }, + { + "epoch": 4.688437775816417, + "grad_norm": 1.21805904446881, + "learning_rate": 5.73833960615456e-06, + "loss": 0.0654, + "step": 1328 + }, + { + "epoch": 4.691968225948808, + "grad_norm": 1.2720158896144405, + "learning_rate": 5.732679467429353e-06, + "loss": 0.0872, + "step": 1329 + }, + { + "epoch": 4.6954986760812005, + "grad_norm": 1.362220253742237, + "learning_rate": 5.72701836901845e-06, + "loss": 0.0763, + "step": 1330 + }, + { + "epoch": 4.699029126213592, + "grad_norm": 1.3376103897361957, + "learning_rate": 5.7213563183369295e-06, + "loss": 0.0718, + "step": 1331 + }, + { + "epoch": 4.702559576345984, + "grad_norm": 1.2102663618571865, + "learning_rate": 5.715693322801117e-06, + "loss": 0.0738, + "step": 1332 + }, + { + "epoch": 4.7060900264783765, + "grad_norm": 1.2277058279665478, + "learning_rate": 5.710029389828572e-06, + "loss": 0.0719, + "step": 1333 + }, + { + "epoch": 4.709620476610768, + "grad_norm": 1.2773556567896192, + "learning_rate": 5.70436452683809e-06, + "loss": 0.0902, + "step": 1334 + }, + { + "epoch": 4.713150926743159, + "grad_norm": 1.4332928684103645, + "learning_rate": 5.698698741249676e-06, + "loss": 0.0709, + "step": 1335 + }, + { + "epoch": 4.716681376875552, + "grad_norm": 1.1461862595511798, + "learning_rate": 5.6930320404845475e-06, + "loss": 0.0689, + "step": 1336 + }, + { + "epoch": 4.720211827007944, + "grad_norm": 1.141265802102587, + "learning_rate": 5.687364431965121e-06, + "loss": 0.0836, + "step": 1337 + }, + { + "epoch": 4.723742277140335, + "grad_norm": 1.503230303094657, + "learning_rate": 5.681695923115002e-06, + "loss": 0.0819, + "step": 1338 + }, + { + "epoch": 4.7272727272727275, + "grad_norm": 1.1626902290316583, + "learning_rate": 5.676026521358973e-06, + "loss": 0.0781, + "step": 1339 + }, + { + "epoch": 4.730803177405119, + "grad_norm": 1.1620016918977858, + "learning_rate": 5.6703562341229886e-06, + "loss": 0.0805, + "step": 1340 + }, + { + "epoch": 4.734333627537511, + "grad_norm": 0.980205333032981, + "learning_rate": 5.664685068834164e-06, + "loss": 0.06, + "step": 1341 + }, + { + "epoch": 4.737864077669903, + "grad_norm": 1.0919120901198711, + "learning_rate": 5.659013032920758e-06, + "loss": 0.0667, + "step": 1342 + }, + { + "epoch": 4.741394527802295, + "grad_norm": 1.058369377556783, + "learning_rate": 5.653340133812179e-06, + "loss": 0.0678, + "step": 1343 + }, + { + "epoch": 4.744924977934687, + "grad_norm": 1.2729504832126617, + "learning_rate": 5.647666378938959e-06, + "loss": 0.0746, + "step": 1344 + }, + { + "epoch": 4.7484554280670785, + "grad_norm": 1.2480081282387157, + "learning_rate": 5.641991775732756e-06, + "loss": 0.0711, + "step": 1345 + }, + { + "epoch": 4.75198587819947, + "grad_norm": 1.1576335449365858, + "learning_rate": 5.636316331626332e-06, + "loss": 0.074, + "step": 1346 + }, + { + "epoch": 4.755516328331862, + "grad_norm": 1.1637647202647143, + "learning_rate": 5.630640054053562e-06, + "loss": 0.0695, + "step": 1347 + }, + { + "epoch": 4.7590467784642545, + "grad_norm": 1.1265709555415027, + "learning_rate": 5.624962950449402e-06, + "loss": 0.0716, + "step": 1348 + }, + { + "epoch": 4.762577228596646, + "grad_norm": 1.2129102271686025, + "learning_rate": 5.6192850282498925e-06, + "loss": 0.0728, + "step": 1349 + }, + { + "epoch": 4.766107678729038, + "grad_norm": 1.4098311899227094, + "learning_rate": 5.6136062948921486e-06, + "loss": 0.0695, + "step": 1350 + }, + { + "epoch": 4.76963812886143, + "grad_norm": 1.0254747813352643, + "learning_rate": 5.607926757814347e-06, + "loss": 0.072, + "step": 1351 + }, + { + "epoch": 4.773168578993822, + "grad_norm": 1.3062783164855614, + "learning_rate": 5.6022464244557175e-06, + "loss": 0.0836, + "step": 1352 + }, + { + "epoch": 4.776699029126213, + "grad_norm": 1.1787618293744981, + "learning_rate": 5.596565302256531e-06, + "loss": 0.0644, + "step": 1353 + }, + { + "epoch": 4.7802294792586055, + "grad_norm": 1.4645406649621868, + "learning_rate": 5.590883398658095e-06, + "loss": 0.0786, + "step": 1354 + }, + { + "epoch": 4.783759929390998, + "grad_norm": 1.3381703719371458, + "learning_rate": 5.585200721102737e-06, + "loss": 0.0855, + "step": 1355 + }, + { + "epoch": 4.787290379523389, + "grad_norm": 1.5162385054662686, + "learning_rate": 5.579517277033799e-06, + "loss": 0.0727, + "step": 1356 + }, + { + "epoch": 4.7908208296557815, + "grad_norm": 1.333030859041725, + "learning_rate": 5.573833073895629e-06, + "loss": 0.0872, + "step": 1357 + }, + { + "epoch": 4.794351279788173, + "grad_norm": 1.2513683261705562, + "learning_rate": 5.568148119133566e-06, + "loss": 0.0661, + "step": 1358 + }, + { + "epoch": 4.797881729920565, + "grad_norm": 1.204520949412178, + "learning_rate": 5.562462420193939e-06, + "loss": 0.0769, + "step": 1359 + }, + { + "epoch": 4.801412180052957, + "grad_norm": 1.2922206835673506, + "learning_rate": 5.556775984524044e-06, + "loss": 0.0718, + "step": 1360 + }, + { + "epoch": 4.804942630185349, + "grad_norm": 1.4591640646940425, + "learning_rate": 5.55108881957215e-06, + "loss": 0.0648, + "step": 1361 + }, + { + "epoch": 4.80847308031774, + "grad_norm": 1.2720780355230428, + "learning_rate": 5.545400932787476e-06, + "loss": 0.0735, + "step": 1362 + }, + { + "epoch": 4.8120035304501325, + "grad_norm": 1.2599249791592424, + "learning_rate": 5.539712331620186e-06, + "loss": 0.062, + "step": 1363 + }, + { + "epoch": 4.815533980582524, + "grad_norm": 1.0986643518131567, + "learning_rate": 5.534023023521384e-06, + "loss": 0.0701, + "step": 1364 + }, + { + "epoch": 4.819064430714916, + "grad_norm": 1.194220635218748, + "learning_rate": 5.528333015943097e-06, + "loss": 0.0663, + "step": 1365 + }, + { + "epoch": 4.8225948808473085, + "grad_norm": 1.1823226864675451, + "learning_rate": 5.522642316338268e-06, + "loss": 0.059, + "step": 1366 + }, + { + "epoch": 4.8261253309797, + "grad_norm": 1.225355888651712, + "learning_rate": 5.516950932160748e-06, + "loss": 0.0755, + "step": 1367 + }, + { + "epoch": 4.829655781112092, + "grad_norm": 1.417142754172787, + "learning_rate": 5.511258870865283e-06, + "loss": 0.0907, + "step": 1368 + }, + { + "epoch": 4.8331862312444835, + "grad_norm": 1.315652572963653, + "learning_rate": 5.50556613990751e-06, + "loss": 0.0807, + "step": 1369 + }, + { + "epoch": 4.836716681376876, + "grad_norm": 1.4848341875180038, + "learning_rate": 5.499872746743935e-06, + "loss": 0.1056, + "step": 1370 + }, + { + "epoch": 4.840247131509267, + "grad_norm": 1.5291378895324677, + "learning_rate": 5.49417869883194e-06, + "loss": 0.0837, + "step": 1371 + }, + { + "epoch": 4.8437775816416595, + "grad_norm": 1.226839957471945, + "learning_rate": 5.488484003629759e-06, + "loss": 0.0779, + "step": 1372 + }, + { + "epoch": 4.847308031774051, + "grad_norm": 1.3463992269178928, + "learning_rate": 5.4827886685964774e-06, + "loss": 0.0976, + "step": 1373 + }, + { + "epoch": 4.850838481906443, + "grad_norm": 1.1494326044050664, + "learning_rate": 5.477092701192014e-06, + "loss": 0.0649, + "step": 1374 + }, + { + "epoch": 4.854368932038835, + "grad_norm": 1.2447420578267399, + "learning_rate": 5.471396108877123e-06, + "loss": 0.072, + "step": 1375 + }, + { + "epoch": 4.857899382171227, + "grad_norm": 1.3612258681021532, + "learning_rate": 5.465698899113371e-06, + "loss": 0.0807, + "step": 1376 + }, + { + "epoch": 4.861429832303619, + "grad_norm": 1.4193052420845504, + "learning_rate": 5.460001079363135e-06, + "loss": 0.0811, + "step": 1377 + }, + { + "epoch": 4.8649602824360105, + "grad_norm": 1.2743805042555498, + "learning_rate": 5.454302657089592e-06, + "loss": 0.08, + "step": 1378 + }, + { + "epoch": 4.868490732568403, + "grad_norm": 1.078809447494733, + "learning_rate": 5.448603639756708e-06, + "loss": 0.062, + "step": 1379 + }, + { + "epoch": 4.872021182700794, + "grad_norm": 1.0799797925482022, + "learning_rate": 5.442904034829226e-06, + "loss": 0.075, + "step": 1380 + }, + { + "epoch": 4.8755516328331865, + "grad_norm": 1.141767854824888, + "learning_rate": 5.437203849772664e-06, + "loss": 0.0636, + "step": 1381 + }, + { + "epoch": 4.879082082965578, + "grad_norm": 1.123958849288234, + "learning_rate": 5.431503092053294e-06, + "loss": 0.077, + "step": 1382 + }, + { + "epoch": 4.88261253309797, + "grad_norm": 0.9850627842600758, + "learning_rate": 5.4258017691381415e-06, + "loss": 0.0602, + "step": 1383 + }, + { + "epoch": 4.886142983230362, + "grad_norm": 1.1111371089369984, + "learning_rate": 5.420099888494972e-06, + "loss": 0.072, + "step": 1384 + }, + { + "epoch": 4.889673433362754, + "grad_norm": 1.116994328139007, + "learning_rate": 5.414397457592278e-06, + "loss": 0.0769, + "step": 1385 + }, + { + "epoch": 4.893203883495145, + "grad_norm": 1.4416102109718356, + "learning_rate": 5.408694483899281e-06, + "loss": 0.0813, + "step": 1386 + }, + { + "epoch": 4.8967343336275375, + "grad_norm": 1.4613051629608635, + "learning_rate": 5.402990974885904e-06, + "loss": 0.0821, + "step": 1387 + }, + { + "epoch": 4.90026478375993, + "grad_norm": 1.1868315059556709, + "learning_rate": 5.3972869380227765e-06, + "loss": 0.0804, + "step": 1388 + }, + { + "epoch": 4.903795233892321, + "grad_norm": 1.1666204562256506, + "learning_rate": 5.3915823807812196e-06, + "loss": 0.0785, + "step": 1389 + }, + { + "epoch": 4.9073256840247135, + "grad_norm": 1.1215733098359313, + "learning_rate": 5.385877310633233e-06, + "loss": 0.0745, + "step": 1390 + }, + { + "epoch": 4.910856134157105, + "grad_norm": 1.585401934451052, + "learning_rate": 5.38017173505149e-06, + "loss": 0.0797, + "step": 1391 + }, + { + "epoch": 4.914386584289497, + "grad_norm": 1.3919507079497462, + "learning_rate": 5.374465661509326e-06, + "loss": 0.0853, + "step": 1392 + }, + { + "epoch": 4.9179170344218885, + "grad_norm": 1.1020309883219317, + "learning_rate": 5.368759097480728e-06, + "loss": 0.0738, + "step": 1393 + }, + { + "epoch": 4.921447484554281, + "grad_norm": 1.2530954184669498, + "learning_rate": 5.363052050440327e-06, + "loss": 0.0649, + "step": 1394 + }, + { + "epoch": 4.924977934686672, + "grad_norm": 1.3563436418670745, + "learning_rate": 5.357344527863385e-06, + "loss": 0.0848, + "step": 1395 + }, + { + "epoch": 4.9285083848190645, + "grad_norm": 1.2161047750969436, + "learning_rate": 5.35163653722579e-06, + "loss": 0.0711, + "step": 1396 + }, + { + "epoch": 4.932038834951456, + "grad_norm": 1.1405877554596378, + "learning_rate": 5.345928086004034e-06, + "loss": 0.0759, + "step": 1397 + }, + { + "epoch": 4.935569285083848, + "grad_norm": 1.5340642291990807, + "learning_rate": 5.340219181675227e-06, + "loss": 0.0756, + "step": 1398 + }, + { + "epoch": 4.9390997352162405, + "grad_norm": 1.4009707021891677, + "learning_rate": 5.334509831717058e-06, + "loss": 0.0773, + "step": 1399 + }, + { + "epoch": 4.942630185348632, + "grad_norm": 1.2971256965481643, + "learning_rate": 5.328800043607809e-06, + "loss": 0.0891, + "step": 1400 + }, + { + "epoch": 4.946160635481024, + "grad_norm": 1.378471603765613, + "learning_rate": 5.323089824826332e-06, + "loss": 0.0812, + "step": 1401 + }, + { + "epoch": 4.9496910856134155, + "grad_norm": 0.9246558166305276, + "learning_rate": 5.317379182852044e-06, + "loss": 0.0652, + "step": 1402 + }, + { + "epoch": 4.953221535745808, + "grad_norm": 1.1283491586166101, + "learning_rate": 5.311668125164916e-06, + "loss": 0.0747, + "step": 1403 + }, + { + "epoch": 4.956751985878199, + "grad_norm": 1.1947879187764758, + "learning_rate": 5.305956659245466e-06, + "loss": 0.0871, + "step": 1404 + }, + { + "epoch": 4.9602824360105915, + "grad_norm": 1.1562516815747672, + "learning_rate": 5.300244792574743e-06, + "loss": 0.0818, + "step": 1405 + }, + { + "epoch": 4.963812886142983, + "grad_norm": 1.2658874903814799, + "learning_rate": 5.294532532634321e-06, + "loss": 0.0784, + "step": 1406 + }, + { + "epoch": 4.967343336275375, + "grad_norm": 1.0490558876913674, + "learning_rate": 5.288819886906293e-06, + "loss": 0.0618, + "step": 1407 + }, + { + "epoch": 4.970873786407767, + "grad_norm": 1.2147121704532486, + "learning_rate": 5.283106862873253e-06, + "loss": 0.0799, + "step": 1408 + }, + { + "epoch": 4.974404236540159, + "grad_norm": 1.1451504862662256, + "learning_rate": 5.2773934680182935e-06, + "loss": 0.0668, + "step": 1409 + }, + { + "epoch": 4.977934686672551, + "grad_norm": 1.1648655228082387, + "learning_rate": 5.271679709824993e-06, + "loss": 0.0742, + "step": 1410 + }, + { + "epoch": 4.9814651368049425, + "grad_norm": 1.1046312634759032, + "learning_rate": 5.265965595777402e-06, + "loss": 0.0814, + "step": 1411 + }, + { + "epoch": 4.984995586937335, + "grad_norm": 1.1273037101841474, + "learning_rate": 5.2602511333600414e-06, + "loss": 0.0785, + "step": 1412 + }, + { + "epoch": 4.988526037069726, + "grad_norm": 1.354622519190956, + "learning_rate": 5.254536330057884e-06, + "loss": 0.0785, + "step": 1413 + }, + { + "epoch": 4.9920564872021185, + "grad_norm": 1.3583127081670356, + "learning_rate": 5.248821193356356e-06, + "loss": 0.0789, + "step": 1414 + }, + { + "epoch": 4.99558693733451, + "grad_norm": 1.2297540102464188, + "learning_rate": 5.243105730741312e-06, + "loss": 0.0782, + "step": 1415 + }, + { + "epoch": 4.999117387466902, + "grad_norm": 1.1685725176323518, + "learning_rate": 5.237389949699042e-06, + "loss": 0.0617, + "step": 1416 + }, + { + "epoch": 5.0026478375992935, + "grad_norm": 0.8257805889567651, + "learning_rate": 5.231673857716244e-06, + "loss": 0.0429, + "step": 1417 + }, + { + "epoch": 5.006178287731686, + "grad_norm": 0.8199217627638553, + "learning_rate": 5.2259574622800305e-06, + "loss": 0.0345, + "step": 1418 + }, + { + "epoch": 5.009708737864078, + "grad_norm": 0.6795639401534437, + "learning_rate": 5.220240770877909e-06, + "loss": 0.0372, + "step": 1419 + }, + { + "epoch": 5.0132391879964695, + "grad_norm": 0.8501746015877016, + "learning_rate": 5.214523790997773e-06, + "loss": 0.0377, + "step": 1420 + }, + { + "epoch": 5.016769638128862, + "grad_norm": 0.8824465805765227, + "learning_rate": 5.208806530127897e-06, + "loss": 0.0373, + "step": 1421 + }, + { + "epoch": 5.020300088261253, + "grad_norm": 0.9205208121497219, + "learning_rate": 5.203088995756921e-06, + "loss": 0.0341, + "step": 1422 + }, + { + "epoch": 5.0238305383936455, + "grad_norm": 0.9604765129499832, + "learning_rate": 5.197371195373843e-06, + "loss": 0.0489, + "step": 1423 + }, + { + "epoch": 5.027360988526037, + "grad_norm": 0.809893364630281, + "learning_rate": 5.1916531364680145e-06, + "loss": 0.0341, + "step": 1424 + }, + { + "epoch": 5.030891438658429, + "grad_norm": 0.7211672164738381, + "learning_rate": 5.1859348265291174e-06, + "loss": 0.0297, + "step": 1425 + }, + { + "epoch": 5.0344218887908205, + "grad_norm": 1.202591035738676, + "learning_rate": 5.1802162730471704e-06, + "loss": 0.0358, + "step": 1426 + }, + { + "epoch": 5.037952338923213, + "grad_norm": 1.6083152186860212, + "learning_rate": 5.174497483512506e-06, + "loss": 0.0415, + "step": 1427 + }, + { + "epoch": 5.041482789055604, + "grad_norm": 1.0297003747245823, + "learning_rate": 5.168778465415767e-06, + "loss": 0.0319, + "step": 1428 + }, + { + "epoch": 5.0450132391879965, + "grad_norm": 0.8436108989916318, + "learning_rate": 5.163059226247895e-06, + "loss": 0.0318, + "step": 1429 + }, + { + "epoch": 5.048543689320389, + "grad_norm": 0.9865353291360064, + "learning_rate": 5.1573397735001254e-06, + "loss": 0.0342, + "step": 1430 + }, + { + "epoch": 5.05207413945278, + "grad_norm": 1.1707497654997088, + "learning_rate": 5.151620114663968e-06, + "loss": 0.0288, + "step": 1431 + }, + { + "epoch": 5.0556045895851724, + "grad_norm": 1.1194925914169904, + "learning_rate": 5.145900257231204e-06, + "loss": 0.0177, + "step": 1432 + }, + { + "epoch": 5.059135039717564, + "grad_norm": 1.0003521185971491, + "learning_rate": 5.140180208693878e-06, + "loss": 0.0279, + "step": 1433 + }, + { + "epoch": 5.062665489849956, + "grad_norm": 1.258560532803691, + "learning_rate": 5.134459976544279e-06, + "loss": 0.0338, + "step": 1434 + }, + { + "epoch": 5.0661959399823475, + "grad_norm": 1.3446204618289492, + "learning_rate": 5.1287395682749444e-06, + "loss": 0.0414, + "step": 1435 + }, + { + "epoch": 5.06972639011474, + "grad_norm": 0.9660695418598767, + "learning_rate": 5.123018991378632e-06, + "loss": 0.0265, + "step": 1436 + }, + { + "epoch": 5.073256840247131, + "grad_norm": 1.2366964428685454, + "learning_rate": 5.117298253348331e-06, + "loss": 0.0352, + "step": 1437 + }, + { + "epoch": 5.0767872903795235, + "grad_norm": 1.8279426747052296, + "learning_rate": 5.111577361677235e-06, + "loss": 0.0318, + "step": 1438 + }, + { + "epoch": 5.080317740511915, + "grad_norm": 1.1506985736787303, + "learning_rate": 5.105856323858739e-06, + "loss": 0.0258, + "step": 1439 + }, + { + "epoch": 5.083848190644307, + "grad_norm": 1.4246107723443555, + "learning_rate": 5.1001351473864344e-06, + "loss": 0.0295, + "step": 1440 + }, + { + "epoch": 5.087378640776699, + "grad_norm": 0.8822363113436165, + "learning_rate": 5.0944138397540895e-06, + "loss": 0.0279, + "step": 1441 + }, + { + "epoch": 5.090909090909091, + "grad_norm": 1.395559247506785, + "learning_rate": 5.088692408455646e-06, + "loss": 0.0333, + "step": 1442 + }, + { + "epoch": 5.094439541041483, + "grad_norm": 1.5182045973236369, + "learning_rate": 5.082970860985206e-06, + "loss": 0.0296, + "step": 1443 + }, + { + "epoch": 5.0979699911738745, + "grad_norm": 0.9242573289174264, + "learning_rate": 5.077249204837026e-06, + "loss": 0.016, + "step": 1444 + }, + { + "epoch": 5.101500441306267, + "grad_norm": 1.936773205826491, + "learning_rate": 5.071527447505506e-06, + "loss": 0.0435, + "step": 1445 + }, + { + "epoch": 5.105030891438658, + "grad_norm": 1.4767038310878702, + "learning_rate": 5.065805596485173e-06, + "loss": 0.0299, + "step": 1446 + }, + { + "epoch": 5.1085613415710505, + "grad_norm": 1.263839063646782, + "learning_rate": 5.060083659270685e-06, + "loss": 0.0388, + "step": 1447 + }, + { + "epoch": 5.112091791703442, + "grad_norm": 1.3779652528942368, + "learning_rate": 5.0543616433568035e-06, + "loss": 0.0306, + "step": 1448 + }, + { + "epoch": 5.115622241835834, + "grad_norm": 1.5639948639539878, + "learning_rate": 5.048639556238399e-06, + "loss": 0.0377, + "step": 1449 + }, + { + "epoch": 5.1191526919682255, + "grad_norm": 1.3453580254033097, + "learning_rate": 5.042917405410436e-06, + "loss": 0.0335, + "step": 1450 + }, + { + "epoch": 5.122683142100618, + "grad_norm": 1.0811653047788332, + "learning_rate": 5.037195198367959e-06, + "loss": 0.0324, + "step": 1451 + }, + { + "epoch": 5.12621359223301, + "grad_norm": 1.339443531084938, + "learning_rate": 5.031472942606087e-06, + "loss": 0.04, + "step": 1452 + }, + { + "epoch": 5.1297440423654015, + "grad_norm": 1.0627572674086125, + "learning_rate": 5.025750645620004e-06, + "loss": 0.0261, + "step": 1453 + }, + { + "epoch": 5.133274492497794, + "grad_norm": 1.4586995836962917, + "learning_rate": 5.02002831490495e-06, + "loss": 0.0232, + "step": 1454 + }, + { + "epoch": 5.136804942630185, + "grad_norm": 1.4485310435323844, + "learning_rate": 5.014305957956203e-06, + "loss": 0.0371, + "step": 1455 + }, + { + "epoch": 5.1403353927625774, + "grad_norm": 1.6366978052444436, + "learning_rate": 5.008583582269078e-06, + "loss": 0.0312, + "step": 1456 + }, + { + "epoch": 5.143865842894969, + "grad_norm": 1.2940871351216503, + "learning_rate": 5.002861195338921e-06, + "loss": 0.0388, + "step": 1457 + }, + { + "epoch": 5.147396293027361, + "grad_norm": 1.0900574715419957, + "learning_rate": 4.997138804661082e-06, + "loss": 0.0366, + "step": 1458 + }, + { + "epoch": 5.1509267431597525, + "grad_norm": 1.0267178840790936, + "learning_rate": 4.991416417730924e-06, + "loss": 0.0261, + "step": 1459 + }, + { + "epoch": 5.154457193292145, + "grad_norm": 1.070899041255336, + "learning_rate": 4.9856940420438e-06, + "loss": 0.0222, + "step": 1460 + }, + { + "epoch": 5.157987643424537, + "grad_norm": 1.218928833068551, + "learning_rate": 4.979971685095053e-06, + "loss": 0.0355, + "step": 1461 + }, + { + "epoch": 5.1615180935569285, + "grad_norm": 1.2131609735900892, + "learning_rate": 4.974249354379997e-06, + "loss": 0.0351, + "step": 1462 + }, + { + "epoch": 5.165048543689321, + "grad_norm": 1.037489526125749, + "learning_rate": 4.9685270573939145e-06, + "loss": 0.0304, + "step": 1463 + }, + { + "epoch": 5.168578993821712, + "grad_norm": 1.0980343468796883, + "learning_rate": 4.9628048016320425e-06, + "loss": 0.0284, + "step": 1464 + }, + { + "epoch": 5.172109443954104, + "grad_norm": 1.25052893487193, + "learning_rate": 4.957082594589566e-06, + "loss": 0.0284, + "step": 1465 + }, + { + "epoch": 5.175639894086496, + "grad_norm": 1.0638136569160561, + "learning_rate": 4.9513604437616015e-06, + "loss": 0.0364, + "step": 1466 + }, + { + "epoch": 5.179170344218888, + "grad_norm": 1.5179351058052895, + "learning_rate": 4.945638356643198e-06, + "loss": 0.0377, + "step": 1467 + }, + { + "epoch": 5.1827007943512795, + "grad_norm": 1.3977740297264558, + "learning_rate": 4.939916340729317e-06, + "loss": 0.0379, + "step": 1468 + }, + { + "epoch": 5.186231244483672, + "grad_norm": 1.0481389634796345, + "learning_rate": 4.934194403514827e-06, + "loss": 0.0322, + "step": 1469 + }, + { + "epoch": 5.189761694616063, + "grad_norm": 1.650605995342595, + "learning_rate": 4.928472552494494e-06, + "loss": 0.0341, + "step": 1470 + }, + { + "epoch": 5.1932921447484555, + "grad_norm": 0.8149965364021505, + "learning_rate": 4.922750795162974e-06, + "loss": 0.0225, + "step": 1471 + }, + { + "epoch": 5.196822594880848, + "grad_norm": 0.9138026684992125, + "learning_rate": 4.9170291390147964e-06, + "loss": 0.0229, + "step": 1472 + }, + { + "epoch": 5.200353045013239, + "grad_norm": 0.9497975937605081, + "learning_rate": 4.911307591544357e-06, + "loss": 0.026, + "step": 1473 + }, + { + "epoch": 5.203883495145631, + "grad_norm": 1.1533504016571094, + "learning_rate": 4.905586160245913e-06, + "loss": 0.0358, + "step": 1474 + }, + { + "epoch": 5.207413945278023, + "grad_norm": 1.0551140620172796, + "learning_rate": 4.899864852613566e-06, + "loss": 0.0328, + "step": 1475 + }, + { + "epoch": 5.210944395410415, + "grad_norm": 1.2335519280933265, + "learning_rate": 4.894143676141262e-06, + "loss": 0.0356, + "step": 1476 + }, + { + "epoch": 5.2144748455428065, + "grad_norm": 1.173364956196987, + "learning_rate": 4.888422638322767e-06, + "loss": 0.0452, + "step": 1477 + }, + { + "epoch": 5.218005295675199, + "grad_norm": 1.180983301454911, + "learning_rate": 4.882701746651671e-06, + "loss": 0.0408, + "step": 1478 + }, + { + "epoch": 5.22153574580759, + "grad_norm": 1.1966562535053094, + "learning_rate": 4.87698100862137e-06, + "loss": 0.0326, + "step": 1479 + }, + { + "epoch": 5.2250661959399824, + "grad_norm": 1.338990761316199, + "learning_rate": 4.871260431725058e-06, + "loss": 0.0329, + "step": 1480 + }, + { + "epoch": 5.228596646072374, + "grad_norm": 1.6702333512118446, + "learning_rate": 4.865540023455722e-06, + "loss": 0.0437, + "step": 1481 + }, + { + "epoch": 5.232127096204766, + "grad_norm": 1.3692897235522883, + "learning_rate": 4.859819791306124e-06, + "loss": 0.0444, + "step": 1482 + }, + { + "epoch": 5.235657546337158, + "grad_norm": 1.1576589977769405, + "learning_rate": 4.854099742768797e-06, + "loss": 0.0318, + "step": 1483 + }, + { + "epoch": 5.23918799646955, + "grad_norm": 1.1872720185083636, + "learning_rate": 4.848379885336033e-06, + "loss": 0.0435, + "step": 1484 + }, + { + "epoch": 5.242718446601942, + "grad_norm": 0.9514120114110632, + "learning_rate": 4.842660226499875e-06, + "loss": 0.0283, + "step": 1485 + }, + { + "epoch": 5.2462488967343335, + "grad_norm": 1.146999340750997, + "learning_rate": 4.8369407737521075e-06, + "loss": 0.0433, + "step": 1486 + }, + { + "epoch": 5.249779346866726, + "grad_norm": 1.1236633423224671, + "learning_rate": 4.831221534584236e-06, + "loss": 0.0267, + "step": 1487 + }, + { + "epoch": 5.253309796999117, + "grad_norm": 0.9158462075377869, + "learning_rate": 4.825502516487497e-06, + "loss": 0.0303, + "step": 1488 + }, + { + "epoch": 5.256840247131509, + "grad_norm": 1.6495025653791489, + "learning_rate": 4.819783726952831e-06, + "loss": 0.0598, + "step": 1489 + }, + { + "epoch": 5.260370697263901, + "grad_norm": 1.3019350952285167, + "learning_rate": 4.814065173470883e-06, + "loss": 0.0351, + "step": 1490 + }, + { + "epoch": 5.263901147396293, + "grad_norm": 1.0264285827572417, + "learning_rate": 4.808346863531987e-06, + "loss": 0.0376, + "step": 1491 + }, + { + "epoch": 5.2674315975286845, + "grad_norm": 1.2892284013851287, + "learning_rate": 4.802628804626158e-06, + "loss": 0.0285, + "step": 1492 + }, + { + "epoch": 5.270962047661077, + "grad_norm": 1.2182684563778106, + "learning_rate": 4.796911004243081e-06, + "loss": 0.0321, + "step": 1493 + }, + { + "epoch": 5.274492497793469, + "grad_norm": 1.5038347502170855, + "learning_rate": 4.791193469872105e-06, + "loss": 0.0404, + "step": 1494 + }, + { + "epoch": 5.2780229479258605, + "grad_norm": 1.146484453359854, + "learning_rate": 4.7854762090022274e-06, + "loss": 0.0332, + "step": 1495 + }, + { + "epoch": 5.281553398058253, + "grad_norm": 1.0414147154504831, + "learning_rate": 4.779759229122093e-06, + "loss": 0.0325, + "step": 1496 + }, + { + "epoch": 5.285083848190644, + "grad_norm": 1.2133343457927983, + "learning_rate": 4.77404253771997e-06, + "loss": 0.0321, + "step": 1497 + }, + { + "epoch": 5.288614298323036, + "grad_norm": 1.578530748046948, + "learning_rate": 4.768326142283757e-06, + "loss": 0.0398, + "step": 1498 + }, + { + "epoch": 5.292144748455428, + "grad_norm": 1.11987396824221, + "learning_rate": 4.762610050300959e-06, + "loss": 0.0293, + "step": 1499 + }, + { + "epoch": 5.29567519858782, + "grad_norm": 1.4485778387815824, + "learning_rate": 4.756894269258688e-06, + "loss": 0.0313, + "step": 1500 + }, + { + "epoch": 5.2992056487202115, + "grad_norm": 1.1628182354166519, + "learning_rate": 4.751178806643646e-06, + "loss": 0.0334, + "step": 1501 + }, + { + "epoch": 5.302736098852604, + "grad_norm": 1.1888595900741459, + "learning_rate": 4.745463669942118e-06, + "loss": 0.0331, + "step": 1502 + }, + { + "epoch": 5.306266548984995, + "grad_norm": 0.9872037193786444, + "learning_rate": 4.739748866639961e-06, + "loss": 0.0229, + "step": 1503 + }, + { + "epoch": 5.3097969991173875, + "grad_norm": 1.595605980821677, + "learning_rate": 4.734034404222601e-06, + "loss": 0.0523, + "step": 1504 + }, + { + "epoch": 5.31332744924978, + "grad_norm": 1.1821169330839858, + "learning_rate": 4.7283202901750095e-06, + "loss": 0.0449, + "step": 1505 + }, + { + "epoch": 5.316857899382171, + "grad_norm": 1.3511834704411476, + "learning_rate": 4.722606531981707e-06, + "loss": 0.0407, + "step": 1506 + }, + { + "epoch": 5.320388349514563, + "grad_norm": 1.0921292172210884, + "learning_rate": 4.716893137126748e-06, + "loss": 0.041, + "step": 1507 + }, + { + "epoch": 5.323918799646955, + "grad_norm": 0.9142044634169543, + "learning_rate": 4.711180113093709e-06, + "loss": 0.0329, + "step": 1508 + }, + { + "epoch": 5.327449249779347, + "grad_norm": 1.1512944606541926, + "learning_rate": 4.70546746736568e-06, + "loss": 0.0354, + "step": 1509 + }, + { + "epoch": 5.3309796999117385, + "grad_norm": 1.2325212146838187, + "learning_rate": 4.699755207425259e-06, + "loss": 0.0355, + "step": 1510 + }, + { + "epoch": 5.334510150044131, + "grad_norm": 1.1782174747122032, + "learning_rate": 4.694043340754535e-06, + "loss": 0.0273, + "step": 1511 + }, + { + "epoch": 5.338040600176522, + "grad_norm": 1.3703485363576169, + "learning_rate": 4.6883318748350846e-06, + "loss": 0.0376, + "step": 1512 + }, + { + "epoch": 5.341571050308914, + "grad_norm": 1.0893592430199348, + "learning_rate": 4.682620817147956e-06, + "loss": 0.0282, + "step": 1513 + }, + { + "epoch": 5.345101500441306, + "grad_norm": 1.0033637518271674, + "learning_rate": 4.676910175173669e-06, + "loss": 0.0291, + "step": 1514 + }, + { + "epoch": 5.348631950573698, + "grad_norm": 1.2921249721136994, + "learning_rate": 4.671199956392194e-06, + "loss": 0.0447, + "step": 1515 + }, + { + "epoch": 5.35216240070609, + "grad_norm": 0.9392181152529292, + "learning_rate": 4.665490168282943e-06, + "loss": 0.0312, + "step": 1516 + }, + { + "epoch": 5.355692850838482, + "grad_norm": 1.1681917089042493, + "learning_rate": 4.659780818324775e-06, + "loss": 0.0388, + "step": 1517 + }, + { + "epoch": 5.359223300970874, + "grad_norm": 1.358092988699173, + "learning_rate": 4.654071913995967e-06, + "loss": 0.0297, + "step": 1518 + }, + { + "epoch": 5.3627537511032655, + "grad_norm": 1.2624061159291087, + "learning_rate": 4.648363462774213e-06, + "loss": 0.0339, + "step": 1519 + }, + { + "epoch": 5.366284201235658, + "grad_norm": 1.3477280900652608, + "learning_rate": 4.642655472136616e-06, + "loss": 0.0415, + "step": 1520 + }, + { + "epoch": 5.369814651368049, + "grad_norm": 1.4129557155678707, + "learning_rate": 4.636947949559673e-06, + "loss": 0.0418, + "step": 1521 + }, + { + "epoch": 5.373345101500441, + "grad_norm": 1.1720035992860682, + "learning_rate": 4.631240902519274e-06, + "loss": 0.0294, + "step": 1522 + }, + { + "epoch": 5.376875551632833, + "grad_norm": 1.2030839625483056, + "learning_rate": 4.625534338490675e-06, + "loss": 0.042, + "step": 1523 + }, + { + "epoch": 5.380406001765225, + "grad_norm": 1.4135472532512512, + "learning_rate": 4.619828264948512e-06, + "loss": 0.0504, + "step": 1524 + }, + { + "epoch": 5.3839364518976165, + "grad_norm": 1.230800021198084, + "learning_rate": 4.614122689366769e-06, + "loss": 0.0296, + "step": 1525 + }, + { + "epoch": 5.387466902030009, + "grad_norm": 1.3910622631166456, + "learning_rate": 4.608417619218782e-06, + "loss": 0.0361, + "step": 1526 + }, + { + "epoch": 5.390997352162401, + "grad_norm": 1.154994374471451, + "learning_rate": 4.6027130619772234e-06, + "loss": 0.0365, + "step": 1527 + }, + { + "epoch": 5.3945278022947925, + "grad_norm": 1.4583007850410203, + "learning_rate": 4.597009025114096e-06, + "loss": 0.0334, + "step": 1528 + }, + { + "epoch": 5.398058252427185, + "grad_norm": 1.0217585804543807, + "learning_rate": 4.591305516100721e-06, + "loss": 0.0288, + "step": 1529 + }, + { + "epoch": 5.401588702559576, + "grad_norm": 1.5658199832092448, + "learning_rate": 4.585602542407723e-06, + "loss": 0.0413, + "step": 1530 + }, + { + "epoch": 5.405119152691968, + "grad_norm": 1.0594187515491122, + "learning_rate": 4.579900111505031e-06, + "loss": 0.0242, + "step": 1531 + }, + { + "epoch": 5.40864960282436, + "grad_norm": 1.2528595289385047, + "learning_rate": 4.57419823086186e-06, + "loss": 0.0248, + "step": 1532 + }, + { + "epoch": 5.412180052956752, + "grad_norm": 1.2539031057166354, + "learning_rate": 4.568496907946708e-06, + "loss": 0.0316, + "step": 1533 + }, + { + "epoch": 5.4157105030891435, + "grad_norm": 1.0230340221613308, + "learning_rate": 4.562796150227337e-06, + "loss": 0.026, + "step": 1534 + }, + { + "epoch": 5.419240953221536, + "grad_norm": 1.0810927793088345, + "learning_rate": 4.557095965170775e-06, + "loss": 0.0333, + "step": 1535 + }, + { + "epoch": 5.422771403353927, + "grad_norm": 1.109625244299689, + "learning_rate": 4.551396360243294e-06, + "loss": 0.0403, + "step": 1536 + }, + { + "epoch": 5.426301853486319, + "grad_norm": 1.07732729782522, + "learning_rate": 4.54569734291041e-06, + "loss": 0.0348, + "step": 1537 + }, + { + "epoch": 5.429832303618712, + "grad_norm": 1.2147514018439594, + "learning_rate": 4.539998920636866e-06, + "loss": 0.0347, + "step": 1538 + }, + { + "epoch": 5.433362753751103, + "grad_norm": 1.2084983817851274, + "learning_rate": 4.53430110088663e-06, + "loss": 0.043, + "step": 1539 + }, + { + "epoch": 5.436893203883495, + "grad_norm": 1.485440381945466, + "learning_rate": 4.528603891122878e-06, + "loss": 0.0432, + "step": 1540 + }, + { + "epoch": 5.440423654015887, + "grad_norm": 1.0939190277342912, + "learning_rate": 4.522907298807986e-06, + "loss": 0.0295, + "step": 1541 + }, + { + "epoch": 5.443954104148279, + "grad_norm": 1.1173523434000188, + "learning_rate": 4.517211331403524e-06, + "loss": 0.0288, + "step": 1542 + }, + { + "epoch": 5.4474845542806705, + "grad_norm": 1.2894250461462553, + "learning_rate": 4.511515996370244e-06, + "loss": 0.0328, + "step": 1543 + }, + { + "epoch": 5.451015004413063, + "grad_norm": 1.3974026499640995, + "learning_rate": 4.5058213011680626e-06, + "loss": 0.0452, + "step": 1544 + }, + { + "epoch": 5.454545454545454, + "grad_norm": 1.1745924829538836, + "learning_rate": 4.5001272532560665e-06, + "loss": 0.0286, + "step": 1545 + }, + { + "epoch": 5.458075904677846, + "grad_norm": 1.3328331786068008, + "learning_rate": 4.494433860092493e-06, + "loss": 0.0245, + "step": 1546 + }, + { + "epoch": 5.461606354810239, + "grad_norm": 1.5297987737670309, + "learning_rate": 4.488741129134718e-06, + "loss": 0.0468, + "step": 1547 + }, + { + "epoch": 5.46513680494263, + "grad_norm": 1.2620300857408455, + "learning_rate": 4.483049067839253e-06, + "loss": 0.0342, + "step": 1548 + }, + { + "epoch": 5.468667255075022, + "grad_norm": 1.3626059834217417, + "learning_rate": 4.477357683661734e-06, + "loss": 0.0345, + "step": 1549 + }, + { + "epoch": 5.472197705207414, + "grad_norm": 1.2893295205393553, + "learning_rate": 4.471666984056904e-06, + "loss": 0.0352, + "step": 1550 + }, + { + "epoch": 5.475728155339806, + "grad_norm": 1.092230821740141, + "learning_rate": 4.465976976478617e-06, + "loss": 0.0265, + "step": 1551 + }, + { + "epoch": 5.4792586054721975, + "grad_norm": 1.0444735170471344, + "learning_rate": 4.460287668379815e-06, + "loss": 0.0294, + "step": 1552 + }, + { + "epoch": 5.48278905560459, + "grad_norm": 0.8608039885250791, + "learning_rate": 4.454599067212526e-06, + "loss": 0.0211, + "step": 1553 + }, + { + "epoch": 5.486319505736981, + "grad_norm": 1.1171371234521095, + "learning_rate": 4.44891118042785e-06, + "loss": 0.0334, + "step": 1554 + }, + { + "epoch": 5.489849955869373, + "grad_norm": 1.3743958463995387, + "learning_rate": 4.443224015475956e-06, + "loss": 0.038, + "step": 1555 + }, + { + "epoch": 5.493380406001766, + "grad_norm": 1.4098411452671888, + "learning_rate": 4.4375375798060615e-06, + "loss": 0.0383, + "step": 1556 + }, + { + "epoch": 5.496910856134157, + "grad_norm": 1.004506255513621, + "learning_rate": 4.431851880866434e-06, + "loss": 0.0297, + "step": 1557 + }, + { + "epoch": 5.500441306266549, + "grad_norm": 1.0735711330953728, + "learning_rate": 4.426166926104374e-06, + "loss": 0.035, + "step": 1558 + }, + { + "epoch": 5.503971756398941, + "grad_norm": 1.1539255627869378, + "learning_rate": 4.420482722966204e-06, + "loss": 0.0293, + "step": 1559 + }, + { + "epoch": 5.507502206531333, + "grad_norm": 1.333537661254262, + "learning_rate": 4.414799278897265e-06, + "loss": 0.0462, + "step": 1560 + }, + { + "epoch": 5.511032656663724, + "grad_norm": 1.5317935595283634, + "learning_rate": 4.409116601341908e-06, + "loss": 0.0413, + "step": 1561 + }, + { + "epoch": 5.514563106796117, + "grad_norm": 0.9407897698825937, + "learning_rate": 4.40343469774347e-06, + "loss": 0.0364, + "step": 1562 + }, + { + "epoch": 5.518093556928508, + "grad_norm": 1.390133537237101, + "learning_rate": 4.397753575544284e-06, + "loss": 0.0361, + "step": 1563 + }, + { + "epoch": 5.5216240070609, + "grad_norm": 1.0828139489400255, + "learning_rate": 4.3920732421856535e-06, + "loss": 0.0311, + "step": 1564 + }, + { + "epoch": 5.525154457193292, + "grad_norm": 1.5186830012147863, + "learning_rate": 4.386393705107853e-06, + "loss": 0.0403, + "step": 1565 + }, + { + "epoch": 5.528684907325684, + "grad_norm": 1.2324115224643393, + "learning_rate": 4.380714971750109e-06, + "loss": 0.0294, + "step": 1566 + }, + { + "epoch": 5.532215357458076, + "grad_norm": 0.9975288959363068, + "learning_rate": 4.3750370495506e-06, + "loss": 0.0267, + "step": 1567 + }, + { + "epoch": 5.535745807590468, + "grad_norm": 1.105707284209736, + "learning_rate": 4.369359945946438e-06, + "loss": 0.0368, + "step": 1568 + }, + { + "epoch": 5.53927625772286, + "grad_norm": 1.282931081461775, + "learning_rate": 4.363683668373668e-06, + "loss": 0.0323, + "step": 1569 + }, + { + "epoch": 5.542806707855251, + "grad_norm": 1.2368549483830427, + "learning_rate": 4.358008224267245e-06, + "loss": 0.0388, + "step": 1570 + }, + { + "epoch": 5.546337157987644, + "grad_norm": 1.0516207620976028, + "learning_rate": 4.352333621061042e-06, + "loss": 0.0381, + "step": 1571 + }, + { + "epoch": 5.549867608120035, + "grad_norm": 0.9213821614358485, + "learning_rate": 4.346659866187824e-06, + "loss": 0.023, + "step": 1572 + }, + { + "epoch": 5.553398058252427, + "grad_norm": 1.2326810195773108, + "learning_rate": 4.340986967079244e-06, + "loss": 0.0397, + "step": 1573 + }, + { + "epoch": 5.556928508384819, + "grad_norm": 1.1037869710766632, + "learning_rate": 4.33531493116584e-06, + "loss": 0.0352, + "step": 1574 + }, + { + "epoch": 5.560458958517211, + "grad_norm": 1.4887296909179542, + "learning_rate": 4.329643765877012e-06, + "loss": 0.0378, + "step": 1575 + }, + { + "epoch": 5.5639894086496025, + "grad_norm": 0.8387454624239772, + "learning_rate": 4.323973478641028e-06, + "loss": 0.0203, + "step": 1576 + }, + { + "epoch": 5.567519858781995, + "grad_norm": 1.3162273803222384, + "learning_rate": 4.318304076884999e-06, + "loss": 0.0425, + "step": 1577 + }, + { + "epoch": 5.571050308914387, + "grad_norm": 1.6143182006142627, + "learning_rate": 4.312635568034879e-06, + "loss": 0.034, + "step": 1578 + }, + { + "epoch": 5.574580759046778, + "grad_norm": 0.842809991267627, + "learning_rate": 4.306967959515454e-06, + "loss": 0.0282, + "step": 1579 + }, + { + "epoch": 5.578111209179171, + "grad_norm": 1.5840796918324767, + "learning_rate": 4.301301258750325e-06, + "loss": 0.0345, + "step": 1580 + }, + { + "epoch": 5.581641659311562, + "grad_norm": 0.9886258766069115, + "learning_rate": 4.295635473161912e-06, + "loss": 0.0261, + "step": 1581 + }, + { + "epoch": 5.585172109443954, + "grad_norm": 1.1976500442964986, + "learning_rate": 4.289970610171428e-06, + "loss": 0.0339, + "step": 1582 + }, + { + "epoch": 5.588702559576346, + "grad_norm": 1.350500394870302, + "learning_rate": 4.284306677198885e-06, + "loss": 0.0362, + "step": 1583 + }, + { + "epoch": 5.592233009708738, + "grad_norm": 1.0870411651838163, + "learning_rate": 4.278643681663071e-06, + "loss": 0.0347, + "step": 1584 + }, + { + "epoch": 5.595763459841129, + "grad_norm": 1.3440201584435856, + "learning_rate": 4.272981630981551e-06, + "loss": 0.0437, + "step": 1585 + }, + { + "epoch": 5.599293909973522, + "grad_norm": 1.2362974360921064, + "learning_rate": 4.26732053257065e-06, + "loss": 0.029, + "step": 1586 + }, + { + "epoch": 5.602824360105913, + "grad_norm": 1.24379998659571, + "learning_rate": 4.261660393845441e-06, + "loss": 0.0369, + "step": 1587 + }, + { + "epoch": 5.606354810238305, + "grad_norm": 1.1929867048652107, + "learning_rate": 4.256001222219751e-06, + "loss": 0.0344, + "step": 1588 + }, + { + "epoch": 5.609885260370698, + "grad_norm": 1.1546309189588573, + "learning_rate": 4.250343025106129e-06, + "loss": 0.0338, + "step": 1589 + }, + { + "epoch": 5.613415710503089, + "grad_norm": 0.9817180886936909, + "learning_rate": 4.244685809915855e-06, + "loss": 0.0264, + "step": 1590 + }, + { + "epoch": 5.616946160635481, + "grad_norm": 1.22779968851127, + "learning_rate": 4.23902958405892e-06, + "loss": 0.0297, + "step": 1591 + }, + { + "epoch": 5.620476610767873, + "grad_norm": 1.277605114754718, + "learning_rate": 4.233374354944019e-06, + "loss": 0.0353, + "step": 1592 + }, + { + "epoch": 5.624007060900265, + "grad_norm": 1.1376160372982294, + "learning_rate": 4.2277201299785414e-06, + "loss": 0.0384, + "step": 1593 + }, + { + "epoch": 5.627537511032656, + "grad_norm": 0.9624800823947393, + "learning_rate": 4.222066916568564e-06, + "loss": 0.0252, + "step": 1594 + }, + { + "epoch": 5.631067961165049, + "grad_norm": 0.9439407579224949, + "learning_rate": 4.2164147221188335e-06, + "loss": 0.026, + "step": 1595 + }, + { + "epoch": 5.63459841129744, + "grad_norm": 1.2337049870840913, + "learning_rate": 4.210763554032768e-06, + "loss": 0.0358, + "step": 1596 + }, + { + "epoch": 5.638128861429832, + "grad_norm": 1.5874250323992236, + "learning_rate": 4.2051134197124354e-06, + "loss": 0.0594, + "step": 1597 + }, + { + "epoch": 5.641659311562224, + "grad_norm": 1.2543667228578632, + "learning_rate": 4.1994643265585545e-06, + "loss": 0.04, + "step": 1598 + }, + { + "epoch": 5.645189761694616, + "grad_norm": 1.1537540952954353, + "learning_rate": 4.193816281970475e-06, + "loss": 0.0339, + "step": 1599 + }, + { + "epoch": 5.648720211827008, + "grad_norm": 1.197121170775597, + "learning_rate": 4.1881692933461835e-06, + "loss": 0.0339, + "step": 1600 + }, + { + "epoch": 5.6522506619594, + "grad_norm": 1.4876713070885859, + "learning_rate": 4.182523368082267e-06, + "loss": 0.0402, + "step": 1601 + }, + { + "epoch": 5.655781112091792, + "grad_norm": 1.0718461472117065, + "learning_rate": 4.1768785135739335e-06, + "loss": 0.0301, + "step": 1602 + }, + { + "epoch": 5.659311562224183, + "grad_norm": 1.0869454673908943, + "learning_rate": 4.171234737214983e-06, + "loss": 0.03, + "step": 1603 + }, + { + "epoch": 5.662842012356576, + "grad_norm": 1.4519457032264655, + "learning_rate": 4.165592046397806e-06, + "loss": 0.04, + "step": 1604 + }, + { + "epoch": 5.666372462488967, + "grad_norm": 1.0193310940338902, + "learning_rate": 4.159950448513367e-06, + "loss": 0.029, + "step": 1605 + }, + { + "epoch": 5.669902912621359, + "grad_norm": 1.0502500926022942, + "learning_rate": 4.154309950951203e-06, + "loss": 0.0319, + "step": 1606 + }, + { + "epoch": 5.673433362753751, + "grad_norm": 0.7470487059238342, + "learning_rate": 4.148670561099406e-06, + "loss": 0.0225, + "step": 1607 + }, + { + "epoch": 5.676963812886143, + "grad_norm": 1.0022247053025022, + "learning_rate": 4.1430322863446225e-06, + "loss": 0.031, + "step": 1608 + }, + { + "epoch": 5.680494263018534, + "grad_norm": 1.1234488887050893, + "learning_rate": 4.137395134072032e-06, + "loss": 0.0419, + "step": 1609 + }, + { + "epoch": 5.684024713150927, + "grad_norm": 1.2288044982633797, + "learning_rate": 4.131759111665349e-06, + "loss": 0.0331, + "step": 1610 + }, + { + "epoch": 5.687555163283319, + "grad_norm": 1.3552076892820886, + "learning_rate": 4.126124226506803e-06, + "loss": 0.0369, + "step": 1611 + }, + { + "epoch": 5.69108561341571, + "grad_norm": 1.3378524414412645, + "learning_rate": 4.120490485977141e-06, + "loss": 0.0456, + "step": 1612 + }, + { + "epoch": 5.694616063548103, + "grad_norm": 0.9875293322389016, + "learning_rate": 4.114857897455602e-06, + "loss": 0.0314, + "step": 1613 + }, + { + "epoch": 5.698146513680494, + "grad_norm": 1.2336698386052185, + "learning_rate": 4.1092264683199255e-06, + "loss": 0.0278, + "step": 1614 + }, + { + "epoch": 5.701676963812886, + "grad_norm": 1.1015989513714806, + "learning_rate": 4.103596205946323e-06, + "loss": 0.036, + "step": 1615 + }, + { + "epoch": 5.705207413945278, + "grad_norm": 1.127307319170273, + "learning_rate": 4.097967117709483e-06, + "loss": 0.0273, + "step": 1616 + }, + { + "epoch": 5.70873786407767, + "grad_norm": 1.2008259758190027, + "learning_rate": 4.092339210982558e-06, + "loss": 0.0388, + "step": 1617 + }, + { + "epoch": 5.712268314210061, + "grad_norm": 1.0615384931131095, + "learning_rate": 4.086712493137146e-06, + "loss": 0.0305, + "step": 1618 + }, + { + "epoch": 5.715798764342454, + "grad_norm": 1.3345584292217139, + "learning_rate": 4.081086971543296e-06, + "loss": 0.0353, + "step": 1619 + }, + { + "epoch": 5.719329214474845, + "grad_norm": 1.2641272529631358, + "learning_rate": 4.075462653569484e-06, + "loss": 0.0449, + "step": 1620 + }, + { + "epoch": 5.722859664607237, + "grad_norm": 1.064733968468094, + "learning_rate": 4.069839546582611e-06, + "loss": 0.0305, + "step": 1621 + }, + { + "epoch": 5.72639011473963, + "grad_norm": 1.6370554966016035, + "learning_rate": 4.064217657947995e-06, + "loss": 0.037, + "step": 1622 + }, + { + "epoch": 5.729920564872021, + "grad_norm": 1.3541877940203568, + "learning_rate": 4.058596995029353e-06, + "loss": 0.0313, + "step": 1623 + }, + { + "epoch": 5.733451015004413, + "grad_norm": 1.5466260987879117, + "learning_rate": 4.0529775651888e-06, + "loss": 0.0388, + "step": 1624 + }, + { + "epoch": 5.736981465136805, + "grad_norm": 1.3844077192235618, + "learning_rate": 4.0473593757868344e-06, + "loss": 0.0388, + "step": 1625 + }, + { + "epoch": 5.740511915269197, + "grad_norm": 0.8866810369065234, + "learning_rate": 4.041742434182332e-06, + "loss": 0.0275, + "step": 1626 + }, + { + "epoch": 5.744042365401588, + "grad_norm": 1.2621787066958736, + "learning_rate": 4.036126747732529e-06, + "loss": 0.036, + "step": 1627 + }, + { + "epoch": 5.747572815533981, + "grad_norm": 1.0474742711936464, + "learning_rate": 4.030512323793026e-06, + "loss": 0.0294, + "step": 1628 + }, + { + "epoch": 5.751103265666372, + "grad_norm": 1.1184697802275934, + "learning_rate": 4.024899169717763e-06, + "loss": 0.0314, + "step": 1629 + }, + { + "epoch": 5.754633715798764, + "grad_norm": 1.0413870709876862, + "learning_rate": 4.019287292859016e-06, + "loss": 0.0313, + "step": 1630 + }, + { + "epoch": 5.758164165931156, + "grad_norm": 1.3799320335195928, + "learning_rate": 4.013676700567394e-06, + "loss": 0.0321, + "step": 1631 + }, + { + "epoch": 5.761694616063548, + "grad_norm": 1.4422001245231622, + "learning_rate": 4.008067400191819e-06, + "loss": 0.0326, + "step": 1632 + }, + { + "epoch": 5.76522506619594, + "grad_norm": 1.1886988415945137, + "learning_rate": 4.002459399079523e-06, + "loss": 0.0311, + "step": 1633 + }, + { + "epoch": 5.768755516328332, + "grad_norm": 1.097815662449813, + "learning_rate": 3.996852704576033e-06, + "loss": 0.0267, + "step": 1634 + }, + { + "epoch": 5.772285966460724, + "grad_norm": 1.2560055985189782, + "learning_rate": 3.991247324025171e-06, + "loss": 0.0361, + "step": 1635 + }, + { + "epoch": 5.775816416593115, + "grad_norm": 1.4054550223897035, + "learning_rate": 3.985643264769029e-06, + "loss": 0.0487, + "step": 1636 + }, + { + "epoch": 5.779346866725508, + "grad_norm": 1.1970955623002149, + "learning_rate": 3.980040534147976e-06, + "loss": 0.0307, + "step": 1637 + }, + { + "epoch": 5.782877316857899, + "grad_norm": 1.374467985258232, + "learning_rate": 3.9744391395006365e-06, + "loss": 0.0306, + "step": 1638 + }, + { + "epoch": 5.786407766990291, + "grad_norm": 1.2549470774404117, + "learning_rate": 3.968839088163888e-06, + "loss": 0.0376, + "step": 1639 + }, + { + "epoch": 5.789938217122684, + "grad_norm": 1.3063679286333543, + "learning_rate": 3.963240387472844e-06, + "loss": 0.0362, + "step": 1640 + }, + { + "epoch": 5.793468667255075, + "grad_norm": 1.094620445518725, + "learning_rate": 3.957643044760855e-06, + "loss": 0.033, + "step": 1641 + }, + { + "epoch": 5.796999117387466, + "grad_norm": 1.324271988433546, + "learning_rate": 3.952047067359488e-06, + "loss": 0.0411, + "step": 1642 + }, + { + "epoch": 5.800529567519859, + "grad_norm": 1.0691146110640006, + "learning_rate": 3.9464524625985245e-06, + "loss": 0.0299, + "step": 1643 + }, + { + "epoch": 5.804060017652251, + "grad_norm": 1.6238152298946706, + "learning_rate": 3.9408592378059435e-06, + "loss": 0.0514, + "step": 1644 + }, + { + "epoch": 5.807590467784642, + "grad_norm": 1.5003014221904643, + "learning_rate": 3.9352674003079225e-06, + "loss": 0.0389, + "step": 1645 + }, + { + "epoch": 5.811120917917035, + "grad_norm": 0.9568160793181512, + "learning_rate": 3.929676957428818e-06, + "loss": 0.0259, + "step": 1646 + }, + { + "epoch": 5.814651368049426, + "grad_norm": 1.2133994572570093, + "learning_rate": 3.924087916491162e-06, + "loss": 0.0346, + "step": 1647 + }, + { + "epoch": 5.818181818181818, + "grad_norm": 1.3800721666905214, + "learning_rate": 3.9185002848156474e-06, + "loss": 0.028, + "step": 1648 + }, + { + "epoch": 5.82171226831421, + "grad_norm": 0.9712702885809616, + "learning_rate": 3.912914069721126e-06, + "loss": 0.0346, + "step": 1649 + }, + { + "epoch": 5.825242718446602, + "grad_norm": 1.1498616687315717, + "learning_rate": 3.907329278524587e-06, + "loss": 0.0416, + "step": 1650 + }, + { + "epoch": 5.828773168578994, + "grad_norm": 1.238604660547264, + "learning_rate": 3.901745918541162e-06, + "loss": 0.0321, + "step": 1651 + }, + { + "epoch": 5.832303618711386, + "grad_norm": 1.0109695573959605, + "learning_rate": 3.8961639970841026e-06, + "loss": 0.0331, + "step": 1652 + }, + { + "epoch": 5.835834068843777, + "grad_norm": 1.036661604672026, + "learning_rate": 3.89058352146478e-06, + "loss": 0.0298, + "step": 1653 + }, + { + "epoch": 5.839364518976169, + "grad_norm": 0.9903235467069647, + "learning_rate": 3.885004498992667e-06, + "loss": 0.0278, + "step": 1654 + }, + { + "epoch": 5.842894969108562, + "grad_norm": 1.1134824628538762, + "learning_rate": 3.879426936975338e-06, + "loss": 0.0267, + "step": 1655 + }, + { + "epoch": 5.846425419240953, + "grad_norm": 0.9773717826570428, + "learning_rate": 3.8738508427184504e-06, + "loss": 0.0318, + "step": 1656 + }, + { + "epoch": 5.849955869373345, + "grad_norm": 1.1990515209367925, + "learning_rate": 3.868276223525745e-06, + "loss": 0.0309, + "step": 1657 + }, + { + "epoch": 5.853486319505737, + "grad_norm": 1.2004136667479008, + "learning_rate": 3.86270308669902e-06, + "loss": 0.0313, + "step": 1658 + }, + { + "epoch": 5.857016769638129, + "grad_norm": 1.3275294136418903, + "learning_rate": 3.857131439538143e-06, + "loss": 0.0287, + "step": 1659 + }, + { + "epoch": 5.86054721977052, + "grad_norm": 1.3029763877555014, + "learning_rate": 3.851561289341023e-06, + "loss": 0.0389, + "step": 1660 + }, + { + "epoch": 5.864077669902913, + "grad_norm": 1.4685841971344737, + "learning_rate": 3.8459926434036135e-06, + "loss": 0.0391, + "step": 1661 + }, + { + "epoch": 5.867608120035305, + "grad_norm": 1.0732193682604763, + "learning_rate": 3.840425509019892e-06, + "loss": 0.0266, + "step": 1662 + }, + { + "epoch": 5.871138570167696, + "grad_norm": 1.2660485416506868, + "learning_rate": 3.834859893481864e-06, + "loss": 0.0423, + "step": 1663 + }, + { + "epoch": 5.874669020300089, + "grad_norm": 1.1287653825694117, + "learning_rate": 3.829295804079536e-06, + "loss": 0.0442, + "step": 1664 + }, + { + "epoch": 5.87819947043248, + "grad_norm": 1.1396467243533388, + "learning_rate": 3.823733248100924e-06, + "loss": 0.0375, + "step": 1665 + }, + { + "epoch": 5.881729920564872, + "grad_norm": 1.1843243501076075, + "learning_rate": 3.818172232832028e-06, + "loss": 0.0385, + "step": 1666 + }, + { + "epoch": 5.885260370697264, + "grad_norm": 1.0277571870629416, + "learning_rate": 3.812612765556838e-06, + "loss": 0.0257, + "step": 1667 + }, + { + "epoch": 5.888790820829656, + "grad_norm": 1.1554476229893942, + "learning_rate": 3.8070548535573083e-06, + "loss": 0.0308, + "step": 1668 + }, + { + "epoch": 5.892321270962047, + "grad_norm": 1.1523365738890718, + "learning_rate": 3.8014985041133627e-06, + "loss": 0.0358, + "step": 1669 + }, + { + "epoch": 5.89585172109444, + "grad_norm": 1.0072551262487832, + "learning_rate": 3.795943724502874e-06, + "loss": 0.0306, + "step": 1670 + }, + { + "epoch": 5.899382171226831, + "grad_norm": 1.0471753503658485, + "learning_rate": 3.790390522001662e-06, + "loss": 0.0332, + "step": 1671 + }, + { + "epoch": 5.902912621359223, + "grad_norm": 1.1227448997020295, + "learning_rate": 3.784838903883478e-06, + "loss": 0.0273, + "step": 1672 + }, + { + "epoch": 5.906443071491616, + "grad_norm": 1.0634270827840409, + "learning_rate": 3.7792888774199976e-06, + "loss": 0.0331, + "step": 1673 + }, + { + "epoch": 5.909973521624007, + "grad_norm": 1.507070277423157, + "learning_rate": 3.773740449880815e-06, + "loss": 0.0435, + "step": 1674 + }, + { + "epoch": 5.913503971756399, + "grad_norm": 1.0627803694881015, + "learning_rate": 3.768193628533427e-06, + "loss": 0.0301, + "step": 1675 + }, + { + "epoch": 5.917034421888791, + "grad_norm": 1.2773031091505942, + "learning_rate": 3.762648420643229e-06, + "loss": 0.0369, + "step": 1676 + }, + { + "epoch": 5.920564872021183, + "grad_norm": 1.051047355239252, + "learning_rate": 3.7571048334735004e-06, + "loss": 0.0259, + "step": 1677 + }, + { + "epoch": 5.924095322153574, + "grad_norm": 1.1959016952670376, + "learning_rate": 3.7515628742854006e-06, + "loss": 0.0302, + "step": 1678 + }, + { + "epoch": 5.927625772285967, + "grad_norm": 1.1190088459841994, + "learning_rate": 3.746022550337953e-06, + "loss": 0.0298, + "step": 1679 + }, + { + "epoch": 5.931156222418358, + "grad_norm": 1.278097130133865, + "learning_rate": 3.7404838688880445e-06, + "loss": 0.0347, + "step": 1680 + }, + { + "epoch": 5.93468667255075, + "grad_norm": 1.926981437205621, + "learning_rate": 3.7349468371904054e-06, + "loss": 0.0385, + "step": 1681 + }, + { + "epoch": 5.938217122683142, + "grad_norm": 1.0453195552294023, + "learning_rate": 3.7294114624976053e-06, + "loss": 0.0291, + "step": 1682 + }, + { + "epoch": 5.941747572815534, + "grad_norm": 0.8142508085949226, + "learning_rate": 3.7238777520600485e-06, + "loss": 0.0189, + "step": 1683 + }, + { + "epoch": 5.945278022947926, + "grad_norm": 1.0955539492932918, + "learning_rate": 3.718345713125953e-06, + "loss": 0.031, + "step": 1684 + }, + { + "epoch": 5.948808473080318, + "grad_norm": 1.1573023005071321, + "learning_rate": 3.7128153529413535e-06, + "loss": 0.0291, + "step": 1685 + }, + { + "epoch": 5.95233892321271, + "grad_norm": 1.1034526336247084, + "learning_rate": 3.707286678750082e-06, + "loss": 0.0361, + "step": 1686 + }, + { + "epoch": 5.955869373345101, + "grad_norm": 1.2699877135354682, + "learning_rate": 3.701759697793761e-06, + "loss": 0.0365, + "step": 1687 + }, + { + "epoch": 5.959399823477494, + "grad_norm": 0.8003447962213704, + "learning_rate": 3.696234417311798e-06, + "loss": 0.0215, + "step": 1688 + }, + { + "epoch": 5.962930273609885, + "grad_norm": 1.3797712899993184, + "learning_rate": 3.690710844541372e-06, + "loss": 0.0374, + "step": 1689 + }, + { + "epoch": 5.966460723742277, + "grad_norm": 1.3359061813036293, + "learning_rate": 3.685188986717427e-06, + "loss": 0.0327, + "step": 1690 + }, + { + "epoch": 5.969991173874669, + "grad_norm": 1.137062617546707, + "learning_rate": 3.6796688510726575e-06, + "loss": 0.0334, + "step": 1691 + }, + { + "epoch": 5.973521624007061, + "grad_norm": 1.2242712032971437, + "learning_rate": 3.674150444837506e-06, + "loss": 0.0338, + "step": 1692 + }, + { + "epoch": 5.977052074139452, + "grad_norm": 1.554983019186223, + "learning_rate": 3.6686337752401446e-06, + "loss": 0.0351, + "step": 1693 + }, + { + "epoch": 5.980582524271845, + "grad_norm": 1.310508704502005, + "learning_rate": 3.6631188495064764e-06, + "loss": 0.0335, + "step": 1694 + }, + { + "epoch": 5.984112974404237, + "grad_norm": 1.5775757496985878, + "learning_rate": 3.6576056748601153e-06, + "loss": 0.0437, + "step": 1695 + }, + { + "epoch": 5.987643424536628, + "grad_norm": 1.5450846303171186, + "learning_rate": 3.652094258522387e-06, + "loss": 0.0412, + "step": 1696 + }, + { + "epoch": 5.991173874669021, + "grad_norm": 1.0538519504564978, + "learning_rate": 3.646584607712308e-06, + "loss": 0.0311, + "step": 1697 + }, + { + "epoch": 5.994704324801412, + "grad_norm": 1.008771230747239, + "learning_rate": 3.6410767296465876e-06, + "loss": 0.0216, + "step": 1698 + }, + { + "epoch": 5.998234774933804, + "grad_norm": 0.9724742616144021, + "learning_rate": 3.6355706315396094e-06, + "loss": 0.035, + "step": 1699 + }, + { + "epoch": 6.001765225066196, + "grad_norm": 0.9059733968308091, + "learning_rate": 3.6300663206034305e-06, + "loss": 0.023, + "step": 1700 + }, + { + "epoch": 6.005295675198588, + "grad_norm": 0.5572513282946998, + "learning_rate": 3.624563804047758e-06, + "loss": 0.019, + "step": 1701 + }, + { + "epoch": 6.008826125330979, + "grad_norm": 0.38137629824196734, + "learning_rate": 3.619063089079957e-06, + "loss": 0.0102, + "step": 1702 + }, + { + "epoch": 6.012356575463372, + "grad_norm": 0.8257608389626776, + "learning_rate": 3.61356418290503e-06, + "loss": 0.0143, + "step": 1703 + }, + { + "epoch": 6.015887025595763, + "grad_norm": 0.7826935351798261, + "learning_rate": 3.608067092725611e-06, + "loss": 0.0147, + "step": 1704 + }, + { + "epoch": 6.019417475728155, + "grad_norm": 0.53989341169862, + "learning_rate": 3.6025718257419532e-06, + "loss": 0.014, + "step": 1705 + }, + { + "epoch": 6.022947925860548, + "grad_norm": 0.4050317440329078, + "learning_rate": 3.5970783891519256e-06, + "loss": 0.0095, + "step": 1706 + }, + { + "epoch": 6.026478375992939, + "grad_norm": 0.6711962858373305, + "learning_rate": 3.591586790150995e-06, + "loss": 0.0147, + "step": 1707 + }, + { + "epoch": 6.030008826125331, + "grad_norm": 0.7051314515732616, + "learning_rate": 3.5860970359322256e-06, + "loss": 0.0209, + "step": 1708 + }, + { + "epoch": 6.033539276257723, + "grad_norm": 0.5497847280536009, + "learning_rate": 3.5806091336862615e-06, + "loss": 0.0139, + "step": 1709 + }, + { + "epoch": 6.037069726390115, + "grad_norm": 0.8770341544382031, + "learning_rate": 3.575123090601324e-06, + "loss": 0.0148, + "step": 1710 + }, + { + "epoch": 6.040600176522506, + "grad_norm": 0.7112573685198552, + "learning_rate": 3.569638913863197e-06, + "loss": 0.0129, + "step": 1711 + }, + { + "epoch": 6.044130626654899, + "grad_norm": 0.6820526086570896, + "learning_rate": 3.564156610655223e-06, + "loss": 0.0143, + "step": 1712 + }, + { + "epoch": 6.04766107678729, + "grad_norm": 0.5320535692203213, + "learning_rate": 3.5586761881582854e-06, + "loss": 0.0126, + "step": 1713 + }, + { + "epoch": 6.051191526919682, + "grad_norm": 1.2453100860743471, + "learning_rate": 3.5531976535508107e-06, + "loss": 0.0256, + "step": 1714 + }, + { + "epoch": 6.054721977052074, + "grad_norm": 0.5401780711388949, + "learning_rate": 3.5477210140087458e-06, + "loss": 0.0089, + "step": 1715 + }, + { + "epoch": 6.058252427184466, + "grad_norm": 0.7966005057927301, + "learning_rate": 3.542246276705559e-06, + "loss": 0.0108, + "step": 1716 + }, + { + "epoch": 6.061782877316858, + "grad_norm": 0.707779856120114, + "learning_rate": 3.5367734488122275e-06, + "loss": 0.0139, + "step": 1717 + }, + { + "epoch": 6.06531332744925, + "grad_norm": 0.6338051837957891, + "learning_rate": 3.5313025374972255e-06, + "loss": 0.0109, + "step": 1718 + }, + { + "epoch": 6.068843777581642, + "grad_norm": 1.0852142010955435, + "learning_rate": 3.5258335499265196e-06, + "loss": 0.0158, + "step": 1719 + }, + { + "epoch": 6.072374227714033, + "grad_norm": 0.9797530018189778, + "learning_rate": 3.520366493263554e-06, + "loss": 0.0142, + "step": 1720 + }, + { + "epoch": 6.075904677846426, + "grad_norm": 0.8075896779981921, + "learning_rate": 3.514901374669244e-06, + "loss": 0.0191, + "step": 1721 + }, + { + "epoch": 6.079435127978817, + "grad_norm": 0.8206530660890795, + "learning_rate": 3.509438201301968e-06, + "loss": 0.0213, + "step": 1722 + }, + { + "epoch": 6.082965578111209, + "grad_norm": 1.0021284693893828, + "learning_rate": 3.5039769803175545e-06, + "loss": 0.0121, + "step": 1723 + }, + { + "epoch": 6.086496028243601, + "grad_norm": 1.0333805013099364, + "learning_rate": 3.498517718869277e-06, + "loss": 0.0173, + "step": 1724 + }, + { + "epoch": 6.090026478375993, + "grad_norm": 0.8521140675368526, + "learning_rate": 3.4930604241078392e-06, + "loss": 0.0124, + "step": 1725 + }, + { + "epoch": 6.093556928508384, + "grad_norm": 1.2483815916293344, + "learning_rate": 3.487605103181373e-06, + "loss": 0.0177, + "step": 1726 + }, + { + "epoch": 6.097087378640777, + "grad_norm": 0.9360861416940088, + "learning_rate": 3.48215176323542e-06, + "loss": 0.0074, + "step": 1727 + }, + { + "epoch": 6.100617828773169, + "grad_norm": 0.7060808263911972, + "learning_rate": 3.4767004114129332e-06, + "loss": 0.0134, + "step": 1728 + }, + { + "epoch": 6.10414827890556, + "grad_norm": 1.0761544174873692, + "learning_rate": 3.4712510548542546e-06, + "loss": 0.0104, + "step": 1729 + }, + { + "epoch": 6.107678729037953, + "grad_norm": 1.1472034051156381, + "learning_rate": 3.4658037006971147e-06, + "loss": 0.0136, + "step": 1730 + }, + { + "epoch": 6.111209179170344, + "grad_norm": 0.7632094720369841, + "learning_rate": 3.4603583560766265e-06, + "loss": 0.0102, + "step": 1731 + }, + { + "epoch": 6.114739629302736, + "grad_norm": 0.6064483592400337, + "learning_rate": 3.4549150281252635e-06, + "loss": 0.0117, + "step": 1732 + }, + { + "epoch": 6.118270079435128, + "grad_norm": 1.272210037416928, + "learning_rate": 3.4494737239728638e-06, + "loss": 0.0137, + "step": 1733 + }, + { + "epoch": 6.12180052956752, + "grad_norm": 0.8375016357416992, + "learning_rate": 3.4440344507466094e-06, + "loss": 0.0146, + "step": 1734 + }, + { + "epoch": 6.125330979699911, + "grad_norm": 0.9642198367217255, + "learning_rate": 3.4385972155710274e-06, + "loss": 0.0171, + "step": 1735 + }, + { + "epoch": 6.128861429832304, + "grad_norm": 0.955436426733389, + "learning_rate": 3.4331620255679697e-06, + "loss": 0.0154, + "step": 1736 + }, + { + "epoch": 6.132391879964696, + "grad_norm": 0.902470959697602, + "learning_rate": 3.427728887856615e-06, + "loss": 0.0116, + "step": 1737 + }, + { + "epoch": 6.135922330097087, + "grad_norm": 1.0263405213556056, + "learning_rate": 3.4222978095534495e-06, + "loss": 0.0236, + "step": 1738 + }, + { + "epoch": 6.13945278022948, + "grad_norm": 1.1486275009132847, + "learning_rate": 3.416868797772264e-06, + "loss": 0.0209, + "step": 1739 + }, + { + "epoch": 6.142983230361871, + "grad_norm": 0.6304921817804658, + "learning_rate": 3.4114418596241417e-06, + "loss": 0.0127, + "step": 1740 + }, + { + "epoch": 6.146513680494263, + "grad_norm": 0.7810344893496806, + "learning_rate": 3.406017002217452e-06, + "loss": 0.0173, + "step": 1741 + }, + { + "epoch": 6.150044130626655, + "grad_norm": 0.8208790759245346, + "learning_rate": 3.4005942326578362e-06, + "loss": 0.0175, + "step": 1742 + }, + { + "epoch": 6.153574580759047, + "grad_norm": 0.7542642677857302, + "learning_rate": 3.3951735580481997e-06, + "loss": 0.0086, + "step": 1743 + }, + { + "epoch": 6.157105030891438, + "grad_norm": 0.8172427239282024, + "learning_rate": 3.3897549854887065e-06, + "loss": 0.013, + "step": 1744 + }, + { + "epoch": 6.160635481023831, + "grad_norm": 0.8454160227750958, + "learning_rate": 3.3843385220767676e-06, + "loss": 0.0103, + "step": 1745 + }, + { + "epoch": 6.164165931156222, + "grad_norm": 0.9255883108475365, + "learning_rate": 3.3789241749070293e-06, + "loss": 0.0132, + "step": 1746 + }, + { + "epoch": 6.167696381288614, + "grad_norm": 0.9976100384192047, + "learning_rate": 3.3735119510713687e-06, + "loss": 0.0174, + "step": 1747 + }, + { + "epoch": 6.171226831421007, + "grad_norm": 0.9387965749473235, + "learning_rate": 3.368101857658879e-06, + "loss": 0.0169, + "step": 1748 + }, + { + "epoch": 6.174757281553398, + "grad_norm": 0.5869647008785216, + "learning_rate": 3.3626939017558636e-06, + "loss": 0.0127, + "step": 1749 + }, + { + "epoch": 6.17828773168579, + "grad_norm": 0.6745171861981384, + "learning_rate": 3.3572880904458267e-06, + "loss": 0.0103, + "step": 1750 + }, + { + "epoch": 6.181818181818182, + "grad_norm": 0.8550220206304635, + "learning_rate": 3.3518844308094646e-06, + "loss": 0.0102, + "step": 1751 + }, + { + "epoch": 6.185348631950574, + "grad_norm": 0.864056295698312, + "learning_rate": 3.3464829299246503e-06, + "loss": 0.0143, + "step": 1752 + }, + { + "epoch": 6.188879082082965, + "grad_norm": 0.678998938780041, + "learning_rate": 3.3410835948664377e-06, + "loss": 0.0131, + "step": 1753 + }, + { + "epoch": 6.192409532215358, + "grad_norm": 0.625889566640382, + "learning_rate": 3.335686432707035e-06, + "loss": 0.0091, + "step": 1754 + }, + { + "epoch": 6.195939982347749, + "grad_norm": 0.5116822458748091, + "learning_rate": 3.3302914505158107e-06, + "loss": 0.0091, + "step": 1755 + }, + { + "epoch": 6.199470432480141, + "grad_norm": 0.535615566109914, + "learning_rate": 3.3248986553592733e-06, + "loss": 0.0075, + "step": 1756 + }, + { + "epoch": 6.203000882612533, + "grad_norm": 0.8457251925825474, + "learning_rate": 3.3195080543010714e-06, + "loss": 0.0171, + "step": 1757 + }, + { + "epoch": 6.206531332744925, + "grad_norm": 0.8773537874248233, + "learning_rate": 3.3141196544019757e-06, + "loss": 0.0139, + "step": 1758 + }, + { + "epoch": 6.210061782877317, + "grad_norm": 0.4394787304626812, + "learning_rate": 3.308733462719873e-06, + "loss": 0.012, + "step": 1759 + }, + { + "epoch": 6.213592233009709, + "grad_norm": 0.7207459401149353, + "learning_rate": 3.3033494863097633e-06, + "loss": 0.0108, + "step": 1760 + }, + { + "epoch": 6.217122683142101, + "grad_norm": 0.8430470099190911, + "learning_rate": 3.29796773222374e-06, + "loss": 0.0188, + "step": 1761 + }, + { + "epoch": 6.220653133274492, + "grad_norm": 0.7135350223099851, + "learning_rate": 3.2925882075109855e-06, + "loss": 0.0154, + "step": 1762 + }, + { + "epoch": 6.224183583406885, + "grad_norm": 0.7612507576414165, + "learning_rate": 3.287210919217767e-06, + "loss": 0.0144, + "step": 1763 + }, + { + "epoch": 6.227714033539276, + "grad_norm": 0.5999384462823261, + "learning_rate": 3.2818358743874163e-06, + "loss": 0.0108, + "step": 1764 + }, + { + "epoch": 6.231244483671668, + "grad_norm": 0.9581020361071976, + "learning_rate": 3.2764630800603314e-06, + "loss": 0.0185, + "step": 1765 + }, + { + "epoch": 6.23477493380406, + "grad_norm": 1.2002976181598761, + "learning_rate": 3.2710925432739593e-06, + "loss": 0.0167, + "step": 1766 + }, + { + "epoch": 6.238305383936452, + "grad_norm": 0.7930404116117287, + "learning_rate": 3.265724271062793e-06, + "loss": 0.0081, + "step": 1767 + }, + { + "epoch": 6.241835834068844, + "grad_norm": 0.7614401436743771, + "learning_rate": 3.2603582704583547e-06, + "loss": 0.013, + "step": 1768 + }, + { + "epoch": 6.245366284201236, + "grad_norm": 0.9465373832461532, + "learning_rate": 3.2549945484891974e-06, + "loss": 0.0089, + "step": 1769 + }, + { + "epoch": 6.248896734333628, + "grad_norm": 0.7159172955428454, + "learning_rate": 3.249633112180884e-06, + "loss": 0.0128, + "step": 1770 + }, + { + "epoch": 6.252427184466019, + "grad_norm": 0.8768339943616019, + "learning_rate": 3.2442739685559878e-06, + "loss": 0.0139, + "step": 1771 + }, + { + "epoch": 6.255957634598412, + "grad_norm": 1.207797522155778, + "learning_rate": 3.238917124634074e-06, + "loss": 0.0211, + "step": 1772 + }, + { + "epoch": 6.259488084730803, + "grad_norm": 1.1776230275164792, + "learning_rate": 3.233562587431699e-06, + "loss": 0.0135, + "step": 1773 + }, + { + "epoch": 6.263018534863195, + "grad_norm": 0.5122161658230125, + "learning_rate": 3.2282103639623995e-06, + "loss": 0.008, + "step": 1774 + }, + { + "epoch": 6.266548984995587, + "grad_norm": 1.4907017011921408, + "learning_rate": 3.2228604612366777e-06, + "loss": 0.0269, + "step": 1775 + }, + { + "epoch": 6.270079435127979, + "grad_norm": 0.8843650375376056, + "learning_rate": 3.2175128862619988e-06, + "loss": 0.0173, + "step": 1776 + }, + { + "epoch": 6.27360988526037, + "grad_norm": 0.7162318124292995, + "learning_rate": 3.2121676460427765e-06, + "loss": 0.0186, + "step": 1777 + }, + { + "epoch": 6.277140335392763, + "grad_norm": 1.0113079378269854, + "learning_rate": 3.2068247475803705e-06, + "loss": 0.0199, + "step": 1778 + }, + { + "epoch": 6.280670785525155, + "grad_norm": 1.2014498113702745, + "learning_rate": 3.2014841978730677e-06, + "loss": 0.0143, + "step": 1779 + }, + { + "epoch": 6.284201235657546, + "grad_norm": 1.0559901724040555, + "learning_rate": 3.1961460039160842e-06, + "loss": 0.0166, + "step": 1780 + }, + { + "epoch": 6.287731685789939, + "grad_norm": 1.0683626849122516, + "learning_rate": 3.190810172701544e-06, + "loss": 0.0206, + "step": 1781 + }, + { + "epoch": 6.29126213592233, + "grad_norm": 0.8452582463543812, + "learning_rate": 3.1854767112184843e-06, + "loss": 0.0119, + "step": 1782 + }, + { + "epoch": 6.294792586054722, + "grad_norm": 0.5381151395894828, + "learning_rate": 3.1801456264528314e-06, + "loss": 0.0107, + "step": 1783 + }, + { + "epoch": 6.298323036187114, + "grad_norm": 0.8504416097040094, + "learning_rate": 3.1748169253874006e-06, + "loss": 0.0137, + "step": 1784 + }, + { + "epoch": 6.301853486319506, + "grad_norm": 1.0335107877381846, + "learning_rate": 3.1694906150018887e-06, + "loss": 0.0123, + "step": 1785 + }, + { + "epoch": 6.305383936451897, + "grad_norm": 1.3020784761987876, + "learning_rate": 3.164166702272855e-06, + "loss": 0.0161, + "step": 1786 + }, + { + "epoch": 6.30891438658429, + "grad_norm": 0.9175792105209791, + "learning_rate": 3.158845194173721e-06, + "loss": 0.0122, + "step": 1787 + }, + { + "epoch": 6.312444836716681, + "grad_norm": 0.5147257686856647, + "learning_rate": 3.1535260976747605e-06, + "loss": 0.0087, + "step": 1788 + }, + { + "epoch": 6.315975286849073, + "grad_norm": 0.7164679178753424, + "learning_rate": 3.1482094197430844e-06, + "loss": 0.0111, + "step": 1789 + }, + { + "epoch": 6.319505736981466, + "grad_norm": 0.5564318663669371, + "learning_rate": 3.1428951673426413e-06, + "loss": 0.0111, + "step": 1790 + }, + { + "epoch": 6.323036187113857, + "grad_norm": 0.965226763168101, + "learning_rate": 3.1375833474341964e-06, + "loss": 0.0206, + "step": 1791 + }, + { + "epoch": 6.326566637246249, + "grad_norm": 0.6621990855082875, + "learning_rate": 3.1322739669753346e-06, + "loss": 0.0106, + "step": 1792 + }, + { + "epoch": 6.330097087378641, + "grad_norm": 1.023712916894823, + "learning_rate": 3.12696703292044e-06, + "loss": 0.0157, + "step": 1793 + }, + { + "epoch": 6.333627537511033, + "grad_norm": 1.3193096444726848, + "learning_rate": 3.1216625522206976e-06, + "loss": 0.0188, + "step": 1794 + }, + { + "epoch": 6.337157987643424, + "grad_norm": 0.7230176002582147, + "learning_rate": 3.116360531824074e-06, + "loss": 0.0123, + "step": 1795 + }, + { + "epoch": 6.340688437775817, + "grad_norm": 0.4968510070957696, + "learning_rate": 3.111060978675319e-06, + "loss": 0.01, + "step": 1796 + }, + { + "epoch": 6.344218887908208, + "grad_norm": 0.9168620334994467, + "learning_rate": 3.1057638997159436e-06, + "loss": 0.0123, + "step": 1797 + }, + { + "epoch": 6.3477493380406, + "grad_norm": 1.045422154798156, + "learning_rate": 3.1004693018842246e-06, + "loss": 0.0176, + "step": 1798 + }, + { + "epoch": 6.351279788172992, + "grad_norm": 0.6898387904541954, + "learning_rate": 3.095177192115186e-06, + "loss": 0.0108, + "step": 1799 + }, + { + "epoch": 6.354810238305384, + "grad_norm": 0.9475975339179177, + "learning_rate": 3.089887577340592e-06, + "loss": 0.0223, + "step": 1800 + }, + { + "epoch": 6.358340688437776, + "grad_norm": 0.7921639177654283, + "learning_rate": 3.0846004644889384e-06, + "loss": 0.014, + "step": 1801 + }, + { + "epoch": 6.361871138570168, + "grad_norm": 0.4986357808151056, + "learning_rate": 3.079315860485447e-06, + "loss": 0.0089, + "step": 1802 + }, + { + "epoch": 6.36540158870256, + "grad_norm": 0.7912713947374417, + "learning_rate": 3.074033772252052e-06, + "loss": 0.0138, + "step": 1803 + }, + { + "epoch": 6.368932038834951, + "grad_norm": 0.6116714173538124, + "learning_rate": 3.068754206707392e-06, + "loss": 0.0107, + "step": 1804 + }, + { + "epoch": 6.372462488967344, + "grad_norm": 0.9156310485947148, + "learning_rate": 3.0634771707668e-06, + "loss": 0.0169, + "step": 1805 + }, + { + "epoch": 6.375992939099735, + "grad_norm": 1.167596910522646, + "learning_rate": 3.0582026713422986e-06, + "loss": 0.02, + "step": 1806 + }, + { + "epoch": 6.379523389232127, + "grad_norm": 0.9901915927758171, + "learning_rate": 3.0529307153425856e-06, + "loss": 0.0165, + "step": 1807 + }, + { + "epoch": 6.383053839364519, + "grad_norm": 1.481008174870301, + "learning_rate": 3.0476613096730297e-06, + "loss": 0.0222, + "step": 1808 + }, + { + "epoch": 6.386584289496911, + "grad_norm": 0.5923893303530244, + "learning_rate": 3.0423944612356547e-06, + "loss": 0.0164, + "step": 1809 + }, + { + "epoch": 6.390114739629302, + "grad_norm": 1.0883877118497178, + "learning_rate": 3.0371301769291417e-06, + "loss": 0.0244, + "step": 1810 + }, + { + "epoch": 6.393645189761695, + "grad_norm": 0.6617820948850102, + "learning_rate": 3.031868463648805e-06, + "loss": 0.0132, + "step": 1811 + }, + { + "epoch": 6.397175639894087, + "grad_norm": 1.1319312444676841, + "learning_rate": 3.0266093282865994e-06, + "loss": 0.0128, + "step": 1812 + }, + { + "epoch": 6.400706090026478, + "grad_norm": 1.0364535054263901, + "learning_rate": 3.021352777731096e-06, + "loss": 0.0135, + "step": 1813 + }, + { + "epoch": 6.404236540158871, + "grad_norm": 0.8231566571712303, + "learning_rate": 3.016098818867488e-06, + "loss": 0.0142, + "step": 1814 + }, + { + "epoch": 6.407766990291262, + "grad_norm": 0.9400515599635602, + "learning_rate": 3.0108474585775643e-06, + "loss": 0.0155, + "step": 1815 + }, + { + "epoch": 6.411297440423654, + "grad_norm": 1.142321312401896, + "learning_rate": 3.0055987037397165e-06, + "loss": 0.015, + "step": 1816 + }, + { + "epoch": 6.414827890556046, + "grad_norm": 0.8314342177085077, + "learning_rate": 3.000352561228924e-06, + "loss": 0.0145, + "step": 1817 + }, + { + "epoch": 6.418358340688438, + "grad_norm": 0.5257384424577614, + "learning_rate": 2.995109037916741e-06, + "loss": 0.0128, + "step": 1818 + }, + { + "epoch": 6.421888790820829, + "grad_norm": 0.9145824282698037, + "learning_rate": 2.9898681406712948e-06, + "loss": 0.0206, + "step": 1819 + }, + { + "epoch": 6.425419240953222, + "grad_norm": 0.8440962701698802, + "learning_rate": 2.984629876357268e-06, + "loss": 0.0162, + "step": 1820 + }, + { + "epoch": 6.428949691085613, + "grad_norm": 0.557696206536458, + "learning_rate": 2.9793942518359e-06, + "loss": 0.0121, + "step": 1821 + }, + { + "epoch": 6.432480141218005, + "grad_norm": 0.8839676879262386, + "learning_rate": 2.9741612739649694e-06, + "loss": 0.0187, + "step": 1822 + }, + { + "epoch": 6.4360105913503975, + "grad_norm": 0.6137056429912303, + "learning_rate": 2.9689309495987866e-06, + "loss": 0.0114, + "step": 1823 + }, + { + "epoch": 6.439541041482789, + "grad_norm": 0.590960637623899, + "learning_rate": 2.9637032855881907e-06, + "loss": 0.011, + "step": 1824 + }, + { + "epoch": 6.443071491615181, + "grad_norm": 0.7340878364330029, + "learning_rate": 2.958478288780533e-06, + "loss": 0.0156, + "step": 1825 + }, + { + "epoch": 6.446601941747573, + "grad_norm": 0.6068939473715877, + "learning_rate": 2.953255966019673e-06, + "loss": 0.0099, + "step": 1826 + }, + { + "epoch": 6.450132391879965, + "grad_norm": 0.956976085667651, + "learning_rate": 2.9480363241459654e-06, + "loss": 0.0153, + "step": 1827 + }, + { + "epoch": 6.453662842012356, + "grad_norm": 1.4101673999870787, + "learning_rate": 2.942819369996257e-06, + "loss": 0.0138, + "step": 1828 + }, + { + "epoch": 6.457193292144749, + "grad_norm": 0.6476705925290829, + "learning_rate": 2.9376051104038693e-06, + "loss": 0.0127, + "step": 1829 + }, + { + "epoch": 6.46072374227714, + "grad_norm": 0.9511940590697429, + "learning_rate": 2.932393552198597e-06, + "loss": 0.0152, + "step": 1830 + }, + { + "epoch": 6.464254192409532, + "grad_norm": 0.8895177754805732, + "learning_rate": 2.9271847022066992e-06, + "loss": 0.0091, + "step": 1831 + }, + { + "epoch": 6.467784642541924, + "grad_norm": 1.0547739786387238, + "learning_rate": 2.9219785672508814e-06, + "loss": 0.0215, + "step": 1832 + }, + { + "epoch": 6.471315092674316, + "grad_norm": 0.5846741192111404, + "learning_rate": 2.9167751541502997e-06, + "loss": 0.0088, + "step": 1833 + }, + { + "epoch": 6.474845542806708, + "grad_norm": 1.0256021473915227, + "learning_rate": 2.9115744697205394e-06, + "loss": 0.0151, + "step": 1834 + }, + { + "epoch": 6.4783759929391, + "grad_norm": 0.8233290956951373, + "learning_rate": 2.906376520773615e-06, + "loss": 0.019, + "step": 1835 + }, + { + "epoch": 6.481906443071492, + "grad_norm": 1.0153922841400518, + "learning_rate": 2.901181314117959e-06, + "loss": 0.0186, + "step": 1836 + }, + { + "epoch": 6.485436893203883, + "grad_norm": 0.6693530662923874, + "learning_rate": 2.8959888565584053e-06, + "loss": 0.0109, + "step": 1837 + }, + { + "epoch": 6.488967343336276, + "grad_norm": 0.8082313047750236, + "learning_rate": 2.8907991548961943e-06, + "loss": 0.0141, + "step": 1838 + }, + { + "epoch": 6.492497793468667, + "grad_norm": 1.0918853626421907, + "learning_rate": 2.8856122159289545e-06, + "loss": 0.0151, + "step": 1839 + }, + { + "epoch": 6.496028243601059, + "grad_norm": 1.0800355527343226, + "learning_rate": 2.880428046450697e-06, + "loss": 0.0174, + "step": 1840 + }, + { + "epoch": 6.499558693733451, + "grad_norm": 0.7973461757381258, + "learning_rate": 2.8752466532517985e-06, + "loss": 0.0148, + "step": 1841 + }, + { + "epoch": 6.503089143865843, + "grad_norm": 1.175354248378571, + "learning_rate": 2.87006804311901e-06, + "loss": 0.0173, + "step": 1842 + }, + { + "epoch": 6.506619593998234, + "grad_norm": 0.423213483711083, + "learning_rate": 2.864892222835426e-06, + "loss": 0.0093, + "step": 1843 + }, + { + "epoch": 6.510150044130627, + "grad_norm": 0.7988133105737514, + "learning_rate": 2.859719199180496e-06, + "loss": 0.0164, + "step": 1844 + }, + { + "epoch": 6.513680494263019, + "grad_norm": 0.71288566470618, + "learning_rate": 2.8545489789300006e-06, + "loss": 0.0083, + "step": 1845 + }, + { + "epoch": 6.51721094439541, + "grad_norm": 0.8606966993768842, + "learning_rate": 2.8493815688560507e-06, + "loss": 0.0156, + "step": 1846 + }, + { + "epoch": 6.5207413945278025, + "grad_norm": 1.023865389920783, + "learning_rate": 2.8442169757270766e-06, + "loss": 0.0173, + "step": 1847 + }, + { + "epoch": 6.524271844660194, + "grad_norm": 0.8342793515620819, + "learning_rate": 2.8390552063078202e-06, + "loss": 0.0138, + "step": 1848 + }, + { + "epoch": 6.527802294792586, + "grad_norm": 0.4755096914752449, + "learning_rate": 2.8338962673593196e-06, + "loss": 0.0062, + "step": 1849 + }, + { + "epoch": 6.531332744924978, + "grad_norm": 0.8143400988585866, + "learning_rate": 2.8287401656389098e-06, + "loss": 0.0194, + "step": 1850 + }, + { + "epoch": 6.53486319505737, + "grad_norm": 1.1739059664716804, + "learning_rate": 2.8235869079002092e-06, + "loss": 0.0191, + "step": 1851 + }, + { + "epoch": 6.538393645189762, + "grad_norm": 0.8089264957792369, + "learning_rate": 2.8184365008931126e-06, + "loss": 0.0114, + "step": 1852 + }, + { + "epoch": 6.541924095322154, + "grad_norm": 0.5052031411791448, + "learning_rate": 2.8132889513637745e-06, + "loss": 0.0073, + "step": 1853 + }, + { + "epoch": 6.545454545454545, + "grad_norm": 0.8255194878701497, + "learning_rate": 2.8081442660546126e-06, + "loss": 0.0146, + "step": 1854 + }, + { + "epoch": 6.548984995586937, + "grad_norm": 0.5988324299599385, + "learning_rate": 2.8030024517042907e-06, + "loss": 0.0058, + "step": 1855 + }, + { + "epoch": 6.5525154457193295, + "grad_norm": 1.1098726586457406, + "learning_rate": 2.797863515047716e-06, + "loss": 0.0232, + "step": 1856 + }, + { + "epoch": 6.556045895851721, + "grad_norm": 0.6982378631775757, + "learning_rate": 2.7927274628160195e-06, + "loss": 0.0128, + "step": 1857 + }, + { + "epoch": 6.559576345984113, + "grad_norm": 0.8093220058810642, + "learning_rate": 2.787594301736556e-06, + "loss": 0.0142, + "step": 1858 + }, + { + "epoch": 6.563106796116505, + "grad_norm": 0.7303991883818645, + "learning_rate": 2.782464038532897e-06, + "loss": 0.018, + "step": 1859 + }, + { + "epoch": 6.566637246248897, + "grad_norm": 0.7863951712032101, + "learning_rate": 2.777336679924816e-06, + "loss": 0.0125, + "step": 1860 + }, + { + "epoch": 6.570167696381288, + "grad_norm": 1.0360494744842748, + "learning_rate": 2.772212232628284e-06, + "loss": 0.0139, + "step": 1861 + }, + { + "epoch": 6.573698146513681, + "grad_norm": 0.6979794461268157, + "learning_rate": 2.767090703355454e-06, + "loss": 0.0112, + "step": 1862 + }, + { + "epoch": 6.577228596646073, + "grad_norm": 1.0339739858435775, + "learning_rate": 2.76197209881466e-06, + "loss": 0.0177, + "step": 1863 + }, + { + "epoch": 6.580759046778464, + "grad_norm": 0.7010699589655166, + "learning_rate": 2.756856425710408e-06, + "loss": 0.0106, + "step": 1864 + }, + { + "epoch": 6.5842894969108565, + "grad_norm": 1.0646218308258766, + "learning_rate": 2.7517436907433563e-06, + "loss": 0.0172, + "step": 1865 + }, + { + "epoch": 6.587819947043248, + "grad_norm": 0.5663703941696809, + "learning_rate": 2.7466339006103226e-06, + "loss": 0.009, + "step": 1866 + }, + { + "epoch": 6.59135039717564, + "grad_norm": 0.7034711845021318, + "learning_rate": 2.741527062004264e-06, + "loss": 0.0157, + "step": 1867 + }, + { + "epoch": 6.594880847308032, + "grad_norm": 1.254655973728525, + "learning_rate": 2.7364231816142733e-06, + "loss": 0.0151, + "step": 1868 + }, + { + "epoch": 6.598411297440424, + "grad_norm": 0.8992473823546595, + "learning_rate": 2.731322266125564e-06, + "loss": 0.0107, + "step": 1869 + }, + { + "epoch": 6.601941747572815, + "grad_norm": 0.9193192739275635, + "learning_rate": 2.7262243222194728e-06, + "loss": 0.0166, + "step": 1870 + }, + { + "epoch": 6.6054721977052075, + "grad_norm": 0.935566870415597, + "learning_rate": 2.7211293565734365e-06, + "loss": 0.0141, + "step": 1871 + }, + { + "epoch": 6.609002647837599, + "grad_norm": 1.201315262896322, + "learning_rate": 2.7160373758609976e-06, + "loss": 0.0145, + "step": 1872 + }, + { + "epoch": 6.612533097969991, + "grad_norm": 0.8316650103557445, + "learning_rate": 2.710948386751787e-06, + "loss": 0.0151, + "step": 1873 + }, + { + "epoch": 6.6160635481023835, + "grad_norm": 0.8104273659997085, + "learning_rate": 2.7058623959115145e-06, + "loss": 0.0111, + "step": 1874 + }, + { + "epoch": 6.619593998234775, + "grad_norm": 0.7340846212780024, + "learning_rate": 2.700779410001965e-06, + "loss": 0.0067, + "step": 1875 + }, + { + "epoch": 6.623124448367167, + "grad_norm": 0.9805680389902972, + "learning_rate": 2.695699435680986e-06, + "loss": 0.0175, + "step": 1876 + }, + { + "epoch": 6.626654898499559, + "grad_norm": 0.8244971366132741, + "learning_rate": 2.6906224796024853e-06, + "loss": 0.0133, + "step": 1877 + }, + { + "epoch": 6.630185348631951, + "grad_norm": 0.8762028315644956, + "learning_rate": 2.6855485484164085e-06, + "loss": 0.0126, + "step": 1878 + }, + { + "epoch": 6.633715798764342, + "grad_norm": 0.9551396508222052, + "learning_rate": 2.6804776487687457e-06, + "loss": 0.016, + "step": 1879 + }, + { + "epoch": 6.6372462488967345, + "grad_norm": 0.9800507714624233, + "learning_rate": 2.6754097873015152e-06, + "loss": 0.0191, + "step": 1880 + }, + { + "epoch": 6.640776699029126, + "grad_norm": 1.0942378282378151, + "learning_rate": 2.6703449706527553e-06, + "loss": 0.0157, + "step": 1881 + }, + { + "epoch": 6.644307149161518, + "grad_norm": 0.951998750383407, + "learning_rate": 2.6652832054565136e-06, + "loss": 0.0219, + "step": 1882 + }, + { + "epoch": 6.64783759929391, + "grad_norm": 0.6120323544731647, + "learning_rate": 2.660224498342844e-06, + "loss": 0.0108, + "step": 1883 + }, + { + "epoch": 6.651368049426302, + "grad_norm": 0.9318417163719993, + "learning_rate": 2.6551688559377966e-06, + "loss": 0.0214, + "step": 1884 + }, + { + "epoch": 6.654898499558694, + "grad_norm": 0.43025322077322287, + "learning_rate": 2.6501162848634023e-06, + "loss": 0.008, + "step": 1885 + }, + { + "epoch": 6.658428949691086, + "grad_norm": 0.5472856280663208, + "learning_rate": 2.645066791737668e-06, + "loss": 0.0053, + "step": 1886 + }, + { + "epoch": 6.661959399823478, + "grad_norm": 0.9614000557393698, + "learning_rate": 2.6400203831745765e-06, + "loss": 0.0134, + "step": 1887 + }, + { + "epoch": 6.665489849955869, + "grad_norm": 0.9238398634240476, + "learning_rate": 2.6349770657840653e-06, + "loss": 0.0212, + "step": 1888 + }, + { + "epoch": 6.6690203000882615, + "grad_norm": 0.5845416842585917, + "learning_rate": 2.629936846172025e-06, + "loss": 0.011, + "step": 1889 + }, + { + "epoch": 6.672550750220653, + "grad_norm": 0.8923501071752507, + "learning_rate": 2.6248997309402846e-06, + "loss": 0.0218, + "step": 1890 + }, + { + "epoch": 6.676081200353045, + "grad_norm": 0.7664771678371024, + "learning_rate": 2.6198657266866105e-06, + "loss": 0.0148, + "step": 1891 + }, + { + "epoch": 6.679611650485437, + "grad_norm": 1.1102112746038413, + "learning_rate": 2.6148348400046952e-06, + "loss": 0.0157, + "step": 1892 + }, + { + "epoch": 6.683142100617829, + "grad_norm": 0.8812808461757405, + "learning_rate": 2.609807077484147e-06, + "loss": 0.0121, + "step": 1893 + }, + { + "epoch": 6.68667255075022, + "grad_norm": 1.3396110449449714, + "learning_rate": 2.6047824457104766e-06, + "loss": 0.0334, + "step": 1894 + }, + { + "epoch": 6.6902030008826125, + "grad_norm": 0.6481928612986103, + "learning_rate": 2.5997609512651014e-06, + "loss": 0.0118, + "step": 1895 + }, + { + "epoch": 6.693733451015005, + "grad_norm": 1.187343268087044, + "learning_rate": 2.5947426007253256e-06, + "loss": 0.0191, + "step": 1896 + }, + { + "epoch": 6.697263901147396, + "grad_norm": 0.7336948290805988, + "learning_rate": 2.589727400664339e-06, + "loss": 0.0141, + "step": 1897 + }, + { + "epoch": 6.7007943512797885, + "grad_norm": 0.8481627858263938, + "learning_rate": 2.5847153576511986e-06, + "loss": 0.0181, + "step": 1898 + }, + { + "epoch": 6.70432480141218, + "grad_norm": 1.011115573530793, + "learning_rate": 2.579706478250833e-06, + "loss": 0.0148, + "step": 1899 + }, + { + "epoch": 6.707855251544572, + "grad_norm": 0.8533148784996211, + "learning_rate": 2.57470076902402e-06, + "loss": 0.0153, + "step": 1900 + }, + { + "epoch": 6.711385701676964, + "grad_norm": 0.7164591734752312, + "learning_rate": 2.569698236527392e-06, + "loss": 0.0132, + "step": 1901 + }, + { + "epoch": 6.714916151809356, + "grad_norm": 0.8649800018686793, + "learning_rate": 2.5646988873134137e-06, + "loss": 0.0197, + "step": 1902 + }, + { + "epoch": 6.718446601941747, + "grad_norm": 0.6278288819065663, + "learning_rate": 2.559702727930386e-06, + "loss": 0.0088, + "step": 1903 + }, + { + "epoch": 6.7219770520741395, + "grad_norm": 0.9383944545936675, + "learning_rate": 2.5547097649224288e-06, + "loss": 0.0175, + "step": 1904 + }, + { + "epoch": 6.725507502206531, + "grad_norm": 0.5390575913706852, + "learning_rate": 2.5497200048294782e-06, + "loss": 0.0077, + "step": 1905 + }, + { + "epoch": 6.729037952338923, + "grad_norm": 0.5010840494966236, + "learning_rate": 2.544733454187269e-06, + "loss": 0.0092, + "step": 1906 + }, + { + "epoch": 6.7325684024713155, + "grad_norm": 0.9902286701426083, + "learning_rate": 2.5397501195273367e-06, + "loss": 0.0152, + "step": 1907 + }, + { + "epoch": 6.736098852603707, + "grad_norm": 0.48327255331507063, + "learning_rate": 2.534770007377004e-06, + "loss": 0.0102, + "step": 1908 + }, + { + "epoch": 6.739629302736099, + "grad_norm": 0.6367480014254878, + "learning_rate": 2.5297931242593754e-06, + "loss": 0.012, + "step": 1909 + }, + { + "epoch": 6.743159752868491, + "grad_norm": 0.7721470526872342, + "learning_rate": 2.5248194766933175e-06, + "loss": 0.0117, + "step": 1910 + }, + { + "epoch": 6.746690203000883, + "grad_norm": 1.4151775285806047, + "learning_rate": 2.5198490711934664e-06, + "loss": 0.0225, + "step": 1911 + }, + { + "epoch": 6.750220653133274, + "grad_norm": 0.6362601610085807, + "learning_rate": 2.5148819142702095e-06, + "loss": 0.0116, + "step": 1912 + }, + { + "epoch": 6.7537511032656665, + "grad_norm": 1.9697456460989717, + "learning_rate": 2.5099180124296808e-06, + "loss": 0.0185, + "step": 1913 + }, + { + "epoch": 6.757281553398058, + "grad_norm": 0.8738668942530013, + "learning_rate": 2.5049573721737465e-06, + "loss": 0.0138, + "step": 1914 + }, + { + "epoch": 6.76081200353045, + "grad_norm": 0.9937770713380689, + "learning_rate": 2.5000000000000015e-06, + "loss": 0.0176, + "step": 1915 + }, + { + "epoch": 6.764342453662842, + "grad_norm": 1.0050732272796867, + "learning_rate": 2.495045902401763e-06, + "loss": 0.014, + "step": 1916 + }, + { + "epoch": 6.767872903795234, + "grad_norm": 1.3347570778606626, + "learning_rate": 2.490095085868059e-06, + "loss": 0.0243, + "step": 1917 + }, + { + "epoch": 6.771403353927626, + "grad_norm": 0.9166621336672969, + "learning_rate": 2.485147556883619e-06, + "loss": 0.0155, + "step": 1918 + }, + { + "epoch": 6.7749338040600176, + "grad_norm": 0.8820158593931967, + "learning_rate": 2.4802033219288628e-06, + "loss": 0.015, + "step": 1919 + }, + { + "epoch": 6.77846425419241, + "grad_norm": 0.7910362079684602, + "learning_rate": 2.4752623874799004e-06, + "loss": 0.0157, + "step": 1920 + }, + { + "epoch": 6.781994704324801, + "grad_norm": 0.6860221705457729, + "learning_rate": 2.4703247600085173e-06, + "loss": 0.0079, + "step": 1921 + }, + { + "epoch": 6.7855251544571935, + "grad_norm": 0.9611363145829765, + "learning_rate": 2.4653904459821687e-06, + "loss": 0.0186, + "step": 1922 + }, + { + "epoch": 6.789055604589585, + "grad_norm": 0.818824247654669, + "learning_rate": 2.4604594518639643e-06, + "loss": 0.0171, + "step": 1923 + }, + { + "epoch": 6.792586054721977, + "grad_norm": 0.8222351344845611, + "learning_rate": 2.4555317841126708e-06, + "loss": 0.0131, + "step": 1924 + }, + { + "epoch": 6.796116504854369, + "grad_norm": 1.0577365510329078, + "learning_rate": 2.450607449182698e-06, + "loss": 0.0168, + "step": 1925 + }, + { + "epoch": 6.799646954986761, + "grad_norm": 0.5482130314322278, + "learning_rate": 2.445686453524086e-06, + "loss": 0.0125, + "step": 1926 + }, + { + "epoch": 6.803177405119152, + "grad_norm": 0.8680466140874611, + "learning_rate": 2.4407688035825063e-06, + "loss": 0.0093, + "step": 1927 + }, + { + "epoch": 6.8067078552515445, + "grad_norm": 0.5131581636918263, + "learning_rate": 2.435854505799241e-06, + "loss": 0.0098, + "step": 1928 + }, + { + "epoch": 6.810238305383937, + "grad_norm": 0.9101697854471263, + "learning_rate": 2.430943566611189e-06, + "loss": 0.0118, + "step": 1929 + }, + { + "epoch": 6.813768755516328, + "grad_norm": 0.5341481163111168, + "learning_rate": 2.426035992450848e-06, + "loss": 0.0092, + "step": 1930 + }, + { + "epoch": 6.8172992056487205, + "grad_norm": 0.9095914189380372, + "learning_rate": 2.4211317897463034e-06, + "loss": 0.0115, + "step": 1931 + }, + { + "epoch": 6.820829655781112, + "grad_norm": 0.631273466763662, + "learning_rate": 2.416230964921229e-06, + "loss": 0.0159, + "step": 1932 + }, + { + "epoch": 6.824360105913504, + "grad_norm": 0.8583325214226256, + "learning_rate": 2.411333524394874e-06, + "loss": 0.0169, + "step": 1933 + }, + { + "epoch": 6.827890556045896, + "grad_norm": 1.0499011742855509, + "learning_rate": 2.406439474582056e-06, + "loss": 0.0172, + "step": 1934 + }, + { + "epoch": 6.831421006178288, + "grad_norm": 0.8256076385783444, + "learning_rate": 2.4015488218931443e-06, + "loss": 0.0083, + "step": 1935 + }, + { + "epoch": 6.834951456310679, + "grad_norm": 0.6873671803502874, + "learning_rate": 2.3966615727340665e-06, + "loss": 0.0083, + "step": 1936 + }, + { + "epoch": 6.8384819064430715, + "grad_norm": 0.6629323929297138, + "learning_rate": 2.3917777335062887e-06, + "loss": 0.0113, + "step": 1937 + }, + { + "epoch": 6.842012356575463, + "grad_norm": 0.6795199672008525, + "learning_rate": 2.3868973106068133e-06, + "loss": 0.0067, + "step": 1938 + }, + { + "epoch": 6.845542806707855, + "grad_norm": 0.8115238056818895, + "learning_rate": 2.3820203104281616e-06, + "loss": 0.0118, + "step": 1939 + }, + { + "epoch": 6.8490732568402475, + "grad_norm": 1.0808659517813288, + "learning_rate": 2.3771467393583774e-06, + "loss": 0.0188, + "step": 1940 + }, + { + "epoch": 6.852603706972639, + "grad_norm": 0.6744109349732157, + "learning_rate": 2.3722766037810115e-06, + "loss": 0.0116, + "step": 1941 + }, + { + "epoch": 6.856134157105031, + "grad_norm": 0.665080482410981, + "learning_rate": 2.3674099100751184e-06, + "loss": 0.012, + "step": 1942 + }, + { + "epoch": 6.8596646072374226, + "grad_norm": 0.915061757849712, + "learning_rate": 2.3625466646152328e-06, + "loss": 0.0152, + "step": 1943 + }, + { + "epoch": 6.863195057369815, + "grad_norm": 0.8925946977323329, + "learning_rate": 2.3576868737713845e-06, + "loss": 0.0178, + "step": 1944 + }, + { + "epoch": 6.866725507502206, + "grad_norm": 0.7297963848627573, + "learning_rate": 2.3528305439090744e-06, + "loss": 0.0121, + "step": 1945 + }, + { + "epoch": 6.8702559576345985, + "grad_norm": 0.8661398087859238, + "learning_rate": 2.3479776813892714e-06, + "loss": 0.0195, + "step": 1946 + }, + { + "epoch": 6.87378640776699, + "grad_norm": 0.6966450196053696, + "learning_rate": 2.3431282925683974e-06, + "loss": 0.0124, + "step": 1947 + }, + { + "epoch": 6.877316857899382, + "grad_norm": 0.3954236101468094, + "learning_rate": 2.3382823837983314e-06, + "loss": 0.006, + "step": 1948 + }, + { + "epoch": 6.880847308031774, + "grad_norm": 0.8363443158804558, + "learning_rate": 2.3334399614263904e-06, + "loss": 0.0077, + "step": 1949 + }, + { + "epoch": 6.884377758164166, + "grad_norm": 0.72788138750247, + "learning_rate": 2.3286010317953267e-06, + "loss": 0.0112, + "step": 1950 + }, + { + "epoch": 6.887908208296558, + "grad_norm": 0.6651500351212859, + "learning_rate": 2.323765601243314e-06, + "loss": 0.01, + "step": 1951 + }, + { + "epoch": 6.8914386584289495, + "grad_norm": 0.9081927876922583, + "learning_rate": 2.3189336761039467e-06, + "loss": 0.0148, + "step": 1952 + }, + { + "epoch": 6.894969108561342, + "grad_norm": 0.8026395554971862, + "learning_rate": 2.3141052627062254e-06, + "loss": 0.0181, + "step": 1953 + }, + { + "epoch": 6.898499558693733, + "grad_norm": 1.5205412843242194, + "learning_rate": 2.309280367374555e-06, + "loss": 0.0166, + "step": 1954 + }, + { + "epoch": 6.9020300088261255, + "grad_norm": 0.9197943922238804, + "learning_rate": 2.3044589964287245e-06, + "loss": 0.0111, + "step": 1955 + }, + { + "epoch": 6.905560458958517, + "grad_norm": 1.056401242029234, + "learning_rate": 2.2996411561839156e-06, + "loss": 0.0186, + "step": 1956 + }, + { + "epoch": 6.909090909090909, + "grad_norm": 0.513006848713881, + "learning_rate": 2.2948268529506768e-06, + "loss": 0.009, + "step": 1957 + }, + { + "epoch": 6.9126213592233015, + "grad_norm": 0.6976548117747184, + "learning_rate": 2.290016093034929e-06, + "loss": 0.0105, + "step": 1958 + }, + { + "epoch": 6.916151809355693, + "grad_norm": 0.9209561646218092, + "learning_rate": 2.285208882737954e-06, + "loss": 0.0178, + "step": 1959 + }, + { + "epoch": 6.919682259488084, + "grad_norm": 0.614648579383868, + "learning_rate": 2.280405228356377e-06, + "loss": 0.0089, + "step": 1960 + }, + { + "epoch": 6.9232127096204765, + "grad_norm": 1.3433422941865805, + "learning_rate": 2.2756051361821712e-06, + "loss": 0.0183, + "step": 1961 + }, + { + "epoch": 6.926743159752869, + "grad_norm": 0.5144979326196759, + "learning_rate": 2.2708086125026423e-06, + "loss": 0.0082, + "step": 1962 + }, + { + "epoch": 6.93027360988526, + "grad_norm": 0.4849054336178533, + "learning_rate": 2.2660156636004237e-06, + "loss": 0.0064, + "step": 1963 + }, + { + "epoch": 6.9338040600176525, + "grad_norm": 0.7074206184245313, + "learning_rate": 2.2612262957534615e-06, + "loss": 0.0082, + "step": 1964 + }, + { + "epoch": 6.937334510150044, + "grad_norm": 0.9798304944848538, + "learning_rate": 2.2564405152350154e-06, + "loss": 0.0187, + "step": 1965 + }, + { + "epoch": 6.940864960282436, + "grad_norm": 0.8255646303651836, + "learning_rate": 2.2516583283136474e-06, + "loss": 0.0153, + "step": 1966 + }, + { + "epoch": 6.9443954104148276, + "grad_norm": 0.8138807670103092, + "learning_rate": 2.2468797412532067e-06, + "loss": 0.0118, + "step": 1967 + }, + { + "epoch": 6.94792586054722, + "grad_norm": 0.8796999317032345, + "learning_rate": 2.2421047603128326e-06, + "loss": 0.0148, + "step": 1968 + }, + { + "epoch": 6.951456310679612, + "grad_norm": 0.7049390667389277, + "learning_rate": 2.2373333917469398e-06, + "loss": 0.0095, + "step": 1969 + }, + { + "epoch": 6.9549867608120035, + "grad_norm": 0.8813131782643769, + "learning_rate": 2.2325656418052126e-06, + "loss": 0.0141, + "step": 1970 + }, + { + "epoch": 6.958517210944396, + "grad_norm": 1.0821379416594645, + "learning_rate": 2.227801516732592e-06, + "loss": 0.0246, + "step": 1971 + }, + { + "epoch": 6.962047661076787, + "grad_norm": 0.8293318732138872, + "learning_rate": 2.223041022769271e-06, + "loss": 0.0165, + "step": 1972 + }, + { + "epoch": 6.9655781112091795, + "grad_norm": 1.285384829264463, + "learning_rate": 2.2182841661506905e-06, + "loss": 0.0219, + "step": 1973 + }, + { + "epoch": 6.969108561341571, + "grad_norm": 1.0247758371866929, + "learning_rate": 2.213530953107525e-06, + "loss": 0.0083, + "step": 1974 + }, + { + "epoch": 6.972639011473963, + "grad_norm": 0.5660432067307707, + "learning_rate": 2.2087813898656775e-06, + "loss": 0.0105, + "step": 1975 + }, + { + "epoch": 6.9761694616063545, + "grad_norm": 0.918646032004212, + "learning_rate": 2.204035482646267e-06, + "loss": 0.0137, + "step": 1976 + }, + { + "epoch": 6.979699911738747, + "grad_norm": 0.6067421613553441, + "learning_rate": 2.1992932376656262e-06, + "loss": 0.0078, + "step": 1977 + }, + { + "epoch": 6.983230361871138, + "grad_norm": 1.1112636609639794, + "learning_rate": 2.1945546611352926e-06, + "loss": 0.0166, + "step": 1978 + }, + { + "epoch": 6.9867608120035305, + "grad_norm": 0.7293652578857724, + "learning_rate": 2.189819759261998e-06, + "loss": 0.0158, + "step": 1979 + }, + { + "epoch": 6.990291262135923, + "grad_norm": 0.778885502392239, + "learning_rate": 2.185088538247656e-06, + "loss": 0.0142, + "step": 1980 + }, + { + "epoch": 6.993821712268314, + "grad_norm": 0.6563819944306686, + "learning_rate": 2.1803610042893646e-06, + "loss": 0.0117, + "step": 1981 + }, + { + "epoch": 6.9973521624007065, + "grad_norm": 0.8609267221411963, + "learning_rate": 2.175637163579391e-06, + "loss": 0.0156, + "step": 1982 + }, + { + "epoch": 7.000882612533098, + "grad_norm": 0.45708502568103765, + "learning_rate": 2.170917022305166e-06, + "loss": 0.0056, + "step": 1983 + }, + { + "epoch": 7.00441306266549, + "grad_norm": 0.41228931911948263, + "learning_rate": 2.166200586649272e-06, + "loss": 0.0045, + "step": 1984 + }, + { + "epoch": 7.0079435127978815, + "grad_norm": 0.860623181544306, + "learning_rate": 2.1614878627894353e-06, + "loss": 0.007, + "step": 1985 + }, + { + "epoch": 7.011473962930274, + "grad_norm": 0.6781060290779626, + "learning_rate": 2.1567788568985277e-06, + "loss": 0.0057, + "step": 1986 + }, + { + "epoch": 7.015004413062665, + "grad_norm": 0.3505429670800696, + "learning_rate": 2.1520735751445484e-06, + "loss": 0.004, + "step": 1987 + }, + { + "epoch": 7.0185348631950575, + "grad_norm": 0.24554003495832402, + "learning_rate": 2.1473720236906133e-06, + "loss": 0.0056, + "step": 1988 + }, + { + "epoch": 7.022065313327449, + "grad_norm": 0.31769004256254246, + "learning_rate": 2.142674208694959e-06, + "loss": 0.0044, + "step": 1989 + }, + { + "epoch": 7.025595763459841, + "grad_norm": 0.6127315581118619, + "learning_rate": 2.137980136310926e-06, + "loss": 0.0069, + "step": 1990 + }, + { + "epoch": 7.029126213592233, + "grad_norm": 0.41860301271983713, + "learning_rate": 2.133289812686953e-06, + "loss": 0.0068, + "step": 1991 + }, + { + "epoch": 7.032656663724625, + "grad_norm": 0.5457155274188568, + "learning_rate": 2.128603243966564e-06, + "loss": 0.0079, + "step": 1992 + }, + { + "epoch": 7.036187113857017, + "grad_norm": 0.31978560007086737, + "learning_rate": 2.12392043628837e-06, + "loss": 0.0043, + "step": 1993 + }, + { + "epoch": 7.0397175639894085, + "grad_norm": 1.4835532814594143, + "learning_rate": 2.119241395786054e-06, + "loss": 0.0086, + "step": 1994 + }, + { + "epoch": 7.043248014121801, + "grad_norm": 0.3890570564427796, + "learning_rate": 2.1145661285883673e-06, + "loss": 0.0069, + "step": 1995 + }, + { + "epoch": 7.046778464254192, + "grad_norm": 0.2527518302515535, + "learning_rate": 2.1098946408191117e-06, + "loss": 0.0024, + "step": 1996 + }, + { + "epoch": 7.0503089143865845, + "grad_norm": 0.3061895282081957, + "learning_rate": 2.105226938597144e-06, + "loss": 0.0056, + "step": 1997 + }, + { + "epoch": 7.053839364518976, + "grad_norm": 0.226779447461953, + "learning_rate": 2.1005630280363652e-06, + "loss": 0.0045, + "step": 1998 + }, + { + "epoch": 7.057369814651368, + "grad_norm": 0.3313997256002707, + "learning_rate": 2.095902915245701e-06, + "loss": 0.0055, + "step": 1999 + }, + { + "epoch": 7.0609002647837595, + "grad_norm": 1.0582597554848623, + "learning_rate": 2.091246606329112e-06, + "loss": 0.011, + "step": 2000 + }, + { + "epoch": 7.064430714916152, + "grad_norm": 0.6157503968251556, + "learning_rate": 2.086594107385569e-06, + "loss": 0.0078, + "step": 2001 + }, + { + "epoch": 7.067961165048544, + "grad_norm": 0.641112230188211, + "learning_rate": 2.081945424509057e-06, + "loss": 0.0114, + "step": 2002 + }, + { + "epoch": 7.0714916151809355, + "grad_norm": 0.2089418496378771, + "learning_rate": 2.077300563788563e-06, + "loss": 0.0022, + "step": 2003 + }, + { + "epoch": 7.075022065313328, + "grad_norm": 0.28204573811633876, + "learning_rate": 2.072659531308063e-06, + "loss": 0.0026, + "step": 2004 + }, + { + "epoch": 7.078552515445719, + "grad_norm": 0.3704385691129231, + "learning_rate": 2.0680223331465223e-06, + "loss": 0.0089, + "step": 2005 + }, + { + "epoch": 7.0820829655781115, + "grad_norm": 0.7818479441474135, + "learning_rate": 2.0633889753778834e-06, + "loss": 0.0096, + "step": 2006 + }, + { + "epoch": 7.085613415710503, + "grad_norm": 0.71981122846549, + "learning_rate": 2.0587594640710606e-06, + "loss": 0.0072, + "step": 2007 + }, + { + "epoch": 7.089143865842895, + "grad_norm": 0.3295872945060594, + "learning_rate": 2.0541338052899234e-06, + "loss": 0.0044, + "step": 2008 + }, + { + "epoch": 7.0926743159752865, + "grad_norm": 0.6149532844647213, + "learning_rate": 2.0495120050933014e-06, + "loss": 0.0059, + "step": 2009 + }, + { + "epoch": 7.096204766107679, + "grad_norm": 0.7142752029242428, + "learning_rate": 2.044894069534968e-06, + "loss": 0.0066, + "step": 2010 + }, + { + "epoch": 7.09973521624007, + "grad_norm": 0.649415396190019, + "learning_rate": 2.040280004663637e-06, + "loss": 0.0071, + "step": 2011 + }, + { + "epoch": 7.1032656663724625, + "grad_norm": 0.35317897996574077, + "learning_rate": 2.0356698165229455e-06, + "loss": 0.0043, + "step": 2012 + }, + { + "epoch": 7.106796116504855, + "grad_norm": 0.3363756212097448, + "learning_rate": 2.0310635111514616e-06, + "loss": 0.0059, + "step": 2013 + }, + { + "epoch": 7.110326566637246, + "grad_norm": 0.24743127059106357, + "learning_rate": 2.0264610945826584e-06, + "loss": 0.0026, + "step": 2014 + }, + { + "epoch": 7.113857016769638, + "grad_norm": 0.39774338042799795, + "learning_rate": 2.0218625728449232e-06, + "loss": 0.0037, + "step": 2015 + }, + { + "epoch": 7.11738746690203, + "grad_norm": 0.682295829514909, + "learning_rate": 2.01726795196154e-06, + "loss": 0.0076, + "step": 2016 + }, + { + "epoch": 7.120917917034422, + "grad_norm": 0.40853123280689735, + "learning_rate": 2.012677237950678e-06, + "loss": 0.0049, + "step": 2017 + }, + { + "epoch": 7.1244483671668135, + "grad_norm": 0.22303413554625812, + "learning_rate": 2.0080904368253954e-06, + "loss": 0.0029, + "step": 2018 + }, + { + "epoch": 7.127978817299206, + "grad_norm": 0.5940883212491138, + "learning_rate": 2.003507554593623e-06, + "loss": 0.0084, + "step": 2019 + }, + { + "epoch": 7.131509267431597, + "grad_norm": 0.655619517788559, + "learning_rate": 1.9989285972581595e-06, + "loss": 0.0098, + "step": 2020 + }, + { + "epoch": 7.1350397175639895, + "grad_norm": 0.5372316904233096, + "learning_rate": 1.994353570816659e-06, + "loss": 0.0068, + "step": 2021 + }, + { + "epoch": 7.138570167696381, + "grad_norm": 0.511811445589073, + "learning_rate": 1.9897824812616297e-06, + "loss": 0.0055, + "step": 2022 + }, + { + "epoch": 7.142100617828773, + "grad_norm": 0.5876305222452937, + "learning_rate": 1.985215334580423e-06, + "loss": 0.0062, + "step": 2023 + }, + { + "epoch": 7.145631067961165, + "grad_norm": 0.30353105587206025, + "learning_rate": 1.9806521367552285e-06, + "loss": 0.0054, + "step": 2024 + }, + { + "epoch": 7.149161518093557, + "grad_norm": 0.5791532385973693, + "learning_rate": 1.976092893763056e-06, + "loss": 0.0074, + "step": 2025 + }, + { + "epoch": 7.152691968225949, + "grad_norm": 0.5283074351080544, + "learning_rate": 1.9715376115757417e-06, + "loss": 0.0084, + "step": 2026 + }, + { + "epoch": 7.1562224183583405, + "grad_norm": 0.23831832979632886, + "learning_rate": 1.9669862961599336e-06, + "loss": 0.0021, + "step": 2027 + }, + { + "epoch": 7.159752868490733, + "grad_norm": 0.2727166369731957, + "learning_rate": 1.96243895347708e-06, + "loss": 0.0034, + "step": 2028 + }, + { + "epoch": 7.163283318623124, + "grad_norm": 0.30047676873329326, + "learning_rate": 1.957895589483426e-06, + "loss": 0.0019, + "step": 2029 + }, + { + "epoch": 7.1668137687555165, + "grad_norm": 0.47331313177520057, + "learning_rate": 1.95335621013001e-06, + "loss": 0.0038, + "step": 2030 + }, + { + "epoch": 7.170344218887908, + "grad_norm": 0.4382765369790411, + "learning_rate": 1.9488208213626456e-06, + "loss": 0.0067, + "step": 2031 + }, + { + "epoch": 7.1738746690203, + "grad_norm": 0.3281418331048503, + "learning_rate": 1.944289429121927e-06, + "loss": 0.004, + "step": 2032 + }, + { + "epoch": 7.1774051191526915, + "grad_norm": 0.8221403748320214, + "learning_rate": 1.939762039343202e-06, + "loss": 0.0122, + "step": 2033 + }, + { + "epoch": 7.180935569285084, + "grad_norm": 0.38500855845583043, + "learning_rate": 1.9352386579565857e-06, + "loss": 0.0033, + "step": 2034 + }, + { + "epoch": 7.184466019417476, + "grad_norm": 0.5352117347019546, + "learning_rate": 1.93071929088694e-06, + "loss": 0.0078, + "step": 2035 + }, + { + "epoch": 7.1879964695498675, + "grad_norm": 0.622601085026879, + "learning_rate": 1.92620394405387e-06, + "loss": 0.007, + "step": 2036 + }, + { + "epoch": 7.19152691968226, + "grad_norm": 0.4168418000092725, + "learning_rate": 1.9216926233717087e-06, + "loss": 0.0062, + "step": 2037 + }, + { + "epoch": 7.195057369814651, + "grad_norm": 0.34094665153335063, + "learning_rate": 1.9171853347495234e-06, + "loss": 0.0043, + "step": 2038 + }, + { + "epoch": 7.198587819947043, + "grad_norm": 0.5910785636524729, + "learning_rate": 1.912682084091096e-06, + "loss": 0.0063, + "step": 2039 + }, + { + "epoch": 7.202118270079435, + "grad_norm": 0.5996828888297475, + "learning_rate": 1.9081828772949223e-06, + "loss": 0.0049, + "step": 2040 + }, + { + "epoch": 7.205648720211827, + "grad_norm": 0.6192614461202353, + "learning_rate": 1.9036877202541976e-06, + "loss": 0.0095, + "step": 2041 + }, + { + "epoch": 7.2091791703442185, + "grad_norm": 0.3319908993779631, + "learning_rate": 1.8991966188568123e-06, + "loss": 0.004, + "step": 2042 + }, + { + "epoch": 7.212709620476611, + "grad_norm": 0.5245089784216443, + "learning_rate": 1.8947095789853486e-06, + "loss": 0.006, + "step": 2043 + }, + { + "epoch": 7.216240070609003, + "grad_norm": 0.34386359219451257, + "learning_rate": 1.8902266065170688e-06, + "loss": 0.003, + "step": 2044 + }, + { + "epoch": 7.2197705207413945, + "grad_norm": 0.5591308282406359, + "learning_rate": 1.8857477073239011e-06, + "loss": 0.0083, + "step": 2045 + }, + { + "epoch": 7.223300970873787, + "grad_norm": 0.5922921780554933, + "learning_rate": 1.8812728872724461e-06, + "loss": 0.009, + "step": 2046 + }, + { + "epoch": 7.226831421006178, + "grad_norm": 0.5073243952558515, + "learning_rate": 1.8768021522239576e-06, + "loss": 0.0058, + "step": 2047 + }, + { + "epoch": 7.23036187113857, + "grad_norm": 0.591952255271864, + "learning_rate": 1.8723355080343407e-06, + "loss": 0.0062, + "step": 2048 + }, + { + "epoch": 7.233892321270962, + "grad_norm": 0.8815277963979592, + "learning_rate": 1.8678729605541384e-06, + "loss": 0.0069, + "step": 2049 + }, + { + "epoch": 7.237422771403354, + "grad_norm": 0.8332313410047736, + "learning_rate": 1.8634145156285311e-06, + "loss": 0.0053, + "step": 2050 + }, + { + "epoch": 7.2409532215357455, + "grad_norm": 0.4309412894897188, + "learning_rate": 1.858960179097325e-06, + "loss": 0.0058, + "step": 2051 + }, + { + "epoch": 7.244483671668138, + "grad_norm": 0.45352455791855145, + "learning_rate": 1.854509956794946e-06, + "loss": 0.008, + "step": 2052 + }, + { + "epoch": 7.248014121800529, + "grad_norm": 0.46547305020230023, + "learning_rate": 1.8500638545504262e-06, + "loss": 0.0062, + "step": 2053 + }, + { + "epoch": 7.2515445719329215, + "grad_norm": 0.25154463049717257, + "learning_rate": 1.8456218781874068e-06, + "loss": 0.0035, + "step": 2054 + }, + { + "epoch": 7.255075022065314, + "grad_norm": 0.2635764755345156, + "learning_rate": 1.8411840335241243e-06, + "loss": 0.0021, + "step": 2055 + }, + { + "epoch": 7.258605472197705, + "grad_norm": 0.7365464708637562, + "learning_rate": 1.8367503263733983e-06, + "loss": 0.0079, + "step": 2056 + }, + { + "epoch": 7.262135922330097, + "grad_norm": 0.3358283271160915, + "learning_rate": 1.8323207625426349e-06, + "loss": 0.0045, + "step": 2057 + }, + { + "epoch": 7.265666372462489, + "grad_norm": 0.34429634247879287, + "learning_rate": 1.8278953478338079e-06, + "loss": 0.0034, + "step": 2058 + }, + { + "epoch": 7.269196822594881, + "grad_norm": 0.6236794692286916, + "learning_rate": 1.8234740880434609e-06, + "loss": 0.0073, + "step": 2059 + }, + { + "epoch": 7.2727272727272725, + "grad_norm": 0.9276781390452721, + "learning_rate": 1.8190569889626935e-06, + "loss": 0.0039, + "step": 2060 + }, + { + "epoch": 7.276257722859665, + "grad_norm": 0.1918395840003646, + "learning_rate": 1.8146440563771568e-06, + "loss": 0.0024, + "step": 2061 + }, + { + "epoch": 7.279788172992056, + "grad_norm": 0.5367038996008521, + "learning_rate": 1.8102352960670399e-06, + "loss": 0.0066, + "step": 2062 + }, + { + "epoch": 7.283318623124448, + "grad_norm": 0.44161984944274896, + "learning_rate": 1.8058307138070723e-06, + "loss": 0.0063, + "step": 2063 + }, + { + "epoch": 7.28684907325684, + "grad_norm": 0.43549246570154365, + "learning_rate": 1.8014303153665103e-06, + "loss": 0.0075, + "step": 2064 + }, + { + "epoch": 7.290379523389232, + "grad_norm": 0.8507443134547099, + "learning_rate": 1.7970341065091246e-06, + "loss": 0.0052, + "step": 2065 + }, + { + "epoch": 7.293909973521624, + "grad_norm": 0.5504989241058721, + "learning_rate": 1.7926420929932054e-06, + "loss": 0.0076, + "step": 2066 + }, + { + "epoch": 7.297440423654016, + "grad_norm": 0.41960112267243754, + "learning_rate": 1.7882542805715436e-06, + "loss": 0.0043, + "step": 2067 + }, + { + "epoch": 7.300970873786408, + "grad_norm": 0.7435710418468212, + "learning_rate": 1.7838706749914308e-06, + "loss": 0.0096, + "step": 2068 + }, + { + "epoch": 7.3045013239187995, + "grad_norm": 0.45829428278591544, + "learning_rate": 1.7794912819946419e-06, + "loss": 0.0057, + "step": 2069 + }, + { + "epoch": 7.308031774051192, + "grad_norm": 0.27029191110441214, + "learning_rate": 1.7751161073174422e-06, + "loss": 0.0037, + "step": 2070 + }, + { + "epoch": 7.311562224183583, + "grad_norm": 0.9489709898055235, + "learning_rate": 1.7707451566905643e-06, + "loss": 0.0047, + "step": 2071 + }, + { + "epoch": 7.315092674315975, + "grad_norm": 0.792153850910637, + "learning_rate": 1.7663784358392122e-06, + "loss": 0.005, + "step": 2072 + }, + { + "epoch": 7.318623124448367, + "grad_norm": 0.4999468230848353, + "learning_rate": 1.762015950483051e-06, + "loss": 0.0043, + "step": 2073 + }, + { + "epoch": 7.322153574580759, + "grad_norm": 0.6331364396729859, + "learning_rate": 1.757657706336192e-06, + "loss": 0.0063, + "step": 2074 + }, + { + "epoch": 7.325684024713151, + "grad_norm": 0.379287331399495, + "learning_rate": 1.7533037091071975e-06, + "loss": 0.004, + "step": 2075 + }, + { + "epoch": 7.329214474845543, + "grad_norm": 0.5018784981519422, + "learning_rate": 1.7489539644990637e-06, + "loss": 0.0047, + "step": 2076 + }, + { + "epoch": 7.332744924977935, + "grad_norm": 0.3670848064767987, + "learning_rate": 1.74460847820922e-06, + "loss": 0.0027, + "step": 2077 + }, + { + "epoch": 7.3362753751103265, + "grad_norm": 0.5766467268399271, + "learning_rate": 1.7402672559295113e-06, + "loss": 0.0111, + "step": 2078 + }, + { + "epoch": 7.339805825242719, + "grad_norm": 0.33369604164571204, + "learning_rate": 1.7359303033462043e-06, + "loss": 0.0022, + "step": 2079 + }, + { + "epoch": 7.34333627537511, + "grad_norm": 0.4138627913785017, + "learning_rate": 1.7315976261399698e-06, + "loss": 0.006, + "step": 2080 + }, + { + "epoch": 7.346866725507502, + "grad_norm": 0.42408393021921204, + "learning_rate": 1.7272692299858812e-06, + "loss": 0.0051, + "step": 2081 + }, + { + "epoch": 7.350397175639894, + "grad_norm": 0.47352239596929324, + "learning_rate": 1.7229451205533986e-06, + "loss": 0.0036, + "step": 2082 + }, + { + "epoch": 7.353927625772286, + "grad_norm": 0.20783717162410653, + "learning_rate": 1.7186253035063738e-06, + "loss": 0.0016, + "step": 2083 + }, + { + "epoch": 7.3574580759046775, + "grad_norm": 0.5651863072814612, + "learning_rate": 1.7143097845030348e-06, + "loss": 0.0051, + "step": 2084 + }, + { + "epoch": 7.36098852603707, + "grad_norm": 0.9995881100743844, + "learning_rate": 1.7099985691959769e-06, + "loss": 0.0092, + "step": 2085 + }, + { + "epoch": 7.364518976169462, + "grad_norm": 0.8554874281693983, + "learning_rate": 1.705691663232159e-06, + "loss": 0.013, + "step": 2086 + }, + { + "epoch": 7.368049426301853, + "grad_norm": 0.5683866729820999, + "learning_rate": 1.7013890722528974e-06, + "loss": 0.0073, + "step": 2087 + }, + { + "epoch": 7.371579876434246, + "grad_norm": 0.36310200090132655, + "learning_rate": 1.697090801893857e-06, + "loss": 0.0045, + "step": 2088 + }, + { + "epoch": 7.375110326566637, + "grad_norm": 0.31785105317224766, + "learning_rate": 1.6927968577850423e-06, + "loss": 0.0042, + "step": 2089 + }, + { + "epoch": 7.378640776699029, + "grad_norm": 0.4559237735316684, + "learning_rate": 1.6885072455507884e-06, + "loss": 0.0048, + "step": 2090 + }, + { + "epoch": 7.382171226831421, + "grad_norm": 0.5468073117410707, + "learning_rate": 1.6842219708097611e-06, + "loss": 0.005, + "step": 2091 + }, + { + "epoch": 7.385701676963813, + "grad_norm": 0.37821943724848195, + "learning_rate": 1.6799410391749416e-06, + "loss": 0.0051, + "step": 2092 + }, + { + "epoch": 7.3892321270962045, + "grad_norm": 0.5523714408477713, + "learning_rate": 1.6756644562536262e-06, + "loss": 0.006, + "step": 2093 + }, + { + "epoch": 7.392762577228597, + "grad_norm": 0.6843443568960861, + "learning_rate": 1.671392227647408e-06, + "loss": 0.0055, + "step": 2094 + }, + { + "epoch": 7.396293027360988, + "grad_norm": 0.6058500299135865, + "learning_rate": 1.667124358952184e-06, + "loss": 0.0069, + "step": 2095 + }, + { + "epoch": 7.39982347749338, + "grad_norm": 0.35647004905274227, + "learning_rate": 1.6628608557581366e-06, + "loss": 0.0031, + "step": 2096 + }, + { + "epoch": 7.403353927625773, + "grad_norm": 0.5782037164835417, + "learning_rate": 1.6586017236497326e-06, + "loss": 0.004, + "step": 2097 + }, + { + "epoch": 7.406884377758164, + "grad_norm": 0.39804157540328444, + "learning_rate": 1.6543469682057105e-06, + "loss": 0.0042, + "step": 2098 + }, + { + "epoch": 7.410414827890556, + "grad_norm": 0.48856311744532227, + "learning_rate": 1.650096594999076e-06, + "loss": 0.0062, + "step": 2099 + }, + { + "epoch": 7.413945278022948, + "grad_norm": 0.8925058797105488, + "learning_rate": 1.6458506095970977e-06, + "loss": 0.0096, + "step": 2100 + }, + { + "epoch": 7.41747572815534, + "grad_norm": 0.6743148786789692, + "learning_rate": 1.6416090175612958e-06, + "loss": 0.0055, + "step": 2101 + }, + { + "epoch": 7.4210061782877315, + "grad_norm": 0.8471780481105783, + "learning_rate": 1.6373718244474374e-06, + "loss": 0.0059, + "step": 2102 + }, + { + "epoch": 7.424536628420124, + "grad_norm": 0.48063744571603145, + "learning_rate": 1.6331390358055217e-06, + "loss": 0.0062, + "step": 2103 + }, + { + "epoch": 7.428067078552515, + "grad_norm": 0.5041068910831926, + "learning_rate": 1.6289106571797857e-06, + "loss": 0.0064, + "step": 2104 + }, + { + "epoch": 7.431597528684907, + "grad_norm": 0.4912848052839785, + "learning_rate": 1.6246866941086897e-06, + "loss": 0.0062, + "step": 2105 + }, + { + "epoch": 7.435127978817299, + "grad_norm": 0.3098764251875192, + "learning_rate": 1.6204671521249033e-06, + "loss": 0.0043, + "step": 2106 + }, + { + "epoch": 7.438658428949691, + "grad_norm": 0.4904119311120319, + "learning_rate": 1.6162520367553126e-06, + "loss": 0.0053, + "step": 2107 + }, + { + "epoch": 7.442188879082083, + "grad_norm": 1.0753952448810695, + "learning_rate": 1.6120413535210023e-06, + "loss": 0.008, + "step": 2108 + }, + { + "epoch": 7.445719329214475, + "grad_norm": 0.4820978080707071, + "learning_rate": 1.6078351079372552e-06, + "loss": 0.0044, + "step": 2109 + }, + { + "epoch": 7.449249779346867, + "grad_norm": 0.277812752090112, + "learning_rate": 1.6036333055135345e-06, + "loss": 0.002, + "step": 2110 + }, + { + "epoch": 7.4527802294792584, + "grad_norm": 0.43817630954988956, + "learning_rate": 1.5994359517534902e-06, + "loss": 0.0036, + "step": 2111 + }, + { + "epoch": 7.456310679611651, + "grad_norm": 0.3239454796498814, + "learning_rate": 1.5952430521549444e-06, + "loss": 0.0033, + "step": 2112 + }, + { + "epoch": 7.459841129744042, + "grad_norm": 0.6587825872176601, + "learning_rate": 1.5910546122098808e-06, + "loss": 0.006, + "step": 2113 + }, + { + "epoch": 7.463371579876434, + "grad_norm": 0.4568883622526341, + "learning_rate": 1.5868706374044473e-06, + "loss": 0.0037, + "step": 2114 + }, + { + "epoch": 7.466902030008826, + "grad_norm": 0.6237094213650858, + "learning_rate": 1.582691133218937e-06, + "loss": 0.0051, + "step": 2115 + }, + { + "epoch": 7.470432480141218, + "grad_norm": 0.6039773932175183, + "learning_rate": 1.578516105127793e-06, + "loss": 0.0064, + "step": 2116 + }, + { + "epoch": 7.4739629302736095, + "grad_norm": 0.4489684614244742, + "learning_rate": 1.5743455585995926e-06, + "loss": 0.0051, + "step": 2117 + }, + { + "epoch": 7.477493380406002, + "grad_norm": 0.4515555231636241, + "learning_rate": 1.570179499097046e-06, + "loss": 0.0052, + "step": 2118 + }, + { + "epoch": 7.481023830538394, + "grad_norm": 0.433850925105127, + "learning_rate": 1.5660179320769792e-06, + "loss": 0.0044, + "step": 2119 + }, + { + "epoch": 7.484554280670785, + "grad_norm": 1.1217413190676948, + "learning_rate": 1.5618608629903415e-06, + "loss": 0.0085, + "step": 2120 + }, + { + "epoch": 7.488084730803178, + "grad_norm": 0.7289810401488171, + "learning_rate": 1.5577082972821865e-06, + "loss": 0.0098, + "step": 2121 + }, + { + "epoch": 7.491615180935569, + "grad_norm": 0.39796283163631685, + "learning_rate": 1.5535602403916723e-06, + "loss": 0.0054, + "step": 2122 + }, + { + "epoch": 7.495145631067961, + "grad_norm": 0.804935952251448, + "learning_rate": 1.5494166977520459e-06, + "loss": 0.0067, + "step": 2123 + }, + { + "epoch": 7.498676081200353, + "grad_norm": 0.7822886291712798, + "learning_rate": 1.5452776747906457e-06, + "loss": 0.0042, + "step": 2124 + }, + { + "epoch": 7.502206531332745, + "grad_norm": 0.3344541594188005, + "learning_rate": 1.541143176928891e-06, + "loss": 0.0028, + "step": 2125 + }, + { + "epoch": 7.5057369814651365, + "grad_norm": 0.49551225743622945, + "learning_rate": 1.5370132095822704e-06, + "loss": 0.0061, + "step": 2126 + }, + { + "epoch": 7.509267431597529, + "grad_norm": 0.45627096677134044, + "learning_rate": 1.532887778160338e-06, + "loss": 0.003, + "step": 2127 + }, + { + "epoch": 7.51279788172992, + "grad_norm": 0.3183685756463059, + "learning_rate": 1.5287668880667107e-06, + "loss": 0.0026, + "step": 2128 + }, + { + "epoch": 7.516328331862312, + "grad_norm": 0.9301888146063189, + "learning_rate": 1.524650544699054e-06, + "loss": 0.0048, + "step": 2129 + }, + { + "epoch": 7.519858781994705, + "grad_norm": 0.42670570471119096, + "learning_rate": 1.5205387534490807e-06, + "loss": 0.003, + "step": 2130 + }, + { + "epoch": 7.523389232127096, + "grad_norm": 0.24197029333035525, + "learning_rate": 1.5164315197025358e-06, + "loss": 0.0019, + "step": 2131 + }, + { + "epoch": 7.526919682259488, + "grad_norm": 0.29609102827844574, + "learning_rate": 1.5123288488392002e-06, + "loss": 0.0036, + "step": 2132 + }, + { + "epoch": 7.53045013239188, + "grad_norm": 0.4654242589811813, + "learning_rate": 1.5082307462328756e-06, + "loss": 0.0048, + "step": 2133 + }, + { + "epoch": 7.533980582524272, + "grad_norm": 0.4961315545286196, + "learning_rate": 1.5041372172513824e-06, + "loss": 0.0052, + "step": 2134 + }, + { + "epoch": 7.5375110326566634, + "grad_norm": 0.3019963213629793, + "learning_rate": 1.5000482672565452e-06, + "loss": 0.0029, + "step": 2135 + }, + { + "epoch": 7.541041482789056, + "grad_norm": 0.2522238559592797, + "learning_rate": 1.495963901604196e-06, + "loss": 0.003, + "step": 2136 + }, + { + "epoch": 7.544571932921447, + "grad_norm": 0.340253131958453, + "learning_rate": 1.4918841256441603e-06, + "loss": 0.0034, + "step": 2137 + }, + { + "epoch": 7.548102383053839, + "grad_norm": 0.6871273378755545, + "learning_rate": 1.487808944720253e-06, + "loss": 0.0056, + "step": 2138 + }, + { + "epoch": 7.551632833186231, + "grad_norm": 0.6863467342256887, + "learning_rate": 1.4837383641702662e-06, + "loss": 0.0051, + "step": 2139 + }, + { + "epoch": 7.555163283318623, + "grad_norm": 0.35607121306724576, + "learning_rate": 1.4796723893259712e-06, + "loss": 0.0039, + "step": 2140 + }, + { + "epoch": 7.558693733451015, + "grad_norm": 0.44120292398141187, + "learning_rate": 1.4756110255131063e-06, + "loss": 0.0045, + "step": 2141 + }, + { + "epoch": 7.562224183583407, + "grad_norm": 0.4252222691913999, + "learning_rate": 1.4715542780513647e-06, + "loss": 0.0024, + "step": 2142 + }, + { + "epoch": 7.565754633715799, + "grad_norm": 0.23716645002838307, + "learning_rate": 1.4675021522543997e-06, + "loss": 0.003, + "step": 2143 + }, + { + "epoch": 7.56928508384819, + "grad_norm": 0.4026023987166713, + "learning_rate": 1.463454653429806e-06, + "loss": 0.0042, + "step": 2144 + }, + { + "epoch": 7.572815533980583, + "grad_norm": 0.6274869013188908, + "learning_rate": 1.4594117868791208e-06, + "loss": 0.0073, + "step": 2145 + }, + { + "epoch": 7.576345984112974, + "grad_norm": 0.5748290272146682, + "learning_rate": 1.455373557897814e-06, + "loss": 0.0081, + "step": 2146 + }, + { + "epoch": 7.579876434245366, + "grad_norm": 0.26146659454989724, + "learning_rate": 1.4513399717752774e-06, + "loss": 0.0027, + "step": 2147 + }, + { + "epoch": 7.583406884377758, + "grad_norm": 0.428745887656463, + "learning_rate": 1.4473110337948248e-06, + "loss": 0.0063, + "step": 2148 + }, + { + "epoch": 7.58693733451015, + "grad_norm": 0.3399658619315668, + "learning_rate": 1.4432867492336817e-06, + "loss": 0.0033, + "step": 2149 + }, + { + "epoch": 7.5904677846425415, + "grad_norm": 0.3176167058612785, + "learning_rate": 1.4392671233629784e-06, + "loss": 0.0027, + "step": 2150 + }, + { + "epoch": 7.593998234774934, + "grad_norm": 0.4076903040549753, + "learning_rate": 1.43525216144774e-06, + "loss": 0.0042, + "step": 2151 + }, + { + "epoch": 7.597528684907326, + "grad_norm": 0.6432264802314678, + "learning_rate": 1.4312418687468855e-06, + "loss": 0.0034, + "step": 2152 + }, + { + "epoch": 7.601059135039717, + "grad_norm": 0.6728612199857228, + "learning_rate": 1.4272362505132175e-06, + "loss": 0.0063, + "step": 2153 + }, + { + "epoch": 7.60458958517211, + "grad_norm": 0.7319672398008487, + "learning_rate": 1.4232353119934179e-06, + "loss": 0.006, + "step": 2154 + }, + { + "epoch": 7.608120035304501, + "grad_norm": 0.5227354963234135, + "learning_rate": 1.4192390584280347e-06, + "loss": 0.0054, + "step": 2155 + }, + { + "epoch": 7.611650485436893, + "grad_norm": 0.6988676685777467, + "learning_rate": 1.4152474950514806e-06, + "loss": 0.0056, + "step": 2156 + }, + { + "epoch": 7.615180935569285, + "grad_norm": 0.6885323373533663, + "learning_rate": 1.4112606270920265e-06, + "loss": 0.0087, + "step": 2157 + }, + { + "epoch": 7.618711385701677, + "grad_norm": 0.5019879241969453, + "learning_rate": 1.4072784597717937e-06, + "loss": 0.0042, + "step": 2158 + }, + { + "epoch": 7.622241835834069, + "grad_norm": 0.27883010329033725, + "learning_rate": 1.4033009983067454e-06, + "loss": 0.0024, + "step": 2159 + }, + { + "epoch": 7.625772285966461, + "grad_norm": 1.0856731046071662, + "learning_rate": 1.3993282479066782e-06, + "loss": 0.0118, + "step": 2160 + }, + { + "epoch": 7.629302736098852, + "grad_norm": 0.47802937491323577, + "learning_rate": 1.3953602137752232e-06, + "loss": 0.0052, + "step": 2161 + }, + { + "epoch": 7.632833186231244, + "grad_norm": 0.5363068516085663, + "learning_rate": 1.3913969011098295e-06, + "loss": 0.005, + "step": 2162 + }, + { + "epoch": 7.636363636363637, + "grad_norm": 0.37760486500333973, + "learning_rate": 1.3874383151017662e-06, + "loss": 0.0025, + "step": 2163 + }, + { + "epoch": 7.639894086496028, + "grad_norm": 0.21912672385632706, + "learning_rate": 1.3834844609361064e-06, + "loss": 0.0022, + "step": 2164 + }, + { + "epoch": 7.64342453662842, + "grad_norm": 1.047823724589565, + "learning_rate": 1.3795353437917281e-06, + "loss": 0.0146, + "step": 2165 + }, + { + "epoch": 7.646954986760812, + "grad_norm": 0.3417774574548769, + "learning_rate": 1.375590968841306e-06, + "loss": 0.0051, + "step": 2166 + }, + { + "epoch": 7.650485436893204, + "grad_norm": 0.4659854663225553, + "learning_rate": 1.371651341251299e-06, + "loss": 0.0056, + "step": 2167 + }, + { + "epoch": 7.654015887025595, + "grad_norm": 0.33725498286038147, + "learning_rate": 1.3677164661819508e-06, + "loss": 0.0056, + "step": 2168 + }, + { + "epoch": 7.657546337157988, + "grad_norm": 0.383037161943809, + "learning_rate": 1.3637863487872826e-06, + "loss": 0.0039, + "step": 2169 + }, + { + "epoch": 7.66107678729038, + "grad_norm": 0.36196553579727175, + "learning_rate": 1.3598609942150765e-06, + "loss": 0.0025, + "step": 2170 + }, + { + "epoch": 7.664607237422771, + "grad_norm": 0.4077063382297464, + "learning_rate": 1.3559404076068844e-06, + "loss": 0.0047, + "step": 2171 + }, + { + "epoch": 7.668137687555163, + "grad_norm": 0.574699551622731, + "learning_rate": 1.3520245940980053e-06, + "loss": 0.0065, + "step": 2172 + }, + { + "epoch": 7.671668137687555, + "grad_norm": 0.5124968672820271, + "learning_rate": 1.3481135588174926e-06, + "loss": 0.0034, + "step": 2173 + }, + { + "epoch": 7.675198587819947, + "grad_norm": 0.4108603142586646, + "learning_rate": 1.3442073068881374e-06, + "loss": 0.0049, + "step": 2174 + }, + { + "epoch": 7.678729037952339, + "grad_norm": 0.5699362175042854, + "learning_rate": 1.3403058434264682e-06, + "loss": 0.0065, + "step": 2175 + }, + { + "epoch": 7.682259488084731, + "grad_norm": 0.6637554996774887, + "learning_rate": 1.3364091735427365e-06, + "loss": 0.004, + "step": 2176 + }, + { + "epoch": 7.685789938217122, + "grad_norm": 0.4638667292662007, + "learning_rate": 1.3325173023409204e-06, + "loss": 0.0078, + "step": 2177 + }, + { + "epoch": 7.689320388349515, + "grad_norm": 0.39021247327811986, + "learning_rate": 1.3286302349187098e-06, + "loss": 0.0059, + "step": 2178 + }, + { + "epoch": 7.692850838481906, + "grad_norm": 0.4360040685046432, + "learning_rate": 1.3247479763675043e-06, + "loss": 0.0064, + "step": 2179 + }, + { + "epoch": 7.696381288614298, + "grad_norm": 0.3448874736236507, + "learning_rate": 1.3208705317724007e-06, + "loss": 0.0037, + "step": 2180 + }, + { + "epoch": 7.699911738746691, + "grad_norm": 0.7225133053075434, + "learning_rate": 1.3169979062121952e-06, + "loss": 0.0101, + "step": 2181 + }, + { + "epoch": 7.703442188879082, + "grad_norm": 0.45324937453542613, + "learning_rate": 1.3131301047593685e-06, + "loss": 0.0055, + "step": 2182 + }, + { + "epoch": 7.706972639011474, + "grad_norm": 0.34433248264962, + "learning_rate": 1.3092671324800876e-06, + "loss": 0.0048, + "step": 2183 + }, + { + "epoch": 7.710503089143866, + "grad_norm": 0.16910912788703633, + "learning_rate": 1.305408994434184e-06, + "loss": 0.0025, + "step": 2184 + }, + { + "epoch": 7.714033539276258, + "grad_norm": 0.6404737290600166, + "learning_rate": 1.3015556956751669e-06, + "loss": 0.0062, + "step": 2185 + }, + { + "epoch": 7.717563989408649, + "grad_norm": 0.4836778682507138, + "learning_rate": 1.2977072412502023e-06, + "loss": 0.0034, + "step": 2186 + }, + { + "epoch": 7.721094439541042, + "grad_norm": 0.3890997906022876, + "learning_rate": 1.2938636362001145e-06, + "loss": 0.0035, + "step": 2187 + }, + { + "epoch": 7.724624889673433, + "grad_norm": 0.7178693749882828, + "learning_rate": 1.29002488555937e-06, + "loss": 0.0045, + "step": 2188 + }, + { + "epoch": 7.728155339805825, + "grad_norm": 1.0996159818911244, + "learning_rate": 1.2861909943560825e-06, + "loss": 0.01, + "step": 2189 + }, + { + "epoch": 7.731685789938217, + "grad_norm": 0.7260175454647014, + "learning_rate": 1.2823619676119974e-06, + "loss": 0.0063, + "step": 2190 + }, + { + "epoch": 7.735216240070609, + "grad_norm": 0.5964615712362218, + "learning_rate": 1.2785378103424917e-06, + "loss": 0.004, + "step": 2191 + }, + { + "epoch": 7.738746690203001, + "grad_norm": 0.4005460749305055, + "learning_rate": 1.2747185275565593e-06, + "loss": 0.0045, + "step": 2192 + }, + { + "epoch": 7.742277140335393, + "grad_norm": 0.34228208553244777, + "learning_rate": 1.270904124256813e-06, + "loss": 0.0015, + "step": 2193 + }, + { + "epoch": 7.745807590467785, + "grad_norm": 0.3952885528564364, + "learning_rate": 1.2670946054394745e-06, + "loss": 0.0047, + "step": 2194 + }, + { + "epoch": 7.749338040600176, + "grad_norm": 0.46135386595585526, + "learning_rate": 1.2632899760943674e-06, + "loss": 0.0055, + "step": 2195 + }, + { + "epoch": 7.752868490732569, + "grad_norm": 0.2982091213230059, + "learning_rate": 1.2594902412049064e-06, + "loss": 0.0037, + "step": 2196 + }, + { + "epoch": 7.75639894086496, + "grad_norm": 0.6119977711701021, + "learning_rate": 1.2556954057481019e-06, + "loss": 0.0066, + "step": 2197 + }, + { + "epoch": 7.759929390997352, + "grad_norm": 0.6071708991663604, + "learning_rate": 1.2519054746945452e-06, + "loss": 0.0052, + "step": 2198 + }, + { + "epoch": 7.763459841129744, + "grad_norm": 0.4223555564442321, + "learning_rate": 1.2481204530083985e-06, + "loss": 0.0032, + "step": 2199 + }, + { + "epoch": 7.766990291262136, + "grad_norm": 0.7016182830548846, + "learning_rate": 1.2443403456474017e-06, + "loss": 0.0069, + "step": 2200 + }, + { + "epoch": 7.770520741394527, + "grad_norm": 0.597751889933163, + "learning_rate": 1.2405651575628497e-06, + "loss": 0.0057, + "step": 2201 + }, + { + "epoch": 7.77405119152692, + "grad_norm": 0.30475088633674524, + "learning_rate": 1.2367948936995998e-06, + "loss": 0.0026, + "step": 2202 + }, + { + "epoch": 7.777581641659312, + "grad_norm": 0.2852173372403375, + "learning_rate": 1.2330295589960578e-06, + "loss": 0.0011, + "step": 2203 + }, + { + "epoch": 7.781112091791703, + "grad_norm": 0.48824360226886665, + "learning_rate": 1.2292691583841737e-06, + "loss": 0.0055, + "step": 2204 + }, + { + "epoch": 7.784642541924096, + "grad_norm": 0.45431460593598044, + "learning_rate": 1.2255136967894316e-06, + "loss": 0.0049, + "step": 2205 + }, + { + "epoch": 7.788172992056487, + "grad_norm": 0.27605778601797426, + "learning_rate": 1.2217631791308486e-06, + "loss": 0.0043, + "step": 2206 + }, + { + "epoch": 7.791703442188879, + "grad_norm": 0.4034269740034327, + "learning_rate": 1.218017610320969e-06, + "loss": 0.0054, + "step": 2207 + }, + { + "epoch": 7.795233892321271, + "grad_norm": 0.3799965537270784, + "learning_rate": 1.214276995265848e-06, + "loss": 0.0041, + "step": 2208 + }, + { + "epoch": 7.798764342453663, + "grad_norm": 0.39654749670958617, + "learning_rate": 1.2105413388650577e-06, + "loss": 0.004, + "step": 2209 + }, + { + "epoch": 7.802294792586054, + "grad_norm": 0.4763001516018584, + "learning_rate": 1.2068106460116746e-06, + "loss": 0.004, + "step": 2210 + }, + { + "epoch": 7.805825242718447, + "grad_norm": 0.6278092270982261, + "learning_rate": 1.2030849215922736e-06, + "loss": 0.005, + "step": 2211 + }, + { + "epoch": 7.809355692850838, + "grad_norm": 0.5054300697480051, + "learning_rate": 1.1993641704869196e-06, + "loss": 0.0039, + "step": 2212 + }, + { + "epoch": 7.81288614298323, + "grad_norm": 0.44107050867656955, + "learning_rate": 1.1956483975691635e-06, + "loss": 0.0069, + "step": 2213 + }, + { + "epoch": 7.816416593115623, + "grad_norm": 0.3448600276654447, + "learning_rate": 1.1919376077060385e-06, + "loss": 0.0027, + "step": 2214 + }, + { + "epoch": 7.819947043248014, + "grad_norm": 0.5068603530960161, + "learning_rate": 1.188231805758049e-06, + "loss": 0.0032, + "step": 2215 + }, + { + "epoch": 7.823477493380406, + "grad_norm": 0.6393213643114429, + "learning_rate": 1.1845309965791673e-06, + "loss": 0.0069, + "step": 2216 + }, + { + "epoch": 7.827007943512798, + "grad_norm": 0.6693695625274589, + "learning_rate": 1.1808351850168232e-06, + "loss": 0.0093, + "step": 2217 + }, + { + "epoch": 7.83053839364519, + "grad_norm": 0.21235573629959953, + "learning_rate": 1.1771443759119028e-06, + "loss": 0.0017, + "step": 2218 + }, + { + "epoch": 7.834068843777581, + "grad_norm": 0.5130911572311384, + "learning_rate": 1.1734585740987403e-06, + "loss": 0.0076, + "step": 2219 + }, + { + "epoch": 7.837599293909974, + "grad_norm": 0.737310584991482, + "learning_rate": 1.1697777844051105e-06, + "loss": 0.0077, + "step": 2220 + }, + { + "epoch": 7.841129744042365, + "grad_norm": 0.5200463629172021, + "learning_rate": 1.166102011652221e-06, + "loss": 0.0052, + "step": 2221 + }, + { + "epoch": 7.844660194174757, + "grad_norm": 0.4023443717143966, + "learning_rate": 1.1624312606547111e-06, + "loss": 0.007, + "step": 2222 + }, + { + "epoch": 7.848190644307149, + "grad_norm": 0.3582518031253343, + "learning_rate": 1.1587655362206412e-06, + "loss": 0.0049, + "step": 2223 + }, + { + "epoch": 7.851721094439541, + "grad_norm": 0.7030847874333035, + "learning_rate": 1.1551048431514894e-06, + "loss": 0.0099, + "step": 2224 + }, + { + "epoch": 7.855251544571933, + "grad_norm": 0.4846525739512421, + "learning_rate": 1.151449186242139e-06, + "loss": 0.0068, + "step": 2225 + }, + { + "epoch": 7.858781994704325, + "grad_norm": 0.2978587917518878, + "learning_rate": 1.147798570280883e-06, + "loss": 0.0041, + "step": 2226 + }, + { + "epoch": 7.862312444836717, + "grad_norm": 0.6656052875640509, + "learning_rate": 1.1441530000494055e-06, + "loss": 0.0077, + "step": 2227 + }, + { + "epoch": 7.865842894969108, + "grad_norm": 0.3358418029208277, + "learning_rate": 1.1405124803227874e-06, + "loss": 0.0023, + "step": 2228 + }, + { + "epoch": 7.869373345101501, + "grad_norm": 1.550405268165845, + "learning_rate": 1.1368770158694887e-06, + "loss": 0.0081, + "step": 2229 + }, + { + "epoch": 7.872903795233892, + "grad_norm": 0.4328313475020204, + "learning_rate": 1.1332466114513512e-06, + "loss": 0.0064, + "step": 2230 + }, + { + "epoch": 7.876434245366284, + "grad_norm": 0.5936128907050019, + "learning_rate": 1.129621271823589e-06, + "loss": 0.0073, + "step": 2231 + }, + { + "epoch": 7.879964695498676, + "grad_norm": 0.6350464288432247, + "learning_rate": 1.1260010017347823e-06, + "loss": 0.0093, + "step": 2232 + }, + { + "epoch": 7.883495145631068, + "grad_norm": 0.4766023549450927, + "learning_rate": 1.1223858059268682e-06, + "loss": 0.0048, + "step": 2233 + }, + { + "epoch": 7.887025595763459, + "grad_norm": 0.5600411625794194, + "learning_rate": 1.1187756891351399e-06, + "loss": 0.0061, + "step": 2234 + }, + { + "epoch": 7.890556045895852, + "grad_norm": 0.21311059490334683, + "learning_rate": 1.115170656088238e-06, + "loss": 0.0029, + "step": 2235 + }, + { + "epoch": 7.894086496028244, + "grad_norm": 0.9066682139378052, + "learning_rate": 1.1115707115081448e-06, + "loss": 0.0119, + "step": 2236 + }, + { + "epoch": 7.897616946160635, + "grad_norm": 0.22247554339306094, + "learning_rate": 1.1079758601101737e-06, + "loss": 0.0018, + "step": 2237 + }, + { + "epoch": 7.901147396293028, + "grad_norm": 0.334551387516309, + "learning_rate": 1.104386106602971e-06, + "loss": 0.0048, + "step": 2238 + }, + { + "epoch": 7.904677846425419, + "grad_norm": 0.7515546717070604, + "learning_rate": 1.1008014556885044e-06, + "loss": 0.0097, + "step": 2239 + }, + { + "epoch": 7.908208296557811, + "grad_norm": 0.45664982979894914, + "learning_rate": 1.0972219120620593e-06, + "loss": 0.0052, + "step": 2240 + }, + { + "epoch": 7.911738746690203, + "grad_norm": 0.668223133779893, + "learning_rate": 1.0936474804122281e-06, + "loss": 0.007, + "step": 2241 + }, + { + "epoch": 7.915269196822595, + "grad_norm": 0.8632954919876714, + "learning_rate": 1.0900781654209085e-06, + "loss": 0.0052, + "step": 2242 + }, + { + "epoch": 7.918799646954986, + "grad_norm": 0.4054310273390406, + "learning_rate": 1.086513971763299e-06, + "loss": 0.0052, + "step": 2243 + }, + { + "epoch": 7.922330097087379, + "grad_norm": 0.5115428123159997, + "learning_rate": 1.0829549041078867e-06, + "loss": 0.0038, + "step": 2244 + }, + { + "epoch": 7.92586054721977, + "grad_norm": 0.4600465651325307, + "learning_rate": 1.0794009671164484e-06, + "loss": 0.0061, + "step": 2245 + }, + { + "epoch": 7.929390997352162, + "grad_norm": 0.14205725953666065, + "learning_rate": 1.0758521654440358e-06, + "loss": 0.001, + "step": 2246 + }, + { + "epoch": 7.932921447484555, + "grad_norm": 0.3238560285218814, + "learning_rate": 1.0723085037389762e-06, + "loss": 0.0038, + "step": 2247 + }, + { + "epoch": 7.936451897616946, + "grad_norm": 0.22288640416993805, + "learning_rate": 1.0687699866428669e-06, + "loss": 0.0034, + "step": 2248 + }, + { + "epoch": 7.939982347749338, + "grad_norm": 0.7063334574262585, + "learning_rate": 1.0652366187905621e-06, + "loss": 0.0062, + "step": 2249 + }, + { + "epoch": 7.94351279788173, + "grad_norm": 0.41010735113634367, + "learning_rate": 1.0617084048101755e-06, + "loss": 0.0072, + "step": 2250 + }, + { + "epoch": 7.947043248014122, + "grad_norm": 1.0195218032535276, + "learning_rate": 1.058185349323067e-06, + "loss": 0.0121, + "step": 2251 + }, + { + "epoch": 7.950573698146513, + "grad_norm": 0.7623529955094425, + "learning_rate": 1.054667456943843e-06, + "loss": 0.0095, + "step": 2252 + }, + { + "epoch": 7.954104148278906, + "grad_norm": 0.2941124621635523, + "learning_rate": 1.0511547322803434e-06, + "loss": 0.0031, + "step": 2253 + }, + { + "epoch": 7.957634598411297, + "grad_norm": 0.8566239355207853, + "learning_rate": 1.0476471799336424e-06, + "loss": 0.0082, + "step": 2254 + }, + { + "epoch": 7.961165048543689, + "grad_norm": 0.35380526618880326, + "learning_rate": 1.0441448044980357e-06, + "loss": 0.004, + "step": 2255 + }, + { + "epoch": 7.964695498676081, + "grad_norm": 0.29047648747887694, + "learning_rate": 1.040647610561042e-06, + "loss": 0.0029, + "step": 2256 + }, + { + "epoch": 7.968225948808473, + "grad_norm": 0.9772008701959095, + "learning_rate": 1.037155602703393e-06, + "loss": 0.0062, + "step": 2257 + }, + { + "epoch": 7.971756398940865, + "grad_norm": 0.6601213507960044, + "learning_rate": 1.0336687854990234e-06, + "loss": 0.0094, + "step": 2258 + }, + { + "epoch": 7.975286849073257, + "grad_norm": 0.5115797673863285, + "learning_rate": 1.0301871635150723e-06, + "loss": 0.0042, + "step": 2259 + }, + { + "epoch": 7.978817299205649, + "grad_norm": 0.5161718797549839, + "learning_rate": 1.0267107413118743e-06, + "loss": 0.0062, + "step": 2260 + }, + { + "epoch": 7.98234774933804, + "grad_norm": 0.4278430468557276, + "learning_rate": 1.0232395234429532e-06, + "loss": 0.004, + "step": 2261 + }, + { + "epoch": 7.985878199470433, + "grad_norm": 0.33574841251337917, + "learning_rate": 1.019773514455012e-06, + "loss": 0.0046, + "step": 2262 + }, + { + "epoch": 7.989408649602824, + "grad_norm": 0.39752839673829643, + "learning_rate": 1.0163127188879352e-06, + "loss": 0.0045, + "step": 2263 + }, + { + "epoch": 7.992939099735216, + "grad_norm": 0.26612656736349266, + "learning_rate": 1.0128571412747785e-06, + "loss": 0.003, + "step": 2264 + }, + { + "epoch": 7.996469549867609, + "grad_norm": 0.3198947899888937, + "learning_rate": 1.0094067861417616e-06, + "loss": 0.0038, + "step": 2265 + }, + { + "epoch": 8.0, + "grad_norm": 0.6763442189183823, + "learning_rate": 1.005961658008262e-06, + "loss": 0.0045, + "step": 2266 + }, + { + "epoch": 8.003530450132391, + "grad_norm": 0.24853108750156552, + "learning_rate": 1.0025217613868143e-06, + "loss": 0.0026, + "step": 2267 + }, + { + "epoch": 8.007060900264785, + "grad_norm": 0.17200389416582687, + "learning_rate": 9.990871007830994e-07, + "loss": 0.0018, + "step": 2268 + }, + { + "epoch": 8.010591350397176, + "grad_norm": 0.1897515244258015, + "learning_rate": 9.956576806959389e-07, + "loss": 0.0024, + "step": 2269 + }, + { + "epoch": 8.014121800529567, + "grad_norm": 0.2281769324721284, + "learning_rate": 9.922335056172893e-07, + "loss": 0.0031, + "step": 2270 + }, + { + "epoch": 8.017652250661959, + "grad_norm": 0.18364419478391805, + "learning_rate": 9.888145800322397e-07, + "loss": 0.0021, + "step": 2271 + }, + { + "epoch": 8.021182700794352, + "grad_norm": 0.19134408561758534, + "learning_rate": 9.85400908419002e-07, + "loss": 0.0025, + "step": 2272 + }, + { + "epoch": 8.024713150926743, + "grad_norm": 0.15884716244851232, + "learning_rate": 9.81992495248908e-07, + "loss": 0.0013, + "step": 2273 + }, + { + "epoch": 8.028243601059135, + "grad_norm": 0.19256570981681082, + "learning_rate": 9.785893449863975e-07, + "loss": 0.001, + "step": 2274 + }, + { + "epoch": 8.031774051191526, + "grad_norm": 0.24642592453759304, + "learning_rate": 9.751914620890208e-07, + "loss": 0.0012, + "step": 2275 + }, + { + "epoch": 8.03530450132392, + "grad_norm": 0.17604961079517817, + "learning_rate": 9.717988510074271e-07, + "loss": 0.0035, + "step": 2276 + }, + { + "epoch": 8.03883495145631, + "grad_norm": 1.105833444888186, + "learning_rate": 9.68411516185363e-07, + "loss": 0.0054, + "step": 2277 + }, + { + "epoch": 8.042365401588702, + "grad_norm": 0.40653507580446874, + "learning_rate": 9.650294620596589e-07, + "loss": 0.003, + "step": 2278 + }, + { + "epoch": 8.045895851721095, + "grad_norm": 0.1525161194281428, + "learning_rate": 9.616526930602326e-07, + "loss": 0.0023, + "step": 2279 + }, + { + "epoch": 8.049426301853487, + "grad_norm": 0.24560582658999638, + "learning_rate": 9.582812136100784e-07, + "loss": 0.0031, + "step": 2280 + }, + { + "epoch": 8.052956751985878, + "grad_norm": 0.27548878336422705, + "learning_rate": 9.549150281252633e-07, + "loss": 0.0029, + "step": 2281 + }, + { + "epoch": 8.05648720211827, + "grad_norm": 0.2592679721816901, + "learning_rate": 9.515541410149159e-07, + "loss": 0.0031, + "step": 2282 + }, + { + "epoch": 8.060017652250663, + "grad_norm": 0.1481045997777656, + "learning_rate": 9.481985566812302e-07, + "loss": 0.0011, + "step": 2283 + }, + { + "epoch": 8.063548102383054, + "grad_norm": 0.36880049997230363, + "learning_rate": 9.448482795194491e-07, + "loss": 0.0028, + "step": 2284 + }, + { + "epoch": 8.067078552515445, + "grad_norm": 0.2699697690321897, + "learning_rate": 9.415033139178697e-07, + "loss": 0.0029, + "step": 2285 + }, + { + "epoch": 8.070609002647837, + "grad_norm": 0.1421409375029426, + "learning_rate": 9.381636642578268e-07, + "loss": 0.0014, + "step": 2286 + }, + { + "epoch": 8.07413945278023, + "grad_norm": 0.14383677190866428, + "learning_rate": 9.34829334913695e-07, + "loss": 0.0014, + "step": 2287 + }, + { + "epoch": 8.077669902912621, + "grad_norm": 0.30276201118616464, + "learning_rate": 9.315003302528797e-07, + "loss": 0.0042, + "step": 2288 + }, + { + "epoch": 8.081200353045013, + "grad_norm": 0.19366371201974647, + "learning_rate": 9.281766546358134e-07, + "loss": 0.0019, + "step": 2289 + }, + { + "epoch": 8.084730803177406, + "grad_norm": 0.4388683409994427, + "learning_rate": 9.248583124159438e-07, + "loss": 0.005, + "step": 2290 + }, + { + "epoch": 8.088261253309797, + "grad_norm": 0.8183042455275201, + "learning_rate": 9.215453079397374e-07, + "loss": 0.0049, + "step": 2291 + }, + { + "epoch": 8.091791703442189, + "grad_norm": 0.1545893653095532, + "learning_rate": 9.182376455466674e-07, + "loss": 0.0012, + "step": 2292 + }, + { + "epoch": 8.09532215357458, + "grad_norm": 0.23980521960414977, + "learning_rate": 9.14935329569211e-07, + "loss": 0.0029, + "step": 2293 + }, + { + "epoch": 8.098852603706973, + "grad_norm": 0.18907755804963491, + "learning_rate": 9.116383643328392e-07, + "loss": 0.0017, + "step": 2294 + }, + { + "epoch": 8.102383053839365, + "grad_norm": 0.21583759411843836, + "learning_rate": 9.083467541560176e-07, + "loss": 0.0021, + "step": 2295 + }, + { + "epoch": 8.105913503971756, + "grad_norm": 0.18173646658720455, + "learning_rate": 9.050605033501969e-07, + "loss": 0.0019, + "step": 2296 + }, + { + "epoch": 8.109443954104147, + "grad_norm": 0.16936974399321558, + "learning_rate": 9.017796162198089e-07, + "loss": 0.0019, + "step": 2297 + }, + { + "epoch": 8.11297440423654, + "grad_norm": 0.777370240719121, + "learning_rate": 8.985040970622571e-07, + "loss": 0.0057, + "step": 2298 + }, + { + "epoch": 8.116504854368932, + "grad_norm": 0.18192094375619758, + "learning_rate": 8.952339501679142e-07, + "loss": 0.0015, + "step": 2299 + }, + { + "epoch": 8.120035304501323, + "grad_norm": 0.280044134167334, + "learning_rate": 8.919691798201191e-07, + "loss": 0.0022, + "step": 2300 + }, + { + "epoch": 8.123565754633717, + "grad_norm": 0.23395525753566201, + "learning_rate": 8.887097902951658e-07, + "loss": 0.0022, + "step": 2301 + }, + { + "epoch": 8.127096204766108, + "grad_norm": 0.2180991995857826, + "learning_rate": 8.854557858623031e-07, + "loss": 0.0027, + "step": 2302 + }, + { + "epoch": 8.1306266548985, + "grad_norm": 0.21170986543988135, + "learning_rate": 8.822071707837215e-07, + "loss": 0.002, + "step": 2303 + }, + { + "epoch": 8.13415710503089, + "grad_norm": 0.2932034467804219, + "learning_rate": 8.789639493145569e-07, + "loss": 0.0031, + "step": 2304 + }, + { + "epoch": 8.137687555163284, + "grad_norm": 0.22396176835225987, + "learning_rate": 8.757261257028776e-07, + "loss": 0.002, + "step": 2305 + }, + { + "epoch": 8.141218005295675, + "grad_norm": 0.23458587328625635, + "learning_rate": 8.724937041896852e-07, + "loss": 0.0017, + "step": 2306 + }, + { + "epoch": 8.144748455428067, + "grad_norm": 0.401547520712631, + "learning_rate": 8.692666890088996e-07, + "loss": 0.0027, + "step": 2307 + }, + { + "epoch": 8.148278905560458, + "grad_norm": 0.22174445297787493, + "learning_rate": 8.660450843873647e-07, + "loss": 0.0018, + "step": 2308 + }, + { + "epoch": 8.151809355692851, + "grad_norm": 0.44049055526900305, + "learning_rate": 8.628288945448365e-07, + "loss": 0.0041, + "step": 2309 + }, + { + "epoch": 8.155339805825243, + "grad_norm": 0.2093298379558397, + "learning_rate": 8.596181236939755e-07, + "loss": 0.0018, + "step": 2310 + }, + { + "epoch": 8.158870255957634, + "grad_norm": 0.4883446265352031, + "learning_rate": 8.564127760403479e-07, + "loss": 0.004, + "step": 2311 + }, + { + "epoch": 8.162400706090027, + "grad_norm": 0.2810855843866889, + "learning_rate": 8.532128557824127e-07, + "loss": 0.0029, + "step": 2312 + }, + { + "epoch": 8.165931156222419, + "grad_norm": 0.1971374569847832, + "learning_rate": 8.500183671115242e-07, + "loss": 0.0019, + "step": 2313 + }, + { + "epoch": 8.16946160635481, + "grad_norm": 0.23859895291760927, + "learning_rate": 8.468293142119199e-07, + "loss": 0.0018, + "step": 2314 + }, + { + "epoch": 8.172992056487201, + "grad_norm": 0.23564465391912243, + "learning_rate": 8.436457012607169e-07, + "loss": 0.0018, + "step": 2315 + }, + { + "epoch": 8.176522506619595, + "grad_norm": 0.27081101994069606, + "learning_rate": 8.404675324279072e-07, + "loss": 0.0038, + "step": 2316 + }, + { + "epoch": 8.180052956751986, + "grad_norm": 0.22165918949813526, + "learning_rate": 8.372948118763536e-07, + "loss": 0.0016, + "step": 2317 + }, + { + "epoch": 8.183583406884377, + "grad_norm": 0.20811376779725055, + "learning_rate": 8.341275437617813e-07, + "loss": 0.0017, + "step": 2318 + }, + { + "epoch": 8.187113857016769, + "grad_norm": 0.11592579520247528, + "learning_rate": 8.309657322327724e-07, + "loss": 0.0008, + "step": 2319 + }, + { + "epoch": 8.190644307149162, + "grad_norm": 0.2175560339567269, + "learning_rate": 8.278093814307637e-07, + "loss": 0.0023, + "step": 2320 + }, + { + "epoch": 8.194174757281553, + "grad_norm": 0.14114434073988247, + "learning_rate": 8.246584954900389e-07, + "loss": 0.0015, + "step": 2321 + }, + { + "epoch": 8.197705207413945, + "grad_norm": 0.17453779503687972, + "learning_rate": 8.215130785377251e-07, + "loss": 0.0012, + "step": 2322 + }, + { + "epoch": 8.201235657546338, + "grad_norm": 0.14444287140217857, + "learning_rate": 8.183731346937817e-07, + "loss": 0.0011, + "step": 2323 + }, + { + "epoch": 8.20476610767873, + "grad_norm": 0.27054153191874764, + "learning_rate": 8.152386680710034e-07, + "loss": 0.0022, + "step": 2324 + }, + { + "epoch": 8.20829655781112, + "grad_norm": 0.06942043670079288, + "learning_rate": 8.121096827750091e-07, + "loss": 0.0006, + "step": 2325 + }, + { + "epoch": 8.211827007943512, + "grad_norm": 0.36264298465417066, + "learning_rate": 8.089861829042406e-07, + "loss": 0.0035, + "step": 2326 + }, + { + "epoch": 8.215357458075905, + "grad_norm": 0.25490063252042, + "learning_rate": 8.058681725499478e-07, + "loss": 0.0028, + "step": 2327 + }, + { + "epoch": 8.218887908208297, + "grad_norm": 0.4143528119088819, + "learning_rate": 8.027556557961968e-07, + "loss": 0.0038, + "step": 2328 + }, + { + "epoch": 8.222418358340688, + "grad_norm": 0.22252949180025564, + "learning_rate": 7.996486367198564e-07, + "loss": 0.0014, + "step": 2329 + }, + { + "epoch": 8.225948808473081, + "grad_norm": 0.19750707214480903, + "learning_rate": 7.965471193905955e-07, + "loss": 0.0017, + "step": 2330 + }, + { + "epoch": 8.229479258605473, + "grad_norm": 0.2073566143672113, + "learning_rate": 7.934511078708724e-07, + "loss": 0.0025, + "step": 2331 + }, + { + "epoch": 8.233009708737864, + "grad_norm": 0.19180475927876756, + "learning_rate": 7.903606062159375e-07, + "loss": 0.0018, + "step": 2332 + }, + { + "epoch": 8.236540158870255, + "grad_norm": 0.4187272830758832, + "learning_rate": 7.872756184738234e-07, + "loss": 0.0032, + "step": 2333 + }, + { + "epoch": 8.240070609002649, + "grad_norm": 0.28287301747844157, + "learning_rate": 7.841961486853417e-07, + "loss": 0.0023, + "step": 2334 + }, + { + "epoch": 8.24360105913504, + "grad_norm": 0.1800624958620469, + "learning_rate": 7.811222008840719e-07, + "loss": 0.0016, + "step": 2335 + }, + { + "epoch": 8.247131509267431, + "grad_norm": 0.11027925354188345, + "learning_rate": 7.780537790963655e-07, + "loss": 0.0008, + "step": 2336 + }, + { + "epoch": 8.250661959399823, + "grad_norm": 0.21049594148491443, + "learning_rate": 7.749908873413342e-07, + "loss": 0.0024, + "step": 2337 + }, + { + "epoch": 8.254192409532216, + "grad_norm": 0.7288569295606722, + "learning_rate": 7.719335296308467e-07, + "loss": 0.0041, + "step": 2338 + }, + { + "epoch": 8.257722859664607, + "grad_norm": 0.2873453010563978, + "learning_rate": 7.6888170996952e-07, + "loss": 0.0026, + "step": 2339 + }, + { + "epoch": 8.261253309796999, + "grad_norm": 0.2853943337684663, + "learning_rate": 7.658354323547224e-07, + "loss": 0.0019, + "step": 2340 + }, + { + "epoch": 8.264783759929392, + "grad_norm": 0.25517021008343965, + "learning_rate": 7.62794700776558e-07, + "loss": 0.0025, + "step": 2341 + }, + { + "epoch": 8.268314210061783, + "grad_norm": 0.18017315866584402, + "learning_rate": 7.597595192178702e-07, + "loss": 0.0014, + "step": 2342 + }, + { + "epoch": 8.271844660194175, + "grad_norm": 0.08411304944031515, + "learning_rate": 7.56729891654232e-07, + "loss": 0.0006, + "step": 2343 + }, + { + "epoch": 8.275375110326566, + "grad_norm": 0.15189911319547994, + "learning_rate": 7.537058220539395e-07, + "loss": 0.0011, + "step": 2344 + }, + { + "epoch": 8.27890556045896, + "grad_norm": 0.13527505469611611, + "learning_rate": 7.506873143780113e-07, + "loss": 0.0006, + "step": 2345 + }, + { + "epoch": 8.28243601059135, + "grad_norm": 0.45781661309914223, + "learning_rate": 7.476743725801805e-07, + "loss": 0.0018, + "step": 2346 + }, + { + "epoch": 8.285966460723742, + "grad_norm": 0.3218056478882174, + "learning_rate": 7.446670006068879e-07, + "loss": 0.0031, + "step": 2347 + }, + { + "epoch": 8.289496910856133, + "grad_norm": 0.20219515246828856, + "learning_rate": 7.416652023972809e-07, + "loss": 0.0016, + "step": 2348 + }, + { + "epoch": 8.293027360988527, + "grad_norm": 0.27395131714336235, + "learning_rate": 7.386689818832049e-07, + "loss": 0.0023, + "step": 2349 + }, + { + "epoch": 8.296557811120918, + "grad_norm": 0.23566533125440195, + "learning_rate": 7.356783429892022e-07, + "loss": 0.0024, + "step": 2350 + }, + { + "epoch": 8.30008826125331, + "grad_norm": 0.9177464183370133, + "learning_rate": 7.32693289632499e-07, + "loss": 0.0021, + "step": 2351 + }, + { + "epoch": 8.303618711385703, + "grad_norm": 0.30051188636071063, + "learning_rate": 7.29713825723009e-07, + "loss": 0.0022, + "step": 2352 + }, + { + "epoch": 8.307149161518094, + "grad_norm": 0.31110073349373807, + "learning_rate": 7.267399551633253e-07, + "loss": 0.0015, + "step": 2353 + }, + { + "epoch": 8.310679611650485, + "grad_norm": 0.45311627750966477, + "learning_rate": 7.237716818487139e-07, + "loss": 0.003, + "step": 2354 + }, + { + "epoch": 8.314210061782877, + "grad_norm": 0.3093508268511413, + "learning_rate": 7.208090096671072e-07, + "loss": 0.0024, + "step": 2355 + }, + { + "epoch": 8.31774051191527, + "grad_norm": 0.38314920555955406, + "learning_rate": 7.178519424991026e-07, + "loss": 0.0026, + "step": 2356 + }, + { + "epoch": 8.321270962047661, + "grad_norm": 0.12785596835387358, + "learning_rate": 7.149004842179563e-07, + "loss": 0.001, + "step": 2357 + }, + { + "epoch": 8.324801412180053, + "grad_norm": 0.5236784701825291, + "learning_rate": 7.119546386895776e-07, + "loss": 0.0039, + "step": 2358 + }, + { + "epoch": 8.328331862312444, + "grad_norm": 0.35786059648745555, + "learning_rate": 7.090144097725249e-07, + "loss": 0.0038, + "step": 2359 + }, + { + "epoch": 8.331862312444837, + "grad_norm": 0.1336306055172873, + "learning_rate": 7.060798013179976e-07, + "loss": 0.0008, + "step": 2360 + }, + { + "epoch": 8.335392762577229, + "grad_norm": 0.3877414458452562, + "learning_rate": 7.031508171698342e-07, + "loss": 0.0029, + "step": 2361 + }, + { + "epoch": 8.33892321270962, + "grad_norm": 0.18209368663759887, + "learning_rate": 7.002274611645083e-07, + "loss": 0.0014, + "step": 2362 + }, + { + "epoch": 8.342453662842013, + "grad_norm": 0.1136880060989946, + "learning_rate": 6.973097371311194e-07, + "loss": 0.0007, + "step": 2363 + }, + { + "epoch": 8.345984112974405, + "grad_norm": 0.04005695638704302, + "learning_rate": 6.943976488913889e-07, + "loss": 0.0003, + "step": 2364 + }, + { + "epoch": 8.349514563106796, + "grad_norm": 0.33955840453369496, + "learning_rate": 6.914912002596591e-07, + "loss": 0.0019, + "step": 2365 + }, + { + "epoch": 8.353045013239187, + "grad_norm": 0.29676504471641846, + "learning_rate": 6.885903950428841e-07, + "loss": 0.0018, + "step": 2366 + }, + { + "epoch": 8.35657546337158, + "grad_norm": 0.22984781111769104, + "learning_rate": 6.856952370406278e-07, + "loss": 0.0017, + "step": 2367 + }, + { + "epoch": 8.360105913503972, + "grad_norm": 0.16768141232216027, + "learning_rate": 6.828057300450536e-07, + "loss": 0.0009, + "step": 2368 + }, + { + "epoch": 8.363636363636363, + "grad_norm": 0.2685251386160339, + "learning_rate": 6.799218778409245e-07, + "loss": 0.0033, + "step": 2369 + }, + { + "epoch": 8.367166813768755, + "grad_norm": 0.14214657038530995, + "learning_rate": 6.770436842055988e-07, + "loss": 0.001, + "step": 2370 + }, + { + "epoch": 8.370697263901148, + "grad_norm": 0.23004510621778682, + "learning_rate": 6.741711529090212e-07, + "loss": 0.0013, + "step": 2371 + }, + { + "epoch": 8.37422771403354, + "grad_norm": 0.31858895501062834, + "learning_rate": 6.713042877137188e-07, + "loss": 0.0021, + "step": 2372 + }, + { + "epoch": 8.37775816416593, + "grad_norm": 0.30684383594925474, + "learning_rate": 6.684430923747992e-07, + "loss": 0.0023, + "step": 2373 + }, + { + "epoch": 8.381288614298324, + "grad_norm": 0.2696907901925904, + "learning_rate": 6.655875706399423e-07, + "loss": 0.0025, + "step": 2374 + }, + { + "epoch": 8.384819064430715, + "grad_norm": 0.27502425340423275, + "learning_rate": 6.627377262493978e-07, + "loss": 0.0019, + "step": 2375 + }, + { + "epoch": 8.388349514563107, + "grad_norm": 0.14956899613896565, + "learning_rate": 6.598935629359759e-07, + "loss": 0.0012, + "step": 2376 + }, + { + "epoch": 8.391879964695498, + "grad_norm": 0.18231538855848042, + "learning_rate": 6.570550844250484e-07, + "loss": 0.0016, + "step": 2377 + }, + { + "epoch": 8.395410414827891, + "grad_norm": 0.4393932704501488, + "learning_rate": 6.542222944345405e-07, + "loss": 0.0017, + "step": 2378 + }, + { + "epoch": 8.398940864960283, + "grad_norm": 0.14897851688953653, + "learning_rate": 6.513951966749271e-07, + "loss": 0.0012, + "step": 2379 + }, + { + "epoch": 8.402471315092674, + "grad_norm": 0.2599365209107397, + "learning_rate": 6.485737948492237e-07, + "loss": 0.0008, + "step": 2380 + }, + { + "epoch": 8.406001765225065, + "grad_norm": 0.14672100354107334, + "learning_rate": 6.457580926529894e-07, + "loss": 0.0008, + "step": 2381 + }, + { + "epoch": 8.409532215357459, + "grad_norm": 0.21461742547697923, + "learning_rate": 6.429480937743165e-07, + "loss": 0.0019, + "step": 2382 + }, + { + "epoch": 8.41306266548985, + "grad_norm": 0.31839675313824506, + "learning_rate": 6.401438018938249e-07, + "loss": 0.0036, + "step": 2383 + }, + { + "epoch": 8.416593115622241, + "grad_norm": 0.25237858441685496, + "learning_rate": 6.37345220684662e-07, + "loss": 0.0018, + "step": 2384 + }, + { + "epoch": 8.420123565754634, + "grad_norm": 0.15649301491939496, + "learning_rate": 6.345523538124926e-07, + "loss": 0.0016, + "step": 2385 + }, + { + "epoch": 8.423654015887026, + "grad_norm": 0.2495658653557097, + "learning_rate": 6.317652049354984e-07, + "loss": 0.0023, + "step": 2386 + }, + { + "epoch": 8.427184466019417, + "grad_norm": 0.20252983558961288, + "learning_rate": 6.289837777043734e-07, + "loss": 0.0015, + "step": 2387 + }, + { + "epoch": 8.430714916151809, + "grad_norm": 0.25409299694821064, + "learning_rate": 6.262080757623118e-07, + "loss": 0.0018, + "step": 2388 + }, + { + "epoch": 8.434245366284202, + "grad_norm": 0.2295421164717151, + "learning_rate": 6.234381027450132e-07, + "loss": 0.0011, + "step": 2389 + }, + { + "epoch": 8.437775816416593, + "grad_norm": 0.23361649059711473, + "learning_rate": 6.206738622806719e-07, + "loss": 0.0029, + "step": 2390 + }, + { + "epoch": 8.441306266548985, + "grad_norm": 0.2740569156607416, + "learning_rate": 6.179153579899749e-07, + "loss": 0.0023, + "step": 2391 + }, + { + "epoch": 8.444836716681376, + "grad_norm": 0.09732231752577755, + "learning_rate": 6.151625934860922e-07, + "loss": 0.0005, + "step": 2392 + }, + { + "epoch": 8.44836716681377, + "grad_norm": 0.43313548329096185, + "learning_rate": 6.124155723746794e-07, + "loss": 0.003, + "step": 2393 + }, + { + "epoch": 8.45189761694616, + "grad_norm": 0.3638789288437288, + "learning_rate": 6.096742982538667e-07, + "loss": 0.0034, + "step": 2394 + }, + { + "epoch": 8.455428067078552, + "grad_norm": 0.2402052814993731, + "learning_rate": 6.069387747142591e-07, + "loss": 0.0016, + "step": 2395 + }, + { + "epoch": 8.458958517210945, + "grad_norm": 0.20801212104488306, + "learning_rate": 6.04209005338926e-07, + "loss": 0.0018, + "step": 2396 + }, + { + "epoch": 8.462488967343337, + "grad_norm": 0.20512007176287805, + "learning_rate": 6.014849937034045e-07, + "loss": 0.0012, + "step": 2397 + }, + { + "epoch": 8.466019417475728, + "grad_norm": 0.40763741883755383, + "learning_rate": 5.987667433756844e-07, + "loss": 0.0011, + "step": 2398 + }, + { + "epoch": 8.46954986760812, + "grad_norm": 0.3547272114198054, + "learning_rate": 5.960542579162132e-07, + "loss": 0.0026, + "step": 2399 + }, + { + "epoch": 8.473080317740513, + "grad_norm": 0.4830481921873673, + "learning_rate": 5.933475408778872e-07, + "loss": 0.0048, + "step": 2400 + }, + { + "epoch": 8.476610767872904, + "grad_norm": 0.2873227829873547, + "learning_rate": 5.906465958060448e-07, + "loss": 0.0025, + "step": 2401 + }, + { + "epoch": 8.480141218005295, + "grad_norm": 0.10973647788511005, + "learning_rate": 5.879514262384661e-07, + "loss": 0.0008, + "step": 2402 + }, + { + "epoch": 8.483671668137688, + "grad_norm": 0.18000349407295368, + "learning_rate": 5.852620357053651e-07, + "loss": 0.0008, + "step": 2403 + }, + { + "epoch": 8.48720211827008, + "grad_norm": 0.15500654179918083, + "learning_rate": 5.825784277293884e-07, + "loss": 0.0012, + "step": 2404 + }, + { + "epoch": 8.490732568402471, + "grad_norm": 0.2394470267743939, + "learning_rate": 5.799006058256046e-07, + "loss": 0.0018, + "step": 2405 + }, + { + "epoch": 8.494263018534863, + "grad_norm": 0.17671509493041077, + "learning_rate": 5.772285735015066e-07, + "loss": 0.0015, + "step": 2406 + }, + { + "epoch": 8.497793468667256, + "grad_norm": 0.41100759288989797, + "learning_rate": 5.745623342570039e-07, + "loss": 0.0027, + "step": 2407 + }, + { + "epoch": 8.501323918799647, + "grad_norm": 0.060079873170144356, + "learning_rate": 5.719018915844154e-07, + "loss": 0.0004, + "step": 2408 + }, + { + "epoch": 8.504854368932039, + "grad_norm": 0.19455753898368652, + "learning_rate": 5.692472489684703e-07, + "loss": 0.0014, + "step": 2409 + }, + { + "epoch": 8.50838481906443, + "grad_norm": 0.05026330756607689, + "learning_rate": 5.665984098862992e-07, + "loss": 0.0004, + "step": 2410 + }, + { + "epoch": 8.511915269196823, + "grad_norm": 0.17755702537645754, + "learning_rate": 5.639553778074336e-07, + "loss": 0.0014, + "step": 2411 + }, + { + "epoch": 8.515445719329215, + "grad_norm": 0.28260547944732306, + "learning_rate": 5.613181561937947e-07, + "loss": 0.0022, + "step": 2412 + }, + { + "epoch": 8.518976169461606, + "grad_norm": 0.21261530279348329, + "learning_rate": 5.586867484996944e-07, + "loss": 0.0017, + "step": 2413 + }, + { + "epoch": 8.522506619593997, + "grad_norm": 0.6563515502551887, + "learning_rate": 5.560611581718311e-07, + "loss": 0.007, + "step": 2414 + }, + { + "epoch": 8.52603706972639, + "grad_norm": 0.2308512537904701, + "learning_rate": 5.534413886492823e-07, + "loss": 0.001, + "step": 2415 + }, + { + "epoch": 8.529567519858782, + "grad_norm": 0.626218369316658, + "learning_rate": 5.508274433635019e-07, + "loss": 0.005, + "step": 2416 + }, + { + "epoch": 8.533097969991173, + "grad_norm": 0.1854736935419076, + "learning_rate": 5.482193257383128e-07, + "loss": 0.0013, + "step": 2417 + }, + { + "epoch": 8.536628420123566, + "grad_norm": 0.16984744885809158, + "learning_rate": 5.456170391899068e-07, + "loss": 0.0012, + "step": 2418 + }, + { + "epoch": 8.540158870255958, + "grad_norm": 0.4940731175456218, + "learning_rate": 5.43020587126838e-07, + "loss": 0.0022, + "step": 2419 + }, + { + "epoch": 8.54368932038835, + "grad_norm": 0.13632806172136128, + "learning_rate": 5.404299729500184e-07, + "loss": 0.0004, + "step": 2420 + }, + { + "epoch": 8.54721977052074, + "grad_norm": 0.17305312780894933, + "learning_rate": 5.378452000527107e-07, + "loss": 0.0014, + "step": 2421 + }, + { + "epoch": 8.550750220653134, + "grad_norm": 0.2671602868697286, + "learning_rate": 5.352662718205292e-07, + "loss": 0.0024, + "step": 2422 + }, + { + "epoch": 8.554280670785525, + "grad_norm": 0.16916724955063736, + "learning_rate": 5.326931916314326e-07, + "loss": 0.0016, + "step": 2423 + }, + { + "epoch": 8.557811120917917, + "grad_norm": 0.15099968995205754, + "learning_rate": 5.301259628557192e-07, + "loss": 0.0006, + "step": 2424 + }, + { + "epoch": 8.56134157105031, + "grad_norm": 0.3518013810769276, + "learning_rate": 5.275645888560233e-07, + "loss": 0.0021, + "step": 2425 + }, + { + "epoch": 8.564872021182701, + "grad_norm": 0.6864935383454425, + "learning_rate": 5.250090729873069e-07, + "loss": 0.0054, + "step": 2426 + }, + { + "epoch": 8.568402471315093, + "grad_norm": 0.1938874094037543, + "learning_rate": 5.224594185968646e-07, + "loss": 0.001, + "step": 2427 + }, + { + "epoch": 8.571932921447484, + "grad_norm": 0.3081794038618721, + "learning_rate": 5.199156290243107e-07, + "loss": 0.0015, + "step": 2428 + }, + { + "epoch": 8.575463371579877, + "grad_norm": 0.5521797611481211, + "learning_rate": 5.173777076015763e-07, + "loss": 0.0049, + "step": 2429 + }, + { + "epoch": 8.578993821712269, + "grad_norm": 0.30884629365073196, + "learning_rate": 5.148456576529082e-07, + "loss": 0.0023, + "step": 2430 + }, + { + "epoch": 8.58252427184466, + "grad_norm": 0.26067110298752055, + "learning_rate": 5.123194824948613e-07, + "loss": 0.0021, + "step": 2431 + }, + { + "epoch": 8.586054721977051, + "grad_norm": 0.40402247681002496, + "learning_rate": 5.097991854362983e-07, + "loss": 0.003, + "step": 2432 + }, + { + "epoch": 8.589585172109444, + "grad_norm": 0.19562217257628803, + "learning_rate": 5.072847697783778e-07, + "loss": 0.0004, + "step": 2433 + }, + { + "epoch": 8.593115622241836, + "grad_norm": 0.11225372111822596, + "learning_rate": 5.047762388145582e-07, + "loss": 0.0009, + "step": 2434 + }, + { + "epoch": 8.596646072374227, + "grad_norm": 0.11691334713625078, + "learning_rate": 5.02273595830589e-07, + "loss": 0.0008, + "step": 2435 + }, + { + "epoch": 8.600176522506619, + "grad_norm": 0.25566650837718297, + "learning_rate": 4.997768441045098e-07, + "loss": 0.0013, + "step": 2436 + }, + { + "epoch": 8.603706972639012, + "grad_norm": 0.1232068969954091, + "learning_rate": 4.972859869066377e-07, + "loss": 0.0011, + "step": 2437 + }, + { + "epoch": 8.607237422771403, + "grad_norm": 0.28692360117061677, + "learning_rate": 4.948010274995746e-07, + "loss": 0.0021, + "step": 2438 + }, + { + "epoch": 8.610767872903795, + "grad_norm": 0.3113362261620494, + "learning_rate": 4.923219691381964e-07, + "loss": 0.0022, + "step": 2439 + }, + { + "epoch": 8.614298323036188, + "grad_norm": 0.1792388235271303, + "learning_rate": 4.898488150696468e-07, + "loss": 0.0015, + "step": 2440 + }, + { + "epoch": 8.61782877316858, + "grad_norm": 0.5603484288542537, + "learning_rate": 4.873815685333389e-07, + "loss": 0.0032, + "step": 2441 + }, + { + "epoch": 8.62135922330097, + "grad_norm": 0.39246357859823605, + "learning_rate": 4.849202327609453e-07, + "loss": 0.0026, + "step": 2442 + }, + { + "epoch": 8.624889673433362, + "grad_norm": 0.3563249501089053, + "learning_rate": 4.824648109763991e-07, + "loss": 0.0025, + "step": 2443 + }, + { + "epoch": 8.628420123565755, + "grad_norm": 0.2563368204395348, + "learning_rate": 4.800153063958851e-07, + "loss": 0.002, + "step": 2444 + }, + { + "epoch": 8.631950573698147, + "grad_norm": 0.24673440514859005, + "learning_rate": 4.775717222278403e-07, + "loss": 0.0016, + "step": 2445 + }, + { + "epoch": 8.635481023830538, + "grad_norm": 0.32484691610868777, + "learning_rate": 4.7513406167294275e-07, + "loss": 0.0022, + "step": 2446 + }, + { + "epoch": 8.639011473962931, + "grad_norm": 0.3688216212244623, + "learning_rate": 4.7270232792411497e-07, + "loss": 0.0026, + "step": 2447 + }, + { + "epoch": 8.642541924095323, + "grad_norm": 0.3625444273697154, + "learning_rate": 4.702765241665158e-07, + "loss": 0.0034, + "step": 2448 + }, + { + "epoch": 8.646072374227714, + "grad_norm": 0.11046606200273991, + "learning_rate": 4.6785665357753474e-07, + "loss": 0.0008, + "step": 2449 + }, + { + "epoch": 8.649602824360105, + "grad_norm": 0.4540780652455824, + "learning_rate": 4.6544271932679244e-07, + "loss": 0.001, + "step": 2450 + }, + { + "epoch": 8.653133274492498, + "grad_norm": 0.19679166889602248, + "learning_rate": 4.630347245761324e-07, + "loss": 0.0011, + "step": 2451 + }, + { + "epoch": 8.65666372462489, + "grad_norm": 0.4194582401831824, + "learning_rate": 4.606326724796195e-07, + "loss": 0.0036, + "step": 2452 + }, + { + "epoch": 8.660194174757281, + "grad_norm": 0.4265182858830472, + "learning_rate": 4.582365661835336e-07, + "loss": 0.0032, + "step": 2453 + }, + { + "epoch": 8.663724624889673, + "grad_norm": 0.5753905459123173, + "learning_rate": 4.558464088263681e-07, + "loss": 0.0034, + "step": 2454 + }, + { + "epoch": 8.667255075022066, + "grad_norm": 0.12291091482736478, + "learning_rate": 4.534622035388214e-07, + "loss": 0.0008, + "step": 2455 + }, + { + "epoch": 8.670785525154457, + "grad_norm": 0.439836604459671, + "learning_rate": 4.510839534437983e-07, + "loss": 0.0032, + "step": 2456 + }, + { + "epoch": 8.674315975286849, + "grad_norm": 0.25597173565024023, + "learning_rate": 4.487116616564047e-07, + "loss": 0.001, + "step": 2457 + }, + { + "epoch": 8.677846425419242, + "grad_norm": 0.14521289104979493, + "learning_rate": 4.4634533128393664e-07, + "loss": 0.0007, + "step": 2458 + }, + { + "epoch": 8.681376875551633, + "grad_norm": 0.7127912122776625, + "learning_rate": 4.439849654258871e-07, + "loss": 0.0045, + "step": 2459 + }, + { + "epoch": 8.684907325684025, + "grad_norm": 0.2037787172535473, + "learning_rate": 4.4163056717393404e-07, + "loss": 0.0015, + "step": 2460 + }, + { + "epoch": 8.688437775816416, + "grad_norm": 0.2659534559734492, + "learning_rate": 4.392821396119407e-07, + "loss": 0.0015, + "step": 2461 + }, + { + "epoch": 8.69196822594881, + "grad_norm": 0.42789915231332093, + "learning_rate": 4.369396858159458e-07, + "loss": 0.0038, + "step": 2462 + }, + { + "epoch": 8.6954986760812, + "grad_norm": 0.318580593112181, + "learning_rate": 4.3460320885416796e-07, + "loss": 0.0031, + "step": 2463 + }, + { + "epoch": 8.699029126213592, + "grad_norm": 0.1741653122983511, + "learning_rate": 4.322727117869951e-07, + "loss": 0.0011, + "step": 2464 + }, + { + "epoch": 8.702559576345983, + "grad_norm": 1.0501139370538595, + "learning_rate": 4.2994819766698337e-07, + "loss": 0.0096, + "step": 2465 + }, + { + "epoch": 8.706090026478376, + "grad_norm": 0.3038884816833782, + "learning_rate": 4.276296695388493e-07, + "loss": 0.0023, + "step": 2466 + }, + { + "epoch": 8.709620476610768, + "grad_norm": 0.2563080845963689, + "learning_rate": 4.253171304394732e-07, + "loss": 0.0023, + "step": 2467 + }, + { + "epoch": 8.71315092674316, + "grad_norm": 0.4940016065425068, + "learning_rate": 4.2301058339788814e-07, + "loss": 0.005, + "step": 2468 + }, + { + "epoch": 8.716681376875552, + "grad_norm": 0.25446229713631, + "learning_rate": 4.2071003143527913e-07, + "loss": 0.0031, + "step": 2469 + }, + { + "epoch": 8.720211827007944, + "grad_norm": 0.14118652374036755, + "learning_rate": 4.184154775649768e-07, + "loss": 0.001, + "step": 2470 + }, + { + "epoch": 8.723742277140335, + "grad_norm": 0.21902515183795912, + "learning_rate": 4.161269247924593e-07, + "loss": 0.0014, + "step": 2471 + }, + { + "epoch": 8.727272727272727, + "grad_norm": 0.2398517536313774, + "learning_rate": 4.13844376115341e-07, + "loss": 0.0017, + "step": 2472 + }, + { + "epoch": 8.73080317740512, + "grad_norm": 0.45979027049035526, + "learning_rate": 4.115678345233748e-07, + "loss": 0.0032, + "step": 2473 + }, + { + "epoch": 8.734333627537511, + "grad_norm": 0.11436710971446869, + "learning_rate": 4.092973029984415e-07, + "loss": 0.0009, + "step": 2474 + }, + { + "epoch": 8.737864077669903, + "grad_norm": 0.2590845682491535, + "learning_rate": 4.070327845145522e-07, + "loss": 0.0022, + "step": 2475 + }, + { + "epoch": 8.741394527802294, + "grad_norm": 0.06602584900768566, + "learning_rate": 4.0477428203784306e-07, + "loss": 0.0004, + "step": 2476 + }, + { + "epoch": 8.744924977934687, + "grad_norm": 0.3607144180394752, + "learning_rate": 4.0252179852656835e-07, + "loss": 0.0026, + "step": 2477 + }, + { + "epoch": 8.748455428067079, + "grad_norm": 0.24492064651146822, + "learning_rate": 4.002753369310974e-07, + "loss": 0.0022, + "step": 2478 + }, + { + "epoch": 8.75198587819947, + "grad_norm": 0.15167328843441252, + "learning_rate": 3.9803490019391545e-07, + "loss": 0.001, + "step": 2479 + }, + { + "epoch": 8.755516328331863, + "grad_norm": 0.23888829343683976, + "learning_rate": 3.9580049124961274e-07, + "loss": 0.0023, + "step": 2480 + }, + { + "epoch": 8.759046778464255, + "grad_norm": 0.25336285045063384, + "learning_rate": 3.935721130248876e-07, + "loss": 0.0019, + "step": 2481 + }, + { + "epoch": 8.762577228596646, + "grad_norm": 0.2005539461446993, + "learning_rate": 3.913497684385359e-07, + "loss": 0.0013, + "step": 2482 + }, + { + "epoch": 8.766107678729037, + "grad_norm": 0.6125834082862576, + "learning_rate": 3.8913346040145015e-07, + "loss": 0.0045, + "step": 2483 + }, + { + "epoch": 8.76963812886143, + "grad_norm": 0.10817954434765194, + "learning_rate": 3.869231918166194e-07, + "loss": 0.0007, + "step": 2484 + }, + { + "epoch": 8.773168578993822, + "grad_norm": 0.11456672734217362, + "learning_rate": 3.8471896557912005e-07, + "loss": 0.0006, + "step": 2485 + }, + { + "epoch": 8.776699029126213, + "grad_norm": 0.26144007596625013, + "learning_rate": 3.8252078457611417e-07, + "loss": 0.0019, + "step": 2486 + }, + { + "epoch": 8.780229479258605, + "grad_norm": 0.5292366344954661, + "learning_rate": 3.803286516868449e-07, + "loss": 0.0035, + "step": 2487 + }, + { + "epoch": 8.783759929390998, + "grad_norm": 0.16024835016764843, + "learning_rate": 3.7814256978263465e-07, + "loss": 0.0013, + "step": 2488 + }, + { + "epoch": 8.78729037952339, + "grad_norm": 0.3492141798964574, + "learning_rate": 3.759625417268803e-07, + "loss": 0.0038, + "step": 2489 + }, + { + "epoch": 8.79082082965578, + "grad_norm": 0.3399955273031793, + "learning_rate": 3.7378857037504715e-07, + "loss": 0.0019, + "step": 2490 + }, + { + "epoch": 8.794351279788174, + "grad_norm": 0.029484393145324004, + "learning_rate": 3.7162065857466844e-07, + "loss": 0.0002, + "step": 2491 + }, + { + "epoch": 8.797881729920565, + "grad_norm": 0.2381636956418484, + "learning_rate": 3.6945880916534114e-07, + "loss": 0.0018, + "step": 2492 + }, + { + "epoch": 8.801412180052957, + "grad_norm": 0.22859243192228584, + "learning_rate": 3.673030249787224e-07, + "loss": 0.0017, + "step": 2493 + }, + { + "epoch": 8.804942630185348, + "grad_norm": 0.2655994105813876, + "learning_rate": 3.6515330883852063e-07, + "loss": 0.002, + "step": 2494 + }, + { + "epoch": 8.808473080317741, + "grad_norm": 0.19567876081519323, + "learning_rate": 3.6300966356050014e-07, + "loss": 0.0011, + "step": 2495 + }, + { + "epoch": 8.812003530450133, + "grad_norm": 0.1491829293756405, + "learning_rate": 3.608720919524733e-07, + "loss": 0.001, + "step": 2496 + }, + { + "epoch": 8.815533980582524, + "grad_norm": 0.1256608383174644, + "learning_rate": 3.587405968142943e-07, + "loss": 0.001, + "step": 2497 + }, + { + "epoch": 8.819064430714917, + "grad_norm": 0.20400969641921504, + "learning_rate": 3.566151809378615e-07, + "loss": 0.0018, + "step": 2498 + }, + { + "epoch": 8.822594880847308, + "grad_norm": 0.4983992815721398, + "learning_rate": 3.544958471071069e-07, + "loss": 0.0038, + "step": 2499 + }, + { + "epoch": 8.8261253309797, + "grad_norm": 0.07548734766916114, + "learning_rate": 3.523825980979989e-07, + "loss": 0.0005, + "step": 2500 + }, + { + "epoch": 8.829655781112091, + "grad_norm": 0.18174336781201914, + "learning_rate": 3.502754366785349e-07, + "loss": 0.0012, + "step": 2501 + }, + { + "epoch": 8.833186231244484, + "grad_norm": 0.40131771932866533, + "learning_rate": 3.481743656087394e-07, + "loss": 0.0033, + "step": 2502 + }, + { + "epoch": 8.836716681376876, + "grad_norm": 0.2159531882628192, + "learning_rate": 3.4607938764065764e-07, + "loss": 0.0016, + "step": 2503 + }, + { + "epoch": 8.840247131509267, + "grad_norm": 0.09839009253189844, + "learning_rate": 3.4399050551835455e-07, + "loss": 0.0005, + "step": 2504 + }, + { + "epoch": 8.843777581641659, + "grad_norm": 0.3794404046050197, + "learning_rate": 3.419077219779127e-07, + "loss": 0.0027, + "step": 2505 + }, + { + "epoch": 8.847308031774052, + "grad_norm": 0.22115620612054773, + "learning_rate": 3.3983103974742415e-07, + "loss": 0.0019, + "step": 2506 + }, + { + "epoch": 8.850838481906443, + "grad_norm": 0.2231146721395509, + "learning_rate": 3.3776046154698927e-07, + "loss": 0.0018, + "step": 2507 + }, + { + "epoch": 8.854368932038835, + "grad_norm": 0.23788235186519993, + "learning_rate": 3.3569599008871466e-07, + "loss": 0.0011, + "step": 2508 + }, + { + "epoch": 8.857899382171226, + "grad_norm": 0.1898593086680488, + "learning_rate": 3.3363762807670787e-07, + "loss": 0.0011, + "step": 2509 + }, + { + "epoch": 8.86142983230362, + "grad_norm": 0.27356851283336764, + "learning_rate": 3.315853782070727e-07, + "loss": 0.0029, + "step": 2510 + }, + { + "epoch": 8.86496028243601, + "grad_norm": 0.17736664561120993, + "learning_rate": 3.2953924316790785e-07, + "loss": 0.0011, + "step": 2511 + }, + { + "epoch": 8.868490732568402, + "grad_norm": 0.1667508721395093, + "learning_rate": 3.2749922563930334e-07, + "loss": 0.0013, + "step": 2512 + }, + { + "epoch": 8.872021182700795, + "grad_norm": 0.09466649960243395, + "learning_rate": 3.2546532829333567e-07, + "loss": 0.0007, + "step": 2513 + }, + { + "epoch": 8.875551632833186, + "grad_norm": 0.6608491466600501, + "learning_rate": 3.234375537940654e-07, + "loss": 0.0035, + "step": 2514 + }, + { + "epoch": 8.879082082965578, + "grad_norm": 0.1280366080993897, + "learning_rate": 3.214159047975324e-07, + "loss": 0.001, + "step": 2515 + }, + { + "epoch": 8.88261253309797, + "grad_norm": 0.4325820538482789, + "learning_rate": 3.194003839517529e-07, + "loss": 0.0035, + "step": 2516 + }, + { + "epoch": 8.886142983230362, + "grad_norm": 0.2528792042444565, + "learning_rate": 3.173909938967179e-07, + "loss": 0.0022, + "step": 2517 + }, + { + "epoch": 8.889673433362754, + "grad_norm": 0.2994178626204928, + "learning_rate": 3.1538773726438764e-07, + "loss": 0.0021, + "step": 2518 + }, + { + "epoch": 8.893203883495145, + "grad_norm": 0.2434053943476249, + "learning_rate": 3.133906166786871e-07, + "loss": 0.0024, + "step": 2519 + }, + { + "epoch": 8.896734333627538, + "grad_norm": 0.4396481508649341, + "learning_rate": 3.1139963475550484e-07, + "loss": 0.0027, + "step": 2520 + }, + { + "epoch": 8.90026478375993, + "grad_norm": 0.1457481912294205, + "learning_rate": 3.094147941026904e-07, + "loss": 0.0012, + "step": 2521 + }, + { + "epoch": 8.903795233892321, + "grad_norm": 0.30718705853747585, + "learning_rate": 3.07436097320048e-07, + "loss": 0.0027, + "step": 2522 + }, + { + "epoch": 8.907325684024713, + "grad_norm": 0.2878212456808529, + "learning_rate": 3.054635469993328e-07, + "loss": 0.0036, + "step": 2523 + }, + { + "epoch": 8.910856134157106, + "grad_norm": 0.16137233079934285, + "learning_rate": 3.034971457242525e-07, + "loss": 0.001, + "step": 2524 + }, + { + "epoch": 8.914386584289497, + "grad_norm": 0.18805181394142312, + "learning_rate": 3.015368960704584e-07, + "loss": 0.0014, + "step": 2525 + }, + { + "epoch": 8.917917034421889, + "grad_norm": 0.354866416552788, + "learning_rate": 2.99582800605544e-07, + "loss": 0.0046, + "step": 2526 + }, + { + "epoch": 8.92144748455428, + "grad_norm": 0.5355491916697537, + "learning_rate": 2.97634861889044e-07, + "loss": 0.0035, + "step": 2527 + }, + { + "epoch": 8.924977934686673, + "grad_norm": 0.30515271570194213, + "learning_rate": 2.9569308247242525e-07, + "loss": 0.0027, + "step": 2528 + }, + { + "epoch": 8.928508384819065, + "grad_norm": 0.22882307158355958, + "learning_rate": 2.937574648990904e-07, + "loss": 0.0014, + "step": 2529 + }, + { + "epoch": 8.932038834951456, + "grad_norm": 0.18560781273265456, + "learning_rate": 2.918280117043709e-07, + "loss": 0.0012, + "step": 2530 + }, + { + "epoch": 8.935569285083847, + "grad_norm": 0.30011718943103793, + "learning_rate": 2.899047254155207e-07, + "loss": 0.0029, + "step": 2531 + }, + { + "epoch": 8.93909973521624, + "grad_norm": 0.14577420590031429, + "learning_rate": 2.879876085517191e-07, + "loss": 0.0007, + "step": 2532 + }, + { + "epoch": 8.942630185348632, + "grad_norm": 0.2852281753044023, + "learning_rate": 2.860766636240636e-07, + "loss": 0.0034, + "step": 2533 + }, + { + "epoch": 8.946160635481023, + "grad_norm": 0.20611415086034895, + "learning_rate": 2.841718931355686e-07, + "loss": 0.0014, + "step": 2534 + }, + { + "epoch": 8.949691085613416, + "grad_norm": 0.35452428403662417, + "learning_rate": 2.8227329958115836e-07, + "loss": 0.0036, + "step": 2535 + }, + { + "epoch": 8.953221535745808, + "grad_norm": 0.19686659087731687, + "learning_rate": 2.803808854476686e-07, + "loss": 0.0008, + "step": 2536 + }, + { + "epoch": 8.9567519858782, + "grad_norm": 0.5172226866251929, + "learning_rate": 2.784946532138411e-07, + "loss": 0.0034, + "step": 2537 + }, + { + "epoch": 8.96028243601059, + "grad_norm": 0.26044896841571946, + "learning_rate": 2.7661460535032045e-07, + "loss": 0.0028, + "step": 2538 + }, + { + "epoch": 8.963812886142984, + "grad_norm": 0.29709143183026526, + "learning_rate": 2.747407443196487e-07, + "loss": 0.0025, + "step": 2539 + }, + { + "epoch": 8.967343336275375, + "grad_norm": 0.3150579731243277, + "learning_rate": 2.7287307257626574e-07, + "loss": 0.0017, + "step": 2540 + }, + { + "epoch": 8.970873786407767, + "grad_norm": 0.23728679614037076, + "learning_rate": 2.710115925665041e-07, + "loss": 0.0019, + "step": 2541 + }, + { + "epoch": 8.97440423654016, + "grad_norm": 0.21456045787548475, + "learning_rate": 2.691563067285874e-07, + "loss": 0.0018, + "step": 2542 + }, + { + "epoch": 8.977934686672551, + "grad_norm": 0.17623670749499756, + "learning_rate": 2.673072174926261e-07, + "loss": 0.0019, + "step": 2543 + }, + { + "epoch": 8.981465136804943, + "grad_norm": 0.24092193766691616, + "learning_rate": 2.6546432728061066e-07, + "loss": 0.0017, + "step": 2544 + }, + { + "epoch": 8.984995586937334, + "grad_norm": 0.20944727969185992, + "learning_rate": 2.636276385064157e-07, + "loss": 0.0016, + "step": 2545 + }, + { + "epoch": 8.988526037069727, + "grad_norm": 0.3706785662272731, + "learning_rate": 2.6179715357579126e-07, + "loss": 0.0033, + "step": 2546 + }, + { + "epoch": 8.992056487202118, + "grad_norm": 0.05008407632325225, + "learning_rate": 2.599728748863628e-07, + "loss": 0.0004, + "step": 2547 + }, + { + "epoch": 8.99558693733451, + "grad_norm": 0.1314941961240597, + "learning_rate": 2.5815480482762443e-07, + "loss": 0.0014, + "step": 2548 + }, + { + "epoch": 8.999117387466901, + "grad_norm": 0.3335427878737382, + "learning_rate": 2.563429457809397e-07, + "loss": 0.0034, + "step": 2549 + }, + { + "epoch": 9.002647837599294, + "grad_norm": 0.12263520591979192, + "learning_rate": 2.545373001195367e-07, + "loss": 0.0009, + "step": 2550 + }, + { + "epoch": 9.006178287731686, + "grad_norm": 0.08474942118698686, + "learning_rate": 2.527378702085037e-07, + "loss": 0.0005, + "step": 2551 + }, + { + "epoch": 9.009708737864077, + "grad_norm": 0.165341681373007, + "learning_rate": 2.5094465840478863e-07, + "loss": 0.0013, + "step": 2552 + }, + { + "epoch": 9.01323918799647, + "grad_norm": 0.10483265735338233, + "learning_rate": 2.4915766705719547e-07, + "loss": 0.0005, + "step": 2553 + }, + { + "epoch": 9.016769638128862, + "grad_norm": 0.054891290050464885, + "learning_rate": 2.473768985063785e-07, + "loss": 0.0004, + "step": 2554 + }, + { + "epoch": 9.020300088261253, + "grad_norm": 0.18081175895307633, + "learning_rate": 2.456023550848424e-07, + "loss": 0.0017, + "step": 2555 + }, + { + "epoch": 9.023830538393645, + "grad_norm": 0.17351763440706353, + "learning_rate": 2.4383403911693747e-07, + "loss": 0.001, + "step": 2556 + }, + { + "epoch": 9.027360988526038, + "grad_norm": 0.1388335895765888, + "learning_rate": 2.4207195291885757e-07, + "loss": 0.0008, + "step": 2557 + }, + { + "epoch": 9.03089143865843, + "grad_norm": 0.20839028455006167, + "learning_rate": 2.403160987986358e-07, + "loss": 0.0017, + "step": 2558 + }, + { + "epoch": 9.03442188879082, + "grad_norm": 0.1238318447195371, + "learning_rate": 2.385664790561448e-07, + "loss": 0.0009, + "step": 2559 + }, + { + "epoch": 9.037952338923212, + "grad_norm": 0.2285911775877918, + "learning_rate": 2.368230959830875e-07, + "loss": 0.0017, + "step": 2560 + }, + { + "epoch": 9.041482789055605, + "grad_norm": 0.08826696284639503, + "learning_rate": 2.35085951863e-07, + "loss": 0.0006, + "step": 2561 + }, + { + "epoch": 9.045013239187996, + "grad_norm": 0.2112045083972482, + "learning_rate": 2.333550489712466e-07, + "loss": 0.001, + "step": 2562 + }, + { + "epoch": 9.048543689320388, + "grad_norm": 0.1567237221089486, + "learning_rate": 2.3163038957501716e-07, + "loss": 0.0008, + "step": 2563 + }, + { + "epoch": 9.052074139452781, + "grad_norm": 0.11034154866148924, + "learning_rate": 2.2991197593332016e-07, + "loss": 0.0008, + "step": 2564 + }, + { + "epoch": 9.055604589585172, + "grad_norm": 0.1018809030327687, + "learning_rate": 2.2819981029698734e-07, + "loss": 0.0007, + "step": 2565 + }, + { + "epoch": 9.059135039717564, + "grad_norm": 0.3088634803219091, + "learning_rate": 2.2649389490866481e-07, + "loss": 0.0024, + "step": 2566 + }, + { + "epoch": 9.062665489849955, + "grad_norm": 0.20995299384628607, + "learning_rate": 2.2479423200281346e-07, + "loss": 0.0015, + "step": 2567 + }, + { + "epoch": 9.066195939982348, + "grad_norm": 0.03092998355176428, + "learning_rate": 2.2310082380569975e-07, + "loss": 0.0002, + "step": 2568 + }, + { + "epoch": 9.06972639011474, + "grad_norm": 0.15610224724692195, + "learning_rate": 2.2141367253540325e-07, + "loss": 0.0013, + "step": 2569 + }, + { + "epoch": 9.073256840247131, + "grad_norm": 0.18013966093470357, + "learning_rate": 2.1973278040180458e-07, + "loss": 0.0012, + "step": 2570 + }, + { + "epoch": 9.076787290379523, + "grad_norm": 0.3263561298083216, + "learning_rate": 2.1805814960658866e-07, + "loss": 0.0019, + "step": 2571 + }, + { + "epoch": 9.080317740511916, + "grad_norm": 0.10682024140357836, + "learning_rate": 2.1638978234323538e-07, + "loss": 0.0006, + "step": 2572 + }, + { + "epoch": 9.083848190644307, + "grad_norm": 0.1361829082786385, + "learning_rate": 2.1472768079702388e-07, + "loss": 0.0008, + "step": 2573 + }, + { + "epoch": 9.087378640776699, + "grad_norm": 0.24818590623137238, + "learning_rate": 2.1307184714502437e-07, + "loss": 0.0014, + "step": 2574 + }, + { + "epoch": 9.090909090909092, + "grad_norm": 0.07406349601090866, + "learning_rate": 2.1142228355609862e-07, + "loss": 0.0005, + "step": 2575 + }, + { + "epoch": 9.094439541041483, + "grad_norm": 0.2369198531663275, + "learning_rate": 2.0977899219089382e-07, + "loss": 0.0018, + "step": 2576 + }, + { + "epoch": 9.097969991173875, + "grad_norm": 0.15096428323078476, + "learning_rate": 2.0814197520184321e-07, + "loss": 0.0012, + "step": 2577 + }, + { + "epoch": 9.101500441306266, + "grad_norm": 0.11588721952610355, + "learning_rate": 2.0651123473316103e-07, + "loss": 0.0008, + "step": 2578 + }, + { + "epoch": 9.105030891438659, + "grad_norm": 0.19704847606811446, + "learning_rate": 2.048867729208409e-07, + "loss": 0.0014, + "step": 2579 + }, + { + "epoch": 9.10856134157105, + "grad_norm": 0.1527465684401655, + "learning_rate": 2.032685918926508e-07, + "loss": 0.001, + "step": 2580 + }, + { + "epoch": 9.112091791703442, + "grad_norm": 0.16426721931514415, + "learning_rate": 2.0165669376813356e-07, + "loss": 0.0013, + "step": 2581 + }, + { + "epoch": 9.115622241835833, + "grad_norm": 0.24009655378824435, + "learning_rate": 2.0005108065860313e-07, + "loss": 0.0018, + "step": 2582 + }, + { + "epoch": 9.119152691968226, + "grad_norm": 0.146658406253024, + "learning_rate": 1.9845175466713828e-07, + "loss": 0.001, + "step": 2583 + }, + { + "epoch": 9.122683142100618, + "grad_norm": 0.04455476214117835, + "learning_rate": 1.9685871788858556e-07, + "loss": 0.0003, + "step": 2584 + }, + { + "epoch": 9.12621359223301, + "grad_norm": 0.21013401942731197, + "learning_rate": 1.9527197240955253e-07, + "loss": 0.0016, + "step": 2585 + }, + { + "epoch": 9.129744042365402, + "grad_norm": 0.16179394839889424, + "learning_rate": 1.9369152030840553e-07, + "loss": 0.0012, + "step": 2586 + }, + { + "epoch": 9.133274492497794, + "grad_norm": 0.163294241652609, + "learning_rate": 1.921173636552698e-07, + "loss": 0.001, + "step": 2587 + }, + { + "epoch": 9.136804942630185, + "grad_norm": 0.21967147094616427, + "learning_rate": 1.9054950451202315e-07, + "loss": 0.0021, + "step": 2588 + }, + { + "epoch": 9.140335392762577, + "grad_norm": 0.13891622772589185, + "learning_rate": 1.8898794493229344e-07, + "loss": 0.0008, + "step": 2589 + }, + { + "epoch": 9.14386584289497, + "grad_norm": 0.10153471557995475, + "learning_rate": 1.8743268696145956e-07, + "loss": 0.0006, + "step": 2590 + }, + { + "epoch": 9.147396293027361, + "grad_norm": 0.14722879420098153, + "learning_rate": 1.8588373263664582e-07, + "loss": 0.001, + "step": 2591 + }, + { + "epoch": 9.150926743159753, + "grad_norm": 0.22857190138480388, + "learning_rate": 1.8434108398671824e-07, + "loss": 0.0015, + "step": 2592 + }, + { + "epoch": 9.154457193292144, + "grad_norm": 0.10174674054618638, + "learning_rate": 1.8280474303228545e-07, + "loss": 0.0005, + "step": 2593 + }, + { + "epoch": 9.157987643424537, + "grad_norm": 0.17206402772206614, + "learning_rate": 1.8127471178569278e-07, + "loss": 0.0012, + "step": 2594 + }, + { + "epoch": 9.161518093556928, + "grad_norm": 0.20441922025218245, + "learning_rate": 1.797509922510221e-07, + "loss": 0.0016, + "step": 2595 + }, + { + "epoch": 9.16504854368932, + "grad_norm": 0.2502519472982651, + "learning_rate": 1.7823358642408694e-07, + "loss": 0.0026, + "step": 2596 + }, + { + "epoch": 9.168578993821713, + "grad_norm": 0.17350039246226323, + "learning_rate": 1.7672249629243022e-07, + "loss": 0.0011, + "step": 2597 + }, + { + "epoch": 9.172109443954104, + "grad_norm": 0.15951018975156395, + "learning_rate": 1.7521772383532476e-07, + "loss": 0.001, + "step": 2598 + }, + { + "epoch": 9.175639894086496, + "grad_norm": 0.13807797047056788, + "learning_rate": 1.7371927102376618e-07, + "loss": 0.0012, + "step": 2599 + }, + { + "epoch": 9.179170344218887, + "grad_norm": 0.20507369136346032, + "learning_rate": 1.722271398204739e-07, + "loss": 0.0019, + "step": 2600 + }, + { + "epoch": 9.18270079435128, + "grad_norm": 0.15538686925285705, + "learning_rate": 1.707413321798862e-07, + "loss": 0.0009, + "step": 2601 + }, + { + "epoch": 9.186231244483672, + "grad_norm": 0.1517699820968831, + "learning_rate": 1.69261850048158e-07, + "loss": 0.0011, + "step": 2602 + }, + { + "epoch": 9.189761694616063, + "grad_norm": 0.16074288654949698, + "learning_rate": 1.6778869536316077e-07, + "loss": 0.0011, + "step": 2603 + }, + { + "epoch": 9.193292144748455, + "grad_norm": 0.12152310832849222, + "learning_rate": 1.663218700544772e-07, + "loss": 0.0005, + "step": 2604 + }, + { + "epoch": 9.196822594880848, + "grad_norm": 0.0846705497796846, + "learning_rate": 1.6486137604339813e-07, + "loss": 0.0006, + "step": 2605 + }, + { + "epoch": 9.20035304501324, + "grad_norm": 0.2231400106271614, + "learning_rate": 1.634072152429239e-07, + "loss": 0.0014, + "step": 2606 + }, + { + "epoch": 9.20388349514563, + "grad_norm": 0.11635640345168247, + "learning_rate": 1.6195938955775702e-07, + "loss": 0.0007, + "step": 2607 + }, + { + "epoch": 9.207413945278024, + "grad_norm": 0.18969894099216586, + "learning_rate": 1.6051790088430495e-07, + "loss": 0.0015, + "step": 2608 + }, + { + "epoch": 9.210944395410415, + "grad_norm": 0.09574826585087719, + "learning_rate": 1.5908275111067184e-07, + "loss": 0.0006, + "step": 2609 + }, + { + "epoch": 9.214474845542806, + "grad_norm": 0.1323448213903271, + "learning_rate": 1.5765394211666062e-07, + "loss": 0.0008, + "step": 2610 + }, + { + "epoch": 9.218005295675198, + "grad_norm": 0.08969921516697855, + "learning_rate": 1.562314757737682e-07, + "loss": 0.0005, + "step": 2611 + }, + { + "epoch": 9.221535745807591, + "grad_norm": 0.27985051271931877, + "learning_rate": 1.5481535394518465e-07, + "loss": 0.002, + "step": 2612 + }, + { + "epoch": 9.225066195939982, + "grad_norm": 0.15963288109922574, + "learning_rate": 1.5340557848578798e-07, + "loss": 0.0012, + "step": 2613 + }, + { + "epoch": 9.228596646072374, + "grad_norm": 0.2315108911996512, + "learning_rate": 1.5200215124214491e-07, + "loss": 0.0017, + "step": 2614 + }, + { + "epoch": 9.232127096204765, + "grad_norm": 0.18641835178367888, + "learning_rate": 1.506050740525078e-07, + "loss": 0.0013, + "step": 2615 + }, + { + "epoch": 9.235657546337158, + "grad_norm": 0.136143645078784, + "learning_rate": 1.4921434874681007e-07, + "loss": 0.0006, + "step": 2616 + }, + { + "epoch": 9.23918799646955, + "grad_norm": 0.06951910821539543, + "learning_rate": 1.478299771466657e-07, + "loss": 0.0005, + "step": 2617 + }, + { + "epoch": 9.242718446601941, + "grad_norm": 0.14361092469595013, + "learning_rate": 1.4645196106536586e-07, + "loss": 0.0011, + "step": 2618 + }, + { + "epoch": 9.246248896734334, + "grad_norm": 0.0649042568508669, + "learning_rate": 1.450803023078784e-07, + "loss": 0.0006, + "step": 2619 + }, + { + "epoch": 9.249779346866726, + "grad_norm": 0.19230008566252968, + "learning_rate": 1.4371500267084337e-07, + "loss": 0.0012, + "step": 2620 + }, + { + "epoch": 9.253309796999117, + "grad_norm": 0.06351400051740833, + "learning_rate": 1.4235606394257085e-07, + "loss": 0.0004, + "step": 2621 + }, + { + "epoch": 9.256840247131509, + "grad_norm": 0.24510456433735373, + "learning_rate": 1.4100348790303975e-07, + "loss": 0.0015, + "step": 2622 + }, + { + "epoch": 9.260370697263902, + "grad_norm": 0.1787517621509372, + "learning_rate": 1.396572763238957e-07, + "loss": 0.0012, + "step": 2623 + }, + { + "epoch": 9.263901147396293, + "grad_norm": 0.14204535268662052, + "learning_rate": 1.383174309684482e-07, + "loss": 0.0008, + "step": 2624 + }, + { + "epoch": 9.267431597528685, + "grad_norm": 0.13827231376718335, + "learning_rate": 1.3698395359166505e-07, + "loss": 0.0009, + "step": 2625 + }, + { + "epoch": 9.270962047661076, + "grad_norm": 0.2364797884387283, + "learning_rate": 1.3565684594017637e-07, + "loss": 0.0017, + "step": 2626 + }, + { + "epoch": 9.274492497793469, + "grad_norm": 0.2187070789564688, + "learning_rate": 1.3433610975226718e-07, + "loss": 0.0019, + "step": 2627 + }, + { + "epoch": 9.27802294792586, + "grad_norm": 0.2955674598617109, + "learning_rate": 1.3302174675787872e-07, + "loss": 0.0017, + "step": 2628 + }, + { + "epoch": 9.281553398058252, + "grad_norm": 0.1407534377071998, + "learning_rate": 1.3171375867860215e-07, + "loss": 0.0008, + "step": 2629 + }, + { + "epoch": 9.285083848190645, + "grad_norm": 0.07644665951066944, + "learning_rate": 1.3041214722768036e-07, + "loss": 0.0005, + "step": 2630 + }, + { + "epoch": 9.288614298323036, + "grad_norm": 0.17587104816736154, + "learning_rate": 1.291169141100035e-07, + "loss": 0.0014, + "step": 2631 + }, + { + "epoch": 9.292144748455428, + "grad_norm": 0.1519242998729314, + "learning_rate": 1.278280610221072e-07, + "loss": 0.0011, + "step": 2632 + }, + { + "epoch": 9.29567519858782, + "grad_norm": 0.3159051363472011, + "learning_rate": 1.265455896521689e-07, + "loss": 0.0029, + "step": 2633 + }, + { + "epoch": 9.299205648720212, + "grad_norm": 0.17668803189246593, + "learning_rate": 1.252695016800093e-07, + "loss": 0.0014, + "step": 2634 + }, + { + "epoch": 9.302736098852604, + "grad_norm": 0.1302181138452725, + "learning_rate": 1.2399979877708746e-07, + "loss": 0.0013, + "step": 2635 + }, + { + "epoch": 9.306266548984995, + "grad_norm": 0.12413701558969119, + "learning_rate": 1.2273648260649806e-07, + "loss": 0.0007, + "step": 2636 + }, + { + "epoch": 9.309796999117388, + "grad_norm": 0.05520997869569197, + "learning_rate": 1.2147955482297025e-07, + "loss": 0.0003, + "step": 2637 + }, + { + "epoch": 9.31332744924978, + "grad_norm": 0.222597819253233, + "learning_rate": 1.2022901707286706e-07, + "loss": 0.0014, + "step": 2638 + }, + { + "epoch": 9.316857899382171, + "grad_norm": 0.1555765351980316, + "learning_rate": 1.1898487099417932e-07, + "loss": 0.0014, + "step": 2639 + }, + { + "epoch": 9.320388349514563, + "grad_norm": 0.14209939133905203, + "learning_rate": 1.1774711821652795e-07, + "loss": 0.001, + "step": 2640 + }, + { + "epoch": 9.323918799646956, + "grad_norm": 0.15873434776567705, + "learning_rate": 1.1651576036115942e-07, + "loss": 0.0008, + "step": 2641 + }, + { + "epoch": 9.327449249779347, + "grad_norm": 0.20975936495005168, + "learning_rate": 1.1529079904094188e-07, + "loss": 0.0015, + "step": 2642 + }, + { + "epoch": 9.330979699911738, + "grad_norm": 0.15062064745536857, + "learning_rate": 1.1407223586036797e-07, + "loss": 0.001, + "step": 2643 + }, + { + "epoch": 9.33451015004413, + "grad_norm": 0.1410179207771357, + "learning_rate": 1.128600724155482e-07, + "loss": 0.0012, + "step": 2644 + }, + { + "epoch": 9.338040600176523, + "grad_norm": 0.2342189694217044, + "learning_rate": 1.1165431029421136e-07, + "loss": 0.0016, + "step": 2645 + }, + { + "epoch": 9.341571050308914, + "grad_norm": 0.1090526503271642, + "learning_rate": 1.1045495107570026e-07, + "loss": 0.0008, + "step": 2646 + }, + { + "epoch": 9.345101500441306, + "grad_norm": 0.13018494638297803, + "learning_rate": 1.0926199633097156e-07, + "loss": 0.0008, + "step": 2647 + }, + { + "epoch": 9.348631950573699, + "grad_norm": 0.09502648667692801, + "learning_rate": 1.0807544762259481e-07, + "loss": 0.0005, + "step": 2648 + }, + { + "epoch": 9.35216240070609, + "grad_norm": 0.22990859008196746, + "learning_rate": 1.0689530650474622e-07, + "loss": 0.002, + "step": 2649 + }, + { + "epoch": 9.355692850838482, + "grad_norm": 0.15673900763814655, + "learning_rate": 1.0572157452321097e-07, + "loss": 0.0009, + "step": 2650 + }, + { + "epoch": 9.359223300970873, + "grad_norm": 0.15287020061998272, + "learning_rate": 1.045542532153776e-07, + "loss": 0.0008, + "step": 2651 + }, + { + "epoch": 9.362753751103266, + "grad_norm": 0.23814278433664388, + "learning_rate": 1.0339334411024026e-07, + "loss": 0.0016, + "step": 2652 + }, + { + "epoch": 9.366284201235658, + "grad_norm": 0.19432505452576432, + "learning_rate": 1.0223884872839207e-07, + "loss": 0.0014, + "step": 2653 + }, + { + "epoch": 9.36981465136805, + "grad_norm": 0.49119268333429744, + "learning_rate": 1.0109076858202504e-07, + "loss": 0.0043, + "step": 2654 + }, + { + "epoch": 9.37334510150044, + "grad_norm": 0.0990764225693036, + "learning_rate": 9.994910517493017e-08, + "loss": 0.0007, + "step": 2655 + }, + { + "epoch": 9.376875551632834, + "grad_norm": 0.049583451141731474, + "learning_rate": 9.881386000249293e-08, + "loss": 0.0003, + "step": 2656 + }, + { + "epoch": 9.380406001765225, + "grad_norm": 0.06043744620675663, + "learning_rate": 9.768503455169165e-08, + "loss": 0.0004, + "step": 2657 + }, + { + "epoch": 9.383936451897616, + "grad_norm": 0.19638269153074006, + "learning_rate": 9.656263030109525e-08, + "loss": 0.0019, + "step": 2658 + }, + { + "epoch": 9.38746690203001, + "grad_norm": 0.046126179076112495, + "learning_rate": 9.544664872086329e-08, + "loss": 0.0003, + "step": 2659 + }, + { + "epoch": 9.390997352162401, + "grad_norm": 0.07639833960259697, + "learning_rate": 9.433709127274149e-08, + "loss": 0.0004, + "step": 2660 + }, + { + "epoch": 9.394527802294792, + "grad_norm": 0.18350889988269745, + "learning_rate": 9.323395941006342e-08, + "loss": 0.0011, + "step": 2661 + }, + { + "epoch": 9.398058252427184, + "grad_norm": 0.17130328754070157, + "learning_rate": 9.213725457774269e-08, + "loss": 0.0008, + "step": 2662 + }, + { + "epoch": 9.401588702559577, + "grad_norm": 0.2948917222621375, + "learning_rate": 9.104697821227693e-08, + "loss": 0.0016, + "step": 2663 + }, + { + "epoch": 9.405119152691968, + "grad_norm": 0.044910807493932205, + "learning_rate": 8.99631317417432e-08, + "loss": 0.0003, + "step": 2664 + }, + { + "epoch": 9.40864960282436, + "grad_norm": 0.12404725195821824, + "learning_rate": 8.888571658579704e-08, + "loss": 0.0008, + "step": 2665 + }, + { + "epoch": 9.412180052956751, + "grad_norm": 0.13707461312365968, + "learning_rate": 8.781473415566788e-08, + "loss": 0.0007, + "step": 2666 + }, + { + "epoch": 9.415710503089144, + "grad_norm": 0.130343134898194, + "learning_rate": 8.675018585416195e-08, + "loss": 0.0008, + "step": 2667 + }, + { + "epoch": 9.419240953221536, + "grad_norm": 0.07574808129816553, + "learning_rate": 8.569207307565664e-08, + "loss": 0.0004, + "step": 2668 + }, + { + "epoch": 9.422771403353927, + "grad_norm": 0.10440473930285722, + "learning_rate": 8.464039720609996e-08, + "loss": 0.0008, + "step": 2669 + }, + { + "epoch": 9.42630185348632, + "grad_norm": 0.2644490635271849, + "learning_rate": 8.359515962300779e-08, + "loss": 0.0022, + "step": 2670 + }, + { + "epoch": 9.429832303618712, + "grad_norm": 0.1861444439126708, + "learning_rate": 8.255636169546499e-08, + "loss": 0.001, + "step": 2671 + }, + { + "epoch": 9.433362753751103, + "grad_norm": 0.2485854838129698, + "learning_rate": 8.152400478412037e-08, + "loss": 0.0018, + "step": 2672 + }, + { + "epoch": 9.436893203883495, + "grad_norm": 0.10515884979181271, + "learning_rate": 8.049809024118615e-08, + "loss": 0.0006, + "step": 2673 + }, + { + "epoch": 9.440423654015888, + "grad_norm": 0.24600480722287937, + "learning_rate": 7.947861941043577e-08, + "loss": 0.0018, + "step": 2674 + }, + { + "epoch": 9.443954104148279, + "grad_norm": 0.127875205051203, + "learning_rate": 7.84655936272033e-08, + "loss": 0.0013, + "step": 2675 + }, + { + "epoch": 9.44748455428067, + "grad_norm": 0.14764256883483606, + "learning_rate": 7.745901421838009e-08, + "loss": 0.001, + "step": 2676 + }, + { + "epoch": 9.451015004413062, + "grad_norm": 0.15098656967990545, + "learning_rate": 7.645888250241485e-08, + "loss": 0.001, + "step": 2677 + }, + { + "epoch": 9.454545454545455, + "grad_norm": 0.23989224568180725, + "learning_rate": 7.546519978931022e-08, + "loss": 0.0019, + "step": 2678 + }, + { + "epoch": 9.458075904677846, + "grad_norm": 0.13953854451710707, + "learning_rate": 7.447796738062119e-08, + "loss": 0.0009, + "step": 2679 + }, + { + "epoch": 9.461606354810238, + "grad_norm": 0.2671549827736585, + "learning_rate": 7.349718656945504e-08, + "loss": 0.0017, + "step": 2680 + }, + { + "epoch": 9.465136804942631, + "grad_norm": 0.039145713839155945, + "learning_rate": 7.252285864046804e-08, + "loss": 0.0002, + "step": 2681 + }, + { + "epoch": 9.468667255075022, + "grad_norm": 0.1354723399136827, + "learning_rate": 7.15549848698649e-08, + "loss": 0.0009, + "step": 2682 + }, + { + "epoch": 9.472197705207414, + "grad_norm": 0.1191227379903423, + "learning_rate": 7.059356652539483e-08, + "loss": 0.0011, + "step": 2683 + }, + { + "epoch": 9.475728155339805, + "grad_norm": 0.12204849110412382, + "learning_rate": 6.963860486635332e-08, + "loss": 0.001, + "step": 2684 + }, + { + "epoch": 9.479258605472198, + "grad_norm": 0.12465976520343455, + "learning_rate": 6.869010114357755e-08, + "loss": 0.0006, + "step": 2685 + }, + { + "epoch": 9.48278905560459, + "grad_norm": 0.3993709958109007, + "learning_rate": 6.774805659944761e-08, + "loss": 0.0041, + "step": 2686 + }, + { + "epoch": 9.486319505736981, + "grad_norm": 0.20061179755491176, + "learning_rate": 6.68124724678798e-08, + "loss": 0.0013, + "step": 2687 + }, + { + "epoch": 9.489849955869373, + "grad_norm": 0.15203686430000965, + "learning_rate": 6.588334997433165e-08, + "loss": 0.0012, + "step": 2688 + }, + { + "epoch": 9.493380406001766, + "grad_norm": 0.20776491165767974, + "learning_rate": 6.496069033579688e-08, + "loss": 0.0016, + "step": 2689 + }, + { + "epoch": 9.496910856134157, + "grad_norm": 0.1837102364539882, + "learning_rate": 6.404449476080099e-08, + "loss": 0.0015, + "step": 2690 + }, + { + "epoch": 9.500441306266548, + "grad_norm": 0.13952670460969452, + "learning_rate": 6.31347644494057e-08, + "loss": 0.0005, + "step": 2691 + }, + { + "epoch": 9.503971756398942, + "grad_norm": 0.24745353767830067, + "learning_rate": 6.223150059320282e-08, + "loss": 0.0017, + "step": 2692 + }, + { + "epoch": 9.507502206531333, + "grad_norm": 0.14766605475983083, + "learning_rate": 6.133470437531486e-08, + "loss": 0.0009, + "step": 2693 + }, + { + "epoch": 9.511032656663724, + "grad_norm": 0.22596411187422832, + "learning_rate": 6.044437697039274e-08, + "loss": 0.0022, + "step": 2694 + }, + { + "epoch": 9.514563106796116, + "grad_norm": 0.04986720848965646, + "learning_rate": 5.9560519544614725e-08, + "loss": 0.0003, + "step": 2695 + }, + { + "epoch": 9.518093556928509, + "grad_norm": 0.1344169464123826, + "learning_rate": 5.8683133255682536e-08, + "loss": 0.0009, + "step": 2696 + }, + { + "epoch": 9.5216240070609, + "grad_norm": 0.3282279542782482, + "learning_rate": 5.781221925282354e-08, + "loss": 0.0022, + "step": 2697 + }, + { + "epoch": 9.525154457193292, + "grad_norm": 0.053478112500456976, + "learning_rate": 5.694777867678858e-08, + "loss": 0.0003, + "step": 2698 + }, + { + "epoch": 9.528684907325683, + "grad_norm": 0.04044414895593534, + "learning_rate": 5.608981265984692e-08, + "loss": 0.0002, + "step": 2699 + }, + { + "epoch": 9.532215357458076, + "grad_norm": 0.17067767299678166, + "learning_rate": 5.523832232578796e-08, + "loss": 0.0012, + "step": 2700 + }, + { + "epoch": 9.535745807590468, + "grad_norm": 0.18898568281610514, + "learning_rate": 5.4393308789920105e-08, + "loss": 0.001, + "step": 2701 + }, + { + "epoch": 9.53927625772286, + "grad_norm": 0.17690931430215734, + "learning_rate": 5.3554773159067986e-08, + "loss": 0.0011, + "step": 2702 + }, + { + "epoch": 9.542806707855252, + "grad_norm": 0.1989568613533846, + "learning_rate": 5.272271653157024e-08, + "loss": 0.0014, + "step": 2703 + }, + { + "epoch": 9.546337157987644, + "grad_norm": 0.16580354747027745, + "learning_rate": 5.1897139997280074e-08, + "loss": 0.0017, + "step": 2704 + }, + { + "epoch": 9.549867608120035, + "grad_norm": 0.25336536817783006, + "learning_rate": 5.10780446375625e-08, + "loss": 0.0022, + "step": 2705 + }, + { + "epoch": 9.553398058252426, + "grad_norm": 0.22413056630100875, + "learning_rate": 5.0265431525293175e-08, + "loss": 0.0016, + "step": 2706 + }, + { + "epoch": 9.55692850838482, + "grad_norm": 0.3167026044604959, + "learning_rate": 4.945930172485736e-08, + "loss": 0.0024, + "step": 2707 + }, + { + "epoch": 9.560458958517211, + "grad_norm": 0.1788131783738274, + "learning_rate": 4.865965629214819e-08, + "loss": 0.0017, + "step": 2708 + }, + { + "epoch": 9.563989408649602, + "grad_norm": 0.1763501585691502, + "learning_rate": 4.786649627456563e-08, + "loss": 0.0012, + "step": 2709 + }, + { + "epoch": 9.567519858781996, + "grad_norm": 0.13495480452604583, + "learning_rate": 4.7079822711015296e-08, + "loss": 0.0011, + "step": 2710 + }, + { + "epoch": 9.571050308914387, + "grad_norm": 0.0576217997707109, + "learning_rate": 4.629963663190462e-08, + "loss": 0.0004, + "step": 2711 + }, + { + "epoch": 9.574580759046778, + "grad_norm": 0.29376065388046413, + "learning_rate": 4.552593905914504e-08, + "loss": 0.002, + "step": 2712 + }, + { + "epoch": 9.57811120917917, + "grad_norm": 0.19534319469967165, + "learning_rate": 4.4758731006149804e-08, + "loss": 0.0017, + "step": 2713 + }, + { + "epoch": 9.581641659311563, + "grad_norm": 0.08294356141254022, + "learning_rate": 4.399801347783062e-08, + "loss": 0.0005, + "step": 2714 + }, + { + "epoch": 9.585172109443954, + "grad_norm": 0.07108955014887412, + "learning_rate": 4.324378747059876e-08, + "loss": 0.0006, + "step": 2715 + }, + { + "epoch": 9.588702559576346, + "grad_norm": 0.1188319035764501, + "learning_rate": 4.249605397236178e-08, + "loss": 0.0007, + "step": 2716 + }, + { + "epoch": 9.592233009708737, + "grad_norm": 0.26450351102420977, + "learning_rate": 4.1754813962523457e-08, + "loss": 0.0022, + "step": 2717 + }, + { + "epoch": 9.59576345984113, + "grad_norm": 0.2729168272236753, + "learning_rate": 4.102006841198325e-08, + "loss": 0.0024, + "step": 2718 + }, + { + "epoch": 9.599293909973522, + "grad_norm": 0.18286660674348681, + "learning_rate": 4.0291818283132445e-08, + "loss": 0.0016, + "step": 2719 + }, + { + "epoch": 9.602824360105913, + "grad_norm": 0.348420318536569, + "learning_rate": 3.957006452985468e-08, + "loss": 0.0019, + "step": 2720 + }, + { + "epoch": 9.606354810238305, + "grad_norm": 0.11120669712955043, + "learning_rate": 3.88548080975254e-08, + "loss": 0.0008, + "step": 2721 + }, + { + "epoch": 9.609885260370698, + "grad_norm": 0.10030101674476403, + "learning_rate": 3.814604992300908e-08, + "loss": 0.0007, + "step": 2722 + }, + { + "epoch": 9.613415710503089, + "grad_norm": 0.16424311508552314, + "learning_rate": 3.744379093465811e-08, + "loss": 0.0012, + "step": 2723 + }, + { + "epoch": 9.61694616063548, + "grad_norm": 0.19510120712688056, + "learning_rate": 3.6748032052313366e-08, + "loss": 0.0008, + "step": 2724 + }, + { + "epoch": 9.620476610767874, + "grad_norm": 0.20167137558094284, + "learning_rate": 3.605877418729975e-08, + "loss": 0.001, + "step": 2725 + }, + { + "epoch": 9.624007060900265, + "grad_norm": 0.14293333305059602, + "learning_rate": 3.537601824242898e-08, + "loss": 0.001, + "step": 2726 + }, + { + "epoch": 9.627537511032656, + "grad_norm": 0.08100352527229188, + "learning_rate": 3.4699765111995134e-08, + "loss": 0.0004, + "step": 2727 + }, + { + "epoch": 9.631067961165048, + "grad_norm": 0.18209330950088104, + "learning_rate": 3.403001568177411e-08, + "loss": 0.0016, + "step": 2728 + }, + { + "epoch": 9.634598411297441, + "grad_norm": 0.23379316151218985, + "learning_rate": 3.336677082902529e-08, + "loss": 0.002, + "step": 2729 + }, + { + "epoch": 9.638128861429832, + "grad_norm": 0.05261566854563365, + "learning_rate": 3.271003142248652e-08, + "loss": 0.0003, + "step": 2730 + }, + { + "epoch": 9.641659311562224, + "grad_norm": 0.1240448240368359, + "learning_rate": 3.205979832237416e-08, + "loss": 0.0006, + "step": 2731 + }, + { + "epoch": 9.645189761694617, + "grad_norm": 0.48223749562745, + "learning_rate": 3.1416072380383575e-08, + "loss": 0.0034, + "step": 2732 + }, + { + "epoch": 9.648720211827008, + "grad_norm": 0.11955744356269654, + "learning_rate": 3.0778854439686426e-08, + "loss": 0.0007, + "step": 2733 + }, + { + "epoch": 9.6522506619594, + "grad_norm": 0.2499050501896369, + "learning_rate": 3.01481453349306e-08, + "loss": 0.0016, + "step": 2734 + }, + { + "epoch": 9.655781112091791, + "grad_norm": 0.1907983812315517, + "learning_rate": 2.952394589223695e-08, + "loss": 0.0019, + "step": 2735 + }, + { + "epoch": 9.659311562224184, + "grad_norm": 0.17861233864942413, + "learning_rate": 2.8906256929200903e-08, + "loss": 0.0008, + "step": 2736 + }, + { + "epoch": 9.662842012356576, + "grad_norm": 0.20702410514575556, + "learning_rate": 2.8295079254890834e-08, + "loss": 0.0012, + "step": 2737 + }, + { + "epoch": 9.666372462488967, + "grad_norm": 0.2322811653803858, + "learning_rate": 2.769041366984582e-08, + "loss": 0.0013, + "step": 2738 + }, + { + "epoch": 9.669902912621358, + "grad_norm": 0.18306726264414455, + "learning_rate": 2.7092260966074536e-08, + "loss": 0.0019, + "step": 2739 + }, + { + "epoch": 9.673433362753752, + "grad_norm": 0.11402199245979718, + "learning_rate": 2.6500621927054716e-08, + "loss": 0.0008, + "step": 2740 + }, + { + "epoch": 9.676963812886143, + "grad_norm": 0.11117198341604853, + "learning_rate": 2.591549732773424e-08, + "loss": 0.0011, + "step": 2741 + }, + { + "epoch": 9.680494263018534, + "grad_norm": 0.21516330254124214, + "learning_rate": 2.533688793452671e-08, + "loss": 0.0019, + "step": 2742 + }, + { + "epoch": 9.684024713150926, + "grad_norm": 0.21619422644213207, + "learning_rate": 2.476479450531255e-08, + "loss": 0.0013, + "step": 2743 + }, + { + "epoch": 9.687555163283319, + "grad_norm": 0.14033317809370138, + "learning_rate": 2.419921778943568e-08, + "loss": 0.0013, + "step": 2744 + }, + { + "epoch": 9.69108561341571, + "grad_norm": 0.2112328172344002, + "learning_rate": 2.3640158527706846e-08, + "loss": 0.0023, + "step": 2745 + }, + { + "epoch": 9.694616063548102, + "grad_norm": 0.18437529503225644, + "learning_rate": 2.3087617452398625e-08, + "loss": 0.0013, + "step": 2746 + }, + { + "epoch": 9.698146513680495, + "grad_norm": 0.17315670831841823, + "learning_rate": 2.254159528724542e-08, + "loss": 0.0012, + "step": 2747 + }, + { + "epoch": 9.701676963812886, + "grad_norm": 0.2141300382090153, + "learning_rate": 2.2002092747443472e-08, + "loss": 0.0009, + "step": 2748 + }, + { + "epoch": 9.705207413945278, + "grad_norm": 0.17412250636188412, + "learning_rate": 2.1469110539650283e-08, + "loss": 0.0006, + "step": 2749 + }, + { + "epoch": 9.70873786407767, + "grad_norm": 0.05663583122288284, + "learning_rate": 2.094264936198187e-08, + "loss": 0.0004, + "step": 2750 + }, + { + "epoch": 9.712268314210062, + "grad_norm": 0.1829310683571671, + "learning_rate": 2.0422709904012738e-08, + "loss": 0.0017, + "step": 2751 + }, + { + "epoch": 9.715798764342454, + "grad_norm": 0.18448217055876792, + "learning_rate": 1.9909292846774784e-08, + "loss": 0.0008, + "step": 2752 + }, + { + "epoch": 9.719329214474845, + "grad_norm": 0.05962778931968933, + "learning_rate": 1.9402398862757854e-08, + "loss": 0.0004, + "step": 2753 + }, + { + "epoch": 9.722859664607238, + "grad_norm": 0.09835860277288541, + "learning_rate": 1.8902028615906953e-08, + "loss": 0.0009, + "step": 2754 + }, + { + "epoch": 9.72639011473963, + "grad_norm": 0.11630618452428126, + "learning_rate": 1.840818276162226e-08, + "loss": 0.0007, + "step": 2755 + }, + { + "epoch": 9.729920564872021, + "grad_norm": 0.07517186432613839, + "learning_rate": 1.7920861946758018e-08, + "loss": 0.0004, + "step": 2756 + }, + { + "epoch": 9.733451015004412, + "grad_norm": 0.079255788382644, + "learning_rate": 1.744006680962085e-08, + "loss": 0.0004, + "step": 2757 + }, + { + "epoch": 9.736981465136806, + "grad_norm": 0.05236121659147581, + "learning_rate": 1.6965797979971442e-08, + "loss": 0.0004, + "step": 2758 + }, + { + "epoch": 9.740511915269197, + "grad_norm": 0.17753842362324607, + "learning_rate": 1.6498056079021773e-08, + "loss": 0.0012, + "step": 2759 + }, + { + "epoch": 9.744042365401588, + "grad_norm": 0.22257179806960997, + "learning_rate": 1.6036841719433428e-08, + "loss": 0.0015, + "step": 2760 + }, + { + "epoch": 9.74757281553398, + "grad_norm": 0.17624469739083465, + "learning_rate": 1.558215550531983e-08, + "loss": 0.0011, + "step": 2761 + }, + { + "epoch": 9.751103265666373, + "grad_norm": 0.1710315893595879, + "learning_rate": 1.513399803224236e-08, + "loss": 0.0008, + "step": 2762 + }, + { + "epoch": 9.754633715798764, + "grad_norm": 0.12431802692502394, + "learning_rate": 1.46923698872109e-08, + "loss": 0.0007, + "step": 2763 + }, + { + "epoch": 9.758164165931156, + "grad_norm": 0.08459693942860483, + "learning_rate": 1.425727164868329e-08, + "loss": 0.0005, + "step": 2764 + }, + { + "epoch": 9.761694616063549, + "grad_norm": 0.15897277728180298, + "learning_rate": 1.382870388656532e-08, + "loss": 0.0014, + "step": 2765 + }, + { + "epoch": 9.76522506619594, + "grad_norm": 0.12340505461782379, + "learning_rate": 1.3406667162206843e-08, + "loss": 0.0008, + "step": 2766 + }, + { + "epoch": 9.768755516328332, + "grad_norm": 0.21932856167592174, + "learning_rate": 1.2991162028405113e-08, + "loss": 0.0019, + "step": 2767 + }, + { + "epoch": 9.772285966460723, + "grad_norm": 0.19830274243369675, + "learning_rate": 1.2582189029400339e-08, + "loss": 0.0013, + "step": 2768 + }, + { + "epoch": 9.775816416593116, + "grad_norm": 0.14728497962416523, + "learning_rate": 1.2179748700879013e-08, + "loss": 0.0009, + "step": 2769 + }, + { + "epoch": 9.779346866725508, + "grad_norm": 0.17740215713776258, + "learning_rate": 1.1783841569968368e-08, + "loss": 0.001, + "step": 2770 + }, + { + "epoch": 9.782877316857899, + "grad_norm": 0.12042809392155396, + "learning_rate": 1.1394468155240258e-08, + "loss": 0.0006, + "step": 2771 + }, + { + "epoch": 9.78640776699029, + "grad_norm": 0.13344026305328172, + "learning_rate": 1.1011628966707822e-08, + "loss": 0.0009, + "step": 2772 + }, + { + "epoch": 9.789938217122684, + "grad_norm": 0.2661203700676195, + "learning_rate": 1.063532450582494e-08, + "loss": 0.0023, + "step": 2773 + }, + { + "epoch": 9.793468667255075, + "grad_norm": 0.14642180316783504, + "learning_rate": 1.0265555265486782e-08, + "loss": 0.001, + "step": 2774 + }, + { + "epoch": 9.796999117387466, + "grad_norm": 0.1568511893465151, + "learning_rate": 9.902321730028142e-09, + "loss": 0.001, + "step": 2775 + }, + { + "epoch": 9.80052956751986, + "grad_norm": 0.34031303347621616, + "learning_rate": 9.545624375223439e-09, + "loss": 0.0021, + "step": 2776 + }, + { + "epoch": 9.804060017652251, + "grad_norm": 0.19157015666080168, + "learning_rate": 9.195463668285609e-09, + "loss": 0.0015, + "step": 2777 + }, + { + "epoch": 9.807590467784642, + "grad_norm": 0.18876962207648357, + "learning_rate": 8.851840067866102e-09, + "loss": 0.0011, + "step": 2778 + }, + { + "epoch": 9.811120917917034, + "grad_norm": 0.13807721742887907, + "learning_rate": 8.514754024052663e-09, + "loss": 0.0005, + "step": 2779 + }, + { + "epoch": 9.814651368049427, + "grad_norm": 0.08211939200650463, + "learning_rate": 8.184205978370996e-09, + "loss": 0.0004, + "step": 2780 + }, + { + "epoch": 9.818181818181818, + "grad_norm": 0.20411340111655185, + "learning_rate": 7.860196363783656e-09, + "loss": 0.0009, + "step": 2781 + }, + { + "epoch": 9.82171226831421, + "grad_norm": 0.22261857928504833, + "learning_rate": 7.542725604687274e-09, + "loss": 0.0014, + "step": 2782 + }, + { + "epoch": 9.825242718446601, + "grad_norm": 0.1592099803301052, + "learning_rate": 7.231794116915325e-09, + "loss": 0.0013, + "step": 2783 + }, + { + "epoch": 9.828773168578994, + "grad_norm": 0.09023419756306375, + "learning_rate": 6.927402307734809e-09, + "loss": 0.0006, + "step": 2784 + }, + { + "epoch": 9.832303618711386, + "grad_norm": 0.27434009735766135, + "learning_rate": 6.629550575847355e-09, + "loss": 0.0016, + "step": 2785 + }, + { + "epoch": 9.835834068843777, + "grad_norm": 0.10018150186665546, + "learning_rate": 6.338239311388661e-09, + "loss": 0.0005, + "step": 2786 + }, + { + "epoch": 9.83936451897617, + "grad_norm": 0.2996165201503127, + "learning_rate": 6.053468895926285e-09, + "loss": 0.0022, + "step": 2787 + }, + { + "epoch": 9.842894969108562, + "grad_norm": 0.16854577196780418, + "learning_rate": 5.775239702461855e-09, + "loss": 0.0007, + "step": 2788 + }, + { + "epoch": 9.846425419240953, + "grad_norm": 0.10354721235776354, + "learning_rate": 5.503552095427745e-09, + "loss": 0.0005, + "step": 2789 + }, + { + "epoch": 9.849955869373344, + "grad_norm": 0.1264150816734169, + "learning_rate": 5.238406430688736e-09, + "loss": 0.0007, + "step": 2790 + }, + { + "epoch": 9.853486319505738, + "grad_norm": 0.19960519157944526, + "learning_rate": 4.9798030555409105e-09, + "loss": 0.0021, + "step": 2791 + }, + { + "epoch": 9.857016769638129, + "grad_norm": 0.1346572697588004, + "learning_rate": 4.727742308710537e-09, + "loss": 0.0011, + "step": 2792 + }, + { + "epoch": 9.86054721977052, + "grad_norm": 0.06040725483847343, + "learning_rate": 4.482224520354627e-09, + "loss": 0.0004, + "step": 2793 + }, + { + "epoch": 9.864077669902912, + "grad_norm": 0.24537535501618093, + "learning_rate": 4.243250012059275e-09, + "loss": 0.0025, + "step": 2794 + }, + { + "epoch": 9.867608120035305, + "grad_norm": 0.21409351147972147, + "learning_rate": 4.010819096841312e-09, + "loss": 0.0016, + "step": 2795 + }, + { + "epoch": 9.871138570167696, + "grad_norm": 0.29925922209397343, + "learning_rate": 3.784932079144987e-09, + "loss": 0.0019, + "step": 2796 + }, + { + "epoch": 9.874669020300088, + "grad_norm": 0.30027833486356653, + "learning_rate": 3.565589254844182e-09, + "loss": 0.0018, + "step": 2797 + }, + { + "epoch": 9.878199470432481, + "grad_norm": 0.0786551395809337, + "learning_rate": 3.352790911241299e-09, + "loss": 0.0006, + "step": 2798 + }, + { + "epoch": 9.881729920564872, + "grad_norm": 0.29794606587220973, + "learning_rate": 3.1465373270650466e-09, + "loss": 0.0027, + "step": 2799 + }, + { + "epoch": 9.885260370697264, + "grad_norm": 0.36390195968617994, + "learning_rate": 2.9468287724737642e-09, + "loss": 0.0035, + "step": 2800 + }, + { + "epoch": 9.888790820829655, + "grad_norm": 0.20260450622382992, + "learning_rate": 2.7536655090504295e-09, + "loss": 0.0018, + "step": 2801 + }, + { + "epoch": 9.892321270962048, + "grad_norm": 0.1716067494499037, + "learning_rate": 2.5670477898070978e-09, + "loss": 0.0013, + "step": 2802 + }, + { + "epoch": 9.89585172109444, + "grad_norm": 0.15526497563258354, + "learning_rate": 2.3869758591810177e-09, + "loss": 0.0009, + "step": 2803 + }, + { + "epoch": 9.899382171226831, + "grad_norm": 0.23145744048826364, + "learning_rate": 2.21344995303574e-09, + "loss": 0.0021, + "step": 2804 + }, + { + "epoch": 9.902912621359224, + "grad_norm": 0.17815058246245158, + "learning_rate": 2.0464702986611184e-09, + "loss": 0.0012, + "step": 2805 + }, + { + "epoch": 9.906443071491616, + "grad_norm": 0.03553776709927562, + "learning_rate": 1.886037114771644e-09, + "loss": 0.0002, + "step": 2806 + }, + { + "epoch": 9.909973521624007, + "grad_norm": 0.1068070294794527, + "learning_rate": 1.7321506115075547e-09, + "loss": 0.0008, + "step": 2807 + }, + { + "epoch": 9.913503971756398, + "grad_norm": 0.16340063680286507, + "learning_rate": 1.5848109904348374e-09, + "loss": 0.0014, + "step": 2808 + }, + { + "epoch": 9.917034421888792, + "grad_norm": 0.05756152809089501, + "learning_rate": 1.44401844454245e-09, + "loss": 0.0004, + "step": 2809 + }, + { + "epoch": 9.920564872021183, + "grad_norm": 0.13510742134638923, + "learning_rate": 1.3097731582450978e-09, + "loss": 0.001, + "step": 2810 + }, + { + "epoch": 9.924095322153574, + "grad_norm": 0.14532237386155003, + "learning_rate": 1.1820753073810144e-09, + "loss": 0.0008, + "step": 2811 + }, + { + "epoch": 9.927625772285966, + "grad_norm": 0.08493585835457969, + "learning_rate": 1.0609250592130693e-09, + "loss": 0.0005, + "step": 2812 + }, + { + "epoch": 9.931156222418359, + "grad_norm": 0.09785423366409138, + "learning_rate": 9.463225724271053e-10, + "loss": 0.0008, + "step": 2813 + }, + { + "epoch": 9.93468667255075, + "grad_norm": 0.06518416769007865, + "learning_rate": 8.382679971336017e-10, + "loss": 0.0004, + "step": 2814 + }, + { + "epoch": 9.938217122683142, + "grad_norm": 0.19310083988622934, + "learning_rate": 7.36761474865455e-10, + "loss": 0.0014, + "step": 2815 + }, + { + "epoch": 9.941747572815533, + "grad_norm": 0.23980946647797793, + "learning_rate": 6.41803138578534e-10, + "loss": 0.0016, + "step": 2816 + }, + { + "epoch": 9.945278022947926, + "grad_norm": 0.26105309077288136, + "learning_rate": 5.533931126527892e-10, + "loss": 0.0017, + "step": 2817 + }, + { + "epoch": 9.948808473080318, + "grad_norm": 0.24907617437212784, + "learning_rate": 4.715315128894782e-10, + "loss": 0.0014, + "step": 2818 + }, + { + "epoch": 9.952338923212709, + "grad_norm": 0.10639203783779039, + "learning_rate": 3.962184465144958e-10, + "loss": 0.0007, + "step": 2819 + }, + { + "epoch": 9.955869373345102, + "grad_norm": 0.14899172228298055, + "learning_rate": 3.274540121744885e-10, + "loss": 0.0013, + "step": 2820 + }, + { + "epoch": 9.959399823477494, + "grad_norm": 0.07394012811062049, + "learning_rate": 2.6523829993907456e-10, + "loss": 0.0004, + "step": 2821 + }, + { + "epoch": 9.962930273609885, + "grad_norm": 0.11402029197303257, + "learning_rate": 2.0957139130028948e-10, + "loss": 0.0007, + "step": 2822 + }, + { + "epoch": 9.966460723742276, + "grad_norm": 0.1848413730393921, + "learning_rate": 1.6045335917258542e-10, + "loss": 0.0016, + "step": 2823 + }, + { + "epoch": 9.96999117387467, + "grad_norm": 0.18455147464263416, + "learning_rate": 1.1788426789227647e-10, + "loss": 0.0015, + "step": 2824 + }, + { + "epoch": 9.973521624007061, + "grad_norm": 0.1423320988684452, + "learning_rate": 8.186417321809359e-11, + "loss": 0.0012, + "step": 2825 + }, + { + "epoch": 9.977052074139452, + "grad_norm": 0.31130541454063626, + "learning_rate": 5.239312232896421e-11, + "loss": 0.002, + "step": 2826 + }, + { + "epoch": 9.980582524271846, + "grad_norm": 0.2359558106083685, + "learning_rate": 2.9471153827898e-11, + "loss": 0.002, + "step": 2827 + }, + { + "epoch": 9.984112974404237, + "grad_norm": 0.12670378160201803, + "learning_rate": 1.309829773865623e-11, + "loss": 0.0005, + "step": 2828 + }, + { + "epoch": 9.987643424536628, + "grad_norm": 0.2583622355409926, + "learning_rate": 3.2745755068619432e-12, + "loss": 0.0016, + "step": 2829 + }, + { + "epoch": 9.99117387466902, + "grad_norm": 0.16532659014255321, + "learning_rate": 0.0, + "loss": 0.0014, + "step": 2830 + } + ], + "logging_steps": 1.0, + "max_steps": 2830, + "num_input_tokens_seen": 0, + "num_train_epochs": 10, + "save_steps": 1415, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 221122765266944.0, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +}