[ { "current_steps": 0, "loss": 2.8643, "learning_rate": 1e-08, "epoch": 0.0015105740181268882 }, { "current_steps": 1, "loss": 3.3851, "learning_rate": 2e-08, "epoch": 0.0030211480362537764 }, { "current_steps": 2, "loss": 2.9606, "learning_rate": 3e-08, "epoch": 0.004531722054380665 }, { "current_steps": 3, "loss": 3.2647, "learning_rate": 4e-08, "epoch": 0.006042296072507553 }, { "current_steps": 4, "loss": 3.3543, "learning_rate": 5e-08, "epoch": 0.0075528700906344415 }, { "current_steps": 5, "loss": 3.2605, "learning_rate": 5e-08, "epoch": 0.00906344410876133 }, { "current_steps": 6, "loss": 3.3266, "learning_rate": 6e-08, "epoch": 0.010574018126888218 }, { "current_steps": 7, "loss": 3.0449, "learning_rate": 7e-08, "epoch": 0.012084592145015106 }, { "current_steps": 8, "loss": 3.2478, "learning_rate": 8e-08, "epoch": 0.013595166163141994 }, { "current_steps": 9, "loss": 3.3969, "learning_rate": 9e-08, "epoch": 0.015105740181268883 }, { "current_steps": 10, "loss": 3.2827, "learning_rate": 1e-07, "epoch": 0.01661631419939577 }, { "current_steps": 11, "loss": 3.0901, "learning_rate": 1.0999999999999999e-07, "epoch": 0.01812688821752266 }, { "current_steps": 12, "loss": 3.1617, "learning_rate": 1.2e-07, "epoch": 0.019637462235649546 }, { "current_steps": 13, "loss": 3.34, "learning_rate": 1.3e-07, "epoch": 0.021148036253776436 }, { "current_steps": 14, "loss": 3.3251, "learning_rate": 1.4e-07, "epoch": 0.022658610271903322 }, { "current_steps": 15, "loss": 3.1552, "learning_rate": 1.5e-07, "epoch": 0.02416918429003021 }, { "current_steps": 16, "loss": 3.2264, "learning_rate": 1.6e-07, "epoch": 0.0256797583081571 }, { "current_steps": 17, "loss": 3.1803, "learning_rate": 1.6e-07, "epoch": 0.027190332326283987 }, { "current_steps": 18, "loss": 3.2943, "learning_rate": 1.7000000000000001e-07, "epoch": 0.028700906344410877 }, { "current_steps": 19, "loss": 3.3261, "learning_rate": 1.8e-07, "epoch": 0.030211480362537766 }, { "current_steps": 20, "loss": 3.1263, "learning_rate": 1.8999999999999998e-07, "epoch": 0.03172205438066465 }, { "current_steps": 21, "loss": 3.1523, "learning_rate": 2e-07, "epoch": 0.03323262839879154 }, { "current_steps": 22, "loss": 3.312, "learning_rate": 2.0999999999999997e-07, "epoch": 0.03474320241691843 }, { "current_steps": 23, "loss": 3.2666, "learning_rate": 2.1999999999999998e-07, "epoch": 0.03625377643504532 }, { "current_steps": 24, "loss": 3.2577, "learning_rate": 2.3e-07, "epoch": 0.0377643504531722 }, { "current_steps": 25, "loss": 3.3348, "learning_rate": 2.4e-07, "epoch": 0.03927492447129909 }, { "current_steps": 26, "loss": 3.2292, "learning_rate": 2.5e-07, "epoch": 0.04078549848942598 }, { "current_steps": 27, "loss": 3.242, "learning_rate": 2.6e-07, "epoch": 0.04229607250755287 }, { "current_steps": 28, "loss": 3.2687, "learning_rate": 2.7e-07, "epoch": 0.04380664652567976 }, { "current_steps": 29, "loss": 2.7037, "learning_rate": 2.8e-07, "epoch": 0.045317220543806644 }, { "current_steps": 30, "loss": 3.1619, "learning_rate": 2.9e-07, "epoch": 0.04682779456193353 }, { "current_steps": 31, "loss": 3.2853, "learning_rate": 3e-07, "epoch": 0.04833836858006042 }, { "current_steps": 32, "loss": 3.3545, "learning_rate": 3.1e-07, "epoch": 0.04984894259818731 }, { "current_steps": 33, "loss": 2.5809, "learning_rate": 3.2e-07, "epoch": 0.0513595166163142 }, { "current_steps": 34, "loss": 2.9552, "learning_rate": 3.3e-07, "epoch": 0.052870090634441085 }, { "current_steps": 35, "loss": 3.1433, "learning_rate": 3.4000000000000003e-07, "epoch": 0.054380664652567974 }, { "current_steps": 36, "loss": 3.3105, "learning_rate": 3.5e-07, "epoch": 0.055891238670694864 }, { "current_steps": 37, "loss": 3.0981, "learning_rate": 3.6e-07, "epoch": 0.05740181268882175 }, { "current_steps": 38, "loss": 3.2201, "learning_rate": 3.7e-07, "epoch": 0.05891238670694864 }, { "current_steps": 39, "loss": 3.0603, "learning_rate": 3.7999999999999996e-07, "epoch": 0.06042296072507553 }, { "current_steps": 40, "loss": 3.3152, "learning_rate": 3.8999999999999997e-07, "epoch": 0.061933534743202415 }, { "current_steps": 41, "loss": 3.4646, "learning_rate": 4e-07, "epoch": 0.0634441087613293 }, { "current_steps": 42, "loss": 3.3562, "learning_rate": 4.0999999999999994e-07, "epoch": 0.0649546827794562 }, { "current_steps": 43, "loss": 3.1192, "learning_rate": 4.1999999999999995e-07, "epoch": 0.06646525679758308 }, { "current_steps": 44, "loss": 3.5199, "learning_rate": 4.2999999999999996e-07, "epoch": 0.06797583081570997 }, { "current_steps": 45, "loss": 3.1833, "learning_rate": 4.3999999999999997e-07, "epoch": 0.06948640483383686 }, { "current_steps": 46, "loss": 3.36, "learning_rate": 4.5e-07, "epoch": 0.07099697885196375 }, { "current_steps": 47, "loss": 2.7168, "learning_rate": 4.6e-07, "epoch": 0.07250755287009064 }, { "current_steps": 48, "loss": 3.1549, "learning_rate": 4.6999999999999995e-07, "epoch": 0.07401812688821752 }, { "current_steps": 49, "loss": 3.2864, "learning_rate": 4.8e-07, "epoch": 0.0755287009063444 }, { "current_steps": 50, "loss": 3.3691, "learning_rate": 4.9e-07, "epoch": 0.0770392749244713 }, { "current_steps": 51, "loss": 3.0839, "learning_rate": 5e-07, "epoch": 0.07854984894259819 }, { "current_steps": 52, "loss": 3.1339, "learning_rate": 5.1e-07, "epoch": 0.08006042296072508 }, { "current_steps": 53, "loss": 3.3284, "learning_rate": 5.2e-07, "epoch": 0.08157099697885196 }, { "current_steps": 54, "loss": 3.1468, "learning_rate": 5.3e-07, "epoch": 0.08308157099697885 }, { "current_steps": 55, "loss": 3.2341, "learning_rate": 5.4e-07, "epoch": 0.08459214501510574 }, { "current_steps": 56, "loss": 3.0192, "learning_rate": 5.5e-07, "epoch": 0.08610271903323263 }, { "current_steps": 57, "loss": 3.1109, "learning_rate": 5.6e-07, "epoch": 0.08761329305135952 }, { "current_steps": 58, "loss": 3.2616, "learning_rate": 5.699999999999999e-07, "epoch": 0.0891238670694864 }, { "current_steps": 59, "loss": 3.287, "learning_rate": 5.8e-07, "epoch": 0.09063444108761329 }, { "current_steps": 60, "loss": 3.2243, "learning_rate": 5.9e-07, "epoch": 0.09214501510574018 }, { "current_steps": 61, "loss": 3.1392, "learning_rate": 6e-07, "epoch": 0.09365558912386707 }, { "current_steps": 62, "loss": 3.2099, "learning_rate": 6.1e-07, "epoch": 0.09516616314199396 }, { "current_steps": 63, "loss": 3.2016, "learning_rate": 6.2e-07, "epoch": 0.09667673716012085 }, { "current_steps": 64, "loss": 3.1605, "learning_rate": 6.3e-07, "epoch": 0.09818731117824774 }, { "current_steps": 65, "loss": 3.4862, "learning_rate": 6.4e-07, "epoch": 0.09969788519637462 }, { "current_steps": 66, "loss": 2.9683, "learning_rate": 6.5e-07, "epoch": 0.10120845921450151 }, { "current_steps": 67, "loss": 3.5057, "learning_rate": 6.6e-07, "epoch": 0.1027190332326284 }, { "current_steps": 68, "loss": 3.2837, "learning_rate": 6.7e-07, "epoch": 0.1042296072507553 }, { "current_steps": 69, "loss": 3.3029, "learning_rate": 6.800000000000001e-07, "epoch": 0.10574018126888217 }, { "current_steps": 70, "loss": 2.7988, "learning_rate": 6.9e-07, "epoch": 0.10725075528700906 }, { "current_steps": 71, "loss": 2.8721, "learning_rate": 7e-07, "epoch": 0.10876132930513595 }, { "current_steps": 72, "loss": 3.0222, "learning_rate": 7.1e-07, "epoch": 0.11027190332326284 }, { "current_steps": 73, "loss": 3.2546, "learning_rate": 7.2e-07, "epoch": 0.11178247734138973 }, { "current_steps": 74, "loss": 3.196, "learning_rate": 7.3e-07, "epoch": 0.11329305135951662 }, { "current_steps": 75, "loss": 3.1073, "learning_rate": 7.4e-07, "epoch": 0.1148036253776435 }, { "current_steps": 76, "loss": 3.4334, "learning_rate": 7.5e-07, "epoch": 0.1163141993957704 }, { "current_steps": 77, "loss": 3.2911, "learning_rate": 7.599999999999999e-07, "epoch": 0.11782477341389729 }, { "current_steps": 78, "loss": 3.2415, "learning_rate": 7.699999999999999e-07, "epoch": 0.11933534743202417 }, { "current_steps": 79, "loss": 3.2162, "learning_rate": 7.799999999999999e-07, "epoch": 0.12084592145015106 }, { "current_steps": 80, "loss": 3.1616, "learning_rate": 7.9e-07, "epoch": 0.12235649546827794 }, { "current_steps": 81, "loss": 3.356, "learning_rate": 8e-07, "epoch": 0.12386706948640483 }, { "current_steps": 82, "loss": 3.2459, "learning_rate": 8.1e-07, "epoch": 0.12537764350453173 }, { "current_steps": 83, "loss": 3.2564, "learning_rate": 8.199999999999999e-07, "epoch": 0.1268882175226586 }, { "current_steps": 84, "loss": 3.2301, "learning_rate": 8.299999999999999e-07, "epoch": 0.1283987915407855 }, { "current_steps": 85, "loss": 3.1873, "learning_rate": 8.399999999999999e-07, "epoch": 0.1299093655589124 }, { "current_steps": 86, "loss": 2.931, "learning_rate": 8.499999999999999e-07, "epoch": 0.13141993957703926 }, { "current_steps": 87, "loss": 3.2395, "learning_rate": 8.599999999999999e-07, "epoch": 0.13293051359516617 }, { "current_steps": 88, "loss": 3.2334, "learning_rate": 8.699999999999999e-07, "epoch": 0.13444108761329304 }, { "current_steps": 89, "loss": 3.408, "learning_rate": 8.799999999999999e-07, "epoch": 0.13595166163141995 }, { "current_steps": 90, "loss": 3.1124, "learning_rate": 8.9e-07, "epoch": 0.13746223564954682 }, { "current_steps": 91, "loss": 3.0723, "learning_rate": 9e-07, "epoch": 0.13897280966767372 }, { "current_steps": 92, "loss": 3.1404, "learning_rate": 9.1e-07, "epoch": 0.1404833836858006 }, { "current_steps": 93, "loss": 3.0665, "learning_rate": 9.2e-07, "epoch": 0.1419939577039275 }, { "current_steps": 94, "loss": 2.6321, "learning_rate": 9.3e-07, "epoch": 0.14350453172205438 }, { "current_steps": 95, "loss": 2.9725, "learning_rate": 9.399999999999999e-07, "epoch": 0.14501510574018128 }, { "current_steps": 96, "loss": 3.1267, "learning_rate": 9.499999999999999e-07, "epoch": 0.14652567975830816 }, { "current_steps": 97, "loss": 3.3175, "learning_rate": 9.6e-07, "epoch": 0.14803625377643503 }, { "current_steps": 98, "loss": 3.277, "learning_rate": 9.7e-07, "epoch": 0.14954682779456194 }, { "current_steps": 99, "loss": 3.3042, "learning_rate": 9.8e-07, "epoch": 0.1510574018126888 }, { "current_steps": 100, "loss": 3.3383, "learning_rate": 9.9e-07, "epoch": 0.15256797583081572 }, { "current_steps": 101, "loss": 2.8985, "learning_rate": 1e-06, "epoch": 0.1540785498489426 }, { "current_steps": 102, "loss": 3.1538, "learning_rate": 9.994697773064687e-07, "epoch": 0.1555891238670695 }, { "current_steps": 103, "loss": 3.3022, "learning_rate": 9.989395546129374e-07, "epoch": 0.15709969788519637 }, { "current_steps": 104, "loss": 2.5211, "learning_rate": 9.98409331919406e-07, "epoch": 0.15861027190332327 }, { "current_steps": 105, "loss": 3.1553, "learning_rate": 9.978791092258748e-07, "epoch": 0.16012084592145015 }, { "current_steps": 106, "loss": 3.097, "learning_rate": 9.973488865323435e-07, "epoch": 0.16163141993957703 }, { "current_steps": 107, "loss": 3.1513, "learning_rate": 9.968186638388122e-07, "epoch": 0.16314199395770393 }, { "current_steps": 108, "loss": 3.1952, "learning_rate": 9.96288441145281e-07, "epoch": 0.1646525679758308 }, { "current_steps": 109, "loss": 3.1655, "learning_rate": 9.957582184517497e-07, "epoch": 0.1661631419939577 }, { "current_steps": 110, "loss": 3.3205, "learning_rate": 9.952279957582184e-07, "epoch": 0.16767371601208458 }, { "current_steps": 111, "loss": 3.1351, "learning_rate": 9.946977730646871e-07, "epoch": 0.1691842900302115 }, { "current_steps": 112, "loss": 2.9261, "learning_rate": 9.941675503711558e-07, "epoch": 0.17069486404833836 }, { "current_steps": 113, "loss": 3.0024, "learning_rate": 9.936373276776245e-07, "epoch": 0.17220543806646527 }, { "current_steps": 114, "loss": 3.0746, "learning_rate": 9.931071049840933e-07, "epoch": 0.17371601208459214 }, { "current_steps": 115, "loss": 3.0436, "learning_rate": 9.92576882290562e-07, "epoch": 0.17522658610271905 }, { "current_steps": 116, "loss": 3.1425, "learning_rate": 9.920466595970307e-07, "epoch": 0.17673716012084592 }, { "current_steps": 117, "loss": 3.3725, "learning_rate": 9.915164369034994e-07, "epoch": 0.1782477341389728 }, { "current_steps": 118, "loss": 2.691, "learning_rate": 9.909862142099681e-07, "epoch": 0.1797583081570997 }, { "current_steps": 119, "loss": 3.1319, "learning_rate": 9.904559915164368e-07, "epoch": 0.18126888217522658 }, { "current_steps": 120, "loss": 3.4188, "learning_rate": 9.899257688229056e-07, "epoch": 0.18277945619335348 }, { "current_steps": 121, "loss": 3.2334, "learning_rate": 9.893955461293743e-07, "epoch": 0.18429003021148035 }, { "current_steps": 122, "loss": 3.3199, "learning_rate": 9.88865323435843e-07, "epoch": 0.18580060422960726 }, { "current_steps": 123, "loss": 2.7292, "learning_rate": 9.883351007423117e-07, "epoch": 0.18731117824773413 }, { "current_steps": 124, "loss": 2.8369, "learning_rate": 9.878048780487804e-07, "epoch": 0.18882175226586104 }, { "current_steps": 125, "loss": 2.8981, "learning_rate": 9.872746553552491e-07, "epoch": 0.1903323262839879 }, { "current_steps": 126, "loss": 3.082, "learning_rate": 9.867444326617179e-07, "epoch": 0.19184290030211482 }, { "current_steps": 127, "loss": 2.9516, "learning_rate": 9.862142099681866e-07, "epoch": 0.1933534743202417 }, { "current_steps": 128, "loss": 3.3629, "learning_rate": 9.856839872746553e-07, "epoch": 0.19486404833836857 }, { "current_steps": 129, "loss": 2.9393, "learning_rate": 9.85153764581124e-07, "epoch": 0.19637462235649547 }, { "current_steps": 130, "loss": 3.2696, "learning_rate": 9.846235418875927e-07, "epoch": 0.19788519637462235 }, { "current_steps": 131, "loss": 2.9017, "learning_rate": 9.840933191940614e-07, "epoch": 0.19939577039274925 }, { "current_steps": 132, "loss": 3.1235, "learning_rate": 9.835630965005301e-07, "epoch": 0.20090634441087613 }, { "current_steps": 133, "loss": 2.9032, "learning_rate": 9.830328738069989e-07, "epoch": 0.20241691842900303 }, { "current_steps": 134, "loss": 3.1998, "learning_rate": 9.825026511134676e-07, "epoch": 0.2039274924471299 }, { "current_steps": 135, "loss": 3.2412, "learning_rate": 9.819724284199363e-07, "epoch": 0.2054380664652568 }, { "current_steps": 136, "loss": 3.3582, "learning_rate": 9.81442205726405e-07, "epoch": 0.20694864048338368 }, { "current_steps": 137, "loss": 3.2789, "learning_rate": 9.809119830328737e-07, "epoch": 0.2084592145015106 }, { "current_steps": 138, "loss": 3.2905, "learning_rate": 9.803817603393424e-07, "epoch": 0.20996978851963746 }, { "current_steps": 139, "loss": 3.0474, "learning_rate": 9.798515376458112e-07, "epoch": 0.21148036253776434 }, { "current_steps": 140, "loss": 3.218, "learning_rate": 9.793213149522799e-07, "epoch": 0.21299093655589124 }, { "current_steps": 141, "loss": 2.9356, "learning_rate": 9.787910922587486e-07, "epoch": 0.21450151057401812 }, { "current_steps": 142, "loss": 3.1152, "learning_rate": 9.782608695652173e-07, "epoch": 0.21601208459214502 }, { "current_steps": 143, "loss": 2.902, "learning_rate": 9.77730646871686e-07, "epoch": 0.2175226586102719 }, { "current_steps": 144, "loss": 3.3406, "learning_rate": 9.772004241781547e-07, "epoch": 0.2190332326283988 }, { "current_steps": 145, "loss": 3.0509, "learning_rate": 9.766702014846235e-07, "epoch": 0.22054380664652568 }, { "current_steps": 146, "loss": 3.2664, "learning_rate": 9.761399787910922e-07, "epoch": 0.22205438066465258 }, { "current_steps": 147, "loss": 3.2172, "learning_rate": 9.756097560975609e-07, "epoch": 0.22356495468277945 }, { "current_steps": 148, "loss": 3.2731, "learning_rate": 9.750795334040296e-07, "epoch": 0.22507552870090636 }, { "current_steps": 149, "loss": 3.1866, "learning_rate": 9.745493107104983e-07, "epoch": 0.22658610271903323 }, { "current_steps": 150, "loss": 3.2318, "learning_rate": 9.74019088016967e-07, "epoch": 0.2280966767371601 }, { "current_steps": 151, "loss": 3.2205, "learning_rate": 9.734888653234357e-07, "epoch": 0.229607250755287 }, { "current_steps": 152, "loss": 3.2561, "learning_rate": 9.729586426299045e-07, "epoch": 0.2311178247734139 }, { "current_steps": 153, "loss": 3.2357, "learning_rate": 9.724284199363732e-07, "epoch": 0.2326283987915408 }, { "current_steps": 154, "loss": 3.1097, "learning_rate": 9.71898197242842e-07, "epoch": 0.23413897280966767 }, { "current_steps": 155, "loss": 3.0489, "learning_rate": 9.713679745493106e-07, "epoch": 0.23564954682779457 }, { "current_steps": 156, "loss": 3.1937, "learning_rate": 9.708377518557793e-07, "epoch": 0.23716012084592145 }, { "current_steps": 157, "loss": 3.1377, "learning_rate": 9.70307529162248e-07, "epoch": 0.23867069486404835 }, { "current_steps": 158, "loss": 3.1246, "learning_rate": 9.697773064687168e-07, "epoch": 0.24018126888217523 }, { "current_steps": 159, "loss": 2.9715, "learning_rate": 9.692470837751855e-07, "epoch": 0.24169184290030213 }, { "current_steps": 160, "loss": 3.3301, "learning_rate": 9.687168610816542e-07, "epoch": 0.243202416918429 }, { "current_steps": 161, "loss": 3.175, "learning_rate": 9.68186638388123e-07, "epoch": 0.24471299093655588 }, { "current_steps": 162, "loss": 3.5311, "learning_rate": 9.676564156945916e-07, "epoch": 0.24622356495468278 }, { "current_steps": 163, "loss": 2.9409, "learning_rate": 9.671261930010603e-07, "epoch": 0.24773413897280966 }, { "current_steps": 164, "loss": 3.1624, "learning_rate": 9.66595970307529e-07, "epoch": 0.24924471299093656 }, { "current_steps": 165, "loss": 3.2766, "learning_rate": 9.660657476139978e-07, "epoch": 0.25075528700906347 }, { "current_steps": 166, "loss": 3.1353, "learning_rate": 9.655355249204665e-07, "epoch": 0.25226586102719034 }, { "current_steps": 167, "loss": 3.1441, "learning_rate": 9.650053022269352e-07, "epoch": 0.2537764350453172 }, { "current_steps": 168, "loss": 2.5206, "learning_rate": 9.64475079533404e-07, "epoch": 0.2552870090634441 }, { "current_steps": 169, "loss": 3.3392, "learning_rate": 9.639448568398726e-07, "epoch": 0.256797583081571 }, { "current_steps": 170, "loss": 3.3648, "learning_rate": 9.634146341463414e-07, "epoch": 0.2583081570996979 }, { "current_steps": 171, "loss": 2.9235, "learning_rate": 9.628844114528103e-07, "epoch": 0.2598187311178248 }, { "current_steps": 172, "loss": 3.2705, "learning_rate": 9.623541887592788e-07, "epoch": 0.26132930513595165 }, { "current_steps": 173, "loss": 3.0963, "learning_rate": 9.618239660657475e-07, "epoch": 0.2628398791540785 }, { "current_steps": 174, "loss": 3.0412, "learning_rate": 9.612937433722162e-07, "epoch": 0.26435045317220546 }, { "current_steps": 175, "loss": 3.0035, "learning_rate": 9.607635206786851e-07, "epoch": 0.26586102719033233 }, { "current_steps": 176, "loss": 2.977, "learning_rate": 9.602332979851536e-07, "epoch": 0.2673716012084592 }, { "current_steps": 177, "loss": 3.1074, "learning_rate": 9.597030752916224e-07, "epoch": 0.2688821752265861 }, { "current_steps": 178, "loss": 3.247, "learning_rate": 9.59172852598091e-07, "epoch": 0.270392749244713 }, { "current_steps": 179, "loss": 3.1697, "learning_rate": 9.5864262990456e-07, "epoch": 0.2719033232628399 }, { "current_steps": 180, "loss": 3.3536, "learning_rate": 9.581124072110285e-07, "epoch": 0.27341389728096677 }, { "current_steps": 181, "loss": 3.1507, "learning_rate": 9.575821845174972e-07, "epoch": 0.27492447129909364 }, { "current_steps": 182, "loss": 3.3054, "learning_rate": 9.57051961823966e-07, "epoch": 0.2764350453172205 }, { "current_steps": 183, "loss": 3.1328, "learning_rate": 9.565217391304349e-07, "epoch": 0.27794561933534745 }, { "current_steps": 184, "loss": 3.2001, "learning_rate": 9.559915164369034e-07, "epoch": 0.2794561933534743 }, { "current_steps": 185, "loss": 2.9704, "learning_rate": 9.55461293743372e-07, "epoch": 0.2809667673716012 }, { "current_steps": 186, "loss": 2.9961, "learning_rate": 9.54931071049841e-07, "epoch": 0.2824773413897281 }, { "current_steps": 187, "loss": 3.2078, "learning_rate": 9.544008483563097e-07, "epoch": 0.283987915407855 }, { "current_steps": 188, "loss": 3.1483, "learning_rate": 9.538706256627782e-07, "epoch": 0.2854984894259819 }, { "current_steps": 189, "loss": 2.4296, "learning_rate": 9.53340402969247e-07, "epoch": 0.28700906344410876 }, { "current_steps": 190, "loss": 3.2875, "learning_rate": 9.528101802757158e-07, "epoch": 0.28851963746223563 }, { "current_steps": 191, "loss": 3.2538, "learning_rate": 9.522799575821845e-07, "epoch": 0.29003021148036257 }, { "current_steps": 192, "loss": 3.0849, "learning_rate": 9.517497348886532e-07, "epoch": 0.29154078549848944 }, { "current_steps": 193, "loss": 3.3613, "learning_rate": 9.512195121951218e-07, "epoch": 0.2930513595166163 }, { "current_steps": 194, "loss": 3.1343, "learning_rate": 9.506892895015906e-07, "epoch": 0.2945619335347432 }, { "current_steps": 195, "loss": 3.3011, "learning_rate": 9.501590668080594e-07, "epoch": 0.29607250755287007 }, { "current_steps": 196, "loss": 3.2206, "learning_rate": 9.496288441145281e-07, "epoch": 0.297583081570997 }, { "current_steps": 197, "loss": 2.3201, "learning_rate": 9.490986214209967e-07, "epoch": 0.2990936555891239 }, { "current_steps": 198, "loss": 3.1189, "learning_rate": 9.485683987274655e-07, "epoch": 0.30060422960725075 }, { "current_steps": 199, "loss": 3.1607, "learning_rate": 9.480381760339342e-07, "epoch": 0.3021148036253776 }, { "current_steps": 200, "loss": 3.2595, "learning_rate": 9.475079533404029e-07, "epoch": 0.30362537764350456 }, { "current_steps": 201, "loss": 3.3786, "learning_rate": 9.469777306468716e-07, "epoch": 0.30513595166163143 }, { "current_steps": 202, "loss": 3.0444, "learning_rate": 9.464475079533404e-07, "epoch": 0.3066465256797583 }, { "current_steps": 203, "loss": 3.261, "learning_rate": 9.459172852598091e-07, "epoch": 0.3081570996978852 }, { "current_steps": 204, "loss": 3.3189, "learning_rate": 9.453870625662778e-07, "epoch": 0.30966767371601206 }, { "current_steps": 205, "loss": 3.4952, "learning_rate": 9.448568398727465e-07, "epoch": 0.311178247734139 }, { "current_steps": 206, "loss": 3.0673, "learning_rate": 9.443266171792152e-07, "epoch": 0.31268882175226587 }, { "current_steps": 207, "loss": 3.2579, "learning_rate": 9.437963944856839e-07, "epoch": 0.31419939577039274 }, { "current_steps": 208, "loss": 3.0032, "learning_rate": 9.432661717921527e-07, "epoch": 0.3157099697885196 }, { "current_steps": 209, "loss": 3.2322, "learning_rate": 9.427359490986214e-07, "epoch": 0.31722054380664655 }, { "current_steps": 210, "loss": 3.0372, "learning_rate": 9.422057264050901e-07, "epoch": 0.3187311178247734 }, { "current_steps": 211, "loss": 3.1053, "learning_rate": 9.416755037115588e-07, "epoch": 0.3202416918429003 }, { "current_steps": 212, "loss": 3.048, "learning_rate": 9.411452810180275e-07, "epoch": 0.3217522658610272 }, { "current_steps": 213, "loss": 3.0653, "learning_rate": 9.406150583244962e-07, "epoch": 0.32326283987915405 }, { "current_steps": 214, "loss": 2.9183, "learning_rate": 9.40084835630965e-07, "epoch": 0.324773413897281 }, { "current_steps": 215, "loss": 3.0895, "learning_rate": 9.395546129374337e-07, "epoch": 0.32628398791540786 }, { "current_steps": 216, "loss": 2.6947, "learning_rate": 9.390243902439024e-07, "epoch": 0.32779456193353473 }, { "current_steps": 217, "loss": 3.0717, "learning_rate": 9.384941675503711e-07, "epoch": 0.3293051359516616 }, { "current_steps": 218, "loss": 2.9181, "learning_rate": 9.379639448568398e-07, "epoch": 0.33081570996978854 }, { "current_steps": 219, "loss": 3.0709, "learning_rate": 9.374337221633085e-07, "epoch": 0.3323262839879154 }, { "current_steps": 220, "loss": 2.9505, "learning_rate": 9.369034994697774e-07, "epoch": 0.3338368580060423 }, { "current_steps": 221, "loss": 2.8928, "learning_rate": 9.36373276776246e-07, "epoch": 0.33534743202416917 }, { "current_steps": 222, "loss": 3.1669, "learning_rate": 9.358430540827147e-07, "epoch": 0.3368580060422961 }, { "current_steps": 223, "loss": 3.1317, "learning_rate": 9.353128313891834e-07, "epoch": 0.338368580060423 }, { "current_steps": 224, "loss": 3.1631, "learning_rate": 9.347826086956522e-07, "epoch": 0.33987915407854985 }, { "current_steps": 225, "loss": 2.6912, "learning_rate": 9.342523860021208e-07, "epoch": 0.3413897280966767 }, { "current_steps": 226, "loss": 3.0521, "learning_rate": 9.337221633085895e-07, "epoch": 0.3429003021148036 }, { "current_steps": 227, "loss": 3.3139, "learning_rate": 9.331919406150583e-07, "epoch": 0.34441087613293053 }, { "current_steps": 228, "loss": 3.0466, "learning_rate": 9.326617179215271e-07, "epoch": 0.3459214501510574 }, { "current_steps": 229, "loss": 3.2494, "learning_rate": 9.321314952279957e-07, "epoch": 0.3474320241691843 }, { "current_steps": 230, "loss": 3.2022, "learning_rate": 9.316012725344644e-07, "epoch": 0.34894259818731116 }, { "current_steps": 231, "loss": 3.1887, "learning_rate": 9.310710498409331e-07, "epoch": 0.3504531722054381 }, { "current_steps": 232, "loss": 3.2125, "learning_rate": 9.305408271474019e-07, "epoch": 0.35196374622356497 }, { "current_steps": 233, "loss": 3.0106, "learning_rate": 9.300106044538706e-07, "epoch": 0.35347432024169184 }, { "current_steps": 234, "loss": 3.1587, "learning_rate": 9.294803817603393e-07, "epoch": 0.3549848942598187 }, { "current_steps": 235, "loss": 3.0159, "learning_rate": 9.289501590668081e-07, "epoch": 0.3564954682779456 }, { "current_steps": 236, "loss": 2.9888, "learning_rate": 9.284199363732768e-07, "epoch": 0.3580060422960725 }, { "current_steps": 237, "loss": 3.3355, "learning_rate": 9.278897136797454e-07, "epoch": 0.3595166163141994 }, { "current_steps": 238, "loss": 3.2464, "learning_rate": 9.273594909862141e-07, "epoch": 0.3610271903323263 }, { "current_steps": 239, "loss": 3.1984, "learning_rate": 9.26829268292683e-07, "epoch": 0.36253776435045315 }, { "current_steps": 240, "loss": 3.1774, "learning_rate": 9.262990455991516e-07, "epoch": 0.3640483383685801 }, { "current_steps": 241, "loss": 3.1891, "learning_rate": 9.257688229056203e-07, "epoch": 0.36555891238670696 }, { "current_steps": 242, "loss": 3.2667, "learning_rate": 9.25238600212089e-07, "epoch": 0.36706948640483383 }, { "current_steps": 243, "loss": 3.0815, "learning_rate": 9.25238600212089e-07, "epoch": 0.3685800604229607 }, { "current_steps": 244, "loss": 2.9868, "learning_rate": 9.247083775185578e-07, "epoch": 0.37009063444108764 }, { "current_steps": 245, "loss": 3.2655, "learning_rate": 9.241781548250264e-07, "epoch": 0.3716012084592145 }, { "current_steps": 246, "loss": 3.2307, "learning_rate": 9.236479321314951e-07, "epoch": 0.3731117824773414 }, { "current_steps": 247, "loss": 3.2197, "learning_rate": 9.231177094379639e-07, "epoch": 0.37462235649546827 }, { "current_steps": 248, "loss": 3.2812, "learning_rate": 9.225874867444327e-07, "epoch": 0.37613293051359514 }, { "current_steps": 249, "loss": 3.0202, "learning_rate": 9.220572640509013e-07, "epoch": 0.3776435045317221 }, { "current_steps": 250, "loss": 2.7464, "learning_rate": 9.2152704135737e-07, "epoch": 0.37915407854984895 }, { "current_steps": 251, "loss": 3.0009, "learning_rate": 9.209968186638388e-07, "epoch": 0.3806646525679758 }, { "current_steps": 252, "loss": 2.9182, "learning_rate": 9.204665959703075e-07, "epoch": 0.3821752265861027 }, { "current_steps": 253, "loss": 3.0978, "learning_rate": 9.199363732767762e-07, "epoch": 0.38368580060422963 }, { "current_steps": 254, "loss": 2.5917, "learning_rate": 9.194061505832449e-07, "epoch": 0.3851963746223565 }, { "current_steps": 255, "loss": 3.2166, "learning_rate": 9.188759278897137e-07, "epoch": 0.3867069486404834 }, { "current_steps": 256, "loss": 3.049, "learning_rate": 9.183457051961824e-07, "epoch": 0.38821752265861026 }, { "current_steps": 257, "loss": 2.9817, "learning_rate": 9.17815482502651e-07, "epoch": 0.38972809667673713 }, { "current_steps": 258, "loss": 3.3222, "learning_rate": 9.172852598091197e-07, "epoch": 0.39123867069486407 }, { "current_steps": 259, "loss": 3.1317, "learning_rate": 9.167550371155886e-07, "epoch": 0.39274924471299094 }, { "current_steps": 260, "loss": 2.9371, "learning_rate": 9.162248144220573e-07, "epoch": 0.3942598187311178 }, { "current_steps": 261, "loss": 3.1652, "learning_rate": 9.156945917285259e-07, "epoch": 0.3957703927492447 }, { "current_steps": 262, "loss": 3.1902, "learning_rate": 9.151643690349946e-07, "epoch": 0.3972809667673716 }, { "current_steps": 263, "loss": 3.1463, "learning_rate": 9.146341463414634e-07, "epoch": 0.3987915407854985 }, { "current_steps": 264, "loss": 2.9862, "learning_rate": 9.141039236479321e-07, "epoch": 0.4003021148036254 }, { "current_steps": 265, "loss": 3.3495, "learning_rate": 9.135737009544007e-07, "epoch": 0.40181268882175225 }, { "current_steps": 266, "loss": 2.5696, "learning_rate": 9.130434782608695e-07, "epoch": 0.4033232628398791 }, { "current_steps": 267, "loss": 2.9964, "learning_rate": 9.125132555673383e-07, "epoch": 0.40483383685800606 }, { "current_steps": 268, "loss": 2.6063, "learning_rate": 9.11983032873807e-07, "epoch": 0.40634441087613293 }, { "current_steps": 269, "loss": 3.5285, "learning_rate": 9.114528101802756e-07, "epoch": 0.4078549848942598 }, { "current_steps": 270, "loss": 3.3055, "learning_rate": 9.109225874867444e-07, "epoch": 0.4093655589123867 }, { "current_steps": 271, "loss": 3.1632, "learning_rate": 9.103923647932131e-07, "epoch": 0.4108761329305136 }, { "current_steps": 272, "loss": 3.0509, "learning_rate": 9.098621420996819e-07, "epoch": 0.4123867069486405 }, { "current_steps": 273, "loss": 3.1983, "learning_rate": 9.093319194061505e-07, "epoch": 0.41389728096676737 }, { "current_steps": 274, "loss": 3.1477, "learning_rate": 9.088016967126193e-07, "epoch": 0.41540785498489424 }, { "current_steps": 275, "loss": 2.8309, "learning_rate": 9.08271474019088e-07, "epoch": 0.4169184290030212 }, { "current_steps": 276, "loss": 2.5193, "learning_rate": 9.077412513255567e-07, "epoch": 0.41842900302114805 }, { "current_steps": 277, "loss": 2.6047, "learning_rate": 9.072110286320253e-07, "epoch": 0.4199395770392749 }, { "current_steps": 278, "loss": 3.1801, "learning_rate": 9.066808059384942e-07, "epoch": 0.4214501510574018 }, { "current_steps": 279, "loss": 3.2664, "learning_rate": 9.061505832449629e-07, "epoch": 0.4229607250755287 }, { "current_steps": 280, "loss": 3.1992, "learning_rate": 9.056203605514315e-07, "epoch": 0.4244712990936556 }, { "current_steps": 281, "loss": 3.003, "learning_rate": 9.050901378579002e-07, "epoch": 0.4259818731117825 }, { "current_steps": 282, "loss": 3.1076, "learning_rate": 9.04559915164369e-07, "epoch": 0.42749244712990936 }, { "current_steps": 283, "loss": 3.032, "learning_rate": 9.040296924708377e-07, "epoch": 0.42900302114803623 }, { "current_steps": 284, "loss": 3.3778, "learning_rate": 9.034994697773063e-07, "epoch": 0.43051359516616317 }, { "current_steps": 285, "loss": 3.0228, "learning_rate": 9.029692470837752e-07, "epoch": 0.43202416918429004 }, { "current_steps": 286, "loss": 3.1486, "learning_rate": 9.024390243902439e-07, "epoch": 0.4335347432024169 }, { "current_steps": 287, "loss": 2.5894, "learning_rate": 9.019088016967126e-07, "epoch": 0.4350453172205438 }, { "current_steps": 288, "loss": 3.0163, "learning_rate": 9.013785790031812e-07, "epoch": 0.43655589123867067 }, { "current_steps": 289, "loss": 2.8523, "learning_rate": 9.0084835630965e-07, "epoch": 0.4380664652567976 }, { "current_steps": 290, "loss": 2.9167, "learning_rate": 9.003181336161188e-07, "epoch": 0.4395770392749245 }, { "current_steps": 291, "loss": 3.3304, "learning_rate": 8.997879109225875e-07, "epoch": 0.44108761329305135 }, { "current_steps": 292, "loss": 2.8962, "learning_rate": 8.992576882290561e-07, "epoch": 0.4425981873111782 }, { "current_steps": 293, "loss": 3.2467, "learning_rate": 8.987274655355249e-07, "epoch": 0.44410876132930516 }, { "current_steps": 294, "loss": 3.0529, "learning_rate": 8.981972428419936e-07, "epoch": 0.44561933534743203 }, { "current_steps": 295, "loss": 3.0481, "learning_rate": 8.976670201484623e-07, "epoch": 0.4471299093655589 }, { "current_steps": 296, "loss": 3.081, "learning_rate": 8.971367974549309e-07, "epoch": 0.4486404833836858 }, { "current_steps": 297, "loss": 3.2012, "learning_rate": 8.966065747613998e-07, "epoch": 0.4501510574018127 }, { "current_steps": 298, "loss": 3.2297, "learning_rate": 8.960763520678685e-07, "epoch": 0.4516616314199396 }, { "current_steps": 299, "loss": 2.7969, "learning_rate": 8.955461293743372e-07, "epoch": 0.45317220543806647 }, { "current_steps": 300, "loss": 3.2338, "learning_rate": 8.950159066808058e-07, "epoch": 0.45468277945619334 }, { "current_steps": 301, "loss": 3.5633, "learning_rate": 8.944856839872746e-07, "epoch": 0.4561933534743202 }, { "current_steps": 302, "loss": 3.2321, "learning_rate": 8.939554612937433e-07, "epoch": 0.45770392749244715 }, { "current_steps": 303, "loss": 3.2112, "learning_rate": 8.934252386002121e-07, "epoch": 0.459214501510574 }, { "current_steps": 304, "loss": 3.13, "learning_rate": 8.928950159066808e-07, "epoch": 0.4607250755287009 }, { "current_steps": 305, "loss": 3.1433, "learning_rate": 8.923647932131495e-07, "epoch": 0.4622356495468278 }, { "current_steps": 306, "loss": 3.2684, "learning_rate": 8.918345705196182e-07, "epoch": 0.4637462235649547 }, { "current_steps": 307, "loss": 3.0937, "learning_rate": 8.913043478260869e-07, "epoch": 0.4652567975830816 }, { "current_steps": 308, "loss": 3.3234, "learning_rate": 8.907741251325556e-07, "epoch": 0.46676737160120846 }, { "current_steps": 309, "loss": 3.3496, "learning_rate": 8.902439024390244e-07, "epoch": 0.46827794561933533 }, { "current_steps": 310, "loss": 3.1118, "learning_rate": 8.897136797454931e-07, "epoch": 0.4697885196374622 }, { "current_steps": 311, "loss": 3.3059, "learning_rate": 8.891834570519618e-07, "epoch": 0.47129909365558914 }, { "current_steps": 312, "loss": 2.9695, "learning_rate": 8.886532343584305e-07, "epoch": 0.472809667673716 }, { "current_steps": 313, "loss": 3.2671, "learning_rate": 8.881230116648992e-07, "epoch": 0.4743202416918429 }, { "current_steps": 314, "loss": 3.0554, "learning_rate": 8.875927889713679e-07, "epoch": 0.47583081570996977 }, { "current_steps": 315, "loss": 3.2973, "learning_rate": 8.870625662778366e-07, "epoch": 0.4773413897280967 }, { "current_steps": 316, "loss": 3.2223, "learning_rate": 8.865323435843054e-07, "epoch": 0.4788519637462236 }, { "current_steps": 317, "loss": 3.4084, "learning_rate": 8.860021208907741e-07, "epoch": 0.48036253776435045 }, { "current_steps": 318, "loss": 3.0606, "learning_rate": 8.854718981972428e-07, "epoch": 0.4818731117824773 }, { "current_steps": 319, "loss": 3.3439, "learning_rate": 8.849416755037116e-07, "epoch": 0.48338368580060426 }, { "current_steps": 320, "loss": 3.2179, "learning_rate": 8.844114528101802e-07, "epoch": 0.48489425981873113 }, { "current_steps": 321, "loss": 3.1217, "learning_rate": 8.838812301166489e-07, "epoch": 0.486404833836858 }, { "current_steps": 322, "loss": 3.1913, "learning_rate": 8.833510074231177e-07, "epoch": 0.4879154078549849 }, { "current_steps": 323, "loss": 3.0094, "learning_rate": 8.828207847295865e-07, "epoch": 0.48942598187311176 }, { "current_steps": 324, "loss": 2.7767, "learning_rate": 8.822905620360551e-07, "epoch": 0.4909365558912387 }, { "current_steps": 325, "loss": 3.084, "learning_rate": 8.817603393425238e-07, "epoch": 0.49244712990936557 }, { "current_steps": 326, "loss": 2.9481, "learning_rate": 8.812301166489925e-07, "epoch": 0.49395770392749244 }, { "current_steps": 327, "loss": 3.0572, "learning_rate": 8.806998939554613e-07, "epoch": 0.4954682779456193 }, { "current_steps": 328, "loss": 3.0843, "learning_rate": 8.8016967126193e-07, "epoch": 0.49697885196374625 }, { "current_steps": 329, "loss": 2.938, "learning_rate": 8.796394485683987e-07, "epoch": 0.4984894259818731 }, { "current_steps": 330, "loss": 2.9294, "learning_rate": 8.791092258748674e-07, "epoch": 0.5 }, { "current_steps": 331, "loss": 2.9109, "learning_rate": 8.785790031813361e-07, "epoch": 0.5015105740181269 }, { "current_steps": 332, "loss": 2.9943, "learning_rate": 8.780487804878048e-07, "epoch": 0.5030211480362538 }, { "current_steps": 333, "loss": 2.7782, "learning_rate": 8.775185577942735e-07, "epoch": 0.5045317220543807 }, { "current_steps": 334, "loss": 3.1677, "learning_rate": 8.769883351007424e-07, "epoch": 0.5060422960725075 }, { "current_steps": 335, "loss": 2.9955, "learning_rate": 8.76458112407211e-07, "epoch": 0.5075528700906344 }, { "current_steps": 336, "loss": 3.3114, "learning_rate": 8.759278897136797e-07, "epoch": 0.5090634441087614 }, { "current_steps": 337, "loss": 3.2823, "learning_rate": 8.753976670201484e-07, "epoch": 0.5105740181268882 }, { "current_steps": 338, "loss": 2.7996, "learning_rate": 8.748674443266172e-07, "epoch": 0.5120845921450151 }, { "current_steps": 339, "loss": 3.1617, "learning_rate": 8.743372216330858e-07, "epoch": 0.513595166163142 }, { "current_steps": 340, "loss": 3.1676, "learning_rate": 8.738069989395545e-07, "epoch": 0.5151057401812689 }, { "current_steps": 341, "loss": 2.4332, "learning_rate": 8.732767762460233e-07, "epoch": 0.5166163141993958 }, { "current_steps": 342, "loss": 3.1305, "learning_rate": 8.727465535524921e-07, "epoch": 0.5181268882175226 }, { "current_steps": 343, "loss": 3.1859, "learning_rate": 8.722163308589607e-07, "epoch": 0.5196374622356495 }, { "current_steps": 344, "loss": 3.0989, "learning_rate": 8.716861081654294e-07, "epoch": 0.5211480362537765 }, { "current_steps": 345, "loss": 2.9217, "learning_rate": 8.711558854718981e-07, "epoch": 0.5226586102719033 }, { "current_steps": 346, "loss": 3.0529, "learning_rate": 8.706256627783669e-07, "epoch": 0.5241691842900302 }, { "current_steps": 347, "loss": 2.9959, "learning_rate": 8.700954400848356e-07, "epoch": 0.525679758308157 }, { "current_steps": 348, "loss": 3.6103, "learning_rate": 8.695652173913043e-07, "epoch": 0.527190332326284 }, { "current_steps": 349, "loss": 3.1724, "learning_rate": 8.69034994697773e-07, "epoch": 0.5287009063444109 }, { "current_steps": 350, "loss": 3.1836, "learning_rate": 8.685047720042418e-07, "epoch": 0.5302114803625377 }, { "current_steps": 351, "loss": 3.2854, "learning_rate": 8.679745493107104e-07, "epoch": 0.5317220543806647 }, { "current_steps": 352, "loss": 3.1328, "learning_rate": 8.674443266171791e-07, "epoch": 0.5332326283987915 }, { "current_steps": 353, "loss": 2.0668, "learning_rate": 8.66914103923648e-07, "epoch": 0.5347432024169184 }, { "current_steps": 354, "loss": 3.0729, "learning_rate": 8.663838812301167e-07, "epoch": 0.5362537764350453 }, { "current_steps": 355, "loss": 3.1418, "learning_rate": 8.658536585365853e-07, "epoch": 0.5377643504531722 }, { "current_steps": 356, "loss": 3.1024, "learning_rate": 8.65323435843054e-07, "epoch": 0.5392749244712991 }, { "current_steps": 357, "loss": 3.0899, "learning_rate": 8.647932131495228e-07, "epoch": 0.540785498489426 }, { "current_steps": 358, "loss": 3.1664, "learning_rate": 8.642629904559915e-07, "epoch": 0.5422960725075529 }, { "current_steps": 359, "loss": 3.0295, "learning_rate": 8.637327677624601e-07, "epoch": 0.5438066465256798 }, { "current_steps": 360, "loss": 3.1117, "learning_rate": 8.632025450689289e-07, "epoch": 0.5453172205438066 }, { "current_steps": 361, "loss": 3.2985, "learning_rate": 8.626723223753977e-07, "epoch": 0.5468277945619335 }, { "current_steps": 362, "loss": 3.1515, "learning_rate": 8.621420996818664e-07, "epoch": 0.5483383685800605 }, { "current_steps": 363, "loss": 2.7235, "learning_rate": 8.61611876988335e-07, "epoch": 0.5498489425981873 }, { "current_steps": 364, "loss": 3.1985, "learning_rate": 8.610816542948037e-07, "epoch": 0.5513595166163142 }, { "current_steps": 365, "loss": 2.8398, "learning_rate": 8.605514316012725e-07, "epoch": 0.552870090634441 }, { "current_steps": 366, "loss": 3.0781, "learning_rate": 8.600212089077413e-07, "epoch": 0.554380664652568 }, { "current_steps": 367, "loss": 2.8402, "learning_rate": 8.594909862142099e-07, "epoch": 0.5558912386706949 }, { "current_steps": 368, "loss": 2.8804, "learning_rate": 8.589607635206787e-07, "epoch": 0.5574018126888217 }, { "current_steps": 369, "loss": 3.1703, "learning_rate": 8.584305408271474e-07, "epoch": 0.5589123867069486 }, { "current_steps": 370, "loss": 3.1955, "learning_rate": 8.579003181336161e-07, "epoch": 0.5604229607250756 }, { "current_steps": 371, "loss": 3.1886, "learning_rate": 8.573700954400847e-07, "epoch": 0.5619335347432024 }, { "current_steps": 372, "loss": 3.0821, "learning_rate": 8.568398727465536e-07, "epoch": 0.5634441087613293 }, { "current_steps": 373, "loss": 2.7845, "learning_rate": 8.563096500530223e-07, "epoch": 0.5649546827794562 }, { "current_steps": 374, "loss": 3.1924, "learning_rate": 8.557794273594909e-07, "epoch": 0.5664652567975831 }, { "current_steps": 375, "loss": 3.2262, "learning_rate": 8.552492046659596e-07, "epoch": 0.56797583081571 }, { "current_steps": 376, "loss": 3.0846, "learning_rate": 8.547189819724284e-07, "epoch": 0.5694864048338368 }, { "current_steps": 377, "loss": 3.2293, "learning_rate": 8.541887592788971e-07, "epoch": 0.5709969788519638 }, { "current_steps": 378, "loss": 2.9524, "learning_rate": 8.536585365853657e-07, "epoch": 0.5725075528700906 }, { "current_steps": 379, "loss": 3.2021, "learning_rate": 8.531283138918345e-07, "epoch": 0.5740181268882175 }, { "current_steps": 380, "loss": 3.255, "learning_rate": 8.525980911983033e-07, "epoch": 0.5755287009063444 }, { "current_steps": 381, "loss": 3.151, "learning_rate": 8.52067868504772e-07, "epoch": 0.5770392749244713 }, { "current_steps": 382, "loss": 3.3928, "learning_rate": 8.515376458112406e-07, "epoch": 0.5785498489425982 }, { "current_steps": 383, "loss": 3.0898, "learning_rate": 8.510074231177093e-07, "epoch": 0.5800604229607251 }, { "current_steps": 384, "loss": 3.045, "learning_rate": 8.504772004241781e-07, "epoch": 0.581570996978852 }, { "current_steps": 385, "loss": 2.9802, "learning_rate": 8.499469777306469e-07, "epoch": 0.5830815709969789 }, { "current_steps": 386, "loss": 3.301, "learning_rate": 8.494167550371155e-07, "epoch": 0.5845921450151057 }, { "current_steps": 387, "loss": 2.794, "learning_rate": 8.488865323435843e-07, "epoch": 0.5861027190332326 }, { "current_steps": 388, "loss": 3.3134, "learning_rate": 8.48356309650053e-07, "epoch": 0.5876132930513596 }, { "current_steps": 389, "loss": 3.1513, "learning_rate": 8.478260869565217e-07, "epoch": 0.5891238670694864 }, { "current_steps": 390, "loss": 3.1984, "learning_rate": 8.472958642629903e-07, "epoch": 0.5906344410876133 }, { "current_steps": 391, "loss": 2.9866, "learning_rate": 8.467656415694592e-07, "epoch": 0.5921450151057401 }, { "current_steps": 392, "loss": 3.0448, "learning_rate": 8.462354188759279e-07, "epoch": 0.5936555891238671 }, { "current_steps": 393, "loss": 3.0681, "learning_rate": 8.457051961823966e-07, "epoch": 0.595166163141994 }, { "current_steps": 394, "loss": 3.1794, "learning_rate": 8.451749734888652e-07, "epoch": 0.5966767371601208 }, { "current_steps": 395, "loss": 3.2612, "learning_rate": 8.44644750795334e-07, "epoch": 0.5981873111782477 }, { "current_steps": 396, "loss": 3.2551, "learning_rate": 8.441145281018027e-07, "epoch": 0.5996978851963746 }, { "current_steps": 397, "loss": 2.9567, "learning_rate": 8.435843054082715e-07, "epoch": 0.6012084592145015 }, { "current_steps": 398, "loss": 3.0785, "learning_rate": 8.430540827147401e-07, "epoch": 0.6027190332326284 }, { "current_steps": 399, "loss": 2.9135, "learning_rate": 8.425238600212089e-07, "epoch": 0.6042296072507553 }, { "current_steps": 400, "loss": 3.1097, "learning_rate": 8.419936373276776e-07, "epoch": 0.6057401812688822 }, { "current_steps": 401, "loss": 3.2021, "learning_rate": 8.414634146341463e-07, "epoch": 0.6072507552870091 }, { "current_steps": 402, "loss": 3.0949, "learning_rate": 8.40933191940615e-07, "epoch": 0.6087613293051359 }, { "current_steps": 403, "loss": 2.9137, "learning_rate": 8.404029692470838e-07, "epoch": 0.6102719033232629 }, { "current_steps": 404, "loss": 3.1532, "learning_rate": 8.398727465535525e-07, "epoch": 0.6117824773413897 }, { "current_steps": 405, "loss": 3.2406, "learning_rate": 8.393425238600212e-07, "epoch": 0.6132930513595166 }, { "current_steps": 406, "loss": 3.0427, "learning_rate": 8.388123011664899e-07, "epoch": 0.6148036253776435 }, { "current_steps": 407, "loss": 3.267, "learning_rate": 8.382820784729586e-07, "epoch": 0.6163141993957704 }, { "current_steps": 408, "loss": 3.0729, "learning_rate": 8.377518557794273e-07, "epoch": 0.6178247734138973 }, { "current_steps": 409, "loss": 3.0987, "learning_rate": 8.37221633085896e-07, "epoch": 0.6193353474320241 }, { "current_steps": 410, "loss": 3.1801, "learning_rate": 8.366914103923648e-07, "epoch": 0.620845921450151 }, { "current_steps": 411, "loss": 3.0017, "learning_rate": 8.361611876988335e-07, "epoch": 0.622356495468278 }, { "current_steps": 412, "loss": 3.3731, "learning_rate": 8.356309650053022e-07, "epoch": 0.6238670694864048 }, { "current_steps": 413, "loss": 2.6755, "learning_rate": 8.351007423117708e-07, "epoch": 0.6253776435045317 }, { "current_steps": 414, "loss": 3.1755, "learning_rate": 8.345705196182396e-07, "epoch": 0.6268882175226587 }, { "current_steps": 415, "loss": 3.1861, "learning_rate": 8.340402969247083e-07, "epoch": 0.6283987915407855 }, { "current_steps": 416, "loss": 3.0512, "learning_rate": 8.335100742311771e-07, "epoch": 0.6299093655589124 }, { "current_steps": 417, "loss": 2.2501, "learning_rate": 8.329798515376459e-07, "epoch": 0.6314199395770392 }, { "current_steps": 418, "loss": 3.1965, "learning_rate": 8.324496288441145e-07, "epoch": 0.6329305135951662 }, { "current_steps": 419, "loss": 3.1893, "learning_rate": 8.319194061505832e-07, "epoch": 0.6344410876132931 }, { "current_steps": 420, "loss": 3.3295, "learning_rate": 8.313891834570519e-07, "epoch": 0.6359516616314199 }, { "current_steps": 421, "loss": 3.0503, "learning_rate": 8.308589607635206e-07, "epoch": 0.6374622356495468 }, { "current_steps": 422, "loss": 3.253, "learning_rate": 8.303287380699894e-07, "epoch": 0.6389728096676737 }, { "current_steps": 423, "loss": 3.3239, "learning_rate": 8.297985153764581e-07, "epoch": 0.6404833836858006 }, { "current_steps": 424, "loss": 3.461, "learning_rate": 8.292682926829268e-07, "epoch": 0.6419939577039275 }, { "current_steps": 425, "loss": 2.8585, "learning_rate": 8.287380699893955e-07, "epoch": 0.6435045317220544 }, { "current_steps": 426, "loss": 3.13, "learning_rate": 8.282078472958642e-07, "epoch": 0.6450151057401813 }, { "current_steps": 427, "loss": 3.1843, "learning_rate": 8.276776246023329e-07, "epoch": 0.6465256797583081 }, { "current_steps": 428, "loss": 3.0517, "learning_rate": 8.271474019088016e-07, "epoch": 0.648036253776435 }, { "current_steps": 429, "loss": 3.1236, "learning_rate": 8.266171792152704e-07, "epoch": 0.649546827794562 }, { "current_steps": 430, "loss": 2.8257, "learning_rate": 8.260869565217391e-07, "epoch": 0.6510574018126888 }, { "current_steps": 431, "loss": 2.9849, "learning_rate": 8.255567338282078e-07, "epoch": 0.6525679758308157 }, { "current_steps": 432, "loss": 3.0392, "learning_rate": 8.250265111346765e-07, "epoch": 0.6540785498489426 }, { "current_steps": 433, "loss": 3.1265, "learning_rate": 8.244962884411452e-07, "epoch": 0.6555891238670695 }, { "current_steps": 434, "loss": 2.9919, "learning_rate": 8.239660657476139e-07, "epoch": 0.6570996978851964 }, { "current_steps": 435, "loss": 2.8665, "learning_rate": 8.234358430540827e-07, "epoch": 0.6586102719033232 }, { "current_steps": 436, "loss": 3.2855, "learning_rate": 8.229056203605515e-07, "epoch": 0.6601208459214502 }, { "current_steps": 437, "loss": 3.2686, "learning_rate": 8.223753976670201e-07, "epoch": 0.6616314199395771 }, { "current_steps": 438, "loss": 3.1965, "learning_rate": 8.218451749734888e-07, "epoch": 0.6631419939577039 }, { "current_steps": 439, "loss": 3.1999, "learning_rate": 8.213149522799575e-07, "epoch": 0.6646525679758308 }, { "current_steps": 440, "loss": 3.0052, "learning_rate": 8.207847295864263e-07, "epoch": 0.6661631419939577 }, { "current_steps": 441, "loss": 3.3237, "learning_rate": 8.20254506892895e-07, "epoch": 0.6676737160120846 }, { "current_steps": 442, "loss": 2.9499, "learning_rate": 8.197242841993637e-07, "epoch": 0.6691842900302115 }, { "current_steps": 443, "loss": 3.0801, "learning_rate": 8.191940615058324e-07, "epoch": 0.6706948640483383 }, { "current_steps": 444, "loss": 3.0574, "learning_rate": 8.186638388123012e-07, "epoch": 0.6722054380664653 }, { "current_steps": 445, "loss": 3.068, "learning_rate": 8.181336161187698e-07, "epoch": 0.6737160120845922 }, { "current_steps": 446, "loss": 3.1562, "learning_rate": 8.176033934252385e-07, "epoch": 0.675226586102719 }, { "current_steps": 447, "loss": 2.9136, "learning_rate": 8.170731707317072e-07, "epoch": 0.676737160120846 }, { "current_steps": 448, "loss": 2.4061, "learning_rate": 8.165429480381761e-07, "epoch": 0.6782477341389728 }, { "current_steps": 449, "loss": 3.3174, "learning_rate": 8.160127253446447e-07, "epoch": 0.6797583081570997 }, { "current_steps": 450, "loss": 2.9948, "learning_rate": 8.154825026511134e-07, "epoch": 0.6812688821752266 }, { "current_steps": 451, "loss": 3.0868, "learning_rate": 8.149522799575822e-07, "epoch": 0.6827794561933535 }, { "current_steps": 452, "loss": 3.0387, "learning_rate": 8.144220572640509e-07, "epoch": 0.6842900302114804 }, { "current_steps": 453, "loss": 3.1065, "learning_rate": 8.138918345705195e-07, "epoch": 0.6858006042296072 }, { "current_steps": 454, "loss": 3.3386, "learning_rate": 8.133616118769883e-07, "epoch": 0.6873111782477341 }, { "current_steps": 455, "loss": 2.8352, "learning_rate": 8.128313891834571e-07, "epoch": 0.6888217522658611 }, { "current_steps": 456, "loss": 2.966, "learning_rate": 8.123011664899258e-07, "epoch": 0.6903323262839879 }, { "current_steps": 457, "loss": 3.1063, "learning_rate": 8.117709437963944e-07, "epoch": 0.6918429003021148 }, { "current_steps": 458, "loss": 3.1141, "learning_rate": 8.112407211028631e-07, "epoch": 0.6933534743202417 }, { "current_steps": 459, "loss": 2.979, "learning_rate": 8.107104984093319e-07, "epoch": 0.6948640483383686 }, { "current_steps": 460, "loss": 3.1482, "learning_rate": 8.101802757158007e-07, "epoch": 0.6963746223564955 }, { "current_steps": 461, "loss": 3.07, "learning_rate": 8.096500530222693e-07, "epoch": 0.6978851963746223 }, { "current_steps": 462, "loss": 2.5385, "learning_rate": 8.09119830328738e-07, "epoch": 0.6993957703927492 }, { "current_steps": 463, "loss": 3.0696, "learning_rate": 8.085896076352068e-07, "epoch": 0.7009063444108762 }, { "current_steps": 464, "loss": 2.9707, "learning_rate": 8.080593849416754e-07, "epoch": 0.702416918429003 }, { "current_steps": 465, "loss": 3.2853, "learning_rate": 8.075291622481441e-07, "epoch": 0.7039274924471299 }, { "current_steps": 466, "loss": 3.1772, "learning_rate": 8.06998939554613e-07, "epoch": 0.7054380664652568 }, { "current_steps": 467, "loss": 2.3754, "learning_rate": 8.064687168610817e-07, "epoch": 0.7069486404833837 }, { "current_steps": 468, "loss": 3.1331, "learning_rate": 8.059384941675503e-07, "epoch": 0.7084592145015106 }, { "current_steps": 469, "loss": 3.2332, "learning_rate": 8.05408271474019e-07, "epoch": 0.7099697885196374 }, { "current_steps": 470, "loss": 3.1364, "learning_rate": 8.048780487804878e-07, "epoch": 0.7114803625377644 }, { "current_steps": 471, "loss": 3.0382, "learning_rate": 8.043478260869565e-07, "epoch": 0.7129909365558912 }, { "current_steps": 472, "loss": 3.0213, "learning_rate": 8.038176033934251e-07, "epoch": 0.7145015105740181 }, { "current_steps": 473, "loss": 3.0728, "learning_rate": 8.032873806998939e-07, "epoch": 0.716012084592145 }, { "current_steps": 474, "loss": 2.8303, "learning_rate": 8.027571580063627e-07, "epoch": 0.7175226586102719 }, { "current_steps": 475, "loss": 2.9884, "learning_rate": 8.022269353128314e-07, "epoch": 0.7190332326283988 }, { "current_steps": 476, "loss": 2.8377, "learning_rate": 8.016967126193e-07, "epoch": 0.7205438066465257 }, { "current_steps": 477, "loss": 3.0181, "learning_rate": 8.011664899257687e-07, "epoch": 0.7220543806646526 }, { "current_steps": 478, "loss": 2.967, "learning_rate": 8.006362672322375e-07, "epoch": 0.7235649546827795 }, { "current_steps": 479, "loss": 3.1362, "learning_rate": 8.001060445387063e-07, "epoch": 0.7250755287009063 }, { "current_steps": 480, "loss": 3.0157, "learning_rate": 7.995758218451749e-07, "epoch": 0.7265861027190332 }, { "current_steps": 481, "loss": 3.2133, "learning_rate": 7.990455991516436e-07, "epoch": 0.7280966767371602 }, { "current_steps": 482, "loss": 3.2363, "learning_rate": 7.985153764581124e-07, "epoch": 0.729607250755287 }, { "current_steps": 483, "loss": 2.9194, "learning_rate": 7.979851537645811e-07, "epoch": 0.7311178247734139 }, { "current_steps": 484, "loss": 3.0577, "learning_rate": 7.974549310710497e-07, "epoch": 0.7326283987915407 }, { "current_steps": 485, "loss": 3.1717, "learning_rate": 7.969247083775186e-07, "epoch": 0.7341389728096677 }, { "current_steps": 486, "loss": 3.2147, "learning_rate": 7.963944856839873e-07, "epoch": 0.7356495468277946 }, { "current_steps": 487, "loss": 3.1182, "learning_rate": 7.95864262990456e-07, "epoch": 0.7371601208459214 }, { "current_steps": 488, "loss": 2.8692, "learning_rate": 7.953340402969246e-07, "epoch": 0.7386706948640483 }, { "current_steps": 489, "loss": 2.9588, "learning_rate": 7.948038176033934e-07, "epoch": 0.7401812688821753 }, { "current_steps": 490, "loss": 2.9832, "learning_rate": 7.942735949098621e-07, "epoch": 0.7416918429003021 }, { "current_steps": 491, "loss": 3.1204, "learning_rate": 7.937433722163309e-07, "epoch": 0.743202416918429 }, { "current_steps": 492, "loss": 3.0391, "learning_rate": 7.932131495227995e-07, "epoch": 0.7447129909365559 }, { "current_steps": 493, "loss": 2.8312, "learning_rate": 7.926829268292683e-07, "epoch": 0.7462235649546828 }, { "current_steps": 494, "loss": 2.8298, "learning_rate": 7.92152704135737e-07, "epoch": 0.7477341389728097 }, { "current_steps": 495, "loss": 2.8368, "learning_rate": 7.916224814422057e-07, "epoch": 0.7492447129909365 }, { "current_steps": 496, "loss": 3.212, "learning_rate": 7.910922587486743e-07, "epoch": 0.7507552870090635 }, { "current_steps": 497, "loss": 3.2808, "learning_rate": 7.905620360551431e-07, "epoch": 0.7522658610271903 }, { "current_steps": 498, "loss": 3.1818, "learning_rate": 7.900318133616119e-07, "epoch": 0.7537764350453172 }, { "current_steps": 499, "loss": 2.937, "learning_rate": 7.895015906680806e-07, "epoch": 0.7552870090634441 }, { "current_steps": 500, "loss": 3.1882, "learning_rate": 7.889713679745493e-07, "epoch": 0.756797583081571 }, { "current_steps": 501, "loss": 3.2112, "learning_rate": 7.88441145281018e-07, "epoch": 0.7583081570996979 }, { "current_steps": 502, "loss": 3.063, "learning_rate": 7.879109225874867e-07, "epoch": 0.7598187311178247 }, { "current_steps": 503, "loss": 2.8865, "learning_rate": 7.873806998939554e-07, "epoch": 0.7613293051359517 }, { "current_steps": 504, "loss": 3.1008, "learning_rate": 7.868504772004242e-07, "epoch": 0.7628398791540786 }, { "current_steps": 505, "loss": 2.8965, "learning_rate": 7.863202545068929e-07, "epoch": 0.7643504531722054 }, { "current_steps": 506, "loss": 2.8766, "learning_rate": 7.857900318133616e-07, "epoch": 0.7658610271903323 }, { "current_steps": 507, "loss": 2.3833, "learning_rate": 7.852598091198302e-07, "epoch": 0.7673716012084593 }, { "current_steps": 508, "loss": 3.1039, "learning_rate": 7.84729586426299e-07, "epoch": 0.7688821752265861 }, { "current_steps": 509, "loss": 2.7846, "learning_rate": 7.841993637327677e-07, "epoch": 0.770392749244713 }, { "current_steps": 510, "loss": 3.0165, "learning_rate": 7.836691410392365e-07, "epoch": 0.7719033232628398 }, { "current_steps": 511, "loss": 3.1822, "learning_rate": 7.831389183457051e-07, "epoch": 0.7734138972809668 }, { "current_steps": 512, "loss": 3.2, "learning_rate": 7.826086956521739e-07, "epoch": 0.7749244712990937 }, { "current_steps": 513, "loss": 3.0028, "learning_rate": 7.820784729586426e-07, "epoch": 0.7764350453172205 }, { "current_steps": 514, "loss": 2.9666, "learning_rate": 7.815482502651113e-07, "epoch": 0.7779456193353474 }, { "current_steps": 515, "loss": 2.7161, "learning_rate": 7.810180275715799e-07, "epoch": 0.7794561933534743 }, { "current_steps": 516, "loss": 2.9436, "learning_rate": 7.804878048780488e-07, "epoch": 0.7809667673716012 }, { "current_steps": 517, "loss": 3.1067, "learning_rate": 7.799575821845175e-07, "epoch": 0.7824773413897281 }, { "current_steps": 518, "loss": 2.9706, "learning_rate": 7.794273594909862e-07, "epoch": 0.783987915407855 }, { "current_steps": 519, "loss": 3.1969, "learning_rate": 7.788971367974549e-07, "epoch": 0.7854984894259819 }, { "current_steps": 520, "loss": 2.5519, "learning_rate": 7.783669141039236e-07, "epoch": 0.7870090634441088 }, { "current_steps": 521, "loss": 3.1815, "learning_rate": 7.778366914103923e-07, "epoch": 0.7885196374622356 }, { "current_steps": 522, "loss": 2.8434, "learning_rate": 7.77306468716861e-07, "epoch": 0.7900302114803626 }, { "current_steps": 523, "loss": 3.1529, "learning_rate": 7.767762460233298e-07, "epoch": 0.7915407854984894 }, { "current_steps": 524, "loss": 2.6969, "learning_rate": 7.762460233297985e-07, "epoch": 0.7930513595166163 }, { "current_steps": 525, "loss": 3.0667, "learning_rate": 7.757158006362672e-07, "epoch": 0.7945619335347432 }, { "current_steps": 526, "loss": 3.2016, "learning_rate": 7.751855779427359e-07, "epoch": 0.7960725075528701 }, { "current_steps": 527, "loss": 3.0715, "learning_rate": 7.746553552492046e-07, "epoch": 0.797583081570997 }, { "current_steps": 528, "loss": 3.1189, "learning_rate": 7.741251325556733e-07, "epoch": 0.7990936555891238 }, { "current_steps": 529, "loss": 3.1317, "learning_rate": 7.735949098621421e-07, "epoch": 0.8006042296072508 }, { "current_steps": 530, "loss": 3.1127, "learning_rate": 7.730646871686108e-07, "epoch": 0.8021148036253777 }, { "current_steps": 531, "loss": 3.3035, "learning_rate": 7.725344644750795e-07, "epoch": 0.8036253776435045 }, { "current_steps": 532, "loss": 3.2779, "learning_rate": 7.720042417815482e-07, "epoch": 0.8051359516616314 }, { "current_steps": 533, "loss": 2.9717, "learning_rate": 7.714740190880169e-07, "epoch": 0.8066465256797583 }, { "current_steps": 534, "loss": 3.1084, "learning_rate": 7.709437963944857e-07, "epoch": 0.8081570996978852 }, { "current_steps": 535, "loss": 3.0763, "learning_rate": 7.704135737009544e-07, "epoch": 0.8096676737160121 }, { "current_steps": 536, "loss": 2.4254, "learning_rate": 7.698833510074231e-07, "epoch": 0.8111782477341389 }, { "current_steps": 537, "loss": 3.2848, "learning_rate": 7.693531283138918e-07, "epoch": 0.8126888217522659 }, { "current_steps": 538, "loss": 2.9759, "learning_rate": 7.688229056203606e-07, "epoch": 0.8141993957703928 }, { "current_steps": 539, "loss": 2.8448, "learning_rate": 7.682926829268292e-07, "epoch": 0.8157099697885196 }, { "current_steps": 540, "loss": 2.8971, "learning_rate": 7.677624602332979e-07, "epoch": 0.8172205438066465 }, { "current_steps": 541, "loss": 3.0558, "learning_rate": 7.672322375397666e-07, "epoch": 0.8187311178247734 }, { "current_steps": 542, "loss": 3.1267, "learning_rate": 7.667020148462355e-07, "epoch": 0.8202416918429003 }, { "current_steps": 543, "loss": 2.6363, "learning_rate": 7.661717921527041e-07, "epoch": 0.8217522658610272 }, { "current_steps": 544, "loss": 2.9708, "learning_rate": 7.656415694591728e-07, "epoch": 0.823262839879154 }, { "current_steps": 545, "loss": 3.2362, "learning_rate": 7.651113467656415e-07, "epoch": 0.824773413897281 }, { "current_steps": 546, "loss": 3.0293, "learning_rate": 7.645811240721103e-07, "epoch": 0.8262839879154078 }, { "current_steps": 547, "loss": 3.1935, "learning_rate": 7.640509013785789e-07, "epoch": 0.8277945619335347 }, { "current_steps": 548, "loss": 3.0003, "learning_rate": 7.635206786850477e-07, "epoch": 0.8293051359516617 }, { "current_steps": 549, "loss": 3.2585, "learning_rate": 7.629904559915165e-07, "epoch": 0.8308157099697885 }, { "current_steps": 550, "loss": 3.1151, "learning_rate": 7.624602332979852e-07, "epoch": 0.8323262839879154 }, { "current_steps": 551, "loss": 2.6131, "learning_rate": 7.619300106044538e-07, "epoch": 0.8338368580060423 }, { "current_steps": 552, "loss": 2.9351, "learning_rate": 7.613997879109225e-07, "epoch": 0.8353474320241692 }, { "current_steps": 553, "loss": 3.0069, "learning_rate": 7.608695652173913e-07, "epoch": 0.8368580060422961 }, { "current_steps": 554, "loss": 2.5323, "learning_rate": 7.6033934252386e-07, "epoch": 0.8383685800604229 }, { "current_steps": 555, "loss": 3.1387, "learning_rate": 7.598091198303287e-07, "epoch": 0.8398791540785498 }, { "current_steps": 556, "loss": 3.0216, "learning_rate": 7.592788971367974e-07, "epoch": 0.8413897280966768 }, { "current_steps": 557, "loss": 3.2079, "learning_rate": 7.587486744432662e-07, "epoch": 0.8429003021148036 }, { "current_steps": 558, "loss": 3.0622, "learning_rate": 7.582184517497348e-07, "epoch": 0.8444108761329305 }, { "current_steps": 559, "loss": 2.4394, "learning_rate": 7.576882290562035e-07, "epoch": 0.8459214501510574 }, { "current_steps": 560, "loss": 3.1407, "learning_rate": 7.571580063626722e-07, "epoch": 0.8474320241691843 }, { "current_steps": 561, "loss": 2.9881, "learning_rate": 7.566277836691411e-07, "epoch": 0.8489425981873112 }, { "current_steps": 562, "loss": 2.9118, "learning_rate": 7.560975609756097e-07, "epoch": 0.850453172205438 }, { "current_steps": 563, "loss": 2.5422, "learning_rate": 7.555673382820784e-07, "epoch": 0.851963746223565 }, { "current_steps": 564, "loss": 3.0906, "learning_rate": 7.550371155885471e-07, "epoch": 0.8534743202416919 }, { "current_steps": 565, "loss": 3.019, "learning_rate": 7.545068928950159e-07, "epoch": 0.8549848942598187 }, { "current_steps": 566, "loss": 3.1581, "learning_rate": 7.539766702014845e-07, "epoch": 0.8564954682779456 }, { "current_steps": 567, "loss": 3.0822, "learning_rate": 7.534464475079533e-07, "epoch": 0.8580060422960725 }, { "current_steps": 568, "loss": 3.1795, "learning_rate": 7.529162248144221e-07, "epoch": 0.8595166163141994 }, { "current_steps": 569, "loss": 3.1357, "learning_rate": 7.523860021208908e-07, "epoch": 0.8610271903323263 }, { "current_steps": 570, "loss": 3.0625, "learning_rate": 7.518557794273594e-07, "epoch": 0.8625377643504532 }, { "current_steps": 571, "loss": 3.013, "learning_rate": 7.513255567338281e-07, "epoch": 0.8640483383685801 }, { "current_steps": 572, "loss": 3.1232, "learning_rate": 7.507953340402969e-07, "epoch": 0.8655589123867069 }, { "current_steps": 573, "loss": 3.0944, "learning_rate": 7.502651113467657e-07, "epoch": 0.8670694864048338 }, { "current_steps": 574, "loss": 3.3296, "learning_rate": 7.497348886532343e-07, "epoch": 0.8685800604229608 }, { "current_steps": 575, "loss": 3.1027, "learning_rate": 7.49204665959703e-07, "epoch": 0.8700906344410876 }, { "current_steps": 576, "loss": 2.9162, "learning_rate": 7.486744432661718e-07, "epoch": 0.8716012084592145 }, { "current_steps": 577, "loss": 3.242, "learning_rate": 7.481442205726405e-07, "epoch": 0.8731117824773413 }, { "current_steps": 578, "loss": 2.8622, "learning_rate": 7.476139978791091e-07, "epoch": 0.8746223564954683 }, { "current_steps": 579, "loss": 3.0473, "learning_rate": 7.470837751855779e-07, "epoch": 0.8761329305135952 }, { "current_steps": 580, "loss": 2.3757, "learning_rate": 7.465535524920467e-07, "epoch": 0.877643504531722 }, { "current_steps": 581, "loss": 3.076, "learning_rate": 7.460233297985154e-07, "epoch": 0.879154078549849 }, { "current_steps": 582, "loss": 3.1744, "learning_rate": 7.45493107104984e-07, "epoch": 0.8806646525679759 }, { "current_steps": 583, "loss": 2.5494, "learning_rate": 7.449628844114528e-07, "epoch": 0.8821752265861027 }, { "current_steps": 584, "loss": 3.0651, "learning_rate": 7.444326617179215e-07, "epoch": 0.8836858006042296 }, { "current_steps": 585, "loss": 3.062, "learning_rate": 7.439024390243903e-07, "epoch": 0.8851963746223565 }, { "current_steps": 586, "loss": 3.0721, "learning_rate": 7.433722163308589e-07, "epoch": 0.8867069486404834 }, { "current_steps": 587, "loss": 2.9384, "learning_rate": 7.428419936373277e-07, "epoch": 0.8882175226586103 }, { "current_steps": 588, "loss": 3.1853, "learning_rate": 7.423117709437964e-07, "epoch": 0.8897280966767371 }, { "current_steps": 589, "loss": 2.8847, "learning_rate": 7.417815482502651e-07, "epoch": 0.8912386706948641 }, { "current_steps": 590, "loss": 3.0281, "learning_rate": 7.412513255567337e-07, "epoch": 0.8927492447129909 }, { "current_steps": 591, "loss": 2.9952, "learning_rate": 7.407211028632025e-07, "epoch": 0.8942598187311178 }, { "current_steps": 592, "loss": 2.9052, "learning_rate": 7.401908801696713e-07, "epoch": 0.8957703927492447 }, { "current_steps": 593, "loss": 3.2017, "learning_rate": 7.3966065747614e-07, "epoch": 0.8972809667673716 }, { "current_steps": 594, "loss": 2.5512, "learning_rate": 7.391304347826086e-07, "epoch": 0.8987915407854985 }, { "current_steps": 595, "loss": 2.5658, "learning_rate": 7.386002120890774e-07, "epoch": 0.9003021148036254 }, { "current_steps": 596, "loss": 3.0919, "learning_rate": 7.380699893955461e-07, "epoch": 0.9018126888217523 }, { "current_steps": 597, "loss": 3.0511, "learning_rate": 7.375397667020147e-07, "epoch": 0.9033232628398792 }, { "current_steps": 598, "loss": 2.8342, "learning_rate": 7.370095440084835e-07, "epoch": 0.904833836858006 }, { "current_steps": 599, "loss": 3.0754, "learning_rate": 7.364793213149523e-07, "epoch": 0.9063444108761329 }, { "current_steps": 600, "loss": 3.1201, "learning_rate": 7.35949098621421e-07, "epoch": 0.9078549848942599 }, { "current_steps": 601, "loss": 2.6998, "learning_rate": 7.354188759278896e-07, "epoch": 0.9093655589123867 }, { "current_steps": 602, "loss": 3.2571, "learning_rate": 7.348886532343584e-07, "epoch": 0.9108761329305136 }, { "current_steps": 603, "loss": 2.898, "learning_rate": 7.343584305408271e-07, "epoch": 0.9123867069486404 }, { "current_steps": 604, "loss": 3.1187, "learning_rate": 7.338282078472959e-07, "epoch": 0.9138972809667674 }, { "current_steps": 605, "loss": 3.0872, "learning_rate": 7.332979851537645e-07, "epoch": 0.9154078549848943 }, { "current_steps": 606, "loss": 3.1769, "learning_rate": 7.327677624602333e-07, "epoch": 0.9169184290030211 }, { "current_steps": 607, "loss": 2.9847, "learning_rate": 7.32237539766702e-07, "epoch": 0.918429003021148 }, { "current_steps": 608, "loss": 3.0472, "learning_rate": 7.317073170731707e-07, "epoch": 0.9199395770392749 }, { "current_steps": 609, "loss": 3.1711, "learning_rate": 7.311770943796393e-07, "epoch": 0.9214501510574018 }, { "current_steps": 610, "loss": 2.9676, "learning_rate": 7.306468716861081e-07, "epoch": 0.9229607250755287 }, { "current_steps": 611, "loss": 3.2103, "learning_rate": 7.301166489925769e-07, "epoch": 0.9244712990936556 }, { "current_steps": 612, "loss": 3.1522, "learning_rate": 7.295864262990456e-07, "epoch": 0.9259818731117825 }, { "current_steps": 613, "loss": 3.0049, "learning_rate": 7.290562036055142e-07, "epoch": 0.9274924471299094 }, { "current_steps": 614, "loss": 3.1888, "learning_rate": 7.28525980911983e-07, "epoch": 0.9290030211480362 }, { "current_steps": 615, "loss": 3.042, "learning_rate": 7.279957582184517e-07, "epoch": 0.9305135951661632 }, { "current_steps": 616, "loss": 3.1997, "learning_rate": 7.274655355249204e-07, "epoch": 0.93202416918429 }, { "current_steps": 617, "loss": 3.084, "learning_rate": 7.269353128313892e-07, "epoch": 0.9335347432024169 }, { "current_steps": 618, "loss": 3.0364, "learning_rate": 7.264050901378579e-07, "epoch": 0.9350453172205438 }, { "current_steps": 619, "loss": 2.993, "learning_rate": 7.258748674443266e-07, "epoch": 0.9365558912386707 }, { "current_steps": 620, "loss": 3.2291, "learning_rate": 7.253446447507953e-07, "epoch": 0.9380664652567976 }, { "current_steps": 621, "loss": 3.1331, "learning_rate": 7.24814422057264e-07, "epoch": 0.9395770392749244 }, { "current_steps": 622, "loss": 3.1535, "learning_rate": 7.242841993637327e-07, "epoch": 0.9410876132930514 }, { "current_steps": 623, "loss": 2.919, "learning_rate": 7.237539766702015e-07, "epoch": 0.9425981873111783 }, { "current_steps": 624, "loss": 3.0084, "learning_rate": 7.232237539766702e-07, "epoch": 0.9441087613293051 }, { "current_steps": 625, "loss": 2.874, "learning_rate": 7.226935312831389e-07, "epoch": 0.945619335347432 }, { "current_steps": 626, "loss": 3.1027, "learning_rate": 7.221633085896076e-07, "epoch": 0.947129909365559 }, { "current_steps": 627, "loss": 3.0754, "learning_rate": 7.216330858960763e-07, "epoch": 0.9486404833836858 }, { "current_steps": 628, "loss": 3.1411, "learning_rate": 7.21102863202545e-07, "epoch": 0.9501510574018127 }, { "current_steps": 629, "loss": 3.1479, "learning_rate": 7.205726405090138e-07, "epoch": 0.9516616314199395 }, { "current_steps": 630, "loss": 3.0268, "learning_rate": 7.200424178154825e-07, "epoch": 0.9531722054380665 }, { "current_steps": 631, "loss": 2.853, "learning_rate": 7.195121951219512e-07, "epoch": 0.9546827794561934 }, { "current_steps": 632, "loss": 2.8375, "learning_rate": 7.1898197242842e-07, "epoch": 0.9561933534743202 }, { "current_steps": 633, "loss": 3.0905, "learning_rate": 7.184517497348886e-07, "epoch": 0.9577039274924471 }, { "current_steps": 634, "loss": 2.9836, "learning_rate": 7.179215270413573e-07, "epoch": 0.959214501510574 }, { "current_steps": 635, "loss": 3.0679, "learning_rate": 7.17391304347826e-07, "epoch": 0.9607250755287009 }, { "current_steps": 636, "loss": 3.0425, "learning_rate": 7.168610816542949e-07, "epoch": 0.9622356495468278 }, { "current_steps": 637, "loss": 3.1853, "learning_rate": 7.163308589607635e-07, "epoch": 0.9637462235649547 }, { "current_steps": 638, "loss": 3.1647, "learning_rate": 7.158006362672322e-07, "epoch": 0.9652567975830816 }, { "current_steps": 639, "loss": 2.5096, "learning_rate": 7.152704135737009e-07, "epoch": 0.9667673716012085 }, { "current_steps": 640, "loss": 2.6806, "learning_rate": 7.147401908801697e-07, "epoch": 0.9682779456193353 }, { "current_steps": 641, "loss": 3.1098, "learning_rate": 7.142099681866383e-07, "epoch": 0.9697885196374623 }, { "current_steps": 642, "loss": 3.1581, "learning_rate": 7.136797454931071e-07, "epoch": 0.9712990936555891 }, { "current_steps": 643, "loss": 3.1607, "learning_rate": 7.131495227995758e-07, "epoch": 0.972809667673716 }, { "current_steps": 644, "loss": 2.9051, "learning_rate": 7.126193001060445e-07, "epoch": 0.974320241691843 }, { "current_steps": 645, "loss": 3.1324, "learning_rate": 7.120890774125132e-07, "epoch": 0.9758308157099698 }, { "current_steps": 646, "loss": 3.1401, "learning_rate": 7.115588547189819e-07, "epoch": 0.9773413897280967 }, { "current_steps": 647, "loss": 3.1494, "learning_rate": 7.110286320254506e-07, "epoch": 0.9788519637462235 }, { "current_steps": 648, "loss": 2.4261, "learning_rate": 7.104984093319194e-07, "epoch": 0.9803625377643505 }, { "current_steps": 649, "loss": 3.1947, "learning_rate": 7.099681866383881e-07, "epoch": 0.9818731117824774 }, { "current_steps": 650, "loss": 2.7172, "learning_rate": 7.094379639448568e-07, "epoch": 0.9833836858006042 }, { "current_steps": 651, "loss": 2.8514, "learning_rate": 7.089077412513256e-07, "epoch": 0.9848942598187311 }, { "current_steps": 652, "loss": 3.079, "learning_rate": 7.083775185577942e-07, "epoch": 0.986404833836858 }, { "current_steps": 653, "loss": 2.8849, "learning_rate": 7.078472958642629e-07, "epoch": 0.9879154078549849 }, { "current_steps": 654, "loss": 3.0467, "learning_rate": 7.073170731707316e-07, "epoch": 0.9894259818731118 }, { "current_steps": 655, "loss": 3.3841, "learning_rate": 7.067868504772005e-07, "epoch": 0.9909365558912386 }, { "current_steps": 656, "loss": 3.1446, "learning_rate": 7.062566277836691e-07, "epoch": 0.9924471299093656 }, { "current_steps": 657, "loss": 3.1988, "learning_rate": 7.057264050901378e-07, "epoch": 0.9939577039274925 }, { "current_steps": 658, "loss": 3.4408, "learning_rate": 7.051961823966065e-07, "epoch": 0.9954682779456193 }, { "current_steps": 659, "loss": 3.0199, "learning_rate": 7.046659597030753e-07, "epoch": 0.9969788519637462 }, { "current_steps": 660, "loss": 3.2084, "learning_rate": 7.041357370095439e-07, "epoch": 0.9984894259818731 }, { "current_steps": 661, "loss": 3.165, "learning_rate": 7.036055143160127e-07, "epoch": 1.0 }, { "current_steps": 662, "loss": 3.137, "learning_rate": 7.030752916224814e-07, "epoch": 1.001510574018127 }, { "current_steps": 663, "loss": 3.2575, "learning_rate": 7.025450689289502e-07, "epoch": 1.0030211480362539 }, { "current_steps": 664, "loss": 3.0359, "learning_rate": 7.020148462354188e-07, "epoch": 1.0045317220543806 }, { "current_steps": 665, "loss": 3.0096, "learning_rate": 7.014846235418875e-07, "epoch": 1.0060422960725075 }, { "current_steps": 666, "loss": 2.9221, "learning_rate": 7.009544008483563e-07, "epoch": 1.0075528700906344 }, { "current_steps": 667, "loss": 2.769, "learning_rate": 7.004241781548251e-07, "epoch": 1.0090634441087614 }, { "current_steps": 668, "loss": 3.2482, "learning_rate": 6.998939554612937e-07, "epoch": 1.0105740181268883 }, { "current_steps": 669, "loss": 2.9021, "learning_rate": 6.993637327677624e-07, "epoch": 1.012084592145015 }, { "current_steps": 670, "loss": 2.5269, "learning_rate": 6.988335100742312e-07, "epoch": 1.013595166163142 }, { "current_steps": 671, "loss": 3.1131, "learning_rate": 6.983032873806999e-07, "epoch": 1.0151057401812689 }, { "current_steps": 672, "loss": 3.0572, "learning_rate": 6.977730646871685e-07, "epoch": 1.0166163141993958 }, { "current_steps": 673, "loss": 3.036, "learning_rate": 6.972428419936372e-07, "epoch": 1.0181268882175227 }, { "current_steps": 674, "loss": 3.1856, "learning_rate": 6.967126193001061e-07, "epoch": 1.0196374622356494 }, { "current_steps": 675, "loss": 3.2084, "learning_rate": 6.961823966065748e-07, "epoch": 1.0211480362537764 }, { "current_steps": 676, "loss": 3.1684, "learning_rate": 6.956521739130434e-07, "epoch": 1.0226586102719033 }, { "current_steps": 677, "loss": 2.9708, "learning_rate": 6.951219512195121e-07, "epoch": 1.0241691842900302 }, { "current_steps": 678, "loss": 2.6613, "learning_rate": 6.945917285259809e-07, "epoch": 1.0256797583081572 }, { "current_steps": 679, "loss": 2.9646, "learning_rate": 6.940615058324496e-07, "epoch": 1.027190332326284 }, { "current_steps": 680, "loss": 3.1718, "learning_rate": 6.935312831389183e-07, "epoch": 1.0287009063444108 }, { "current_steps": 681, "loss": 2.6423, "learning_rate": 6.93001060445387e-07, "epoch": 1.0302114803625377 }, { "current_steps": 682, "loss": 3.0778, "learning_rate": 6.924708377518558e-07, "epoch": 1.0317220543806647 }, { "current_steps": 683, "loss": 3.0156, "learning_rate": 6.919406150583245e-07, "epoch": 1.0332326283987916 }, { "current_steps": 684, "loss": 2.9816, "learning_rate": 6.914103923647931e-07, "epoch": 1.0347432024169185 }, { "current_steps": 685, "loss": 2.938, "learning_rate": 6.908801696712619e-07, "epoch": 1.0362537764350452 }, { "current_steps": 686, "loss": 3.1688, "learning_rate": 6.903499469777307e-07, "epoch": 1.0377643504531722 }, { "current_steps": 687, "loss": 2.8528, "learning_rate": 6.898197242841993e-07, "epoch": 1.039274924471299 }, { "current_steps": 688, "loss": 3.1124, "learning_rate": 6.89289501590668e-07, "epoch": 1.040785498489426 }, { "current_steps": 689, "loss": 2.5208, "learning_rate": 6.887592788971368e-07, "epoch": 1.042296072507553 }, { "current_steps": 690, "loss": 3.2518, "learning_rate": 6.882290562036055e-07, "epoch": 1.0438066465256797 }, { "current_steps": 691, "loss": 3.0917, "learning_rate": 6.876988335100741e-07, "epoch": 1.0453172205438066 }, { "current_steps": 692, "loss": 3.0708, "learning_rate": 6.871686108165429e-07, "epoch": 1.0468277945619335 }, { "current_steps": 693, "loss": 2.637, "learning_rate": 6.866383881230117e-07, "epoch": 1.0483383685800605 }, { "current_steps": 694, "loss": 3.0982, "learning_rate": 6.861081654294804e-07, "epoch": 1.0498489425981874 }, { "current_steps": 695, "loss": 3.2218, "learning_rate": 6.85577942735949e-07, "epoch": 1.051359516616314 }, { "current_steps": 696, "loss": 3.1743, "learning_rate": 6.850477200424177e-07, "epoch": 1.052870090634441 }, { "current_steps": 697, "loss": 3.0273, "learning_rate": 6.845174973488865e-07, "epoch": 1.054380664652568 }, { "current_steps": 698, "loss": 2.8772, "learning_rate": 6.839872746553553e-07, "epoch": 1.055891238670695 }, { "current_steps": 699, "loss": 3.18, "learning_rate": 6.834570519618239e-07, "epoch": 1.0574018126888218 }, { "current_steps": 700, "loss": 3.0082, "learning_rate": 6.829268292682927e-07, "epoch": 1.0589123867069485 }, { "current_steps": 701, "loss": 3.1256, "learning_rate": 6.823966065747614e-07, "epoch": 1.0604229607250755 }, { "current_steps": 702, "loss": 2.3957, "learning_rate": 6.818663838812301e-07, "epoch": 1.0619335347432024 }, { "current_steps": 703, "loss": 3.0132, "learning_rate": 6.813361611876987e-07, "epoch": 1.0634441087613293 }, { "current_steps": 704, "loss": 3.1058, "learning_rate": 6.808059384941675e-07, "epoch": 1.0649546827794563 }, { "current_steps": 705, "loss": 2.9103, "learning_rate": 6.802757158006363e-07, "epoch": 1.0664652567975832 }, { "current_steps": 706, "loss": 2.504, "learning_rate": 6.79745493107105e-07, "epoch": 1.06797583081571 }, { "current_steps": 707, "loss": 2.9153, "learning_rate": 6.792152704135736e-07, "epoch": 1.0694864048338368 }, { "current_steps": 708, "loss": 3.2544, "learning_rate": 6.786850477200424e-07, "epoch": 1.0709969788519638 }, { "current_steps": 709, "loss": 3.0341, "learning_rate": 6.781548250265111e-07, "epoch": 1.0725075528700907 }, { "current_steps": 710, "loss": 3.0629, "learning_rate": 6.776246023329798e-07, "epoch": 1.0740181268882174 }, { "current_steps": 711, "loss": 2.8091, "learning_rate": 6.770943796394485e-07, "epoch": 1.0755287009063443 }, { "current_steps": 712, "loss": 3.0339, "learning_rate": 6.765641569459173e-07, "epoch": 1.0770392749244713 }, { "current_steps": 713, "loss": 3.0739, "learning_rate": 6.76033934252386e-07, "epoch": 1.0785498489425982 }, { "current_steps": 714, "loss": 3.2144, "learning_rate": 6.755037115588547e-07, "epoch": 1.0800604229607251 }, { "current_steps": 715, "loss": 3.1725, "learning_rate": 6.749734888653234e-07, "epoch": 1.081570996978852 }, { "current_steps": 716, "loss": 3.2246, "learning_rate": 6.744432661717921e-07, "epoch": 1.0830815709969788 }, { "current_steps": 717, "loss": 2.9513, "learning_rate": 6.739130434782609e-07, "epoch": 1.0845921450151057 }, { "current_steps": 718, "loss": 3.1187, "learning_rate": 6.733828207847296e-07, "epoch": 1.0861027190332326 }, { "current_steps": 719, "loss": 2.8206, "learning_rate": 6.728525980911983e-07, "epoch": 1.0876132930513596 }, { "current_steps": 720, "loss": 3.1289, "learning_rate": 6.72322375397667e-07, "epoch": 1.0891238670694865 }, { "current_steps": 721, "loss": 3.0412, "learning_rate": 6.717921527041357e-07, "epoch": 1.0906344410876132 }, { "current_steps": 722, "loss": 3.0754, "learning_rate": 6.712619300106044e-07, "epoch": 1.0921450151057401 }, { "current_steps": 723, "loss": 2.9696, "learning_rate": 6.707317073170731e-07, "epoch": 1.093655589123867 }, { "current_steps": 724, "loss": 3.1556, "learning_rate": 6.702014846235419e-07, "epoch": 1.095166163141994 }, { "current_steps": 725, "loss": 3.1364, "learning_rate": 6.696712619300106e-07, "epoch": 1.096676737160121 }, { "current_steps": 726, "loss": 2.9792, "learning_rate": 6.691410392364793e-07, "epoch": 1.0981873111782476 }, { "current_steps": 727, "loss": 2.7867, "learning_rate": 6.68610816542948e-07, "epoch": 1.0996978851963746 }, { "current_steps": 728, "loss": 3.1406, "learning_rate": 6.680805938494167e-07, "epoch": 1.1012084592145015 }, { "current_steps": 729, "loss": 2.8349, "learning_rate": 6.675503711558854e-07, "epoch": 1.1027190332326284 }, { "current_steps": 730, "loss": 2.939, "learning_rate": 6.670201484623541e-07, "epoch": 1.1042296072507554 }, { "current_steps": 731, "loss": 3.0505, "learning_rate": 6.664899257688229e-07, "epoch": 1.105740181268882 }, { "current_steps": 732, "loss": 2.6417, "learning_rate": 6.659597030752916e-07, "epoch": 1.107250755287009 }, { "current_steps": 733, "loss": 3.1132, "learning_rate": 6.654294803817603e-07, "epoch": 1.108761329305136 }, { "current_steps": 734, "loss": 3.0482, "learning_rate": 6.648992576882291e-07, "epoch": 1.1102719033232629 }, { "current_steps": 735, "loss": 3.0637, "learning_rate": 6.643690349946977e-07, "epoch": 1.1117824773413898 }, { "current_steps": 736, "loss": 2.9517, "learning_rate": 6.638388123011665e-07, "epoch": 1.1132930513595167 }, { "current_steps": 737, "loss": 3.083, "learning_rate": 6.633085896076352e-07, "epoch": 1.1148036253776434 }, { "current_steps": 738, "loss": 3.0229, "learning_rate": 6.627783669141039e-07, "epoch": 1.1163141993957704 }, { "current_steps": 739, "loss": 3.4434, "learning_rate": 6.622481442205726e-07, "epoch": 1.1178247734138973 }, { "current_steps": 740, "loss": 3.03, "learning_rate": 6.617179215270413e-07, "epoch": 1.1193353474320242 }, { "current_steps": 741, "loss": 2.8607, "learning_rate": 6.6118769883351e-07, "epoch": 1.1208459214501512 }, { "current_steps": 742, "loss": 3.2391, "learning_rate": 6.606574761399787e-07, "epoch": 1.1223564954682779 }, { "current_steps": 743, "loss": 3.097, "learning_rate": 6.601272534464475e-07, "epoch": 1.1238670694864048 }, { "current_steps": 744, "loss": 3.1365, "learning_rate": 6.595970307529162e-07, "epoch": 1.1253776435045317 }, { "current_steps": 745, "loss": 3.1356, "learning_rate": 6.590668080593849e-07, "epoch": 1.1268882175226587 }, { "current_steps": 746, "loss": 3.0501, "learning_rate": 6.585365853658536e-07, "epoch": 1.1283987915407856 }, { "current_steps": 747, "loss": 3.2082, "learning_rate": 6.580063626723223e-07, "epoch": 1.1299093655589123 }, { "current_steps": 748, "loss": 3.0707, "learning_rate": 6.57476139978791e-07, "epoch": 1.1314199395770392 }, { "current_steps": 749, "loss": 3.1677, "learning_rate": 6.569459172852599e-07, "epoch": 1.1329305135951662 }, { "current_steps": 750, "loss": 2.8567, "learning_rate": 6.564156945917285e-07, "epoch": 1.134441087613293 }, { "current_steps": 751, "loss": 2.6226, "learning_rate": 6.558854718981972e-07, "epoch": 1.13595166163142 }, { "current_steps": 752, "loss": 3.0405, "learning_rate": 6.553552492046659e-07, "epoch": 1.1374622356495467 }, { "current_steps": 753, "loss": 3.2305, "learning_rate": 6.548250265111347e-07, "epoch": 1.1389728096676737 }, { "current_steps": 754, "loss": 3.1328, "learning_rate": 6.542948038176033e-07, "epoch": 1.1404833836858006 }, { "current_steps": 755, "loss": 3.0304, "learning_rate": 6.537645811240721e-07, "epoch": 1.1419939577039275 }, { "current_steps": 756, "loss": 3.2476, "learning_rate": 6.532343584305408e-07, "epoch": 1.1435045317220545 }, { "current_steps": 757, "loss": 2.9137, "learning_rate": 6.527041357370096e-07, "epoch": 1.1450151057401814 }, { "current_steps": 758, "loss": 3.0561, "learning_rate": 6.521739130434782e-07, "epoch": 1.146525679758308 }, { "current_steps": 759, "loss": 2.9514, "learning_rate": 6.516436903499469e-07, "epoch": 1.148036253776435 }, { "current_steps": 760, "loss": 3.2839, "learning_rate": 6.511134676564156e-07, "epoch": 1.149546827794562 }, { "current_steps": 761, "loss": 2.473, "learning_rate": 6.505832449628845e-07, "epoch": 1.151057401812689 }, { "current_steps": 762, "loss": 2.9458, "learning_rate": 6.500530222693531e-07, "epoch": 1.1525679758308156 }, { "current_steps": 763, "loss": 3.2217, "learning_rate": 6.495227995758218e-07, "epoch": 1.1540785498489425 }, { "current_steps": 764, "loss": 2.8589, "learning_rate": 6.489925768822906e-07, "epoch": 1.1555891238670695 }, { "current_steps": 765, "loss": 2.8266, "learning_rate": 6.484623541887593e-07, "epoch": 1.1570996978851964 }, { "current_steps": 766, "loss": 3.1491, "learning_rate": 6.479321314952279e-07, "epoch": 1.1586102719033233 }, { "current_steps": 767, "loss": 2.9012, "learning_rate": 6.474019088016966e-07, "epoch": 1.1601208459214503 }, { "current_steps": 768, "loss": 2.8355, "learning_rate": 6.468716861081655e-07, "epoch": 1.161631419939577 }, { "current_steps": 769, "loss": 3.0604, "learning_rate": 6.463414634146342e-07, "epoch": 1.163141993957704 }, { "current_steps": 770, "loss": 2.8673, "learning_rate": 6.458112407211028e-07, "epoch": 1.1646525679758308 }, { "current_steps": 771, "loss": 3.0344, "learning_rate": 6.452810180275715e-07, "epoch": 1.1661631419939578 }, { "current_steps": 772, "loss": 2.7393, "learning_rate": 6.447507953340403e-07, "epoch": 1.1676737160120845 }, { "current_steps": 773, "loss": 2.9999, "learning_rate": 6.44220572640509e-07, "epoch": 1.1691842900302114 }, { "current_steps": 774, "loss": 2.8167, "learning_rate": 6.436903499469777e-07, "epoch": 1.1706948640483383 }, { "current_steps": 775, "loss": 3.1691, "learning_rate": 6.431601272534464e-07, "epoch": 1.1722054380664653 }, { "current_steps": 776, "loss": 3.1307, "learning_rate": 6.426299045599152e-07, "epoch": 1.1737160120845922 }, { "current_steps": 777, "loss": 3.053, "learning_rate": 6.420996818663838e-07, "epoch": 1.1752265861027191 }, { "current_steps": 778, "loss": 2.8401, "learning_rate": 6.415694591728525e-07, "epoch": 1.1767371601208458 }, { "current_steps": 779, "loss": 3.0846, "learning_rate": 6.410392364793212e-07, "epoch": 1.1782477341389728 }, { "current_steps": 780, "loss": 2.9075, "learning_rate": 6.405090137857901e-07, "epoch": 1.1797583081570997 }, { "current_steps": 781, "loss": 3.008, "learning_rate": 6.399787910922587e-07, "epoch": 1.1812688821752266 }, { "current_steps": 782, "loss": 2.9316, "learning_rate": 6.394485683987274e-07, "epoch": 1.1827794561933536 }, { "current_steps": 783, "loss": 2.9682, "learning_rate": 6.389183457051962e-07, "epoch": 1.1842900302114803 }, { "current_steps": 784, "loss": 3.1512, "learning_rate": 6.383881230116649e-07, "epoch": 1.1858006042296072 }, { "current_steps": 785, "loss": 2.3582, "learning_rate": 6.378579003181335e-07, "epoch": 1.1873111782477341 }, { "current_steps": 786, "loss": 3.1125, "learning_rate": 6.373276776246022e-07, "epoch": 1.188821752265861 }, { "current_steps": 787, "loss": 3.2427, "learning_rate": 6.367974549310711e-07, "epoch": 1.190332326283988 }, { "current_steps": 788, "loss": 2.8622, "learning_rate": 6.362672322375398e-07, "epoch": 1.191842900302115 }, { "current_steps": 789, "loss": 2.6609, "learning_rate": 6.357370095440084e-07, "epoch": 1.1933534743202416 }, { "current_steps": 790, "loss": 3.0324, "learning_rate": 6.352067868504771e-07, "epoch": 1.1948640483383686 }, { "current_steps": 791, "loss": 2.9873, "learning_rate": 6.346765641569459e-07, "epoch": 1.1963746223564955 }, { "current_steps": 792, "loss": 3.0322, "learning_rate": 6.341463414634146e-07, "epoch": 1.1978851963746224 }, { "current_steps": 793, "loss": 2.9698, "learning_rate": 6.336161187698833e-07, "epoch": 1.1993957703927491 }, { "current_steps": 794, "loss": 2.9074, "learning_rate": 6.33085896076352e-07, "epoch": 1.200906344410876 }, { "current_steps": 795, "loss": 2.7794, "learning_rate": 6.325556733828208e-07, "epoch": 1.202416918429003 }, { "current_steps": 796, "loss": 2.8889, "learning_rate": 6.320254506892895e-07, "epoch": 1.20392749244713 }, { "current_steps": 797, "loss": 2.9311, "learning_rate": 6.314952279957581e-07, "epoch": 1.2054380664652569 }, { "current_steps": 798, "loss": 2.9702, "learning_rate": 6.309650053022269e-07, "epoch": 1.2069486404833838 }, { "current_steps": 799, "loss": 2.0709, "learning_rate": 6.304347826086957e-07, "epoch": 1.2084592145015105 }, { "current_steps": 800, "loss": 3.211, "learning_rate": 6.299045599151644e-07, "epoch": 1.2099697885196374 }, { "current_steps": 801, "loss": 3.3029, "learning_rate": 6.29374337221633e-07, "epoch": 1.2114803625377644 }, { "current_steps": 802, "loss": 2.9971, "learning_rate": 6.288441145281018e-07, "epoch": 1.2129909365558913 }, { "current_steps": 803, "loss": 2.7264, "learning_rate": 6.283138918345705e-07, "epoch": 1.214501510574018 }, { "current_steps": 804, "loss": 2.8374, "learning_rate": 6.277836691410392e-07, "epoch": 1.216012084592145 }, { "current_steps": 805, "loss": 3.144, "learning_rate": 6.272534464475078e-07, "epoch": 1.2175226586102719 }, { "current_steps": 806, "loss": 3.0111, "learning_rate": 6.267232237539767e-07, "epoch": 1.2190332326283988 }, { "current_steps": 807, "loss": 2.9968, "learning_rate": 6.261930010604454e-07, "epoch": 1.2205438066465257 }, { "current_steps": 808, "loss": 3.141, "learning_rate": 6.256627783669141e-07, "epoch": 1.2220543806646527 }, { "current_steps": 809, "loss": 3.0366, "learning_rate": 6.251325556733827e-07, "epoch": 1.2235649546827794 }, { "current_steps": 810, "loss": 3.059, "learning_rate": 6.246023329798515e-07, "epoch": 1.2250755287009063 }, { "current_steps": 811, "loss": 2.8797, "learning_rate": 6.240721102863203e-07, "epoch": 1.2265861027190332 }, { "current_steps": 812, "loss": 2.5681, "learning_rate": 6.23541887592789e-07, "epoch": 1.2280966767371602 }, { "current_steps": 813, "loss": 3.088, "learning_rate": 6.230116648992576e-07, "epoch": 1.229607250755287 }, { "current_steps": 814, "loss": 2.4583, "learning_rate": 6.224814422057264e-07, "epoch": 1.2311178247734138 }, { "current_steps": 815, "loss": 2.9792, "learning_rate": 6.219512195121951e-07, "epoch": 1.2326283987915407 }, { "current_steps": 816, "loss": 3.0863, "learning_rate": 6.214209968186638e-07, "epoch": 1.2341389728096677 }, { "current_steps": 817, "loss": 3.0954, "learning_rate": 6.208907741251325e-07, "epoch": 1.2356495468277946 }, { "current_steps": 818, "loss": 3.2413, "learning_rate": 6.203605514316013e-07, "epoch": 1.2371601208459215 }, { "current_steps": 819, "loss": 3.0499, "learning_rate": 6.1983032873807e-07, "epoch": 1.2386706948640485 }, { "current_steps": 820, "loss": 3.1518, "learning_rate": 6.193001060445386e-07, "epoch": 1.2401812688821752 }, { "current_steps": 821, "loss": 3.2022, "learning_rate": 6.187698833510074e-07, "epoch": 1.241691842900302 }, { "current_steps": 822, "loss": 3.0871, "learning_rate": 6.182396606574761e-07, "epoch": 1.243202416918429 }, { "current_steps": 823, "loss": 2.957, "learning_rate": 6.177094379639448e-07, "epoch": 1.244712990936556 }, { "current_steps": 824, "loss": 3.1851, "learning_rate": 6.171792152704135e-07, "epoch": 1.2462235649546827 }, { "current_steps": 825, "loss": 2.6717, "learning_rate": 6.166489925768823e-07, "epoch": 1.2477341389728096 }, { "current_steps": 826, "loss": 3.0164, "learning_rate": 6.16118769883351e-07, "epoch": 1.2492447129909365 }, { "current_steps": 827, "loss": 2.977, "learning_rate": 6.155885471898197e-07, "epoch": 1.2507552870090635 }, { "current_steps": 828, "loss": 2.9581, "learning_rate": 6.150583244962883e-07, "epoch": 1.2522658610271904 }, { "current_steps": 829, "loss": 3.2141, "learning_rate": 6.145281018027571e-07, "epoch": 1.2537764350453173 }, { "current_steps": 830, "loss": 3.0109, "learning_rate": 6.139978791092259e-07, "epoch": 1.255287009063444 }, { "current_steps": 831, "loss": 3.0534, "learning_rate": 6.134676564156946e-07, "epoch": 1.256797583081571 }, { "current_steps": 832, "loss": 3.0423, "learning_rate": 6.129374337221633e-07, "epoch": 1.258308157099698 }, { "current_steps": 833, "loss": 3.0218, "learning_rate": 6.12407211028632e-07, "epoch": 1.2598187311178248 }, { "current_steps": 834, "loss": 2.9106, "learning_rate": 6.118769883351007e-07, "epoch": 1.2613293051359515 }, { "current_steps": 835, "loss": 3.1955, "learning_rate": 6.113467656415694e-07, "epoch": 1.2628398791540785 }, { "current_steps": 836, "loss": 3.0351, "learning_rate": 6.108165429480381e-07, "epoch": 1.2643504531722054 }, { "current_steps": 837, "loss": 3.0073, "learning_rate": 6.102863202545069e-07, "epoch": 1.2658610271903323 }, { "current_steps": 838, "loss": 2.7255, "learning_rate": 6.097560975609756e-07, "epoch": 1.2673716012084593 }, { "current_steps": 839, "loss": 2.8498, "learning_rate": 6.092258748674443e-07, "epoch": 1.2688821752265862 }, { "current_steps": 840, "loss": 3.0627, "learning_rate": 6.08695652173913e-07, "epoch": 1.2703927492447131 }, { "current_steps": 841, "loss": 3.053, "learning_rate": 6.081654294803817e-07, "epoch": 1.2719033232628398 }, { "current_steps": 842, "loss": 2.9029, "learning_rate": 6.076352067868504e-07, "epoch": 1.2734138972809668 }, { "current_steps": 843, "loss": 3.1912, "learning_rate": 6.071049840933192e-07, "epoch": 1.2749244712990937 }, { "current_steps": 844, "loss": 2.9187, "learning_rate": 6.065747613997879e-07, "epoch": 1.2764350453172204 }, { "current_steps": 845, "loss": 2.9065, "learning_rate": 6.060445387062566e-07, "epoch": 1.2779456193353473 }, { "current_steps": 846, "loss": 3.0209, "learning_rate": 6.055143160127253e-07, "epoch": 1.2794561933534743 }, { "current_steps": 847, "loss": 3.0331, "learning_rate": 6.049840933191941e-07, "epoch": 1.2809667673716012 }, { "current_steps": 848, "loss": 3.1613, "learning_rate": 6.044538706256627e-07, "epoch": 1.2824773413897281 }, { "current_steps": 849, "loss": 2.6895, "learning_rate": 6.039236479321315e-07, "epoch": 1.283987915407855 }, { "current_steps": 850, "loss": 3.0909, "learning_rate": 6.033934252386002e-07, "epoch": 1.285498489425982 }, { "current_steps": 851, "loss": 3.0953, "learning_rate": 6.02863202545069e-07, "epoch": 1.2870090634441087 }, { "current_steps": 852, "loss": 3.0476, "learning_rate": 6.023329798515376e-07, "epoch": 1.2885196374622356 }, { "current_steps": 853, "loss": 2.8508, "learning_rate": 6.018027571580063e-07, "epoch": 1.2900302114803626 }, { "current_steps": 854, "loss": 3.4764, "learning_rate": 6.01272534464475e-07, "epoch": 1.2915407854984895 }, { "current_steps": 855, "loss": 3.099, "learning_rate": 6.007423117709439e-07, "epoch": 1.2930513595166162 }, { "current_steps": 856, "loss": 3.069, "learning_rate": 6.002120890774125e-07, "epoch": 1.2945619335347431 }, { "current_steps": 857, "loss": 2.58, "learning_rate": 5.996818663838812e-07, "epoch": 1.29607250755287 }, { "current_steps": 858, "loss": 2.3271, "learning_rate": 5.991516436903499e-07, "epoch": 1.297583081570997 }, { "current_steps": 859, "loss": 2.9752, "learning_rate": 5.986214209968187e-07, "epoch": 1.299093655589124 }, { "current_steps": 860, "loss": 3.0935, "learning_rate": 5.980911983032873e-07, "epoch": 1.3006042296072509 }, { "current_steps": 861, "loss": 3.0636, "learning_rate": 5.97560975609756e-07, "epoch": 1.3021148036253776 }, { "current_steps": 862, "loss": 3.2768, "learning_rate": 5.970307529162248e-07, "epoch": 1.3036253776435045 }, { "current_steps": 863, "loss": 2.9215, "learning_rate": 5.965005302226936e-07, "epoch": 1.3051359516616314 }, { "current_steps": 864, "loss": 2.7811, "learning_rate": 5.959703075291622e-07, "epoch": 1.3066465256797584 }, { "current_steps": 865, "loss": 2.9675, "learning_rate": 5.954400848356309e-07, "epoch": 1.308157099697885 }, { "current_steps": 866, "loss": 2.8599, "learning_rate": 5.949098621420997e-07, "epoch": 1.309667673716012 }, { "current_steps": 867, "loss": 3.0658, "learning_rate": 5.943796394485684e-07, "epoch": 1.311178247734139 }, { "current_steps": 868, "loss": 2.9285, "learning_rate": 5.938494167550371e-07, "epoch": 1.3126888217522659 }, { "current_steps": 869, "loss": 2.8511, "learning_rate": 5.933191940615058e-07, "epoch": 1.3141993957703928 }, { "current_steps": 870, "loss": 3.236, "learning_rate": 5.927889713679746e-07, "epoch": 1.3157099697885197 }, { "current_steps": 871, "loss": 2.8618, "learning_rate": 5.922587486744432e-07, "epoch": 1.3172205438066467 }, { "current_steps": 872, "loss": 2.9207, "learning_rate": 5.917285259809119e-07, "epoch": 1.3187311178247734 }, { "current_steps": 873, "loss": 3.115, "learning_rate": 5.911983032873806e-07, "epoch": 1.3202416918429003 }, { "current_steps": 874, "loss": 3.0432, "learning_rate": 5.906680805938495e-07, "epoch": 1.3217522658610272 }, { "current_steps": 875, "loss": 2.9716, "learning_rate": 5.901378579003181e-07, "epoch": 1.323262839879154 }, { "current_steps": 876, "loss": 3.0914, "learning_rate": 5.896076352067868e-07, "epoch": 1.3247734138972809 }, { "current_steps": 877, "loss": 3.002, "learning_rate": 5.890774125132555e-07, "epoch": 1.3262839879154078 }, { "current_steps": 878, "loss": 2.9328, "learning_rate": 5.885471898197243e-07, "epoch": 1.3277945619335347 }, { "current_steps": 879, "loss": 3.1159, "learning_rate": 5.880169671261929e-07, "epoch": 1.3293051359516617 }, { "current_steps": 880, "loss": 2.632, "learning_rate": 5.874867444326616e-07, "epoch": 1.3308157099697886 }, { "current_steps": 881, "loss": 3.2234, "learning_rate": 5.869565217391305e-07, "epoch": 1.3323262839879155 }, { "current_steps": 882, "loss": 2.7351, "learning_rate": 5.864262990455992e-07, "epoch": 1.3338368580060422 }, { "current_steps": 883, "loss": 2.9956, "learning_rate": 5.858960763520678e-07, "epoch": 1.3353474320241692 }, { "current_steps": 884, "loss": 2.8181, "learning_rate": 5.853658536585365e-07, "epoch": 1.336858006042296 }, { "current_steps": 885, "loss": 3.0941, "learning_rate": 5.848356309650053e-07, "epoch": 1.338368580060423 }, { "current_steps": 886, "loss": 3.093, "learning_rate": 5.84305408271474e-07, "epoch": 1.3398791540785497 }, { "current_steps": 887, "loss": 3.1539, "learning_rate": 5.837751855779427e-07, "epoch": 1.3413897280966767 }, { "current_steps": 888, "loss": 2.798, "learning_rate": 5.832449628844114e-07, "epoch": 1.3429003021148036 }, { "current_steps": 889, "loss": 2.8593, "learning_rate": 5.827147401908802e-07, "epoch": 1.3444108761329305 }, { "current_steps": 890, "loss": 2.9718, "learning_rate": 5.821845174973489e-07, "epoch": 1.3459214501510575 }, { "current_steps": 891, "loss": 2.9314, "learning_rate": 5.816542948038175e-07, "epoch": 1.3474320241691844 }, { "current_steps": 892, "loss": 3.0187, "learning_rate": 5.811240721102862e-07, "epoch": 1.348942598187311 }, { "current_steps": 893, "loss": 3.0841, "learning_rate": 5.805938494167551e-07, "epoch": 1.350453172205438 }, { "current_steps": 894, "loss": 3.0654, "learning_rate": 5.800636267232238e-07, "epoch": 1.351963746223565 }, { "current_steps": 895, "loss": 3.0979, "learning_rate": 5.795334040296924e-07, "epoch": 1.353474320241692 }, { "current_steps": 896, "loss": 3.09, "learning_rate": 5.790031813361611e-07, "epoch": 1.3549848942598186 }, { "current_steps": 897, "loss": 2.9968, "learning_rate": 5.784729586426299e-07, "epoch": 1.3564954682779455 }, { "current_steps": 898, "loss": 2.4133, "learning_rate": 5.779427359490986e-07, "epoch": 1.3580060422960725 }, { "current_steps": 899, "loss": 3.1315, "learning_rate": 5.774125132555672e-07, "epoch": 1.3595166163141994 }, { "current_steps": 900, "loss": 3.0132, "learning_rate": 5.768822905620361e-07, "epoch": 1.3610271903323263 }, { "current_steps": 901, "loss": 2.8193, "learning_rate": 5.763520678685048e-07, "epoch": 1.3625377643504533 }, { "current_steps": 902, "loss": 3.0323, "learning_rate": 5.758218451749735e-07, "epoch": 1.3640483383685802 }, { "current_steps": 903, "loss": 2.8813, "learning_rate": 5.752916224814421e-07, "epoch": 1.365558912386707 }, { "current_steps": 904, "loss": 2.9499, "learning_rate": 5.747613997879109e-07, "epoch": 1.3670694864048338 }, { "current_steps": 905, "loss": 2.9082, "learning_rate": 5.742311770943796e-07, "epoch": 1.3685800604229608 }, { "current_steps": 906, "loss": 3.1331, "learning_rate": 5.737009544008484e-07, "epoch": 1.3700906344410877 }, { "current_steps": 907, "loss": 3.0406, "learning_rate": 5.73170731707317e-07, "epoch": 1.3716012084592144 }, { "current_steps": 908, "loss": 3.0628, "learning_rate": 5.726405090137858e-07, "epoch": 1.3731117824773413 }, { "current_steps": 909, "loss": 3.0879, "learning_rate": 5.721102863202545e-07, "epoch": 1.3746223564954683 }, { "current_steps": 910, "loss": 3.1388, "learning_rate": 5.715800636267231e-07, "epoch": 1.3761329305135952 }, { "current_steps": 911, "loss": 2.8779, "learning_rate": 5.710498409331918e-07, "epoch": 1.3776435045317221 }, { "current_steps": 912, "loss": 2.9396, "learning_rate": 5.705196182396607e-07, "epoch": 1.379154078549849 }, { "current_steps": 913, "loss": 3.0501, "learning_rate": 5.699893955461294e-07, "epoch": 1.3806646525679758 }, { "current_steps": 914, "loss": 3.077, "learning_rate": 5.69459172852598e-07, "epoch": 1.3821752265861027 }, { "current_steps": 915, "loss": 2.8821, "learning_rate": 5.689289501590668e-07, "epoch": 1.3836858006042296 }, { "current_steps": 916, "loss": 3.1957, "learning_rate": 5.683987274655355e-07, "epoch": 1.3851963746223566 }, { "current_steps": 917, "loss": 2.9614, "learning_rate": 5.678685047720042e-07, "epoch": 1.3867069486404833 }, { "current_steps": 918, "loss": 3.086, "learning_rate": 5.673382820784728e-07, "epoch": 1.3882175226586102 }, { "current_steps": 919, "loss": 2.2982, "learning_rate": 5.668080593849417e-07, "epoch": 1.3897280966767371 }, { "current_steps": 920, "loss": 3.0439, "learning_rate": 5.662778366914104e-07, "epoch": 1.391238670694864 }, { "current_steps": 921, "loss": 3.0418, "learning_rate": 5.657476139978791e-07, "epoch": 1.392749244712991 }, { "current_steps": 922, "loss": 2.9936, "learning_rate": 5.652173913043477e-07, "epoch": 1.394259818731118 }, { "current_steps": 923, "loss": 2.9119, "learning_rate": 5.646871686108165e-07, "epoch": 1.3957703927492446 }, { "current_steps": 924, "loss": 2.8972, "learning_rate": 5.641569459172853e-07, "epoch": 1.3972809667673716 }, { "current_steps": 925, "loss": 3.0952, "learning_rate": 5.63626723223754e-07, "epoch": 1.3987915407854985 }, { "current_steps": 926, "loss": 3.0398, "learning_rate": 5.630965005302226e-07, "epoch": 1.4003021148036254 }, { "current_steps": 927, "loss": 3.1679, "learning_rate": 5.625662778366914e-07, "epoch": 1.4018126888217521 }, { "current_steps": 928, "loss": 3.007, "learning_rate": 5.620360551431601e-07, "epoch": 1.403323262839879 }, { "current_steps": 929, "loss": 3.0324, "learning_rate": 5.615058324496288e-07, "epoch": 1.404833836858006 }, { "current_steps": 930, "loss": 3.2281, "learning_rate": 5.609756097560975e-07, "epoch": 1.406344410876133 }, { "current_steps": 931, "loss": 3.0227, "learning_rate": 5.604453870625663e-07, "epoch": 1.4078549848942599 }, { "current_steps": 932, "loss": 2.8318, "learning_rate": 5.59915164369035e-07, "epoch": 1.4093655589123868 }, { "current_steps": 933, "loss": 2.8935, "learning_rate": 5.593849416755037e-07, "epoch": 1.4108761329305137 }, { "current_steps": 934, "loss": 3.0167, "learning_rate": 5.588547189819724e-07, "epoch": 1.4123867069486404 }, { "current_steps": 935, "loss": 2.9921, "learning_rate": 5.583244962884411e-07, "epoch": 1.4138972809667674 }, { "current_steps": 936, "loss": 3.1317, "learning_rate": 5.577942735949098e-07, "epoch": 1.4154078549848943 }, { "current_steps": 937, "loss": 3.0075, "learning_rate": 5.572640509013786e-07, "epoch": 1.4169184290030212 }, { "current_steps": 938, "loss": 3.1235, "learning_rate": 5.567338282078473e-07, "epoch": 1.418429003021148 }, { "current_steps": 939, "loss": 2.6289, "learning_rate": 5.56203605514316e-07, "epoch": 1.4199395770392749 }, { "current_steps": 940, "loss": 2.9834, "learning_rate": 5.556733828207847e-07, "epoch": 1.4214501510574018 }, { "current_steps": 941, "loss": 3.1223, "learning_rate": 5.551431601272534e-07, "epoch": 1.4229607250755287 }, { "current_steps": 942, "loss": 3.1114, "learning_rate": 5.546129374337221e-07, "epoch": 1.4244712990936557 }, { "current_steps": 943, "loss": 2.9695, "learning_rate": 5.540827147401909e-07, "epoch": 1.4259818731117826 }, { "current_steps": 944, "loss": 2.5181, "learning_rate": 5.535524920466596e-07, "epoch": 1.4274924471299093 }, { "current_steps": 945, "loss": 3.0749, "learning_rate": 5.530222693531283e-07, "epoch": 1.4290030211480362 }, { "current_steps": 946, "loss": 2.8732, "learning_rate": 5.52492046659597e-07, "epoch": 1.4305135951661632 }, { "current_steps": 947, "loss": 3.0371, "learning_rate": 5.519618239660657e-07, "epoch": 1.43202416918429 }, { "current_steps": 948, "loss": 2.8945, "learning_rate": 5.514316012725344e-07, "epoch": 1.4335347432024168 }, { "current_steps": 949, "loss": 3.1701, "learning_rate": 5.509013785790033e-07, "epoch": 1.4350453172205437 }, { "current_steps": 950, "loss": 2.9467, "learning_rate": 5.503711558854719e-07, "epoch": 1.4365558912386707 }, { "current_steps": 951, "loss": 2.9566, "learning_rate": 5.498409331919406e-07, "epoch": 1.4380664652567976 }, { "current_steps": 952, "loss": 2.9671, "learning_rate": 5.493107104984093e-07, "epoch": 1.4395770392749245 }, { "current_steps": 953, "loss": 2.823, "learning_rate": 5.487804878048781e-07, "epoch": 1.4410876132930515 }, { "current_steps": 954, "loss": 2.5544, "learning_rate": 5.482502651113467e-07, "epoch": 1.4425981873111782 }, { "current_steps": 955, "loss": 2.8079, "learning_rate": 5.477200424178154e-07, "epoch": 1.444108761329305 }, { "current_steps": 956, "loss": 2.9745, "learning_rate": 5.471898197242842e-07, "epoch": 1.445619335347432 }, { "current_steps": 957, "loss": 2.9221, "learning_rate": 5.46659597030753e-07, "epoch": 1.447129909365559 }, { "current_steps": 958, "loss": 3.1318, "learning_rate": 5.461293743372216e-07, "epoch": 1.4486404833836857 }, { "current_steps": 959, "loss": 2.9277, "learning_rate": 5.455991516436903e-07, "epoch": 1.4501510574018126 }, { "current_steps": 960, "loss": 2.9046, "learning_rate": 5.45068928950159e-07, "epoch": 1.4516616314199395 }, { "current_steps": 961, "loss": 3.0094, "learning_rate": 5.445387062566277e-07, "epoch": 1.4531722054380665 }, { "current_steps": 962, "loss": 2.9582, "learning_rate": 5.440084835630965e-07, "epoch": 1.4546827794561934 }, { "current_steps": 963, "loss": 2.9235, "learning_rate": 5.434782608695652e-07, "epoch": 1.4561933534743203 }, { "current_steps": 964, "loss": 3.1794, "learning_rate": 5.42948038176034e-07, "epoch": 1.4577039274924473 }, { "current_steps": 965, "loss": 2.944, "learning_rate": 5.424178154825026e-07, "epoch": 1.459214501510574 }, { "current_steps": 966, "loss": 3.0503, "learning_rate": 5.418875927889713e-07, "epoch": 1.460725075528701 }, { "current_steps": 967, "loss": 2.824, "learning_rate": 5.4135737009544e-07, "epoch": 1.4622356495468278 }, { "current_steps": 968, "loss": 2.8722, "learning_rate": 5.408271474019089e-07, "epoch": 1.4637462235649548 }, { "current_steps": 969, "loss": 3.0602, "learning_rate": 5.402969247083775e-07, "epoch": 1.4652567975830815 }, { "current_steps": 970, "loss": 2.4757, "learning_rate": 5.397667020148462e-07, "epoch": 1.4667673716012084 }, { "current_steps": 971, "loss": 3.0635, "learning_rate": 5.392364793213149e-07, "epoch": 1.4682779456193353 }, { "current_steps": 972, "loss": 3.1136, "learning_rate": 5.387062566277837e-07, "epoch": 1.4697885196374623 }, { "current_steps": 973, "loss": 2.9535, "learning_rate": 5.381760339342523e-07, "epoch": 1.4712990936555892 }, { "current_steps": 974, "loss": 2.9776, "learning_rate": 5.37645811240721e-07, "epoch": 1.4728096676737161 }, { "current_steps": 975, "loss": 3.0826, "learning_rate": 5.371155885471898e-07, "epoch": 1.4743202416918428 }, { "current_steps": 976, "loss": 3.0639, "learning_rate": 5.365853658536586e-07, "epoch": 1.4758308157099698 }, { "current_steps": 977, "loss": 3.0708, "learning_rate": 5.360551431601272e-07, "epoch": 1.4773413897280967 }, { "current_steps": 978, "loss": 2.9469, "learning_rate": 5.355249204665959e-07, "epoch": 1.4788519637462236 }, { "current_steps": 979, "loss": 2.9673, "learning_rate": 5.349946977730647e-07, "epoch": 1.4803625377643503 }, { "current_steps": 980, "loss": 3.0806, "learning_rate": 5.344644750795334e-07, "epoch": 1.4818731117824773 }, { "current_steps": 981, "loss": 3.1313, "learning_rate": 5.339342523860021e-07, "epoch": 1.4833836858006042 }, { "current_steps": 982, "loss": 2.8451, "learning_rate": 5.334040296924708e-07, "epoch": 1.4848942598187311 }, { "current_steps": 983, "loss": 2.4288, "learning_rate": 5.328738069989396e-07, "epoch": 1.486404833836858 }, { "current_steps": 984, "loss": 3.0014, "learning_rate": 5.323435843054083e-07, "epoch": 1.487915407854985 }, { "current_steps": 985, "loss": 3.05, "learning_rate": 5.318133616118769e-07, "epoch": 1.4894259818731117 }, { "current_steps": 986, "loss": 3.0385, "learning_rate": 5.312831389183456e-07, "epoch": 1.4909365558912386 }, { "current_steps": 987, "loss": 3.0729, "learning_rate": 5.307529162248145e-07, "epoch": 1.4924471299093656 }, { "current_steps": 988, "loss": 2.8984, "learning_rate": 5.302226935312832e-07, "epoch": 1.4939577039274925 }, { "current_steps": 989, "loss": 3.0152, "learning_rate": 5.296924708377518e-07, "epoch": 1.4954682779456192 }, { "current_steps": 990, "loss": 2.9607, "learning_rate": 5.291622481442205e-07, "epoch": 1.4969788519637461 }, { "current_steps": 991, "loss": 2.8356, "learning_rate": 5.286320254506893e-07, "epoch": 1.498489425981873 }, { "current_steps": 992, "loss": 2.9473, "learning_rate": 5.28101802757158e-07, "epoch": 1.5 }, { "current_steps": 993, "loss": 2.9858, "learning_rate": 5.275715800636266e-07, "epoch": 1.501510574018127 }, { "current_steps": 994, "loss": 3.157, "learning_rate": 5.270413573700954e-07, "epoch": 1.5030211480362539 }, { "current_steps": 995, "loss": 3.1751, "learning_rate": 5.265111346765642e-07, "epoch": 1.5045317220543808 }, { "current_steps": 996, "loss": 2.8955, "learning_rate": 5.259809119830329e-07, "epoch": 1.5060422960725075 }, { "current_steps": 997, "loss": 3.0465, "learning_rate": 5.254506892895015e-07, "epoch": 1.5075528700906344 }, { "current_steps": 998, "loss": 2.6712, "learning_rate": 5.249204665959703e-07, "epoch": 1.5090634441087614 }, { "current_steps": 999, "loss": 2.9198, "learning_rate": 5.24390243902439e-07, "epoch": 1.510574018126888 }, { "current_steps": 1000, "loss": 2.9792, "learning_rate": 5.238600212089077e-07, "epoch": 1.512084592145015 }, { "current_steps": 1001, "loss": 2.5173, "learning_rate": 5.233297985153764e-07, "epoch": 1.513595166163142 }, { "current_steps": 1002, "loss": 3.2622, "learning_rate": 5.227995758218452e-07, "epoch": 1.5151057401812689 }, { "current_steps": 1003, "loss": 2.9889, "learning_rate": 5.222693531283139e-07, "epoch": 1.5166163141993958 }, { "current_steps": 1004, "loss": 2.9991, "learning_rate": 5.217391304347825e-07, "epoch": 1.5181268882175227 }, { "current_steps": 1005, "loss": 2.9715, "learning_rate": 5.212089077412512e-07, "epoch": 1.5196374622356497 }, { "current_steps": 1006, "loss": 3.1659, "learning_rate": 5.206786850477201e-07, "epoch": 1.5211480362537766 }, { "current_steps": 1007, "loss": 3.0155, "learning_rate": 5.201484623541888e-07, "epoch": 1.5226586102719033 }, { "current_steps": 1008, "loss": 3.0391, "learning_rate": 5.196182396606574e-07, "epoch": 1.5241691842900302 }, { "current_steps": 1009, "loss": 3.0686, "learning_rate": 5.190880169671261e-07, "epoch": 1.525679758308157 }, { "current_steps": 1010, "loss": 3.1287, "learning_rate": 5.185577942735949e-07, "epoch": 1.5271903323262839 }, { "current_steps": 1011, "loss": 3.0787, "learning_rate": 5.180275715800636e-07, "epoch": 1.5287009063444108 }, { "current_steps": 1012, "loss": 3.0413, "learning_rate": 5.174973488865322e-07, "epoch": 1.5302114803625377 }, { "current_steps": 1013, "loss": 2.9218, "learning_rate": 5.169671261930011e-07, "epoch": 1.5317220543806647 }, { "current_steps": 1014, "loss": 2.8721, "learning_rate": 5.164369034994698e-07, "epoch": 1.5332326283987916 }, { "current_steps": 1015, "loss": 2.922, "learning_rate": 5.159066808059385e-07, "epoch": 1.5347432024169185 }, { "current_steps": 1016, "loss": 3.1057, "learning_rate": 5.153764581124071e-07, "epoch": 1.5362537764350455 }, { "current_steps": 1017, "loss": 2.9731, "learning_rate": 5.148462354188759e-07, "epoch": 1.5377643504531722 }, { "current_steps": 1018, "loss": 3.1715, "learning_rate": 5.143160127253446e-07, "epoch": 1.539274924471299 }, { "current_steps": 1019, "loss": 3.045, "learning_rate": 5.137857900318134e-07, "epoch": 1.540785498489426 }, { "current_steps": 1020, "loss": 3.3725, "learning_rate": 5.13255567338282e-07, "epoch": 1.5422960725075527 }, { "current_steps": 1021, "loss": 2.7582, "learning_rate": 5.127253446447508e-07, "epoch": 1.5438066465256797 }, { "current_steps": 1022, "loss": 2.7718, "learning_rate": 5.121951219512195e-07, "epoch": 1.5453172205438066 }, { "current_steps": 1023, "loss": 2.9662, "learning_rate": 5.116648992576882e-07, "epoch": 1.5468277945619335 }, { "current_steps": 1024, "loss": 2.865, "learning_rate": 5.111346765641568e-07, "epoch": 1.5483383685800605 }, { "current_steps": 1025, "loss": 3.0375, "learning_rate": 5.106044538706257e-07, "epoch": 1.5498489425981874 }, { "current_steps": 1026, "loss": 3.0481, "learning_rate": 5.100742311770944e-07, "epoch": 1.5513595166163143 }, { "current_steps": 1027, "loss": 2.8947, "learning_rate": 5.095440084835631e-07, "epoch": 1.552870090634441 }, { "current_steps": 1028, "loss": 3.0734, "learning_rate": 5.090137857900317e-07, "epoch": 1.554380664652568 }, { "current_steps": 1029, "loss": 2.9264, "learning_rate": 5.084835630965005e-07, "epoch": 1.555891238670695 }, { "current_steps": 1030, "loss": 3.3168, "learning_rate": 5.079533404029692e-07, "epoch": 1.5574018126888216 }, { "current_steps": 1031, "loss": 3.019, "learning_rate": 5.07423117709438e-07, "epoch": 1.5589123867069485 }, { "current_steps": 1032, "loss": 2.8949, "learning_rate": 5.068928950159067e-07, "epoch": 1.5604229607250755 }, { "current_steps": 1033, "loss": 3.0329, "learning_rate": 5.063626723223754e-07, "epoch": 1.5619335347432024 }, { "current_steps": 1034, "loss": 2.9624, "learning_rate": 5.058324496288441e-07, "epoch": 1.5634441087613293 }, { "current_steps": 1035, "loss": 3.3392, "learning_rate": 5.053022269353128e-07, "epoch": 1.5649546827794563 }, { "current_steps": 1036, "loss": 2.2515, "learning_rate": 5.047720042417815e-07, "epoch": 1.5664652567975832 }, { "current_steps": 1037, "loss": 3.0141, "learning_rate": 5.042417815482503e-07, "epoch": 1.5679758308157101 }, { "current_steps": 1038, "loss": 3.0273, "learning_rate": 5.03711558854719e-07, "epoch": 1.5694864048338368 }, { "current_steps": 1039, "loss": 2.9199, "learning_rate": 5.031813361611877e-07, "epoch": 1.5709969788519638 }, { "current_steps": 1040, "loss": 3.1826, "learning_rate": 5.026511134676564e-07, "epoch": 1.5725075528700905 }, { "current_steps": 1041, "loss": 3.0795, "learning_rate": 5.021208907741251e-07, "epoch": 1.5740181268882174 }, { "current_steps": 1042, "loss": 3.0395, "learning_rate": 5.015906680805938e-07, "epoch": 1.5755287009063443 }, { "current_steps": 1043, "loss": 2.8528, "learning_rate": 5.010604453870624e-07, "epoch": 1.5770392749244713 }, { "current_steps": 1044, "loss": 2.7252, "learning_rate": 5.005302226935313e-07, "epoch": 1.5785498489425982 }, { "current_steps": 1045, "loss": 2.9675, "learning_rate": 5e-07, "epoch": 1.5800604229607251 }, { "current_steps": 1046, "loss": 2.956, "learning_rate": 4.994697773064687e-07, "epoch": 1.581570996978852 }, { "current_steps": 1047, "loss": 3.01, "learning_rate": 4.989395546129374e-07, "epoch": 1.583081570996979 }, { "current_steps": 1048, "loss": 2.9625, "learning_rate": 4.984093319194061e-07, "epoch": 1.5845921450151057 }, { "current_steps": 1049, "loss": 2.8837, "learning_rate": 4.978791092258748e-07, "epoch": 1.5861027190332326 }, { "current_steps": 1050, "loss": 3.0228, "learning_rate": 4.973488865323436e-07, "epoch": 1.5876132930513596 }, { "current_steps": 1051, "loss": 2.232, "learning_rate": 4.968186638388123e-07, "epoch": 1.5891238670694863 }, { "current_steps": 1052, "loss": 3.1003, "learning_rate": 4.96288441145281e-07, "epoch": 1.5906344410876132 }, { "current_steps": 1053, "loss": 3.1487, "learning_rate": 4.957582184517497e-07, "epoch": 1.5921450151057401 }, { "current_steps": 1054, "loss": 1.836, "learning_rate": 4.952279957582184e-07, "epoch": 1.593655589123867 }, { "current_steps": 1055, "loss": 3.1404, "learning_rate": 4.946977730646871e-07, "epoch": 1.595166163141994 }, { "current_steps": 1056, "loss": 3.036, "learning_rate": 4.941675503711559e-07, "epoch": 1.596676737160121 }, { "current_steps": 1057, "loss": 3.2197, "learning_rate": 4.936373276776246e-07, "epoch": 1.5981873111782479 }, { "current_steps": 1058, "loss": 3.0912, "learning_rate": 4.931071049840933e-07, "epoch": 1.5996978851963746 }, { "current_steps": 1059, "loss": 3.0635, "learning_rate": 4.92576882290562e-07, "epoch": 1.6012084592145015 }, { "current_steps": 1060, "loss": 2.8236, "learning_rate": 4.920466595970307e-07, "epoch": 1.6027190332326284 }, { "current_steps": 1061, "loss": 2.8799, "learning_rate": 4.915164369034994e-07, "epoch": 1.6042296072507551 }, { "current_steps": 1062, "loss": 3.0009, "learning_rate": 4.909862142099681e-07, "epoch": 1.605740181268882 }, { "current_steps": 1063, "loss": 2.7792, "learning_rate": 4.904559915164369e-07, "epoch": 1.607250755287009 }, { "current_steps": 1064, "loss": 2.4307, "learning_rate": 4.899257688229056e-07, "epoch": 1.608761329305136 }, { "current_steps": 1065, "loss": 3.0087, "learning_rate": 4.893955461293743e-07, "epoch": 1.6102719033232629 }, { "current_steps": 1066, "loss": 2.9968, "learning_rate": 4.88865323435843e-07, "epoch": 1.6117824773413898 }, { "current_steps": 1067, "loss": 3.0844, "learning_rate": 4.883351007423117e-07, "epoch": 1.6132930513595167 }, { "current_steps": 1068, "loss": 2.7457, "learning_rate": 4.878048780487804e-07, "epoch": 1.6148036253776437 }, { "current_steps": 1069, "loss": 3.1845, "learning_rate": 4.872746553552492e-07, "epoch": 1.6163141993957704 }, { "current_steps": 1070, "loss": 2.9618, "learning_rate": 4.867444326617179e-07, "epoch": 1.6178247734138973 }, { "current_steps": 1071, "loss": 3.1695, "learning_rate": 4.862142099681866e-07, "epoch": 1.619335347432024 }, { "current_steps": 1072, "loss": 3.048, "learning_rate": 4.856839872746553e-07, "epoch": 1.620845921450151 }, { "current_steps": 1073, "loss": 2.3976, "learning_rate": 4.85153764581124e-07, "epoch": 1.6223564954682779 }, { "current_steps": 1074, "loss": 2.8795, "learning_rate": 4.846235418875927e-07, "epoch": 1.6238670694864048 }, { "current_steps": 1075, "loss": 2.9775, "learning_rate": 4.840933191940615e-07, "epoch": 1.6253776435045317 }, { "current_steps": 1076, "loss": 3.0759, "learning_rate": 4.835630965005302e-07, "epoch": 1.6268882175226587 }, { "current_steps": 1077, "loss": 2.9119, "learning_rate": 4.830328738069989e-07, "epoch": 1.6283987915407856 }, { "current_steps": 1078, "loss": 2.9232, "learning_rate": 4.825026511134676e-07, "epoch": 1.6299093655589125 }, { "current_steps": 1079, "loss": 3.234, "learning_rate": 4.819724284199363e-07, "epoch": 1.6314199395770392 }, { "current_steps": 1080, "loss": 3.0302, "learning_rate": 4.814422057264051e-07, "epoch": 1.6329305135951662 }, { "current_steps": 1081, "loss": 2.9832, "learning_rate": 4.809119830328737e-07, "epoch": 1.634441087613293 }, { "current_steps": 1082, "loss": 3.0889, "learning_rate": 4.803817603393426e-07, "epoch": 1.6359516616314198 }, { "current_steps": 1083, "loss": 2.8111, "learning_rate": 4.798515376458112e-07, "epoch": 1.6374622356495467 }, { "current_steps": 1084, "loss": 2.8503, "learning_rate": 4.7932131495228e-07, "epoch": 1.6389728096676737 }, { "current_steps": 1085, "loss": 3.0871, "learning_rate": 4.787910922587486e-07, "epoch": 1.6404833836858006 }, { "current_steps": 1086, "loss": 2.9174, "learning_rate": 4.782608695652174e-07, "epoch": 1.6419939577039275 }, { "current_steps": 1087, "loss": 2.8806, "learning_rate": 4.77730646871686e-07, "epoch": 1.6435045317220545 }, { "current_steps": 1088, "loss": 3.13, "learning_rate": 4.772004241781549e-07, "epoch": 1.6450151057401814 }, { "current_steps": 1089, "loss": 2.6956, "learning_rate": 4.766702014846235e-07, "epoch": 1.646525679758308 }, { "current_steps": 1090, "loss": 2.8828, "learning_rate": 4.7613997879109224e-07, "epoch": 1.648036253776435 }, { "current_steps": 1091, "loss": 2.895, "learning_rate": 4.756097560975609e-07, "epoch": 1.649546827794562 }, { "current_steps": 1092, "loss": 2.8119, "learning_rate": 4.750795334040297e-07, "epoch": 1.6510574018126887 }, { "current_steps": 1093, "loss": 2.85, "learning_rate": 4.7454931071049834e-07, "epoch": 1.6525679758308156 }, { "current_steps": 1094, "loss": 3.1631, "learning_rate": 4.740190880169671e-07, "epoch": 1.6540785498489425 }, { "current_steps": 1095, "loss": 2.9764, "learning_rate": 4.734888653234358e-07, "epoch": 1.6555891238670695 }, { "current_steps": 1096, "loss": 3.0942, "learning_rate": 4.7295864262990454e-07, "epoch": 1.6570996978851964 }, { "current_steps": 1097, "loss": 2.8275, "learning_rate": 4.7242841993637326e-07, "epoch": 1.6586102719033233 }, { "current_steps": 1098, "loss": 2.9924, "learning_rate": 4.7189819724284197e-07, "epoch": 1.6601208459214503 }, { "current_steps": 1099, "loss": 2.9311, "learning_rate": 4.713679745493107e-07, "epoch": 1.6616314199395772 }, { "current_steps": 1100, "loss": 2.8092, "learning_rate": 4.708377518557794e-07, "epoch": 1.663141993957704 }, { "current_steps": 1101, "loss": 2.8916, "learning_rate": 4.703075291622481e-07, "epoch": 1.6646525679758308 }, { "current_steps": 1102, "loss": 3.2377, "learning_rate": 4.6977730646871684e-07, "epoch": 1.6661631419939575 }, { "current_steps": 1103, "loss": 3.0164, "learning_rate": 4.6924708377518555e-07, "epoch": 1.6676737160120845 }, { "current_steps": 1104, "loss": 3.0071, "learning_rate": 4.6871686108165427e-07, "epoch": 1.6691842900302114 }, { "current_steps": 1105, "loss": 3.116, "learning_rate": 4.68186638388123e-07, "epoch": 1.6706948640483383 }, { "current_steps": 1106, "loss": 3.2985, "learning_rate": 4.676564156945917e-07, "epoch": 1.6722054380664653 }, { "current_steps": 1107, "loss": 2.9589, "learning_rate": 4.671261930010604e-07, "epoch": 1.6737160120845922 }, { "current_steps": 1108, "loss": 2.9321, "learning_rate": 4.6659597030752913e-07, "epoch": 1.6752265861027191 }, { "current_steps": 1109, "loss": 3.1626, "learning_rate": 4.6606574761399785e-07, "epoch": 1.676737160120846 }, { "current_steps": 1110, "loss": 2.995, "learning_rate": 4.6553552492046656e-07, "epoch": 1.6782477341389728 }, { "current_steps": 1111, "loss": 2.5659, "learning_rate": 4.650053022269353e-07, "epoch": 1.6797583081570997 }, { "current_steps": 1112, "loss": 2.7322, "learning_rate": 4.6447507953340405e-07, "epoch": 1.6812688821752266 }, { "current_steps": 1113, "loss": 2.6716, "learning_rate": 4.639448568398727e-07, "epoch": 1.6827794561933533 }, { "current_steps": 1114, "loss": 2.9964, "learning_rate": 4.634146341463415e-07, "epoch": 1.6842900302114803 }, { "current_steps": 1115, "loss": 2.9699, "learning_rate": 4.6288441145281014e-07, "epoch": 1.6858006042296072 }, { "current_steps": 1116, "loss": 3.0185, "learning_rate": 4.623541887592789e-07, "epoch": 1.6873111782477341 }, { "current_steps": 1117, "loss": 2.8103, "learning_rate": 4.6182396606574757e-07, "epoch": 1.688821752265861 }, { "current_steps": 1118, "loss": 2.9484, "learning_rate": 4.6129374337221634e-07, "epoch": 1.690332326283988 }, { "current_steps": 1119, "loss": 3.0841, "learning_rate": 4.60763520678685e-07, "epoch": 1.691842900302115 }, { "current_steps": 1120, "loss": 2.9487, "learning_rate": 4.602332979851538e-07, "epoch": 1.6933534743202419 }, { "current_steps": 1121, "loss": 3.3735, "learning_rate": 4.5970307529162244e-07, "epoch": 1.6948640483383686 }, { "current_steps": 1122, "loss": 3.0413, "learning_rate": 4.591728525980912e-07, "epoch": 1.6963746223564955 }, { "current_steps": 1123, "loss": 2.9268, "learning_rate": 4.5864262990455987e-07, "epoch": 1.6978851963746222 }, { "current_steps": 1124, "loss": 3.07, "learning_rate": 4.5811240721102864e-07, "epoch": 1.6993957703927491 }, { "current_steps": 1125, "loss": 2.9197, "learning_rate": 4.575821845174973e-07, "epoch": 1.700906344410876 }, { "current_steps": 1126, "loss": 3.0171, "learning_rate": 4.5705196182396607e-07, "epoch": 1.702416918429003 }, { "current_steps": 1127, "loss": 3.1647, "learning_rate": 4.5652173913043473e-07, "epoch": 1.70392749244713 }, { "current_steps": 1128, "loss": 2.8764, "learning_rate": 4.559915164369035e-07, "epoch": 1.7054380664652569 }, { "current_steps": 1129, "loss": 3.005, "learning_rate": 4.554612937433722e-07, "epoch": 1.7069486404833838 }, { "current_steps": 1130, "loss": 3.1907, "learning_rate": 4.5493107104984093e-07, "epoch": 1.7084592145015107 }, { "current_steps": 1131, "loss": 3.0095, "learning_rate": 4.5440084835630965e-07, "epoch": 1.7099697885196374 }, { "current_steps": 1132, "loss": 2.9422, "learning_rate": 4.5387062566277836e-07, "epoch": 1.7114803625377644 }, { "current_steps": 1133, "loss": 3.0652, "learning_rate": 4.533404029692471e-07, "epoch": 1.712990936555891 }, { "current_steps": 1134, "loss": 2.9805, "learning_rate": 4.5281018027571574e-07, "epoch": 1.714501510574018 }, { "current_steps": 1135, "loss": 3.1721, "learning_rate": 4.522799575821845e-07, "epoch": 1.716012084592145 }, { "current_steps": 1136, "loss": 2.9939, "learning_rate": 4.517497348886532e-07, "epoch": 1.7175226586102719 }, { "current_steps": 1137, "loss": 3.0265, "learning_rate": 4.5121951219512194e-07, "epoch": 1.7190332326283988 }, { "current_steps": 1138, "loss": 3.2744, "learning_rate": 4.506892895015906e-07, "epoch": 1.7205438066465257 }, { "current_steps": 1139, "loss": 3.0499, "learning_rate": 4.501590668080594e-07, "epoch": 1.7220543806646527 }, { "current_steps": 1140, "loss": 2.9634, "learning_rate": 4.4962884411452804e-07, "epoch": 1.7235649546827796 }, { "current_steps": 1141, "loss": 3.0803, "learning_rate": 4.490986214209968e-07, "epoch": 1.7250755287009063 }, { "current_steps": 1142, "loss": 2.9291, "learning_rate": 4.4856839872746547e-07, "epoch": 1.7265861027190332 }, { "current_steps": 1143, "loss": 3.019, "learning_rate": 4.4803817603393424e-07, "epoch": 1.7280966767371602 }, { "current_steps": 1144, "loss": 2.9313, "learning_rate": 4.475079533404029e-07, "epoch": 1.7296072507552869 }, { "current_steps": 1145, "loss": 2.7374, "learning_rate": 4.4697773064687167e-07, "epoch": 1.7311178247734138 }, { "current_steps": 1146, "loss": 2.4116, "learning_rate": 4.464475079533404e-07, "epoch": 1.7326283987915407 }, { "current_steps": 1147, "loss": 2.6954, "learning_rate": 4.459172852598091e-07, "epoch": 1.7341389728096677 }, { "current_steps": 1148, "loss": 2.9735, "learning_rate": 4.453870625662778e-07, "epoch": 1.7356495468277946 }, { "current_steps": 1149, "loss": 2.9407, "learning_rate": 4.4485683987274653e-07, "epoch": 1.7371601208459215 }, { "current_steps": 1150, "loss": 3.0389, "learning_rate": 4.4432661717921525e-07, "epoch": 1.7386706948640485 }, { "current_steps": 1151, "loss": 2.4312, "learning_rate": 4.4379639448568397e-07, "epoch": 1.7401812688821754 }, { "current_steps": 1152, "loss": 2.9282, "learning_rate": 4.432661717921527e-07, "epoch": 1.741691842900302 }, { "current_steps": 1153, "loss": 2.918, "learning_rate": 4.427359490986214e-07, "epoch": 1.743202416918429 }, { "current_steps": 1154, "loss": 3.0313, "learning_rate": 4.422057264050901e-07, "epoch": 1.7447129909365557 }, { "current_steps": 1155, "loss": 2.8787, "learning_rate": 4.4167550371155883e-07, "epoch": 1.7462235649546827 }, { "current_steps": 1156, "loss": 3.3246, "learning_rate": 4.4114528101802755e-07, "epoch": 1.7477341389728096 }, { "current_steps": 1157, "loss": 3.0239, "learning_rate": 4.4061505832449626e-07, "epoch": 1.7492447129909365 }, { "current_steps": 1158, "loss": 2.8333, "learning_rate": 4.40084835630965e-07, "epoch": 1.7507552870090635 }, { "current_steps": 1159, "loss": 3.0921, "learning_rate": 4.395546129374337e-07, "epoch": 1.7522658610271904 }, { "current_steps": 1160, "loss": 3.0326, "learning_rate": 4.390243902439024e-07, "epoch": 1.7537764350453173 }, { "current_steps": 1161, "loss": 3.024, "learning_rate": 4.384941675503712e-07, "epoch": 1.7552870090634443 }, { "current_steps": 1162, "loss": 3.0421, "learning_rate": 4.3796394485683984e-07, "epoch": 1.756797583081571 }, { "current_steps": 1163, "loss": 2.9477, "learning_rate": 4.374337221633086e-07, "epoch": 1.758308157099698 }, { "current_steps": 1164, "loss": 3.0383, "learning_rate": 4.3690349946977727e-07, "epoch": 1.7598187311178246 }, { "current_steps": 1165, "loss": 2.8124, "learning_rate": 4.3637327677624604e-07, "epoch": 1.7613293051359515 }, { "current_steps": 1166, "loss": 3.0573, "learning_rate": 4.358430540827147e-07, "epoch": 1.7628398791540785 }, { "current_steps": 1167, "loss": 2.9941, "learning_rate": 4.3531283138918347e-07, "epoch": 1.7643504531722054 }, { "current_steps": 1168, "loss": 3.0391, "learning_rate": 4.3478260869565214e-07, "epoch": 1.7658610271903323 }, { "current_steps": 1169, "loss": 3.0691, "learning_rate": 4.342523860021209e-07, "epoch": 1.7673716012084593 }, { "current_steps": 1170, "loss": 2.916, "learning_rate": 4.3372216330858957e-07, "epoch": 1.7688821752265862 }, { "current_steps": 1171, "loss": 3.2599, "learning_rate": 4.3319194061505834e-07, "epoch": 1.7703927492447131 }, { "current_steps": 1172, "loss": 3.1476, "learning_rate": 4.32661717921527e-07, "epoch": 1.7719033232628398 }, { "current_steps": 1173, "loss": 2.758, "learning_rate": 4.3213149522799577e-07, "epoch": 1.7734138972809668 }, { "current_steps": 1174, "loss": 3.0326, "learning_rate": 4.3160127253446443e-07, "epoch": 1.7749244712990937 }, { "current_steps": 1175, "loss": 3.0573, "learning_rate": 4.310710498409332e-07, "epoch": 1.7764350453172204 }, { "current_steps": 1176, "loss": 3.0204, "learning_rate": 4.3054082714740186e-07, "epoch": 1.7779456193353473 }, { "current_steps": 1177, "loss": 2.9169, "learning_rate": 4.3001060445387063e-07, "epoch": 1.7794561933534743 }, { "current_steps": 1178, "loss": 3.0008, "learning_rate": 4.2948038176033935e-07, "epoch": 1.7809667673716012 }, { "current_steps": 1179, "loss": 3.0732, "learning_rate": 4.2895015906680806e-07, "epoch": 1.7824773413897281 }, { "current_steps": 1180, "loss": 2.9596, "learning_rate": 4.284199363732768e-07, "epoch": 1.783987915407855 }, { "current_steps": 1181, "loss": 2.971, "learning_rate": 4.2788971367974544e-07, "epoch": 1.785498489425982 }, { "current_steps": 1182, "loss": 3.1206, "learning_rate": 4.273594909862142e-07, "epoch": 1.787009063444109 }, { "current_steps": 1183, "loss": 3.1017, "learning_rate": 4.268292682926829e-07, "epoch": 1.7885196374622356 }, { "current_steps": 1184, "loss": 2.8115, "learning_rate": 4.2629904559915164e-07, "epoch": 1.7900302114803626 }, { "current_steps": 1185, "loss": 2.9832, "learning_rate": 4.257688229056203e-07, "epoch": 1.7915407854984893 }, { "current_steps": 1186, "loss": 2.9295, "learning_rate": 4.252386002120891e-07, "epoch": 1.7930513595166162 }, { "current_steps": 1187, "loss": 3.0374, "learning_rate": 4.2470837751855774e-07, "epoch": 1.7945619335347431 }, { "current_steps": 1188, "loss": 3.0179, "learning_rate": 4.241781548250265e-07, "epoch": 1.79607250755287 }, { "current_steps": 1189, "loss": 2.6478, "learning_rate": 4.2364793213149517e-07, "epoch": 1.797583081570997 }, { "current_steps": 1190, "loss": 3.1365, "learning_rate": 4.2311770943796394e-07, "epoch": 1.799093655589124 }, { "current_steps": 1191, "loss": 2.9728, "learning_rate": 4.225874867444326e-07, "epoch": 1.8006042296072509 }, { "current_steps": 1192, "loss": 3.0082, "learning_rate": 4.2205726405090137e-07, "epoch": 1.8021148036253778 }, { "current_steps": 1193, "loss": 3.0707, "learning_rate": 4.2152704135737003e-07, "epoch": 1.8036253776435045 }, { "current_steps": 1194, "loss": 3.0781, "learning_rate": 4.209968186638388e-07, "epoch": 1.8051359516616314 }, { "current_steps": 1195, "loss": 2.3737, "learning_rate": 4.204665959703075e-07, "epoch": 1.8066465256797581 }, { "current_steps": 1196, "loss": 2.7853, "learning_rate": 4.1993637327677623e-07, "epoch": 1.808157099697885 }, { "current_steps": 1197, "loss": 2.69, "learning_rate": 4.1940615058324495e-07, "epoch": 1.809667673716012 }, { "current_steps": 1198, "loss": 2.9768, "learning_rate": 4.1887592788971367e-07, "epoch": 1.811178247734139 }, { "current_steps": 1199, "loss": 2.9662, "learning_rate": 4.183457051961824e-07, "epoch": 1.8126888217522659 }, { "current_steps": 1200, "loss": 3.1874, "learning_rate": 4.178154825026511e-07, "epoch": 1.8141993957703928 }, { "current_steps": 1201, "loss": 2.9907, "learning_rate": 4.172852598091198e-07, "epoch": 1.8157099697885197 }, { "current_steps": 1202, "loss": 3.2888, "learning_rate": 4.1675503711558853e-07, "epoch": 1.8172205438066467 }, { "current_steps": 1203, "loss": 2.8668, "learning_rate": 4.1622481442205724e-07, "epoch": 1.8187311178247734 }, { "current_steps": 1204, "loss": 2.8886, "learning_rate": 4.1569459172852596e-07, "epoch": 1.8202416918429003 }, { "current_steps": 1205, "loss": 2.9962, "learning_rate": 4.151643690349947e-07, "epoch": 1.8217522658610272 }, { "current_steps": 1206, "loss": 2.9163, "learning_rate": 4.146341463414634e-07, "epoch": 1.823262839879154 }, { "current_steps": 1207, "loss": 2.8536, "learning_rate": 4.141039236479321e-07, "epoch": 1.8247734138972809 }, { "current_steps": 1208, "loss": 2.8376, "learning_rate": 4.135737009544008e-07, "epoch": 1.8262839879154078 }, { "current_steps": 1209, "loss": 2.9377, "learning_rate": 4.1304347826086954e-07, "epoch": 1.8277945619335347 }, { "current_steps": 1210, "loss": 2.7267, "learning_rate": 4.1251325556733826e-07, "epoch": 1.8293051359516617 }, { "current_steps": 1211, "loss": 2.8558, "learning_rate": 4.1198303287380697e-07, "epoch": 1.8308157099697886 }, { "current_steps": 1212, "loss": 3.2157, "learning_rate": 4.1145281018027574e-07, "epoch": 1.8323262839879155 }, { "current_steps": 1213, "loss": 2.5958, "learning_rate": 4.109225874867444e-07, "epoch": 1.8338368580060425 }, { "current_steps": 1214, "loss": 2.9366, "learning_rate": 4.1039236479321317e-07, "epoch": 1.8353474320241692 }, { "current_steps": 1215, "loss": 2.971, "learning_rate": 4.0986214209968183e-07, "epoch": 1.836858006042296 }, { "current_steps": 1216, "loss": 2.69, "learning_rate": 4.093319194061506e-07, "epoch": 1.8383685800604228 }, { "current_steps": 1217, "loss": 3.0503, "learning_rate": 4.0880169671261927e-07, "epoch": 1.8398791540785497 }, { "current_steps": 1218, "loss": 3.1028, "learning_rate": 4.0827147401908804e-07, "epoch": 1.8413897280966767 }, { "current_steps": 1219, "loss": 3.1497, "learning_rate": 4.077412513255567e-07, "epoch": 1.8429003021148036 }, { "current_steps": 1220, "loss": 3.0073, "learning_rate": 4.0721102863202547e-07, "epoch": 1.8444108761329305 }, { "current_steps": 1221, "loss": 3.0223, "learning_rate": 4.0668080593849413e-07, "epoch": 1.8459214501510575 }, { "current_steps": 1222, "loss": 3.1263, "learning_rate": 4.061505832449629e-07, "epoch": 1.8474320241691844 }, { "current_steps": 1223, "loss": 2.9748, "learning_rate": 4.0562036055143156e-07, "epoch": 1.8489425981873113 }, { "current_steps": 1224, "loss": 3.0741, "learning_rate": 4.0509013785790033e-07, "epoch": 1.850453172205438 }, { "current_steps": 1225, "loss": 3.1341, "learning_rate": 4.04559915164369e-07, "epoch": 1.851963746223565 }, { "current_steps": 1226, "loss": 2.9934, "learning_rate": 4.040296924708377e-07, "epoch": 1.853474320241692 }, { "current_steps": 1227, "loss": 2.8055, "learning_rate": 4.034994697773065e-07, "epoch": 1.8549848942598186 }, { "current_steps": 1228, "loss": 2.9672, "learning_rate": 4.0296924708377514e-07, "epoch": 1.8564954682779455 }, { "current_steps": 1229, "loss": 3.0207, "learning_rate": 4.024390243902439e-07, "epoch": 1.8580060422960725 }, { "current_steps": 1230, "loss": 2.9925, "learning_rate": 4.0190880169671257e-07, "epoch": 1.8595166163141994 }, { "current_steps": 1231, "loss": 2.2767, "learning_rate": 4.0137857900318134e-07, "epoch": 1.8610271903323263 }, { "current_steps": 1232, "loss": 2.8183, "learning_rate": 4.0084835630965e-07, "epoch": 1.8625377643504533 }, { "current_steps": 1233, "loss": 3.1208, "learning_rate": 4.003181336161188e-07, "epoch": 1.8640483383685802 }, { "current_steps": 1234, "loss": 2.8493, "learning_rate": 3.9978791092258744e-07, "epoch": 1.865558912386707 }, { "current_steps": 1235, "loss": 3.2308, "learning_rate": 3.992576882290562e-07, "epoch": 1.8670694864048338 }, { "current_steps": 1236, "loss": 3.1253, "learning_rate": 3.9872746553552487e-07, "epoch": 1.8685800604229608 }, { "current_steps": 1237, "loss": 2.9741, "learning_rate": 3.9819724284199364e-07, "epoch": 1.8700906344410875 }, { "current_steps": 1238, "loss": 3.0487, "learning_rate": 3.976670201484623e-07, "epoch": 1.8716012084592144 }, { "current_steps": 1239, "loss": 2.8604, "learning_rate": 3.9713679745493107e-07, "epoch": 1.8731117824773413 }, { "current_steps": 1240, "loss": 2.9669, "learning_rate": 3.9660657476139973e-07, "epoch": 1.8746223564954683 }, { "current_steps": 1241, "loss": 2.7471, "learning_rate": 3.960763520678685e-07, "epoch": 1.8761329305135952 }, { "current_steps": 1242, "loss": 3.0211, "learning_rate": 3.9554612937433716e-07, "epoch": 1.8776435045317221 }, { "current_steps": 1243, "loss": 3.1709, "learning_rate": 3.9501590668080593e-07, "epoch": 1.879154078549849 }, { "current_steps": 1244, "loss": 2.9585, "learning_rate": 3.9448568398727465e-07, "epoch": 1.880664652567976 }, { "current_steps": 1245, "loss": 2.2015, "learning_rate": 3.9395546129374336e-07, "epoch": 1.8821752265861027 }, { "current_steps": 1246, "loss": 2.9052, "learning_rate": 3.934252386002121e-07, "epoch": 1.8836858006042296 }, { "current_steps": 1247, "loss": 2.9724, "learning_rate": 3.928950159066808e-07, "epoch": 1.8851963746223563 }, { "current_steps": 1248, "loss": 3.0223, "learning_rate": 3.923647932131495e-07, "epoch": 1.8867069486404833 }, { "current_steps": 1249, "loss": 2.9257, "learning_rate": 3.9183457051961823e-07, "epoch": 1.8882175226586102 }, { "current_steps": 1250, "loss": 2.9357, "learning_rate": 3.9130434782608694e-07, "epoch": 1.8897280966767371 }, { "current_steps": 1251, "loss": 3.0655, "learning_rate": 3.9077412513255566e-07, "epoch": 1.891238670694864 }, { "current_steps": 1252, "loss": 2.9263, "learning_rate": 3.902439024390244e-07, "epoch": 1.892749244712991 }, { "current_steps": 1253, "loss": 3.0746, "learning_rate": 3.897136797454931e-07, "epoch": 1.894259818731118 }, { "current_steps": 1254, "loss": 2.9773, "learning_rate": 3.891834570519618e-07, "epoch": 1.8957703927492449 }, { "current_steps": 1255, "loss": 3.0119, "learning_rate": 3.886532343584305e-07, "epoch": 1.8972809667673716 }, { "current_steps": 1256, "loss": 2.7599, "learning_rate": 3.8812301166489924e-07, "epoch": 1.8987915407854985 }, { "current_steps": 1257, "loss": 2.9744, "learning_rate": 3.8759278897136795e-07, "epoch": 1.9003021148036254 }, { "current_steps": 1258, "loss": 3.1058, "learning_rate": 3.8706256627783667e-07, "epoch": 1.9018126888217521 }, { "current_steps": 1259, "loss": 3.071, "learning_rate": 3.865323435843054e-07, "epoch": 1.903323262839879 }, { "current_steps": 1260, "loss": 2.8942, "learning_rate": 3.860021208907741e-07, "epoch": 1.904833836858006 }, { "current_steps": 1261, "loss": 2.7528, "learning_rate": 3.8547189819724287e-07, "epoch": 1.906344410876133 }, { "current_steps": 1262, "loss": 2.3892, "learning_rate": 3.8494167550371153e-07, "epoch": 1.9078549848942599 }, { "current_steps": 1263, "loss": 3.075, "learning_rate": 3.844114528101803e-07, "epoch": 1.9093655589123868 }, { "current_steps": 1264, "loss": 2.9926, "learning_rate": 3.8388123011664897e-07, "epoch": 1.9108761329305137 }, { "current_steps": 1265, "loss": 3.1426, "learning_rate": 3.8335100742311773e-07, "epoch": 1.9123867069486404 }, { "current_steps": 1266, "loss": 2.9099, "learning_rate": 3.828207847295864e-07, "epoch": 1.9138972809667674 }, { "current_steps": 1267, "loss": 2.9872, "learning_rate": 3.8229056203605517e-07, "epoch": 1.9154078549848943 }, { "current_steps": 1268, "loss": 3.1473, "learning_rate": 3.8176033934252383e-07, "epoch": 1.916918429003021 }, { "current_steps": 1269, "loss": 3.0227, "learning_rate": 3.812301166489926e-07, "epoch": 1.918429003021148 }, { "current_steps": 1270, "loss": 2.9583, "learning_rate": 3.8069989395546126e-07, "epoch": 1.9199395770392749 }, { "current_steps": 1271, "loss": 3.1131, "learning_rate": 3.8016967126193e-07, "epoch": 1.9214501510574018 }, { "current_steps": 1272, "loss": 2.8828, "learning_rate": 3.796394485683987e-07, "epoch": 1.9229607250755287 }, { "current_steps": 1273, "loss": 3.0638, "learning_rate": 3.791092258748674e-07, "epoch": 1.9244712990936557 }, { "current_steps": 1274, "loss": 2.876, "learning_rate": 3.785790031813361e-07, "epoch": 1.9259818731117826 }, { "current_steps": 1275, "loss": 3.2298, "learning_rate": 3.7804878048780484e-07, "epoch": 1.9274924471299095 }, { "current_steps": 1276, "loss": 2.9171, "learning_rate": 3.7751855779427356e-07, "epoch": 1.9290030211480362 }, { "current_steps": 1277, "loss": 3.0993, "learning_rate": 3.7698833510074227e-07, "epoch": 1.9305135951661632 }, { "current_steps": 1278, "loss": 3.2426, "learning_rate": 3.7645811240721104e-07, "epoch": 1.9320241691842899 }, { "current_steps": 1279, "loss": 2.8922, "learning_rate": 3.759278897136797e-07, "epoch": 1.9335347432024168 }, { "current_steps": 1280, "loss": 3.1178, "learning_rate": 3.7539766702014847e-07, "epoch": 1.9350453172205437 }, { "current_steps": 1281, "loss": 3.5744, "learning_rate": 3.7486744432661714e-07, "epoch": 1.9365558912386707 }, { "current_steps": 1282, "loss": 2.8252, "learning_rate": 3.743372216330859e-07, "epoch": 1.9380664652567976 }, { "current_steps": 1283, "loss": 2.998, "learning_rate": 3.7380699893955457e-07, "epoch": 1.9395770392749245 }, { "current_steps": 1284, "loss": 2.6686, "learning_rate": 3.7327677624602334e-07, "epoch": 1.9410876132930515 }, { "current_steps": 1285, "loss": 3.123, "learning_rate": 3.72746553552492e-07, "epoch": 1.9425981873111784 }, { "current_steps": 1286, "loss": 3.0199, "learning_rate": 3.7221633085896077e-07, "epoch": 1.944108761329305 }, { "current_steps": 1287, "loss": 2.8993, "learning_rate": 3.7168610816542943e-07, "epoch": 1.945619335347432 }, { "current_steps": 1288, "loss": 3.1056, "learning_rate": 3.711558854718982e-07, "epoch": 1.947129909365559 }, { "current_steps": 1289, "loss": 3.0076, "learning_rate": 3.7062566277836686e-07, "epoch": 1.9486404833836857 }, { "current_steps": 1290, "loss": 2.8532, "learning_rate": 3.7009544008483563e-07, "epoch": 1.9501510574018126 }, { "current_steps": 1291, "loss": 2.9681, "learning_rate": 3.695652173913043e-07, "epoch": 1.9516616314199395 }, { "current_steps": 1292, "loss": 2.186, "learning_rate": 3.6903499469777306e-07, "epoch": 1.9531722054380665 }, { "current_steps": 1293, "loss": 2.8652, "learning_rate": 3.685047720042417e-07, "epoch": 1.9546827794561934 }, { "current_steps": 1294, "loss": 2.4061, "learning_rate": 3.679745493107105e-07, "epoch": 1.9561933534743203 }, { "current_steps": 1295, "loss": 2.9065, "learning_rate": 3.674443266171792e-07, "epoch": 1.9577039274924473 }, { "current_steps": 1296, "loss": 2.9456, "learning_rate": 3.6691410392364793e-07, "epoch": 1.959214501510574 }, { "current_steps": 1297, "loss": 2.8736, "learning_rate": 3.6638388123011664e-07, "epoch": 1.960725075528701 }, { "current_steps": 1298, "loss": 2.9118, "learning_rate": 3.6585365853658536e-07, "epoch": 1.9622356495468278 }, { "current_steps": 1299, "loss": 2.8376, "learning_rate": 3.653234358430541e-07, "epoch": 1.9637462235649545 }, { "current_steps": 1300, "loss": 3.0716, "learning_rate": 3.647932131495228e-07, "epoch": 1.9652567975830815 }, { "current_steps": 1301, "loss": 2.2068, "learning_rate": 3.642629904559915e-07, "epoch": 1.9667673716012084 }, { "current_steps": 1302, "loss": 2.8887, "learning_rate": 3.637327677624602e-07, "epoch": 1.9682779456193353 }, { "current_steps": 1303, "loss": 3.0633, "learning_rate": 3.6320254506892894e-07, "epoch": 1.9697885196374623 }, { "current_steps": 1304, "loss": 3.1106, "learning_rate": 3.6267232237539765e-07, "epoch": 1.9712990936555892 }, { "current_steps": 1305, "loss": 3.075, "learning_rate": 3.6214209968186637e-07, "epoch": 1.9728096676737161 }, { "current_steps": 1306, "loss": 3.0428, "learning_rate": 3.616118769883351e-07, "epoch": 1.974320241691843 }, { "current_steps": 1307, "loss": 3.0262, "learning_rate": 3.610816542948038e-07, "epoch": 1.9758308157099698 }, { "current_steps": 1308, "loss": 2.8567, "learning_rate": 3.605514316012725e-07, "epoch": 1.9773413897280967 }, { "current_steps": 1309, "loss": 2.8223, "learning_rate": 3.6002120890774123e-07, "epoch": 1.9788519637462234 }, { "current_steps": 1310, "loss": 2.9899, "learning_rate": 3.5949098621421e-07, "epoch": 1.9803625377643503 }, { "current_steps": 1311, "loss": 2.9231, "learning_rate": 3.5896076352067866e-07, "epoch": 1.9818731117824773 }, { "current_steps": 1312, "loss": 2.9539, "learning_rate": 3.5843054082714743e-07, "epoch": 1.9833836858006042 }, { "current_steps": 1313, "loss": 2.9124, "learning_rate": 3.579003181336161e-07, "epoch": 1.9848942598187311 }, { "current_steps": 1314, "loss": 2.9618, "learning_rate": 3.5737009544008487e-07, "epoch": 1.986404833836858 }, { "current_steps": 1315, "loss": 2.3594, "learning_rate": 3.5683987274655353e-07, "epoch": 1.987915407854985 }, { "current_steps": 1316, "loss": 2.7551, "learning_rate": 3.5630965005302224e-07, "epoch": 1.989425981873112 }, { "current_steps": 1317, "loss": 2.8081, "learning_rate": 3.5577942735949096e-07, "epoch": 1.9909365558912386 }, { "current_steps": 1318, "loss": 3.1055, "learning_rate": 3.552492046659597e-07, "epoch": 1.9924471299093656 }, { "current_steps": 1319, "loss": 2.9898, "learning_rate": 3.547189819724284e-07, "epoch": 1.9939577039274925 }, { "current_steps": 1320, "loss": 3.066, "learning_rate": 3.541887592788971e-07, "epoch": 1.9954682779456192 }, { "current_steps": 1321, "loss": 2.865, "learning_rate": 3.536585365853658e-07, "epoch": 1.9969788519637461 }, { "current_steps": 1322, "loss": 2.8501, "learning_rate": 3.5312831389183454e-07, "epoch": 1.998489425981873 }, { "current_steps": 1323, "loss": 3.094, "learning_rate": 3.5259809119830326e-07, "epoch": 2.0 }, { "current_steps": 1324, "loss": 2.6459, "learning_rate": 3.5206786850477197e-07, "epoch": 2.001510574018127 }, { "current_steps": 1325, "loss": 3.0059, "learning_rate": 3.515376458112407e-07, "epoch": 2.003021148036254 }, { "current_steps": 1326, "loss": 3.1677, "learning_rate": 3.510074231177094e-07, "epoch": 2.004531722054381 }, { "current_steps": 1327, "loss": 2.4981, "learning_rate": 3.5047720042417817e-07, "epoch": 2.0060422960725077 }, { "current_steps": 1328, "loss": 2.7343, "learning_rate": 3.4994697773064683e-07, "epoch": 2.0075528700906347 }, { "current_steps": 1329, "loss": 2.8042, "learning_rate": 3.494167550371156e-07, "epoch": 2.009063444108761 }, { "current_steps": 1330, "loss": 2.9556, "learning_rate": 3.4888653234358427e-07, "epoch": 2.010574018126888 }, { "current_steps": 1331, "loss": 3.1188, "learning_rate": 3.4835630965005304e-07, "epoch": 2.012084592145015 }, { "current_steps": 1332, "loss": 2.9043, "learning_rate": 3.478260869565217e-07, "epoch": 2.013595166163142 }, { "current_steps": 1333, "loss": 3.0621, "learning_rate": 3.4729586426299047e-07, "epoch": 2.015105740181269 }, { "current_steps": 1334, "loss": 2.9419, "learning_rate": 3.4676564156945913e-07, "epoch": 2.016616314199396 }, { "current_steps": 1335, "loss": 3.0796, "learning_rate": 3.462354188759279e-07, "epoch": 2.0181268882175227 }, { "current_steps": 1336, "loss": 2.8447, "learning_rate": 3.4570519618239656e-07, "epoch": 2.0196374622356497 }, { "current_steps": 1337, "loss": 3.1167, "learning_rate": 3.4517497348886533e-07, "epoch": 2.0211480362537766 }, { "current_steps": 1338, "loss": 3.2094, "learning_rate": 3.44644750795334e-07, "epoch": 2.0226586102719035 }, { "current_steps": 1339, "loss": 2.9261, "learning_rate": 3.4411452810180276e-07, "epoch": 2.02416918429003 }, { "current_steps": 1340, "loss": 2.8645, "learning_rate": 3.435843054082714e-07, "epoch": 2.025679758308157 }, { "current_steps": 1341, "loss": 2.9348, "learning_rate": 3.430540827147402e-07, "epoch": 2.027190332326284 }, { "current_steps": 1342, "loss": 2.9588, "learning_rate": 3.4252386002120886e-07, "epoch": 2.028700906344411 }, { "current_steps": 1343, "loss": 3.0665, "learning_rate": 3.419936373276776e-07, "epoch": 2.0302114803625377 }, { "current_steps": 1344, "loss": 3.0973, "learning_rate": 3.4146341463414634e-07, "epoch": 2.0317220543806647 }, { "current_steps": 1345, "loss": 2.9658, "learning_rate": 3.4093319194061506e-07, "epoch": 2.0332326283987916 }, { "current_steps": 1346, "loss": 2.9395, "learning_rate": 3.4040296924708377e-07, "epoch": 2.0347432024169185 }, { "current_steps": 1347, "loss": 3.0696, "learning_rate": 3.398727465535525e-07, "epoch": 2.0362537764350455 }, { "current_steps": 1348, "loss": 2.9523, "learning_rate": 3.393425238600212e-07, "epoch": 2.0377643504531724 }, { "current_steps": 1349, "loss": 3.1074, "learning_rate": 3.388123011664899e-07, "epoch": 2.039274924471299 }, { "current_steps": 1350, "loss": 3.0615, "learning_rate": 3.3828207847295864e-07, "epoch": 2.040785498489426 }, { "current_steps": 1351, "loss": 2.9761, "learning_rate": 3.3775185577942735e-07, "epoch": 2.0422960725075527 }, { "current_steps": 1352, "loss": 2.6124, "learning_rate": 3.3722163308589607e-07, "epoch": 2.0438066465256797 }, { "current_steps": 1353, "loss": 2.8982, "learning_rate": 3.366914103923648e-07, "epoch": 2.0453172205438066 }, { "current_steps": 1354, "loss": 2.8904, "learning_rate": 3.361611876988335e-07, "epoch": 2.0468277945619335 }, { "current_steps": 1355, "loss": 3.1752, "learning_rate": 3.356309650053022e-07, "epoch": 2.0483383685800605 }, { "current_steps": 1356, "loss": 2.9027, "learning_rate": 3.3510074231177093e-07, "epoch": 2.0498489425981874 }, { "current_steps": 1357, "loss": 2.9032, "learning_rate": 3.3457051961823965e-07, "epoch": 2.0513595166163143 }, { "current_steps": 1358, "loss": 2.9716, "learning_rate": 3.3404029692470836e-07, "epoch": 2.0528700906344413 }, { "current_steps": 1359, "loss": 2.9792, "learning_rate": 3.3351007423117703e-07, "epoch": 2.054380664652568 }, { "current_steps": 1360, "loss": 2.8627, "learning_rate": 3.329798515376458e-07, "epoch": 2.0558912386706947 }, { "current_steps": 1361, "loss": 3.2493, "learning_rate": 3.3244962884411456e-07, "epoch": 2.0574018126888216 }, { "current_steps": 1362, "loss": 3.0259, "learning_rate": 3.3191940615058323e-07, "epoch": 2.0589123867069485 }, { "current_steps": 1363, "loss": 2.9988, "learning_rate": 3.3138918345705194e-07, "epoch": 2.0604229607250755 }, { "current_steps": 1364, "loss": 3.1801, "learning_rate": 3.3085896076352066e-07, "epoch": 2.0619335347432024 }, { "current_steps": 1365, "loss": 2.504, "learning_rate": 3.303287380699894e-07, "epoch": 2.0634441087613293 }, { "current_steps": 1366, "loss": 2.4585, "learning_rate": 3.297985153764581e-07, "epoch": 2.0649546827794563 }, { "current_steps": 1367, "loss": 1.8878, "learning_rate": 3.292682926829268e-07, "epoch": 2.066465256797583 }, { "current_steps": 1368, "loss": 2.8641, "learning_rate": 3.287380699893955e-07, "epoch": 2.06797583081571 }, { "current_steps": 1369, "loss": 2.9, "learning_rate": 3.2820784729586424e-07, "epoch": 2.069486404833837 }, { "current_steps": 1370, "loss": 3.0279, "learning_rate": 3.2767762460233295e-07, "epoch": 2.0709969788519635 }, { "current_steps": 1371, "loss": 3.0085, "learning_rate": 3.2714740190880167e-07, "epoch": 2.0725075528700905 }, { "current_steps": 1372, "loss": 2.8525, "learning_rate": 3.266171792152704e-07, "epoch": 2.0740181268882174 }, { "current_steps": 1373, "loss": 2.9782, "learning_rate": 3.260869565217391e-07, "epoch": 2.0755287009063443 }, { "current_steps": 1374, "loss": 2.4857, "learning_rate": 3.255567338282078e-07, "epoch": 2.0770392749244713 }, { "current_steps": 1375, "loss": 2.9204, "learning_rate": 3.2502651113467653e-07, "epoch": 2.078549848942598 }, { "current_steps": 1376, "loss": 3.0433, "learning_rate": 3.244962884411453e-07, "epoch": 2.080060422960725 }, { "current_steps": 1377, "loss": 2.9001, "learning_rate": 3.2396606574761397e-07, "epoch": 2.081570996978852 }, { "current_steps": 1378, "loss": 2.8464, "learning_rate": 3.2343584305408273e-07, "epoch": 2.083081570996979 }, { "current_steps": 1379, "loss": 2.8298, "learning_rate": 3.229056203605514e-07, "epoch": 2.084592145015106 }, { "current_steps": 1380, "loss": 2.9859, "learning_rate": 3.2237539766702017e-07, "epoch": 2.0861027190332324 }, { "current_steps": 1381, "loss": 2.3218, "learning_rate": 3.2184517497348883e-07, "epoch": 2.0876132930513593 }, { "current_steps": 1382, "loss": 2.9048, "learning_rate": 3.213149522799576e-07, "epoch": 2.0891238670694863 }, { "current_steps": 1383, "loss": 2.7674, "learning_rate": 3.2078472958642626e-07, "epoch": 2.090634441087613 }, { "current_steps": 1384, "loss": 3.087, "learning_rate": 3.2025450689289503e-07, "epoch": 2.09214501510574 }, { "current_steps": 1385, "loss": 2.9192, "learning_rate": 3.197242841993637e-07, "epoch": 2.093655589123867 }, { "current_steps": 1386, "loss": 3.0376, "learning_rate": 3.1919406150583246e-07, "epoch": 2.095166163141994 }, { "current_steps": 1387, "loss": 2.79, "learning_rate": 3.186638388123011e-07, "epoch": 2.096676737160121 }, { "current_steps": 1388, "loss": 2.8143, "learning_rate": 3.181336161187699e-07, "epoch": 2.098187311178248 }, { "current_steps": 1389, "loss": 3.077, "learning_rate": 3.1760339342523856e-07, "epoch": 2.099697885196375 }, { "current_steps": 1390, "loss": 2.8832, "learning_rate": 3.170731707317073e-07, "epoch": 2.1012084592145017 }, { "current_steps": 1391, "loss": 3.012, "learning_rate": 3.16542948038176e-07, "epoch": 2.102719033232628 }, { "current_steps": 1392, "loss": 3.0334, "learning_rate": 3.1601272534464476e-07, "epoch": 2.104229607250755 }, { "current_steps": 1393, "loss": 3.0294, "learning_rate": 3.1548250265111347e-07, "epoch": 2.105740181268882 }, { "current_steps": 1394, "loss": 3.0511, "learning_rate": 3.149522799575822e-07, "epoch": 2.107250755287009 }, { "current_steps": 1395, "loss": 2.4776, "learning_rate": 3.144220572640509e-07, "epoch": 2.108761329305136 }, { "current_steps": 1396, "loss": 3.2582, "learning_rate": 3.138918345705196e-07, "epoch": 2.110271903323263 }, { "current_steps": 1397, "loss": 2.9826, "learning_rate": 3.1336161187698834e-07, "epoch": 2.11178247734139 }, { "current_steps": 1398, "loss": 3.002, "learning_rate": 3.1283138918345705e-07, "epoch": 2.1132930513595167 }, { "current_steps": 1399, "loss": 2.7149, "learning_rate": 3.1230116648992577e-07, "epoch": 2.1148036253776437 }, { "current_steps": 1400, "loss": 2.768, "learning_rate": 3.117709437963945e-07, "epoch": 2.1163141993957706 }, { "current_steps": 1401, "loss": 3.1165, "learning_rate": 3.112407211028632e-07, "epoch": 2.117824773413897 }, { "current_steps": 1402, "loss": 2.9925, "learning_rate": 3.107104984093319e-07, "epoch": 2.119335347432024 }, { "current_steps": 1403, "loss": 2.8689, "learning_rate": 3.1018027571580063e-07, "epoch": 2.120845921450151 }, { "current_steps": 1404, "loss": 3.1892, "learning_rate": 3.096500530222693e-07, "epoch": 2.122356495468278 }, { "current_steps": 1405, "loss": 3.1442, "learning_rate": 3.0911983032873806e-07, "epoch": 2.123867069486405 }, { "current_steps": 1406, "loss": 2.429, "learning_rate": 3.085896076352067e-07, "epoch": 2.1253776435045317 }, { "current_steps": 1407, "loss": 3.0257, "learning_rate": 3.080593849416755e-07, "epoch": 2.1268882175226587 }, { "current_steps": 1408, "loss": 2.8809, "learning_rate": 3.0752916224814416e-07, "epoch": 2.1283987915407856 }, { "current_steps": 1409, "loss": 2.9666, "learning_rate": 3.069989395546129e-07, "epoch": 2.1299093655589125 }, { "current_steps": 1410, "loss": 2.8281, "learning_rate": 3.0646871686108164e-07, "epoch": 2.1314199395770395 }, { "current_steps": 1411, "loss": 3.168, "learning_rate": 3.0593849416755036e-07, "epoch": 2.1329305135951664 }, { "current_steps": 1412, "loss": 2.8572, "learning_rate": 3.054082714740191e-07, "epoch": 2.134441087613293 }, { "current_steps": 1413, "loss": 3.003, "learning_rate": 3.048780487804878e-07, "epoch": 2.13595166163142 }, { "current_steps": 1414, "loss": 2.77, "learning_rate": 3.043478260869565e-07, "epoch": 2.1374622356495467 }, { "current_steps": 1415, "loss": 3.0762, "learning_rate": 3.038176033934252e-07, "epoch": 2.1389728096676737 }, { "current_steps": 1416, "loss": 3.132, "learning_rate": 3.0328738069989394e-07, "epoch": 2.1404833836858006 }, { "current_steps": 1417, "loss": 3.0142, "learning_rate": 3.0275715800636265e-07, "epoch": 2.1419939577039275 }, { "current_steps": 1418, "loss": 3.117, "learning_rate": 3.0222693531283137e-07, "epoch": 2.1435045317220545 }, { "current_steps": 1419, "loss": 2.8927, "learning_rate": 3.016967126193001e-07, "epoch": 2.1450151057401814 }, { "current_steps": 1420, "loss": 2.9227, "learning_rate": 3.011664899257688e-07, "epoch": 2.1465256797583083 }, { "current_steps": 1421, "loss": 2.9973, "learning_rate": 3.006362672322375e-07, "epoch": 2.148036253776435 }, { "current_steps": 1422, "loss": 2.8912, "learning_rate": 3.0010604453870623e-07, "epoch": 2.1495468277945617 }, { "current_steps": 1423, "loss": 2.9818, "learning_rate": 2.9957582184517495e-07, "epoch": 2.1510574018126887 }, { "current_steps": 1424, "loss": 2.4071, "learning_rate": 2.9904559915164366e-07, "epoch": 2.1525679758308156 }, { "current_steps": 1425, "loss": 2.8711, "learning_rate": 2.985153764581124e-07, "epoch": 2.1540785498489425 }, { "current_steps": 1426, "loss": 3.1843, "learning_rate": 2.979851537645811e-07, "epoch": 2.1555891238670695 }, { "current_steps": 1427, "loss": 3.0332, "learning_rate": 2.9745493107104986e-07, "epoch": 2.1570996978851964 }, { "current_steps": 1428, "loss": 2.981, "learning_rate": 2.9692470837751853e-07, "epoch": 2.1586102719033233 }, { "current_steps": 1429, "loss": 3.0876, "learning_rate": 2.963944856839873e-07, "epoch": 2.1601208459214503 }, { "current_steps": 1430, "loss": 3.1866, "learning_rate": 2.9586426299045596e-07, "epoch": 2.161631419939577 }, { "current_steps": 1431, "loss": 2.9309, "learning_rate": 2.9533404029692473e-07, "epoch": 2.163141993957704 }, { "current_steps": 1432, "loss": 3.0347, "learning_rate": 2.948038176033934e-07, "epoch": 2.1646525679758306 }, { "current_steps": 1433, "loss": 2.9721, "learning_rate": 2.9427359490986216e-07, "epoch": 2.1661631419939575 }, { "current_steps": 1434, "loss": 3.0952, "learning_rate": 2.937433722163308e-07, "epoch": 2.1676737160120845 }, { "current_steps": 1435, "loss": 3.0751, "learning_rate": 2.932131495227996e-07, "epoch": 2.1691842900302114 }, { "current_steps": 1436, "loss": 2.927, "learning_rate": 2.9268292682926825e-07, "epoch": 2.1706948640483383 }, { "current_steps": 1437, "loss": 2.5054, "learning_rate": 2.92152704135737e-07, "epoch": 2.1722054380664653 }, { "current_steps": 1438, "loss": 3.0626, "learning_rate": 2.916224814422057e-07, "epoch": 2.173716012084592 }, { "current_steps": 1439, "loss": 2.7461, "learning_rate": 2.9109225874867446e-07, "epoch": 2.175226586102719 }, { "current_steps": 1440, "loss": 2.9681, "learning_rate": 2.905620360551431e-07, "epoch": 2.176737160120846 }, { "current_steps": 1441, "loss": 2.8997, "learning_rate": 2.900318133616119e-07, "epoch": 2.178247734138973 }, { "current_steps": 1442, "loss": 3.1472, "learning_rate": 2.8950159066808055e-07, "epoch": 2.1797583081570995 }, { "current_steps": 1443, "loss": 2.9166, "learning_rate": 2.889713679745493e-07, "epoch": 2.1812688821752264 }, { "current_steps": 1444, "loss": 3.0052, "learning_rate": 2.8844114528101803e-07, "epoch": 2.1827794561933533 }, { "current_steps": 1445, "loss": 3.1859, "learning_rate": 2.8791092258748675e-07, "epoch": 2.1842900302114803 }, { "current_steps": 1446, "loss": 2.9732, "learning_rate": 2.8738069989395547e-07, "epoch": 2.185800604229607 }, { "current_steps": 1447, "loss": 2.9773, "learning_rate": 2.868504772004242e-07, "epoch": 2.187311178247734 }, { "current_steps": 1448, "loss": 2.9693, "learning_rate": 2.863202545068929e-07, "epoch": 2.188821752265861 }, { "current_steps": 1449, "loss": 2.9008, "learning_rate": 2.8579003181336156e-07, "epoch": 2.190332326283988 }, { "current_steps": 1450, "loss": 3.093, "learning_rate": 2.8525980911983033e-07, "epoch": 2.191842900302115 }, { "current_steps": 1451, "loss": 2.838, "learning_rate": 2.84729586426299e-07, "epoch": 2.193353474320242 }, { "current_steps": 1452, "loss": 3.1251, "learning_rate": 2.8419936373276776e-07, "epoch": 2.194864048338369 }, { "current_steps": 1453, "loss": 3.1252, "learning_rate": 2.836691410392364e-07, "epoch": 2.1963746223564953 }, { "current_steps": 1454, "loss": 2.5353, "learning_rate": 2.831389183457052e-07, "epoch": 2.197885196374622 }, { "current_steps": 1455, "loss": 2.911, "learning_rate": 2.8260869565217386e-07, "epoch": 2.199395770392749 }, { "current_steps": 1456, "loss": 3.0334, "learning_rate": 2.820784729586426e-07, "epoch": 2.200906344410876 }, { "current_steps": 1457, "loss": 3.0871, "learning_rate": 2.815482502651113e-07, "epoch": 2.202416918429003 }, { "current_steps": 1458, "loss": 3.115, "learning_rate": 2.8101802757158006e-07, "epoch": 2.20392749244713 }, { "current_steps": 1459, "loss": 3.124, "learning_rate": 2.8048780487804877e-07, "epoch": 2.205438066465257 }, { "current_steps": 1460, "loss": 3.0949, "learning_rate": 2.799575821845175e-07, "epoch": 2.206948640483384 }, { "current_steps": 1461, "loss": 2.8581, "learning_rate": 2.794273594909862e-07, "epoch": 2.2084592145015107 }, { "current_steps": 1462, "loss": 2.9666, "learning_rate": 2.788971367974549e-07, "epoch": 2.2099697885196377 }, { "current_steps": 1463, "loss": 3.033, "learning_rate": 2.7836691410392364e-07, "epoch": 2.211480362537764 }, { "current_steps": 1464, "loss": 2.8995, "learning_rate": 2.7783669141039235e-07, "epoch": 2.212990936555891 }, { "current_steps": 1465, "loss": 3.0165, "learning_rate": 2.7730646871686107e-07, "epoch": 2.214501510574018 }, { "current_steps": 1466, "loss": 2.4119, "learning_rate": 2.767762460233298e-07, "epoch": 2.216012084592145 }, { "current_steps": 1467, "loss": 2.6998, "learning_rate": 2.762460233297985e-07, "epoch": 2.217522658610272 }, { "current_steps": 1468, "loss": 2.9498, "learning_rate": 2.757158006362672e-07, "epoch": 2.219033232628399 }, { "current_steps": 1469, "loss": 2.9901, "learning_rate": 2.7518557794273593e-07, "epoch": 2.2205438066465257 }, { "current_steps": 1470, "loss": 3.0045, "learning_rate": 2.7465535524920465e-07, "epoch": 2.2220543806646527 }, { "current_steps": 1471, "loss": 2.9205, "learning_rate": 2.7412513255567336e-07, "epoch": 2.2235649546827796 }, { "current_steps": 1472, "loss": 2.7696, "learning_rate": 2.735949098621421e-07, "epoch": 2.2250755287009065 }, { "current_steps": 1473, "loss": 3.2643, "learning_rate": 2.730646871686108e-07, "epoch": 2.2265861027190335 }, { "current_steps": 1474, "loss": 2.905, "learning_rate": 2.725344644750795e-07, "epoch": 2.22809667673716 }, { "current_steps": 1475, "loss": 2.9568, "learning_rate": 2.7200424178154823e-07, "epoch": 2.229607250755287 }, { "current_steps": 1476, "loss": 2.9691, "learning_rate": 2.71474019088017e-07, "epoch": 2.231117824773414 }, { "current_steps": 1477, "loss": 3.0642, "learning_rate": 2.7094379639448566e-07, "epoch": 2.2326283987915407 }, { "current_steps": 1478, "loss": 2.6817, "learning_rate": 2.7041357370095443e-07, "epoch": 2.2341389728096677 }, { "current_steps": 1479, "loss": 2.8455, "learning_rate": 2.698833510074231e-07, "epoch": 2.2356495468277946 }, { "current_steps": 1480, "loss": 2.7563, "learning_rate": 2.6935312831389186e-07, "epoch": 2.2371601208459215 }, { "current_steps": 1481, "loss": 3.0935, "learning_rate": 2.688229056203605e-07, "epoch": 2.2386706948640485 }, { "current_steps": 1482, "loss": 2.1801, "learning_rate": 2.682926829268293e-07, "epoch": 2.2401812688821754 }, { "current_steps": 1483, "loss": 3.0399, "learning_rate": 2.6776246023329795e-07, "epoch": 2.2416918429003023 }, { "current_steps": 1484, "loss": 2.7325, "learning_rate": 2.672322375397667e-07, "epoch": 2.243202416918429 }, { "current_steps": 1485, "loss": 2.9746, "learning_rate": 2.667020148462354e-07, "epoch": 2.2447129909365557 }, { "current_steps": 1486, "loss": 2.9142, "learning_rate": 2.6617179215270415e-07, "epoch": 2.2462235649546827 }, { "current_steps": 1487, "loss": 2.8804, "learning_rate": 2.656415694591728e-07, "epoch": 2.2477341389728096 }, { "current_steps": 1488, "loss": 3.2683, "learning_rate": 2.651113467656416e-07, "epoch": 2.2492447129909365 }, { "current_steps": 1489, "loss": 3.1025, "learning_rate": 2.6458112407211025e-07, "epoch": 2.2507552870090635 }, { "current_steps": 1490, "loss": 3.1795, "learning_rate": 2.64050901378579e-07, "epoch": 2.2522658610271904 }, { "current_steps": 1491, "loss": 2.7756, "learning_rate": 2.635206786850477e-07, "epoch": 2.2537764350453173 }, { "current_steps": 1492, "loss": 3.0307, "learning_rate": 2.6299045599151645e-07, "epoch": 2.2552870090634443 }, { "current_steps": 1493, "loss": 3.0719, "learning_rate": 2.6246023329798517e-07, "epoch": 2.256797583081571 }, { "current_steps": 1494, "loss": 3.0361, "learning_rate": 2.6193001060445383e-07, "epoch": 2.258308157099698 }, { "current_steps": 1495, "loss": 3.0995, "learning_rate": 2.613997879109226e-07, "epoch": 2.2598187311178246 }, { "current_steps": 1496, "loss": 2.9234, "learning_rate": 2.6086956521739126e-07, "epoch": 2.2613293051359515 }, { "current_steps": 1497, "loss": 2.8779, "learning_rate": 2.6033934252386003e-07, "epoch": 2.2628398791540785 }, { "current_steps": 1498, "loss": 2.9209, "learning_rate": 2.598091198303287e-07, "epoch": 2.2643504531722054 }, { "current_steps": 1499, "loss": 2.9894, "learning_rate": 2.5927889713679746e-07, "epoch": 2.2658610271903323 }, { "current_steps": 1500, "loss": 3.0221, "learning_rate": 2.587486744432661e-07, "epoch": 2.2673716012084593 }, { "current_steps": 1501, "loss": 2.7052, "learning_rate": 2.582184517497349e-07, "epoch": 2.268882175226586 }, { "current_steps": 1502, "loss": 2.7213, "learning_rate": 2.5768822905620356e-07, "epoch": 2.270392749244713 }, { "current_steps": 1503, "loss": 2.8471, "learning_rate": 2.571580063626723e-07, "epoch": 2.27190332326284 }, { "current_steps": 1504, "loss": 2.9732, "learning_rate": 2.56627783669141e-07, "epoch": 2.2734138972809665 }, { "current_steps": 1505, "loss": 2.8034, "learning_rate": 2.5609756097560976e-07, "epoch": 2.2749244712990935 }, { "current_steps": 1506, "loss": 3.2221, "learning_rate": 2.555673382820784e-07, "epoch": 2.2764350453172204 }, { "current_steps": 1507, "loss": 2.816, "learning_rate": 2.550371155885472e-07, "epoch": 2.2779456193353473 }, { "current_steps": 1508, "loss": 2.9744, "learning_rate": 2.5450689289501585e-07, "epoch": 2.2794561933534743 }, { "current_steps": 1509, "loss": 3.0932, "learning_rate": 2.539766702014846e-07, "epoch": 2.280966767371601 }, { "current_steps": 1510, "loss": 2.7198, "learning_rate": 2.5344644750795334e-07, "epoch": 2.282477341389728 }, { "current_steps": 1511, "loss": 3.0977, "learning_rate": 2.5291622481442205e-07, "epoch": 2.283987915407855 }, { "current_steps": 1512, "loss": 2.8511, "learning_rate": 2.5238600212089077e-07, "epoch": 2.285498489425982 }, { "current_steps": 1513, "loss": 3.0167, "learning_rate": 2.518557794273595e-07, "epoch": 2.287009063444109 }, { "current_steps": 1514, "loss": 3.0085, "learning_rate": 2.513255567338282e-07, "epoch": 2.288519637462236 }, { "current_steps": 1515, "loss": 3.1017, "learning_rate": 2.507953340402969e-07, "epoch": 2.290030211480363 }, { "current_steps": 1516, "loss": 3.0252, "learning_rate": 2.5026511134676563e-07, "epoch": 2.2915407854984893 }, { "current_steps": 1517, "loss": 2.811, "learning_rate": 2.4973488865323435e-07, "epoch": 2.293051359516616 }, { "current_steps": 1518, "loss": 2.7579, "learning_rate": 2.4920466595970306e-07, "epoch": 2.294561933534743 }, { "current_steps": 1519, "loss": 2.888, "learning_rate": 2.486744432661718e-07, "epoch": 2.29607250755287 }, { "current_steps": 1520, "loss": 2.9921, "learning_rate": 2.481442205726405e-07, "epoch": 2.297583081570997 }, { "current_steps": 1521, "loss": 2.9381, "learning_rate": 2.476139978791092e-07, "epoch": 2.299093655589124 }, { "current_steps": 1522, "loss": 3.2375, "learning_rate": 2.470837751855779e-07, "epoch": 2.300604229607251 }, { "current_steps": 1523, "loss": 2.9129, "learning_rate": 2.4655355249204664e-07, "epoch": 2.302114803625378 }, { "current_steps": 1524, "loss": 2.9648, "learning_rate": 2.4602332979851536e-07, "epoch": 2.3036253776435047 }, { "current_steps": 1525, "loss": 2.9395, "learning_rate": 2.454931071049841e-07, "epoch": 2.305135951661631 }, { "current_steps": 1526, "loss": 2.9632, "learning_rate": 2.449628844114528e-07, "epoch": 2.306646525679758 }, { "current_steps": 1527, "loss": 3.0419, "learning_rate": 2.444326617179215e-07, "epoch": 2.308157099697885 }, { "current_steps": 1528, "loss": 2.7947, "learning_rate": 2.439024390243902e-07, "epoch": 2.309667673716012 }, { "current_steps": 1529, "loss": 3.0674, "learning_rate": 2.4337221633085894e-07, "epoch": 2.311178247734139 }, { "current_steps": 1530, "loss": 2.9957, "learning_rate": 2.4284199363732765e-07, "epoch": 2.312688821752266 }, { "current_steps": 1531, "loss": 2.9322, "learning_rate": 2.4231177094379637e-07, "epoch": 2.314199395770393 }, { "current_steps": 1532, "loss": 2.9148, "learning_rate": 2.417815482502651e-07, "epoch": 2.3157099697885197 }, { "current_steps": 1533, "loss": 2.9209, "learning_rate": 2.412513255567338e-07, "epoch": 2.3172205438066467 }, { "current_steps": 1534, "loss": 3.0346, "learning_rate": 2.4072110286320257e-07, "epoch": 2.3187311178247736 }, { "current_steps": 1535, "loss": 3.0656, "learning_rate": 2.401908801696713e-07, "epoch": 2.3202416918429005 }, { "current_steps": 1536, "loss": 3.0893, "learning_rate": 2.3966065747614e-07, "epoch": 2.321752265861027 }, { "current_steps": 1537, "loss": 3.0211, "learning_rate": 2.391304347826087e-07, "epoch": 2.323262839879154 }, { "current_steps": 1538, "loss": 2.9478, "learning_rate": 2.3860021208907743e-07, "epoch": 2.324773413897281 }, { "current_steps": 1539, "loss": 2.8997, "learning_rate": 2.3806998939554612e-07, "epoch": 2.326283987915408 }, { "current_steps": 1540, "loss": 2.598, "learning_rate": 2.3753976670201484e-07, "epoch": 2.3277945619335347 }, { "current_steps": 1541, "loss": 2.798, "learning_rate": 2.3700954400848355e-07, "epoch": 2.3293051359516617 }, { "current_steps": 1542, "loss": 2.9423, "learning_rate": 2.3647932131495227e-07, "epoch": 2.3308157099697886 }, { "current_steps": 1543, "loss": 3.0865, "learning_rate": 2.3594909862142099e-07, "epoch": 2.3323262839879155 }, { "current_steps": 1544, "loss": 2.6892, "learning_rate": 2.354188759278897e-07, "epoch": 2.3338368580060425 }, { "current_steps": 1545, "loss": 3.0822, "learning_rate": 2.3488865323435842e-07, "epoch": 2.335347432024169 }, { "current_steps": 1546, "loss": 3.1105, "learning_rate": 2.3435843054082713e-07, "epoch": 2.336858006042296 }, { "current_steps": 1547, "loss": 3.0123, "learning_rate": 2.3382820784729585e-07, "epoch": 2.338368580060423 }, { "current_steps": 1548, "loss": 2.8546, "learning_rate": 2.3329798515376457e-07, "epoch": 2.3398791540785497 }, { "current_steps": 1549, "loss": 2.9216, "learning_rate": 2.3276776246023328e-07, "epoch": 2.3413897280966767 }, { "current_steps": 1550, "loss": 2.9761, "learning_rate": 2.3223753976670202e-07, "epoch": 2.3429003021148036 }, { "current_steps": 1551, "loss": 2.7588, "learning_rate": 2.3170731707317074e-07, "epoch": 2.3444108761329305 }, { "current_steps": 1552, "loss": 3.024, "learning_rate": 2.3117709437963946e-07, "epoch": 2.3459214501510575 }, { "current_steps": 1553, "loss": 2.9984, "learning_rate": 2.3064687168610817e-07, "epoch": 2.3474320241691844 }, { "current_steps": 1554, "loss": 3.0442, "learning_rate": 2.301166489925769e-07, "epoch": 2.3489425981873113 }, { "current_steps": 1555, "loss": 2.9495, "learning_rate": 2.295864262990456e-07, "epoch": 2.3504531722054383 }, { "current_steps": 1556, "loss": 3.0512, "learning_rate": 2.2905620360551432e-07, "epoch": 2.351963746223565 }, { "current_steps": 1557, "loss": 3.0835, "learning_rate": 2.2852598091198303e-07, "epoch": 2.3534743202416917 }, { "current_steps": 1558, "loss": 2.2911, "learning_rate": 2.2799575821845175e-07, "epoch": 2.3549848942598186 }, { "current_steps": 1559, "loss": 3.0784, "learning_rate": 2.2746553552492047e-07, "epoch": 2.3564954682779455 }, { "current_steps": 1560, "loss": 3.0934, "learning_rate": 2.2693531283138918e-07, "epoch": 2.3580060422960725 }, { "current_steps": 1561, "loss": 2.499, "learning_rate": 2.2640509013785787e-07, "epoch": 2.3595166163141994 }, { "current_steps": 1562, "loss": 3.2606, "learning_rate": 2.258748674443266e-07, "epoch": 2.3610271903323263 }, { "current_steps": 1563, "loss": 2.9355, "learning_rate": 2.253446447507953e-07, "epoch": 2.3625377643504533 }, { "current_steps": 1564, "loss": 3.0685, "learning_rate": 2.2481442205726402e-07, "epoch": 2.36404833836858 }, { "current_steps": 1565, "loss": 2.87, "learning_rate": 2.2428419936373274e-07, "epoch": 2.365558912386707 }, { "current_steps": 1566, "loss": 2.9643, "learning_rate": 2.2375397667020145e-07, "epoch": 2.3670694864048336 }, { "current_steps": 1567, "loss": 2.3665, "learning_rate": 2.232237539766702e-07, "epoch": 2.3685800604229605 }, { "current_steps": 1568, "loss": 3.0814, "learning_rate": 2.226935312831389e-07, "epoch": 2.3700906344410875 }, { "current_steps": 1569, "loss": 2.9984, "learning_rate": 2.2216330858960763e-07, "epoch": 2.3716012084592144 }, { "current_steps": 1570, "loss": 2.8213, "learning_rate": 2.2163308589607634e-07, "epoch": 2.3731117824773413 }, { "current_steps": 1571, "loss": 2.8461, "learning_rate": 2.2110286320254506e-07, "epoch": 2.3746223564954683 }, { "current_steps": 1572, "loss": 3.0246, "learning_rate": 2.2057264050901377e-07, "epoch": 2.376132930513595 }, { "current_steps": 1573, "loss": 2.9836, "learning_rate": 2.200424178154825e-07, "epoch": 2.377643504531722 }, { "current_steps": 1574, "loss": 3.2213, "learning_rate": 2.195121951219512e-07, "epoch": 2.379154078549849 }, { "current_steps": 1575, "loss": 2.9788, "learning_rate": 2.1898197242841992e-07, "epoch": 2.380664652567976 }, { "current_steps": 1576, "loss": 3.2071, "learning_rate": 2.1845174973488864e-07, "epoch": 2.382175226586103 }, { "current_steps": 1577, "loss": 2.9546, "learning_rate": 2.1792152704135735e-07, "epoch": 2.38368580060423 }, { "current_steps": 1578, "loss": 2.5353, "learning_rate": 2.1739130434782607e-07, "epoch": 2.3851963746223563 }, { "current_steps": 1579, "loss": 3.0954, "learning_rate": 2.1686108165429478e-07, "epoch": 2.3867069486404833 }, { "current_steps": 1580, "loss": 3.0557, "learning_rate": 2.163308589607635e-07, "epoch": 2.38821752265861 }, { "current_steps": 1581, "loss": 2.9073, "learning_rate": 2.1580063626723222e-07, "epoch": 2.389728096676737 }, { "current_steps": 1582, "loss": 3.1207, "learning_rate": 2.1527041357370093e-07, "epoch": 2.391238670694864 }, { "current_steps": 1583, "loss": 3.0087, "learning_rate": 2.1474019088016967e-07, "epoch": 2.392749244712991 }, { "current_steps": 1584, "loss": 2.9414, "learning_rate": 2.142099681866384e-07, "epoch": 2.394259818731118 }, { "current_steps": 1585, "loss": 2.7435, "learning_rate": 2.136797454931071e-07, "epoch": 2.395770392749245 }, { "current_steps": 1586, "loss": 3.0774, "learning_rate": 2.1314952279957582e-07, "epoch": 2.397280966767372 }, { "current_steps": 1587, "loss": 2.927, "learning_rate": 2.1261930010604454e-07, "epoch": 2.3987915407854983 }, { "current_steps": 1588, "loss": 2.8824, "learning_rate": 2.1208907741251325e-07, "epoch": 2.400302114803625 }, { "current_steps": 1589, "loss": 2.8408, "learning_rate": 2.1155885471898197e-07, "epoch": 2.401812688821752 }, { "current_steps": 1590, "loss": 2.9424, "learning_rate": 2.1102863202545068e-07, "epoch": 2.403323262839879 }, { "current_steps": 1591, "loss": 2.9665, "learning_rate": 2.104984093319194e-07, "epoch": 2.404833836858006 }, { "current_steps": 1592, "loss": 2.9924, "learning_rate": 2.0996818663838812e-07, "epoch": 2.406344410876133 }, { "current_steps": 1593, "loss": 2.7694, "learning_rate": 2.0943796394485683e-07, "epoch": 2.40785498489426 }, { "current_steps": 1594, "loss": 2.8801, "learning_rate": 2.0890774125132555e-07, "epoch": 2.409365558912387 }, { "current_steps": 1595, "loss": 3.0329, "learning_rate": 2.0837751855779426e-07, "epoch": 2.4108761329305137 }, { "current_steps": 1596, "loss": 2.5639, "learning_rate": 2.0784729586426298e-07, "epoch": 2.4123867069486407 }, { "current_steps": 1597, "loss": 3.0518, "learning_rate": 2.073170731707317e-07, "epoch": 2.4138972809667676 }, { "current_steps": 1598, "loss": 2.9236, "learning_rate": 2.067868504772004e-07, "epoch": 2.415407854984894 }, { "current_steps": 1599, "loss": 2.8624, "learning_rate": 2.0625662778366913e-07, "epoch": 2.416918429003021 }, { "current_steps": 1600, "loss": 2.9981, "learning_rate": 2.0572640509013787e-07, "epoch": 2.418429003021148 }, { "current_steps": 1601, "loss": 2.8338, "learning_rate": 2.0519618239660659e-07, "epoch": 2.419939577039275 }, { "current_steps": 1602, "loss": 3.1091, "learning_rate": 2.046659597030753e-07, "epoch": 2.421450151057402 }, { "current_steps": 1603, "loss": 2.9502, "learning_rate": 2.0413573700954402e-07, "epoch": 2.4229607250755287 }, { "current_steps": 1604, "loss": 2.9127, "learning_rate": 2.0360551431601273e-07, "epoch": 2.4244712990936557 }, { "current_steps": 1605, "loss": 3.1176, "learning_rate": 2.0307529162248145e-07, "epoch": 2.4259818731117826 }, { "current_steps": 1606, "loss": 3.1529, "learning_rate": 2.0254506892895017e-07, "epoch": 2.4274924471299095 }, { "current_steps": 1607, "loss": 2.9226, "learning_rate": 2.0201484623541885e-07, "epoch": 2.429003021148036 }, { "current_steps": 1608, "loss": 2.9755, "learning_rate": 2.0148462354188757e-07, "epoch": 2.430513595166163 }, { "current_steps": 1609, "loss": 2.3473, "learning_rate": 2.0095440084835629e-07, "epoch": 2.43202416918429 }, { "current_steps": 1610, "loss": 3.0144, "learning_rate": 2.00424178154825e-07, "epoch": 2.433534743202417 }, { "current_steps": 1611, "loss": 3.0732, "learning_rate": 1.9989395546129372e-07, "epoch": 2.4350453172205437 }, { "current_steps": 1612, "loss": 2.9595, "learning_rate": 1.9936373276776243e-07, "epoch": 2.4365558912386707 }, { "current_steps": 1613, "loss": 2.4932, "learning_rate": 1.9883351007423115e-07, "epoch": 2.4380664652567976 }, { "current_steps": 1614, "loss": 3.1989, "learning_rate": 1.9830328738069987e-07, "epoch": 2.4395770392749245 }, { "current_steps": 1615, "loss": 2.9832, "learning_rate": 1.9777306468716858e-07, "epoch": 2.4410876132930515 }, { "current_steps": 1616, "loss": 3.0117, "learning_rate": 1.9724284199363732e-07, "epoch": 2.4425981873111784 }, { "current_steps": 1617, "loss": 2.25, "learning_rate": 1.9671261930010604e-07, "epoch": 2.4441087613293053 }, { "current_steps": 1618, "loss": 3.0307, "learning_rate": 1.9618239660657476e-07, "epoch": 2.4456193353474323 }, { "current_steps": 1619, "loss": 2.8028, "learning_rate": 1.9565217391304347e-07, "epoch": 2.4471299093655587 }, { "current_steps": 1620, "loss": 2.8706, "learning_rate": 1.951219512195122e-07, "epoch": 2.4486404833836857 }, { "current_steps": 1621, "loss": 3.0146, "learning_rate": 1.945917285259809e-07, "epoch": 2.4501510574018126 }, { "current_steps": 1622, "loss": 3.0197, "learning_rate": 1.9406150583244962e-07, "epoch": 2.4516616314199395 }, { "current_steps": 1623, "loss": 2.9147, "learning_rate": 1.9353128313891834e-07, "epoch": 2.4531722054380665 }, { "current_steps": 1624, "loss": 2.9364, "learning_rate": 1.9300106044538705e-07, "epoch": 2.4546827794561934 }, { "current_steps": 1625, "loss": 2.9045, "learning_rate": 1.9247083775185577e-07, "epoch": 2.4561933534743203 }, { "current_steps": 1626, "loss": 2.9614, "learning_rate": 1.9194061505832448e-07, "epoch": 2.4577039274924473 }, { "current_steps": 1627, "loss": 3.018, "learning_rate": 1.914103923647932e-07, "epoch": 2.459214501510574 }, { "current_steps": 1628, "loss": 2.9721, "learning_rate": 1.9088016967126191e-07, "epoch": 2.4607250755287007 }, { "current_steps": 1629, "loss": 2.3248, "learning_rate": 1.9034994697773063e-07, "epoch": 2.4622356495468276 }, { "current_steps": 1630, "loss": 2.765, "learning_rate": 1.8981972428419935e-07, "epoch": 2.4637462235649545 }, { "current_steps": 1631, "loss": 2.8288, "learning_rate": 1.8928950159066806e-07, "epoch": 2.4652567975830815 }, { "current_steps": 1632, "loss": 2.7947, "learning_rate": 1.8875927889713678e-07, "epoch": 2.4667673716012084 }, { "current_steps": 1633, "loss": 2.9891, "learning_rate": 1.8822905620360552e-07, "epoch": 2.4682779456193353 }, { "current_steps": 1634, "loss": 3.082, "learning_rate": 1.8769883351007424e-07, "epoch": 2.4697885196374623 }, { "current_steps": 1635, "loss": 2.9986, "learning_rate": 1.8716861081654295e-07, "epoch": 2.471299093655589 }, { "current_steps": 1636, "loss": 2.9933, "learning_rate": 1.8663838812301167e-07, "epoch": 2.472809667673716 }, { "current_steps": 1637, "loss": 3.0289, "learning_rate": 1.8610816542948038e-07, "epoch": 2.474320241691843 }, { "current_steps": 1638, "loss": 2.9133, "learning_rate": 1.855779427359491e-07, "epoch": 2.47583081570997 }, { "current_steps": 1639, "loss": 2.5725, "learning_rate": 1.8504772004241782e-07, "epoch": 2.477341389728097 }, { "current_steps": 1640, "loss": 3.0534, "learning_rate": 1.8451749734888653e-07, "epoch": 2.4788519637462234 }, { "current_steps": 1641, "loss": 2.9152, "learning_rate": 1.8398727465535525e-07, "epoch": 2.4803625377643503 }, { "current_steps": 1642, "loss": 3.1145, "learning_rate": 1.8345705196182396e-07, "epoch": 2.4818731117824773 }, { "current_steps": 1643, "loss": 3.0405, "learning_rate": 1.8292682926829268e-07, "epoch": 2.483383685800604 }, { "current_steps": 1644, "loss": 2.8501, "learning_rate": 1.823966065747614e-07, "epoch": 2.484894259818731 }, { "current_steps": 1645, "loss": 2.6631, "learning_rate": 1.818663838812301e-07, "epoch": 2.486404833836858 }, { "current_steps": 1646, "loss": 2.9754, "learning_rate": 1.8133616118769883e-07, "epoch": 2.487915407854985 }, { "current_steps": 1647, "loss": 2.7988, "learning_rate": 1.8080593849416754e-07, "epoch": 2.489425981873112 }, { "current_steps": 1648, "loss": 3.1064, "learning_rate": 1.8027571580063626e-07, "epoch": 2.490936555891239 }, { "current_steps": 1649, "loss": 2.9314, "learning_rate": 1.79745493107105e-07, "epoch": 2.4924471299093653 }, { "current_steps": 1650, "loss": 2.7982, "learning_rate": 1.7921527041357372e-07, "epoch": 2.4939577039274923 }, { "current_steps": 1651, "loss": 2.9565, "learning_rate": 1.7868504772004243e-07, "epoch": 2.495468277945619 }, { "current_steps": 1652, "loss": 2.4873, "learning_rate": 1.7815482502651112e-07, "epoch": 2.496978851963746 }, { "current_steps": 1653, "loss": 2.7168, "learning_rate": 1.7762460233297984e-07, "epoch": 2.498489425981873 }, { "current_steps": 1654, "loss": 2.9843, "learning_rate": 1.7709437963944855e-07, "epoch": 2.5 }, { "current_steps": 1655, "loss": 2.9685, "learning_rate": 1.7656415694591727e-07, "epoch": 2.501510574018127 }, { "current_steps": 1656, "loss": 2.3108, "learning_rate": 1.7603393425238599e-07, "epoch": 2.503021148036254 }, { "current_steps": 1657, "loss": 3.1197, "learning_rate": 1.755037115588547e-07, "epoch": 2.504531722054381 }, { "current_steps": 1658, "loss": 3.0199, "learning_rate": 1.7497348886532342e-07, "epoch": 2.5060422960725077 }, { "current_steps": 1659, "loss": 3.1118, "learning_rate": 1.7444326617179213e-07, "epoch": 2.5075528700906347 }, { "current_steps": 1660, "loss": 3.025, "learning_rate": 1.7391304347826085e-07, "epoch": 2.5090634441087616 }, { "current_steps": 1661, "loss": 2.8618, "learning_rate": 1.7338282078472956e-07, "epoch": 2.510574018126888 }, { "current_steps": 1662, "loss": 3.043, "learning_rate": 1.7285259809119828e-07, "epoch": 2.512084592145015 }, { "current_steps": 1663, "loss": 2.9807, "learning_rate": 1.72322375397667e-07, "epoch": 2.513595166163142 }, { "current_steps": 1664, "loss": 2.8322, "learning_rate": 1.717921527041357e-07, "epoch": 2.515105740181269 }, { "current_steps": 1665, "loss": 2.9425, "learning_rate": 1.7126193001060443e-07, "epoch": 2.516616314199396 }, { "current_steps": 1666, "loss": 2.9077, "learning_rate": 1.7073170731707317e-07, "epoch": 2.5181268882175227 }, { "current_steps": 1667, "loss": 3.0191, "learning_rate": 1.7020148462354189e-07, "epoch": 2.5196374622356497 }, { "current_steps": 1668, "loss": 2.9413, "learning_rate": 1.696712619300106e-07, "epoch": 2.5211480362537766 }, { "current_steps": 1669, "loss": 2.8285, "learning_rate": 1.6914103923647932e-07, "epoch": 2.522658610271903 }, { "current_steps": 1670, "loss": 2.8357, "learning_rate": 1.6861081654294803e-07, "epoch": 2.52416918429003 }, { "current_steps": 1671, "loss": 2.6727, "learning_rate": 1.6808059384941675e-07, "epoch": 2.525679758308157 }, { "current_steps": 1672, "loss": 3.0076, "learning_rate": 1.6755037115588547e-07, "epoch": 2.527190332326284 }, { "current_steps": 1673, "loss": 3.0603, "learning_rate": 1.6702014846235418e-07, "epoch": 2.528700906344411 }, { "current_steps": 1674, "loss": 2.9549, "learning_rate": 1.664899257688229e-07, "epoch": 2.5302114803625377 }, { "current_steps": 1675, "loss": 2.9827, "learning_rate": 1.6595970307529161e-07, "epoch": 2.5317220543806647 }, { "current_steps": 1676, "loss": 2.9776, "learning_rate": 1.6542948038176033e-07, "epoch": 2.5332326283987916 }, { "current_steps": 1677, "loss": 2.8078, "learning_rate": 1.6489925768822905e-07, "epoch": 2.5347432024169185 }, { "current_steps": 1678, "loss": 3.1163, "learning_rate": 1.6436903499469776e-07, "epoch": 2.5362537764350455 }, { "current_steps": 1679, "loss": 2.9042, "learning_rate": 1.6383881230116648e-07, "epoch": 2.5377643504531724 }, { "current_steps": 1680, "loss": 2.6071, "learning_rate": 1.633085896076352e-07, "epoch": 2.5392749244712993 }, { "current_steps": 1681, "loss": 2.921, "learning_rate": 1.627783669141039e-07, "epoch": 2.5407854984894263 }, { "current_steps": 1682, "loss": 3.0543, "learning_rate": 1.6224814422057265e-07, "epoch": 2.5422960725075527 }, { "current_steps": 1683, "loss": 2.9539, "learning_rate": 1.6171792152704137e-07, "epoch": 2.5438066465256797 }, { "current_steps": 1684, "loss": 3.1212, "learning_rate": 1.6118769883351008e-07, "epoch": 2.5453172205438066 }, { "current_steps": 1685, "loss": 2.934, "learning_rate": 1.606574761399788e-07, "epoch": 2.5468277945619335 }, { "current_steps": 1686, "loss": 2.9337, "learning_rate": 1.6012725344644751e-07, "epoch": 2.5483383685800605 }, { "current_steps": 1687, "loss": 2.9206, "learning_rate": 1.5959703075291623e-07, "epoch": 2.5498489425981874 }, { "current_steps": 1688, "loss": 2.8159, "learning_rate": 1.5906680805938495e-07, "epoch": 2.5513595166163143 }, { "current_steps": 1689, "loss": 2.7428, "learning_rate": 1.5853658536585366e-07, "epoch": 2.552870090634441 }, { "current_steps": 1690, "loss": 2.8544, "learning_rate": 1.5800636267232238e-07, "epoch": 2.5543806646525677 }, { "current_steps": 1691, "loss": 2.9312, "learning_rate": 1.574761399787911e-07, "epoch": 2.5558912386706947 }, { "current_steps": 1692, "loss": 3.0904, "learning_rate": 1.569459172852598e-07, "epoch": 2.5574018126888216 }, { "current_steps": 1693, "loss": 2.9235, "learning_rate": 1.5641569459172853e-07, "epoch": 2.5589123867069485 }, { "current_steps": 1694, "loss": 2.9325, "learning_rate": 1.5588547189819724e-07, "epoch": 2.5604229607250755 }, { "current_steps": 1695, "loss": 2.953, "learning_rate": 1.5535524920466596e-07, "epoch": 2.5619335347432024 }, { "current_steps": 1696, "loss": 2.8501, "learning_rate": 1.5482502651113465e-07, "epoch": 2.5634441087613293 }, { "current_steps": 1697, "loss": 3.1307, "learning_rate": 1.5429480381760336e-07, "epoch": 2.5649546827794563 }, { "current_steps": 1698, "loss": 3.072, "learning_rate": 1.5376458112407208e-07, "epoch": 2.566465256797583 }, { "current_steps": 1699, "loss": 2.9131, "learning_rate": 1.5323435843054082e-07, "epoch": 2.56797583081571 }, { "current_steps": 1700, "loss": 3.2172, "learning_rate": 1.5270413573700954e-07, "epoch": 2.569486404833837 }, { "current_steps": 1701, "loss": 2.9902, "learning_rate": 1.5217391304347825e-07, "epoch": 2.570996978851964 }, { "current_steps": 1702, "loss": 2.6912, "learning_rate": 1.5164369034994697e-07, "epoch": 2.5725075528700905 }, { "current_steps": 1703, "loss": 2.9426, "learning_rate": 1.5111346765641568e-07, "epoch": 2.5740181268882174 }, { "current_steps": 1704, "loss": 3.0671, "learning_rate": 1.505832449628844e-07, "epoch": 2.5755287009063443 }, { "current_steps": 1705, "loss": 2.9862, "learning_rate": 1.5005302226935312e-07, "epoch": 2.5770392749244713 }, { "current_steps": 1706, "loss": 2.9179, "learning_rate": 1.4952279957582183e-07, "epoch": 2.578549848942598 }, { "current_steps": 1707, "loss": 3.1907, "learning_rate": 1.4899257688229055e-07, "epoch": 2.580060422960725 }, { "current_steps": 1708, "loss": 2.9362, "learning_rate": 1.4846235418875926e-07, "epoch": 2.581570996978852 }, { "current_steps": 1709, "loss": 2.8519, "learning_rate": 1.4793213149522798e-07, "epoch": 2.583081570996979 }, { "current_steps": 1710, "loss": 3.1644, "learning_rate": 1.474019088016967e-07, "epoch": 2.5845921450151055 }, { "current_steps": 1711, "loss": 2.9998, "learning_rate": 1.468716861081654e-07, "epoch": 2.5861027190332324 }, { "current_steps": 1712, "loss": 2.8967, "learning_rate": 1.4634146341463413e-07, "epoch": 2.5876132930513593 }, { "current_steps": 1713, "loss": 3.0513, "learning_rate": 1.4581124072110284e-07, "epoch": 2.5891238670694863 }, { "current_steps": 1714, "loss": 2.9667, "learning_rate": 1.4528101802757156e-07, "epoch": 2.590634441087613 }, { "current_steps": 1715, "loss": 2.8657, "learning_rate": 1.4475079533404028e-07, "epoch": 2.59214501510574 }, { "current_steps": 1716, "loss": 2.8812, "learning_rate": 1.4422057264050902e-07, "epoch": 2.593655589123867 }, { "current_steps": 1717, "loss": 3.0895, "learning_rate": 1.4369034994697773e-07, "epoch": 2.595166163141994 }, { "current_steps": 1718, "loss": 2.9389, "learning_rate": 1.4316012725344645e-07, "epoch": 2.596676737160121 }, { "current_steps": 1719, "loss": 2.7815, "learning_rate": 1.4262990455991517e-07, "epoch": 2.598187311178248 }, { "current_steps": 1720, "loss": 2.9379, "learning_rate": 1.4209968186638388e-07, "epoch": 2.599697885196375 }, { "current_steps": 1721, "loss": 3.013, "learning_rate": 1.415694591728526e-07, "epoch": 2.6012084592145017 }, { "current_steps": 1722, "loss": 3.0051, "learning_rate": 1.410392364793213e-07, "epoch": 2.6027190332326287 }, { "current_steps": 1723, "loss": 2.7263, "learning_rate": 1.4050901378579003e-07, "epoch": 2.604229607250755 }, { "current_steps": 1724, "loss": 2.8186, "learning_rate": 1.3997879109225874e-07, "epoch": 2.605740181268882 }, { "current_steps": 1725, "loss": 3.1671, "learning_rate": 1.3944856839872746e-07, "epoch": 2.607250755287009 }, { "current_steps": 1726, "loss": 3.0972, "learning_rate": 1.3891834570519618e-07, "epoch": 2.608761329305136 }, { "current_steps": 1727, "loss": 3.0065, "learning_rate": 1.383881230116649e-07, "epoch": 2.610271903323263 }, { "current_steps": 1728, "loss": 3.0505, "learning_rate": 1.378579003181336e-07, "epoch": 2.61178247734139 }, { "current_steps": 1729, "loss": 3.1039, "learning_rate": 1.3732767762460232e-07, "epoch": 2.6132930513595167 }, { "current_steps": 1730, "loss": 2.92, "learning_rate": 1.3679745493107104e-07, "epoch": 2.6148036253776437 }, { "current_steps": 1731, "loss": 2.0313, "learning_rate": 1.3626723223753976e-07, "epoch": 2.61631419939577 }, { "current_steps": 1732, "loss": 2.8514, "learning_rate": 1.357370095440085e-07, "epoch": 2.617824773413897 }, { "current_steps": 1733, "loss": 2.8458, "learning_rate": 1.3520678685047721e-07, "epoch": 2.619335347432024 }, { "current_steps": 1734, "loss": 2.8333, "learning_rate": 1.3467656415694593e-07, "epoch": 2.620845921450151 }, { "current_steps": 1735, "loss": 2.0155, "learning_rate": 1.3414634146341465e-07, "epoch": 2.622356495468278 }, { "current_steps": 1736, "loss": 2.8929, "learning_rate": 1.3361611876988336e-07, "epoch": 2.623867069486405 }, { "current_steps": 1737, "loss": 2.9456, "learning_rate": 1.3308589607635208e-07, "epoch": 2.6253776435045317 }, { "current_steps": 1738, "loss": 3.0525, "learning_rate": 1.325556733828208e-07, "epoch": 2.6268882175226587 }, { "current_steps": 1739, "loss": 3.019, "learning_rate": 1.320254506892895e-07, "epoch": 2.6283987915407856 }, { "current_steps": 1740, "loss": 2.5144, "learning_rate": 1.3149522799575822e-07, "epoch": 2.6299093655589125 }, { "current_steps": 1741, "loss": 3.0884, "learning_rate": 1.3096500530222691e-07, "epoch": 2.6314199395770395 }, { "current_steps": 1742, "loss": 3.1782, "learning_rate": 1.3043478260869563e-07, "epoch": 2.6329305135951664 }, { "current_steps": 1743, "loss": 2.9949, "learning_rate": 1.2990455991516435e-07, "epoch": 2.6344410876132933 }, { "current_steps": 1744, "loss": 2.936, "learning_rate": 1.2937433722163306e-07, "epoch": 2.63595166163142 }, { "current_steps": 1745, "loss": 2.9181, "learning_rate": 1.2884411452810178e-07, "epoch": 2.6374622356495467 }, { "current_steps": 1746, "loss": 2.7088, "learning_rate": 1.283138918345705e-07, "epoch": 2.6389728096676737 }, { "current_steps": 1747, "loss": 2.8018, "learning_rate": 1.277836691410392e-07, "epoch": 2.6404833836858006 }, { "current_steps": 1748, "loss": 2.3942, "learning_rate": 1.2725344644750793e-07, "epoch": 2.6419939577039275 }, { "current_steps": 1749, "loss": 3.1531, "learning_rate": 1.2672322375397667e-07, "epoch": 2.6435045317220545 }, { "current_steps": 1750, "loss": 2.8847, "learning_rate": 1.2619300106044538e-07, "epoch": 2.6450151057401814 }, { "current_steps": 1751, "loss": 3.0701, "learning_rate": 1.256627783669141e-07, "epoch": 2.646525679758308 }, { "current_steps": 1752, "loss": 3.3033, "learning_rate": 1.2513255567338282e-07, "epoch": 2.648036253776435 }, { "current_steps": 1753, "loss": 2.7969, "learning_rate": 1.2460233297985153e-07, "epoch": 2.6495468277945617 }, { "current_steps": 1754, "loss": 2.8172, "learning_rate": 1.2407211028632025e-07, "epoch": 2.6510574018126887 }, { "current_steps": 1755, "loss": 2.0426, "learning_rate": 1.2354188759278896e-07, "epoch": 2.6525679758308156 }, { "current_steps": 1756, "loss": 3.0481, "learning_rate": 1.2301166489925768e-07, "epoch": 2.6540785498489425 }, { "current_steps": 1757, "loss": 2.8373, "learning_rate": 1.224814422057264e-07, "epoch": 2.6555891238670695 }, { "current_steps": 1758, "loss": 2.8724, "learning_rate": 1.219512195121951e-07, "epoch": 2.6570996978851964 }, { "current_steps": 1759, "loss": 2.9434, "learning_rate": 1.2142099681866383e-07, "epoch": 2.6586102719033233 }, { "current_steps": 1760, "loss": 2.9696, "learning_rate": 1.2089077412513254e-07, "epoch": 2.6601208459214503 }, { "current_steps": 1761, "loss": 3.091, "learning_rate": 1.2036055143160128e-07, "epoch": 2.661631419939577 }, { "current_steps": 1762, "loss": 2.9475, "learning_rate": 1.1983032873807e-07, "epoch": 2.663141993957704 }, { "current_steps": 1763, "loss": 3.1059, "learning_rate": 1.1930010604453872e-07, "epoch": 2.664652567975831 }, { "current_steps": 1764, "loss": 2.9254, "learning_rate": 1.1876988335100742e-07, "epoch": 2.6661631419939575 }, { "current_steps": 1765, "loss": 2.9885, "learning_rate": 1.1823966065747613e-07, "epoch": 2.6676737160120845 }, { "current_steps": 1766, "loss": 2.7631, "learning_rate": 1.1770943796394485e-07, "epoch": 2.6691842900302114 }, { "current_steps": 1767, "loss": 2.939, "learning_rate": 1.1717921527041357e-07, "epoch": 2.6706948640483383 }, { "current_steps": 1768, "loss": 2.8795, "learning_rate": 1.1664899257688228e-07, "epoch": 2.6722054380664653 }, { "current_steps": 1769, "loss": 2.994, "learning_rate": 1.1611876988335101e-07, "epoch": 2.673716012084592 }, { "current_steps": 1770, "loss": 2.8866, "learning_rate": 1.1558854718981973e-07, "epoch": 2.675226586102719 }, { "current_steps": 1771, "loss": 3.0901, "learning_rate": 1.1505832449628844e-07, "epoch": 2.676737160120846 }, { "current_steps": 1772, "loss": 3.0214, "learning_rate": 1.1452810180275716e-07, "epoch": 2.6782477341389725 }, { "current_steps": 1773, "loss": 2.9535, "learning_rate": 1.1399787910922588e-07, "epoch": 2.6797583081570995 }, { "current_steps": 1774, "loss": 2.8239, "learning_rate": 1.1346765641569459e-07, "epoch": 2.6812688821752264 }, { "current_steps": 1775, "loss": 2.9065, "learning_rate": 1.129374337221633e-07, "epoch": 2.6827794561933533 }, { "current_steps": 1776, "loss": 2.9284, "learning_rate": 1.1240721102863201e-07, "epoch": 2.6842900302114803 }, { "current_steps": 1777, "loss": 2.9564, "learning_rate": 1.1187698833510073e-07, "epoch": 2.685800604229607 }, { "current_steps": 1778, "loss": 2.7768, "learning_rate": 1.1134676564156945e-07, "epoch": 2.687311178247734 }, { "current_steps": 1779, "loss": 2.8876, "learning_rate": 1.1081654294803817e-07, "epoch": 2.688821752265861 }, { "current_steps": 1780, "loss": 2.4335, "learning_rate": 1.1028632025450689e-07, "epoch": 2.690332326283988 }, { "current_steps": 1781, "loss": 2.9975, "learning_rate": 1.097560975609756e-07, "epoch": 2.691842900302115 }, { "current_steps": 1782, "loss": 3.1829, "learning_rate": 1.0922587486744432e-07, "epoch": 2.693353474320242 }, { "current_steps": 1783, "loss": 2.8911, "learning_rate": 1.0869565217391303e-07, "epoch": 2.694864048338369 }, { "current_steps": 1784, "loss": 3.0422, "learning_rate": 1.0816542948038175e-07, "epoch": 2.6963746223564957 }, { "current_steps": 1785, "loss": 2.9348, "learning_rate": 1.0763520678685047e-07, "epoch": 2.697885196374622 }, { "current_steps": 1786, "loss": 2.9154, "learning_rate": 1.071049840933192e-07, "epoch": 2.699395770392749 }, { "current_steps": 1787, "loss": 3.0889, "learning_rate": 1.0657476139978791e-07, "epoch": 2.700906344410876 }, { "current_steps": 1788, "loss": 3.1454, "learning_rate": 1.0604453870625663e-07, "epoch": 2.702416918429003 }, { "current_steps": 1789, "loss": 2.8741, "learning_rate": 1.0551431601272534e-07, "epoch": 2.70392749244713 }, { "current_steps": 1790, "loss": 2.8628, "learning_rate": 1.0498409331919406e-07, "epoch": 2.705438066465257 }, { "current_steps": 1791, "loss": 2.9693, "learning_rate": 1.0445387062566277e-07, "epoch": 2.706948640483384 }, { "current_steps": 1792, "loss": 2.9256, "learning_rate": 1.0392364793213149e-07, "epoch": 2.7084592145015107 }, { "current_steps": 1793, "loss": 2.8001, "learning_rate": 1.033934252386002e-07, "epoch": 2.709969788519637 }, { "current_steps": 1794, "loss": 2.961, "learning_rate": 1.0286320254506894e-07, "epoch": 2.711480362537764 }, { "current_steps": 1795, "loss": 2.9523, "learning_rate": 1.0233297985153765e-07, "epoch": 2.712990936555891 }, { "current_steps": 1796, "loss": 2.9713, "learning_rate": 1.0180275715800637e-07, "epoch": 2.714501510574018 }, { "current_steps": 1797, "loss": 3.019, "learning_rate": 1.0127253446447508e-07, "epoch": 2.716012084592145 }, { "current_steps": 1798, "loss": 2.8769, "learning_rate": 1.0074231177094379e-07, "epoch": 2.717522658610272 }, { "current_steps": 1799, "loss": 3.0721, "learning_rate": 1.002120890774125e-07, "epoch": 2.719033232628399 }, { "current_steps": 1800, "loss": 3.0198, "learning_rate": 9.968186638388122e-08, "epoch": 2.7205438066465257 }, { "current_steps": 1801, "loss": 3.0286, "learning_rate": 9.915164369034993e-08, "epoch": 2.7220543806646527 }, { "current_steps": 1802, "loss": 2.8577, "learning_rate": 9.862142099681866e-08, "epoch": 2.7235649546827796 }, { "current_steps": 1803, "loss": 2.9513, "learning_rate": 9.809119830328738e-08, "epoch": 2.7250755287009065 }, { "current_steps": 1804, "loss": 2.9736, "learning_rate": 9.75609756097561e-08, "epoch": 2.7265861027190335 }, { "current_steps": 1805, "loss": 2.9948, "learning_rate": 9.703075291622481e-08, "epoch": 2.7280966767371604 }, { "current_steps": 1806, "loss": 2.4052, "learning_rate": 9.650053022269353e-08, "epoch": 2.729607250755287 }, { "current_steps": 1807, "loss": 2.3485, "learning_rate": 9.597030752916224e-08, "epoch": 2.731117824773414 }, { "current_steps": 1808, "loss": 2.8374, "learning_rate": 9.544008483563096e-08, "epoch": 2.7326283987915407 }, { "current_steps": 1809, "loss": 2.9591, "learning_rate": 9.490986214209967e-08, "epoch": 2.7341389728096677 }, { "current_steps": 1810, "loss": 2.8868, "learning_rate": 9.437963944856839e-08, "epoch": 2.7356495468277946 }, { "current_steps": 1811, "loss": 2.9672, "learning_rate": 9.384941675503712e-08, "epoch": 2.7371601208459215 }, { "current_steps": 1812, "loss": 3.1142, "learning_rate": 9.331919406150583e-08, "epoch": 2.7386706948640485 }, { "current_steps": 1813, "loss": 2.899, "learning_rate": 9.278897136797455e-08, "epoch": 2.7401812688821754 }, { "current_steps": 1814, "loss": 2.7175, "learning_rate": 9.225874867444327e-08, "epoch": 2.741691842900302 }, { "current_steps": 1815, "loss": 2.8848, "learning_rate": 9.172852598091198e-08, "epoch": 2.743202416918429 }, { "current_steps": 1816, "loss": 2.8628, "learning_rate": 9.11983032873807e-08, "epoch": 2.7447129909365557 }, { "current_steps": 1817, "loss": 2.9828, "learning_rate": 9.066808059384941e-08, "epoch": 2.7462235649546827 }, { "current_steps": 1818, "loss": 2.6791, "learning_rate": 9.013785790031813e-08, "epoch": 2.7477341389728096 }, { "current_steps": 1819, "loss": 3.0182, "learning_rate": 8.960763520678686e-08, "epoch": 2.7492447129909365 }, { "current_steps": 1820, "loss": 2.6572, "learning_rate": 8.907741251325556e-08, "epoch": 2.7507552870090635 }, { "current_steps": 1821, "loss": 3.0155, "learning_rate": 8.854718981972428e-08, "epoch": 2.7522658610271904 }, { "current_steps": 1822, "loss": 2.2547, "learning_rate": 8.801696712619299e-08, "epoch": 2.7537764350453173 }, { "current_steps": 1823, "loss": 2.9473, "learning_rate": 8.748674443266171e-08, "epoch": 2.7552870090634443 }, { "current_steps": 1824, "loss": 2.9914, "learning_rate": 8.695652173913042e-08, "epoch": 2.756797583081571 }, { "current_steps": 1825, "loss": 3.1624, "learning_rate": 8.642629904559914e-08, "epoch": 2.758308157099698 }, { "current_steps": 1826, "loss": 3.0589, "learning_rate": 8.589607635206786e-08, "epoch": 2.7598187311178246 }, { "current_steps": 1827, "loss": 3.0378, "learning_rate": 8.536585365853659e-08, "epoch": 2.7613293051359515 }, { "current_steps": 1828, "loss": 2.979, "learning_rate": 8.48356309650053e-08, "epoch": 2.7628398791540785 }, { "current_steps": 1829, "loss": 2.9768, "learning_rate": 8.430540827147402e-08, "epoch": 2.7643504531722054 }, { "current_steps": 1830, "loss": 2.9681, "learning_rate": 8.377518557794273e-08, "epoch": 2.7658610271903323 }, { "current_steps": 1831, "loss": 3.1236, "learning_rate": 8.324496288441145e-08, "epoch": 2.7673716012084593 }, { "current_steps": 1832, "loss": 3.0861, "learning_rate": 8.271474019088016e-08, "epoch": 2.768882175226586 }, { "current_steps": 1833, "loss": 3.1181, "learning_rate": 8.218451749734888e-08, "epoch": 2.770392749244713 }, { "current_steps": 1834, "loss": 2.7771, "learning_rate": 8.16542948038176e-08, "epoch": 2.7719033232628396 }, { "current_steps": 1835, "loss": 3.0001, "learning_rate": 8.112407211028633e-08, "epoch": 2.7734138972809665 }, { "current_steps": 1836, "loss": 2.904, "learning_rate": 8.059384941675504e-08, "epoch": 2.7749244712990935 }, { "current_steps": 1837, "loss": 2.9935, "learning_rate": 8.006362672322376e-08, "epoch": 2.7764350453172204 }, { "current_steps": 1838, "loss": 2.8819, "learning_rate": 7.953340402969247e-08, "epoch": 2.7779456193353473 }, { "current_steps": 1839, "loss": 2.8291, "learning_rate": 7.900318133616119e-08, "epoch": 2.7794561933534743 }, { "current_steps": 1840, "loss": 3.0213, "learning_rate": 7.84729586426299e-08, "epoch": 2.780966767371601 }, { "current_steps": 1841, "loss": 3.0364, "learning_rate": 7.794273594909862e-08, "epoch": 2.782477341389728 }, { "current_steps": 1842, "loss": 3.1144, "learning_rate": 7.741251325556732e-08, "epoch": 2.783987915407855 }, { "current_steps": 1843, "loss": 2.8542, "learning_rate": 7.688229056203604e-08, "epoch": 2.785498489425982 }, { "current_steps": 1844, "loss": 3.028, "learning_rate": 7.635206786850477e-08, "epoch": 2.787009063444109 }, { "current_steps": 1845, "loss": 2.8336, "learning_rate": 7.582184517497348e-08, "epoch": 2.788519637462236 }, { "current_steps": 1846, "loss": 2.7862, "learning_rate": 7.52916224814422e-08, "epoch": 2.790030211480363 }, { "current_steps": 1847, "loss": 3.0405, "learning_rate": 7.476139978791092e-08, "epoch": 2.7915407854984893 }, { "current_steps": 1848, "loss": 2.8597, "learning_rate": 7.423117709437963e-08, "epoch": 2.793051359516616 }, { "current_steps": 1849, "loss": 2.946, "learning_rate": 7.370095440084835e-08, "epoch": 2.794561933534743 }, { "current_steps": 1850, "loss": 3.1885, "learning_rate": 7.317073170731706e-08, "epoch": 2.79607250755287 }, { "current_steps": 1851, "loss": 3.1006, "learning_rate": 7.264050901378578e-08, "epoch": 2.797583081570997 }, { "current_steps": 1852, "loss": 2.8335, "learning_rate": 7.211028632025451e-08, "epoch": 2.799093655589124 }, { "current_steps": 1853, "loss": 2.8269, "learning_rate": 7.158006362672322e-08, "epoch": 2.800604229607251 }, { "current_steps": 1854, "loss": 2.6435, "learning_rate": 7.104984093319194e-08, "epoch": 2.802114803625378 }, { "current_steps": 1855, "loss": 2.9359, "learning_rate": 7.051961823966066e-08, "epoch": 2.8036253776435043 }, { "current_steps": 1856, "loss": 3.0477, "learning_rate": 6.998939554612937e-08, "epoch": 2.805135951661631 }, { "current_steps": 1857, "loss": 3.1709, "learning_rate": 6.945917285259809e-08, "epoch": 2.806646525679758 }, { "current_steps": 1858, "loss": 2.8944, "learning_rate": 6.89289501590668e-08, "epoch": 2.808157099697885 }, { "current_steps": 1859, "loss": 2.8094, "learning_rate": 6.839872746553552e-08, "epoch": 2.809667673716012 }, { "current_steps": 1860, "loss": 2.9316, "learning_rate": 6.786850477200425e-08, "epoch": 2.811178247734139 }, { "current_steps": 1861, "loss": 3.0675, "learning_rate": 6.733828207847296e-08, "epoch": 2.812688821752266 }, { "current_steps": 1862, "loss": 2.958, "learning_rate": 6.680805938494168e-08, "epoch": 2.814199395770393 }, { "current_steps": 1863, "loss": 3.0425, "learning_rate": 6.62778366914104e-08, "epoch": 2.8157099697885197 }, { "current_steps": 1864, "loss": 2.7715, "learning_rate": 6.574761399787911e-08, "epoch": 2.8172205438066467 }, { "current_steps": 1865, "loss": 3.0176, "learning_rate": 6.521739130434782e-08, "epoch": 2.8187311178247736 }, { "current_steps": 1866, "loss": 3.0609, "learning_rate": 6.468716861081653e-08, "epoch": 2.8202416918429005 }, { "current_steps": 1867, "loss": 2.612, "learning_rate": 6.415694591728525e-08, "epoch": 2.8217522658610275 }, { "current_steps": 1868, "loss": 2.9679, "learning_rate": 6.362672322375396e-08, "epoch": 2.823262839879154 }, { "current_steps": 1869, "loss": 2.5519, "learning_rate": 6.309650053022269e-08, "epoch": 2.824773413897281 }, { "current_steps": 1870, "loss": 2.9077, "learning_rate": 6.256627783669141e-08, "epoch": 2.826283987915408 }, { "current_steps": 1871, "loss": 2.9207, "learning_rate": 6.203605514316012e-08, "epoch": 2.8277945619335347 }, { "current_steps": 1872, "loss": 2.7965, "learning_rate": 6.150583244962884e-08, "epoch": 2.8293051359516617 }, { "current_steps": 1873, "loss": 2.881, "learning_rate": 6.097560975609756e-08, "epoch": 2.8308157099697886 }, { "current_steps": 1874, "loss": 2.963, "learning_rate": 6.044538706256627e-08, "epoch": 2.8323262839879155 }, { "current_steps": 1875, "loss": 3.058, "learning_rate": 5.9915164369035e-08, "epoch": 2.8338368580060425 }, { "current_steps": 1876, "loss": 3.0952, "learning_rate": 5.938494167550371e-08, "epoch": 2.835347432024169 }, { "current_steps": 1877, "loss": 3.166, "learning_rate": 5.8854718981972425e-08, "epoch": 2.836858006042296 }, { "current_steps": 1878, "loss": 2.8453, "learning_rate": 5.832449628844114e-08, "epoch": 2.838368580060423 }, { "current_steps": 1879, "loss": 2.3267, "learning_rate": 5.7794273594909864e-08, "epoch": 2.8398791540785497 }, { "current_steps": 1880, "loss": 3.3411, "learning_rate": 5.726405090137858e-08, "epoch": 2.8413897280966767 }, { "current_steps": 1881, "loss": 2.727, "learning_rate": 5.6733828207847296e-08, "epoch": 2.8429003021148036 }, { "current_steps": 1882, "loss": 3.0315, "learning_rate": 5.6203605514316005e-08, "epoch": 2.8444108761329305 }, { "current_steps": 1883, "loss": 2.8649, "learning_rate": 5.567338282078473e-08, "epoch": 2.8459214501510575 }, { "current_steps": 1884, "loss": 3.0212, "learning_rate": 5.514316012725344e-08, "epoch": 2.8474320241691844 }, { "current_steps": 1885, "loss": 3.1805, "learning_rate": 5.461293743372216e-08, "epoch": 2.8489425981873113 }, { "current_steps": 1886, "loss": 3.0182, "learning_rate": 5.4082714740190875e-08, "epoch": 2.8504531722054383 }, { "current_steps": 1887, "loss": 3.0725, "learning_rate": 5.35524920466596e-08, "epoch": 2.851963746223565 }, { "current_steps": 1888, "loss": 2.0851, "learning_rate": 5.302226935312831e-08, "epoch": 2.853474320241692 }, { "current_steps": 1889, "loss": 2.9063, "learning_rate": 5.249204665959703e-08, "epoch": 2.8549848942598186 }, { "current_steps": 1890, "loss": 2.9657, "learning_rate": 5.1961823966065745e-08, "epoch": 2.8564954682779455 }, { "current_steps": 1891, "loss": 2.9003, "learning_rate": 5.143160127253447e-08, "epoch": 2.8580060422960725 }, { "current_steps": 1892, "loss": 3.1238, "learning_rate": 5.0901378579003183e-08, "epoch": 2.8595166163141994 }, { "current_steps": 1893, "loss": 3.0727, "learning_rate": 5.037115588547189e-08, "epoch": 2.8610271903323263 }, { "current_steps": 1894, "loss": 2.9672, "learning_rate": 4.984093319194061e-08, "epoch": 2.8625377643504533 }, { "current_steps": 1895, "loss": 3.0155, "learning_rate": 4.931071049840933e-08, "epoch": 2.86404833836858 }, { "current_steps": 1896, "loss": 2.9195, "learning_rate": 4.878048780487805e-08, "epoch": 2.8655589123867067 }, { "current_steps": 1897, "loss": 3.0422, "learning_rate": 4.825026511134676e-08, "epoch": 2.8670694864048336 }, { "current_steps": 1898, "loss": 2.7132, "learning_rate": 4.772004241781548e-08, "epoch": 2.8685800604229605 }, { "current_steps": 1899, "loss": 3.126, "learning_rate": 4.7189819724284195e-08, "epoch": 2.8700906344410875 }, { "current_steps": 1900, "loss": 2.8712, "learning_rate": 4.665959703075292e-08, "epoch": 2.8716012084592144 }, { "current_steps": 1901, "loss": 2.9313, "learning_rate": 4.612937433722163e-08, "epoch": 2.8731117824773413 }, { "current_steps": 1902, "loss": 3.2139, "learning_rate": 4.559915164369035e-08, "epoch": 2.8746223564954683 }, { "current_steps": 1903, "loss": 2.8807, "learning_rate": 4.5068928950159065e-08, "epoch": 2.876132930513595 }, { "current_steps": 1904, "loss": 2.8274, "learning_rate": 4.453870625662778e-08, "epoch": 2.877643504531722 }, { "current_steps": 1905, "loss": 2.9904, "learning_rate": 4.4008483563096496e-08, "epoch": 2.879154078549849 }, { "current_steps": 1906, "loss": 3.1017, "learning_rate": 4.347826086956521e-08, "epoch": 2.880664652567976 }, { "current_steps": 1907, "loss": 2.9465, "learning_rate": 4.294803817603393e-08, "epoch": 2.882175226586103 }, { "current_steps": 1908, "loss": 2.963, "learning_rate": 4.241781548250265e-08, "epoch": 2.88368580060423 }, { "current_steps": 1909, "loss": 3.0362, "learning_rate": 4.1887592788971367e-08, "epoch": 2.8851963746223563 }, { "current_steps": 1910, "loss": 2.824, "learning_rate": 4.135737009544008e-08, "epoch": 2.8867069486404833 }, { "current_steps": 1911, "loss": 2.8822, "learning_rate": 4.08271474019088e-08, "epoch": 2.88821752265861 }, { "current_steps": 1912, "loss": 2.8811, "learning_rate": 4.029692470837752e-08, "epoch": 2.889728096676737 }, { "current_steps": 1913, "loss": 3.0816, "learning_rate": 3.9766702014846237e-08, "epoch": 2.891238670694864 }, { "current_steps": 1914, "loss": 3.0799, "learning_rate": 3.923647932131495e-08, "epoch": 2.892749244712991 }, { "current_steps": 1915, "loss": 2.89, "learning_rate": 3.870625662778366e-08, "epoch": 2.894259818731118 }, { "current_steps": 1916, "loss": 3.1869, "learning_rate": 3.8176033934252384e-08, "epoch": 2.895770392749245 }, { "current_steps": 1917, "loss": 3.1516, "learning_rate": 3.76458112407211e-08, "epoch": 2.8972809667673713 }, { "current_steps": 1918, "loss": 3.0378, "learning_rate": 3.7115588547189816e-08, "epoch": 2.8987915407854983 }, { "current_steps": 1919, "loss": 2.8611, "learning_rate": 3.658536585365853e-08, "epoch": 2.900302114803625 }, { "current_steps": 1920, "loss": 2.9668, "learning_rate": 3.6055143160127254e-08, "epoch": 2.901812688821752 }, { "current_steps": 1921, "loss": 2.9424, "learning_rate": 3.552492046659597e-08, "epoch": 2.903323262839879 }, { "current_steps": 1922, "loss": 2.9069, "learning_rate": 3.4994697773064686e-08, "epoch": 2.904833836858006 }, { "current_steps": 1923, "loss": 2.9543, "learning_rate": 3.44644750795334e-08, "epoch": 2.906344410876133 }, { "current_steps": 1924, "loss": 2.7606, "learning_rate": 3.3934252386002124e-08, "epoch": 2.90785498489426 }, { "current_steps": 1925, "loss": 3.0176, "learning_rate": 3.340402969247084e-08, "epoch": 2.909365558912387 }, { "current_steps": 1926, "loss": 2.9565, "learning_rate": 3.2873806998939556e-08, "epoch": 2.9108761329305137 }, { "current_steps": 1927, "loss": 2.94, "learning_rate": 3.2343584305408265e-08, "epoch": 2.9123867069486407 }, { "current_steps": 1928, "loss": 3.0043, "learning_rate": 3.181336161187698e-08, "epoch": 2.9138972809667676 }, { "current_steps": 1929, "loss": 2.9566, "learning_rate": 3.1283138918345704e-08, "epoch": 2.9154078549848945 }, { "current_steps": 1930, "loss": 3.0867, "learning_rate": 3.075291622481442e-08, "epoch": 2.916918429003021 }, { "current_steps": 1931, "loss": 3.0064, "learning_rate": 3.0222693531283136e-08, "epoch": 2.918429003021148 }, { "current_steps": 1932, "loss": 2.7964, "learning_rate": 2.9692470837751855e-08, "epoch": 2.919939577039275 }, { "current_steps": 1933, "loss": 2.9417, "learning_rate": 2.916224814422057e-08, "epoch": 2.921450151057402 }, { "current_steps": 1934, "loss": 2.9919, "learning_rate": 2.863202545068929e-08, "epoch": 2.9229607250755287 }, { "current_steps": 1935, "loss": 3.0771, "learning_rate": 2.8101802757158002e-08, "epoch": 2.9244712990936557 }, { "current_steps": 1936, "loss": 2.8613, "learning_rate": 2.757158006362672e-08, "epoch": 2.9259818731117826 }, { "current_steps": 1937, "loss": 2.9584, "learning_rate": 2.7041357370095437e-08, "epoch": 2.9274924471299095 }, { "current_steps": 1938, "loss": 2.9161, "learning_rate": 2.6511134676564157e-08, "epoch": 2.929003021148036 }, { "current_steps": 1939, "loss": 3.0336, "learning_rate": 2.5980911983032873e-08, "epoch": 2.930513595166163 }, { "current_steps": 1940, "loss": 2.9381, "learning_rate": 2.5450689289501592e-08, "epoch": 2.93202416918429 }, { "current_steps": 1941, "loss": 3.0685, "learning_rate": 2.4920466595970304e-08, "epoch": 2.933534743202417 }, { "current_steps": 1942, "loss": 3.0419, "learning_rate": 2.4390243902439023e-08, "epoch": 2.9350453172205437 }, { "current_steps": 1943, "loss": 2.7357, "learning_rate": 2.386002120890774e-08, "epoch": 2.9365558912386707 }, { "current_steps": 1944, "loss": 3.0399, "learning_rate": 2.332979851537646e-08, "epoch": 2.9380664652567976 }, { "current_steps": 1945, "loss": 2.3534, "learning_rate": 2.2799575821845174e-08, "epoch": 2.9395770392749245 }, { "current_steps": 1946, "loss": 2.8138, "learning_rate": 2.226935312831389e-08, "epoch": 2.9410876132930515 }, { "current_steps": 1947, "loss": 2.9336, "learning_rate": 2.1739130434782606e-08, "epoch": 2.9425981873111784 }, { "current_steps": 1948, "loss": 3.1749, "learning_rate": 2.1208907741251325e-08, "epoch": 2.9441087613293053 }, { "current_steps": 1949, "loss": 2.857, "learning_rate": 2.067868504772004e-08, "epoch": 2.9456193353474323 }, { "current_steps": 1950, "loss": 3.0774, "learning_rate": 2.014846235418876e-08, "epoch": 2.947129909365559 }, { "current_steps": 1951, "loss": 3.0129, "learning_rate": 1.9618239660657476e-08, "epoch": 2.9486404833836857 }, { "current_steps": 1952, "loss": 2.8487, "learning_rate": 1.9088016967126192e-08, "epoch": 2.9501510574018126 }, { "current_steps": 1953, "loss": 2.9382, "learning_rate": 1.8557794273594908e-08, "epoch": 2.9516616314199395 }, { "current_steps": 1954, "loss": 3.1172, "learning_rate": 1.8027571580063627e-08, "epoch": 2.9531722054380665 }, { "current_steps": 1955, "loss": 3.1302, "learning_rate": 1.7497348886532343e-08, "epoch": 2.9546827794561934 }, { "current_steps": 1956, "loss": 2.8669, "learning_rate": 1.6967126193001062e-08, "epoch": 2.9561933534743203 }, { "current_steps": 1957, "loss": 3.0425, "learning_rate": 1.6436903499469778e-08, "epoch": 2.9577039274924473 }, { "current_steps": 1958, "loss": 2.9269, "learning_rate": 1.590668080593849e-08, "epoch": 2.9592145015105737 }, { "current_steps": 1959, "loss": 2.8536, "learning_rate": 1.537645811240721e-08, "epoch": 2.9607250755287007 }, { "current_steps": 1960, "loss": 3.0508, "learning_rate": 1.4846235418875927e-08, "epoch": 2.9622356495468276 }, { "current_steps": 1961, "loss": 2.984, "learning_rate": 1.4316012725344645e-08, "epoch": 2.9637462235649545 }, { "current_steps": 1962, "loss": 2.6814, "learning_rate": 1.378579003181336e-08, "epoch": 2.9652567975830815 }, { "current_steps": 1963, "loss": 2.935, "learning_rate": 1.3255567338282078e-08, "epoch": 2.9667673716012084 }, { "current_steps": 1964, "loss": 2.8209, "learning_rate": 1.2725344644750796e-08, "epoch": 2.9682779456193353 }, { "current_steps": 1965, "loss": 2.9674, "learning_rate": 1.2195121951219512e-08, "epoch": 2.9697885196374623 }, { "current_steps": 1966, "loss": 2.8591, "learning_rate": 1.166489925768823e-08, "epoch": 2.971299093655589 }, { "current_steps": 1967, "loss": 3.0003, "learning_rate": 1.1134676564156945e-08, "epoch": 2.972809667673716 }, { "current_steps": 1968, "loss": 3.0823, "learning_rate": 1.0604453870625663e-08, "epoch": 2.974320241691843 }, { "current_steps": 1969, "loss": 2.8632, "learning_rate": 1.007423117709438e-08, "epoch": 2.97583081570997 }, { "current_steps": 1970, "loss": 3.1056, "learning_rate": 9.544008483563096e-09, "epoch": 2.977341389728097 }, { "current_steps": 1971, "loss": 2.7785, "learning_rate": 9.013785790031814e-09, "epoch": 2.9788519637462234 }, { "current_steps": 1972, "loss": 3.0876, "learning_rate": 8.483563096500531e-09, "epoch": 2.9803625377643503 }, { "current_steps": 1973, "loss": 2.9398, "learning_rate": 7.953340402969245e-09, "epoch": 2.9818731117824773 }, { "current_steps": 1974, "loss": 2.6686, "learning_rate": 7.423117709437964e-09, "epoch": 2.983383685800604 }, { "current_steps": 1975, "loss": 2.2615, "learning_rate": 6.89289501590668e-09, "epoch": 2.984894259818731 }, { "current_steps": 1976, "loss": 3.0279, "learning_rate": 6.362672322375398e-09, "epoch": 2.986404833836858 }, { "current_steps": 1977, "loss": 2.4458, "learning_rate": 5.832449628844115e-09, "epoch": 2.987915407854985 }, { "current_steps": 1978, "loss": 2.6705, "learning_rate": 5.302226935312831e-09, "epoch": 2.989425981873112 }, { "current_steps": 1979, "loss": 2.9036, "learning_rate": 4.772004241781548e-09, "epoch": 2.9909365558912384 }, { "current_steps": 1980, "loss": 3.007, "learning_rate": 4.2417815482502656e-09, "epoch": 2.9924471299093653 }, { "current_steps": 1981, "loss": 3.0469, "learning_rate": 3.711558854718982e-09, "epoch": 2.9939577039274923 }, { "current_steps": 1982, "loss": 3.1243, "learning_rate": 3.181336161187699e-09, "epoch": 2.995468277945619 }, { "current_steps": 1983, "loss": 2.9443, "learning_rate": 2.6511134676564157e-09, "epoch": 2.996978851963746 }, { "current_steps": 1984, "loss": 2.9877, "learning_rate": 2.1208907741251328e-09, "epoch": 2.998489425981873 }, { "current_steps": 1985, "loss": 3.1269, "learning_rate": 1.5906680805938495e-09, "epoch": 3.0 }, { "current_steps": 1985, "loss": 3.1269, "learning_rate": 1.5906680805938495e-09, "epoch": 3.0 } ]