diff --git "a/illad_llama3/training_graph.json" "b/illad_llama3/training_graph.json" new file mode 100644--- /dev/null +++ "b/illad_llama3/training_graph.json" @@ -0,0 +1,11924 @@ +[ + { + "current_steps": 0, + "loss": 2.8643, + "learning_rate": 1e-08, + "epoch": 0.0015105740181268882 + }, + { + "current_steps": 1, + "loss": 3.3851, + "learning_rate": 2e-08, + "epoch": 0.0030211480362537764 + }, + { + "current_steps": 2, + "loss": 2.9606, + "learning_rate": 3e-08, + "epoch": 0.004531722054380665 + }, + { + "current_steps": 3, + "loss": 3.2647, + "learning_rate": 4e-08, + "epoch": 0.006042296072507553 + }, + { + "current_steps": 4, + "loss": 3.3543, + "learning_rate": 5e-08, + "epoch": 0.0075528700906344415 + }, + { + "current_steps": 5, + "loss": 3.2605, + "learning_rate": 5e-08, + "epoch": 0.00906344410876133 + }, + { + "current_steps": 6, + "loss": 3.3266, + "learning_rate": 6e-08, + "epoch": 0.010574018126888218 + }, + { + "current_steps": 7, + "loss": 3.0449, + "learning_rate": 7e-08, + "epoch": 0.012084592145015106 + }, + { + "current_steps": 8, + "loss": 3.2478, + "learning_rate": 8e-08, + "epoch": 0.013595166163141994 + }, + { + "current_steps": 9, + "loss": 3.3969, + "learning_rate": 9e-08, + "epoch": 0.015105740181268883 + }, + { + "current_steps": 10, + "loss": 3.2827, + "learning_rate": 1e-07, + "epoch": 0.01661631419939577 + }, + { + "current_steps": 11, + "loss": 3.0901, + "learning_rate": 1.0999999999999999e-07, + "epoch": 0.01812688821752266 + }, + { + "current_steps": 12, + "loss": 3.1617, + "learning_rate": 1.2e-07, + "epoch": 0.019637462235649546 + }, + { + "current_steps": 13, + "loss": 3.34, + "learning_rate": 1.3e-07, + "epoch": 0.021148036253776436 + }, + { + "current_steps": 14, + "loss": 3.3251, + "learning_rate": 1.4e-07, + "epoch": 0.022658610271903322 + }, + { + "current_steps": 15, + "loss": 3.1552, + "learning_rate": 1.5e-07, + "epoch": 0.02416918429003021 + }, + { + "current_steps": 16, + "loss": 3.2264, + "learning_rate": 1.6e-07, + "epoch": 0.0256797583081571 + }, + { + "current_steps": 17, + "loss": 3.1803, + "learning_rate": 1.6e-07, + "epoch": 0.027190332326283987 + }, + { + "current_steps": 18, + "loss": 3.2943, + "learning_rate": 1.7000000000000001e-07, + "epoch": 0.028700906344410877 + }, + { + "current_steps": 19, + "loss": 3.3261, + "learning_rate": 1.8e-07, + "epoch": 0.030211480362537766 + }, + { + "current_steps": 20, + "loss": 3.1263, + "learning_rate": 1.8999999999999998e-07, + "epoch": 0.03172205438066465 + }, + { + "current_steps": 21, + "loss": 3.1523, + "learning_rate": 2e-07, + "epoch": 0.03323262839879154 + }, + { + "current_steps": 22, + "loss": 3.312, + "learning_rate": 2.0999999999999997e-07, + "epoch": 0.03474320241691843 + }, + { + "current_steps": 23, + "loss": 3.2666, + "learning_rate": 2.1999999999999998e-07, + "epoch": 0.03625377643504532 + }, + { + "current_steps": 24, + "loss": 3.2577, + "learning_rate": 2.3e-07, + "epoch": 0.0377643504531722 + }, + { + "current_steps": 25, + "loss": 3.3348, + "learning_rate": 2.4e-07, + "epoch": 0.03927492447129909 + }, + { + "current_steps": 26, + "loss": 3.2292, + "learning_rate": 2.5e-07, + "epoch": 0.04078549848942598 + }, + { + "current_steps": 27, + "loss": 3.242, + "learning_rate": 2.6e-07, + "epoch": 0.04229607250755287 + }, + { + "current_steps": 28, + "loss": 3.2687, + "learning_rate": 2.7e-07, + "epoch": 0.04380664652567976 + }, + { + "current_steps": 29, + "loss": 2.7037, + "learning_rate": 2.8e-07, + "epoch": 0.045317220543806644 + }, + { + "current_steps": 30, + "loss": 3.1619, + "learning_rate": 2.9e-07, + "epoch": 0.04682779456193353 + }, + { + "current_steps": 31, + "loss": 3.2853, + "learning_rate": 3e-07, + "epoch": 0.04833836858006042 + }, + { + "current_steps": 32, + "loss": 3.3545, + "learning_rate": 3.1e-07, + "epoch": 0.04984894259818731 + }, + { + "current_steps": 33, + "loss": 2.5809, + "learning_rate": 3.2e-07, + "epoch": 0.0513595166163142 + }, + { + "current_steps": 34, + "loss": 2.9552, + "learning_rate": 3.3e-07, + "epoch": 0.052870090634441085 + }, + { + "current_steps": 35, + "loss": 3.1433, + "learning_rate": 3.4000000000000003e-07, + "epoch": 0.054380664652567974 + }, + { + "current_steps": 36, + "loss": 3.3105, + "learning_rate": 3.5e-07, + "epoch": 0.055891238670694864 + }, + { + "current_steps": 37, + "loss": 3.0981, + "learning_rate": 3.6e-07, + "epoch": 0.05740181268882175 + }, + { + "current_steps": 38, + "loss": 3.2201, + "learning_rate": 3.7e-07, + "epoch": 0.05891238670694864 + }, + { + "current_steps": 39, + "loss": 3.0603, + "learning_rate": 3.7999999999999996e-07, + "epoch": 0.06042296072507553 + }, + { + "current_steps": 40, + "loss": 3.3152, + "learning_rate": 3.8999999999999997e-07, + "epoch": 0.061933534743202415 + }, + { + "current_steps": 41, + "loss": 3.4646, + "learning_rate": 4e-07, + "epoch": 0.0634441087613293 + }, + { + "current_steps": 42, + "loss": 3.3562, + "learning_rate": 4.0999999999999994e-07, + "epoch": 0.0649546827794562 + }, + { + "current_steps": 43, + "loss": 3.1192, + "learning_rate": 4.1999999999999995e-07, + "epoch": 0.06646525679758308 + }, + { + "current_steps": 44, + "loss": 3.5199, + "learning_rate": 4.2999999999999996e-07, + "epoch": 0.06797583081570997 + }, + { + "current_steps": 45, + "loss": 3.1833, + "learning_rate": 4.3999999999999997e-07, + "epoch": 0.06948640483383686 + }, + { + "current_steps": 46, + "loss": 3.36, + "learning_rate": 4.5e-07, + "epoch": 0.07099697885196375 + }, + { + "current_steps": 47, + "loss": 2.7168, + "learning_rate": 4.6e-07, + "epoch": 0.07250755287009064 + }, + { + "current_steps": 48, + "loss": 3.1549, + "learning_rate": 4.6999999999999995e-07, + "epoch": 0.07401812688821752 + }, + { + "current_steps": 49, + "loss": 3.2864, + "learning_rate": 4.8e-07, + "epoch": 0.0755287009063444 + }, + { + "current_steps": 50, + "loss": 3.3691, + "learning_rate": 4.9e-07, + "epoch": 0.0770392749244713 + }, + { + "current_steps": 51, + "loss": 3.0839, + "learning_rate": 5e-07, + "epoch": 0.07854984894259819 + }, + { + "current_steps": 52, + "loss": 3.1339, + "learning_rate": 5.1e-07, + "epoch": 0.08006042296072508 + }, + { + "current_steps": 53, + "loss": 3.3284, + "learning_rate": 5.2e-07, + "epoch": 0.08157099697885196 + }, + { + "current_steps": 54, + "loss": 3.1468, + "learning_rate": 5.3e-07, + "epoch": 0.08308157099697885 + }, + { + "current_steps": 55, + "loss": 3.2341, + "learning_rate": 5.4e-07, + "epoch": 0.08459214501510574 + }, + { + "current_steps": 56, + "loss": 3.0192, + "learning_rate": 5.5e-07, + "epoch": 0.08610271903323263 + }, + { + "current_steps": 57, + "loss": 3.1109, + "learning_rate": 5.6e-07, + "epoch": 0.08761329305135952 + }, + { + "current_steps": 58, + "loss": 3.2616, + "learning_rate": 5.699999999999999e-07, + "epoch": 0.0891238670694864 + }, + { + "current_steps": 59, + "loss": 3.287, + "learning_rate": 5.8e-07, + "epoch": 0.09063444108761329 + }, + { + "current_steps": 60, + "loss": 3.2243, + "learning_rate": 5.9e-07, + "epoch": 0.09214501510574018 + }, + { + "current_steps": 61, + "loss": 3.1392, + "learning_rate": 6e-07, + "epoch": 0.09365558912386707 + }, + { + "current_steps": 62, + "loss": 3.2099, + "learning_rate": 6.1e-07, + "epoch": 0.09516616314199396 + }, + { + "current_steps": 63, + "loss": 3.2016, + "learning_rate": 6.2e-07, + "epoch": 0.09667673716012085 + }, + { + "current_steps": 64, + "loss": 3.1605, + "learning_rate": 6.3e-07, + "epoch": 0.09818731117824774 + }, + { + "current_steps": 65, + "loss": 3.4862, + "learning_rate": 6.4e-07, + "epoch": 0.09969788519637462 + }, + { + "current_steps": 66, + "loss": 2.9683, + "learning_rate": 6.5e-07, + "epoch": 0.10120845921450151 + }, + { + "current_steps": 67, + "loss": 3.5057, + "learning_rate": 6.6e-07, + "epoch": 0.1027190332326284 + }, + { + "current_steps": 68, + "loss": 3.2837, + "learning_rate": 6.7e-07, + "epoch": 0.1042296072507553 + }, + { + "current_steps": 69, + "loss": 3.3029, + "learning_rate": 6.800000000000001e-07, + "epoch": 0.10574018126888217 + }, + { + "current_steps": 70, + "loss": 2.7988, + "learning_rate": 6.9e-07, + "epoch": 0.10725075528700906 + }, + { + "current_steps": 71, + "loss": 2.8721, + "learning_rate": 7e-07, + "epoch": 0.10876132930513595 + }, + { + "current_steps": 72, + "loss": 3.0222, + "learning_rate": 7.1e-07, + "epoch": 0.11027190332326284 + }, + { + "current_steps": 73, + "loss": 3.2546, + "learning_rate": 7.2e-07, + "epoch": 0.11178247734138973 + }, + { + "current_steps": 74, + "loss": 3.196, + "learning_rate": 7.3e-07, + "epoch": 0.11329305135951662 + }, + { + "current_steps": 75, + "loss": 3.1073, + "learning_rate": 7.4e-07, + "epoch": 0.1148036253776435 + }, + { + "current_steps": 76, + "loss": 3.4334, + "learning_rate": 7.5e-07, + "epoch": 0.1163141993957704 + }, + { + "current_steps": 77, + "loss": 3.2911, + "learning_rate": 7.599999999999999e-07, + "epoch": 0.11782477341389729 + }, + { + "current_steps": 78, + "loss": 3.2415, + "learning_rate": 7.699999999999999e-07, + "epoch": 0.11933534743202417 + }, + { + "current_steps": 79, + "loss": 3.2162, + "learning_rate": 7.799999999999999e-07, + "epoch": 0.12084592145015106 + }, + { + "current_steps": 80, + "loss": 3.1616, + "learning_rate": 7.9e-07, + "epoch": 0.12235649546827794 + }, + { + "current_steps": 81, + "loss": 3.356, + "learning_rate": 8e-07, + "epoch": 0.12386706948640483 + }, + { + "current_steps": 82, + "loss": 3.2459, + "learning_rate": 8.1e-07, + "epoch": 0.12537764350453173 + }, + { + "current_steps": 83, + "loss": 3.2564, + "learning_rate": 8.199999999999999e-07, + "epoch": 0.1268882175226586 + }, + { + "current_steps": 84, + "loss": 3.2301, + "learning_rate": 8.299999999999999e-07, + "epoch": 0.1283987915407855 + }, + { + "current_steps": 85, + "loss": 3.1873, + "learning_rate": 8.399999999999999e-07, + "epoch": 0.1299093655589124 + }, + { + "current_steps": 86, + "loss": 2.931, + "learning_rate": 8.499999999999999e-07, + "epoch": 0.13141993957703926 + }, + { + "current_steps": 87, + "loss": 3.2395, + "learning_rate": 8.599999999999999e-07, + "epoch": 0.13293051359516617 + }, + { + "current_steps": 88, + "loss": 3.2334, + "learning_rate": 8.699999999999999e-07, + "epoch": 0.13444108761329304 + }, + { + "current_steps": 89, + "loss": 3.408, + "learning_rate": 8.799999999999999e-07, + "epoch": 0.13595166163141995 + }, + { + "current_steps": 90, + "loss": 3.1124, + "learning_rate": 8.9e-07, + "epoch": 0.13746223564954682 + }, + { + "current_steps": 91, + "loss": 3.0723, + "learning_rate": 9e-07, + "epoch": 0.13897280966767372 + }, + { + "current_steps": 92, + "loss": 3.1404, + "learning_rate": 9.1e-07, + "epoch": 0.1404833836858006 + }, + { + "current_steps": 93, + "loss": 3.0665, + "learning_rate": 9.2e-07, + "epoch": 0.1419939577039275 + }, + { + "current_steps": 94, + "loss": 2.6321, + "learning_rate": 9.3e-07, + "epoch": 0.14350453172205438 + }, + { + "current_steps": 95, + "loss": 2.9725, + "learning_rate": 9.399999999999999e-07, + "epoch": 0.14501510574018128 + }, + { + "current_steps": 96, + "loss": 3.1267, + "learning_rate": 9.499999999999999e-07, + "epoch": 0.14652567975830816 + }, + { + "current_steps": 97, + "loss": 3.3175, + "learning_rate": 9.6e-07, + "epoch": 0.14803625377643503 + }, + { + "current_steps": 98, + "loss": 3.277, + "learning_rate": 9.7e-07, + "epoch": 0.14954682779456194 + }, + { + "current_steps": 99, + "loss": 3.3042, + "learning_rate": 9.8e-07, + "epoch": 0.1510574018126888 + }, + { + "current_steps": 100, + "loss": 3.3383, + "learning_rate": 9.9e-07, + "epoch": 0.15256797583081572 + }, + { + "current_steps": 101, + "loss": 2.8985, + "learning_rate": 1e-06, + "epoch": 0.1540785498489426 + }, + { + "current_steps": 102, + "loss": 3.1538, + "learning_rate": 9.994697773064687e-07, + "epoch": 0.1555891238670695 + }, + { + "current_steps": 103, + "loss": 3.3022, + "learning_rate": 9.989395546129374e-07, + "epoch": 0.15709969788519637 + }, + { + "current_steps": 104, + "loss": 2.5211, + "learning_rate": 9.98409331919406e-07, + "epoch": 0.15861027190332327 + }, + { + "current_steps": 105, + "loss": 3.1553, + "learning_rate": 9.978791092258748e-07, + "epoch": 0.16012084592145015 + }, + { + "current_steps": 106, + "loss": 3.097, + "learning_rate": 9.973488865323435e-07, + "epoch": 0.16163141993957703 + }, + { + "current_steps": 107, + "loss": 3.1513, + "learning_rate": 9.968186638388122e-07, + "epoch": 0.16314199395770393 + }, + { + "current_steps": 108, + "loss": 3.1952, + "learning_rate": 9.96288441145281e-07, + "epoch": 0.1646525679758308 + }, + { + "current_steps": 109, + "loss": 3.1655, + "learning_rate": 9.957582184517497e-07, + "epoch": 0.1661631419939577 + }, + { + "current_steps": 110, + "loss": 3.3205, + "learning_rate": 9.952279957582184e-07, + "epoch": 0.16767371601208458 + }, + { + "current_steps": 111, + "loss": 3.1351, + "learning_rate": 9.946977730646871e-07, + "epoch": 0.1691842900302115 + }, + { + "current_steps": 112, + "loss": 2.9261, + "learning_rate": 9.941675503711558e-07, + "epoch": 0.17069486404833836 + }, + { + "current_steps": 113, + "loss": 3.0024, + "learning_rate": 9.936373276776245e-07, + "epoch": 0.17220543806646527 + }, + { + "current_steps": 114, + "loss": 3.0746, + "learning_rate": 9.931071049840933e-07, + "epoch": 0.17371601208459214 + }, + { + "current_steps": 115, + "loss": 3.0436, + "learning_rate": 9.92576882290562e-07, + "epoch": 0.17522658610271905 + }, + { + "current_steps": 116, + "loss": 3.1425, + "learning_rate": 9.920466595970307e-07, + "epoch": 0.17673716012084592 + }, + { + "current_steps": 117, + "loss": 3.3725, + "learning_rate": 9.915164369034994e-07, + "epoch": 0.1782477341389728 + }, + { + "current_steps": 118, + "loss": 2.691, + "learning_rate": 9.909862142099681e-07, + "epoch": 0.1797583081570997 + }, + { + "current_steps": 119, + "loss": 3.1319, + "learning_rate": 9.904559915164368e-07, + "epoch": 0.18126888217522658 + }, + { + "current_steps": 120, + "loss": 3.4188, + "learning_rate": 9.899257688229056e-07, + "epoch": 0.18277945619335348 + }, + { + "current_steps": 121, + "loss": 3.2334, + "learning_rate": 9.893955461293743e-07, + "epoch": 0.18429003021148035 + }, + { + "current_steps": 122, + "loss": 3.3199, + "learning_rate": 9.88865323435843e-07, + "epoch": 0.18580060422960726 + }, + { + "current_steps": 123, + "loss": 2.7292, + "learning_rate": 9.883351007423117e-07, + "epoch": 0.18731117824773413 + }, + { + "current_steps": 124, + "loss": 2.8369, + "learning_rate": 9.878048780487804e-07, + "epoch": 0.18882175226586104 + }, + { + "current_steps": 125, + "loss": 2.8981, + "learning_rate": 9.872746553552491e-07, + "epoch": 0.1903323262839879 + }, + { + "current_steps": 126, + "loss": 3.082, + "learning_rate": 9.867444326617179e-07, + "epoch": 0.19184290030211482 + }, + { + "current_steps": 127, + "loss": 2.9516, + "learning_rate": 9.862142099681866e-07, + "epoch": 0.1933534743202417 + }, + { + "current_steps": 128, + "loss": 3.3629, + "learning_rate": 9.856839872746553e-07, + "epoch": 0.19486404833836857 + }, + { + "current_steps": 129, + "loss": 2.9393, + "learning_rate": 9.85153764581124e-07, + "epoch": 0.19637462235649547 + }, + { + "current_steps": 130, + "loss": 3.2696, + "learning_rate": 9.846235418875927e-07, + "epoch": 0.19788519637462235 + }, + { + "current_steps": 131, + "loss": 2.9017, + "learning_rate": 9.840933191940614e-07, + "epoch": 0.19939577039274925 + }, + { + "current_steps": 132, + "loss": 3.1235, + "learning_rate": 9.835630965005301e-07, + "epoch": 0.20090634441087613 + }, + { + "current_steps": 133, + "loss": 2.9032, + "learning_rate": 9.830328738069989e-07, + "epoch": 0.20241691842900303 + }, + { + "current_steps": 134, + "loss": 3.1998, + "learning_rate": 9.825026511134676e-07, + "epoch": 0.2039274924471299 + }, + { + "current_steps": 135, + "loss": 3.2412, + "learning_rate": 9.819724284199363e-07, + "epoch": 0.2054380664652568 + }, + { + "current_steps": 136, + "loss": 3.3582, + "learning_rate": 9.81442205726405e-07, + "epoch": 0.20694864048338368 + }, + { + "current_steps": 137, + "loss": 3.2789, + "learning_rate": 9.809119830328737e-07, + "epoch": 0.2084592145015106 + }, + { + "current_steps": 138, + "loss": 3.2905, + "learning_rate": 9.803817603393424e-07, + "epoch": 0.20996978851963746 + }, + { + "current_steps": 139, + "loss": 3.0474, + "learning_rate": 9.798515376458112e-07, + "epoch": 0.21148036253776434 + }, + { + "current_steps": 140, + "loss": 3.218, + "learning_rate": 9.793213149522799e-07, + "epoch": 0.21299093655589124 + }, + { + "current_steps": 141, + "loss": 2.9356, + "learning_rate": 9.787910922587486e-07, + "epoch": 0.21450151057401812 + }, + { + "current_steps": 142, + "loss": 3.1152, + "learning_rate": 9.782608695652173e-07, + "epoch": 0.21601208459214502 + }, + { + "current_steps": 143, + "loss": 2.902, + "learning_rate": 9.77730646871686e-07, + "epoch": 0.2175226586102719 + }, + { + "current_steps": 144, + "loss": 3.3406, + "learning_rate": 9.772004241781547e-07, + "epoch": 0.2190332326283988 + }, + { + "current_steps": 145, + "loss": 3.0509, + "learning_rate": 9.766702014846235e-07, + "epoch": 0.22054380664652568 + }, + { + "current_steps": 146, + "loss": 3.2664, + "learning_rate": 9.761399787910922e-07, + "epoch": 0.22205438066465258 + }, + { + "current_steps": 147, + "loss": 3.2172, + "learning_rate": 9.756097560975609e-07, + "epoch": 0.22356495468277945 + }, + { + "current_steps": 148, + "loss": 3.2731, + "learning_rate": 9.750795334040296e-07, + "epoch": 0.22507552870090636 + }, + { + "current_steps": 149, + "loss": 3.1866, + "learning_rate": 9.745493107104983e-07, + "epoch": 0.22658610271903323 + }, + { + "current_steps": 150, + "loss": 3.2318, + "learning_rate": 9.74019088016967e-07, + "epoch": 0.2280966767371601 + }, + { + "current_steps": 151, + "loss": 3.2205, + "learning_rate": 9.734888653234357e-07, + "epoch": 0.229607250755287 + }, + { + "current_steps": 152, + "loss": 3.2561, + "learning_rate": 9.729586426299045e-07, + "epoch": 0.2311178247734139 + }, + { + "current_steps": 153, + "loss": 3.2357, + "learning_rate": 9.724284199363732e-07, + "epoch": 0.2326283987915408 + }, + { + "current_steps": 154, + "loss": 3.1097, + "learning_rate": 9.71898197242842e-07, + "epoch": 0.23413897280966767 + }, + { + "current_steps": 155, + "loss": 3.0489, + "learning_rate": 9.713679745493106e-07, + "epoch": 0.23564954682779457 + }, + { + "current_steps": 156, + "loss": 3.1937, + "learning_rate": 9.708377518557793e-07, + "epoch": 0.23716012084592145 + }, + { + "current_steps": 157, + "loss": 3.1377, + "learning_rate": 9.70307529162248e-07, + "epoch": 0.23867069486404835 + }, + { + "current_steps": 158, + "loss": 3.1246, + "learning_rate": 9.697773064687168e-07, + "epoch": 0.24018126888217523 + }, + { + "current_steps": 159, + "loss": 2.9715, + "learning_rate": 9.692470837751855e-07, + "epoch": 0.24169184290030213 + }, + { + "current_steps": 160, + "loss": 3.3301, + "learning_rate": 9.687168610816542e-07, + "epoch": 0.243202416918429 + }, + { + "current_steps": 161, + "loss": 3.175, + "learning_rate": 9.68186638388123e-07, + "epoch": 0.24471299093655588 + }, + { + "current_steps": 162, + "loss": 3.5311, + "learning_rate": 9.676564156945916e-07, + "epoch": 0.24622356495468278 + }, + { + "current_steps": 163, + "loss": 2.9409, + "learning_rate": 9.671261930010603e-07, + "epoch": 0.24773413897280966 + }, + { + "current_steps": 164, + "loss": 3.1624, + "learning_rate": 9.66595970307529e-07, + "epoch": 0.24924471299093656 + }, + { + "current_steps": 165, + "loss": 3.2766, + "learning_rate": 9.660657476139978e-07, + "epoch": 0.25075528700906347 + }, + { + "current_steps": 166, + "loss": 3.1353, + "learning_rate": 9.655355249204665e-07, + "epoch": 0.25226586102719034 + }, + { + "current_steps": 167, + "loss": 3.1441, + "learning_rate": 9.650053022269352e-07, + "epoch": 0.2537764350453172 + }, + { + "current_steps": 168, + "loss": 2.5206, + "learning_rate": 9.64475079533404e-07, + "epoch": 0.2552870090634441 + }, + { + "current_steps": 169, + "loss": 3.3392, + "learning_rate": 9.639448568398726e-07, + "epoch": 0.256797583081571 + }, + { + "current_steps": 170, + "loss": 3.3648, + "learning_rate": 9.634146341463414e-07, + "epoch": 0.2583081570996979 + }, + { + "current_steps": 171, + "loss": 2.9235, + "learning_rate": 9.628844114528103e-07, + "epoch": 0.2598187311178248 + }, + { + "current_steps": 172, + "loss": 3.2705, + "learning_rate": 9.623541887592788e-07, + "epoch": 0.26132930513595165 + }, + { + "current_steps": 173, + "loss": 3.0963, + "learning_rate": 9.618239660657475e-07, + "epoch": 0.2628398791540785 + }, + { + "current_steps": 174, + "loss": 3.0412, + "learning_rate": 9.612937433722162e-07, + "epoch": 0.26435045317220546 + }, + { + "current_steps": 175, + "loss": 3.0035, + "learning_rate": 9.607635206786851e-07, + "epoch": 0.26586102719033233 + }, + { + "current_steps": 176, + "loss": 2.977, + "learning_rate": 9.602332979851536e-07, + "epoch": 0.2673716012084592 + }, + { + "current_steps": 177, + "loss": 3.1074, + "learning_rate": 9.597030752916224e-07, + "epoch": 0.2688821752265861 + }, + { + "current_steps": 178, + "loss": 3.247, + "learning_rate": 9.59172852598091e-07, + "epoch": 0.270392749244713 + }, + { + "current_steps": 179, + "loss": 3.1697, + "learning_rate": 9.5864262990456e-07, + "epoch": 0.2719033232628399 + }, + { + "current_steps": 180, + "loss": 3.3536, + "learning_rate": 9.581124072110285e-07, + "epoch": 0.27341389728096677 + }, + { + "current_steps": 181, + "loss": 3.1507, + "learning_rate": 9.575821845174972e-07, + "epoch": 0.27492447129909364 + }, + { + "current_steps": 182, + "loss": 3.3054, + "learning_rate": 9.57051961823966e-07, + "epoch": 0.2764350453172205 + }, + { + "current_steps": 183, + "loss": 3.1328, + "learning_rate": 9.565217391304349e-07, + "epoch": 0.27794561933534745 + }, + { + "current_steps": 184, + "loss": 3.2001, + "learning_rate": 9.559915164369034e-07, + "epoch": 0.2794561933534743 + }, + { + "current_steps": 185, + "loss": 2.9704, + "learning_rate": 9.55461293743372e-07, + "epoch": 0.2809667673716012 + }, + { + "current_steps": 186, + "loss": 2.9961, + "learning_rate": 9.54931071049841e-07, + "epoch": 0.2824773413897281 + }, + { + "current_steps": 187, + "loss": 3.2078, + "learning_rate": 9.544008483563097e-07, + "epoch": 0.283987915407855 + }, + { + "current_steps": 188, + "loss": 3.1483, + "learning_rate": 9.538706256627782e-07, + "epoch": 0.2854984894259819 + }, + { + "current_steps": 189, + "loss": 2.4296, + "learning_rate": 9.53340402969247e-07, + "epoch": 0.28700906344410876 + }, + { + "current_steps": 190, + "loss": 3.2875, + "learning_rate": 9.528101802757158e-07, + "epoch": 0.28851963746223563 + }, + { + "current_steps": 191, + "loss": 3.2538, + "learning_rate": 9.522799575821845e-07, + "epoch": 0.29003021148036257 + }, + { + "current_steps": 192, + "loss": 3.0849, + "learning_rate": 9.517497348886532e-07, + "epoch": 0.29154078549848944 + }, + { + "current_steps": 193, + "loss": 3.3613, + "learning_rate": 9.512195121951218e-07, + "epoch": 0.2930513595166163 + }, + { + "current_steps": 194, + "loss": 3.1343, + "learning_rate": 9.506892895015906e-07, + "epoch": 0.2945619335347432 + }, + { + "current_steps": 195, + "loss": 3.3011, + "learning_rate": 9.501590668080594e-07, + "epoch": 0.29607250755287007 + }, + { + "current_steps": 196, + "loss": 3.2206, + "learning_rate": 9.496288441145281e-07, + "epoch": 0.297583081570997 + }, + { + "current_steps": 197, + "loss": 2.3201, + "learning_rate": 9.490986214209967e-07, + "epoch": 0.2990936555891239 + }, + { + "current_steps": 198, + "loss": 3.1189, + "learning_rate": 9.485683987274655e-07, + "epoch": 0.30060422960725075 + }, + { + "current_steps": 199, + "loss": 3.1607, + "learning_rate": 9.480381760339342e-07, + "epoch": 0.3021148036253776 + }, + { + "current_steps": 200, + "loss": 3.2595, + "learning_rate": 9.475079533404029e-07, + "epoch": 0.30362537764350456 + }, + { + "current_steps": 201, + "loss": 3.3786, + "learning_rate": 9.469777306468716e-07, + "epoch": 0.30513595166163143 + }, + { + "current_steps": 202, + "loss": 3.0444, + "learning_rate": 9.464475079533404e-07, + "epoch": 0.3066465256797583 + }, + { + "current_steps": 203, + "loss": 3.261, + "learning_rate": 9.459172852598091e-07, + "epoch": 0.3081570996978852 + }, + { + "current_steps": 204, + "loss": 3.3189, + "learning_rate": 9.453870625662778e-07, + "epoch": 0.30966767371601206 + }, + { + "current_steps": 205, + "loss": 3.4952, + "learning_rate": 9.448568398727465e-07, + "epoch": 0.311178247734139 + }, + { + "current_steps": 206, + "loss": 3.0673, + "learning_rate": 9.443266171792152e-07, + "epoch": 0.31268882175226587 + }, + { + "current_steps": 207, + "loss": 3.2579, + "learning_rate": 9.437963944856839e-07, + "epoch": 0.31419939577039274 + }, + { + "current_steps": 208, + "loss": 3.0032, + "learning_rate": 9.432661717921527e-07, + "epoch": 0.3157099697885196 + }, + { + "current_steps": 209, + "loss": 3.2322, + "learning_rate": 9.427359490986214e-07, + "epoch": 0.31722054380664655 + }, + { + "current_steps": 210, + "loss": 3.0372, + "learning_rate": 9.422057264050901e-07, + "epoch": 0.3187311178247734 + }, + { + "current_steps": 211, + "loss": 3.1053, + "learning_rate": 9.416755037115588e-07, + "epoch": 0.3202416918429003 + }, + { + "current_steps": 212, + "loss": 3.048, + "learning_rate": 9.411452810180275e-07, + "epoch": 0.3217522658610272 + }, + { + "current_steps": 213, + "loss": 3.0653, + "learning_rate": 9.406150583244962e-07, + "epoch": 0.32326283987915405 + }, + { + "current_steps": 214, + "loss": 2.9183, + "learning_rate": 9.40084835630965e-07, + "epoch": 0.324773413897281 + }, + { + "current_steps": 215, + "loss": 3.0895, + "learning_rate": 9.395546129374337e-07, + "epoch": 0.32628398791540786 + }, + { + "current_steps": 216, + "loss": 2.6947, + "learning_rate": 9.390243902439024e-07, + "epoch": 0.32779456193353473 + }, + { + "current_steps": 217, + "loss": 3.0717, + "learning_rate": 9.384941675503711e-07, + "epoch": 0.3293051359516616 + }, + { + "current_steps": 218, + "loss": 2.9181, + "learning_rate": 9.379639448568398e-07, + "epoch": 0.33081570996978854 + }, + { + "current_steps": 219, + "loss": 3.0709, + "learning_rate": 9.374337221633085e-07, + "epoch": 0.3323262839879154 + }, + { + "current_steps": 220, + "loss": 2.9505, + "learning_rate": 9.369034994697774e-07, + "epoch": 0.3338368580060423 + }, + { + "current_steps": 221, + "loss": 2.8928, + "learning_rate": 9.36373276776246e-07, + "epoch": 0.33534743202416917 + }, + { + "current_steps": 222, + "loss": 3.1669, + "learning_rate": 9.358430540827147e-07, + "epoch": 0.3368580060422961 + }, + { + "current_steps": 223, + "loss": 3.1317, + "learning_rate": 9.353128313891834e-07, + "epoch": 0.338368580060423 + }, + { + "current_steps": 224, + "loss": 3.1631, + "learning_rate": 9.347826086956522e-07, + "epoch": 0.33987915407854985 + }, + { + "current_steps": 225, + "loss": 2.6912, + "learning_rate": 9.342523860021208e-07, + "epoch": 0.3413897280966767 + }, + { + "current_steps": 226, + "loss": 3.0521, + "learning_rate": 9.337221633085895e-07, + "epoch": 0.3429003021148036 + }, + { + "current_steps": 227, + "loss": 3.3139, + "learning_rate": 9.331919406150583e-07, + "epoch": 0.34441087613293053 + }, + { + "current_steps": 228, + "loss": 3.0466, + "learning_rate": 9.326617179215271e-07, + "epoch": 0.3459214501510574 + }, + { + "current_steps": 229, + "loss": 3.2494, + "learning_rate": 9.321314952279957e-07, + "epoch": 0.3474320241691843 + }, + { + "current_steps": 230, + "loss": 3.2022, + "learning_rate": 9.316012725344644e-07, + "epoch": 0.34894259818731116 + }, + { + "current_steps": 231, + "loss": 3.1887, + "learning_rate": 9.310710498409331e-07, + "epoch": 0.3504531722054381 + }, + { + "current_steps": 232, + "loss": 3.2125, + "learning_rate": 9.305408271474019e-07, + "epoch": 0.35196374622356497 + }, + { + "current_steps": 233, + "loss": 3.0106, + "learning_rate": 9.300106044538706e-07, + "epoch": 0.35347432024169184 + }, + { + "current_steps": 234, + "loss": 3.1587, + "learning_rate": 9.294803817603393e-07, + "epoch": 0.3549848942598187 + }, + { + "current_steps": 235, + "loss": 3.0159, + "learning_rate": 9.289501590668081e-07, + "epoch": 0.3564954682779456 + }, + { + "current_steps": 236, + "loss": 2.9888, + "learning_rate": 9.284199363732768e-07, + "epoch": 0.3580060422960725 + }, + { + "current_steps": 237, + "loss": 3.3355, + "learning_rate": 9.278897136797454e-07, + "epoch": 0.3595166163141994 + }, + { + "current_steps": 238, + "loss": 3.2464, + "learning_rate": 9.273594909862141e-07, + "epoch": 0.3610271903323263 + }, + { + "current_steps": 239, + "loss": 3.1984, + "learning_rate": 9.26829268292683e-07, + "epoch": 0.36253776435045315 + }, + { + "current_steps": 240, + "loss": 3.1774, + "learning_rate": 9.262990455991516e-07, + "epoch": 0.3640483383685801 + }, + { + "current_steps": 241, + "loss": 3.1891, + "learning_rate": 9.257688229056203e-07, + "epoch": 0.36555891238670696 + }, + { + "current_steps": 242, + "loss": 3.2667, + "learning_rate": 9.25238600212089e-07, + "epoch": 0.36706948640483383 + }, + { + "current_steps": 243, + "loss": 3.0815, + "learning_rate": 9.25238600212089e-07, + "epoch": 0.3685800604229607 + }, + { + "current_steps": 244, + "loss": 2.9868, + "learning_rate": 9.247083775185578e-07, + "epoch": 0.37009063444108764 + }, + { + "current_steps": 245, + "loss": 3.2655, + "learning_rate": 9.241781548250264e-07, + "epoch": 0.3716012084592145 + }, + { + "current_steps": 246, + "loss": 3.2307, + "learning_rate": 9.236479321314951e-07, + "epoch": 0.3731117824773414 + }, + { + "current_steps": 247, + "loss": 3.2197, + "learning_rate": 9.231177094379639e-07, + "epoch": 0.37462235649546827 + }, + { + "current_steps": 248, + "loss": 3.2812, + "learning_rate": 9.225874867444327e-07, + "epoch": 0.37613293051359514 + }, + { + "current_steps": 249, + "loss": 3.0202, + "learning_rate": 9.220572640509013e-07, + "epoch": 0.3776435045317221 + }, + { + "current_steps": 250, + "loss": 2.7464, + "learning_rate": 9.2152704135737e-07, + "epoch": 0.37915407854984895 + }, + { + "current_steps": 251, + "loss": 3.0009, + "learning_rate": 9.209968186638388e-07, + "epoch": 0.3806646525679758 + }, + { + "current_steps": 252, + "loss": 2.9182, + "learning_rate": 9.204665959703075e-07, + "epoch": 0.3821752265861027 + }, + { + "current_steps": 253, + "loss": 3.0978, + "learning_rate": 9.199363732767762e-07, + "epoch": 0.38368580060422963 + }, + { + "current_steps": 254, + "loss": 2.5917, + "learning_rate": 9.194061505832449e-07, + "epoch": 0.3851963746223565 + }, + { + "current_steps": 255, + "loss": 3.2166, + "learning_rate": 9.188759278897137e-07, + "epoch": 0.3867069486404834 + }, + { + "current_steps": 256, + "loss": 3.049, + "learning_rate": 9.183457051961824e-07, + "epoch": 0.38821752265861026 + }, + { + "current_steps": 257, + "loss": 2.9817, + "learning_rate": 9.17815482502651e-07, + "epoch": 0.38972809667673713 + }, + { + "current_steps": 258, + "loss": 3.3222, + "learning_rate": 9.172852598091197e-07, + "epoch": 0.39123867069486407 + }, + { + "current_steps": 259, + "loss": 3.1317, + "learning_rate": 9.167550371155886e-07, + "epoch": 0.39274924471299094 + }, + { + "current_steps": 260, + "loss": 2.9371, + "learning_rate": 9.162248144220573e-07, + "epoch": 0.3942598187311178 + }, + { + "current_steps": 261, + "loss": 3.1652, + "learning_rate": 9.156945917285259e-07, + "epoch": 0.3957703927492447 + }, + { + "current_steps": 262, + "loss": 3.1902, + "learning_rate": 9.151643690349946e-07, + "epoch": 0.3972809667673716 + }, + { + "current_steps": 263, + "loss": 3.1463, + "learning_rate": 9.146341463414634e-07, + "epoch": 0.3987915407854985 + }, + { + "current_steps": 264, + "loss": 2.9862, + "learning_rate": 9.141039236479321e-07, + "epoch": 0.4003021148036254 + }, + { + "current_steps": 265, + "loss": 3.3495, + "learning_rate": 9.135737009544007e-07, + "epoch": 0.40181268882175225 + }, + { + "current_steps": 266, + "loss": 2.5696, + "learning_rate": 9.130434782608695e-07, + "epoch": 0.4033232628398791 + }, + { + "current_steps": 267, + "loss": 2.9964, + "learning_rate": 9.125132555673383e-07, + "epoch": 0.40483383685800606 + }, + { + "current_steps": 268, + "loss": 2.6063, + "learning_rate": 9.11983032873807e-07, + "epoch": 0.40634441087613293 + }, + { + "current_steps": 269, + "loss": 3.5285, + "learning_rate": 9.114528101802756e-07, + "epoch": 0.4078549848942598 + }, + { + "current_steps": 270, + "loss": 3.3055, + "learning_rate": 9.109225874867444e-07, + "epoch": 0.4093655589123867 + }, + { + "current_steps": 271, + "loss": 3.1632, + "learning_rate": 9.103923647932131e-07, + "epoch": 0.4108761329305136 + }, + { + "current_steps": 272, + "loss": 3.0509, + "learning_rate": 9.098621420996819e-07, + "epoch": 0.4123867069486405 + }, + { + "current_steps": 273, + "loss": 3.1983, + "learning_rate": 9.093319194061505e-07, + "epoch": 0.41389728096676737 + }, + { + "current_steps": 274, + "loss": 3.1477, + "learning_rate": 9.088016967126193e-07, + "epoch": 0.41540785498489424 + }, + { + "current_steps": 275, + "loss": 2.8309, + "learning_rate": 9.08271474019088e-07, + "epoch": 0.4169184290030212 + }, + { + "current_steps": 276, + "loss": 2.5193, + "learning_rate": 9.077412513255567e-07, + "epoch": 0.41842900302114805 + }, + { + "current_steps": 277, + "loss": 2.6047, + "learning_rate": 9.072110286320253e-07, + "epoch": 0.4199395770392749 + }, + { + "current_steps": 278, + "loss": 3.1801, + "learning_rate": 9.066808059384942e-07, + "epoch": 0.4214501510574018 + }, + { + "current_steps": 279, + "loss": 3.2664, + "learning_rate": 9.061505832449629e-07, + "epoch": 0.4229607250755287 + }, + { + "current_steps": 280, + "loss": 3.1992, + "learning_rate": 9.056203605514315e-07, + "epoch": 0.4244712990936556 + }, + { + "current_steps": 281, + "loss": 3.003, + "learning_rate": 9.050901378579002e-07, + "epoch": 0.4259818731117825 + }, + { + "current_steps": 282, + "loss": 3.1076, + "learning_rate": 9.04559915164369e-07, + "epoch": 0.42749244712990936 + }, + { + "current_steps": 283, + "loss": 3.032, + "learning_rate": 9.040296924708377e-07, + "epoch": 0.42900302114803623 + }, + { + "current_steps": 284, + "loss": 3.3778, + "learning_rate": 9.034994697773063e-07, + "epoch": 0.43051359516616317 + }, + { + "current_steps": 285, + "loss": 3.0228, + "learning_rate": 9.029692470837752e-07, + "epoch": 0.43202416918429004 + }, + { + "current_steps": 286, + "loss": 3.1486, + "learning_rate": 9.024390243902439e-07, + "epoch": 0.4335347432024169 + }, + { + "current_steps": 287, + "loss": 2.5894, + "learning_rate": 9.019088016967126e-07, + "epoch": 0.4350453172205438 + }, + { + "current_steps": 288, + "loss": 3.0163, + "learning_rate": 9.013785790031812e-07, + "epoch": 0.43655589123867067 + }, + { + "current_steps": 289, + "loss": 2.8523, + "learning_rate": 9.0084835630965e-07, + "epoch": 0.4380664652567976 + }, + { + "current_steps": 290, + "loss": 2.9167, + "learning_rate": 9.003181336161188e-07, + "epoch": 0.4395770392749245 + }, + { + "current_steps": 291, + "loss": 3.3304, + "learning_rate": 8.997879109225875e-07, + "epoch": 0.44108761329305135 + }, + { + "current_steps": 292, + "loss": 2.8962, + "learning_rate": 8.992576882290561e-07, + "epoch": 0.4425981873111782 + }, + { + "current_steps": 293, + "loss": 3.2467, + "learning_rate": 8.987274655355249e-07, + "epoch": 0.44410876132930516 + }, + { + "current_steps": 294, + "loss": 3.0529, + "learning_rate": 8.981972428419936e-07, + "epoch": 0.44561933534743203 + }, + { + "current_steps": 295, + "loss": 3.0481, + "learning_rate": 8.976670201484623e-07, + "epoch": 0.4471299093655589 + }, + { + "current_steps": 296, + "loss": 3.081, + "learning_rate": 8.971367974549309e-07, + "epoch": 0.4486404833836858 + }, + { + "current_steps": 297, + "loss": 3.2012, + "learning_rate": 8.966065747613998e-07, + "epoch": 0.4501510574018127 + }, + { + "current_steps": 298, + "loss": 3.2297, + "learning_rate": 8.960763520678685e-07, + "epoch": 0.4516616314199396 + }, + { + "current_steps": 299, + "loss": 2.7969, + "learning_rate": 8.955461293743372e-07, + "epoch": 0.45317220543806647 + }, + { + "current_steps": 300, + "loss": 3.2338, + "learning_rate": 8.950159066808058e-07, + "epoch": 0.45468277945619334 + }, + { + "current_steps": 301, + "loss": 3.5633, + "learning_rate": 8.944856839872746e-07, + "epoch": 0.4561933534743202 + }, + { + "current_steps": 302, + "loss": 3.2321, + "learning_rate": 8.939554612937433e-07, + "epoch": 0.45770392749244715 + }, + { + "current_steps": 303, + "loss": 3.2112, + "learning_rate": 8.934252386002121e-07, + "epoch": 0.459214501510574 + }, + { + "current_steps": 304, + "loss": 3.13, + "learning_rate": 8.928950159066808e-07, + "epoch": 0.4607250755287009 + }, + { + "current_steps": 305, + "loss": 3.1433, + "learning_rate": 8.923647932131495e-07, + "epoch": 0.4622356495468278 + }, + { + "current_steps": 306, + "loss": 3.2684, + "learning_rate": 8.918345705196182e-07, + "epoch": 0.4637462235649547 + }, + { + "current_steps": 307, + "loss": 3.0937, + "learning_rate": 8.913043478260869e-07, + "epoch": 0.4652567975830816 + }, + { + "current_steps": 308, + "loss": 3.3234, + "learning_rate": 8.907741251325556e-07, + "epoch": 0.46676737160120846 + }, + { + "current_steps": 309, + "loss": 3.3496, + "learning_rate": 8.902439024390244e-07, + "epoch": 0.46827794561933533 + }, + { + "current_steps": 310, + "loss": 3.1118, + "learning_rate": 8.897136797454931e-07, + "epoch": 0.4697885196374622 + }, + { + "current_steps": 311, + "loss": 3.3059, + "learning_rate": 8.891834570519618e-07, + "epoch": 0.47129909365558914 + }, + { + "current_steps": 312, + "loss": 2.9695, + "learning_rate": 8.886532343584305e-07, + "epoch": 0.472809667673716 + }, + { + "current_steps": 313, + "loss": 3.2671, + "learning_rate": 8.881230116648992e-07, + "epoch": 0.4743202416918429 + }, + { + "current_steps": 314, + "loss": 3.0554, + "learning_rate": 8.875927889713679e-07, + "epoch": 0.47583081570996977 + }, + { + "current_steps": 315, + "loss": 3.2973, + "learning_rate": 8.870625662778366e-07, + "epoch": 0.4773413897280967 + }, + { + "current_steps": 316, + "loss": 3.2223, + "learning_rate": 8.865323435843054e-07, + "epoch": 0.4788519637462236 + }, + { + "current_steps": 317, + "loss": 3.4084, + "learning_rate": 8.860021208907741e-07, + "epoch": 0.48036253776435045 + }, + { + "current_steps": 318, + "loss": 3.0606, + "learning_rate": 8.854718981972428e-07, + "epoch": 0.4818731117824773 + }, + { + "current_steps": 319, + "loss": 3.3439, + "learning_rate": 8.849416755037116e-07, + "epoch": 0.48338368580060426 + }, + { + "current_steps": 320, + "loss": 3.2179, + "learning_rate": 8.844114528101802e-07, + "epoch": 0.48489425981873113 + }, + { + "current_steps": 321, + "loss": 3.1217, + "learning_rate": 8.838812301166489e-07, + "epoch": 0.486404833836858 + }, + { + "current_steps": 322, + "loss": 3.1913, + "learning_rate": 8.833510074231177e-07, + "epoch": 0.4879154078549849 + }, + { + "current_steps": 323, + "loss": 3.0094, + "learning_rate": 8.828207847295865e-07, + "epoch": 0.48942598187311176 + }, + { + "current_steps": 324, + "loss": 2.7767, + "learning_rate": 8.822905620360551e-07, + "epoch": 0.4909365558912387 + }, + { + "current_steps": 325, + "loss": 3.084, + "learning_rate": 8.817603393425238e-07, + "epoch": 0.49244712990936557 + }, + { + "current_steps": 326, + "loss": 2.9481, + "learning_rate": 8.812301166489925e-07, + "epoch": 0.49395770392749244 + }, + { + "current_steps": 327, + "loss": 3.0572, + "learning_rate": 8.806998939554613e-07, + "epoch": 0.4954682779456193 + }, + { + "current_steps": 328, + "loss": 3.0843, + "learning_rate": 8.8016967126193e-07, + "epoch": 0.49697885196374625 + }, + { + "current_steps": 329, + "loss": 2.938, + "learning_rate": 8.796394485683987e-07, + "epoch": 0.4984894259818731 + }, + { + "current_steps": 330, + "loss": 2.9294, + "learning_rate": 8.791092258748674e-07, + "epoch": 0.5 + }, + { + "current_steps": 331, + "loss": 2.9109, + "learning_rate": 8.785790031813361e-07, + "epoch": 0.5015105740181269 + }, + { + "current_steps": 332, + "loss": 2.9943, + "learning_rate": 8.780487804878048e-07, + "epoch": 0.5030211480362538 + }, + { + "current_steps": 333, + "loss": 2.7782, + "learning_rate": 8.775185577942735e-07, + "epoch": 0.5045317220543807 + }, + { + "current_steps": 334, + "loss": 3.1677, + "learning_rate": 8.769883351007424e-07, + "epoch": 0.5060422960725075 + }, + { + "current_steps": 335, + "loss": 2.9955, + "learning_rate": 8.76458112407211e-07, + "epoch": 0.5075528700906344 + }, + { + "current_steps": 336, + "loss": 3.3114, + "learning_rate": 8.759278897136797e-07, + "epoch": 0.5090634441087614 + }, + { + "current_steps": 337, + "loss": 3.2823, + "learning_rate": 8.753976670201484e-07, + "epoch": 0.5105740181268882 + }, + { + "current_steps": 338, + "loss": 2.7996, + "learning_rate": 8.748674443266172e-07, + "epoch": 0.5120845921450151 + }, + { + "current_steps": 339, + "loss": 3.1617, + "learning_rate": 8.743372216330858e-07, + "epoch": 0.513595166163142 + }, + { + "current_steps": 340, + "loss": 3.1676, + "learning_rate": 8.738069989395545e-07, + "epoch": 0.5151057401812689 + }, + { + "current_steps": 341, + "loss": 2.4332, + "learning_rate": 8.732767762460233e-07, + "epoch": 0.5166163141993958 + }, + { + "current_steps": 342, + "loss": 3.1305, + "learning_rate": 8.727465535524921e-07, + "epoch": 0.5181268882175226 + }, + { + "current_steps": 343, + "loss": 3.1859, + "learning_rate": 8.722163308589607e-07, + "epoch": 0.5196374622356495 + }, + { + "current_steps": 344, + "loss": 3.0989, + "learning_rate": 8.716861081654294e-07, + "epoch": 0.5211480362537765 + }, + { + "current_steps": 345, + "loss": 2.9217, + "learning_rate": 8.711558854718981e-07, + "epoch": 0.5226586102719033 + }, + { + "current_steps": 346, + "loss": 3.0529, + "learning_rate": 8.706256627783669e-07, + "epoch": 0.5241691842900302 + }, + { + "current_steps": 347, + "loss": 2.9959, + "learning_rate": 8.700954400848356e-07, + "epoch": 0.525679758308157 + }, + { + "current_steps": 348, + "loss": 3.6103, + "learning_rate": 8.695652173913043e-07, + "epoch": 0.527190332326284 + }, + { + "current_steps": 349, + "loss": 3.1724, + "learning_rate": 8.69034994697773e-07, + "epoch": 0.5287009063444109 + }, + { + "current_steps": 350, + "loss": 3.1836, + "learning_rate": 8.685047720042418e-07, + "epoch": 0.5302114803625377 + }, + { + "current_steps": 351, + "loss": 3.2854, + "learning_rate": 8.679745493107104e-07, + "epoch": 0.5317220543806647 + }, + { + "current_steps": 352, + "loss": 3.1328, + "learning_rate": 8.674443266171791e-07, + "epoch": 0.5332326283987915 + }, + { + "current_steps": 353, + "loss": 2.0668, + "learning_rate": 8.66914103923648e-07, + "epoch": 0.5347432024169184 + }, + { + "current_steps": 354, + "loss": 3.0729, + "learning_rate": 8.663838812301167e-07, + "epoch": 0.5362537764350453 + }, + { + "current_steps": 355, + "loss": 3.1418, + "learning_rate": 8.658536585365853e-07, + "epoch": 0.5377643504531722 + }, + { + "current_steps": 356, + "loss": 3.1024, + "learning_rate": 8.65323435843054e-07, + "epoch": 0.5392749244712991 + }, + { + "current_steps": 357, + "loss": 3.0899, + "learning_rate": 8.647932131495228e-07, + "epoch": 0.540785498489426 + }, + { + "current_steps": 358, + "loss": 3.1664, + "learning_rate": 8.642629904559915e-07, + "epoch": 0.5422960725075529 + }, + { + "current_steps": 359, + "loss": 3.0295, + "learning_rate": 8.637327677624601e-07, + "epoch": 0.5438066465256798 + }, + { + "current_steps": 360, + "loss": 3.1117, + "learning_rate": 8.632025450689289e-07, + "epoch": 0.5453172205438066 + }, + { + "current_steps": 361, + "loss": 3.2985, + "learning_rate": 8.626723223753977e-07, + "epoch": 0.5468277945619335 + }, + { + "current_steps": 362, + "loss": 3.1515, + "learning_rate": 8.621420996818664e-07, + "epoch": 0.5483383685800605 + }, + { + "current_steps": 363, + "loss": 2.7235, + "learning_rate": 8.61611876988335e-07, + "epoch": 0.5498489425981873 + }, + { + "current_steps": 364, + "loss": 3.1985, + "learning_rate": 8.610816542948037e-07, + "epoch": 0.5513595166163142 + }, + { + "current_steps": 365, + "loss": 2.8398, + "learning_rate": 8.605514316012725e-07, + "epoch": 0.552870090634441 + }, + { + "current_steps": 366, + "loss": 3.0781, + "learning_rate": 8.600212089077413e-07, + "epoch": 0.554380664652568 + }, + { + "current_steps": 367, + "loss": 2.8402, + "learning_rate": 8.594909862142099e-07, + "epoch": 0.5558912386706949 + }, + { + "current_steps": 368, + "loss": 2.8804, + "learning_rate": 8.589607635206787e-07, + "epoch": 0.5574018126888217 + }, + { + "current_steps": 369, + "loss": 3.1703, + "learning_rate": 8.584305408271474e-07, + "epoch": 0.5589123867069486 + }, + { + "current_steps": 370, + "loss": 3.1955, + "learning_rate": 8.579003181336161e-07, + "epoch": 0.5604229607250756 + }, + { + "current_steps": 371, + "loss": 3.1886, + "learning_rate": 8.573700954400847e-07, + "epoch": 0.5619335347432024 + }, + { + "current_steps": 372, + "loss": 3.0821, + "learning_rate": 8.568398727465536e-07, + "epoch": 0.5634441087613293 + }, + { + "current_steps": 373, + "loss": 2.7845, + "learning_rate": 8.563096500530223e-07, + "epoch": 0.5649546827794562 + }, + { + "current_steps": 374, + "loss": 3.1924, + "learning_rate": 8.557794273594909e-07, + "epoch": 0.5664652567975831 + }, + { + "current_steps": 375, + "loss": 3.2262, + "learning_rate": 8.552492046659596e-07, + "epoch": 0.56797583081571 + }, + { + "current_steps": 376, + "loss": 3.0846, + "learning_rate": 8.547189819724284e-07, + "epoch": 0.5694864048338368 + }, + { + "current_steps": 377, + "loss": 3.2293, + "learning_rate": 8.541887592788971e-07, + "epoch": 0.5709969788519638 + }, + { + "current_steps": 378, + "loss": 2.9524, + "learning_rate": 8.536585365853657e-07, + "epoch": 0.5725075528700906 + }, + { + "current_steps": 379, + "loss": 3.2021, + "learning_rate": 8.531283138918345e-07, + "epoch": 0.5740181268882175 + }, + { + "current_steps": 380, + "loss": 3.255, + "learning_rate": 8.525980911983033e-07, + "epoch": 0.5755287009063444 + }, + { + "current_steps": 381, + "loss": 3.151, + "learning_rate": 8.52067868504772e-07, + "epoch": 0.5770392749244713 + }, + { + "current_steps": 382, + "loss": 3.3928, + "learning_rate": 8.515376458112406e-07, + "epoch": 0.5785498489425982 + }, + { + "current_steps": 383, + "loss": 3.0898, + "learning_rate": 8.510074231177093e-07, + "epoch": 0.5800604229607251 + }, + { + "current_steps": 384, + "loss": 3.045, + "learning_rate": 8.504772004241781e-07, + "epoch": 0.581570996978852 + }, + { + "current_steps": 385, + "loss": 2.9802, + "learning_rate": 8.499469777306469e-07, + "epoch": 0.5830815709969789 + }, + { + "current_steps": 386, + "loss": 3.301, + "learning_rate": 8.494167550371155e-07, + "epoch": 0.5845921450151057 + }, + { + "current_steps": 387, + "loss": 2.794, + "learning_rate": 8.488865323435843e-07, + "epoch": 0.5861027190332326 + }, + { + "current_steps": 388, + "loss": 3.3134, + "learning_rate": 8.48356309650053e-07, + "epoch": 0.5876132930513596 + }, + { + "current_steps": 389, + "loss": 3.1513, + "learning_rate": 8.478260869565217e-07, + "epoch": 0.5891238670694864 + }, + { + "current_steps": 390, + "loss": 3.1984, + "learning_rate": 8.472958642629903e-07, + "epoch": 0.5906344410876133 + }, + { + "current_steps": 391, + "loss": 2.9866, + "learning_rate": 8.467656415694592e-07, + "epoch": 0.5921450151057401 + }, + { + "current_steps": 392, + "loss": 3.0448, + "learning_rate": 8.462354188759279e-07, + "epoch": 0.5936555891238671 + }, + { + "current_steps": 393, + "loss": 3.0681, + "learning_rate": 8.457051961823966e-07, + "epoch": 0.595166163141994 + }, + { + "current_steps": 394, + "loss": 3.1794, + "learning_rate": 8.451749734888652e-07, + "epoch": 0.5966767371601208 + }, + { + "current_steps": 395, + "loss": 3.2612, + "learning_rate": 8.44644750795334e-07, + "epoch": 0.5981873111782477 + }, + { + "current_steps": 396, + "loss": 3.2551, + "learning_rate": 8.441145281018027e-07, + "epoch": 0.5996978851963746 + }, + { + "current_steps": 397, + "loss": 2.9567, + "learning_rate": 8.435843054082715e-07, + "epoch": 0.6012084592145015 + }, + { + "current_steps": 398, + "loss": 3.0785, + "learning_rate": 8.430540827147401e-07, + "epoch": 0.6027190332326284 + }, + { + "current_steps": 399, + "loss": 2.9135, + "learning_rate": 8.425238600212089e-07, + "epoch": 0.6042296072507553 + }, + { + "current_steps": 400, + "loss": 3.1097, + "learning_rate": 8.419936373276776e-07, + "epoch": 0.6057401812688822 + }, + { + "current_steps": 401, + "loss": 3.2021, + "learning_rate": 8.414634146341463e-07, + "epoch": 0.6072507552870091 + }, + { + "current_steps": 402, + "loss": 3.0949, + "learning_rate": 8.40933191940615e-07, + "epoch": 0.6087613293051359 + }, + { + "current_steps": 403, + "loss": 2.9137, + "learning_rate": 8.404029692470838e-07, + "epoch": 0.6102719033232629 + }, + { + "current_steps": 404, + "loss": 3.1532, + "learning_rate": 8.398727465535525e-07, + "epoch": 0.6117824773413897 + }, + { + "current_steps": 405, + "loss": 3.2406, + "learning_rate": 8.393425238600212e-07, + "epoch": 0.6132930513595166 + }, + { + "current_steps": 406, + "loss": 3.0427, + "learning_rate": 8.388123011664899e-07, + "epoch": 0.6148036253776435 + }, + { + "current_steps": 407, + "loss": 3.267, + "learning_rate": 8.382820784729586e-07, + "epoch": 0.6163141993957704 + }, + { + "current_steps": 408, + "loss": 3.0729, + "learning_rate": 8.377518557794273e-07, + "epoch": 0.6178247734138973 + }, + { + "current_steps": 409, + "loss": 3.0987, + "learning_rate": 8.37221633085896e-07, + "epoch": 0.6193353474320241 + }, + { + "current_steps": 410, + "loss": 3.1801, + "learning_rate": 8.366914103923648e-07, + "epoch": 0.620845921450151 + }, + { + "current_steps": 411, + "loss": 3.0017, + "learning_rate": 8.361611876988335e-07, + "epoch": 0.622356495468278 + }, + { + "current_steps": 412, + "loss": 3.3731, + "learning_rate": 8.356309650053022e-07, + "epoch": 0.6238670694864048 + }, + { + "current_steps": 413, + "loss": 2.6755, + "learning_rate": 8.351007423117708e-07, + "epoch": 0.6253776435045317 + }, + { + "current_steps": 414, + "loss": 3.1755, + "learning_rate": 8.345705196182396e-07, + "epoch": 0.6268882175226587 + }, + { + "current_steps": 415, + "loss": 3.1861, + "learning_rate": 8.340402969247083e-07, + "epoch": 0.6283987915407855 + }, + { + "current_steps": 416, + "loss": 3.0512, + "learning_rate": 8.335100742311771e-07, + "epoch": 0.6299093655589124 + }, + { + "current_steps": 417, + "loss": 2.2501, + "learning_rate": 8.329798515376459e-07, + "epoch": 0.6314199395770392 + }, + { + "current_steps": 418, + "loss": 3.1965, + "learning_rate": 8.324496288441145e-07, + "epoch": 0.6329305135951662 + }, + { + "current_steps": 419, + "loss": 3.1893, + "learning_rate": 8.319194061505832e-07, + "epoch": 0.6344410876132931 + }, + { + "current_steps": 420, + "loss": 3.3295, + "learning_rate": 8.313891834570519e-07, + "epoch": 0.6359516616314199 + }, + { + "current_steps": 421, + "loss": 3.0503, + "learning_rate": 8.308589607635206e-07, + "epoch": 0.6374622356495468 + }, + { + "current_steps": 422, + "loss": 3.253, + "learning_rate": 8.303287380699894e-07, + "epoch": 0.6389728096676737 + }, + { + "current_steps": 423, + "loss": 3.3239, + "learning_rate": 8.297985153764581e-07, + "epoch": 0.6404833836858006 + }, + { + "current_steps": 424, + "loss": 3.461, + "learning_rate": 8.292682926829268e-07, + "epoch": 0.6419939577039275 + }, + { + "current_steps": 425, + "loss": 2.8585, + "learning_rate": 8.287380699893955e-07, + "epoch": 0.6435045317220544 + }, + { + "current_steps": 426, + "loss": 3.13, + "learning_rate": 8.282078472958642e-07, + "epoch": 0.6450151057401813 + }, + { + "current_steps": 427, + "loss": 3.1843, + "learning_rate": 8.276776246023329e-07, + "epoch": 0.6465256797583081 + }, + { + "current_steps": 428, + "loss": 3.0517, + "learning_rate": 8.271474019088016e-07, + "epoch": 0.648036253776435 + }, + { + "current_steps": 429, + "loss": 3.1236, + "learning_rate": 8.266171792152704e-07, + "epoch": 0.649546827794562 + }, + { + "current_steps": 430, + "loss": 2.8257, + "learning_rate": 8.260869565217391e-07, + "epoch": 0.6510574018126888 + }, + { + "current_steps": 431, + "loss": 2.9849, + "learning_rate": 8.255567338282078e-07, + "epoch": 0.6525679758308157 + }, + { + "current_steps": 432, + "loss": 3.0392, + "learning_rate": 8.250265111346765e-07, + "epoch": 0.6540785498489426 + }, + { + "current_steps": 433, + "loss": 3.1265, + "learning_rate": 8.244962884411452e-07, + "epoch": 0.6555891238670695 + }, + { + "current_steps": 434, + "loss": 2.9919, + "learning_rate": 8.239660657476139e-07, + "epoch": 0.6570996978851964 + }, + { + "current_steps": 435, + "loss": 2.8665, + "learning_rate": 8.234358430540827e-07, + "epoch": 0.6586102719033232 + }, + { + "current_steps": 436, + "loss": 3.2855, + "learning_rate": 8.229056203605515e-07, + "epoch": 0.6601208459214502 + }, + { + "current_steps": 437, + "loss": 3.2686, + "learning_rate": 8.223753976670201e-07, + "epoch": 0.6616314199395771 + }, + { + "current_steps": 438, + "loss": 3.1965, + "learning_rate": 8.218451749734888e-07, + "epoch": 0.6631419939577039 + }, + { + "current_steps": 439, + "loss": 3.1999, + "learning_rate": 8.213149522799575e-07, + "epoch": 0.6646525679758308 + }, + { + "current_steps": 440, + "loss": 3.0052, + "learning_rate": 8.207847295864263e-07, + "epoch": 0.6661631419939577 + }, + { + "current_steps": 441, + "loss": 3.3237, + "learning_rate": 8.20254506892895e-07, + "epoch": 0.6676737160120846 + }, + { + "current_steps": 442, + "loss": 2.9499, + "learning_rate": 8.197242841993637e-07, + "epoch": 0.6691842900302115 + }, + { + "current_steps": 443, + "loss": 3.0801, + "learning_rate": 8.191940615058324e-07, + "epoch": 0.6706948640483383 + }, + { + "current_steps": 444, + "loss": 3.0574, + "learning_rate": 8.186638388123012e-07, + "epoch": 0.6722054380664653 + }, + { + "current_steps": 445, + "loss": 3.068, + "learning_rate": 8.181336161187698e-07, + "epoch": 0.6737160120845922 + }, + { + "current_steps": 446, + "loss": 3.1562, + "learning_rate": 8.176033934252385e-07, + "epoch": 0.675226586102719 + }, + { + "current_steps": 447, + "loss": 2.9136, + "learning_rate": 8.170731707317072e-07, + "epoch": 0.676737160120846 + }, + { + "current_steps": 448, + "loss": 2.4061, + "learning_rate": 8.165429480381761e-07, + "epoch": 0.6782477341389728 + }, + { + "current_steps": 449, + "loss": 3.3174, + "learning_rate": 8.160127253446447e-07, + "epoch": 0.6797583081570997 + }, + { + "current_steps": 450, + "loss": 2.9948, + "learning_rate": 8.154825026511134e-07, + "epoch": 0.6812688821752266 + }, + { + "current_steps": 451, + "loss": 3.0868, + "learning_rate": 8.149522799575822e-07, + "epoch": 0.6827794561933535 + }, + { + "current_steps": 452, + "loss": 3.0387, + "learning_rate": 8.144220572640509e-07, + "epoch": 0.6842900302114804 + }, + { + "current_steps": 453, + "loss": 3.1065, + "learning_rate": 8.138918345705195e-07, + "epoch": 0.6858006042296072 + }, + { + "current_steps": 454, + "loss": 3.3386, + "learning_rate": 8.133616118769883e-07, + "epoch": 0.6873111782477341 + }, + { + "current_steps": 455, + "loss": 2.8352, + "learning_rate": 8.128313891834571e-07, + "epoch": 0.6888217522658611 + }, + { + "current_steps": 456, + "loss": 2.966, + "learning_rate": 8.123011664899258e-07, + "epoch": 0.6903323262839879 + }, + { + "current_steps": 457, + "loss": 3.1063, + "learning_rate": 8.117709437963944e-07, + "epoch": 0.6918429003021148 + }, + { + "current_steps": 458, + "loss": 3.1141, + "learning_rate": 8.112407211028631e-07, + "epoch": 0.6933534743202417 + }, + { + "current_steps": 459, + "loss": 2.979, + "learning_rate": 8.107104984093319e-07, + "epoch": 0.6948640483383686 + }, + { + "current_steps": 460, + "loss": 3.1482, + "learning_rate": 8.101802757158007e-07, + "epoch": 0.6963746223564955 + }, + { + "current_steps": 461, + "loss": 3.07, + "learning_rate": 8.096500530222693e-07, + "epoch": 0.6978851963746223 + }, + { + "current_steps": 462, + "loss": 2.5385, + "learning_rate": 8.09119830328738e-07, + "epoch": 0.6993957703927492 + }, + { + "current_steps": 463, + "loss": 3.0696, + "learning_rate": 8.085896076352068e-07, + "epoch": 0.7009063444108762 + }, + { + "current_steps": 464, + "loss": 2.9707, + "learning_rate": 8.080593849416754e-07, + "epoch": 0.702416918429003 + }, + { + "current_steps": 465, + "loss": 3.2853, + "learning_rate": 8.075291622481441e-07, + "epoch": 0.7039274924471299 + }, + { + "current_steps": 466, + "loss": 3.1772, + "learning_rate": 8.06998939554613e-07, + "epoch": 0.7054380664652568 + }, + { + "current_steps": 467, + "loss": 2.3754, + "learning_rate": 8.064687168610817e-07, + "epoch": 0.7069486404833837 + }, + { + "current_steps": 468, + "loss": 3.1331, + "learning_rate": 8.059384941675503e-07, + "epoch": 0.7084592145015106 + }, + { + "current_steps": 469, + "loss": 3.2332, + "learning_rate": 8.05408271474019e-07, + "epoch": 0.7099697885196374 + }, + { + "current_steps": 470, + "loss": 3.1364, + "learning_rate": 8.048780487804878e-07, + "epoch": 0.7114803625377644 + }, + { + "current_steps": 471, + "loss": 3.0382, + "learning_rate": 8.043478260869565e-07, + "epoch": 0.7129909365558912 + }, + { + "current_steps": 472, + "loss": 3.0213, + "learning_rate": 8.038176033934251e-07, + "epoch": 0.7145015105740181 + }, + { + "current_steps": 473, + "loss": 3.0728, + "learning_rate": 8.032873806998939e-07, + "epoch": 0.716012084592145 + }, + { + "current_steps": 474, + "loss": 2.8303, + "learning_rate": 8.027571580063627e-07, + "epoch": 0.7175226586102719 + }, + { + "current_steps": 475, + "loss": 2.9884, + "learning_rate": 8.022269353128314e-07, + "epoch": 0.7190332326283988 + }, + { + "current_steps": 476, + "loss": 2.8377, + "learning_rate": 8.016967126193e-07, + "epoch": 0.7205438066465257 + }, + { + "current_steps": 477, + "loss": 3.0181, + "learning_rate": 8.011664899257687e-07, + "epoch": 0.7220543806646526 + }, + { + "current_steps": 478, + "loss": 2.967, + "learning_rate": 8.006362672322375e-07, + "epoch": 0.7235649546827795 + }, + { + "current_steps": 479, + "loss": 3.1362, + "learning_rate": 8.001060445387063e-07, + "epoch": 0.7250755287009063 + }, + { + "current_steps": 480, + "loss": 3.0157, + "learning_rate": 7.995758218451749e-07, + "epoch": 0.7265861027190332 + }, + { + "current_steps": 481, + "loss": 3.2133, + "learning_rate": 7.990455991516436e-07, + "epoch": 0.7280966767371602 + }, + { + "current_steps": 482, + "loss": 3.2363, + "learning_rate": 7.985153764581124e-07, + "epoch": 0.729607250755287 + }, + { + "current_steps": 483, + "loss": 2.9194, + "learning_rate": 7.979851537645811e-07, + "epoch": 0.7311178247734139 + }, + { + "current_steps": 484, + "loss": 3.0577, + "learning_rate": 7.974549310710497e-07, + "epoch": 0.7326283987915407 + }, + { + "current_steps": 485, + "loss": 3.1717, + "learning_rate": 7.969247083775186e-07, + "epoch": 0.7341389728096677 + }, + { + "current_steps": 486, + "loss": 3.2147, + "learning_rate": 7.963944856839873e-07, + "epoch": 0.7356495468277946 + }, + { + "current_steps": 487, + "loss": 3.1182, + "learning_rate": 7.95864262990456e-07, + "epoch": 0.7371601208459214 + }, + { + "current_steps": 488, + "loss": 2.8692, + "learning_rate": 7.953340402969246e-07, + "epoch": 0.7386706948640483 + }, + { + "current_steps": 489, + "loss": 2.9588, + "learning_rate": 7.948038176033934e-07, + "epoch": 0.7401812688821753 + }, + { + "current_steps": 490, + "loss": 2.9832, + "learning_rate": 7.942735949098621e-07, + "epoch": 0.7416918429003021 + }, + { + "current_steps": 491, + "loss": 3.1204, + "learning_rate": 7.937433722163309e-07, + "epoch": 0.743202416918429 + }, + { + "current_steps": 492, + "loss": 3.0391, + "learning_rate": 7.932131495227995e-07, + "epoch": 0.7447129909365559 + }, + { + "current_steps": 493, + "loss": 2.8312, + "learning_rate": 7.926829268292683e-07, + "epoch": 0.7462235649546828 + }, + { + "current_steps": 494, + "loss": 2.8298, + "learning_rate": 7.92152704135737e-07, + "epoch": 0.7477341389728097 + }, + { + "current_steps": 495, + "loss": 2.8368, + "learning_rate": 7.916224814422057e-07, + "epoch": 0.7492447129909365 + }, + { + "current_steps": 496, + "loss": 3.212, + "learning_rate": 7.910922587486743e-07, + "epoch": 0.7507552870090635 + }, + { + "current_steps": 497, + "loss": 3.2808, + "learning_rate": 7.905620360551431e-07, + "epoch": 0.7522658610271903 + }, + { + "current_steps": 498, + "loss": 3.1818, + "learning_rate": 7.900318133616119e-07, + "epoch": 0.7537764350453172 + }, + { + "current_steps": 499, + "loss": 2.937, + "learning_rate": 7.895015906680806e-07, + "epoch": 0.7552870090634441 + }, + { + "current_steps": 500, + "loss": 3.1882, + "learning_rate": 7.889713679745493e-07, + "epoch": 0.756797583081571 + }, + { + "current_steps": 501, + "loss": 3.2112, + "learning_rate": 7.88441145281018e-07, + "epoch": 0.7583081570996979 + }, + { + "current_steps": 502, + "loss": 3.063, + "learning_rate": 7.879109225874867e-07, + "epoch": 0.7598187311178247 + }, + { + "current_steps": 503, + "loss": 2.8865, + "learning_rate": 7.873806998939554e-07, + "epoch": 0.7613293051359517 + }, + { + "current_steps": 504, + "loss": 3.1008, + "learning_rate": 7.868504772004242e-07, + "epoch": 0.7628398791540786 + }, + { + "current_steps": 505, + "loss": 2.8965, + "learning_rate": 7.863202545068929e-07, + "epoch": 0.7643504531722054 + }, + { + "current_steps": 506, + "loss": 2.8766, + "learning_rate": 7.857900318133616e-07, + "epoch": 0.7658610271903323 + }, + { + "current_steps": 507, + "loss": 2.3833, + "learning_rate": 7.852598091198302e-07, + "epoch": 0.7673716012084593 + }, + { + "current_steps": 508, + "loss": 3.1039, + "learning_rate": 7.84729586426299e-07, + "epoch": 0.7688821752265861 + }, + { + "current_steps": 509, + "loss": 2.7846, + "learning_rate": 7.841993637327677e-07, + "epoch": 0.770392749244713 + }, + { + "current_steps": 510, + "loss": 3.0165, + "learning_rate": 7.836691410392365e-07, + "epoch": 0.7719033232628398 + }, + { + "current_steps": 511, + "loss": 3.1822, + "learning_rate": 7.831389183457051e-07, + "epoch": 0.7734138972809668 + }, + { + "current_steps": 512, + "loss": 3.2, + "learning_rate": 7.826086956521739e-07, + "epoch": 0.7749244712990937 + }, + { + "current_steps": 513, + "loss": 3.0028, + "learning_rate": 7.820784729586426e-07, + "epoch": 0.7764350453172205 + }, + { + "current_steps": 514, + "loss": 2.9666, + "learning_rate": 7.815482502651113e-07, + "epoch": 0.7779456193353474 + }, + { + "current_steps": 515, + "loss": 2.7161, + "learning_rate": 7.810180275715799e-07, + "epoch": 0.7794561933534743 + }, + { + "current_steps": 516, + "loss": 2.9436, + "learning_rate": 7.804878048780488e-07, + "epoch": 0.7809667673716012 + }, + { + "current_steps": 517, + "loss": 3.1067, + "learning_rate": 7.799575821845175e-07, + "epoch": 0.7824773413897281 + }, + { + "current_steps": 518, + "loss": 2.9706, + "learning_rate": 7.794273594909862e-07, + "epoch": 0.783987915407855 + }, + { + "current_steps": 519, + "loss": 3.1969, + "learning_rate": 7.788971367974549e-07, + "epoch": 0.7854984894259819 + }, + { + "current_steps": 520, + "loss": 2.5519, + "learning_rate": 7.783669141039236e-07, + "epoch": 0.7870090634441088 + }, + { + "current_steps": 521, + "loss": 3.1815, + "learning_rate": 7.778366914103923e-07, + "epoch": 0.7885196374622356 + }, + { + "current_steps": 522, + "loss": 2.8434, + "learning_rate": 7.77306468716861e-07, + "epoch": 0.7900302114803626 + }, + { + "current_steps": 523, + "loss": 3.1529, + "learning_rate": 7.767762460233298e-07, + "epoch": 0.7915407854984894 + }, + { + "current_steps": 524, + "loss": 2.6969, + "learning_rate": 7.762460233297985e-07, + "epoch": 0.7930513595166163 + }, + { + "current_steps": 525, + "loss": 3.0667, + "learning_rate": 7.757158006362672e-07, + "epoch": 0.7945619335347432 + }, + { + "current_steps": 526, + "loss": 3.2016, + "learning_rate": 7.751855779427359e-07, + "epoch": 0.7960725075528701 + }, + { + "current_steps": 527, + "loss": 3.0715, + "learning_rate": 7.746553552492046e-07, + "epoch": 0.797583081570997 + }, + { + "current_steps": 528, + "loss": 3.1189, + "learning_rate": 7.741251325556733e-07, + "epoch": 0.7990936555891238 + }, + { + "current_steps": 529, + "loss": 3.1317, + "learning_rate": 7.735949098621421e-07, + "epoch": 0.8006042296072508 + }, + { + "current_steps": 530, + "loss": 3.1127, + "learning_rate": 7.730646871686108e-07, + "epoch": 0.8021148036253777 + }, + { + "current_steps": 531, + "loss": 3.3035, + "learning_rate": 7.725344644750795e-07, + "epoch": 0.8036253776435045 + }, + { + "current_steps": 532, + "loss": 3.2779, + "learning_rate": 7.720042417815482e-07, + "epoch": 0.8051359516616314 + }, + { + "current_steps": 533, + "loss": 2.9717, + "learning_rate": 7.714740190880169e-07, + "epoch": 0.8066465256797583 + }, + { + "current_steps": 534, + "loss": 3.1084, + "learning_rate": 7.709437963944857e-07, + "epoch": 0.8081570996978852 + }, + { + "current_steps": 535, + "loss": 3.0763, + "learning_rate": 7.704135737009544e-07, + "epoch": 0.8096676737160121 + }, + { + "current_steps": 536, + "loss": 2.4254, + "learning_rate": 7.698833510074231e-07, + "epoch": 0.8111782477341389 + }, + { + "current_steps": 537, + "loss": 3.2848, + "learning_rate": 7.693531283138918e-07, + "epoch": 0.8126888217522659 + }, + { + "current_steps": 538, + "loss": 2.9759, + "learning_rate": 7.688229056203606e-07, + "epoch": 0.8141993957703928 + }, + { + "current_steps": 539, + "loss": 2.8448, + "learning_rate": 7.682926829268292e-07, + "epoch": 0.8157099697885196 + }, + { + "current_steps": 540, + "loss": 2.8971, + "learning_rate": 7.677624602332979e-07, + "epoch": 0.8172205438066465 + }, + { + "current_steps": 541, + "loss": 3.0558, + "learning_rate": 7.672322375397666e-07, + "epoch": 0.8187311178247734 + }, + { + "current_steps": 542, + "loss": 3.1267, + "learning_rate": 7.667020148462355e-07, + "epoch": 0.8202416918429003 + }, + { + "current_steps": 543, + "loss": 2.6363, + "learning_rate": 7.661717921527041e-07, + "epoch": 0.8217522658610272 + }, + { + "current_steps": 544, + "loss": 2.9708, + "learning_rate": 7.656415694591728e-07, + "epoch": 0.823262839879154 + }, + { + "current_steps": 545, + "loss": 3.2362, + "learning_rate": 7.651113467656415e-07, + "epoch": 0.824773413897281 + }, + { + "current_steps": 546, + "loss": 3.0293, + "learning_rate": 7.645811240721103e-07, + "epoch": 0.8262839879154078 + }, + { + "current_steps": 547, + "loss": 3.1935, + "learning_rate": 7.640509013785789e-07, + "epoch": 0.8277945619335347 + }, + { + "current_steps": 548, + "loss": 3.0003, + "learning_rate": 7.635206786850477e-07, + "epoch": 0.8293051359516617 + }, + { + "current_steps": 549, + "loss": 3.2585, + "learning_rate": 7.629904559915165e-07, + "epoch": 0.8308157099697885 + }, + { + "current_steps": 550, + "loss": 3.1151, + "learning_rate": 7.624602332979852e-07, + "epoch": 0.8323262839879154 + }, + { + "current_steps": 551, + "loss": 2.6131, + "learning_rate": 7.619300106044538e-07, + "epoch": 0.8338368580060423 + }, + { + "current_steps": 552, + "loss": 2.9351, + "learning_rate": 7.613997879109225e-07, + "epoch": 0.8353474320241692 + }, + { + "current_steps": 553, + "loss": 3.0069, + "learning_rate": 7.608695652173913e-07, + "epoch": 0.8368580060422961 + }, + { + "current_steps": 554, + "loss": 2.5323, + "learning_rate": 7.6033934252386e-07, + "epoch": 0.8383685800604229 + }, + { + "current_steps": 555, + "loss": 3.1387, + "learning_rate": 7.598091198303287e-07, + "epoch": 0.8398791540785498 + }, + { + "current_steps": 556, + "loss": 3.0216, + "learning_rate": 7.592788971367974e-07, + "epoch": 0.8413897280966768 + }, + { + "current_steps": 557, + "loss": 3.2079, + "learning_rate": 7.587486744432662e-07, + "epoch": 0.8429003021148036 + }, + { + "current_steps": 558, + "loss": 3.0622, + "learning_rate": 7.582184517497348e-07, + "epoch": 0.8444108761329305 + }, + { + "current_steps": 559, + "loss": 2.4394, + "learning_rate": 7.576882290562035e-07, + "epoch": 0.8459214501510574 + }, + { + "current_steps": 560, + "loss": 3.1407, + "learning_rate": 7.571580063626722e-07, + "epoch": 0.8474320241691843 + }, + { + "current_steps": 561, + "loss": 2.9881, + "learning_rate": 7.566277836691411e-07, + "epoch": 0.8489425981873112 + }, + { + "current_steps": 562, + "loss": 2.9118, + "learning_rate": 7.560975609756097e-07, + "epoch": 0.850453172205438 + }, + { + "current_steps": 563, + "loss": 2.5422, + "learning_rate": 7.555673382820784e-07, + "epoch": 0.851963746223565 + }, + { + "current_steps": 564, + "loss": 3.0906, + "learning_rate": 7.550371155885471e-07, + "epoch": 0.8534743202416919 + }, + { + "current_steps": 565, + "loss": 3.019, + "learning_rate": 7.545068928950159e-07, + "epoch": 0.8549848942598187 + }, + { + "current_steps": 566, + "loss": 3.1581, + "learning_rate": 7.539766702014845e-07, + "epoch": 0.8564954682779456 + }, + { + "current_steps": 567, + "loss": 3.0822, + "learning_rate": 7.534464475079533e-07, + "epoch": 0.8580060422960725 + }, + { + "current_steps": 568, + "loss": 3.1795, + "learning_rate": 7.529162248144221e-07, + "epoch": 0.8595166163141994 + }, + { + "current_steps": 569, + "loss": 3.1357, + "learning_rate": 7.523860021208908e-07, + "epoch": 0.8610271903323263 + }, + { + "current_steps": 570, + "loss": 3.0625, + "learning_rate": 7.518557794273594e-07, + "epoch": 0.8625377643504532 + }, + { + "current_steps": 571, + "loss": 3.013, + "learning_rate": 7.513255567338281e-07, + "epoch": 0.8640483383685801 + }, + { + "current_steps": 572, + "loss": 3.1232, + "learning_rate": 7.507953340402969e-07, + "epoch": 0.8655589123867069 + }, + { + "current_steps": 573, + "loss": 3.0944, + "learning_rate": 7.502651113467657e-07, + "epoch": 0.8670694864048338 + }, + { + "current_steps": 574, + "loss": 3.3296, + "learning_rate": 7.497348886532343e-07, + "epoch": 0.8685800604229608 + }, + { + "current_steps": 575, + "loss": 3.1027, + "learning_rate": 7.49204665959703e-07, + "epoch": 0.8700906344410876 + }, + { + "current_steps": 576, + "loss": 2.9162, + "learning_rate": 7.486744432661718e-07, + "epoch": 0.8716012084592145 + }, + { + "current_steps": 577, + "loss": 3.242, + "learning_rate": 7.481442205726405e-07, + "epoch": 0.8731117824773413 + }, + { + "current_steps": 578, + "loss": 2.8622, + "learning_rate": 7.476139978791091e-07, + "epoch": 0.8746223564954683 + }, + { + "current_steps": 579, + "loss": 3.0473, + "learning_rate": 7.470837751855779e-07, + "epoch": 0.8761329305135952 + }, + { + "current_steps": 580, + "loss": 2.3757, + "learning_rate": 7.465535524920467e-07, + "epoch": 0.877643504531722 + }, + { + "current_steps": 581, + "loss": 3.076, + "learning_rate": 7.460233297985154e-07, + "epoch": 0.879154078549849 + }, + { + "current_steps": 582, + "loss": 3.1744, + "learning_rate": 7.45493107104984e-07, + "epoch": 0.8806646525679759 + }, + { + "current_steps": 583, + "loss": 2.5494, + "learning_rate": 7.449628844114528e-07, + "epoch": 0.8821752265861027 + }, + { + "current_steps": 584, + "loss": 3.0651, + "learning_rate": 7.444326617179215e-07, + "epoch": 0.8836858006042296 + }, + { + "current_steps": 585, + "loss": 3.062, + "learning_rate": 7.439024390243903e-07, + "epoch": 0.8851963746223565 + }, + { + "current_steps": 586, + "loss": 3.0721, + "learning_rate": 7.433722163308589e-07, + "epoch": 0.8867069486404834 + }, + { + "current_steps": 587, + "loss": 2.9384, + "learning_rate": 7.428419936373277e-07, + "epoch": 0.8882175226586103 + }, + { + "current_steps": 588, + "loss": 3.1853, + "learning_rate": 7.423117709437964e-07, + "epoch": 0.8897280966767371 + }, + { + "current_steps": 589, + "loss": 2.8847, + "learning_rate": 7.417815482502651e-07, + "epoch": 0.8912386706948641 + }, + { + "current_steps": 590, + "loss": 3.0281, + "learning_rate": 7.412513255567337e-07, + "epoch": 0.8927492447129909 + }, + { + "current_steps": 591, + "loss": 2.9952, + "learning_rate": 7.407211028632025e-07, + "epoch": 0.8942598187311178 + }, + { + "current_steps": 592, + "loss": 2.9052, + "learning_rate": 7.401908801696713e-07, + "epoch": 0.8957703927492447 + }, + { + "current_steps": 593, + "loss": 3.2017, + "learning_rate": 7.3966065747614e-07, + "epoch": 0.8972809667673716 + }, + { + "current_steps": 594, + "loss": 2.5512, + "learning_rate": 7.391304347826086e-07, + "epoch": 0.8987915407854985 + }, + { + "current_steps": 595, + "loss": 2.5658, + "learning_rate": 7.386002120890774e-07, + "epoch": 0.9003021148036254 + }, + { + "current_steps": 596, + "loss": 3.0919, + "learning_rate": 7.380699893955461e-07, + "epoch": 0.9018126888217523 + }, + { + "current_steps": 597, + "loss": 3.0511, + "learning_rate": 7.375397667020147e-07, + "epoch": 0.9033232628398792 + }, + { + "current_steps": 598, + "loss": 2.8342, + "learning_rate": 7.370095440084835e-07, + "epoch": 0.904833836858006 + }, + { + "current_steps": 599, + "loss": 3.0754, + "learning_rate": 7.364793213149523e-07, + "epoch": 0.9063444108761329 + }, + { + "current_steps": 600, + "loss": 3.1201, + "learning_rate": 7.35949098621421e-07, + "epoch": 0.9078549848942599 + }, + { + "current_steps": 601, + "loss": 2.6998, + "learning_rate": 7.354188759278896e-07, + "epoch": 0.9093655589123867 + }, + { + "current_steps": 602, + "loss": 3.2571, + "learning_rate": 7.348886532343584e-07, + "epoch": 0.9108761329305136 + }, + { + "current_steps": 603, + "loss": 2.898, + "learning_rate": 7.343584305408271e-07, + "epoch": 0.9123867069486404 + }, + { + "current_steps": 604, + "loss": 3.1187, + "learning_rate": 7.338282078472959e-07, + "epoch": 0.9138972809667674 + }, + { + "current_steps": 605, + "loss": 3.0872, + "learning_rate": 7.332979851537645e-07, + "epoch": 0.9154078549848943 + }, + { + "current_steps": 606, + "loss": 3.1769, + "learning_rate": 7.327677624602333e-07, + "epoch": 0.9169184290030211 + }, + { + "current_steps": 607, + "loss": 2.9847, + "learning_rate": 7.32237539766702e-07, + "epoch": 0.918429003021148 + }, + { + "current_steps": 608, + "loss": 3.0472, + "learning_rate": 7.317073170731707e-07, + "epoch": 0.9199395770392749 + }, + { + "current_steps": 609, + "loss": 3.1711, + "learning_rate": 7.311770943796393e-07, + "epoch": 0.9214501510574018 + }, + { + "current_steps": 610, + "loss": 2.9676, + "learning_rate": 7.306468716861081e-07, + "epoch": 0.9229607250755287 + }, + { + "current_steps": 611, + "loss": 3.2103, + "learning_rate": 7.301166489925769e-07, + "epoch": 0.9244712990936556 + }, + { + "current_steps": 612, + "loss": 3.1522, + "learning_rate": 7.295864262990456e-07, + "epoch": 0.9259818731117825 + }, + { + "current_steps": 613, + "loss": 3.0049, + "learning_rate": 7.290562036055142e-07, + "epoch": 0.9274924471299094 + }, + { + "current_steps": 614, + "loss": 3.1888, + "learning_rate": 7.28525980911983e-07, + "epoch": 0.9290030211480362 + }, + { + "current_steps": 615, + "loss": 3.042, + "learning_rate": 7.279957582184517e-07, + "epoch": 0.9305135951661632 + }, + { + "current_steps": 616, + "loss": 3.1997, + "learning_rate": 7.274655355249204e-07, + "epoch": 0.93202416918429 + }, + { + "current_steps": 617, + "loss": 3.084, + "learning_rate": 7.269353128313892e-07, + "epoch": 0.9335347432024169 + }, + { + "current_steps": 618, + "loss": 3.0364, + "learning_rate": 7.264050901378579e-07, + "epoch": 0.9350453172205438 + }, + { + "current_steps": 619, + "loss": 2.993, + "learning_rate": 7.258748674443266e-07, + "epoch": 0.9365558912386707 + }, + { + "current_steps": 620, + "loss": 3.2291, + "learning_rate": 7.253446447507953e-07, + "epoch": 0.9380664652567976 + }, + { + "current_steps": 621, + "loss": 3.1331, + "learning_rate": 7.24814422057264e-07, + "epoch": 0.9395770392749244 + }, + { + "current_steps": 622, + "loss": 3.1535, + "learning_rate": 7.242841993637327e-07, + "epoch": 0.9410876132930514 + }, + { + "current_steps": 623, + "loss": 2.919, + "learning_rate": 7.237539766702015e-07, + "epoch": 0.9425981873111783 + }, + { + "current_steps": 624, + "loss": 3.0084, + "learning_rate": 7.232237539766702e-07, + "epoch": 0.9441087613293051 + }, + { + "current_steps": 625, + "loss": 2.874, + "learning_rate": 7.226935312831389e-07, + "epoch": 0.945619335347432 + }, + { + "current_steps": 626, + "loss": 3.1027, + "learning_rate": 7.221633085896076e-07, + "epoch": 0.947129909365559 + }, + { + "current_steps": 627, + "loss": 3.0754, + "learning_rate": 7.216330858960763e-07, + "epoch": 0.9486404833836858 + }, + { + "current_steps": 628, + "loss": 3.1411, + "learning_rate": 7.21102863202545e-07, + "epoch": 0.9501510574018127 + }, + { + "current_steps": 629, + "loss": 3.1479, + "learning_rate": 7.205726405090138e-07, + "epoch": 0.9516616314199395 + }, + { + "current_steps": 630, + "loss": 3.0268, + "learning_rate": 7.200424178154825e-07, + "epoch": 0.9531722054380665 + }, + { + "current_steps": 631, + "loss": 2.853, + "learning_rate": 7.195121951219512e-07, + "epoch": 0.9546827794561934 + }, + { + "current_steps": 632, + "loss": 2.8375, + "learning_rate": 7.1898197242842e-07, + "epoch": 0.9561933534743202 + }, + { + "current_steps": 633, + "loss": 3.0905, + "learning_rate": 7.184517497348886e-07, + "epoch": 0.9577039274924471 + }, + { + "current_steps": 634, + "loss": 2.9836, + "learning_rate": 7.179215270413573e-07, + "epoch": 0.959214501510574 + }, + { + "current_steps": 635, + "loss": 3.0679, + "learning_rate": 7.17391304347826e-07, + "epoch": 0.9607250755287009 + }, + { + "current_steps": 636, + "loss": 3.0425, + "learning_rate": 7.168610816542949e-07, + "epoch": 0.9622356495468278 + }, + { + "current_steps": 637, + "loss": 3.1853, + "learning_rate": 7.163308589607635e-07, + "epoch": 0.9637462235649547 + }, + { + "current_steps": 638, + "loss": 3.1647, + "learning_rate": 7.158006362672322e-07, + "epoch": 0.9652567975830816 + }, + { + "current_steps": 639, + "loss": 2.5096, + "learning_rate": 7.152704135737009e-07, + "epoch": 0.9667673716012085 + }, + { + "current_steps": 640, + "loss": 2.6806, + "learning_rate": 7.147401908801697e-07, + "epoch": 0.9682779456193353 + }, + { + "current_steps": 641, + "loss": 3.1098, + "learning_rate": 7.142099681866383e-07, + "epoch": 0.9697885196374623 + }, + { + "current_steps": 642, + "loss": 3.1581, + "learning_rate": 7.136797454931071e-07, + "epoch": 0.9712990936555891 + }, + { + "current_steps": 643, + "loss": 3.1607, + "learning_rate": 7.131495227995758e-07, + "epoch": 0.972809667673716 + }, + { + "current_steps": 644, + "loss": 2.9051, + "learning_rate": 7.126193001060445e-07, + "epoch": 0.974320241691843 + }, + { + "current_steps": 645, + "loss": 3.1324, + "learning_rate": 7.120890774125132e-07, + "epoch": 0.9758308157099698 + }, + { + "current_steps": 646, + "loss": 3.1401, + "learning_rate": 7.115588547189819e-07, + "epoch": 0.9773413897280967 + }, + { + "current_steps": 647, + "loss": 3.1494, + "learning_rate": 7.110286320254506e-07, + "epoch": 0.9788519637462235 + }, + { + "current_steps": 648, + "loss": 2.4261, + "learning_rate": 7.104984093319194e-07, + "epoch": 0.9803625377643505 + }, + { + "current_steps": 649, + "loss": 3.1947, + "learning_rate": 7.099681866383881e-07, + "epoch": 0.9818731117824774 + }, + { + "current_steps": 650, + "loss": 2.7172, + "learning_rate": 7.094379639448568e-07, + "epoch": 0.9833836858006042 + }, + { + "current_steps": 651, + "loss": 2.8514, + "learning_rate": 7.089077412513256e-07, + "epoch": 0.9848942598187311 + }, + { + "current_steps": 652, + "loss": 3.079, + "learning_rate": 7.083775185577942e-07, + "epoch": 0.986404833836858 + }, + { + "current_steps": 653, + "loss": 2.8849, + "learning_rate": 7.078472958642629e-07, + "epoch": 0.9879154078549849 + }, + { + "current_steps": 654, + "loss": 3.0467, + "learning_rate": 7.073170731707316e-07, + "epoch": 0.9894259818731118 + }, + { + "current_steps": 655, + "loss": 3.3841, + "learning_rate": 7.067868504772005e-07, + "epoch": 0.9909365558912386 + }, + { + "current_steps": 656, + "loss": 3.1446, + "learning_rate": 7.062566277836691e-07, + "epoch": 0.9924471299093656 + }, + { + "current_steps": 657, + "loss": 3.1988, + "learning_rate": 7.057264050901378e-07, + "epoch": 0.9939577039274925 + }, + { + "current_steps": 658, + "loss": 3.4408, + "learning_rate": 7.051961823966065e-07, + "epoch": 0.9954682779456193 + }, + { + "current_steps": 659, + "loss": 3.0199, + "learning_rate": 7.046659597030753e-07, + "epoch": 0.9969788519637462 + }, + { + "current_steps": 660, + "loss": 3.2084, + "learning_rate": 7.041357370095439e-07, + "epoch": 0.9984894259818731 + }, + { + "current_steps": 661, + "loss": 3.165, + "learning_rate": 7.036055143160127e-07, + "epoch": 1.0 + }, + { + "current_steps": 662, + "loss": 3.137, + "learning_rate": 7.030752916224814e-07, + "epoch": 1.001510574018127 + }, + { + "current_steps": 663, + "loss": 3.2575, + "learning_rate": 7.025450689289502e-07, + "epoch": 1.0030211480362539 + }, + { + "current_steps": 664, + "loss": 3.0359, + "learning_rate": 7.020148462354188e-07, + "epoch": 1.0045317220543806 + }, + { + "current_steps": 665, + "loss": 3.0096, + "learning_rate": 7.014846235418875e-07, + "epoch": 1.0060422960725075 + }, + { + "current_steps": 666, + "loss": 2.9221, + "learning_rate": 7.009544008483563e-07, + "epoch": 1.0075528700906344 + }, + { + "current_steps": 667, + "loss": 2.769, + "learning_rate": 7.004241781548251e-07, + "epoch": 1.0090634441087614 + }, + { + "current_steps": 668, + "loss": 3.2482, + "learning_rate": 6.998939554612937e-07, + "epoch": 1.0105740181268883 + }, + { + "current_steps": 669, + "loss": 2.9021, + "learning_rate": 6.993637327677624e-07, + "epoch": 1.012084592145015 + }, + { + "current_steps": 670, + "loss": 2.5269, + "learning_rate": 6.988335100742312e-07, + "epoch": 1.013595166163142 + }, + { + "current_steps": 671, + "loss": 3.1131, + "learning_rate": 6.983032873806999e-07, + "epoch": 1.0151057401812689 + }, + { + "current_steps": 672, + "loss": 3.0572, + "learning_rate": 6.977730646871685e-07, + "epoch": 1.0166163141993958 + }, + { + "current_steps": 673, + "loss": 3.036, + "learning_rate": 6.972428419936372e-07, + "epoch": 1.0181268882175227 + }, + { + "current_steps": 674, + "loss": 3.1856, + "learning_rate": 6.967126193001061e-07, + "epoch": 1.0196374622356494 + }, + { + "current_steps": 675, + "loss": 3.2084, + "learning_rate": 6.961823966065748e-07, + "epoch": 1.0211480362537764 + }, + { + "current_steps": 676, + "loss": 3.1684, + "learning_rate": 6.956521739130434e-07, + "epoch": 1.0226586102719033 + }, + { + "current_steps": 677, + "loss": 2.9708, + "learning_rate": 6.951219512195121e-07, + "epoch": 1.0241691842900302 + }, + { + "current_steps": 678, + "loss": 2.6613, + "learning_rate": 6.945917285259809e-07, + "epoch": 1.0256797583081572 + }, + { + "current_steps": 679, + "loss": 2.9646, + "learning_rate": 6.940615058324496e-07, + "epoch": 1.027190332326284 + }, + { + "current_steps": 680, + "loss": 3.1718, + "learning_rate": 6.935312831389183e-07, + "epoch": 1.0287009063444108 + }, + { + "current_steps": 681, + "loss": 2.6423, + "learning_rate": 6.93001060445387e-07, + "epoch": 1.0302114803625377 + }, + { + "current_steps": 682, + "loss": 3.0778, + "learning_rate": 6.924708377518558e-07, + "epoch": 1.0317220543806647 + }, + { + "current_steps": 683, + "loss": 3.0156, + "learning_rate": 6.919406150583245e-07, + "epoch": 1.0332326283987916 + }, + { + "current_steps": 684, + "loss": 2.9816, + "learning_rate": 6.914103923647931e-07, + "epoch": 1.0347432024169185 + }, + { + "current_steps": 685, + "loss": 2.938, + "learning_rate": 6.908801696712619e-07, + "epoch": 1.0362537764350452 + }, + { + "current_steps": 686, + "loss": 3.1688, + "learning_rate": 6.903499469777307e-07, + "epoch": 1.0377643504531722 + }, + { + "current_steps": 687, + "loss": 2.8528, + "learning_rate": 6.898197242841993e-07, + "epoch": 1.039274924471299 + }, + { + "current_steps": 688, + "loss": 3.1124, + "learning_rate": 6.89289501590668e-07, + "epoch": 1.040785498489426 + }, + { + "current_steps": 689, + "loss": 2.5208, + "learning_rate": 6.887592788971368e-07, + "epoch": 1.042296072507553 + }, + { + "current_steps": 690, + "loss": 3.2518, + "learning_rate": 6.882290562036055e-07, + "epoch": 1.0438066465256797 + }, + { + "current_steps": 691, + "loss": 3.0917, + "learning_rate": 6.876988335100741e-07, + "epoch": 1.0453172205438066 + }, + { + "current_steps": 692, + "loss": 3.0708, + "learning_rate": 6.871686108165429e-07, + "epoch": 1.0468277945619335 + }, + { + "current_steps": 693, + "loss": 2.637, + "learning_rate": 6.866383881230117e-07, + "epoch": 1.0483383685800605 + }, + { + "current_steps": 694, + "loss": 3.0982, + "learning_rate": 6.861081654294804e-07, + "epoch": 1.0498489425981874 + }, + { + "current_steps": 695, + "loss": 3.2218, + "learning_rate": 6.85577942735949e-07, + "epoch": 1.051359516616314 + }, + { + "current_steps": 696, + "loss": 3.1743, + "learning_rate": 6.850477200424177e-07, + "epoch": 1.052870090634441 + }, + { + "current_steps": 697, + "loss": 3.0273, + "learning_rate": 6.845174973488865e-07, + "epoch": 1.054380664652568 + }, + { + "current_steps": 698, + "loss": 2.8772, + "learning_rate": 6.839872746553553e-07, + "epoch": 1.055891238670695 + }, + { + "current_steps": 699, + "loss": 3.18, + "learning_rate": 6.834570519618239e-07, + "epoch": 1.0574018126888218 + }, + { + "current_steps": 700, + "loss": 3.0082, + "learning_rate": 6.829268292682927e-07, + "epoch": 1.0589123867069485 + }, + { + "current_steps": 701, + "loss": 3.1256, + "learning_rate": 6.823966065747614e-07, + "epoch": 1.0604229607250755 + }, + { + "current_steps": 702, + "loss": 2.3957, + "learning_rate": 6.818663838812301e-07, + "epoch": 1.0619335347432024 + }, + { + "current_steps": 703, + "loss": 3.0132, + "learning_rate": 6.813361611876987e-07, + "epoch": 1.0634441087613293 + }, + { + "current_steps": 704, + "loss": 3.1058, + "learning_rate": 6.808059384941675e-07, + "epoch": 1.0649546827794563 + }, + { + "current_steps": 705, + "loss": 2.9103, + "learning_rate": 6.802757158006363e-07, + "epoch": 1.0664652567975832 + }, + { + "current_steps": 706, + "loss": 2.504, + "learning_rate": 6.79745493107105e-07, + "epoch": 1.06797583081571 + }, + { + "current_steps": 707, + "loss": 2.9153, + "learning_rate": 6.792152704135736e-07, + "epoch": 1.0694864048338368 + }, + { + "current_steps": 708, + "loss": 3.2544, + "learning_rate": 6.786850477200424e-07, + "epoch": 1.0709969788519638 + }, + { + "current_steps": 709, + "loss": 3.0341, + "learning_rate": 6.781548250265111e-07, + "epoch": 1.0725075528700907 + }, + { + "current_steps": 710, + "loss": 3.0629, + "learning_rate": 6.776246023329798e-07, + "epoch": 1.0740181268882174 + }, + { + "current_steps": 711, + "loss": 2.8091, + "learning_rate": 6.770943796394485e-07, + "epoch": 1.0755287009063443 + }, + { + "current_steps": 712, + "loss": 3.0339, + "learning_rate": 6.765641569459173e-07, + "epoch": 1.0770392749244713 + }, + { + "current_steps": 713, + "loss": 3.0739, + "learning_rate": 6.76033934252386e-07, + "epoch": 1.0785498489425982 + }, + { + "current_steps": 714, + "loss": 3.2144, + "learning_rate": 6.755037115588547e-07, + "epoch": 1.0800604229607251 + }, + { + "current_steps": 715, + "loss": 3.1725, + "learning_rate": 6.749734888653234e-07, + "epoch": 1.081570996978852 + }, + { + "current_steps": 716, + "loss": 3.2246, + "learning_rate": 6.744432661717921e-07, + "epoch": 1.0830815709969788 + }, + { + "current_steps": 717, + "loss": 2.9513, + "learning_rate": 6.739130434782609e-07, + "epoch": 1.0845921450151057 + }, + { + "current_steps": 718, + "loss": 3.1187, + "learning_rate": 6.733828207847296e-07, + "epoch": 1.0861027190332326 + }, + { + "current_steps": 719, + "loss": 2.8206, + "learning_rate": 6.728525980911983e-07, + "epoch": 1.0876132930513596 + }, + { + "current_steps": 720, + "loss": 3.1289, + "learning_rate": 6.72322375397667e-07, + "epoch": 1.0891238670694865 + }, + { + "current_steps": 721, + "loss": 3.0412, + "learning_rate": 6.717921527041357e-07, + "epoch": 1.0906344410876132 + }, + { + "current_steps": 722, + "loss": 3.0754, + "learning_rate": 6.712619300106044e-07, + "epoch": 1.0921450151057401 + }, + { + "current_steps": 723, + "loss": 2.9696, + "learning_rate": 6.707317073170731e-07, + "epoch": 1.093655589123867 + }, + { + "current_steps": 724, + "loss": 3.1556, + "learning_rate": 6.702014846235419e-07, + "epoch": 1.095166163141994 + }, + { + "current_steps": 725, + "loss": 3.1364, + "learning_rate": 6.696712619300106e-07, + "epoch": 1.096676737160121 + }, + { + "current_steps": 726, + "loss": 2.9792, + "learning_rate": 6.691410392364793e-07, + "epoch": 1.0981873111782476 + }, + { + "current_steps": 727, + "loss": 2.7867, + "learning_rate": 6.68610816542948e-07, + "epoch": 1.0996978851963746 + }, + { + "current_steps": 728, + "loss": 3.1406, + "learning_rate": 6.680805938494167e-07, + "epoch": 1.1012084592145015 + }, + { + "current_steps": 729, + "loss": 2.8349, + "learning_rate": 6.675503711558854e-07, + "epoch": 1.1027190332326284 + }, + { + "current_steps": 730, + "loss": 2.939, + "learning_rate": 6.670201484623541e-07, + "epoch": 1.1042296072507554 + }, + { + "current_steps": 731, + "loss": 3.0505, + "learning_rate": 6.664899257688229e-07, + "epoch": 1.105740181268882 + }, + { + "current_steps": 732, + "loss": 2.6417, + "learning_rate": 6.659597030752916e-07, + "epoch": 1.107250755287009 + }, + { + "current_steps": 733, + "loss": 3.1132, + "learning_rate": 6.654294803817603e-07, + "epoch": 1.108761329305136 + }, + { + "current_steps": 734, + "loss": 3.0482, + "learning_rate": 6.648992576882291e-07, + "epoch": 1.1102719033232629 + }, + { + "current_steps": 735, + "loss": 3.0637, + "learning_rate": 6.643690349946977e-07, + "epoch": 1.1117824773413898 + }, + { + "current_steps": 736, + "loss": 2.9517, + "learning_rate": 6.638388123011665e-07, + "epoch": 1.1132930513595167 + }, + { + "current_steps": 737, + "loss": 3.083, + "learning_rate": 6.633085896076352e-07, + "epoch": 1.1148036253776434 + }, + { + "current_steps": 738, + "loss": 3.0229, + "learning_rate": 6.627783669141039e-07, + "epoch": 1.1163141993957704 + }, + { + "current_steps": 739, + "loss": 3.4434, + "learning_rate": 6.622481442205726e-07, + "epoch": 1.1178247734138973 + }, + { + "current_steps": 740, + "loss": 3.03, + "learning_rate": 6.617179215270413e-07, + "epoch": 1.1193353474320242 + }, + { + "current_steps": 741, + "loss": 2.8607, + "learning_rate": 6.6118769883351e-07, + "epoch": 1.1208459214501512 + }, + { + "current_steps": 742, + "loss": 3.2391, + "learning_rate": 6.606574761399787e-07, + "epoch": 1.1223564954682779 + }, + { + "current_steps": 743, + "loss": 3.097, + "learning_rate": 6.601272534464475e-07, + "epoch": 1.1238670694864048 + }, + { + "current_steps": 744, + "loss": 3.1365, + "learning_rate": 6.595970307529162e-07, + "epoch": 1.1253776435045317 + }, + { + "current_steps": 745, + "loss": 3.1356, + "learning_rate": 6.590668080593849e-07, + "epoch": 1.1268882175226587 + }, + { + "current_steps": 746, + "loss": 3.0501, + "learning_rate": 6.585365853658536e-07, + "epoch": 1.1283987915407856 + }, + { + "current_steps": 747, + "loss": 3.2082, + "learning_rate": 6.580063626723223e-07, + "epoch": 1.1299093655589123 + }, + { + "current_steps": 748, + "loss": 3.0707, + "learning_rate": 6.57476139978791e-07, + "epoch": 1.1314199395770392 + }, + { + "current_steps": 749, + "loss": 3.1677, + "learning_rate": 6.569459172852599e-07, + "epoch": 1.1329305135951662 + }, + { + "current_steps": 750, + "loss": 2.8567, + "learning_rate": 6.564156945917285e-07, + "epoch": 1.134441087613293 + }, + { + "current_steps": 751, + "loss": 2.6226, + "learning_rate": 6.558854718981972e-07, + "epoch": 1.13595166163142 + }, + { + "current_steps": 752, + "loss": 3.0405, + "learning_rate": 6.553552492046659e-07, + "epoch": 1.1374622356495467 + }, + { + "current_steps": 753, + "loss": 3.2305, + "learning_rate": 6.548250265111347e-07, + "epoch": 1.1389728096676737 + }, + { + "current_steps": 754, + "loss": 3.1328, + "learning_rate": 6.542948038176033e-07, + "epoch": 1.1404833836858006 + }, + { + "current_steps": 755, + "loss": 3.0304, + "learning_rate": 6.537645811240721e-07, + "epoch": 1.1419939577039275 + }, + { + "current_steps": 756, + "loss": 3.2476, + "learning_rate": 6.532343584305408e-07, + "epoch": 1.1435045317220545 + }, + { + "current_steps": 757, + "loss": 2.9137, + "learning_rate": 6.527041357370096e-07, + "epoch": 1.1450151057401814 + }, + { + "current_steps": 758, + "loss": 3.0561, + "learning_rate": 6.521739130434782e-07, + "epoch": 1.146525679758308 + }, + { + "current_steps": 759, + "loss": 2.9514, + "learning_rate": 6.516436903499469e-07, + "epoch": 1.148036253776435 + }, + { + "current_steps": 760, + "loss": 3.2839, + "learning_rate": 6.511134676564156e-07, + "epoch": 1.149546827794562 + }, + { + "current_steps": 761, + "loss": 2.473, + "learning_rate": 6.505832449628845e-07, + "epoch": 1.151057401812689 + }, + { + "current_steps": 762, + "loss": 2.9458, + "learning_rate": 6.500530222693531e-07, + "epoch": 1.1525679758308156 + }, + { + "current_steps": 763, + "loss": 3.2217, + "learning_rate": 6.495227995758218e-07, + "epoch": 1.1540785498489425 + }, + { + "current_steps": 764, + "loss": 2.8589, + "learning_rate": 6.489925768822906e-07, + "epoch": 1.1555891238670695 + }, + { + "current_steps": 765, + "loss": 2.8266, + "learning_rate": 6.484623541887593e-07, + "epoch": 1.1570996978851964 + }, + { + "current_steps": 766, + "loss": 3.1491, + "learning_rate": 6.479321314952279e-07, + "epoch": 1.1586102719033233 + }, + { + "current_steps": 767, + "loss": 2.9012, + "learning_rate": 6.474019088016966e-07, + "epoch": 1.1601208459214503 + }, + { + "current_steps": 768, + "loss": 2.8355, + "learning_rate": 6.468716861081655e-07, + "epoch": 1.161631419939577 + }, + { + "current_steps": 769, + "loss": 3.0604, + "learning_rate": 6.463414634146342e-07, + "epoch": 1.163141993957704 + }, + { + "current_steps": 770, + "loss": 2.8673, + "learning_rate": 6.458112407211028e-07, + "epoch": 1.1646525679758308 + }, + { + "current_steps": 771, + "loss": 3.0344, + "learning_rate": 6.452810180275715e-07, + "epoch": 1.1661631419939578 + }, + { + "current_steps": 772, + "loss": 2.7393, + "learning_rate": 6.447507953340403e-07, + "epoch": 1.1676737160120845 + }, + { + "current_steps": 773, + "loss": 2.9999, + "learning_rate": 6.44220572640509e-07, + "epoch": 1.1691842900302114 + }, + { + "current_steps": 774, + "loss": 2.8167, + "learning_rate": 6.436903499469777e-07, + "epoch": 1.1706948640483383 + }, + { + "current_steps": 775, + "loss": 3.1691, + "learning_rate": 6.431601272534464e-07, + "epoch": 1.1722054380664653 + }, + { + "current_steps": 776, + "loss": 3.1307, + "learning_rate": 6.426299045599152e-07, + "epoch": 1.1737160120845922 + }, + { + "current_steps": 777, + "loss": 3.053, + "learning_rate": 6.420996818663838e-07, + "epoch": 1.1752265861027191 + }, + { + "current_steps": 778, + "loss": 2.8401, + "learning_rate": 6.415694591728525e-07, + "epoch": 1.1767371601208458 + }, + { + "current_steps": 779, + "loss": 3.0846, + "learning_rate": 6.410392364793212e-07, + "epoch": 1.1782477341389728 + }, + { + "current_steps": 780, + "loss": 2.9075, + "learning_rate": 6.405090137857901e-07, + "epoch": 1.1797583081570997 + }, + { + "current_steps": 781, + "loss": 3.008, + "learning_rate": 6.399787910922587e-07, + "epoch": 1.1812688821752266 + }, + { + "current_steps": 782, + "loss": 2.9316, + "learning_rate": 6.394485683987274e-07, + "epoch": 1.1827794561933536 + }, + { + "current_steps": 783, + "loss": 2.9682, + "learning_rate": 6.389183457051962e-07, + "epoch": 1.1842900302114803 + }, + { + "current_steps": 784, + "loss": 3.1512, + "learning_rate": 6.383881230116649e-07, + "epoch": 1.1858006042296072 + }, + { + "current_steps": 785, + "loss": 2.3582, + "learning_rate": 6.378579003181335e-07, + "epoch": 1.1873111782477341 + }, + { + "current_steps": 786, + "loss": 3.1125, + "learning_rate": 6.373276776246022e-07, + "epoch": 1.188821752265861 + }, + { + "current_steps": 787, + "loss": 3.2427, + "learning_rate": 6.367974549310711e-07, + "epoch": 1.190332326283988 + }, + { + "current_steps": 788, + "loss": 2.8622, + "learning_rate": 6.362672322375398e-07, + "epoch": 1.191842900302115 + }, + { + "current_steps": 789, + "loss": 2.6609, + "learning_rate": 6.357370095440084e-07, + "epoch": 1.1933534743202416 + }, + { + "current_steps": 790, + "loss": 3.0324, + "learning_rate": 6.352067868504771e-07, + "epoch": 1.1948640483383686 + }, + { + "current_steps": 791, + "loss": 2.9873, + "learning_rate": 6.346765641569459e-07, + "epoch": 1.1963746223564955 + }, + { + "current_steps": 792, + "loss": 3.0322, + "learning_rate": 6.341463414634146e-07, + "epoch": 1.1978851963746224 + }, + { + "current_steps": 793, + "loss": 2.9698, + "learning_rate": 6.336161187698833e-07, + "epoch": 1.1993957703927491 + }, + { + "current_steps": 794, + "loss": 2.9074, + "learning_rate": 6.33085896076352e-07, + "epoch": 1.200906344410876 + }, + { + "current_steps": 795, + "loss": 2.7794, + "learning_rate": 6.325556733828208e-07, + "epoch": 1.202416918429003 + }, + { + "current_steps": 796, + "loss": 2.8889, + "learning_rate": 6.320254506892895e-07, + "epoch": 1.20392749244713 + }, + { + "current_steps": 797, + "loss": 2.9311, + "learning_rate": 6.314952279957581e-07, + "epoch": 1.2054380664652569 + }, + { + "current_steps": 798, + "loss": 2.9702, + "learning_rate": 6.309650053022269e-07, + "epoch": 1.2069486404833838 + }, + { + "current_steps": 799, + "loss": 2.0709, + "learning_rate": 6.304347826086957e-07, + "epoch": 1.2084592145015105 + }, + { + "current_steps": 800, + "loss": 3.211, + "learning_rate": 6.299045599151644e-07, + "epoch": 1.2099697885196374 + }, + { + "current_steps": 801, + "loss": 3.3029, + "learning_rate": 6.29374337221633e-07, + "epoch": 1.2114803625377644 + }, + { + "current_steps": 802, + "loss": 2.9971, + "learning_rate": 6.288441145281018e-07, + "epoch": 1.2129909365558913 + }, + { + "current_steps": 803, + "loss": 2.7264, + "learning_rate": 6.283138918345705e-07, + "epoch": 1.214501510574018 + }, + { + "current_steps": 804, + "loss": 2.8374, + "learning_rate": 6.277836691410392e-07, + "epoch": 1.216012084592145 + }, + { + "current_steps": 805, + "loss": 3.144, + "learning_rate": 6.272534464475078e-07, + "epoch": 1.2175226586102719 + }, + { + "current_steps": 806, + "loss": 3.0111, + "learning_rate": 6.267232237539767e-07, + "epoch": 1.2190332326283988 + }, + { + "current_steps": 807, + "loss": 2.9968, + "learning_rate": 6.261930010604454e-07, + "epoch": 1.2205438066465257 + }, + { + "current_steps": 808, + "loss": 3.141, + "learning_rate": 6.256627783669141e-07, + "epoch": 1.2220543806646527 + }, + { + "current_steps": 809, + "loss": 3.0366, + "learning_rate": 6.251325556733827e-07, + "epoch": 1.2235649546827794 + }, + { + "current_steps": 810, + "loss": 3.059, + "learning_rate": 6.246023329798515e-07, + "epoch": 1.2250755287009063 + }, + { + "current_steps": 811, + "loss": 2.8797, + "learning_rate": 6.240721102863203e-07, + "epoch": 1.2265861027190332 + }, + { + "current_steps": 812, + "loss": 2.5681, + "learning_rate": 6.23541887592789e-07, + "epoch": 1.2280966767371602 + }, + { + "current_steps": 813, + "loss": 3.088, + "learning_rate": 6.230116648992576e-07, + "epoch": 1.229607250755287 + }, + { + "current_steps": 814, + "loss": 2.4583, + "learning_rate": 6.224814422057264e-07, + "epoch": 1.2311178247734138 + }, + { + "current_steps": 815, + "loss": 2.9792, + "learning_rate": 6.219512195121951e-07, + "epoch": 1.2326283987915407 + }, + { + "current_steps": 816, + "loss": 3.0863, + "learning_rate": 6.214209968186638e-07, + "epoch": 1.2341389728096677 + }, + { + "current_steps": 817, + "loss": 3.0954, + "learning_rate": 6.208907741251325e-07, + "epoch": 1.2356495468277946 + }, + { + "current_steps": 818, + "loss": 3.2413, + "learning_rate": 6.203605514316013e-07, + "epoch": 1.2371601208459215 + }, + { + "current_steps": 819, + "loss": 3.0499, + "learning_rate": 6.1983032873807e-07, + "epoch": 1.2386706948640485 + }, + { + "current_steps": 820, + "loss": 3.1518, + "learning_rate": 6.193001060445386e-07, + "epoch": 1.2401812688821752 + }, + { + "current_steps": 821, + "loss": 3.2022, + "learning_rate": 6.187698833510074e-07, + "epoch": 1.241691842900302 + }, + { + "current_steps": 822, + "loss": 3.0871, + "learning_rate": 6.182396606574761e-07, + "epoch": 1.243202416918429 + }, + { + "current_steps": 823, + "loss": 2.957, + "learning_rate": 6.177094379639448e-07, + "epoch": 1.244712990936556 + }, + { + "current_steps": 824, + "loss": 3.1851, + "learning_rate": 6.171792152704135e-07, + "epoch": 1.2462235649546827 + }, + { + "current_steps": 825, + "loss": 2.6717, + "learning_rate": 6.166489925768823e-07, + "epoch": 1.2477341389728096 + }, + { + "current_steps": 826, + "loss": 3.0164, + "learning_rate": 6.16118769883351e-07, + "epoch": 1.2492447129909365 + }, + { + "current_steps": 827, + "loss": 2.977, + "learning_rate": 6.155885471898197e-07, + "epoch": 1.2507552870090635 + }, + { + "current_steps": 828, + "loss": 2.9581, + "learning_rate": 6.150583244962883e-07, + "epoch": 1.2522658610271904 + }, + { + "current_steps": 829, + "loss": 3.2141, + "learning_rate": 6.145281018027571e-07, + "epoch": 1.2537764350453173 + }, + { + "current_steps": 830, + "loss": 3.0109, + "learning_rate": 6.139978791092259e-07, + "epoch": 1.255287009063444 + }, + { + "current_steps": 831, + "loss": 3.0534, + "learning_rate": 6.134676564156946e-07, + "epoch": 1.256797583081571 + }, + { + "current_steps": 832, + "loss": 3.0423, + "learning_rate": 6.129374337221633e-07, + "epoch": 1.258308157099698 + }, + { + "current_steps": 833, + "loss": 3.0218, + "learning_rate": 6.12407211028632e-07, + "epoch": 1.2598187311178248 + }, + { + "current_steps": 834, + "loss": 2.9106, + "learning_rate": 6.118769883351007e-07, + "epoch": 1.2613293051359515 + }, + { + "current_steps": 835, + "loss": 3.1955, + "learning_rate": 6.113467656415694e-07, + "epoch": 1.2628398791540785 + }, + { + "current_steps": 836, + "loss": 3.0351, + "learning_rate": 6.108165429480381e-07, + "epoch": 1.2643504531722054 + }, + { + "current_steps": 837, + "loss": 3.0073, + "learning_rate": 6.102863202545069e-07, + "epoch": 1.2658610271903323 + }, + { + "current_steps": 838, + "loss": 2.7255, + "learning_rate": 6.097560975609756e-07, + "epoch": 1.2673716012084593 + }, + { + "current_steps": 839, + "loss": 2.8498, + "learning_rate": 6.092258748674443e-07, + "epoch": 1.2688821752265862 + }, + { + "current_steps": 840, + "loss": 3.0627, + "learning_rate": 6.08695652173913e-07, + "epoch": 1.2703927492447131 + }, + { + "current_steps": 841, + "loss": 3.053, + "learning_rate": 6.081654294803817e-07, + "epoch": 1.2719033232628398 + }, + { + "current_steps": 842, + "loss": 2.9029, + "learning_rate": 6.076352067868504e-07, + "epoch": 1.2734138972809668 + }, + { + "current_steps": 843, + "loss": 3.1912, + "learning_rate": 6.071049840933192e-07, + "epoch": 1.2749244712990937 + }, + { + "current_steps": 844, + "loss": 2.9187, + "learning_rate": 6.065747613997879e-07, + "epoch": 1.2764350453172204 + }, + { + "current_steps": 845, + "loss": 2.9065, + "learning_rate": 6.060445387062566e-07, + "epoch": 1.2779456193353473 + }, + { + "current_steps": 846, + "loss": 3.0209, + "learning_rate": 6.055143160127253e-07, + "epoch": 1.2794561933534743 + }, + { + "current_steps": 847, + "loss": 3.0331, + "learning_rate": 6.049840933191941e-07, + "epoch": 1.2809667673716012 + }, + { + "current_steps": 848, + "loss": 3.1613, + "learning_rate": 6.044538706256627e-07, + "epoch": 1.2824773413897281 + }, + { + "current_steps": 849, + "loss": 2.6895, + "learning_rate": 6.039236479321315e-07, + "epoch": 1.283987915407855 + }, + { + "current_steps": 850, + "loss": 3.0909, + "learning_rate": 6.033934252386002e-07, + "epoch": 1.285498489425982 + }, + { + "current_steps": 851, + "loss": 3.0953, + "learning_rate": 6.02863202545069e-07, + "epoch": 1.2870090634441087 + }, + { + "current_steps": 852, + "loss": 3.0476, + "learning_rate": 6.023329798515376e-07, + "epoch": 1.2885196374622356 + }, + { + "current_steps": 853, + "loss": 2.8508, + "learning_rate": 6.018027571580063e-07, + "epoch": 1.2900302114803626 + }, + { + "current_steps": 854, + "loss": 3.4764, + "learning_rate": 6.01272534464475e-07, + "epoch": 1.2915407854984895 + }, + { + "current_steps": 855, + "loss": 3.099, + "learning_rate": 6.007423117709439e-07, + "epoch": 1.2930513595166162 + }, + { + "current_steps": 856, + "loss": 3.069, + "learning_rate": 6.002120890774125e-07, + "epoch": 1.2945619335347431 + }, + { + "current_steps": 857, + "loss": 2.58, + "learning_rate": 5.996818663838812e-07, + "epoch": 1.29607250755287 + }, + { + "current_steps": 858, + "loss": 2.3271, + "learning_rate": 5.991516436903499e-07, + "epoch": 1.297583081570997 + }, + { + "current_steps": 859, + "loss": 2.9752, + "learning_rate": 5.986214209968187e-07, + "epoch": 1.299093655589124 + }, + { + "current_steps": 860, + "loss": 3.0935, + "learning_rate": 5.980911983032873e-07, + "epoch": 1.3006042296072509 + }, + { + "current_steps": 861, + "loss": 3.0636, + "learning_rate": 5.97560975609756e-07, + "epoch": 1.3021148036253776 + }, + { + "current_steps": 862, + "loss": 3.2768, + "learning_rate": 5.970307529162248e-07, + "epoch": 1.3036253776435045 + }, + { + "current_steps": 863, + "loss": 2.9215, + "learning_rate": 5.965005302226936e-07, + "epoch": 1.3051359516616314 + }, + { + "current_steps": 864, + "loss": 2.7811, + "learning_rate": 5.959703075291622e-07, + "epoch": 1.3066465256797584 + }, + { + "current_steps": 865, + "loss": 2.9675, + "learning_rate": 5.954400848356309e-07, + "epoch": 1.308157099697885 + }, + { + "current_steps": 866, + "loss": 2.8599, + "learning_rate": 5.949098621420997e-07, + "epoch": 1.309667673716012 + }, + { + "current_steps": 867, + "loss": 3.0658, + "learning_rate": 5.943796394485684e-07, + "epoch": 1.311178247734139 + }, + { + "current_steps": 868, + "loss": 2.9285, + "learning_rate": 5.938494167550371e-07, + "epoch": 1.3126888217522659 + }, + { + "current_steps": 869, + "loss": 2.8511, + "learning_rate": 5.933191940615058e-07, + "epoch": 1.3141993957703928 + }, + { + "current_steps": 870, + "loss": 3.236, + "learning_rate": 5.927889713679746e-07, + "epoch": 1.3157099697885197 + }, + { + "current_steps": 871, + "loss": 2.8618, + "learning_rate": 5.922587486744432e-07, + "epoch": 1.3172205438066467 + }, + { + "current_steps": 872, + "loss": 2.9207, + "learning_rate": 5.917285259809119e-07, + "epoch": 1.3187311178247734 + }, + { + "current_steps": 873, + "loss": 3.115, + "learning_rate": 5.911983032873806e-07, + "epoch": 1.3202416918429003 + }, + { + "current_steps": 874, + "loss": 3.0432, + "learning_rate": 5.906680805938495e-07, + "epoch": 1.3217522658610272 + }, + { + "current_steps": 875, + "loss": 2.9716, + "learning_rate": 5.901378579003181e-07, + "epoch": 1.323262839879154 + }, + { + "current_steps": 876, + "loss": 3.0914, + "learning_rate": 5.896076352067868e-07, + "epoch": 1.3247734138972809 + }, + { + "current_steps": 877, + "loss": 3.002, + "learning_rate": 5.890774125132555e-07, + "epoch": 1.3262839879154078 + }, + { + "current_steps": 878, + "loss": 2.9328, + "learning_rate": 5.885471898197243e-07, + "epoch": 1.3277945619335347 + }, + { + "current_steps": 879, + "loss": 3.1159, + "learning_rate": 5.880169671261929e-07, + "epoch": 1.3293051359516617 + }, + { + "current_steps": 880, + "loss": 2.632, + "learning_rate": 5.874867444326616e-07, + "epoch": 1.3308157099697886 + }, + { + "current_steps": 881, + "loss": 3.2234, + "learning_rate": 5.869565217391305e-07, + "epoch": 1.3323262839879155 + }, + { + "current_steps": 882, + "loss": 2.7351, + "learning_rate": 5.864262990455992e-07, + "epoch": 1.3338368580060422 + }, + { + "current_steps": 883, + "loss": 2.9956, + "learning_rate": 5.858960763520678e-07, + "epoch": 1.3353474320241692 + }, + { + "current_steps": 884, + "loss": 2.8181, + "learning_rate": 5.853658536585365e-07, + "epoch": 1.336858006042296 + }, + { + "current_steps": 885, + "loss": 3.0941, + "learning_rate": 5.848356309650053e-07, + "epoch": 1.338368580060423 + }, + { + "current_steps": 886, + "loss": 3.093, + "learning_rate": 5.84305408271474e-07, + "epoch": 1.3398791540785497 + }, + { + "current_steps": 887, + "loss": 3.1539, + "learning_rate": 5.837751855779427e-07, + "epoch": 1.3413897280966767 + }, + { + "current_steps": 888, + "loss": 2.798, + "learning_rate": 5.832449628844114e-07, + "epoch": 1.3429003021148036 + }, + { + "current_steps": 889, + "loss": 2.8593, + "learning_rate": 5.827147401908802e-07, + "epoch": 1.3444108761329305 + }, + { + "current_steps": 890, + "loss": 2.9718, + "learning_rate": 5.821845174973489e-07, + "epoch": 1.3459214501510575 + }, + { + "current_steps": 891, + "loss": 2.9314, + "learning_rate": 5.816542948038175e-07, + "epoch": 1.3474320241691844 + }, + { + "current_steps": 892, + "loss": 3.0187, + "learning_rate": 5.811240721102862e-07, + "epoch": 1.348942598187311 + }, + { + "current_steps": 893, + "loss": 3.0841, + "learning_rate": 5.805938494167551e-07, + "epoch": 1.350453172205438 + }, + { + "current_steps": 894, + "loss": 3.0654, + "learning_rate": 5.800636267232238e-07, + "epoch": 1.351963746223565 + }, + { + "current_steps": 895, + "loss": 3.0979, + "learning_rate": 5.795334040296924e-07, + "epoch": 1.353474320241692 + }, + { + "current_steps": 896, + "loss": 3.09, + "learning_rate": 5.790031813361611e-07, + "epoch": 1.3549848942598186 + }, + { + "current_steps": 897, + "loss": 2.9968, + "learning_rate": 5.784729586426299e-07, + "epoch": 1.3564954682779455 + }, + { + "current_steps": 898, + "loss": 2.4133, + "learning_rate": 5.779427359490986e-07, + "epoch": 1.3580060422960725 + }, + { + "current_steps": 899, + "loss": 3.1315, + "learning_rate": 5.774125132555672e-07, + "epoch": 1.3595166163141994 + }, + { + "current_steps": 900, + "loss": 3.0132, + "learning_rate": 5.768822905620361e-07, + "epoch": 1.3610271903323263 + }, + { + "current_steps": 901, + "loss": 2.8193, + "learning_rate": 5.763520678685048e-07, + "epoch": 1.3625377643504533 + }, + { + "current_steps": 902, + "loss": 3.0323, + "learning_rate": 5.758218451749735e-07, + "epoch": 1.3640483383685802 + }, + { + "current_steps": 903, + "loss": 2.8813, + "learning_rate": 5.752916224814421e-07, + "epoch": 1.365558912386707 + }, + { + "current_steps": 904, + "loss": 2.9499, + "learning_rate": 5.747613997879109e-07, + "epoch": 1.3670694864048338 + }, + { + "current_steps": 905, + "loss": 2.9082, + "learning_rate": 5.742311770943796e-07, + "epoch": 1.3685800604229608 + }, + { + "current_steps": 906, + "loss": 3.1331, + "learning_rate": 5.737009544008484e-07, + "epoch": 1.3700906344410877 + }, + { + "current_steps": 907, + "loss": 3.0406, + "learning_rate": 5.73170731707317e-07, + "epoch": 1.3716012084592144 + }, + { + "current_steps": 908, + "loss": 3.0628, + "learning_rate": 5.726405090137858e-07, + "epoch": 1.3731117824773413 + }, + { + "current_steps": 909, + "loss": 3.0879, + "learning_rate": 5.721102863202545e-07, + "epoch": 1.3746223564954683 + }, + { + "current_steps": 910, + "loss": 3.1388, + "learning_rate": 5.715800636267231e-07, + "epoch": 1.3761329305135952 + }, + { + "current_steps": 911, + "loss": 2.8779, + "learning_rate": 5.710498409331918e-07, + "epoch": 1.3776435045317221 + }, + { + "current_steps": 912, + "loss": 2.9396, + "learning_rate": 5.705196182396607e-07, + "epoch": 1.379154078549849 + }, + { + "current_steps": 913, + "loss": 3.0501, + "learning_rate": 5.699893955461294e-07, + "epoch": 1.3806646525679758 + }, + { + "current_steps": 914, + "loss": 3.077, + "learning_rate": 5.69459172852598e-07, + "epoch": 1.3821752265861027 + }, + { + "current_steps": 915, + "loss": 2.8821, + "learning_rate": 5.689289501590668e-07, + "epoch": 1.3836858006042296 + }, + { + "current_steps": 916, + "loss": 3.1957, + "learning_rate": 5.683987274655355e-07, + "epoch": 1.3851963746223566 + }, + { + "current_steps": 917, + "loss": 2.9614, + "learning_rate": 5.678685047720042e-07, + "epoch": 1.3867069486404833 + }, + { + "current_steps": 918, + "loss": 3.086, + "learning_rate": 5.673382820784728e-07, + "epoch": 1.3882175226586102 + }, + { + "current_steps": 919, + "loss": 2.2982, + "learning_rate": 5.668080593849417e-07, + "epoch": 1.3897280966767371 + }, + { + "current_steps": 920, + "loss": 3.0439, + "learning_rate": 5.662778366914104e-07, + "epoch": 1.391238670694864 + }, + { + "current_steps": 921, + "loss": 3.0418, + "learning_rate": 5.657476139978791e-07, + "epoch": 1.392749244712991 + }, + { + "current_steps": 922, + "loss": 2.9936, + "learning_rate": 5.652173913043477e-07, + "epoch": 1.394259818731118 + }, + { + "current_steps": 923, + "loss": 2.9119, + "learning_rate": 5.646871686108165e-07, + "epoch": 1.3957703927492446 + }, + { + "current_steps": 924, + "loss": 2.8972, + "learning_rate": 5.641569459172853e-07, + "epoch": 1.3972809667673716 + }, + { + "current_steps": 925, + "loss": 3.0952, + "learning_rate": 5.63626723223754e-07, + "epoch": 1.3987915407854985 + }, + { + "current_steps": 926, + "loss": 3.0398, + "learning_rate": 5.630965005302226e-07, + "epoch": 1.4003021148036254 + }, + { + "current_steps": 927, + "loss": 3.1679, + "learning_rate": 5.625662778366914e-07, + "epoch": 1.4018126888217521 + }, + { + "current_steps": 928, + "loss": 3.007, + "learning_rate": 5.620360551431601e-07, + "epoch": 1.403323262839879 + }, + { + "current_steps": 929, + "loss": 3.0324, + "learning_rate": 5.615058324496288e-07, + "epoch": 1.404833836858006 + }, + { + "current_steps": 930, + "loss": 3.2281, + "learning_rate": 5.609756097560975e-07, + "epoch": 1.406344410876133 + }, + { + "current_steps": 931, + "loss": 3.0227, + "learning_rate": 5.604453870625663e-07, + "epoch": 1.4078549848942599 + }, + { + "current_steps": 932, + "loss": 2.8318, + "learning_rate": 5.59915164369035e-07, + "epoch": 1.4093655589123868 + }, + { + "current_steps": 933, + "loss": 2.8935, + "learning_rate": 5.593849416755037e-07, + "epoch": 1.4108761329305137 + }, + { + "current_steps": 934, + "loss": 3.0167, + "learning_rate": 5.588547189819724e-07, + "epoch": 1.4123867069486404 + }, + { + "current_steps": 935, + "loss": 2.9921, + "learning_rate": 5.583244962884411e-07, + "epoch": 1.4138972809667674 + }, + { + "current_steps": 936, + "loss": 3.1317, + "learning_rate": 5.577942735949098e-07, + "epoch": 1.4154078549848943 + }, + { + "current_steps": 937, + "loss": 3.0075, + "learning_rate": 5.572640509013786e-07, + "epoch": 1.4169184290030212 + }, + { + "current_steps": 938, + "loss": 3.1235, + "learning_rate": 5.567338282078473e-07, + "epoch": 1.418429003021148 + }, + { + "current_steps": 939, + "loss": 2.6289, + "learning_rate": 5.56203605514316e-07, + "epoch": 1.4199395770392749 + }, + { + "current_steps": 940, + "loss": 2.9834, + "learning_rate": 5.556733828207847e-07, + "epoch": 1.4214501510574018 + }, + { + "current_steps": 941, + "loss": 3.1223, + "learning_rate": 5.551431601272534e-07, + "epoch": 1.4229607250755287 + }, + { + "current_steps": 942, + "loss": 3.1114, + "learning_rate": 5.546129374337221e-07, + "epoch": 1.4244712990936557 + }, + { + "current_steps": 943, + "loss": 2.9695, + "learning_rate": 5.540827147401909e-07, + "epoch": 1.4259818731117826 + }, + { + "current_steps": 944, + "loss": 2.5181, + "learning_rate": 5.535524920466596e-07, + "epoch": 1.4274924471299093 + }, + { + "current_steps": 945, + "loss": 3.0749, + "learning_rate": 5.530222693531283e-07, + "epoch": 1.4290030211480362 + }, + { + "current_steps": 946, + "loss": 2.8732, + "learning_rate": 5.52492046659597e-07, + "epoch": 1.4305135951661632 + }, + { + "current_steps": 947, + "loss": 3.0371, + "learning_rate": 5.519618239660657e-07, + "epoch": 1.43202416918429 + }, + { + "current_steps": 948, + "loss": 2.8945, + "learning_rate": 5.514316012725344e-07, + "epoch": 1.4335347432024168 + }, + { + "current_steps": 949, + "loss": 3.1701, + "learning_rate": 5.509013785790033e-07, + "epoch": 1.4350453172205437 + }, + { + "current_steps": 950, + "loss": 2.9467, + "learning_rate": 5.503711558854719e-07, + "epoch": 1.4365558912386707 + }, + { + "current_steps": 951, + "loss": 2.9566, + "learning_rate": 5.498409331919406e-07, + "epoch": 1.4380664652567976 + }, + { + "current_steps": 952, + "loss": 2.9671, + "learning_rate": 5.493107104984093e-07, + "epoch": 1.4395770392749245 + }, + { + "current_steps": 953, + "loss": 2.823, + "learning_rate": 5.487804878048781e-07, + "epoch": 1.4410876132930515 + }, + { + "current_steps": 954, + "loss": 2.5544, + "learning_rate": 5.482502651113467e-07, + "epoch": 1.4425981873111782 + }, + { + "current_steps": 955, + "loss": 2.8079, + "learning_rate": 5.477200424178154e-07, + "epoch": 1.444108761329305 + }, + { + "current_steps": 956, + "loss": 2.9745, + "learning_rate": 5.471898197242842e-07, + "epoch": 1.445619335347432 + }, + { + "current_steps": 957, + "loss": 2.9221, + "learning_rate": 5.46659597030753e-07, + "epoch": 1.447129909365559 + }, + { + "current_steps": 958, + "loss": 3.1318, + "learning_rate": 5.461293743372216e-07, + "epoch": 1.4486404833836857 + }, + { + "current_steps": 959, + "loss": 2.9277, + "learning_rate": 5.455991516436903e-07, + "epoch": 1.4501510574018126 + }, + { + "current_steps": 960, + "loss": 2.9046, + "learning_rate": 5.45068928950159e-07, + "epoch": 1.4516616314199395 + }, + { + "current_steps": 961, + "loss": 3.0094, + "learning_rate": 5.445387062566277e-07, + "epoch": 1.4531722054380665 + }, + { + "current_steps": 962, + "loss": 2.9582, + "learning_rate": 5.440084835630965e-07, + "epoch": 1.4546827794561934 + }, + { + "current_steps": 963, + "loss": 2.9235, + "learning_rate": 5.434782608695652e-07, + "epoch": 1.4561933534743203 + }, + { + "current_steps": 964, + "loss": 3.1794, + "learning_rate": 5.42948038176034e-07, + "epoch": 1.4577039274924473 + }, + { + "current_steps": 965, + "loss": 2.944, + "learning_rate": 5.424178154825026e-07, + "epoch": 1.459214501510574 + }, + { + "current_steps": 966, + "loss": 3.0503, + "learning_rate": 5.418875927889713e-07, + "epoch": 1.460725075528701 + }, + { + "current_steps": 967, + "loss": 2.824, + "learning_rate": 5.4135737009544e-07, + "epoch": 1.4622356495468278 + }, + { + "current_steps": 968, + "loss": 2.8722, + "learning_rate": 5.408271474019089e-07, + "epoch": 1.4637462235649548 + }, + { + "current_steps": 969, + "loss": 3.0602, + "learning_rate": 5.402969247083775e-07, + "epoch": 1.4652567975830815 + }, + { + "current_steps": 970, + "loss": 2.4757, + "learning_rate": 5.397667020148462e-07, + "epoch": 1.4667673716012084 + }, + { + "current_steps": 971, + "loss": 3.0635, + "learning_rate": 5.392364793213149e-07, + "epoch": 1.4682779456193353 + }, + { + "current_steps": 972, + "loss": 3.1136, + "learning_rate": 5.387062566277837e-07, + "epoch": 1.4697885196374623 + }, + { + "current_steps": 973, + "loss": 2.9535, + "learning_rate": 5.381760339342523e-07, + "epoch": 1.4712990936555892 + }, + { + "current_steps": 974, + "loss": 2.9776, + "learning_rate": 5.37645811240721e-07, + "epoch": 1.4728096676737161 + }, + { + "current_steps": 975, + "loss": 3.0826, + "learning_rate": 5.371155885471898e-07, + "epoch": 1.4743202416918428 + }, + { + "current_steps": 976, + "loss": 3.0639, + "learning_rate": 5.365853658536586e-07, + "epoch": 1.4758308157099698 + }, + { + "current_steps": 977, + "loss": 3.0708, + "learning_rate": 5.360551431601272e-07, + "epoch": 1.4773413897280967 + }, + { + "current_steps": 978, + "loss": 2.9469, + "learning_rate": 5.355249204665959e-07, + "epoch": 1.4788519637462236 + }, + { + "current_steps": 979, + "loss": 2.9673, + "learning_rate": 5.349946977730647e-07, + "epoch": 1.4803625377643503 + }, + { + "current_steps": 980, + "loss": 3.0806, + "learning_rate": 5.344644750795334e-07, + "epoch": 1.4818731117824773 + }, + { + "current_steps": 981, + "loss": 3.1313, + "learning_rate": 5.339342523860021e-07, + "epoch": 1.4833836858006042 + }, + { + "current_steps": 982, + "loss": 2.8451, + "learning_rate": 5.334040296924708e-07, + "epoch": 1.4848942598187311 + }, + { + "current_steps": 983, + "loss": 2.4288, + "learning_rate": 5.328738069989396e-07, + "epoch": 1.486404833836858 + }, + { + "current_steps": 984, + "loss": 3.0014, + "learning_rate": 5.323435843054083e-07, + "epoch": 1.487915407854985 + }, + { + "current_steps": 985, + "loss": 3.05, + "learning_rate": 5.318133616118769e-07, + "epoch": 1.4894259818731117 + }, + { + "current_steps": 986, + "loss": 3.0385, + "learning_rate": 5.312831389183456e-07, + "epoch": 1.4909365558912386 + }, + { + "current_steps": 987, + "loss": 3.0729, + "learning_rate": 5.307529162248145e-07, + "epoch": 1.4924471299093656 + }, + { + "current_steps": 988, + "loss": 2.8984, + "learning_rate": 5.302226935312832e-07, + "epoch": 1.4939577039274925 + }, + { + "current_steps": 989, + "loss": 3.0152, + "learning_rate": 5.296924708377518e-07, + "epoch": 1.4954682779456192 + }, + { + "current_steps": 990, + "loss": 2.9607, + "learning_rate": 5.291622481442205e-07, + "epoch": 1.4969788519637461 + }, + { + "current_steps": 991, + "loss": 2.8356, + "learning_rate": 5.286320254506893e-07, + "epoch": 1.498489425981873 + }, + { + "current_steps": 992, + "loss": 2.9473, + "learning_rate": 5.28101802757158e-07, + "epoch": 1.5 + }, + { + "current_steps": 993, + "loss": 2.9858, + "learning_rate": 5.275715800636266e-07, + "epoch": 1.501510574018127 + }, + { + "current_steps": 994, + "loss": 3.157, + "learning_rate": 5.270413573700954e-07, + "epoch": 1.5030211480362539 + }, + { + "current_steps": 995, + "loss": 3.1751, + "learning_rate": 5.265111346765642e-07, + "epoch": 1.5045317220543808 + }, + { + "current_steps": 996, + "loss": 2.8955, + "learning_rate": 5.259809119830329e-07, + "epoch": 1.5060422960725075 + }, + { + "current_steps": 997, + "loss": 3.0465, + "learning_rate": 5.254506892895015e-07, + "epoch": 1.5075528700906344 + }, + { + "current_steps": 998, + "loss": 2.6712, + "learning_rate": 5.249204665959703e-07, + "epoch": 1.5090634441087614 + }, + { + "current_steps": 999, + "loss": 2.9198, + "learning_rate": 5.24390243902439e-07, + "epoch": 1.510574018126888 + }, + { + "current_steps": 1000, + "loss": 2.9792, + "learning_rate": 5.238600212089077e-07, + "epoch": 1.512084592145015 + }, + { + "current_steps": 1001, + "loss": 2.5173, + "learning_rate": 5.233297985153764e-07, + "epoch": 1.513595166163142 + }, + { + "current_steps": 1002, + "loss": 3.2622, + "learning_rate": 5.227995758218452e-07, + "epoch": 1.5151057401812689 + }, + { + "current_steps": 1003, + "loss": 2.9889, + "learning_rate": 5.222693531283139e-07, + "epoch": 1.5166163141993958 + }, + { + "current_steps": 1004, + "loss": 2.9991, + "learning_rate": 5.217391304347825e-07, + "epoch": 1.5181268882175227 + }, + { + "current_steps": 1005, + "loss": 2.9715, + "learning_rate": 5.212089077412512e-07, + "epoch": 1.5196374622356497 + }, + { + "current_steps": 1006, + "loss": 3.1659, + "learning_rate": 5.206786850477201e-07, + "epoch": 1.5211480362537766 + }, + { + "current_steps": 1007, + "loss": 3.0155, + "learning_rate": 5.201484623541888e-07, + "epoch": 1.5226586102719033 + }, + { + "current_steps": 1008, + "loss": 3.0391, + "learning_rate": 5.196182396606574e-07, + "epoch": 1.5241691842900302 + }, + { + "current_steps": 1009, + "loss": 3.0686, + "learning_rate": 5.190880169671261e-07, + "epoch": 1.525679758308157 + }, + { + "current_steps": 1010, + "loss": 3.1287, + "learning_rate": 5.185577942735949e-07, + "epoch": 1.5271903323262839 + }, + { + "current_steps": 1011, + "loss": 3.0787, + "learning_rate": 5.180275715800636e-07, + "epoch": 1.5287009063444108 + }, + { + "current_steps": 1012, + "loss": 3.0413, + "learning_rate": 5.174973488865322e-07, + "epoch": 1.5302114803625377 + }, + { + "current_steps": 1013, + "loss": 2.9218, + "learning_rate": 5.169671261930011e-07, + "epoch": 1.5317220543806647 + }, + { + "current_steps": 1014, + "loss": 2.8721, + "learning_rate": 5.164369034994698e-07, + "epoch": 1.5332326283987916 + }, + { + "current_steps": 1015, + "loss": 2.922, + "learning_rate": 5.159066808059385e-07, + "epoch": 1.5347432024169185 + }, + { + "current_steps": 1016, + "loss": 3.1057, + "learning_rate": 5.153764581124071e-07, + "epoch": 1.5362537764350455 + }, + { + "current_steps": 1017, + "loss": 2.9731, + "learning_rate": 5.148462354188759e-07, + "epoch": 1.5377643504531722 + }, + { + "current_steps": 1018, + "loss": 3.1715, + "learning_rate": 5.143160127253446e-07, + "epoch": 1.539274924471299 + }, + { + "current_steps": 1019, + "loss": 3.045, + "learning_rate": 5.137857900318134e-07, + "epoch": 1.540785498489426 + }, + { + "current_steps": 1020, + "loss": 3.3725, + "learning_rate": 5.13255567338282e-07, + "epoch": 1.5422960725075527 + }, + { + "current_steps": 1021, + "loss": 2.7582, + "learning_rate": 5.127253446447508e-07, + "epoch": 1.5438066465256797 + }, + { + "current_steps": 1022, + "loss": 2.7718, + "learning_rate": 5.121951219512195e-07, + "epoch": 1.5453172205438066 + }, + { + "current_steps": 1023, + "loss": 2.9662, + "learning_rate": 5.116648992576882e-07, + "epoch": 1.5468277945619335 + }, + { + "current_steps": 1024, + "loss": 2.865, + "learning_rate": 5.111346765641568e-07, + "epoch": 1.5483383685800605 + }, + { + "current_steps": 1025, + "loss": 3.0375, + "learning_rate": 5.106044538706257e-07, + "epoch": 1.5498489425981874 + }, + { + "current_steps": 1026, + "loss": 3.0481, + "learning_rate": 5.100742311770944e-07, + "epoch": 1.5513595166163143 + }, + { + "current_steps": 1027, + "loss": 2.8947, + "learning_rate": 5.095440084835631e-07, + "epoch": 1.552870090634441 + }, + { + "current_steps": 1028, + "loss": 3.0734, + "learning_rate": 5.090137857900317e-07, + "epoch": 1.554380664652568 + }, + { + "current_steps": 1029, + "loss": 2.9264, + "learning_rate": 5.084835630965005e-07, + "epoch": 1.555891238670695 + }, + { + "current_steps": 1030, + "loss": 3.3168, + "learning_rate": 5.079533404029692e-07, + "epoch": 1.5574018126888216 + }, + { + "current_steps": 1031, + "loss": 3.019, + "learning_rate": 5.07423117709438e-07, + "epoch": 1.5589123867069485 + }, + { + "current_steps": 1032, + "loss": 2.8949, + "learning_rate": 5.068928950159067e-07, + "epoch": 1.5604229607250755 + }, + { + "current_steps": 1033, + "loss": 3.0329, + "learning_rate": 5.063626723223754e-07, + "epoch": 1.5619335347432024 + }, + { + "current_steps": 1034, + "loss": 2.9624, + "learning_rate": 5.058324496288441e-07, + "epoch": 1.5634441087613293 + }, + { + "current_steps": 1035, + "loss": 3.3392, + "learning_rate": 5.053022269353128e-07, + "epoch": 1.5649546827794563 + }, + { + "current_steps": 1036, + "loss": 2.2515, + "learning_rate": 5.047720042417815e-07, + "epoch": 1.5664652567975832 + }, + { + "current_steps": 1037, + "loss": 3.0141, + "learning_rate": 5.042417815482503e-07, + "epoch": 1.5679758308157101 + }, + { + "current_steps": 1038, + "loss": 3.0273, + "learning_rate": 5.03711558854719e-07, + "epoch": 1.5694864048338368 + }, + { + "current_steps": 1039, + "loss": 2.9199, + "learning_rate": 5.031813361611877e-07, + "epoch": 1.5709969788519638 + }, + { + "current_steps": 1040, + "loss": 3.1826, + "learning_rate": 5.026511134676564e-07, + "epoch": 1.5725075528700905 + }, + { + "current_steps": 1041, + "loss": 3.0795, + "learning_rate": 5.021208907741251e-07, + "epoch": 1.5740181268882174 + }, + { + "current_steps": 1042, + "loss": 3.0395, + "learning_rate": 5.015906680805938e-07, + "epoch": 1.5755287009063443 + }, + { + "current_steps": 1043, + "loss": 2.8528, + "learning_rate": 5.010604453870624e-07, + "epoch": 1.5770392749244713 + }, + { + "current_steps": 1044, + "loss": 2.7252, + "learning_rate": 5.005302226935313e-07, + "epoch": 1.5785498489425982 + }, + { + "current_steps": 1045, + "loss": 2.9675, + "learning_rate": 5e-07, + "epoch": 1.5800604229607251 + }, + { + "current_steps": 1046, + "loss": 2.956, + "learning_rate": 4.994697773064687e-07, + "epoch": 1.581570996978852 + }, + { + "current_steps": 1047, + "loss": 3.01, + "learning_rate": 4.989395546129374e-07, + "epoch": 1.583081570996979 + }, + { + "current_steps": 1048, + "loss": 2.9625, + "learning_rate": 4.984093319194061e-07, + "epoch": 1.5845921450151057 + }, + { + "current_steps": 1049, + "loss": 2.8837, + "learning_rate": 4.978791092258748e-07, + "epoch": 1.5861027190332326 + }, + { + "current_steps": 1050, + "loss": 3.0228, + "learning_rate": 4.973488865323436e-07, + "epoch": 1.5876132930513596 + }, + { + "current_steps": 1051, + "loss": 2.232, + "learning_rate": 4.968186638388123e-07, + "epoch": 1.5891238670694863 + }, + { + "current_steps": 1052, + "loss": 3.1003, + "learning_rate": 4.96288441145281e-07, + "epoch": 1.5906344410876132 + }, + { + "current_steps": 1053, + "loss": 3.1487, + "learning_rate": 4.957582184517497e-07, + "epoch": 1.5921450151057401 + }, + { + "current_steps": 1054, + "loss": 1.836, + "learning_rate": 4.952279957582184e-07, + "epoch": 1.593655589123867 + }, + { + "current_steps": 1055, + "loss": 3.1404, + "learning_rate": 4.946977730646871e-07, + "epoch": 1.595166163141994 + }, + { + "current_steps": 1056, + "loss": 3.036, + "learning_rate": 4.941675503711559e-07, + "epoch": 1.596676737160121 + }, + { + "current_steps": 1057, + "loss": 3.2197, + "learning_rate": 4.936373276776246e-07, + "epoch": 1.5981873111782479 + }, + { + "current_steps": 1058, + "loss": 3.0912, + "learning_rate": 4.931071049840933e-07, + "epoch": 1.5996978851963746 + }, + { + "current_steps": 1059, + "loss": 3.0635, + "learning_rate": 4.92576882290562e-07, + "epoch": 1.6012084592145015 + }, + { + "current_steps": 1060, + "loss": 2.8236, + "learning_rate": 4.920466595970307e-07, + "epoch": 1.6027190332326284 + }, + { + "current_steps": 1061, + "loss": 2.8799, + "learning_rate": 4.915164369034994e-07, + "epoch": 1.6042296072507551 + }, + { + "current_steps": 1062, + "loss": 3.0009, + "learning_rate": 4.909862142099681e-07, + "epoch": 1.605740181268882 + }, + { + "current_steps": 1063, + "loss": 2.7792, + "learning_rate": 4.904559915164369e-07, + "epoch": 1.607250755287009 + }, + { + "current_steps": 1064, + "loss": 2.4307, + "learning_rate": 4.899257688229056e-07, + "epoch": 1.608761329305136 + }, + { + "current_steps": 1065, + "loss": 3.0087, + "learning_rate": 4.893955461293743e-07, + "epoch": 1.6102719033232629 + }, + { + "current_steps": 1066, + "loss": 2.9968, + "learning_rate": 4.88865323435843e-07, + "epoch": 1.6117824773413898 + }, + { + "current_steps": 1067, + "loss": 3.0844, + "learning_rate": 4.883351007423117e-07, + "epoch": 1.6132930513595167 + }, + { + "current_steps": 1068, + "loss": 2.7457, + "learning_rate": 4.878048780487804e-07, + "epoch": 1.6148036253776437 + }, + { + "current_steps": 1069, + "loss": 3.1845, + "learning_rate": 4.872746553552492e-07, + "epoch": 1.6163141993957704 + }, + { + "current_steps": 1070, + "loss": 2.9618, + "learning_rate": 4.867444326617179e-07, + "epoch": 1.6178247734138973 + }, + { + "current_steps": 1071, + "loss": 3.1695, + "learning_rate": 4.862142099681866e-07, + "epoch": 1.619335347432024 + }, + { + "current_steps": 1072, + "loss": 3.048, + "learning_rate": 4.856839872746553e-07, + "epoch": 1.620845921450151 + }, + { + "current_steps": 1073, + "loss": 2.3976, + "learning_rate": 4.85153764581124e-07, + "epoch": 1.6223564954682779 + }, + { + "current_steps": 1074, + "loss": 2.8795, + "learning_rate": 4.846235418875927e-07, + "epoch": 1.6238670694864048 + }, + { + "current_steps": 1075, + "loss": 2.9775, + "learning_rate": 4.840933191940615e-07, + "epoch": 1.6253776435045317 + }, + { + "current_steps": 1076, + "loss": 3.0759, + "learning_rate": 4.835630965005302e-07, + "epoch": 1.6268882175226587 + }, + { + "current_steps": 1077, + "loss": 2.9119, + "learning_rate": 4.830328738069989e-07, + "epoch": 1.6283987915407856 + }, + { + "current_steps": 1078, + "loss": 2.9232, + "learning_rate": 4.825026511134676e-07, + "epoch": 1.6299093655589125 + }, + { + "current_steps": 1079, + "loss": 3.234, + "learning_rate": 4.819724284199363e-07, + "epoch": 1.6314199395770392 + }, + { + "current_steps": 1080, + "loss": 3.0302, + "learning_rate": 4.814422057264051e-07, + "epoch": 1.6329305135951662 + }, + { + "current_steps": 1081, + "loss": 2.9832, + "learning_rate": 4.809119830328737e-07, + "epoch": 1.634441087613293 + }, + { + "current_steps": 1082, + "loss": 3.0889, + "learning_rate": 4.803817603393426e-07, + "epoch": 1.6359516616314198 + }, + { + "current_steps": 1083, + "loss": 2.8111, + "learning_rate": 4.798515376458112e-07, + "epoch": 1.6374622356495467 + }, + { + "current_steps": 1084, + "loss": 2.8503, + "learning_rate": 4.7932131495228e-07, + "epoch": 1.6389728096676737 + }, + { + "current_steps": 1085, + "loss": 3.0871, + "learning_rate": 4.787910922587486e-07, + "epoch": 1.6404833836858006 + }, + { + "current_steps": 1086, + "loss": 2.9174, + "learning_rate": 4.782608695652174e-07, + "epoch": 1.6419939577039275 + }, + { + "current_steps": 1087, + "loss": 2.8806, + "learning_rate": 4.77730646871686e-07, + "epoch": 1.6435045317220545 + }, + { + "current_steps": 1088, + "loss": 3.13, + "learning_rate": 4.772004241781549e-07, + "epoch": 1.6450151057401814 + }, + { + "current_steps": 1089, + "loss": 2.6956, + "learning_rate": 4.766702014846235e-07, + "epoch": 1.646525679758308 + }, + { + "current_steps": 1090, + "loss": 2.8828, + "learning_rate": 4.7613997879109224e-07, + "epoch": 1.648036253776435 + }, + { + "current_steps": 1091, + "loss": 2.895, + "learning_rate": 4.756097560975609e-07, + "epoch": 1.649546827794562 + }, + { + "current_steps": 1092, + "loss": 2.8119, + "learning_rate": 4.750795334040297e-07, + "epoch": 1.6510574018126887 + }, + { + "current_steps": 1093, + "loss": 2.85, + "learning_rate": 4.7454931071049834e-07, + "epoch": 1.6525679758308156 + }, + { + "current_steps": 1094, + "loss": 3.1631, + "learning_rate": 4.740190880169671e-07, + "epoch": 1.6540785498489425 + }, + { + "current_steps": 1095, + "loss": 2.9764, + "learning_rate": 4.734888653234358e-07, + "epoch": 1.6555891238670695 + }, + { + "current_steps": 1096, + "loss": 3.0942, + "learning_rate": 4.7295864262990454e-07, + "epoch": 1.6570996978851964 + }, + { + "current_steps": 1097, + "loss": 2.8275, + "learning_rate": 4.7242841993637326e-07, + "epoch": 1.6586102719033233 + }, + { + "current_steps": 1098, + "loss": 2.9924, + "learning_rate": 4.7189819724284197e-07, + "epoch": 1.6601208459214503 + }, + { + "current_steps": 1099, + "loss": 2.9311, + "learning_rate": 4.713679745493107e-07, + "epoch": 1.6616314199395772 + }, + { + "current_steps": 1100, + "loss": 2.8092, + "learning_rate": 4.708377518557794e-07, + "epoch": 1.663141993957704 + }, + { + "current_steps": 1101, + "loss": 2.8916, + "learning_rate": 4.703075291622481e-07, + "epoch": 1.6646525679758308 + }, + { + "current_steps": 1102, + "loss": 3.2377, + "learning_rate": 4.6977730646871684e-07, + "epoch": 1.6661631419939575 + }, + { + "current_steps": 1103, + "loss": 3.0164, + "learning_rate": 4.6924708377518555e-07, + "epoch": 1.6676737160120845 + }, + { + "current_steps": 1104, + "loss": 3.0071, + "learning_rate": 4.6871686108165427e-07, + "epoch": 1.6691842900302114 + }, + { + "current_steps": 1105, + "loss": 3.116, + "learning_rate": 4.68186638388123e-07, + "epoch": 1.6706948640483383 + }, + { + "current_steps": 1106, + "loss": 3.2985, + "learning_rate": 4.676564156945917e-07, + "epoch": 1.6722054380664653 + }, + { + "current_steps": 1107, + "loss": 2.9589, + "learning_rate": 4.671261930010604e-07, + "epoch": 1.6737160120845922 + }, + { + "current_steps": 1108, + "loss": 2.9321, + "learning_rate": 4.6659597030752913e-07, + "epoch": 1.6752265861027191 + }, + { + "current_steps": 1109, + "loss": 3.1626, + "learning_rate": 4.6606574761399785e-07, + "epoch": 1.676737160120846 + }, + { + "current_steps": 1110, + "loss": 2.995, + "learning_rate": 4.6553552492046656e-07, + "epoch": 1.6782477341389728 + }, + { + "current_steps": 1111, + "loss": 2.5659, + "learning_rate": 4.650053022269353e-07, + "epoch": 1.6797583081570997 + }, + { + "current_steps": 1112, + "loss": 2.7322, + "learning_rate": 4.6447507953340405e-07, + "epoch": 1.6812688821752266 + }, + { + "current_steps": 1113, + "loss": 2.6716, + "learning_rate": 4.639448568398727e-07, + "epoch": 1.6827794561933533 + }, + { + "current_steps": 1114, + "loss": 2.9964, + "learning_rate": 4.634146341463415e-07, + "epoch": 1.6842900302114803 + }, + { + "current_steps": 1115, + "loss": 2.9699, + "learning_rate": 4.6288441145281014e-07, + "epoch": 1.6858006042296072 + }, + { + "current_steps": 1116, + "loss": 3.0185, + "learning_rate": 4.623541887592789e-07, + "epoch": 1.6873111782477341 + }, + { + "current_steps": 1117, + "loss": 2.8103, + "learning_rate": 4.6182396606574757e-07, + "epoch": 1.688821752265861 + }, + { + "current_steps": 1118, + "loss": 2.9484, + "learning_rate": 4.6129374337221634e-07, + "epoch": 1.690332326283988 + }, + { + "current_steps": 1119, + "loss": 3.0841, + "learning_rate": 4.60763520678685e-07, + "epoch": 1.691842900302115 + }, + { + "current_steps": 1120, + "loss": 2.9487, + "learning_rate": 4.602332979851538e-07, + "epoch": 1.6933534743202419 + }, + { + "current_steps": 1121, + "loss": 3.3735, + "learning_rate": 4.5970307529162244e-07, + "epoch": 1.6948640483383686 + }, + { + "current_steps": 1122, + "loss": 3.0413, + "learning_rate": 4.591728525980912e-07, + "epoch": 1.6963746223564955 + }, + { + "current_steps": 1123, + "loss": 2.9268, + "learning_rate": 4.5864262990455987e-07, + "epoch": 1.6978851963746222 + }, + { + "current_steps": 1124, + "loss": 3.07, + "learning_rate": 4.5811240721102864e-07, + "epoch": 1.6993957703927491 + }, + { + "current_steps": 1125, + "loss": 2.9197, + "learning_rate": 4.575821845174973e-07, + "epoch": 1.700906344410876 + }, + { + "current_steps": 1126, + "loss": 3.0171, + "learning_rate": 4.5705196182396607e-07, + "epoch": 1.702416918429003 + }, + { + "current_steps": 1127, + "loss": 3.1647, + "learning_rate": 4.5652173913043473e-07, + "epoch": 1.70392749244713 + }, + { + "current_steps": 1128, + "loss": 2.8764, + "learning_rate": 4.559915164369035e-07, + "epoch": 1.7054380664652569 + }, + { + "current_steps": 1129, + "loss": 3.005, + "learning_rate": 4.554612937433722e-07, + "epoch": 1.7069486404833838 + }, + { + "current_steps": 1130, + "loss": 3.1907, + "learning_rate": 4.5493107104984093e-07, + "epoch": 1.7084592145015107 + }, + { + "current_steps": 1131, + "loss": 3.0095, + "learning_rate": 4.5440084835630965e-07, + "epoch": 1.7099697885196374 + }, + { + "current_steps": 1132, + "loss": 2.9422, + "learning_rate": 4.5387062566277836e-07, + "epoch": 1.7114803625377644 + }, + { + "current_steps": 1133, + "loss": 3.0652, + "learning_rate": 4.533404029692471e-07, + "epoch": 1.712990936555891 + }, + { + "current_steps": 1134, + "loss": 2.9805, + "learning_rate": 4.5281018027571574e-07, + "epoch": 1.714501510574018 + }, + { + "current_steps": 1135, + "loss": 3.1721, + "learning_rate": 4.522799575821845e-07, + "epoch": 1.716012084592145 + }, + { + "current_steps": 1136, + "loss": 2.9939, + "learning_rate": 4.517497348886532e-07, + "epoch": 1.7175226586102719 + }, + { + "current_steps": 1137, + "loss": 3.0265, + "learning_rate": 4.5121951219512194e-07, + "epoch": 1.7190332326283988 + }, + { + "current_steps": 1138, + "loss": 3.2744, + "learning_rate": 4.506892895015906e-07, + "epoch": 1.7205438066465257 + }, + { + "current_steps": 1139, + "loss": 3.0499, + "learning_rate": 4.501590668080594e-07, + "epoch": 1.7220543806646527 + }, + { + "current_steps": 1140, + "loss": 2.9634, + "learning_rate": 4.4962884411452804e-07, + "epoch": 1.7235649546827796 + }, + { + "current_steps": 1141, + "loss": 3.0803, + "learning_rate": 4.490986214209968e-07, + "epoch": 1.7250755287009063 + }, + { + "current_steps": 1142, + "loss": 2.9291, + "learning_rate": 4.4856839872746547e-07, + "epoch": 1.7265861027190332 + }, + { + "current_steps": 1143, + "loss": 3.019, + "learning_rate": 4.4803817603393424e-07, + "epoch": 1.7280966767371602 + }, + { + "current_steps": 1144, + "loss": 2.9313, + "learning_rate": 4.475079533404029e-07, + "epoch": 1.7296072507552869 + }, + { + "current_steps": 1145, + "loss": 2.7374, + "learning_rate": 4.4697773064687167e-07, + "epoch": 1.7311178247734138 + }, + { + "current_steps": 1146, + "loss": 2.4116, + "learning_rate": 4.464475079533404e-07, + "epoch": 1.7326283987915407 + }, + { + "current_steps": 1147, + "loss": 2.6954, + "learning_rate": 4.459172852598091e-07, + "epoch": 1.7341389728096677 + }, + { + "current_steps": 1148, + "loss": 2.9735, + "learning_rate": 4.453870625662778e-07, + "epoch": 1.7356495468277946 + }, + { + "current_steps": 1149, + "loss": 2.9407, + "learning_rate": 4.4485683987274653e-07, + "epoch": 1.7371601208459215 + }, + { + "current_steps": 1150, + "loss": 3.0389, + "learning_rate": 4.4432661717921525e-07, + "epoch": 1.7386706948640485 + }, + { + "current_steps": 1151, + "loss": 2.4312, + "learning_rate": 4.4379639448568397e-07, + "epoch": 1.7401812688821754 + }, + { + "current_steps": 1152, + "loss": 2.9282, + "learning_rate": 4.432661717921527e-07, + "epoch": 1.741691842900302 + }, + { + "current_steps": 1153, + "loss": 2.918, + "learning_rate": 4.427359490986214e-07, + "epoch": 1.743202416918429 + }, + { + "current_steps": 1154, + "loss": 3.0313, + "learning_rate": 4.422057264050901e-07, + "epoch": 1.7447129909365557 + }, + { + "current_steps": 1155, + "loss": 2.8787, + "learning_rate": 4.4167550371155883e-07, + "epoch": 1.7462235649546827 + }, + { + "current_steps": 1156, + "loss": 3.3246, + "learning_rate": 4.4114528101802755e-07, + "epoch": 1.7477341389728096 + }, + { + "current_steps": 1157, + "loss": 3.0239, + "learning_rate": 4.4061505832449626e-07, + "epoch": 1.7492447129909365 + }, + { + "current_steps": 1158, + "loss": 2.8333, + "learning_rate": 4.40084835630965e-07, + "epoch": 1.7507552870090635 + }, + { + "current_steps": 1159, + "loss": 3.0921, + "learning_rate": 4.395546129374337e-07, + "epoch": 1.7522658610271904 + }, + { + "current_steps": 1160, + "loss": 3.0326, + "learning_rate": 4.390243902439024e-07, + "epoch": 1.7537764350453173 + }, + { + "current_steps": 1161, + "loss": 3.024, + "learning_rate": 4.384941675503712e-07, + "epoch": 1.7552870090634443 + }, + { + "current_steps": 1162, + "loss": 3.0421, + "learning_rate": 4.3796394485683984e-07, + "epoch": 1.756797583081571 + }, + { + "current_steps": 1163, + "loss": 2.9477, + "learning_rate": 4.374337221633086e-07, + "epoch": 1.758308157099698 + }, + { + "current_steps": 1164, + "loss": 3.0383, + "learning_rate": 4.3690349946977727e-07, + "epoch": 1.7598187311178246 + }, + { + "current_steps": 1165, + "loss": 2.8124, + "learning_rate": 4.3637327677624604e-07, + "epoch": 1.7613293051359515 + }, + { + "current_steps": 1166, + "loss": 3.0573, + "learning_rate": 4.358430540827147e-07, + "epoch": 1.7628398791540785 + }, + { + "current_steps": 1167, + "loss": 2.9941, + "learning_rate": 4.3531283138918347e-07, + "epoch": 1.7643504531722054 + }, + { + "current_steps": 1168, + "loss": 3.0391, + "learning_rate": 4.3478260869565214e-07, + "epoch": 1.7658610271903323 + }, + { + "current_steps": 1169, + "loss": 3.0691, + "learning_rate": 4.342523860021209e-07, + "epoch": 1.7673716012084593 + }, + { + "current_steps": 1170, + "loss": 2.916, + "learning_rate": 4.3372216330858957e-07, + "epoch": 1.7688821752265862 + }, + { + "current_steps": 1171, + "loss": 3.2599, + "learning_rate": 4.3319194061505834e-07, + "epoch": 1.7703927492447131 + }, + { + "current_steps": 1172, + "loss": 3.1476, + "learning_rate": 4.32661717921527e-07, + "epoch": 1.7719033232628398 + }, + { + "current_steps": 1173, + "loss": 2.758, + "learning_rate": 4.3213149522799577e-07, + "epoch": 1.7734138972809668 + }, + { + "current_steps": 1174, + "loss": 3.0326, + "learning_rate": 4.3160127253446443e-07, + "epoch": 1.7749244712990937 + }, + { + "current_steps": 1175, + "loss": 3.0573, + "learning_rate": 4.310710498409332e-07, + "epoch": 1.7764350453172204 + }, + { + "current_steps": 1176, + "loss": 3.0204, + "learning_rate": 4.3054082714740186e-07, + "epoch": 1.7779456193353473 + }, + { + "current_steps": 1177, + "loss": 2.9169, + "learning_rate": 4.3001060445387063e-07, + "epoch": 1.7794561933534743 + }, + { + "current_steps": 1178, + "loss": 3.0008, + "learning_rate": 4.2948038176033935e-07, + "epoch": 1.7809667673716012 + }, + { + "current_steps": 1179, + "loss": 3.0732, + "learning_rate": 4.2895015906680806e-07, + "epoch": 1.7824773413897281 + }, + { + "current_steps": 1180, + "loss": 2.9596, + "learning_rate": 4.284199363732768e-07, + "epoch": 1.783987915407855 + }, + { + "current_steps": 1181, + "loss": 2.971, + "learning_rate": 4.2788971367974544e-07, + "epoch": 1.785498489425982 + }, + { + "current_steps": 1182, + "loss": 3.1206, + "learning_rate": 4.273594909862142e-07, + "epoch": 1.787009063444109 + }, + { + "current_steps": 1183, + "loss": 3.1017, + "learning_rate": 4.268292682926829e-07, + "epoch": 1.7885196374622356 + }, + { + "current_steps": 1184, + "loss": 2.8115, + "learning_rate": 4.2629904559915164e-07, + "epoch": 1.7900302114803626 + }, + { + "current_steps": 1185, + "loss": 2.9832, + "learning_rate": 4.257688229056203e-07, + "epoch": 1.7915407854984893 + }, + { + "current_steps": 1186, + "loss": 2.9295, + "learning_rate": 4.252386002120891e-07, + "epoch": 1.7930513595166162 + }, + { + "current_steps": 1187, + "loss": 3.0374, + "learning_rate": 4.2470837751855774e-07, + "epoch": 1.7945619335347431 + }, + { + "current_steps": 1188, + "loss": 3.0179, + "learning_rate": 4.241781548250265e-07, + "epoch": 1.79607250755287 + }, + { + "current_steps": 1189, + "loss": 2.6478, + "learning_rate": 4.2364793213149517e-07, + "epoch": 1.797583081570997 + }, + { + "current_steps": 1190, + "loss": 3.1365, + "learning_rate": 4.2311770943796394e-07, + "epoch": 1.799093655589124 + }, + { + "current_steps": 1191, + "loss": 2.9728, + "learning_rate": 4.225874867444326e-07, + "epoch": 1.8006042296072509 + }, + { + "current_steps": 1192, + "loss": 3.0082, + "learning_rate": 4.2205726405090137e-07, + "epoch": 1.8021148036253778 + }, + { + "current_steps": 1193, + "loss": 3.0707, + "learning_rate": 4.2152704135737003e-07, + "epoch": 1.8036253776435045 + }, + { + "current_steps": 1194, + "loss": 3.0781, + "learning_rate": 4.209968186638388e-07, + "epoch": 1.8051359516616314 + }, + { + "current_steps": 1195, + "loss": 2.3737, + "learning_rate": 4.204665959703075e-07, + "epoch": 1.8066465256797581 + }, + { + "current_steps": 1196, + "loss": 2.7853, + "learning_rate": 4.1993637327677623e-07, + "epoch": 1.808157099697885 + }, + { + "current_steps": 1197, + "loss": 2.69, + "learning_rate": 4.1940615058324495e-07, + "epoch": 1.809667673716012 + }, + { + "current_steps": 1198, + "loss": 2.9768, + "learning_rate": 4.1887592788971367e-07, + "epoch": 1.811178247734139 + }, + { + "current_steps": 1199, + "loss": 2.9662, + "learning_rate": 4.183457051961824e-07, + "epoch": 1.8126888217522659 + }, + { + "current_steps": 1200, + "loss": 3.1874, + "learning_rate": 4.178154825026511e-07, + "epoch": 1.8141993957703928 + }, + { + "current_steps": 1201, + "loss": 2.9907, + "learning_rate": 4.172852598091198e-07, + "epoch": 1.8157099697885197 + }, + { + "current_steps": 1202, + "loss": 3.2888, + "learning_rate": 4.1675503711558853e-07, + "epoch": 1.8172205438066467 + }, + { + "current_steps": 1203, + "loss": 2.8668, + "learning_rate": 4.1622481442205724e-07, + "epoch": 1.8187311178247734 + }, + { + "current_steps": 1204, + "loss": 2.8886, + "learning_rate": 4.1569459172852596e-07, + "epoch": 1.8202416918429003 + }, + { + "current_steps": 1205, + "loss": 2.9962, + "learning_rate": 4.151643690349947e-07, + "epoch": 1.8217522658610272 + }, + { + "current_steps": 1206, + "loss": 2.9163, + "learning_rate": 4.146341463414634e-07, + "epoch": 1.823262839879154 + }, + { + "current_steps": 1207, + "loss": 2.8536, + "learning_rate": 4.141039236479321e-07, + "epoch": 1.8247734138972809 + }, + { + "current_steps": 1208, + "loss": 2.8376, + "learning_rate": 4.135737009544008e-07, + "epoch": 1.8262839879154078 + }, + { + "current_steps": 1209, + "loss": 2.9377, + "learning_rate": 4.1304347826086954e-07, + "epoch": 1.8277945619335347 + }, + { + "current_steps": 1210, + "loss": 2.7267, + "learning_rate": 4.1251325556733826e-07, + "epoch": 1.8293051359516617 + }, + { + "current_steps": 1211, + "loss": 2.8558, + "learning_rate": 4.1198303287380697e-07, + "epoch": 1.8308157099697886 + }, + { + "current_steps": 1212, + "loss": 3.2157, + "learning_rate": 4.1145281018027574e-07, + "epoch": 1.8323262839879155 + }, + { + "current_steps": 1213, + "loss": 2.5958, + "learning_rate": 4.109225874867444e-07, + "epoch": 1.8338368580060425 + }, + { + "current_steps": 1214, + "loss": 2.9366, + "learning_rate": 4.1039236479321317e-07, + "epoch": 1.8353474320241692 + }, + { + "current_steps": 1215, + "loss": 2.971, + "learning_rate": 4.0986214209968183e-07, + "epoch": 1.836858006042296 + }, + { + "current_steps": 1216, + "loss": 2.69, + "learning_rate": 4.093319194061506e-07, + "epoch": 1.8383685800604228 + }, + { + "current_steps": 1217, + "loss": 3.0503, + "learning_rate": 4.0880169671261927e-07, + "epoch": 1.8398791540785497 + }, + { + "current_steps": 1218, + "loss": 3.1028, + "learning_rate": 4.0827147401908804e-07, + "epoch": 1.8413897280966767 + }, + { + "current_steps": 1219, + "loss": 3.1497, + "learning_rate": 4.077412513255567e-07, + "epoch": 1.8429003021148036 + }, + { + "current_steps": 1220, + "loss": 3.0073, + "learning_rate": 4.0721102863202547e-07, + "epoch": 1.8444108761329305 + }, + { + "current_steps": 1221, + "loss": 3.0223, + "learning_rate": 4.0668080593849413e-07, + "epoch": 1.8459214501510575 + }, + { + "current_steps": 1222, + "loss": 3.1263, + "learning_rate": 4.061505832449629e-07, + "epoch": 1.8474320241691844 + }, + { + "current_steps": 1223, + "loss": 2.9748, + "learning_rate": 4.0562036055143156e-07, + "epoch": 1.8489425981873113 + }, + { + "current_steps": 1224, + "loss": 3.0741, + "learning_rate": 4.0509013785790033e-07, + "epoch": 1.850453172205438 + }, + { + "current_steps": 1225, + "loss": 3.1341, + "learning_rate": 4.04559915164369e-07, + "epoch": 1.851963746223565 + }, + { + "current_steps": 1226, + "loss": 2.9934, + "learning_rate": 4.040296924708377e-07, + "epoch": 1.853474320241692 + }, + { + "current_steps": 1227, + "loss": 2.8055, + "learning_rate": 4.034994697773065e-07, + "epoch": 1.8549848942598186 + }, + { + "current_steps": 1228, + "loss": 2.9672, + "learning_rate": 4.0296924708377514e-07, + "epoch": 1.8564954682779455 + }, + { + "current_steps": 1229, + "loss": 3.0207, + "learning_rate": 4.024390243902439e-07, + "epoch": 1.8580060422960725 + }, + { + "current_steps": 1230, + "loss": 2.9925, + "learning_rate": 4.0190880169671257e-07, + "epoch": 1.8595166163141994 + }, + { + "current_steps": 1231, + "loss": 2.2767, + "learning_rate": 4.0137857900318134e-07, + "epoch": 1.8610271903323263 + }, + { + "current_steps": 1232, + "loss": 2.8183, + "learning_rate": 4.0084835630965e-07, + "epoch": 1.8625377643504533 + }, + { + "current_steps": 1233, + "loss": 3.1208, + "learning_rate": 4.003181336161188e-07, + "epoch": 1.8640483383685802 + }, + { + "current_steps": 1234, + "loss": 2.8493, + "learning_rate": 3.9978791092258744e-07, + "epoch": 1.865558912386707 + }, + { + "current_steps": 1235, + "loss": 3.2308, + "learning_rate": 3.992576882290562e-07, + "epoch": 1.8670694864048338 + }, + { + "current_steps": 1236, + "loss": 3.1253, + "learning_rate": 3.9872746553552487e-07, + "epoch": 1.8685800604229608 + }, + { + "current_steps": 1237, + "loss": 2.9741, + "learning_rate": 3.9819724284199364e-07, + "epoch": 1.8700906344410875 + }, + { + "current_steps": 1238, + "loss": 3.0487, + "learning_rate": 3.976670201484623e-07, + "epoch": 1.8716012084592144 + }, + { + "current_steps": 1239, + "loss": 2.8604, + "learning_rate": 3.9713679745493107e-07, + "epoch": 1.8731117824773413 + }, + { + "current_steps": 1240, + "loss": 2.9669, + "learning_rate": 3.9660657476139973e-07, + "epoch": 1.8746223564954683 + }, + { + "current_steps": 1241, + "loss": 2.7471, + "learning_rate": 3.960763520678685e-07, + "epoch": 1.8761329305135952 + }, + { + "current_steps": 1242, + "loss": 3.0211, + "learning_rate": 3.9554612937433716e-07, + "epoch": 1.8776435045317221 + }, + { + "current_steps": 1243, + "loss": 3.1709, + "learning_rate": 3.9501590668080593e-07, + "epoch": 1.879154078549849 + }, + { + "current_steps": 1244, + "loss": 2.9585, + "learning_rate": 3.9448568398727465e-07, + "epoch": 1.880664652567976 + }, + { + "current_steps": 1245, + "loss": 2.2015, + "learning_rate": 3.9395546129374336e-07, + "epoch": 1.8821752265861027 + }, + { + "current_steps": 1246, + "loss": 2.9052, + "learning_rate": 3.934252386002121e-07, + "epoch": 1.8836858006042296 + }, + { + "current_steps": 1247, + "loss": 2.9724, + "learning_rate": 3.928950159066808e-07, + "epoch": 1.8851963746223563 + }, + { + "current_steps": 1248, + "loss": 3.0223, + "learning_rate": 3.923647932131495e-07, + "epoch": 1.8867069486404833 + }, + { + "current_steps": 1249, + "loss": 2.9257, + "learning_rate": 3.9183457051961823e-07, + "epoch": 1.8882175226586102 + }, + { + "current_steps": 1250, + "loss": 2.9357, + "learning_rate": 3.9130434782608694e-07, + "epoch": 1.8897280966767371 + }, + { + "current_steps": 1251, + "loss": 3.0655, + "learning_rate": 3.9077412513255566e-07, + "epoch": 1.891238670694864 + }, + { + "current_steps": 1252, + "loss": 2.9263, + "learning_rate": 3.902439024390244e-07, + "epoch": 1.892749244712991 + }, + { + "current_steps": 1253, + "loss": 3.0746, + "learning_rate": 3.897136797454931e-07, + "epoch": 1.894259818731118 + }, + { + "current_steps": 1254, + "loss": 2.9773, + "learning_rate": 3.891834570519618e-07, + "epoch": 1.8957703927492449 + }, + { + "current_steps": 1255, + "loss": 3.0119, + "learning_rate": 3.886532343584305e-07, + "epoch": 1.8972809667673716 + }, + { + "current_steps": 1256, + "loss": 2.7599, + "learning_rate": 3.8812301166489924e-07, + "epoch": 1.8987915407854985 + }, + { + "current_steps": 1257, + "loss": 2.9744, + "learning_rate": 3.8759278897136795e-07, + "epoch": 1.9003021148036254 + }, + { + "current_steps": 1258, + "loss": 3.1058, + "learning_rate": 3.8706256627783667e-07, + "epoch": 1.9018126888217521 + }, + { + "current_steps": 1259, + "loss": 3.071, + "learning_rate": 3.865323435843054e-07, + "epoch": 1.903323262839879 + }, + { + "current_steps": 1260, + "loss": 2.8942, + "learning_rate": 3.860021208907741e-07, + "epoch": 1.904833836858006 + }, + { + "current_steps": 1261, + "loss": 2.7528, + "learning_rate": 3.8547189819724287e-07, + "epoch": 1.906344410876133 + }, + { + "current_steps": 1262, + "loss": 2.3892, + "learning_rate": 3.8494167550371153e-07, + "epoch": 1.9078549848942599 + }, + { + "current_steps": 1263, + "loss": 3.075, + "learning_rate": 3.844114528101803e-07, + "epoch": 1.9093655589123868 + }, + { + "current_steps": 1264, + "loss": 2.9926, + "learning_rate": 3.8388123011664897e-07, + "epoch": 1.9108761329305137 + }, + { + "current_steps": 1265, + "loss": 3.1426, + "learning_rate": 3.8335100742311773e-07, + "epoch": 1.9123867069486404 + }, + { + "current_steps": 1266, + "loss": 2.9099, + "learning_rate": 3.828207847295864e-07, + "epoch": 1.9138972809667674 + }, + { + "current_steps": 1267, + "loss": 2.9872, + "learning_rate": 3.8229056203605517e-07, + "epoch": 1.9154078549848943 + }, + { + "current_steps": 1268, + "loss": 3.1473, + "learning_rate": 3.8176033934252383e-07, + "epoch": 1.916918429003021 + }, + { + "current_steps": 1269, + "loss": 3.0227, + "learning_rate": 3.812301166489926e-07, + "epoch": 1.918429003021148 + }, + { + "current_steps": 1270, + "loss": 2.9583, + "learning_rate": 3.8069989395546126e-07, + "epoch": 1.9199395770392749 + }, + { + "current_steps": 1271, + "loss": 3.1131, + "learning_rate": 3.8016967126193e-07, + "epoch": 1.9214501510574018 + }, + { + "current_steps": 1272, + "loss": 2.8828, + "learning_rate": 3.796394485683987e-07, + "epoch": 1.9229607250755287 + }, + { + "current_steps": 1273, + "loss": 3.0638, + "learning_rate": 3.791092258748674e-07, + "epoch": 1.9244712990936557 + }, + { + "current_steps": 1274, + "loss": 2.876, + "learning_rate": 3.785790031813361e-07, + "epoch": 1.9259818731117826 + }, + { + "current_steps": 1275, + "loss": 3.2298, + "learning_rate": 3.7804878048780484e-07, + "epoch": 1.9274924471299095 + }, + { + "current_steps": 1276, + "loss": 2.9171, + "learning_rate": 3.7751855779427356e-07, + "epoch": 1.9290030211480362 + }, + { + "current_steps": 1277, + "loss": 3.0993, + "learning_rate": 3.7698833510074227e-07, + "epoch": 1.9305135951661632 + }, + { + "current_steps": 1278, + "loss": 3.2426, + "learning_rate": 3.7645811240721104e-07, + "epoch": 1.9320241691842899 + }, + { + "current_steps": 1279, + "loss": 2.8922, + "learning_rate": 3.759278897136797e-07, + "epoch": 1.9335347432024168 + }, + { + "current_steps": 1280, + "loss": 3.1178, + "learning_rate": 3.7539766702014847e-07, + "epoch": 1.9350453172205437 + }, + { + "current_steps": 1281, + "loss": 3.5744, + "learning_rate": 3.7486744432661714e-07, + "epoch": 1.9365558912386707 + }, + { + "current_steps": 1282, + "loss": 2.8252, + "learning_rate": 3.743372216330859e-07, + "epoch": 1.9380664652567976 + }, + { + "current_steps": 1283, + "loss": 2.998, + "learning_rate": 3.7380699893955457e-07, + "epoch": 1.9395770392749245 + }, + { + "current_steps": 1284, + "loss": 2.6686, + "learning_rate": 3.7327677624602334e-07, + "epoch": 1.9410876132930515 + }, + { + "current_steps": 1285, + "loss": 3.123, + "learning_rate": 3.72746553552492e-07, + "epoch": 1.9425981873111784 + }, + { + "current_steps": 1286, + "loss": 3.0199, + "learning_rate": 3.7221633085896077e-07, + "epoch": 1.944108761329305 + }, + { + "current_steps": 1287, + "loss": 2.8993, + "learning_rate": 3.7168610816542943e-07, + "epoch": 1.945619335347432 + }, + { + "current_steps": 1288, + "loss": 3.1056, + "learning_rate": 3.711558854718982e-07, + "epoch": 1.947129909365559 + }, + { + "current_steps": 1289, + "loss": 3.0076, + "learning_rate": 3.7062566277836686e-07, + "epoch": 1.9486404833836857 + }, + { + "current_steps": 1290, + "loss": 2.8532, + "learning_rate": 3.7009544008483563e-07, + "epoch": 1.9501510574018126 + }, + { + "current_steps": 1291, + "loss": 2.9681, + "learning_rate": 3.695652173913043e-07, + "epoch": 1.9516616314199395 + }, + { + "current_steps": 1292, + "loss": 2.186, + "learning_rate": 3.6903499469777306e-07, + "epoch": 1.9531722054380665 + }, + { + "current_steps": 1293, + "loss": 2.8652, + "learning_rate": 3.685047720042417e-07, + "epoch": 1.9546827794561934 + }, + { + "current_steps": 1294, + "loss": 2.4061, + "learning_rate": 3.679745493107105e-07, + "epoch": 1.9561933534743203 + }, + { + "current_steps": 1295, + "loss": 2.9065, + "learning_rate": 3.674443266171792e-07, + "epoch": 1.9577039274924473 + }, + { + "current_steps": 1296, + "loss": 2.9456, + "learning_rate": 3.6691410392364793e-07, + "epoch": 1.959214501510574 + }, + { + "current_steps": 1297, + "loss": 2.8736, + "learning_rate": 3.6638388123011664e-07, + "epoch": 1.960725075528701 + }, + { + "current_steps": 1298, + "loss": 2.9118, + "learning_rate": 3.6585365853658536e-07, + "epoch": 1.9622356495468278 + }, + { + "current_steps": 1299, + "loss": 2.8376, + "learning_rate": 3.653234358430541e-07, + "epoch": 1.9637462235649545 + }, + { + "current_steps": 1300, + "loss": 3.0716, + "learning_rate": 3.647932131495228e-07, + "epoch": 1.9652567975830815 + }, + { + "current_steps": 1301, + "loss": 2.2068, + "learning_rate": 3.642629904559915e-07, + "epoch": 1.9667673716012084 + }, + { + "current_steps": 1302, + "loss": 2.8887, + "learning_rate": 3.637327677624602e-07, + "epoch": 1.9682779456193353 + }, + { + "current_steps": 1303, + "loss": 3.0633, + "learning_rate": 3.6320254506892894e-07, + "epoch": 1.9697885196374623 + }, + { + "current_steps": 1304, + "loss": 3.1106, + "learning_rate": 3.6267232237539765e-07, + "epoch": 1.9712990936555892 + }, + { + "current_steps": 1305, + "loss": 3.075, + "learning_rate": 3.6214209968186637e-07, + "epoch": 1.9728096676737161 + }, + { + "current_steps": 1306, + "loss": 3.0428, + "learning_rate": 3.616118769883351e-07, + "epoch": 1.974320241691843 + }, + { + "current_steps": 1307, + "loss": 3.0262, + "learning_rate": 3.610816542948038e-07, + "epoch": 1.9758308157099698 + }, + { + "current_steps": 1308, + "loss": 2.8567, + "learning_rate": 3.605514316012725e-07, + "epoch": 1.9773413897280967 + }, + { + "current_steps": 1309, + "loss": 2.8223, + "learning_rate": 3.6002120890774123e-07, + "epoch": 1.9788519637462234 + }, + { + "current_steps": 1310, + "loss": 2.9899, + "learning_rate": 3.5949098621421e-07, + "epoch": 1.9803625377643503 + }, + { + "current_steps": 1311, + "loss": 2.9231, + "learning_rate": 3.5896076352067866e-07, + "epoch": 1.9818731117824773 + }, + { + "current_steps": 1312, + "loss": 2.9539, + "learning_rate": 3.5843054082714743e-07, + "epoch": 1.9833836858006042 + }, + { + "current_steps": 1313, + "loss": 2.9124, + "learning_rate": 3.579003181336161e-07, + "epoch": 1.9848942598187311 + }, + { + "current_steps": 1314, + "loss": 2.9618, + "learning_rate": 3.5737009544008487e-07, + "epoch": 1.986404833836858 + }, + { + "current_steps": 1315, + "loss": 2.3594, + "learning_rate": 3.5683987274655353e-07, + "epoch": 1.987915407854985 + }, + { + "current_steps": 1316, + "loss": 2.7551, + "learning_rate": 3.5630965005302224e-07, + "epoch": 1.989425981873112 + }, + { + "current_steps": 1317, + "loss": 2.8081, + "learning_rate": 3.5577942735949096e-07, + "epoch": 1.9909365558912386 + }, + { + "current_steps": 1318, + "loss": 3.1055, + "learning_rate": 3.552492046659597e-07, + "epoch": 1.9924471299093656 + }, + { + "current_steps": 1319, + "loss": 2.9898, + "learning_rate": 3.547189819724284e-07, + "epoch": 1.9939577039274925 + }, + { + "current_steps": 1320, + "loss": 3.066, + "learning_rate": 3.541887592788971e-07, + "epoch": 1.9954682779456192 + }, + { + "current_steps": 1321, + "loss": 2.865, + "learning_rate": 3.536585365853658e-07, + "epoch": 1.9969788519637461 + }, + { + "current_steps": 1322, + "loss": 2.8501, + "learning_rate": 3.5312831389183454e-07, + "epoch": 1.998489425981873 + }, + { + "current_steps": 1323, + "loss": 3.094, + "learning_rate": 3.5259809119830326e-07, + "epoch": 2.0 + }, + { + "current_steps": 1324, + "loss": 2.6459, + "learning_rate": 3.5206786850477197e-07, + "epoch": 2.001510574018127 + }, + { + "current_steps": 1325, + "loss": 3.0059, + "learning_rate": 3.515376458112407e-07, + "epoch": 2.003021148036254 + }, + { + "current_steps": 1326, + "loss": 3.1677, + "learning_rate": 3.510074231177094e-07, + "epoch": 2.004531722054381 + }, + { + "current_steps": 1327, + "loss": 2.4981, + "learning_rate": 3.5047720042417817e-07, + "epoch": 2.0060422960725077 + }, + { + "current_steps": 1328, + "loss": 2.7343, + "learning_rate": 3.4994697773064683e-07, + "epoch": 2.0075528700906347 + }, + { + "current_steps": 1329, + "loss": 2.8042, + "learning_rate": 3.494167550371156e-07, + "epoch": 2.009063444108761 + }, + { + "current_steps": 1330, + "loss": 2.9556, + "learning_rate": 3.4888653234358427e-07, + "epoch": 2.010574018126888 + }, + { + "current_steps": 1331, + "loss": 3.1188, + "learning_rate": 3.4835630965005304e-07, + "epoch": 2.012084592145015 + }, + { + "current_steps": 1332, + "loss": 2.9043, + "learning_rate": 3.478260869565217e-07, + "epoch": 2.013595166163142 + }, + { + "current_steps": 1333, + "loss": 3.0621, + "learning_rate": 3.4729586426299047e-07, + "epoch": 2.015105740181269 + }, + { + "current_steps": 1334, + "loss": 2.9419, + "learning_rate": 3.4676564156945913e-07, + "epoch": 2.016616314199396 + }, + { + "current_steps": 1335, + "loss": 3.0796, + "learning_rate": 3.462354188759279e-07, + "epoch": 2.0181268882175227 + }, + { + "current_steps": 1336, + "loss": 2.8447, + "learning_rate": 3.4570519618239656e-07, + "epoch": 2.0196374622356497 + }, + { + "current_steps": 1337, + "loss": 3.1167, + "learning_rate": 3.4517497348886533e-07, + "epoch": 2.0211480362537766 + }, + { + "current_steps": 1338, + "loss": 3.2094, + "learning_rate": 3.44644750795334e-07, + "epoch": 2.0226586102719035 + }, + { + "current_steps": 1339, + "loss": 2.9261, + "learning_rate": 3.4411452810180276e-07, + "epoch": 2.02416918429003 + }, + { + "current_steps": 1340, + "loss": 2.8645, + "learning_rate": 3.435843054082714e-07, + "epoch": 2.025679758308157 + }, + { + "current_steps": 1341, + "loss": 2.9348, + "learning_rate": 3.430540827147402e-07, + "epoch": 2.027190332326284 + }, + { + "current_steps": 1342, + "loss": 2.9588, + "learning_rate": 3.4252386002120886e-07, + "epoch": 2.028700906344411 + }, + { + "current_steps": 1343, + "loss": 3.0665, + "learning_rate": 3.419936373276776e-07, + "epoch": 2.0302114803625377 + }, + { + "current_steps": 1344, + "loss": 3.0973, + "learning_rate": 3.4146341463414634e-07, + "epoch": 2.0317220543806647 + }, + { + "current_steps": 1345, + "loss": 2.9658, + "learning_rate": 3.4093319194061506e-07, + "epoch": 2.0332326283987916 + }, + { + "current_steps": 1346, + "loss": 2.9395, + "learning_rate": 3.4040296924708377e-07, + "epoch": 2.0347432024169185 + }, + { + "current_steps": 1347, + "loss": 3.0696, + "learning_rate": 3.398727465535525e-07, + "epoch": 2.0362537764350455 + }, + { + "current_steps": 1348, + "loss": 2.9523, + "learning_rate": 3.393425238600212e-07, + "epoch": 2.0377643504531724 + }, + { + "current_steps": 1349, + "loss": 3.1074, + "learning_rate": 3.388123011664899e-07, + "epoch": 2.039274924471299 + }, + { + "current_steps": 1350, + "loss": 3.0615, + "learning_rate": 3.3828207847295864e-07, + "epoch": 2.040785498489426 + }, + { + "current_steps": 1351, + "loss": 2.9761, + "learning_rate": 3.3775185577942735e-07, + "epoch": 2.0422960725075527 + }, + { + "current_steps": 1352, + "loss": 2.6124, + "learning_rate": 3.3722163308589607e-07, + "epoch": 2.0438066465256797 + }, + { + "current_steps": 1353, + "loss": 2.8982, + "learning_rate": 3.366914103923648e-07, + "epoch": 2.0453172205438066 + }, + { + "current_steps": 1354, + "loss": 2.8904, + "learning_rate": 3.361611876988335e-07, + "epoch": 2.0468277945619335 + }, + { + "current_steps": 1355, + "loss": 3.1752, + "learning_rate": 3.356309650053022e-07, + "epoch": 2.0483383685800605 + }, + { + "current_steps": 1356, + "loss": 2.9027, + "learning_rate": 3.3510074231177093e-07, + "epoch": 2.0498489425981874 + }, + { + "current_steps": 1357, + "loss": 2.9032, + "learning_rate": 3.3457051961823965e-07, + "epoch": 2.0513595166163143 + }, + { + "current_steps": 1358, + "loss": 2.9716, + "learning_rate": 3.3404029692470836e-07, + "epoch": 2.0528700906344413 + }, + { + "current_steps": 1359, + "loss": 2.9792, + "learning_rate": 3.3351007423117703e-07, + "epoch": 2.054380664652568 + }, + { + "current_steps": 1360, + "loss": 2.8627, + "learning_rate": 3.329798515376458e-07, + "epoch": 2.0558912386706947 + }, + { + "current_steps": 1361, + "loss": 3.2493, + "learning_rate": 3.3244962884411456e-07, + "epoch": 2.0574018126888216 + }, + { + "current_steps": 1362, + "loss": 3.0259, + "learning_rate": 3.3191940615058323e-07, + "epoch": 2.0589123867069485 + }, + { + "current_steps": 1363, + "loss": 2.9988, + "learning_rate": 3.3138918345705194e-07, + "epoch": 2.0604229607250755 + }, + { + "current_steps": 1364, + "loss": 3.1801, + "learning_rate": 3.3085896076352066e-07, + "epoch": 2.0619335347432024 + }, + { + "current_steps": 1365, + "loss": 2.504, + "learning_rate": 3.303287380699894e-07, + "epoch": 2.0634441087613293 + }, + { + "current_steps": 1366, + "loss": 2.4585, + "learning_rate": 3.297985153764581e-07, + "epoch": 2.0649546827794563 + }, + { + "current_steps": 1367, + "loss": 1.8878, + "learning_rate": 3.292682926829268e-07, + "epoch": 2.066465256797583 + }, + { + "current_steps": 1368, + "loss": 2.8641, + "learning_rate": 3.287380699893955e-07, + "epoch": 2.06797583081571 + }, + { + "current_steps": 1369, + "loss": 2.9, + "learning_rate": 3.2820784729586424e-07, + "epoch": 2.069486404833837 + }, + { + "current_steps": 1370, + "loss": 3.0279, + "learning_rate": 3.2767762460233295e-07, + "epoch": 2.0709969788519635 + }, + { + "current_steps": 1371, + "loss": 3.0085, + "learning_rate": 3.2714740190880167e-07, + "epoch": 2.0725075528700905 + }, + { + "current_steps": 1372, + "loss": 2.8525, + "learning_rate": 3.266171792152704e-07, + "epoch": 2.0740181268882174 + }, + { + "current_steps": 1373, + "loss": 2.9782, + "learning_rate": 3.260869565217391e-07, + "epoch": 2.0755287009063443 + }, + { + "current_steps": 1374, + "loss": 2.4857, + "learning_rate": 3.255567338282078e-07, + "epoch": 2.0770392749244713 + }, + { + "current_steps": 1375, + "loss": 2.9204, + "learning_rate": 3.2502651113467653e-07, + "epoch": 2.078549848942598 + }, + { + "current_steps": 1376, + "loss": 3.0433, + "learning_rate": 3.244962884411453e-07, + "epoch": 2.080060422960725 + }, + { + "current_steps": 1377, + "loss": 2.9001, + "learning_rate": 3.2396606574761397e-07, + "epoch": 2.081570996978852 + }, + { + "current_steps": 1378, + "loss": 2.8464, + "learning_rate": 3.2343584305408273e-07, + "epoch": 2.083081570996979 + }, + { + "current_steps": 1379, + "loss": 2.8298, + "learning_rate": 3.229056203605514e-07, + "epoch": 2.084592145015106 + }, + { + "current_steps": 1380, + "loss": 2.9859, + "learning_rate": 3.2237539766702017e-07, + "epoch": 2.0861027190332324 + }, + { + "current_steps": 1381, + "loss": 2.3218, + "learning_rate": 3.2184517497348883e-07, + "epoch": 2.0876132930513593 + }, + { + "current_steps": 1382, + "loss": 2.9048, + "learning_rate": 3.213149522799576e-07, + "epoch": 2.0891238670694863 + }, + { + "current_steps": 1383, + "loss": 2.7674, + "learning_rate": 3.2078472958642626e-07, + "epoch": 2.090634441087613 + }, + { + "current_steps": 1384, + "loss": 3.087, + "learning_rate": 3.2025450689289503e-07, + "epoch": 2.09214501510574 + }, + { + "current_steps": 1385, + "loss": 2.9192, + "learning_rate": 3.197242841993637e-07, + "epoch": 2.093655589123867 + }, + { + "current_steps": 1386, + "loss": 3.0376, + "learning_rate": 3.1919406150583246e-07, + "epoch": 2.095166163141994 + }, + { + "current_steps": 1387, + "loss": 2.79, + "learning_rate": 3.186638388123011e-07, + "epoch": 2.096676737160121 + }, + { + "current_steps": 1388, + "loss": 2.8143, + "learning_rate": 3.181336161187699e-07, + "epoch": 2.098187311178248 + }, + { + "current_steps": 1389, + "loss": 3.077, + "learning_rate": 3.1760339342523856e-07, + "epoch": 2.099697885196375 + }, + { + "current_steps": 1390, + "loss": 2.8832, + "learning_rate": 3.170731707317073e-07, + "epoch": 2.1012084592145017 + }, + { + "current_steps": 1391, + "loss": 3.012, + "learning_rate": 3.16542948038176e-07, + "epoch": 2.102719033232628 + }, + { + "current_steps": 1392, + "loss": 3.0334, + "learning_rate": 3.1601272534464476e-07, + "epoch": 2.104229607250755 + }, + { + "current_steps": 1393, + "loss": 3.0294, + "learning_rate": 3.1548250265111347e-07, + "epoch": 2.105740181268882 + }, + { + "current_steps": 1394, + "loss": 3.0511, + "learning_rate": 3.149522799575822e-07, + "epoch": 2.107250755287009 + }, + { + "current_steps": 1395, + "loss": 2.4776, + "learning_rate": 3.144220572640509e-07, + "epoch": 2.108761329305136 + }, + { + "current_steps": 1396, + "loss": 3.2582, + "learning_rate": 3.138918345705196e-07, + "epoch": 2.110271903323263 + }, + { + "current_steps": 1397, + "loss": 2.9826, + "learning_rate": 3.1336161187698834e-07, + "epoch": 2.11178247734139 + }, + { + "current_steps": 1398, + "loss": 3.002, + "learning_rate": 3.1283138918345705e-07, + "epoch": 2.1132930513595167 + }, + { + "current_steps": 1399, + "loss": 2.7149, + "learning_rate": 3.1230116648992577e-07, + "epoch": 2.1148036253776437 + }, + { + "current_steps": 1400, + "loss": 2.768, + "learning_rate": 3.117709437963945e-07, + "epoch": 2.1163141993957706 + }, + { + "current_steps": 1401, + "loss": 3.1165, + "learning_rate": 3.112407211028632e-07, + "epoch": 2.117824773413897 + }, + { + "current_steps": 1402, + "loss": 2.9925, + "learning_rate": 3.107104984093319e-07, + "epoch": 2.119335347432024 + }, + { + "current_steps": 1403, + "loss": 2.8689, + "learning_rate": 3.1018027571580063e-07, + "epoch": 2.120845921450151 + }, + { + "current_steps": 1404, + "loss": 3.1892, + "learning_rate": 3.096500530222693e-07, + "epoch": 2.122356495468278 + }, + { + "current_steps": 1405, + "loss": 3.1442, + "learning_rate": 3.0911983032873806e-07, + "epoch": 2.123867069486405 + }, + { + "current_steps": 1406, + "loss": 2.429, + "learning_rate": 3.085896076352067e-07, + "epoch": 2.1253776435045317 + }, + { + "current_steps": 1407, + "loss": 3.0257, + "learning_rate": 3.080593849416755e-07, + "epoch": 2.1268882175226587 + }, + { + "current_steps": 1408, + "loss": 2.8809, + "learning_rate": 3.0752916224814416e-07, + "epoch": 2.1283987915407856 + }, + { + "current_steps": 1409, + "loss": 2.9666, + "learning_rate": 3.069989395546129e-07, + "epoch": 2.1299093655589125 + }, + { + "current_steps": 1410, + "loss": 2.8281, + "learning_rate": 3.0646871686108164e-07, + "epoch": 2.1314199395770395 + }, + { + "current_steps": 1411, + "loss": 3.168, + "learning_rate": 3.0593849416755036e-07, + "epoch": 2.1329305135951664 + }, + { + "current_steps": 1412, + "loss": 2.8572, + "learning_rate": 3.054082714740191e-07, + "epoch": 2.134441087613293 + }, + { + "current_steps": 1413, + "loss": 3.003, + "learning_rate": 3.048780487804878e-07, + "epoch": 2.13595166163142 + }, + { + "current_steps": 1414, + "loss": 2.77, + "learning_rate": 3.043478260869565e-07, + "epoch": 2.1374622356495467 + }, + { + "current_steps": 1415, + "loss": 3.0762, + "learning_rate": 3.038176033934252e-07, + "epoch": 2.1389728096676737 + }, + { + "current_steps": 1416, + "loss": 3.132, + "learning_rate": 3.0328738069989394e-07, + "epoch": 2.1404833836858006 + }, + { + "current_steps": 1417, + "loss": 3.0142, + "learning_rate": 3.0275715800636265e-07, + "epoch": 2.1419939577039275 + }, + { + "current_steps": 1418, + "loss": 3.117, + "learning_rate": 3.0222693531283137e-07, + "epoch": 2.1435045317220545 + }, + { + "current_steps": 1419, + "loss": 2.8927, + "learning_rate": 3.016967126193001e-07, + "epoch": 2.1450151057401814 + }, + { + "current_steps": 1420, + "loss": 2.9227, + "learning_rate": 3.011664899257688e-07, + "epoch": 2.1465256797583083 + }, + { + "current_steps": 1421, + "loss": 2.9973, + "learning_rate": 3.006362672322375e-07, + "epoch": 2.148036253776435 + }, + { + "current_steps": 1422, + "loss": 2.8912, + "learning_rate": 3.0010604453870623e-07, + "epoch": 2.1495468277945617 + }, + { + "current_steps": 1423, + "loss": 2.9818, + "learning_rate": 2.9957582184517495e-07, + "epoch": 2.1510574018126887 + }, + { + "current_steps": 1424, + "loss": 2.4071, + "learning_rate": 2.9904559915164366e-07, + "epoch": 2.1525679758308156 + }, + { + "current_steps": 1425, + "loss": 2.8711, + "learning_rate": 2.985153764581124e-07, + "epoch": 2.1540785498489425 + }, + { + "current_steps": 1426, + "loss": 3.1843, + "learning_rate": 2.979851537645811e-07, + "epoch": 2.1555891238670695 + }, + { + "current_steps": 1427, + "loss": 3.0332, + "learning_rate": 2.9745493107104986e-07, + "epoch": 2.1570996978851964 + }, + { + "current_steps": 1428, + "loss": 2.981, + "learning_rate": 2.9692470837751853e-07, + "epoch": 2.1586102719033233 + }, + { + "current_steps": 1429, + "loss": 3.0876, + "learning_rate": 2.963944856839873e-07, + "epoch": 2.1601208459214503 + }, + { + "current_steps": 1430, + "loss": 3.1866, + "learning_rate": 2.9586426299045596e-07, + "epoch": 2.161631419939577 + }, + { + "current_steps": 1431, + "loss": 2.9309, + "learning_rate": 2.9533404029692473e-07, + "epoch": 2.163141993957704 + }, + { + "current_steps": 1432, + "loss": 3.0347, + "learning_rate": 2.948038176033934e-07, + "epoch": 2.1646525679758306 + }, + { + "current_steps": 1433, + "loss": 2.9721, + "learning_rate": 2.9427359490986216e-07, + "epoch": 2.1661631419939575 + }, + { + "current_steps": 1434, + "loss": 3.0952, + "learning_rate": 2.937433722163308e-07, + "epoch": 2.1676737160120845 + }, + { + "current_steps": 1435, + "loss": 3.0751, + "learning_rate": 2.932131495227996e-07, + "epoch": 2.1691842900302114 + }, + { + "current_steps": 1436, + "loss": 2.927, + "learning_rate": 2.9268292682926825e-07, + "epoch": 2.1706948640483383 + }, + { + "current_steps": 1437, + "loss": 2.5054, + "learning_rate": 2.92152704135737e-07, + "epoch": 2.1722054380664653 + }, + { + "current_steps": 1438, + "loss": 3.0626, + "learning_rate": 2.916224814422057e-07, + "epoch": 2.173716012084592 + }, + { + "current_steps": 1439, + "loss": 2.7461, + "learning_rate": 2.9109225874867446e-07, + "epoch": 2.175226586102719 + }, + { + "current_steps": 1440, + "loss": 2.9681, + "learning_rate": 2.905620360551431e-07, + "epoch": 2.176737160120846 + }, + { + "current_steps": 1441, + "loss": 2.8997, + "learning_rate": 2.900318133616119e-07, + "epoch": 2.178247734138973 + }, + { + "current_steps": 1442, + "loss": 3.1472, + "learning_rate": 2.8950159066808055e-07, + "epoch": 2.1797583081570995 + }, + { + "current_steps": 1443, + "loss": 2.9166, + "learning_rate": 2.889713679745493e-07, + "epoch": 2.1812688821752264 + }, + { + "current_steps": 1444, + "loss": 3.0052, + "learning_rate": 2.8844114528101803e-07, + "epoch": 2.1827794561933533 + }, + { + "current_steps": 1445, + "loss": 3.1859, + "learning_rate": 2.8791092258748675e-07, + "epoch": 2.1842900302114803 + }, + { + "current_steps": 1446, + "loss": 2.9732, + "learning_rate": 2.8738069989395547e-07, + "epoch": 2.185800604229607 + }, + { + "current_steps": 1447, + "loss": 2.9773, + "learning_rate": 2.868504772004242e-07, + "epoch": 2.187311178247734 + }, + { + "current_steps": 1448, + "loss": 2.9693, + "learning_rate": 2.863202545068929e-07, + "epoch": 2.188821752265861 + }, + { + "current_steps": 1449, + "loss": 2.9008, + "learning_rate": 2.8579003181336156e-07, + "epoch": 2.190332326283988 + }, + { + "current_steps": 1450, + "loss": 3.093, + "learning_rate": 2.8525980911983033e-07, + "epoch": 2.191842900302115 + }, + { + "current_steps": 1451, + "loss": 2.838, + "learning_rate": 2.84729586426299e-07, + "epoch": 2.193353474320242 + }, + { + "current_steps": 1452, + "loss": 3.1251, + "learning_rate": 2.8419936373276776e-07, + "epoch": 2.194864048338369 + }, + { + "current_steps": 1453, + "loss": 3.1252, + "learning_rate": 2.836691410392364e-07, + "epoch": 2.1963746223564953 + }, + { + "current_steps": 1454, + "loss": 2.5353, + "learning_rate": 2.831389183457052e-07, + "epoch": 2.197885196374622 + }, + { + "current_steps": 1455, + "loss": 2.911, + "learning_rate": 2.8260869565217386e-07, + "epoch": 2.199395770392749 + }, + { + "current_steps": 1456, + "loss": 3.0334, + "learning_rate": 2.820784729586426e-07, + "epoch": 2.200906344410876 + }, + { + "current_steps": 1457, + "loss": 3.0871, + "learning_rate": 2.815482502651113e-07, + "epoch": 2.202416918429003 + }, + { + "current_steps": 1458, + "loss": 3.115, + "learning_rate": 2.8101802757158006e-07, + "epoch": 2.20392749244713 + }, + { + "current_steps": 1459, + "loss": 3.124, + "learning_rate": 2.8048780487804877e-07, + "epoch": 2.205438066465257 + }, + { + "current_steps": 1460, + "loss": 3.0949, + "learning_rate": 2.799575821845175e-07, + "epoch": 2.206948640483384 + }, + { + "current_steps": 1461, + "loss": 2.8581, + "learning_rate": 2.794273594909862e-07, + "epoch": 2.2084592145015107 + }, + { + "current_steps": 1462, + "loss": 2.9666, + "learning_rate": 2.788971367974549e-07, + "epoch": 2.2099697885196377 + }, + { + "current_steps": 1463, + "loss": 3.033, + "learning_rate": 2.7836691410392364e-07, + "epoch": 2.211480362537764 + }, + { + "current_steps": 1464, + "loss": 2.8995, + "learning_rate": 2.7783669141039235e-07, + "epoch": 2.212990936555891 + }, + { + "current_steps": 1465, + "loss": 3.0165, + "learning_rate": 2.7730646871686107e-07, + "epoch": 2.214501510574018 + }, + { + "current_steps": 1466, + "loss": 2.4119, + "learning_rate": 2.767762460233298e-07, + "epoch": 2.216012084592145 + }, + { + "current_steps": 1467, + "loss": 2.6998, + "learning_rate": 2.762460233297985e-07, + "epoch": 2.217522658610272 + }, + { + "current_steps": 1468, + "loss": 2.9498, + "learning_rate": 2.757158006362672e-07, + "epoch": 2.219033232628399 + }, + { + "current_steps": 1469, + "loss": 2.9901, + "learning_rate": 2.7518557794273593e-07, + "epoch": 2.2205438066465257 + }, + { + "current_steps": 1470, + "loss": 3.0045, + "learning_rate": 2.7465535524920465e-07, + "epoch": 2.2220543806646527 + }, + { + "current_steps": 1471, + "loss": 2.9205, + "learning_rate": 2.7412513255567336e-07, + "epoch": 2.2235649546827796 + }, + { + "current_steps": 1472, + "loss": 2.7696, + "learning_rate": 2.735949098621421e-07, + "epoch": 2.2250755287009065 + }, + { + "current_steps": 1473, + "loss": 3.2643, + "learning_rate": 2.730646871686108e-07, + "epoch": 2.2265861027190335 + }, + { + "current_steps": 1474, + "loss": 2.905, + "learning_rate": 2.725344644750795e-07, + "epoch": 2.22809667673716 + }, + { + "current_steps": 1475, + "loss": 2.9568, + "learning_rate": 2.7200424178154823e-07, + "epoch": 2.229607250755287 + }, + { + "current_steps": 1476, + "loss": 2.9691, + "learning_rate": 2.71474019088017e-07, + "epoch": 2.231117824773414 + }, + { + "current_steps": 1477, + "loss": 3.0642, + "learning_rate": 2.7094379639448566e-07, + "epoch": 2.2326283987915407 + }, + { + "current_steps": 1478, + "loss": 2.6817, + "learning_rate": 2.7041357370095443e-07, + "epoch": 2.2341389728096677 + }, + { + "current_steps": 1479, + "loss": 2.8455, + "learning_rate": 2.698833510074231e-07, + "epoch": 2.2356495468277946 + }, + { + "current_steps": 1480, + "loss": 2.7563, + "learning_rate": 2.6935312831389186e-07, + "epoch": 2.2371601208459215 + }, + { + "current_steps": 1481, + "loss": 3.0935, + "learning_rate": 2.688229056203605e-07, + "epoch": 2.2386706948640485 + }, + { + "current_steps": 1482, + "loss": 2.1801, + "learning_rate": 2.682926829268293e-07, + "epoch": 2.2401812688821754 + }, + { + "current_steps": 1483, + "loss": 3.0399, + "learning_rate": 2.6776246023329795e-07, + "epoch": 2.2416918429003023 + }, + { + "current_steps": 1484, + "loss": 2.7325, + "learning_rate": 2.672322375397667e-07, + "epoch": 2.243202416918429 + }, + { + "current_steps": 1485, + "loss": 2.9746, + "learning_rate": 2.667020148462354e-07, + "epoch": 2.2447129909365557 + }, + { + "current_steps": 1486, + "loss": 2.9142, + "learning_rate": 2.6617179215270415e-07, + "epoch": 2.2462235649546827 + }, + { + "current_steps": 1487, + "loss": 2.8804, + "learning_rate": 2.656415694591728e-07, + "epoch": 2.2477341389728096 + }, + { + "current_steps": 1488, + "loss": 3.2683, + "learning_rate": 2.651113467656416e-07, + "epoch": 2.2492447129909365 + }, + { + "current_steps": 1489, + "loss": 3.1025, + "learning_rate": 2.6458112407211025e-07, + "epoch": 2.2507552870090635 + }, + { + "current_steps": 1490, + "loss": 3.1795, + "learning_rate": 2.64050901378579e-07, + "epoch": 2.2522658610271904 + }, + { + "current_steps": 1491, + "loss": 2.7756, + "learning_rate": 2.635206786850477e-07, + "epoch": 2.2537764350453173 + }, + { + "current_steps": 1492, + "loss": 3.0307, + "learning_rate": 2.6299045599151645e-07, + "epoch": 2.2552870090634443 + }, + { + "current_steps": 1493, + "loss": 3.0719, + "learning_rate": 2.6246023329798517e-07, + "epoch": 2.256797583081571 + }, + { + "current_steps": 1494, + "loss": 3.0361, + "learning_rate": 2.6193001060445383e-07, + "epoch": 2.258308157099698 + }, + { + "current_steps": 1495, + "loss": 3.0995, + "learning_rate": 2.613997879109226e-07, + "epoch": 2.2598187311178246 + }, + { + "current_steps": 1496, + "loss": 2.9234, + "learning_rate": 2.6086956521739126e-07, + "epoch": 2.2613293051359515 + }, + { + "current_steps": 1497, + "loss": 2.8779, + "learning_rate": 2.6033934252386003e-07, + "epoch": 2.2628398791540785 + }, + { + "current_steps": 1498, + "loss": 2.9209, + "learning_rate": 2.598091198303287e-07, + "epoch": 2.2643504531722054 + }, + { + "current_steps": 1499, + "loss": 2.9894, + "learning_rate": 2.5927889713679746e-07, + "epoch": 2.2658610271903323 + }, + { + "current_steps": 1500, + "loss": 3.0221, + "learning_rate": 2.587486744432661e-07, + "epoch": 2.2673716012084593 + }, + { + "current_steps": 1501, + "loss": 2.7052, + "learning_rate": 2.582184517497349e-07, + "epoch": 2.268882175226586 + }, + { + "current_steps": 1502, + "loss": 2.7213, + "learning_rate": 2.5768822905620356e-07, + "epoch": 2.270392749244713 + }, + { + "current_steps": 1503, + "loss": 2.8471, + "learning_rate": 2.571580063626723e-07, + "epoch": 2.27190332326284 + }, + { + "current_steps": 1504, + "loss": 2.9732, + "learning_rate": 2.56627783669141e-07, + "epoch": 2.2734138972809665 + }, + { + "current_steps": 1505, + "loss": 2.8034, + "learning_rate": 2.5609756097560976e-07, + "epoch": 2.2749244712990935 + }, + { + "current_steps": 1506, + "loss": 3.2221, + "learning_rate": 2.555673382820784e-07, + "epoch": 2.2764350453172204 + }, + { + "current_steps": 1507, + "loss": 2.816, + "learning_rate": 2.550371155885472e-07, + "epoch": 2.2779456193353473 + }, + { + "current_steps": 1508, + "loss": 2.9744, + "learning_rate": 2.5450689289501585e-07, + "epoch": 2.2794561933534743 + }, + { + "current_steps": 1509, + "loss": 3.0932, + "learning_rate": 2.539766702014846e-07, + "epoch": 2.280966767371601 + }, + { + "current_steps": 1510, + "loss": 2.7198, + "learning_rate": 2.5344644750795334e-07, + "epoch": 2.282477341389728 + }, + { + "current_steps": 1511, + "loss": 3.0977, + "learning_rate": 2.5291622481442205e-07, + "epoch": 2.283987915407855 + }, + { + "current_steps": 1512, + "loss": 2.8511, + "learning_rate": 2.5238600212089077e-07, + "epoch": 2.285498489425982 + }, + { + "current_steps": 1513, + "loss": 3.0167, + "learning_rate": 2.518557794273595e-07, + "epoch": 2.287009063444109 + }, + { + "current_steps": 1514, + "loss": 3.0085, + "learning_rate": 2.513255567338282e-07, + "epoch": 2.288519637462236 + }, + { + "current_steps": 1515, + "loss": 3.1017, + "learning_rate": 2.507953340402969e-07, + "epoch": 2.290030211480363 + }, + { + "current_steps": 1516, + "loss": 3.0252, + "learning_rate": 2.5026511134676563e-07, + "epoch": 2.2915407854984893 + }, + { + "current_steps": 1517, + "loss": 2.811, + "learning_rate": 2.4973488865323435e-07, + "epoch": 2.293051359516616 + }, + { + "current_steps": 1518, + "loss": 2.7579, + "learning_rate": 2.4920466595970306e-07, + "epoch": 2.294561933534743 + }, + { + "current_steps": 1519, + "loss": 2.888, + "learning_rate": 2.486744432661718e-07, + "epoch": 2.29607250755287 + }, + { + "current_steps": 1520, + "loss": 2.9921, + "learning_rate": 2.481442205726405e-07, + "epoch": 2.297583081570997 + }, + { + "current_steps": 1521, + "loss": 2.9381, + "learning_rate": 2.476139978791092e-07, + "epoch": 2.299093655589124 + }, + { + "current_steps": 1522, + "loss": 3.2375, + "learning_rate": 2.470837751855779e-07, + "epoch": 2.300604229607251 + }, + { + "current_steps": 1523, + "loss": 2.9129, + "learning_rate": 2.4655355249204664e-07, + "epoch": 2.302114803625378 + }, + { + "current_steps": 1524, + "loss": 2.9648, + "learning_rate": 2.4602332979851536e-07, + "epoch": 2.3036253776435047 + }, + { + "current_steps": 1525, + "loss": 2.9395, + "learning_rate": 2.454931071049841e-07, + "epoch": 2.305135951661631 + }, + { + "current_steps": 1526, + "loss": 2.9632, + "learning_rate": 2.449628844114528e-07, + "epoch": 2.306646525679758 + }, + { + "current_steps": 1527, + "loss": 3.0419, + "learning_rate": 2.444326617179215e-07, + "epoch": 2.308157099697885 + }, + { + "current_steps": 1528, + "loss": 2.7947, + "learning_rate": 2.439024390243902e-07, + "epoch": 2.309667673716012 + }, + { + "current_steps": 1529, + "loss": 3.0674, + "learning_rate": 2.4337221633085894e-07, + "epoch": 2.311178247734139 + }, + { + "current_steps": 1530, + "loss": 2.9957, + "learning_rate": 2.4284199363732765e-07, + "epoch": 2.312688821752266 + }, + { + "current_steps": 1531, + "loss": 2.9322, + "learning_rate": 2.4231177094379637e-07, + "epoch": 2.314199395770393 + }, + { + "current_steps": 1532, + "loss": 2.9148, + "learning_rate": 2.417815482502651e-07, + "epoch": 2.3157099697885197 + }, + { + "current_steps": 1533, + "loss": 2.9209, + "learning_rate": 2.412513255567338e-07, + "epoch": 2.3172205438066467 + }, + { + "current_steps": 1534, + "loss": 3.0346, + "learning_rate": 2.4072110286320257e-07, + "epoch": 2.3187311178247736 + }, + { + "current_steps": 1535, + "loss": 3.0656, + "learning_rate": 2.401908801696713e-07, + "epoch": 2.3202416918429005 + }, + { + "current_steps": 1536, + "loss": 3.0893, + "learning_rate": 2.3966065747614e-07, + "epoch": 2.321752265861027 + }, + { + "current_steps": 1537, + "loss": 3.0211, + "learning_rate": 2.391304347826087e-07, + "epoch": 2.323262839879154 + }, + { + "current_steps": 1538, + "loss": 2.9478, + "learning_rate": 2.3860021208907743e-07, + "epoch": 2.324773413897281 + }, + { + "current_steps": 1539, + "loss": 2.8997, + "learning_rate": 2.3806998939554612e-07, + "epoch": 2.326283987915408 + }, + { + "current_steps": 1540, + "loss": 2.598, + "learning_rate": 2.3753976670201484e-07, + "epoch": 2.3277945619335347 + }, + { + "current_steps": 1541, + "loss": 2.798, + "learning_rate": 2.3700954400848355e-07, + "epoch": 2.3293051359516617 + }, + { + "current_steps": 1542, + "loss": 2.9423, + "learning_rate": 2.3647932131495227e-07, + "epoch": 2.3308157099697886 + }, + { + "current_steps": 1543, + "loss": 3.0865, + "learning_rate": 2.3594909862142099e-07, + "epoch": 2.3323262839879155 + }, + { + "current_steps": 1544, + "loss": 2.6892, + "learning_rate": 2.354188759278897e-07, + "epoch": 2.3338368580060425 + }, + { + "current_steps": 1545, + "loss": 3.0822, + "learning_rate": 2.3488865323435842e-07, + "epoch": 2.335347432024169 + }, + { + "current_steps": 1546, + "loss": 3.1105, + "learning_rate": 2.3435843054082713e-07, + "epoch": 2.336858006042296 + }, + { + "current_steps": 1547, + "loss": 3.0123, + "learning_rate": 2.3382820784729585e-07, + "epoch": 2.338368580060423 + }, + { + "current_steps": 1548, + "loss": 2.8546, + "learning_rate": 2.3329798515376457e-07, + "epoch": 2.3398791540785497 + }, + { + "current_steps": 1549, + "loss": 2.9216, + "learning_rate": 2.3276776246023328e-07, + "epoch": 2.3413897280966767 + }, + { + "current_steps": 1550, + "loss": 2.9761, + "learning_rate": 2.3223753976670202e-07, + "epoch": 2.3429003021148036 + }, + { + "current_steps": 1551, + "loss": 2.7588, + "learning_rate": 2.3170731707317074e-07, + "epoch": 2.3444108761329305 + }, + { + "current_steps": 1552, + "loss": 3.024, + "learning_rate": 2.3117709437963946e-07, + "epoch": 2.3459214501510575 + }, + { + "current_steps": 1553, + "loss": 2.9984, + "learning_rate": 2.3064687168610817e-07, + "epoch": 2.3474320241691844 + }, + { + "current_steps": 1554, + "loss": 3.0442, + "learning_rate": 2.301166489925769e-07, + "epoch": 2.3489425981873113 + }, + { + "current_steps": 1555, + "loss": 2.9495, + "learning_rate": 2.295864262990456e-07, + "epoch": 2.3504531722054383 + }, + { + "current_steps": 1556, + "loss": 3.0512, + "learning_rate": 2.2905620360551432e-07, + "epoch": 2.351963746223565 + }, + { + "current_steps": 1557, + "loss": 3.0835, + "learning_rate": 2.2852598091198303e-07, + "epoch": 2.3534743202416917 + }, + { + "current_steps": 1558, + "loss": 2.2911, + "learning_rate": 2.2799575821845175e-07, + "epoch": 2.3549848942598186 + }, + { + "current_steps": 1559, + "loss": 3.0784, + "learning_rate": 2.2746553552492047e-07, + "epoch": 2.3564954682779455 + }, + { + "current_steps": 1560, + "loss": 3.0934, + "learning_rate": 2.2693531283138918e-07, + "epoch": 2.3580060422960725 + }, + { + "current_steps": 1561, + "loss": 2.499, + "learning_rate": 2.2640509013785787e-07, + "epoch": 2.3595166163141994 + }, + { + "current_steps": 1562, + "loss": 3.2606, + "learning_rate": 2.258748674443266e-07, + "epoch": 2.3610271903323263 + }, + { + "current_steps": 1563, + "loss": 2.9355, + "learning_rate": 2.253446447507953e-07, + "epoch": 2.3625377643504533 + }, + { + "current_steps": 1564, + "loss": 3.0685, + "learning_rate": 2.2481442205726402e-07, + "epoch": 2.36404833836858 + }, + { + "current_steps": 1565, + "loss": 2.87, + "learning_rate": 2.2428419936373274e-07, + "epoch": 2.365558912386707 + }, + { + "current_steps": 1566, + "loss": 2.9643, + "learning_rate": 2.2375397667020145e-07, + "epoch": 2.3670694864048336 + }, + { + "current_steps": 1567, + "loss": 2.3665, + "learning_rate": 2.232237539766702e-07, + "epoch": 2.3685800604229605 + }, + { + "current_steps": 1568, + "loss": 3.0814, + "learning_rate": 2.226935312831389e-07, + "epoch": 2.3700906344410875 + }, + { + "current_steps": 1569, + "loss": 2.9984, + "learning_rate": 2.2216330858960763e-07, + "epoch": 2.3716012084592144 + }, + { + "current_steps": 1570, + "loss": 2.8213, + "learning_rate": 2.2163308589607634e-07, + "epoch": 2.3731117824773413 + }, + { + "current_steps": 1571, + "loss": 2.8461, + "learning_rate": 2.2110286320254506e-07, + "epoch": 2.3746223564954683 + }, + { + "current_steps": 1572, + "loss": 3.0246, + "learning_rate": 2.2057264050901377e-07, + "epoch": 2.376132930513595 + }, + { + "current_steps": 1573, + "loss": 2.9836, + "learning_rate": 2.200424178154825e-07, + "epoch": 2.377643504531722 + }, + { + "current_steps": 1574, + "loss": 3.2213, + "learning_rate": 2.195121951219512e-07, + "epoch": 2.379154078549849 + }, + { + "current_steps": 1575, + "loss": 2.9788, + "learning_rate": 2.1898197242841992e-07, + "epoch": 2.380664652567976 + }, + { + "current_steps": 1576, + "loss": 3.2071, + "learning_rate": 2.1845174973488864e-07, + "epoch": 2.382175226586103 + }, + { + "current_steps": 1577, + "loss": 2.9546, + "learning_rate": 2.1792152704135735e-07, + "epoch": 2.38368580060423 + }, + { + "current_steps": 1578, + "loss": 2.5353, + "learning_rate": 2.1739130434782607e-07, + "epoch": 2.3851963746223563 + }, + { + "current_steps": 1579, + "loss": 3.0954, + "learning_rate": 2.1686108165429478e-07, + "epoch": 2.3867069486404833 + }, + { + "current_steps": 1580, + "loss": 3.0557, + "learning_rate": 2.163308589607635e-07, + "epoch": 2.38821752265861 + }, + { + "current_steps": 1581, + "loss": 2.9073, + "learning_rate": 2.1580063626723222e-07, + "epoch": 2.389728096676737 + }, + { + "current_steps": 1582, + "loss": 3.1207, + "learning_rate": 2.1527041357370093e-07, + "epoch": 2.391238670694864 + }, + { + "current_steps": 1583, + "loss": 3.0087, + "learning_rate": 2.1474019088016967e-07, + "epoch": 2.392749244712991 + }, + { + "current_steps": 1584, + "loss": 2.9414, + "learning_rate": 2.142099681866384e-07, + "epoch": 2.394259818731118 + }, + { + "current_steps": 1585, + "loss": 2.7435, + "learning_rate": 2.136797454931071e-07, + "epoch": 2.395770392749245 + }, + { + "current_steps": 1586, + "loss": 3.0774, + "learning_rate": 2.1314952279957582e-07, + "epoch": 2.397280966767372 + }, + { + "current_steps": 1587, + "loss": 2.927, + "learning_rate": 2.1261930010604454e-07, + "epoch": 2.3987915407854983 + }, + { + "current_steps": 1588, + "loss": 2.8824, + "learning_rate": 2.1208907741251325e-07, + "epoch": 2.400302114803625 + }, + { + "current_steps": 1589, + "loss": 2.8408, + "learning_rate": 2.1155885471898197e-07, + "epoch": 2.401812688821752 + }, + { + "current_steps": 1590, + "loss": 2.9424, + "learning_rate": 2.1102863202545068e-07, + "epoch": 2.403323262839879 + }, + { + "current_steps": 1591, + "loss": 2.9665, + "learning_rate": 2.104984093319194e-07, + "epoch": 2.404833836858006 + }, + { + "current_steps": 1592, + "loss": 2.9924, + "learning_rate": 2.0996818663838812e-07, + "epoch": 2.406344410876133 + }, + { + "current_steps": 1593, + "loss": 2.7694, + "learning_rate": 2.0943796394485683e-07, + "epoch": 2.40785498489426 + }, + { + "current_steps": 1594, + "loss": 2.8801, + "learning_rate": 2.0890774125132555e-07, + "epoch": 2.409365558912387 + }, + { + "current_steps": 1595, + "loss": 3.0329, + "learning_rate": 2.0837751855779426e-07, + "epoch": 2.4108761329305137 + }, + { + "current_steps": 1596, + "loss": 2.5639, + "learning_rate": 2.0784729586426298e-07, + "epoch": 2.4123867069486407 + }, + { + "current_steps": 1597, + "loss": 3.0518, + "learning_rate": 2.073170731707317e-07, + "epoch": 2.4138972809667676 + }, + { + "current_steps": 1598, + "loss": 2.9236, + "learning_rate": 2.067868504772004e-07, + "epoch": 2.415407854984894 + }, + { + "current_steps": 1599, + "loss": 2.8624, + "learning_rate": 2.0625662778366913e-07, + "epoch": 2.416918429003021 + }, + { + "current_steps": 1600, + "loss": 2.9981, + "learning_rate": 2.0572640509013787e-07, + "epoch": 2.418429003021148 + }, + { + "current_steps": 1601, + "loss": 2.8338, + "learning_rate": 2.0519618239660659e-07, + "epoch": 2.419939577039275 + }, + { + "current_steps": 1602, + "loss": 3.1091, + "learning_rate": 2.046659597030753e-07, + "epoch": 2.421450151057402 + }, + { + "current_steps": 1603, + "loss": 2.9502, + "learning_rate": 2.0413573700954402e-07, + "epoch": 2.4229607250755287 + }, + { + "current_steps": 1604, + "loss": 2.9127, + "learning_rate": 2.0360551431601273e-07, + "epoch": 2.4244712990936557 + }, + { + "current_steps": 1605, + "loss": 3.1176, + "learning_rate": 2.0307529162248145e-07, + "epoch": 2.4259818731117826 + }, + { + "current_steps": 1606, + "loss": 3.1529, + "learning_rate": 2.0254506892895017e-07, + "epoch": 2.4274924471299095 + }, + { + "current_steps": 1607, + "loss": 2.9226, + "learning_rate": 2.0201484623541885e-07, + "epoch": 2.429003021148036 + }, + { + "current_steps": 1608, + "loss": 2.9755, + "learning_rate": 2.0148462354188757e-07, + "epoch": 2.430513595166163 + }, + { + "current_steps": 1609, + "loss": 2.3473, + "learning_rate": 2.0095440084835629e-07, + "epoch": 2.43202416918429 + }, + { + "current_steps": 1610, + "loss": 3.0144, + "learning_rate": 2.00424178154825e-07, + "epoch": 2.433534743202417 + }, + { + "current_steps": 1611, + "loss": 3.0732, + "learning_rate": 1.9989395546129372e-07, + "epoch": 2.4350453172205437 + }, + { + "current_steps": 1612, + "loss": 2.9595, + "learning_rate": 1.9936373276776243e-07, + "epoch": 2.4365558912386707 + }, + { + "current_steps": 1613, + "loss": 2.4932, + "learning_rate": 1.9883351007423115e-07, + "epoch": 2.4380664652567976 + }, + { + "current_steps": 1614, + "loss": 3.1989, + "learning_rate": 1.9830328738069987e-07, + "epoch": 2.4395770392749245 + }, + { + "current_steps": 1615, + "loss": 2.9832, + "learning_rate": 1.9777306468716858e-07, + "epoch": 2.4410876132930515 + }, + { + "current_steps": 1616, + "loss": 3.0117, + "learning_rate": 1.9724284199363732e-07, + "epoch": 2.4425981873111784 + }, + { + "current_steps": 1617, + "loss": 2.25, + "learning_rate": 1.9671261930010604e-07, + "epoch": 2.4441087613293053 + }, + { + "current_steps": 1618, + "loss": 3.0307, + "learning_rate": 1.9618239660657476e-07, + "epoch": 2.4456193353474323 + }, + { + "current_steps": 1619, + "loss": 2.8028, + "learning_rate": 1.9565217391304347e-07, + "epoch": 2.4471299093655587 + }, + { + "current_steps": 1620, + "loss": 2.8706, + "learning_rate": 1.951219512195122e-07, + "epoch": 2.4486404833836857 + }, + { + "current_steps": 1621, + "loss": 3.0146, + "learning_rate": 1.945917285259809e-07, + "epoch": 2.4501510574018126 + }, + { + "current_steps": 1622, + "loss": 3.0197, + "learning_rate": 1.9406150583244962e-07, + "epoch": 2.4516616314199395 + }, + { + "current_steps": 1623, + "loss": 2.9147, + "learning_rate": 1.9353128313891834e-07, + "epoch": 2.4531722054380665 + }, + { + "current_steps": 1624, + "loss": 2.9364, + "learning_rate": 1.9300106044538705e-07, + "epoch": 2.4546827794561934 + }, + { + "current_steps": 1625, + "loss": 2.9045, + "learning_rate": 1.9247083775185577e-07, + "epoch": 2.4561933534743203 + }, + { + "current_steps": 1626, + "loss": 2.9614, + "learning_rate": 1.9194061505832448e-07, + "epoch": 2.4577039274924473 + }, + { + "current_steps": 1627, + "loss": 3.018, + "learning_rate": 1.914103923647932e-07, + "epoch": 2.459214501510574 + }, + { + "current_steps": 1628, + "loss": 2.9721, + "learning_rate": 1.9088016967126191e-07, + "epoch": 2.4607250755287007 + }, + { + "current_steps": 1629, + "loss": 2.3248, + "learning_rate": 1.9034994697773063e-07, + "epoch": 2.4622356495468276 + }, + { + "current_steps": 1630, + "loss": 2.765, + "learning_rate": 1.8981972428419935e-07, + "epoch": 2.4637462235649545 + }, + { + "current_steps": 1631, + "loss": 2.8288, + "learning_rate": 1.8928950159066806e-07, + "epoch": 2.4652567975830815 + }, + { + "current_steps": 1632, + "loss": 2.7947, + "learning_rate": 1.8875927889713678e-07, + "epoch": 2.4667673716012084 + }, + { + "current_steps": 1633, + "loss": 2.9891, + "learning_rate": 1.8822905620360552e-07, + "epoch": 2.4682779456193353 + }, + { + "current_steps": 1634, + "loss": 3.082, + "learning_rate": 1.8769883351007424e-07, + "epoch": 2.4697885196374623 + }, + { + "current_steps": 1635, + "loss": 2.9986, + "learning_rate": 1.8716861081654295e-07, + "epoch": 2.471299093655589 + }, + { + "current_steps": 1636, + "loss": 2.9933, + "learning_rate": 1.8663838812301167e-07, + "epoch": 2.472809667673716 + }, + { + "current_steps": 1637, + "loss": 3.0289, + "learning_rate": 1.8610816542948038e-07, + "epoch": 2.474320241691843 + }, + { + "current_steps": 1638, + "loss": 2.9133, + "learning_rate": 1.855779427359491e-07, + "epoch": 2.47583081570997 + }, + { + "current_steps": 1639, + "loss": 2.5725, + "learning_rate": 1.8504772004241782e-07, + "epoch": 2.477341389728097 + }, + { + "current_steps": 1640, + "loss": 3.0534, + "learning_rate": 1.8451749734888653e-07, + "epoch": 2.4788519637462234 + }, + { + "current_steps": 1641, + "loss": 2.9152, + "learning_rate": 1.8398727465535525e-07, + "epoch": 2.4803625377643503 + }, + { + "current_steps": 1642, + "loss": 3.1145, + "learning_rate": 1.8345705196182396e-07, + "epoch": 2.4818731117824773 + }, + { + "current_steps": 1643, + "loss": 3.0405, + "learning_rate": 1.8292682926829268e-07, + "epoch": 2.483383685800604 + }, + { + "current_steps": 1644, + "loss": 2.8501, + "learning_rate": 1.823966065747614e-07, + "epoch": 2.484894259818731 + }, + { + "current_steps": 1645, + "loss": 2.6631, + "learning_rate": 1.818663838812301e-07, + "epoch": 2.486404833836858 + }, + { + "current_steps": 1646, + "loss": 2.9754, + "learning_rate": 1.8133616118769883e-07, + "epoch": 2.487915407854985 + }, + { + "current_steps": 1647, + "loss": 2.7988, + "learning_rate": 1.8080593849416754e-07, + "epoch": 2.489425981873112 + }, + { + "current_steps": 1648, + "loss": 3.1064, + "learning_rate": 1.8027571580063626e-07, + "epoch": 2.490936555891239 + }, + { + "current_steps": 1649, + "loss": 2.9314, + "learning_rate": 1.79745493107105e-07, + "epoch": 2.4924471299093653 + }, + { + "current_steps": 1650, + "loss": 2.7982, + "learning_rate": 1.7921527041357372e-07, + "epoch": 2.4939577039274923 + }, + { + "current_steps": 1651, + "loss": 2.9565, + "learning_rate": 1.7868504772004243e-07, + "epoch": 2.495468277945619 + }, + { + "current_steps": 1652, + "loss": 2.4873, + "learning_rate": 1.7815482502651112e-07, + "epoch": 2.496978851963746 + }, + { + "current_steps": 1653, + "loss": 2.7168, + "learning_rate": 1.7762460233297984e-07, + "epoch": 2.498489425981873 + }, + { + "current_steps": 1654, + "loss": 2.9843, + "learning_rate": 1.7709437963944855e-07, + "epoch": 2.5 + }, + { + "current_steps": 1655, + "loss": 2.9685, + "learning_rate": 1.7656415694591727e-07, + "epoch": 2.501510574018127 + }, + { + "current_steps": 1656, + "loss": 2.3108, + "learning_rate": 1.7603393425238599e-07, + "epoch": 2.503021148036254 + }, + { + "current_steps": 1657, + "loss": 3.1197, + "learning_rate": 1.755037115588547e-07, + "epoch": 2.504531722054381 + }, + { + "current_steps": 1658, + "loss": 3.0199, + "learning_rate": 1.7497348886532342e-07, + "epoch": 2.5060422960725077 + }, + { + "current_steps": 1659, + "loss": 3.1118, + "learning_rate": 1.7444326617179213e-07, + "epoch": 2.5075528700906347 + }, + { + "current_steps": 1660, + "loss": 3.025, + "learning_rate": 1.7391304347826085e-07, + "epoch": 2.5090634441087616 + }, + { + "current_steps": 1661, + "loss": 2.8618, + "learning_rate": 1.7338282078472956e-07, + "epoch": 2.510574018126888 + }, + { + "current_steps": 1662, + "loss": 3.043, + "learning_rate": 1.7285259809119828e-07, + "epoch": 2.512084592145015 + }, + { + "current_steps": 1663, + "loss": 2.9807, + "learning_rate": 1.72322375397667e-07, + "epoch": 2.513595166163142 + }, + { + "current_steps": 1664, + "loss": 2.8322, + "learning_rate": 1.717921527041357e-07, + "epoch": 2.515105740181269 + }, + { + "current_steps": 1665, + "loss": 2.9425, + "learning_rate": 1.7126193001060443e-07, + "epoch": 2.516616314199396 + }, + { + "current_steps": 1666, + "loss": 2.9077, + "learning_rate": 1.7073170731707317e-07, + "epoch": 2.5181268882175227 + }, + { + "current_steps": 1667, + "loss": 3.0191, + "learning_rate": 1.7020148462354189e-07, + "epoch": 2.5196374622356497 + }, + { + "current_steps": 1668, + "loss": 2.9413, + "learning_rate": 1.696712619300106e-07, + "epoch": 2.5211480362537766 + }, + { + "current_steps": 1669, + "loss": 2.8285, + "learning_rate": 1.6914103923647932e-07, + "epoch": 2.522658610271903 + }, + { + "current_steps": 1670, + "loss": 2.8357, + "learning_rate": 1.6861081654294803e-07, + "epoch": 2.52416918429003 + }, + { + "current_steps": 1671, + "loss": 2.6727, + "learning_rate": 1.6808059384941675e-07, + "epoch": 2.525679758308157 + }, + { + "current_steps": 1672, + "loss": 3.0076, + "learning_rate": 1.6755037115588547e-07, + "epoch": 2.527190332326284 + }, + { + "current_steps": 1673, + "loss": 3.0603, + "learning_rate": 1.6702014846235418e-07, + "epoch": 2.528700906344411 + }, + { + "current_steps": 1674, + "loss": 2.9549, + "learning_rate": 1.664899257688229e-07, + "epoch": 2.5302114803625377 + }, + { + "current_steps": 1675, + "loss": 2.9827, + "learning_rate": 1.6595970307529161e-07, + "epoch": 2.5317220543806647 + }, + { + "current_steps": 1676, + "loss": 2.9776, + "learning_rate": 1.6542948038176033e-07, + "epoch": 2.5332326283987916 + }, + { + "current_steps": 1677, + "loss": 2.8078, + "learning_rate": 1.6489925768822905e-07, + "epoch": 2.5347432024169185 + }, + { + "current_steps": 1678, + "loss": 3.1163, + "learning_rate": 1.6436903499469776e-07, + "epoch": 2.5362537764350455 + }, + { + "current_steps": 1679, + "loss": 2.9042, + "learning_rate": 1.6383881230116648e-07, + "epoch": 2.5377643504531724 + }, + { + "current_steps": 1680, + "loss": 2.6071, + "learning_rate": 1.633085896076352e-07, + "epoch": 2.5392749244712993 + }, + { + "current_steps": 1681, + "loss": 2.921, + "learning_rate": 1.627783669141039e-07, + "epoch": 2.5407854984894263 + }, + { + "current_steps": 1682, + "loss": 3.0543, + "learning_rate": 1.6224814422057265e-07, + "epoch": 2.5422960725075527 + }, + { + "current_steps": 1683, + "loss": 2.9539, + "learning_rate": 1.6171792152704137e-07, + "epoch": 2.5438066465256797 + }, + { + "current_steps": 1684, + "loss": 3.1212, + "learning_rate": 1.6118769883351008e-07, + "epoch": 2.5453172205438066 + }, + { + "current_steps": 1685, + "loss": 2.934, + "learning_rate": 1.606574761399788e-07, + "epoch": 2.5468277945619335 + }, + { + "current_steps": 1686, + "loss": 2.9337, + "learning_rate": 1.6012725344644751e-07, + "epoch": 2.5483383685800605 + }, + { + "current_steps": 1687, + "loss": 2.9206, + "learning_rate": 1.5959703075291623e-07, + "epoch": 2.5498489425981874 + }, + { + "current_steps": 1688, + "loss": 2.8159, + "learning_rate": 1.5906680805938495e-07, + "epoch": 2.5513595166163143 + }, + { + "current_steps": 1689, + "loss": 2.7428, + "learning_rate": 1.5853658536585366e-07, + "epoch": 2.552870090634441 + }, + { + "current_steps": 1690, + "loss": 2.8544, + "learning_rate": 1.5800636267232238e-07, + "epoch": 2.5543806646525677 + }, + { + "current_steps": 1691, + "loss": 2.9312, + "learning_rate": 1.574761399787911e-07, + "epoch": 2.5558912386706947 + }, + { + "current_steps": 1692, + "loss": 3.0904, + "learning_rate": 1.569459172852598e-07, + "epoch": 2.5574018126888216 + }, + { + "current_steps": 1693, + "loss": 2.9235, + "learning_rate": 1.5641569459172853e-07, + "epoch": 2.5589123867069485 + }, + { + "current_steps": 1694, + "loss": 2.9325, + "learning_rate": 1.5588547189819724e-07, + "epoch": 2.5604229607250755 + }, + { + "current_steps": 1695, + "loss": 2.953, + "learning_rate": 1.5535524920466596e-07, + "epoch": 2.5619335347432024 + }, + { + "current_steps": 1696, + "loss": 2.8501, + "learning_rate": 1.5482502651113465e-07, + "epoch": 2.5634441087613293 + }, + { + "current_steps": 1697, + "loss": 3.1307, + "learning_rate": 1.5429480381760336e-07, + "epoch": 2.5649546827794563 + }, + { + "current_steps": 1698, + "loss": 3.072, + "learning_rate": 1.5376458112407208e-07, + "epoch": 2.566465256797583 + }, + { + "current_steps": 1699, + "loss": 2.9131, + "learning_rate": 1.5323435843054082e-07, + "epoch": 2.56797583081571 + }, + { + "current_steps": 1700, + "loss": 3.2172, + "learning_rate": 1.5270413573700954e-07, + "epoch": 2.569486404833837 + }, + { + "current_steps": 1701, + "loss": 2.9902, + "learning_rate": 1.5217391304347825e-07, + "epoch": 2.570996978851964 + }, + { + "current_steps": 1702, + "loss": 2.6912, + "learning_rate": 1.5164369034994697e-07, + "epoch": 2.5725075528700905 + }, + { + "current_steps": 1703, + "loss": 2.9426, + "learning_rate": 1.5111346765641568e-07, + "epoch": 2.5740181268882174 + }, + { + "current_steps": 1704, + "loss": 3.0671, + "learning_rate": 1.505832449628844e-07, + "epoch": 2.5755287009063443 + }, + { + "current_steps": 1705, + "loss": 2.9862, + "learning_rate": 1.5005302226935312e-07, + "epoch": 2.5770392749244713 + }, + { + "current_steps": 1706, + "loss": 2.9179, + "learning_rate": 1.4952279957582183e-07, + "epoch": 2.578549848942598 + }, + { + "current_steps": 1707, + "loss": 3.1907, + "learning_rate": 1.4899257688229055e-07, + "epoch": 2.580060422960725 + }, + { + "current_steps": 1708, + "loss": 2.9362, + "learning_rate": 1.4846235418875926e-07, + "epoch": 2.581570996978852 + }, + { + "current_steps": 1709, + "loss": 2.8519, + "learning_rate": 1.4793213149522798e-07, + "epoch": 2.583081570996979 + }, + { + "current_steps": 1710, + "loss": 3.1644, + "learning_rate": 1.474019088016967e-07, + "epoch": 2.5845921450151055 + }, + { + "current_steps": 1711, + "loss": 2.9998, + "learning_rate": 1.468716861081654e-07, + "epoch": 2.5861027190332324 + }, + { + "current_steps": 1712, + "loss": 2.8967, + "learning_rate": 1.4634146341463413e-07, + "epoch": 2.5876132930513593 + }, + { + "current_steps": 1713, + "loss": 3.0513, + "learning_rate": 1.4581124072110284e-07, + "epoch": 2.5891238670694863 + }, + { + "current_steps": 1714, + "loss": 2.9667, + "learning_rate": 1.4528101802757156e-07, + "epoch": 2.590634441087613 + }, + { + "current_steps": 1715, + "loss": 2.8657, + "learning_rate": 1.4475079533404028e-07, + "epoch": 2.59214501510574 + }, + { + "current_steps": 1716, + "loss": 2.8812, + "learning_rate": 1.4422057264050902e-07, + "epoch": 2.593655589123867 + }, + { + "current_steps": 1717, + "loss": 3.0895, + "learning_rate": 1.4369034994697773e-07, + "epoch": 2.595166163141994 + }, + { + "current_steps": 1718, + "loss": 2.9389, + "learning_rate": 1.4316012725344645e-07, + "epoch": 2.596676737160121 + }, + { + "current_steps": 1719, + "loss": 2.7815, + "learning_rate": 1.4262990455991517e-07, + "epoch": 2.598187311178248 + }, + { + "current_steps": 1720, + "loss": 2.9379, + "learning_rate": 1.4209968186638388e-07, + "epoch": 2.599697885196375 + }, + { + "current_steps": 1721, + "loss": 3.013, + "learning_rate": 1.415694591728526e-07, + "epoch": 2.6012084592145017 + }, + { + "current_steps": 1722, + "loss": 3.0051, + "learning_rate": 1.410392364793213e-07, + "epoch": 2.6027190332326287 + }, + { + "current_steps": 1723, + "loss": 2.7263, + "learning_rate": 1.4050901378579003e-07, + "epoch": 2.604229607250755 + }, + { + "current_steps": 1724, + "loss": 2.8186, + "learning_rate": 1.3997879109225874e-07, + "epoch": 2.605740181268882 + }, + { + "current_steps": 1725, + "loss": 3.1671, + "learning_rate": 1.3944856839872746e-07, + "epoch": 2.607250755287009 + }, + { + "current_steps": 1726, + "loss": 3.0972, + "learning_rate": 1.3891834570519618e-07, + "epoch": 2.608761329305136 + }, + { + "current_steps": 1727, + "loss": 3.0065, + "learning_rate": 1.383881230116649e-07, + "epoch": 2.610271903323263 + }, + { + "current_steps": 1728, + "loss": 3.0505, + "learning_rate": 1.378579003181336e-07, + "epoch": 2.61178247734139 + }, + { + "current_steps": 1729, + "loss": 3.1039, + "learning_rate": 1.3732767762460232e-07, + "epoch": 2.6132930513595167 + }, + { + "current_steps": 1730, + "loss": 2.92, + "learning_rate": 1.3679745493107104e-07, + "epoch": 2.6148036253776437 + }, + { + "current_steps": 1731, + "loss": 2.0313, + "learning_rate": 1.3626723223753976e-07, + "epoch": 2.61631419939577 + }, + { + "current_steps": 1732, + "loss": 2.8514, + "learning_rate": 1.357370095440085e-07, + "epoch": 2.617824773413897 + }, + { + "current_steps": 1733, + "loss": 2.8458, + "learning_rate": 1.3520678685047721e-07, + "epoch": 2.619335347432024 + }, + { + "current_steps": 1734, + "loss": 2.8333, + "learning_rate": 1.3467656415694593e-07, + "epoch": 2.620845921450151 + }, + { + "current_steps": 1735, + "loss": 2.0155, + "learning_rate": 1.3414634146341465e-07, + "epoch": 2.622356495468278 + }, + { + "current_steps": 1736, + "loss": 2.8929, + "learning_rate": 1.3361611876988336e-07, + "epoch": 2.623867069486405 + }, + { + "current_steps": 1737, + "loss": 2.9456, + "learning_rate": 1.3308589607635208e-07, + "epoch": 2.6253776435045317 + }, + { + "current_steps": 1738, + "loss": 3.0525, + "learning_rate": 1.325556733828208e-07, + "epoch": 2.6268882175226587 + }, + { + "current_steps": 1739, + "loss": 3.019, + "learning_rate": 1.320254506892895e-07, + "epoch": 2.6283987915407856 + }, + { + "current_steps": 1740, + "loss": 2.5144, + "learning_rate": 1.3149522799575822e-07, + "epoch": 2.6299093655589125 + }, + { + "current_steps": 1741, + "loss": 3.0884, + "learning_rate": 1.3096500530222691e-07, + "epoch": 2.6314199395770395 + }, + { + "current_steps": 1742, + "loss": 3.1782, + "learning_rate": 1.3043478260869563e-07, + "epoch": 2.6329305135951664 + }, + { + "current_steps": 1743, + "loss": 2.9949, + "learning_rate": 1.2990455991516435e-07, + "epoch": 2.6344410876132933 + }, + { + "current_steps": 1744, + "loss": 2.936, + "learning_rate": 1.2937433722163306e-07, + "epoch": 2.63595166163142 + }, + { + "current_steps": 1745, + "loss": 2.9181, + "learning_rate": 1.2884411452810178e-07, + "epoch": 2.6374622356495467 + }, + { + "current_steps": 1746, + "loss": 2.7088, + "learning_rate": 1.283138918345705e-07, + "epoch": 2.6389728096676737 + }, + { + "current_steps": 1747, + "loss": 2.8018, + "learning_rate": 1.277836691410392e-07, + "epoch": 2.6404833836858006 + }, + { + "current_steps": 1748, + "loss": 2.3942, + "learning_rate": 1.2725344644750793e-07, + "epoch": 2.6419939577039275 + }, + { + "current_steps": 1749, + "loss": 3.1531, + "learning_rate": 1.2672322375397667e-07, + "epoch": 2.6435045317220545 + }, + { + "current_steps": 1750, + "loss": 2.8847, + "learning_rate": 1.2619300106044538e-07, + "epoch": 2.6450151057401814 + }, + { + "current_steps": 1751, + "loss": 3.0701, + "learning_rate": 1.256627783669141e-07, + "epoch": 2.646525679758308 + }, + { + "current_steps": 1752, + "loss": 3.3033, + "learning_rate": 1.2513255567338282e-07, + "epoch": 2.648036253776435 + }, + { + "current_steps": 1753, + "loss": 2.7969, + "learning_rate": 1.2460233297985153e-07, + "epoch": 2.6495468277945617 + }, + { + "current_steps": 1754, + "loss": 2.8172, + "learning_rate": 1.2407211028632025e-07, + "epoch": 2.6510574018126887 + }, + { + "current_steps": 1755, + "loss": 2.0426, + "learning_rate": 1.2354188759278896e-07, + "epoch": 2.6525679758308156 + }, + { + "current_steps": 1756, + "loss": 3.0481, + "learning_rate": 1.2301166489925768e-07, + "epoch": 2.6540785498489425 + }, + { + "current_steps": 1757, + "loss": 2.8373, + "learning_rate": 1.224814422057264e-07, + "epoch": 2.6555891238670695 + }, + { + "current_steps": 1758, + "loss": 2.8724, + "learning_rate": 1.219512195121951e-07, + "epoch": 2.6570996978851964 + }, + { + "current_steps": 1759, + "loss": 2.9434, + "learning_rate": 1.2142099681866383e-07, + "epoch": 2.6586102719033233 + }, + { + "current_steps": 1760, + "loss": 2.9696, + "learning_rate": 1.2089077412513254e-07, + "epoch": 2.6601208459214503 + }, + { + "current_steps": 1761, + "loss": 3.091, + "learning_rate": 1.2036055143160128e-07, + "epoch": 2.661631419939577 + }, + { + "current_steps": 1762, + "loss": 2.9475, + "learning_rate": 1.1983032873807e-07, + "epoch": 2.663141993957704 + }, + { + "current_steps": 1763, + "loss": 3.1059, + "learning_rate": 1.1930010604453872e-07, + "epoch": 2.664652567975831 + }, + { + "current_steps": 1764, + "loss": 2.9254, + "learning_rate": 1.1876988335100742e-07, + "epoch": 2.6661631419939575 + }, + { + "current_steps": 1765, + "loss": 2.9885, + "learning_rate": 1.1823966065747613e-07, + "epoch": 2.6676737160120845 + }, + { + "current_steps": 1766, + "loss": 2.7631, + "learning_rate": 1.1770943796394485e-07, + "epoch": 2.6691842900302114 + }, + { + "current_steps": 1767, + "loss": 2.939, + "learning_rate": 1.1717921527041357e-07, + "epoch": 2.6706948640483383 + }, + { + "current_steps": 1768, + "loss": 2.8795, + "learning_rate": 1.1664899257688228e-07, + "epoch": 2.6722054380664653 + }, + { + "current_steps": 1769, + "loss": 2.994, + "learning_rate": 1.1611876988335101e-07, + "epoch": 2.673716012084592 + }, + { + "current_steps": 1770, + "loss": 2.8866, + "learning_rate": 1.1558854718981973e-07, + "epoch": 2.675226586102719 + }, + { + "current_steps": 1771, + "loss": 3.0901, + "learning_rate": 1.1505832449628844e-07, + "epoch": 2.676737160120846 + }, + { + "current_steps": 1772, + "loss": 3.0214, + "learning_rate": 1.1452810180275716e-07, + "epoch": 2.6782477341389725 + }, + { + "current_steps": 1773, + "loss": 2.9535, + "learning_rate": 1.1399787910922588e-07, + "epoch": 2.6797583081570995 + }, + { + "current_steps": 1774, + "loss": 2.8239, + "learning_rate": 1.1346765641569459e-07, + "epoch": 2.6812688821752264 + }, + { + "current_steps": 1775, + "loss": 2.9065, + "learning_rate": 1.129374337221633e-07, + "epoch": 2.6827794561933533 + }, + { + "current_steps": 1776, + "loss": 2.9284, + "learning_rate": 1.1240721102863201e-07, + "epoch": 2.6842900302114803 + }, + { + "current_steps": 1777, + "loss": 2.9564, + "learning_rate": 1.1187698833510073e-07, + "epoch": 2.685800604229607 + }, + { + "current_steps": 1778, + "loss": 2.7768, + "learning_rate": 1.1134676564156945e-07, + "epoch": 2.687311178247734 + }, + { + "current_steps": 1779, + "loss": 2.8876, + "learning_rate": 1.1081654294803817e-07, + "epoch": 2.688821752265861 + }, + { + "current_steps": 1780, + "loss": 2.4335, + "learning_rate": 1.1028632025450689e-07, + "epoch": 2.690332326283988 + }, + { + "current_steps": 1781, + "loss": 2.9975, + "learning_rate": 1.097560975609756e-07, + "epoch": 2.691842900302115 + }, + { + "current_steps": 1782, + "loss": 3.1829, + "learning_rate": 1.0922587486744432e-07, + "epoch": 2.693353474320242 + }, + { + "current_steps": 1783, + "loss": 2.8911, + "learning_rate": 1.0869565217391303e-07, + "epoch": 2.694864048338369 + }, + { + "current_steps": 1784, + "loss": 3.0422, + "learning_rate": 1.0816542948038175e-07, + "epoch": 2.6963746223564957 + }, + { + "current_steps": 1785, + "loss": 2.9348, + "learning_rate": 1.0763520678685047e-07, + "epoch": 2.697885196374622 + }, + { + "current_steps": 1786, + "loss": 2.9154, + "learning_rate": 1.071049840933192e-07, + "epoch": 2.699395770392749 + }, + { + "current_steps": 1787, + "loss": 3.0889, + "learning_rate": 1.0657476139978791e-07, + "epoch": 2.700906344410876 + }, + { + "current_steps": 1788, + "loss": 3.1454, + "learning_rate": 1.0604453870625663e-07, + "epoch": 2.702416918429003 + }, + { + "current_steps": 1789, + "loss": 2.8741, + "learning_rate": 1.0551431601272534e-07, + "epoch": 2.70392749244713 + }, + { + "current_steps": 1790, + "loss": 2.8628, + "learning_rate": 1.0498409331919406e-07, + "epoch": 2.705438066465257 + }, + { + "current_steps": 1791, + "loss": 2.9693, + "learning_rate": 1.0445387062566277e-07, + "epoch": 2.706948640483384 + }, + { + "current_steps": 1792, + "loss": 2.9256, + "learning_rate": 1.0392364793213149e-07, + "epoch": 2.7084592145015107 + }, + { + "current_steps": 1793, + "loss": 2.8001, + "learning_rate": 1.033934252386002e-07, + "epoch": 2.709969788519637 + }, + { + "current_steps": 1794, + "loss": 2.961, + "learning_rate": 1.0286320254506894e-07, + "epoch": 2.711480362537764 + }, + { + "current_steps": 1795, + "loss": 2.9523, + "learning_rate": 1.0233297985153765e-07, + "epoch": 2.712990936555891 + }, + { + "current_steps": 1796, + "loss": 2.9713, + "learning_rate": 1.0180275715800637e-07, + "epoch": 2.714501510574018 + }, + { + "current_steps": 1797, + "loss": 3.019, + "learning_rate": 1.0127253446447508e-07, + "epoch": 2.716012084592145 + }, + { + "current_steps": 1798, + "loss": 2.8769, + "learning_rate": 1.0074231177094379e-07, + "epoch": 2.717522658610272 + }, + { + "current_steps": 1799, + "loss": 3.0721, + "learning_rate": 1.002120890774125e-07, + "epoch": 2.719033232628399 + }, + { + "current_steps": 1800, + "loss": 3.0198, + "learning_rate": 9.968186638388122e-08, + "epoch": 2.7205438066465257 + }, + { + "current_steps": 1801, + "loss": 3.0286, + "learning_rate": 9.915164369034993e-08, + "epoch": 2.7220543806646527 + }, + { + "current_steps": 1802, + "loss": 2.8577, + "learning_rate": 9.862142099681866e-08, + "epoch": 2.7235649546827796 + }, + { + "current_steps": 1803, + "loss": 2.9513, + "learning_rate": 9.809119830328738e-08, + "epoch": 2.7250755287009065 + }, + { + "current_steps": 1804, + "loss": 2.9736, + "learning_rate": 9.75609756097561e-08, + "epoch": 2.7265861027190335 + }, + { + "current_steps": 1805, + "loss": 2.9948, + "learning_rate": 9.703075291622481e-08, + "epoch": 2.7280966767371604 + }, + { + "current_steps": 1806, + "loss": 2.4052, + "learning_rate": 9.650053022269353e-08, + "epoch": 2.729607250755287 + }, + { + "current_steps": 1807, + "loss": 2.3485, + "learning_rate": 9.597030752916224e-08, + "epoch": 2.731117824773414 + }, + { + "current_steps": 1808, + "loss": 2.8374, + "learning_rate": 9.544008483563096e-08, + "epoch": 2.7326283987915407 + }, + { + "current_steps": 1809, + "loss": 2.9591, + "learning_rate": 9.490986214209967e-08, + "epoch": 2.7341389728096677 + }, + { + "current_steps": 1810, + "loss": 2.8868, + "learning_rate": 9.437963944856839e-08, + "epoch": 2.7356495468277946 + }, + { + "current_steps": 1811, + "loss": 2.9672, + "learning_rate": 9.384941675503712e-08, + "epoch": 2.7371601208459215 + }, + { + "current_steps": 1812, + "loss": 3.1142, + "learning_rate": 9.331919406150583e-08, + "epoch": 2.7386706948640485 + }, + { + "current_steps": 1813, + "loss": 2.899, + "learning_rate": 9.278897136797455e-08, + "epoch": 2.7401812688821754 + }, + { + "current_steps": 1814, + "loss": 2.7175, + "learning_rate": 9.225874867444327e-08, + "epoch": 2.741691842900302 + }, + { + "current_steps": 1815, + "loss": 2.8848, + "learning_rate": 9.172852598091198e-08, + "epoch": 2.743202416918429 + }, + { + "current_steps": 1816, + "loss": 2.8628, + "learning_rate": 9.11983032873807e-08, + "epoch": 2.7447129909365557 + }, + { + "current_steps": 1817, + "loss": 2.9828, + "learning_rate": 9.066808059384941e-08, + "epoch": 2.7462235649546827 + }, + { + "current_steps": 1818, + "loss": 2.6791, + "learning_rate": 9.013785790031813e-08, + "epoch": 2.7477341389728096 + }, + { + "current_steps": 1819, + "loss": 3.0182, + "learning_rate": 8.960763520678686e-08, + "epoch": 2.7492447129909365 + }, + { + "current_steps": 1820, + "loss": 2.6572, + "learning_rate": 8.907741251325556e-08, + "epoch": 2.7507552870090635 + }, + { + "current_steps": 1821, + "loss": 3.0155, + "learning_rate": 8.854718981972428e-08, + "epoch": 2.7522658610271904 + }, + { + "current_steps": 1822, + "loss": 2.2547, + "learning_rate": 8.801696712619299e-08, + "epoch": 2.7537764350453173 + }, + { + "current_steps": 1823, + "loss": 2.9473, + "learning_rate": 8.748674443266171e-08, + "epoch": 2.7552870090634443 + }, + { + "current_steps": 1824, + "loss": 2.9914, + "learning_rate": 8.695652173913042e-08, + "epoch": 2.756797583081571 + }, + { + "current_steps": 1825, + "loss": 3.1624, + "learning_rate": 8.642629904559914e-08, + "epoch": 2.758308157099698 + }, + { + "current_steps": 1826, + "loss": 3.0589, + "learning_rate": 8.589607635206786e-08, + "epoch": 2.7598187311178246 + }, + { + "current_steps": 1827, + "loss": 3.0378, + "learning_rate": 8.536585365853659e-08, + "epoch": 2.7613293051359515 + }, + { + "current_steps": 1828, + "loss": 2.979, + "learning_rate": 8.48356309650053e-08, + "epoch": 2.7628398791540785 + }, + { + "current_steps": 1829, + "loss": 2.9768, + "learning_rate": 8.430540827147402e-08, + "epoch": 2.7643504531722054 + }, + { + "current_steps": 1830, + "loss": 2.9681, + "learning_rate": 8.377518557794273e-08, + "epoch": 2.7658610271903323 + }, + { + "current_steps": 1831, + "loss": 3.1236, + "learning_rate": 8.324496288441145e-08, + "epoch": 2.7673716012084593 + }, + { + "current_steps": 1832, + "loss": 3.0861, + "learning_rate": 8.271474019088016e-08, + "epoch": 2.768882175226586 + }, + { + "current_steps": 1833, + "loss": 3.1181, + "learning_rate": 8.218451749734888e-08, + "epoch": 2.770392749244713 + }, + { + "current_steps": 1834, + "loss": 2.7771, + "learning_rate": 8.16542948038176e-08, + "epoch": 2.7719033232628396 + }, + { + "current_steps": 1835, + "loss": 3.0001, + "learning_rate": 8.112407211028633e-08, + "epoch": 2.7734138972809665 + }, + { + "current_steps": 1836, + "loss": 2.904, + "learning_rate": 8.059384941675504e-08, + "epoch": 2.7749244712990935 + }, + { + "current_steps": 1837, + "loss": 2.9935, + "learning_rate": 8.006362672322376e-08, + "epoch": 2.7764350453172204 + }, + { + "current_steps": 1838, + "loss": 2.8819, + "learning_rate": 7.953340402969247e-08, + "epoch": 2.7779456193353473 + }, + { + "current_steps": 1839, + "loss": 2.8291, + "learning_rate": 7.900318133616119e-08, + "epoch": 2.7794561933534743 + }, + { + "current_steps": 1840, + "loss": 3.0213, + "learning_rate": 7.84729586426299e-08, + "epoch": 2.780966767371601 + }, + { + "current_steps": 1841, + "loss": 3.0364, + "learning_rate": 7.794273594909862e-08, + "epoch": 2.782477341389728 + }, + { + "current_steps": 1842, + "loss": 3.1144, + "learning_rate": 7.741251325556732e-08, + "epoch": 2.783987915407855 + }, + { + "current_steps": 1843, + "loss": 2.8542, + "learning_rate": 7.688229056203604e-08, + "epoch": 2.785498489425982 + }, + { + "current_steps": 1844, + "loss": 3.028, + "learning_rate": 7.635206786850477e-08, + "epoch": 2.787009063444109 + }, + { + "current_steps": 1845, + "loss": 2.8336, + "learning_rate": 7.582184517497348e-08, + "epoch": 2.788519637462236 + }, + { + "current_steps": 1846, + "loss": 2.7862, + "learning_rate": 7.52916224814422e-08, + "epoch": 2.790030211480363 + }, + { + "current_steps": 1847, + "loss": 3.0405, + "learning_rate": 7.476139978791092e-08, + "epoch": 2.7915407854984893 + }, + { + "current_steps": 1848, + "loss": 2.8597, + "learning_rate": 7.423117709437963e-08, + "epoch": 2.793051359516616 + }, + { + "current_steps": 1849, + "loss": 2.946, + "learning_rate": 7.370095440084835e-08, + "epoch": 2.794561933534743 + }, + { + "current_steps": 1850, + "loss": 3.1885, + "learning_rate": 7.317073170731706e-08, + "epoch": 2.79607250755287 + }, + { + "current_steps": 1851, + "loss": 3.1006, + "learning_rate": 7.264050901378578e-08, + "epoch": 2.797583081570997 + }, + { + "current_steps": 1852, + "loss": 2.8335, + "learning_rate": 7.211028632025451e-08, + "epoch": 2.799093655589124 + }, + { + "current_steps": 1853, + "loss": 2.8269, + "learning_rate": 7.158006362672322e-08, + "epoch": 2.800604229607251 + }, + { + "current_steps": 1854, + "loss": 2.6435, + "learning_rate": 7.104984093319194e-08, + "epoch": 2.802114803625378 + }, + { + "current_steps": 1855, + "loss": 2.9359, + "learning_rate": 7.051961823966066e-08, + "epoch": 2.8036253776435043 + }, + { + "current_steps": 1856, + "loss": 3.0477, + "learning_rate": 6.998939554612937e-08, + "epoch": 2.805135951661631 + }, + { + "current_steps": 1857, + "loss": 3.1709, + "learning_rate": 6.945917285259809e-08, + "epoch": 2.806646525679758 + }, + { + "current_steps": 1858, + "loss": 2.8944, + "learning_rate": 6.89289501590668e-08, + "epoch": 2.808157099697885 + }, + { + "current_steps": 1859, + "loss": 2.8094, + "learning_rate": 6.839872746553552e-08, + "epoch": 2.809667673716012 + }, + { + "current_steps": 1860, + "loss": 2.9316, + "learning_rate": 6.786850477200425e-08, + "epoch": 2.811178247734139 + }, + { + "current_steps": 1861, + "loss": 3.0675, + "learning_rate": 6.733828207847296e-08, + "epoch": 2.812688821752266 + }, + { + "current_steps": 1862, + "loss": 2.958, + "learning_rate": 6.680805938494168e-08, + "epoch": 2.814199395770393 + }, + { + "current_steps": 1863, + "loss": 3.0425, + "learning_rate": 6.62778366914104e-08, + "epoch": 2.8157099697885197 + }, + { + "current_steps": 1864, + "loss": 2.7715, + "learning_rate": 6.574761399787911e-08, + "epoch": 2.8172205438066467 + }, + { + "current_steps": 1865, + "loss": 3.0176, + "learning_rate": 6.521739130434782e-08, + "epoch": 2.8187311178247736 + }, + { + "current_steps": 1866, + "loss": 3.0609, + "learning_rate": 6.468716861081653e-08, + "epoch": 2.8202416918429005 + }, + { + "current_steps": 1867, + "loss": 2.612, + "learning_rate": 6.415694591728525e-08, + "epoch": 2.8217522658610275 + }, + { + "current_steps": 1868, + "loss": 2.9679, + "learning_rate": 6.362672322375396e-08, + "epoch": 2.823262839879154 + }, + { + "current_steps": 1869, + "loss": 2.5519, + "learning_rate": 6.309650053022269e-08, + "epoch": 2.824773413897281 + }, + { + "current_steps": 1870, + "loss": 2.9077, + "learning_rate": 6.256627783669141e-08, + "epoch": 2.826283987915408 + }, + { + "current_steps": 1871, + "loss": 2.9207, + "learning_rate": 6.203605514316012e-08, + "epoch": 2.8277945619335347 + }, + { + "current_steps": 1872, + "loss": 2.7965, + "learning_rate": 6.150583244962884e-08, + "epoch": 2.8293051359516617 + }, + { + "current_steps": 1873, + "loss": 2.881, + "learning_rate": 6.097560975609756e-08, + "epoch": 2.8308157099697886 + }, + { + "current_steps": 1874, + "loss": 2.963, + "learning_rate": 6.044538706256627e-08, + "epoch": 2.8323262839879155 + }, + { + "current_steps": 1875, + "loss": 3.058, + "learning_rate": 5.9915164369035e-08, + "epoch": 2.8338368580060425 + }, + { + "current_steps": 1876, + "loss": 3.0952, + "learning_rate": 5.938494167550371e-08, + "epoch": 2.835347432024169 + }, + { + "current_steps": 1877, + "loss": 3.166, + "learning_rate": 5.8854718981972425e-08, + "epoch": 2.836858006042296 + }, + { + "current_steps": 1878, + "loss": 2.8453, + "learning_rate": 5.832449628844114e-08, + "epoch": 2.838368580060423 + }, + { + "current_steps": 1879, + "loss": 2.3267, + "learning_rate": 5.7794273594909864e-08, + "epoch": 2.8398791540785497 + }, + { + "current_steps": 1880, + "loss": 3.3411, + "learning_rate": 5.726405090137858e-08, + "epoch": 2.8413897280966767 + }, + { + "current_steps": 1881, + "loss": 2.727, + "learning_rate": 5.6733828207847296e-08, + "epoch": 2.8429003021148036 + }, + { + "current_steps": 1882, + "loss": 3.0315, + "learning_rate": 5.6203605514316005e-08, + "epoch": 2.8444108761329305 + }, + { + "current_steps": 1883, + "loss": 2.8649, + "learning_rate": 5.567338282078473e-08, + "epoch": 2.8459214501510575 + }, + { + "current_steps": 1884, + "loss": 3.0212, + "learning_rate": 5.514316012725344e-08, + "epoch": 2.8474320241691844 + }, + { + "current_steps": 1885, + "loss": 3.1805, + "learning_rate": 5.461293743372216e-08, + "epoch": 2.8489425981873113 + }, + { + "current_steps": 1886, + "loss": 3.0182, + "learning_rate": 5.4082714740190875e-08, + "epoch": 2.8504531722054383 + }, + { + "current_steps": 1887, + "loss": 3.0725, + "learning_rate": 5.35524920466596e-08, + "epoch": 2.851963746223565 + }, + { + "current_steps": 1888, + "loss": 2.0851, + "learning_rate": 5.302226935312831e-08, + "epoch": 2.853474320241692 + }, + { + "current_steps": 1889, + "loss": 2.9063, + "learning_rate": 5.249204665959703e-08, + "epoch": 2.8549848942598186 + }, + { + "current_steps": 1890, + "loss": 2.9657, + "learning_rate": 5.1961823966065745e-08, + "epoch": 2.8564954682779455 + }, + { + "current_steps": 1891, + "loss": 2.9003, + "learning_rate": 5.143160127253447e-08, + "epoch": 2.8580060422960725 + }, + { + "current_steps": 1892, + "loss": 3.1238, + "learning_rate": 5.0901378579003183e-08, + "epoch": 2.8595166163141994 + }, + { + "current_steps": 1893, + "loss": 3.0727, + "learning_rate": 5.037115588547189e-08, + "epoch": 2.8610271903323263 + }, + { + "current_steps": 1894, + "loss": 2.9672, + "learning_rate": 4.984093319194061e-08, + "epoch": 2.8625377643504533 + }, + { + "current_steps": 1895, + "loss": 3.0155, + "learning_rate": 4.931071049840933e-08, + "epoch": 2.86404833836858 + }, + { + "current_steps": 1896, + "loss": 2.9195, + "learning_rate": 4.878048780487805e-08, + "epoch": 2.8655589123867067 + }, + { + "current_steps": 1897, + "loss": 3.0422, + "learning_rate": 4.825026511134676e-08, + "epoch": 2.8670694864048336 + }, + { + "current_steps": 1898, + "loss": 2.7132, + "learning_rate": 4.772004241781548e-08, + "epoch": 2.8685800604229605 + }, + { + "current_steps": 1899, + "loss": 3.126, + "learning_rate": 4.7189819724284195e-08, + "epoch": 2.8700906344410875 + }, + { + "current_steps": 1900, + "loss": 2.8712, + "learning_rate": 4.665959703075292e-08, + "epoch": 2.8716012084592144 + }, + { + "current_steps": 1901, + "loss": 2.9313, + "learning_rate": 4.612937433722163e-08, + "epoch": 2.8731117824773413 + }, + { + "current_steps": 1902, + "loss": 3.2139, + "learning_rate": 4.559915164369035e-08, + "epoch": 2.8746223564954683 + }, + { + "current_steps": 1903, + "loss": 2.8807, + "learning_rate": 4.5068928950159065e-08, + "epoch": 2.876132930513595 + }, + { + "current_steps": 1904, + "loss": 2.8274, + "learning_rate": 4.453870625662778e-08, + "epoch": 2.877643504531722 + }, + { + "current_steps": 1905, + "loss": 2.9904, + "learning_rate": 4.4008483563096496e-08, + "epoch": 2.879154078549849 + }, + { + "current_steps": 1906, + "loss": 3.1017, + "learning_rate": 4.347826086956521e-08, + "epoch": 2.880664652567976 + }, + { + "current_steps": 1907, + "loss": 2.9465, + "learning_rate": 4.294803817603393e-08, + "epoch": 2.882175226586103 + }, + { + "current_steps": 1908, + "loss": 2.963, + "learning_rate": 4.241781548250265e-08, + "epoch": 2.88368580060423 + }, + { + "current_steps": 1909, + "loss": 3.0362, + "learning_rate": 4.1887592788971367e-08, + "epoch": 2.8851963746223563 + }, + { + "current_steps": 1910, + "loss": 2.824, + "learning_rate": 4.135737009544008e-08, + "epoch": 2.8867069486404833 + }, + { + "current_steps": 1911, + "loss": 2.8822, + "learning_rate": 4.08271474019088e-08, + "epoch": 2.88821752265861 + }, + { + "current_steps": 1912, + "loss": 2.8811, + "learning_rate": 4.029692470837752e-08, + "epoch": 2.889728096676737 + }, + { + "current_steps": 1913, + "loss": 3.0816, + "learning_rate": 3.9766702014846237e-08, + "epoch": 2.891238670694864 + }, + { + "current_steps": 1914, + "loss": 3.0799, + "learning_rate": 3.923647932131495e-08, + "epoch": 2.892749244712991 + }, + { + "current_steps": 1915, + "loss": 2.89, + "learning_rate": 3.870625662778366e-08, + "epoch": 2.894259818731118 + }, + { + "current_steps": 1916, + "loss": 3.1869, + "learning_rate": 3.8176033934252384e-08, + "epoch": 2.895770392749245 + }, + { + "current_steps": 1917, + "loss": 3.1516, + "learning_rate": 3.76458112407211e-08, + "epoch": 2.8972809667673713 + }, + { + "current_steps": 1918, + "loss": 3.0378, + "learning_rate": 3.7115588547189816e-08, + "epoch": 2.8987915407854983 + }, + { + "current_steps": 1919, + "loss": 2.8611, + "learning_rate": 3.658536585365853e-08, + "epoch": 2.900302114803625 + }, + { + "current_steps": 1920, + "loss": 2.9668, + "learning_rate": 3.6055143160127254e-08, + "epoch": 2.901812688821752 + }, + { + "current_steps": 1921, + "loss": 2.9424, + "learning_rate": 3.552492046659597e-08, + "epoch": 2.903323262839879 + }, + { + "current_steps": 1922, + "loss": 2.9069, + "learning_rate": 3.4994697773064686e-08, + "epoch": 2.904833836858006 + }, + { + "current_steps": 1923, + "loss": 2.9543, + "learning_rate": 3.44644750795334e-08, + "epoch": 2.906344410876133 + }, + { + "current_steps": 1924, + "loss": 2.7606, + "learning_rate": 3.3934252386002124e-08, + "epoch": 2.90785498489426 + }, + { + "current_steps": 1925, + "loss": 3.0176, + "learning_rate": 3.340402969247084e-08, + "epoch": 2.909365558912387 + }, + { + "current_steps": 1926, + "loss": 2.9565, + "learning_rate": 3.2873806998939556e-08, + "epoch": 2.9108761329305137 + }, + { + "current_steps": 1927, + "loss": 2.94, + "learning_rate": 3.2343584305408265e-08, + "epoch": 2.9123867069486407 + }, + { + "current_steps": 1928, + "loss": 3.0043, + "learning_rate": 3.181336161187698e-08, + "epoch": 2.9138972809667676 + }, + { + "current_steps": 1929, + "loss": 2.9566, + "learning_rate": 3.1283138918345704e-08, + "epoch": 2.9154078549848945 + }, + { + "current_steps": 1930, + "loss": 3.0867, + "learning_rate": 3.075291622481442e-08, + "epoch": 2.916918429003021 + }, + { + "current_steps": 1931, + "loss": 3.0064, + "learning_rate": 3.0222693531283136e-08, + "epoch": 2.918429003021148 + }, + { + "current_steps": 1932, + "loss": 2.7964, + "learning_rate": 2.9692470837751855e-08, + "epoch": 2.919939577039275 + }, + { + "current_steps": 1933, + "loss": 2.9417, + "learning_rate": 2.916224814422057e-08, + "epoch": 2.921450151057402 + }, + { + "current_steps": 1934, + "loss": 2.9919, + "learning_rate": 2.863202545068929e-08, + "epoch": 2.9229607250755287 + }, + { + "current_steps": 1935, + "loss": 3.0771, + "learning_rate": 2.8101802757158002e-08, + "epoch": 2.9244712990936557 + }, + { + "current_steps": 1936, + "loss": 2.8613, + "learning_rate": 2.757158006362672e-08, + "epoch": 2.9259818731117826 + }, + { + "current_steps": 1937, + "loss": 2.9584, + "learning_rate": 2.7041357370095437e-08, + "epoch": 2.9274924471299095 + }, + { + "current_steps": 1938, + "loss": 2.9161, + "learning_rate": 2.6511134676564157e-08, + "epoch": 2.929003021148036 + }, + { + "current_steps": 1939, + "loss": 3.0336, + "learning_rate": 2.5980911983032873e-08, + "epoch": 2.930513595166163 + }, + { + "current_steps": 1940, + "loss": 2.9381, + "learning_rate": 2.5450689289501592e-08, + "epoch": 2.93202416918429 + }, + { + "current_steps": 1941, + "loss": 3.0685, + "learning_rate": 2.4920466595970304e-08, + "epoch": 2.933534743202417 + }, + { + "current_steps": 1942, + "loss": 3.0419, + "learning_rate": 2.4390243902439023e-08, + "epoch": 2.9350453172205437 + }, + { + "current_steps": 1943, + "loss": 2.7357, + "learning_rate": 2.386002120890774e-08, + "epoch": 2.9365558912386707 + }, + { + "current_steps": 1944, + "loss": 3.0399, + "learning_rate": 2.332979851537646e-08, + "epoch": 2.9380664652567976 + }, + { + "current_steps": 1945, + "loss": 2.3534, + "learning_rate": 2.2799575821845174e-08, + "epoch": 2.9395770392749245 + }, + { + "current_steps": 1946, + "loss": 2.8138, + "learning_rate": 2.226935312831389e-08, + "epoch": 2.9410876132930515 + }, + { + "current_steps": 1947, + "loss": 2.9336, + "learning_rate": 2.1739130434782606e-08, + "epoch": 2.9425981873111784 + }, + { + "current_steps": 1948, + "loss": 3.1749, + "learning_rate": 2.1208907741251325e-08, + "epoch": 2.9441087613293053 + }, + { + "current_steps": 1949, + "loss": 2.857, + "learning_rate": 2.067868504772004e-08, + "epoch": 2.9456193353474323 + }, + { + "current_steps": 1950, + "loss": 3.0774, + "learning_rate": 2.014846235418876e-08, + "epoch": 2.947129909365559 + }, + { + "current_steps": 1951, + "loss": 3.0129, + "learning_rate": 1.9618239660657476e-08, + "epoch": 2.9486404833836857 + }, + { + "current_steps": 1952, + "loss": 2.8487, + "learning_rate": 1.9088016967126192e-08, + "epoch": 2.9501510574018126 + }, + { + "current_steps": 1953, + "loss": 2.9382, + "learning_rate": 1.8557794273594908e-08, + "epoch": 2.9516616314199395 + }, + { + "current_steps": 1954, + "loss": 3.1172, + "learning_rate": 1.8027571580063627e-08, + "epoch": 2.9531722054380665 + }, + { + "current_steps": 1955, + "loss": 3.1302, + "learning_rate": 1.7497348886532343e-08, + "epoch": 2.9546827794561934 + }, + { + "current_steps": 1956, + "loss": 2.8669, + "learning_rate": 1.6967126193001062e-08, + "epoch": 2.9561933534743203 + }, + { + "current_steps": 1957, + "loss": 3.0425, + "learning_rate": 1.6436903499469778e-08, + "epoch": 2.9577039274924473 + }, + { + "current_steps": 1958, + "loss": 2.9269, + "learning_rate": 1.590668080593849e-08, + "epoch": 2.9592145015105737 + }, + { + "current_steps": 1959, + "loss": 2.8536, + "learning_rate": 1.537645811240721e-08, + "epoch": 2.9607250755287007 + }, + { + "current_steps": 1960, + "loss": 3.0508, + "learning_rate": 1.4846235418875927e-08, + "epoch": 2.9622356495468276 + }, + { + "current_steps": 1961, + "loss": 2.984, + "learning_rate": 1.4316012725344645e-08, + "epoch": 2.9637462235649545 + }, + { + "current_steps": 1962, + "loss": 2.6814, + "learning_rate": 1.378579003181336e-08, + "epoch": 2.9652567975830815 + }, + { + "current_steps": 1963, + "loss": 2.935, + "learning_rate": 1.3255567338282078e-08, + "epoch": 2.9667673716012084 + }, + { + "current_steps": 1964, + "loss": 2.8209, + "learning_rate": 1.2725344644750796e-08, + "epoch": 2.9682779456193353 + }, + { + "current_steps": 1965, + "loss": 2.9674, + "learning_rate": 1.2195121951219512e-08, + "epoch": 2.9697885196374623 + }, + { + "current_steps": 1966, + "loss": 2.8591, + "learning_rate": 1.166489925768823e-08, + "epoch": 2.971299093655589 + }, + { + "current_steps": 1967, + "loss": 3.0003, + "learning_rate": 1.1134676564156945e-08, + "epoch": 2.972809667673716 + }, + { + "current_steps": 1968, + "loss": 3.0823, + "learning_rate": 1.0604453870625663e-08, + "epoch": 2.974320241691843 + }, + { + "current_steps": 1969, + "loss": 2.8632, + "learning_rate": 1.007423117709438e-08, + "epoch": 2.97583081570997 + }, + { + "current_steps": 1970, + "loss": 3.1056, + "learning_rate": 9.544008483563096e-09, + "epoch": 2.977341389728097 + }, + { + "current_steps": 1971, + "loss": 2.7785, + "learning_rate": 9.013785790031814e-09, + "epoch": 2.9788519637462234 + }, + { + "current_steps": 1972, + "loss": 3.0876, + "learning_rate": 8.483563096500531e-09, + "epoch": 2.9803625377643503 + }, + { + "current_steps": 1973, + "loss": 2.9398, + "learning_rate": 7.953340402969245e-09, + "epoch": 2.9818731117824773 + }, + { + "current_steps": 1974, + "loss": 2.6686, + "learning_rate": 7.423117709437964e-09, + "epoch": 2.983383685800604 + }, + { + "current_steps": 1975, + "loss": 2.2615, + "learning_rate": 6.89289501590668e-09, + "epoch": 2.984894259818731 + }, + { + "current_steps": 1976, + "loss": 3.0279, + "learning_rate": 6.362672322375398e-09, + "epoch": 2.986404833836858 + }, + { + "current_steps": 1977, + "loss": 2.4458, + "learning_rate": 5.832449628844115e-09, + "epoch": 2.987915407854985 + }, + { + "current_steps": 1978, + "loss": 2.6705, + "learning_rate": 5.302226935312831e-09, + "epoch": 2.989425981873112 + }, + { + "current_steps": 1979, + "loss": 2.9036, + "learning_rate": 4.772004241781548e-09, + "epoch": 2.9909365558912384 + }, + { + "current_steps": 1980, + "loss": 3.007, + "learning_rate": 4.2417815482502656e-09, + "epoch": 2.9924471299093653 + }, + { + "current_steps": 1981, + "loss": 3.0469, + "learning_rate": 3.711558854718982e-09, + "epoch": 2.9939577039274923 + }, + { + "current_steps": 1982, + "loss": 3.1243, + "learning_rate": 3.181336161187699e-09, + "epoch": 2.995468277945619 + }, + { + "current_steps": 1983, + "loss": 2.9443, + "learning_rate": 2.6511134676564157e-09, + "epoch": 2.996978851963746 + }, + { + "current_steps": 1984, + "loss": 2.9877, + "learning_rate": 2.1208907741251328e-09, + "epoch": 2.998489425981873 + }, + { + "current_steps": 1985, + "loss": 3.1269, + "learning_rate": 1.5906680805938495e-09, + "epoch": 3.0 + }, + { + "current_steps": 1985, + "loss": 3.1269, + "learning_rate": 1.5906680805938495e-09, + "epoch": 3.0 + } +] \ No newline at end of file