[ { "current_steps": 0, "loss": 2.4951, "learning_rate": 1e-08, "epoch": 0.0019455252918287938 }, { "current_steps": 1, "loss": 1.6208, "learning_rate": 2e-08, "epoch": 0.0038910505836575876 }, { "current_steps": 2, "loss": 2.2859, "learning_rate": 3e-08, "epoch": 0.005836575875486381 }, { "current_steps": 3, "loss": 2.2251, "learning_rate": 4e-08, "epoch": 0.007782101167315175 }, { "current_steps": 4, "loss": 2.2389, "learning_rate": 5e-08, "epoch": 0.009727626459143969 }, { "current_steps": 5, "loss": 2.4305, "learning_rate": 6e-08, "epoch": 0.011673151750972763 }, { "current_steps": 6, "loss": 2.2879, "learning_rate": 7e-08, "epoch": 0.013618677042801557 }, { "current_steps": 7, "loss": 2.3366, "learning_rate": 8e-08, "epoch": 0.01556420233463035 }, { "current_steps": 8, "loss": 2.1907, "learning_rate": 9e-08, "epoch": 0.017509727626459144 }, { "current_steps": 9, "loss": 2.3006, "learning_rate": 1e-07, "epoch": 0.019455252918287938 }, { "current_steps": 10, "loss": 2.5113, "learning_rate": 1e-07, "epoch": 0.021400778210116732 }, { "current_steps": 11, "loss": 2.0847, "learning_rate": 1.0999999999999999e-07, "epoch": 0.023346303501945526 }, { "current_steps": 12, "loss": 2.3379, "learning_rate": 1.2e-07, "epoch": 0.02529182879377432 }, { "current_steps": 13, "loss": 2.6107, "learning_rate": 1.2e-07, "epoch": 0.027237354085603113 }, { "current_steps": 14, "loss": 1.9509, "learning_rate": 1.3e-07, "epoch": 0.029182879377431907 }, { "current_steps": 15, "loss": 2.1483, "learning_rate": 1.4e-07, "epoch": 0.0311284046692607 }, { "current_steps": 16, "loss": 2.6173, "learning_rate": 1.5e-07, "epoch": 0.033073929961089495 }, { "current_steps": 17, "loss": 2.5054, "learning_rate": 1.6e-07, "epoch": 0.03501945525291829 }, { "current_steps": 18, "loss": 2.5205, "learning_rate": 1.7000000000000001e-07, "epoch": 0.03696498054474708 }, { "current_steps": 19, "loss": 2.3719, "learning_rate": 1.8e-07, "epoch": 0.038910505836575876 }, { "current_steps": 20, "loss": 2.2282, "learning_rate": 1.8999999999999998e-07, "epoch": 0.04085603112840467 }, { "current_steps": 21, "loss": 2.3871, "learning_rate": 2e-07, "epoch": 0.042801556420233464 }, { "current_steps": 22, "loss": 1.8128, "learning_rate": 2.0999999999999997e-07, "epoch": 0.04474708171206226 }, { "current_steps": 23, "loss": 2.5305, "learning_rate": 2.1999999999999998e-07, "epoch": 0.04669260700389105 }, { "current_steps": 24, "loss": 2.6122, "learning_rate": 2.3e-07, "epoch": 0.048638132295719845 }, { "current_steps": 25, "loss": 2.0801, "learning_rate": 2.4e-07, "epoch": 0.05058365758754864 }, { "current_steps": 26, "loss": 2.432, "learning_rate": 2.5e-07, "epoch": 0.05252918287937743 }, { "current_steps": 27, "loss": 2.4446, "learning_rate": 2.6e-07, "epoch": 0.054474708171206226 }, { "current_steps": 28, "loss": 2.531, "learning_rate": 2.7e-07, "epoch": 0.05642023346303502 }, { "current_steps": 29, "loss": 2.227, "learning_rate": 2.8e-07, "epoch": 0.058365758754863814 }, { "current_steps": 30, "loss": 2.1831, "learning_rate": 2.9e-07, "epoch": 0.06031128404669261 }, { "current_steps": 31, "loss": 2.472, "learning_rate": 3e-07, "epoch": 0.0622568093385214 }, { "current_steps": 32, "loss": 2.3757, "learning_rate": 3.1e-07, "epoch": 0.06420233463035019 }, { "current_steps": 33, "loss": 2.1377, "learning_rate": 3.2e-07, "epoch": 0.06614785992217899 }, { "current_steps": 34, "loss": 2.166, "learning_rate": 3.3e-07, "epoch": 0.06809338521400778 }, { "current_steps": 35, "loss": 2.3693, "learning_rate": 3.4000000000000003e-07, "epoch": 0.07003891050583658 }, { "current_steps": 36, "loss": 2.46, "learning_rate": 3.5e-07, "epoch": 0.07198443579766536 }, { "current_steps": 37, "loss": 2.2548, "learning_rate": 3.6e-07, "epoch": 0.07392996108949416 }, { "current_steps": 38, "loss": 2.393, "learning_rate": 3.7e-07, "epoch": 0.07587548638132295 }, { "current_steps": 39, "loss": 2.4823, "learning_rate": 3.7999999999999996e-07, "epoch": 0.07782101167315175 }, { "current_steps": 40, "loss": 2.35, "learning_rate": 3.8999999999999997e-07, "epoch": 0.07976653696498054 }, { "current_steps": 41, "loss": 2.2729, "learning_rate": 4e-07, "epoch": 0.08171206225680934 }, { "current_steps": 42, "loss": 2.502, "learning_rate": 4.0999999999999994e-07, "epoch": 0.08365758754863813 }, { "current_steps": 43, "loss": 2.3356, "learning_rate": 4.1999999999999995e-07, "epoch": 0.08560311284046693 }, { "current_steps": 44, "loss": 2.6155, "learning_rate": 4.2999999999999996e-07, "epoch": 0.08754863813229571 }, { "current_steps": 45, "loss": 2.5189, "learning_rate": 4.3999999999999997e-07, "epoch": 0.08949416342412451 }, { "current_steps": 46, "loss": 1.931, "learning_rate": 4.5e-07, "epoch": 0.0914396887159533 }, { "current_steps": 47, "loss": 2.1135, "learning_rate": 4.6e-07, "epoch": 0.0933852140077821 }, { "current_steps": 48, "loss": 1.6983, "learning_rate": 4.6999999999999995e-07, "epoch": 0.09533073929961089 }, { "current_steps": 49, "loss": 2.3501, "learning_rate": 4.8e-07, "epoch": 0.09727626459143969 }, { "current_steps": 50, "loss": 2.6536, "learning_rate": 4.9e-07, "epoch": 0.09922178988326848 }, { "current_steps": 51, "loss": 2.0913, "learning_rate": 5e-07, "epoch": 0.10116731517509728 }, { "current_steps": 52, "loss": 2.3891, "learning_rate": 5.1e-07, "epoch": 0.10311284046692606 }, { "current_steps": 53, "loss": 2.209, "learning_rate": 5.2e-07, "epoch": 0.10505836575875487 }, { "current_steps": 54, "loss": 2.4225, "learning_rate": 5.3e-07, "epoch": 0.10700389105058365 }, { "current_steps": 55, "loss": 2.1131, "learning_rate": 5.4e-07, "epoch": 0.10894941634241245 }, { "current_steps": 56, "loss": 2.0144, "learning_rate": 5.5e-07, "epoch": 0.11089494163424124 }, { "current_steps": 57, "loss": 2.2635, "learning_rate": 5.6e-07, "epoch": 0.11284046692607004 }, { "current_steps": 58, "loss": 2.2203, "learning_rate": 5.699999999999999e-07, "epoch": 0.11478599221789883 }, { "current_steps": 59, "loss": 2.4986, "learning_rate": 5.8e-07, "epoch": 0.11673151750972763 }, { "current_steps": 60, "loss": 2.5164, "learning_rate": 5.9e-07, "epoch": 0.11867704280155641 }, { "current_steps": 61, "loss": 2.3796, "learning_rate": 6e-07, "epoch": 0.12062256809338522 }, { "current_steps": 62, "loss": 2.4926, "learning_rate": 6.1e-07, "epoch": 0.122568093385214 }, { "current_steps": 63, "loss": 2.1988, "learning_rate": 6.2e-07, "epoch": 0.1245136186770428 }, { "current_steps": 64, "loss": 2.2953, "learning_rate": 6.3e-07, "epoch": 0.1264591439688716 }, { "current_steps": 65, "loss": 2.6159, "learning_rate": 6.4e-07, "epoch": 0.12840466926070038 }, { "current_steps": 66, "loss": 2.5671, "learning_rate": 6.5e-07, "epoch": 0.1303501945525292 }, { "current_steps": 67, "loss": 2.2536, "learning_rate": 6.6e-07, "epoch": 0.13229571984435798 }, { "current_steps": 68, "loss": 2.0765, "learning_rate": 6.7e-07, "epoch": 0.13424124513618677 }, { "current_steps": 69, "loss": 2.1845, "learning_rate": 6.800000000000001e-07, "epoch": 0.13618677042801555 }, { "current_steps": 70, "loss": 2.4424, "learning_rate": 6.9e-07, "epoch": 0.13813229571984437 }, { "current_steps": 71, "loss": 2.117, "learning_rate": 7e-07, "epoch": 0.14007782101167315 }, { "current_steps": 72, "loss": 2.1983, "learning_rate": 7.1e-07, "epoch": 0.14202334630350194 }, { "current_steps": 73, "loss": 2.575, "learning_rate": 7.2e-07, "epoch": 0.14396887159533073 }, { "current_steps": 74, "loss": 2.61, "learning_rate": 7.3e-07, "epoch": 0.14591439688715954 }, { "current_steps": 75, "loss": 2.3598, "learning_rate": 7.4e-07, "epoch": 0.14785992217898833 }, { "current_steps": 76, "loss": 2.4275, "learning_rate": 7.5e-07, "epoch": 0.14980544747081712 }, { "current_steps": 77, "loss": 2.5497, "learning_rate": 7.599999999999999e-07, "epoch": 0.1517509727626459 }, { "current_steps": 78, "loss": 2.1736, "learning_rate": 7.699999999999999e-07, "epoch": 0.15369649805447472 }, { "current_steps": 79, "loss": 2.266, "learning_rate": 7.799999999999999e-07, "epoch": 0.1556420233463035 }, { "current_steps": 80, "loss": 2.4923, "learning_rate": 7.9e-07, "epoch": 0.1575875486381323 }, { "current_steps": 81, "loss": 2.383, "learning_rate": 8e-07, "epoch": 0.15953307392996108 }, { "current_steps": 82, "loss": 2.296, "learning_rate": 8.1e-07, "epoch": 0.1614785992217899 }, { "current_steps": 83, "loss": 2.4952, "learning_rate": 8.199999999999999e-07, "epoch": 0.16342412451361868 }, { "current_steps": 84, "loss": 2.1596, "learning_rate": 8.299999999999999e-07, "epoch": 0.16536964980544747 }, { "current_steps": 85, "loss": 2.306, "learning_rate": 8.399999999999999e-07, "epoch": 0.16731517509727625 }, { "current_steps": 86, "loss": 2.3771, "learning_rate": 8.499999999999999e-07, "epoch": 0.16926070038910507 }, { "current_steps": 87, "loss": 2.4864, "learning_rate": 8.599999999999999e-07, "epoch": 0.17120622568093385 }, { "current_steps": 88, "loss": 2.5982, "learning_rate": 8.699999999999999e-07, "epoch": 0.17315175097276264 }, { "current_steps": 89, "loss": 2.1604, "learning_rate": 8.799999999999999e-07, "epoch": 0.17509727626459143 }, { "current_steps": 90, "loss": 2.4259, "learning_rate": 8.9e-07, "epoch": 0.17704280155642024 }, { "current_steps": 91, "loss": 2.321, "learning_rate": 9e-07, "epoch": 0.17898832684824903 }, { "current_steps": 92, "loss": 2.3068, "learning_rate": 9.1e-07, "epoch": 0.18093385214007782 }, { "current_steps": 93, "loss": 2.4903, "learning_rate": 9.2e-07, "epoch": 0.1828793774319066 }, { "current_steps": 94, "loss": 2.4446, "learning_rate": 9.3e-07, "epoch": 0.18482490272373542 }, { "current_steps": 95, "loss": 2.2668, "learning_rate": 9.399999999999999e-07, "epoch": 0.1867704280155642 }, { "current_steps": 96, "loss": 2.3091, "learning_rate": 9.499999999999999e-07, "epoch": 0.188715953307393 }, { "current_steps": 97, "loss": 2.2757, "learning_rate": 9.499999999999999e-07, "epoch": 0.19066147859922178 }, { "current_steps": 98, "loss": 2.3616, "learning_rate": 9.6e-07, "epoch": 0.1926070038910506 }, { "current_steps": 99, "loss": 2.0339, "learning_rate": 9.7e-07, "epoch": 0.19455252918287938 }, { "current_steps": 100, "loss": 1.9707, "learning_rate": 9.8e-07, "epoch": 0.19649805447470817 }, { "current_steps": 101, "loss": 2.3426, "learning_rate": 9.9e-07, "epoch": 0.19844357976653695 }, { "current_steps": 102, "loss": 2.4402, "learning_rate": 1e-06, "epoch": 0.20038910505836577 }, { "current_steps": 103, "loss": 2.4034, "learning_rate": 9.995951417004047e-07, "epoch": 0.20233463035019456 }, { "current_steps": 104, "loss": 2.3913, "learning_rate": 9.991902834008098e-07, "epoch": 0.20428015564202334 }, { "current_steps": 105, "loss": 1.8581, "learning_rate": 9.987854251012145e-07, "epoch": 0.20622568093385213 }, { "current_steps": 106, "loss": 2.463, "learning_rate": 9.983805668016193e-07, "epoch": 0.20817120622568094 }, { "current_steps": 107, "loss": 2.3199, "learning_rate": 9.979757085020243e-07, "epoch": 0.21011673151750973 }, { "current_steps": 108, "loss": 2.4528, "learning_rate": 9.975708502024291e-07, "epoch": 0.21206225680933852 }, { "current_steps": 109, "loss": 2.5069, "learning_rate": 9.97165991902834e-07, "epoch": 0.2140077821011673 }, { "current_steps": 110, "loss": 1.8039, "learning_rate": 9.967611336032387e-07, "epoch": 0.21595330739299612 }, { "current_steps": 111, "loss": 2.428, "learning_rate": 9.963562753036437e-07, "epoch": 0.2178988326848249 }, { "current_steps": 112, "loss": 2.355, "learning_rate": 9.959514170040485e-07, "epoch": 0.2198443579766537 }, { "current_steps": 113, "loss": 2.3553, "learning_rate": 9.955465587044533e-07, "epoch": 0.22178988326848248 }, { "current_steps": 114, "loss": 2.277, "learning_rate": 9.951417004048583e-07, "epoch": 0.2237354085603113 }, { "current_steps": 115, "loss": 2.3615, "learning_rate": 9.947368421052631e-07, "epoch": 0.22568093385214008 }, { "current_steps": 116, "loss": 2.0366, "learning_rate": 9.94331983805668e-07, "epoch": 0.22762645914396887 }, { "current_steps": 117, "loss": 2.5816, "learning_rate": 9.93927125506073e-07, "epoch": 0.22957198443579765 }, { "current_steps": 118, "loss": 2.4922, "learning_rate": 9.935222672064777e-07, "epoch": 0.23151750972762647 }, { "current_steps": 119, "loss": 2.5298, "learning_rate": 9.931174089068825e-07, "epoch": 0.23346303501945526 }, { "current_steps": 120, "loss": 2.5506, "learning_rate": 9.927125506072875e-07, "epoch": 0.23540856031128404 }, { "current_steps": 121, "loss": 2.3781, "learning_rate": 9.923076923076923e-07, "epoch": 0.23735408560311283 }, { "current_steps": 122, "loss": 2.3625, "learning_rate": 9.91902834008097e-07, "epoch": 0.23929961089494164 }, { "current_steps": 123, "loss": 2.2802, "learning_rate": 9.91497975708502e-07, "epoch": 0.24124513618677043 }, { "current_steps": 124, "loss": 2.6978, "learning_rate": 9.91093117408907e-07, "epoch": 0.24319066147859922 }, { "current_steps": 125, "loss": 2.3335, "learning_rate": 9.906882591093117e-07, "epoch": 0.245136186770428 }, { "current_steps": 126, "loss": 2.4709, "learning_rate": 9.902834008097165e-07, "epoch": 0.24708171206225682 }, { "current_steps": 127, "loss": 2.7977, "learning_rate": 9.898785425101213e-07, "epoch": 0.2490272373540856 }, { "current_steps": 128, "loss": 2.6136, "learning_rate": 9.894736842105263e-07, "epoch": 0.2509727626459144 }, { "current_steps": 129, "loss": 2.6186, "learning_rate": 9.89068825910931e-07, "epoch": 0.2529182879377432 }, { "current_steps": 130, "loss": 2.3878, "learning_rate": 9.886639676113359e-07, "epoch": 0.25486381322957197 }, { "current_steps": 131, "loss": 2.0557, "learning_rate": 9.882591093117409e-07, "epoch": 0.25680933852140075 }, { "current_steps": 132, "loss": 2.2438, "learning_rate": 9.878542510121457e-07, "epoch": 0.2587548638132296 }, { "current_steps": 133, "loss": 2.3587, "learning_rate": 9.874493927125505e-07, "epoch": 0.2607003891050584 }, { "current_steps": 134, "loss": 2.4023, "learning_rate": 9.870445344129555e-07, "epoch": 0.26264591439688717 }, { "current_steps": 135, "loss": 2.3285, "learning_rate": 9.866396761133603e-07, "epoch": 0.26459143968871596 }, { "current_steps": 136, "loss": 2.23, "learning_rate": 9.862348178137653e-07, "epoch": 0.26653696498054474 }, { "current_steps": 137, "loss": 2.3833, "learning_rate": 9.8582995951417e-07, "epoch": 0.26848249027237353 }, { "current_steps": 138, "loss": 2.608, "learning_rate": 9.854251012145749e-07, "epoch": 0.2704280155642023 }, { "current_steps": 139, "loss": 2.338, "learning_rate": 9.850202429149799e-07, "epoch": 0.2723735408560311 }, { "current_steps": 140, "loss": 2.4685, "learning_rate": 9.846153846153847e-07, "epoch": 0.27431906614785995 }, { "current_steps": 141, "loss": 2.2811, "learning_rate": 9.842105263157894e-07, "epoch": 0.27626459143968873 }, { "current_steps": 142, "loss": 2.3012, "learning_rate": 9.838056680161942e-07, "epoch": 0.2782101167315175 }, { "current_steps": 143, "loss": 2.6326, "learning_rate": 9.83400809716599e-07, "epoch": 0.2801556420233463 }, { "current_steps": 144, "loss": 2.2647, "learning_rate": 9.82995951417004e-07, "epoch": 0.2821011673151751 }, { "current_steps": 145, "loss": 2.1206, "learning_rate": 9.825910931174088e-07, "epoch": 0.2840466926070039 }, { "current_steps": 146, "loss": 2.4821, "learning_rate": 9.821862348178136e-07, "epoch": 0.28599221789883267 }, { "current_steps": 147, "loss": 2.3386, "learning_rate": 9.817813765182186e-07, "epoch": 0.28793774319066145 }, { "current_steps": 148, "loss": 2.3018, "learning_rate": 9.813765182186234e-07, "epoch": 0.2898832684824903 }, { "current_steps": 149, "loss": 2.301, "learning_rate": 9.809716599190282e-07, "epoch": 0.2918287937743191 }, { "current_steps": 150, "loss": 2.2471, "learning_rate": 9.805668016194332e-07, "epoch": 0.29377431906614787 }, { "current_steps": 151, "loss": 1.8782, "learning_rate": 9.80161943319838e-07, "epoch": 0.29571984435797666 }, { "current_steps": 152, "loss": 1.9819, "learning_rate": 9.797570850202428e-07, "epoch": 0.29766536964980544 }, { "current_steps": 153, "loss": 2.457, "learning_rate": 9.793522267206478e-07, "epoch": 0.29961089494163423 }, { "current_steps": 154, "loss": 2.3722, "learning_rate": 9.789473684210526e-07, "epoch": 0.301556420233463 }, { "current_steps": 155, "loss": 2.285, "learning_rate": 9.785425101214574e-07, "epoch": 0.3035019455252918 }, { "current_steps": 156, "loss": 1.963, "learning_rate": 9.781376518218622e-07, "epoch": 0.30544747081712065 }, { "current_steps": 157, "loss": 2.1686, "learning_rate": 9.777327935222672e-07, "epoch": 0.30739299610894943 }, { "current_steps": 158, "loss": 1.6872, "learning_rate": 9.77327935222672e-07, "epoch": 0.3093385214007782 }, { "current_steps": 159, "loss": 1.9588, "learning_rate": 9.769230769230768e-07, "epoch": 0.311284046692607 }, { "current_steps": 160, "loss": 2.3249, "learning_rate": 9.765182186234818e-07, "epoch": 0.3132295719844358 }, { "current_steps": 161, "loss": 2.477, "learning_rate": 9.761133603238866e-07, "epoch": 0.3151750972762646 }, { "current_steps": 162, "loss": 2.3452, "learning_rate": 9.757085020242914e-07, "epoch": 0.31712062256809337 }, { "current_steps": 163, "loss": 2.3242, "learning_rate": 9.753036437246964e-07, "epoch": 0.31906614785992216 }, { "current_steps": 164, "loss": 1.7265, "learning_rate": 9.748987854251012e-07, "epoch": 0.321011673151751 }, { "current_steps": 165, "loss": 2.1968, "learning_rate": 9.74493927125506e-07, "epoch": 0.3229571984435798 }, { "current_steps": 166, "loss": 2.3578, "learning_rate": 9.74089068825911e-07, "epoch": 0.32490272373540857 }, { "current_steps": 167, "loss": 2.1535, "learning_rate": 9.736842105263158e-07, "epoch": 0.32684824902723736 }, { "current_steps": 168, "loss": 2.4985, "learning_rate": 9.732793522267206e-07, "epoch": 0.32879377431906615 }, { "current_steps": 169, "loss": 2.3781, "learning_rate": 9.728744939271256e-07, "epoch": 0.33073929961089493 }, { "current_steps": 170, "loss": 2.3188, "learning_rate": 9.724696356275304e-07, "epoch": 0.3326848249027237 }, { "current_steps": 171, "loss": 2.5229, "learning_rate": 9.720647773279352e-07, "epoch": 0.3346303501945525 }, { "current_steps": 172, "loss": 2.3015, "learning_rate": 9.7165991902834e-07, "epoch": 0.33657587548638135 }, { "current_steps": 173, "loss": 2.2615, "learning_rate": 9.71255060728745e-07, "epoch": 0.33852140077821014 }, { "current_steps": 174, "loss": 2.1693, "learning_rate": 9.708502024291498e-07, "epoch": 0.3404669260700389 }, { "current_steps": 175, "loss": 2.4246, "learning_rate": 9.704453441295545e-07, "epoch": 0.3424124513618677 }, { "current_steps": 176, "loss": 1.9115, "learning_rate": 9.700404858299596e-07, "epoch": 0.3443579766536965 }, { "current_steps": 177, "loss": 2.5014, "learning_rate": 9.696356275303643e-07, "epoch": 0.3463035019455253 }, { "current_steps": 178, "loss": 2.7675, "learning_rate": 9.692307692307691e-07, "epoch": 0.34824902723735407 }, { "current_steps": 179, "loss": 2.5294, "learning_rate": 9.688259109311741e-07, "epoch": 0.35019455252918286 }, { "current_steps": 180, "loss": 2.0494, "learning_rate": 9.68421052631579e-07, "epoch": 0.3521400778210117 }, { "current_steps": 181, "loss": 2.3663, "learning_rate": 9.680161943319837e-07, "epoch": 0.3540856031128405 }, { "current_steps": 182, "loss": 2.0957, "learning_rate": 9.676113360323887e-07, "epoch": 0.3560311284046693 }, { "current_steps": 183, "loss": 2.3879, "learning_rate": 9.672064777327935e-07, "epoch": 0.35797665369649806 }, { "current_steps": 184, "loss": 2.0649, "learning_rate": 9.668016194331983e-07, "epoch": 0.35992217898832685 }, { "current_steps": 185, "loss": 2.2803, "learning_rate": 9.663967611336033e-07, "epoch": 0.36186770428015563 }, { "current_steps": 186, "loss": 2.5002, "learning_rate": 9.65991902834008e-07, "epoch": 0.3638132295719844 }, { "current_steps": 187, "loss": 2.4765, "learning_rate": 9.65587044534413e-07, "epoch": 0.3657587548638132 }, { "current_steps": 188, "loss": 2.4841, "learning_rate": 9.651821862348177e-07, "epoch": 0.36770428015564205 }, { "current_steps": 189, "loss": 1.8182, "learning_rate": 9.647773279352225e-07, "epoch": 0.36964980544747084 }, { "current_steps": 190, "loss": 2.1865, "learning_rate": 9.643724696356275e-07, "epoch": 0.3715953307392996 }, { "current_steps": 191, "loss": 2.545, "learning_rate": 9.639676113360323e-07, "epoch": 0.3735408560311284 }, { "current_steps": 192, "loss": 2.3072, "learning_rate": 9.63562753036437e-07, "epoch": 0.3754863813229572 }, { "current_steps": 193, "loss": 2.357, "learning_rate": 9.63157894736842e-07, "epoch": 0.377431906614786 }, { "current_steps": 194, "loss": 2.6136, "learning_rate": 9.62753036437247e-07, "epoch": 0.37937743190661477 }, { "current_steps": 195, "loss": 2.4342, "learning_rate": 9.623481781376517e-07, "epoch": 0.38132295719844356 }, { "current_steps": 196, "loss": 2.6144, "learning_rate": 9.619433198380567e-07, "epoch": 0.3832684824902724 }, { "current_steps": 197, "loss": 2.4396, "learning_rate": 9.615384615384615e-07, "epoch": 0.3852140077821012 }, { "current_steps": 198, "loss": 2.2339, "learning_rate": 9.611336032388663e-07, "epoch": 0.38715953307393 }, { "current_steps": 199, "loss": 2.0303, "learning_rate": 9.607287449392713e-07, "epoch": 0.38910505836575876 }, { "current_steps": 200, "loss": 2.5061, "learning_rate": 9.60323886639676e-07, "epoch": 0.39105058365758755 }, { "current_steps": 201, "loss": 1.8428, "learning_rate": 9.59919028340081e-07, "epoch": 0.39299610894941633 }, { "current_steps": 202, "loss": 2.1775, "learning_rate": 9.595141700404857e-07, "epoch": 0.3949416342412451 }, { "current_steps": 203, "loss": 2.2355, "learning_rate": 9.591093117408907e-07, "epoch": 0.3968871595330739 }, { "current_steps": 204, "loss": 2.1281, "learning_rate": 9.587044534412955e-07, "epoch": 0.39883268482490275 }, { "current_steps": 205, "loss": 2.1935, "learning_rate": 9.582995951417003e-07, "epoch": 0.40077821011673154 }, { "current_steps": 206, "loss": 2.2388, "learning_rate": 9.578947368421053e-07, "epoch": 0.4027237354085603 }, { "current_steps": 207, "loss": 2.3995, "learning_rate": 9.5748987854251e-07, "epoch": 0.4046692607003891 }, { "current_steps": 208, "loss": 2.3024, "learning_rate": 9.570850202429149e-07, "epoch": 0.4066147859922179 }, { "current_steps": 209, "loss": 2.3294, "learning_rate": 9.566801619433199e-07, "epoch": 0.4085603112840467 }, { "current_steps": 210, "loss": 2.4474, "learning_rate": 9.562753036437247e-07, "epoch": 0.41050583657587547 }, { "current_steps": 211, "loss": 2.3296, "learning_rate": 9.558704453441294e-07, "epoch": 0.41245136186770426 }, { "current_steps": 212, "loss": 2.3921, "learning_rate": 9.554655870445344e-07, "epoch": 0.4143968871595331 }, { "current_steps": 213, "loss": 1.9843, "learning_rate": 9.550607287449392e-07, "epoch": 0.4163424124513619 }, { "current_steps": 214, "loss": 2.3329, "learning_rate": 9.54655870445344e-07, "epoch": 0.4182879377431907 }, { "current_steps": 215, "loss": 2.5095, "learning_rate": 9.54251012145749e-07, "epoch": 0.42023346303501946 }, { "current_steps": 216, "loss": 2.2189, "learning_rate": 9.538461538461538e-07, "epoch": 0.42217898832684825 }, { "current_steps": 217, "loss": 2.2471, "learning_rate": 9.534412955465586e-07, "epoch": 0.42412451361867703 }, { "current_steps": 218, "loss": 2.3316, "learning_rate": 9.530364372469635e-07, "epoch": 0.4260700389105058 }, { "current_steps": 219, "loss": 2.6414, "learning_rate": 9.530364372469635e-07, "epoch": 0.4280155642023346 }, { "current_steps": 220, "loss": 2.3409, "learning_rate": 9.526315789473683e-07, "epoch": 0.42996108949416345 }, { "current_steps": 221, "loss": 2.416, "learning_rate": 9.522267206477732e-07, "epoch": 0.43190661478599224 }, { "current_steps": 222, "loss": 2.7432, "learning_rate": 9.518218623481781e-07, "epoch": 0.433852140077821 }, { "current_steps": 223, "loss": 2.0675, "learning_rate": 9.514170040485829e-07, "epoch": 0.4357976653696498 }, { "current_steps": 224, "loss": 2.4429, "learning_rate": 9.510121457489878e-07, "epoch": 0.4377431906614786 }, { "current_steps": 225, "loss": 2.4737, "learning_rate": 9.506072874493927e-07, "epoch": 0.4396887159533074 }, { "current_steps": 226, "loss": 2.5619, "learning_rate": 9.502024291497975e-07, "epoch": 0.44163424124513617 }, { "current_steps": 227, "loss": 2.1069, "learning_rate": 9.497975708502024e-07, "epoch": 0.44357976653696496 }, { "current_steps": 228, "loss": 2.2065, "learning_rate": 9.493927125506073e-07, "epoch": 0.4455252918287938 }, { "current_steps": 229, "loss": 2.4871, "learning_rate": 9.489878542510121e-07, "epoch": 0.4474708171206226 }, { "current_steps": 230, "loss": 2.3297, "learning_rate": 9.48582995951417e-07, "epoch": 0.4494163424124514 }, { "current_steps": 231, "loss": 2.0088, "learning_rate": 9.481781376518219e-07, "epoch": 0.45136186770428016 }, { "current_steps": 232, "loss": 2.412, "learning_rate": 9.477732793522267e-07, "epoch": 0.45330739299610895 }, { "current_steps": 233, "loss": 2.4837, "learning_rate": 9.473684210526315e-07, "epoch": 0.45525291828793774 }, { "current_steps": 234, "loss": 2.468, "learning_rate": 9.469635627530364e-07, "epoch": 0.4571984435797665 }, { "current_steps": 235, "loss": 2.2149, "learning_rate": 9.465587044534412e-07, "epoch": 0.4591439688715953 }, { "current_steps": 236, "loss": 2.3014, "learning_rate": 9.461538461538461e-07, "epoch": 0.46108949416342415 }, { "current_steps": 237, "loss": 2.3058, "learning_rate": 9.45748987854251e-07, "epoch": 0.46303501945525294 }, { "current_steps": 238, "loss": 2.3221, "learning_rate": 9.453441295546558e-07, "epoch": 0.4649805447470817 }, { "current_steps": 239, "loss": 2.426, "learning_rate": 9.449392712550607e-07, "epoch": 0.4669260700389105 }, { "current_steps": 240, "loss": 1.7225, "learning_rate": 9.445344129554656e-07, "epoch": 0.4688715953307393 }, { "current_steps": 241, "loss": 2.16, "learning_rate": 9.441295546558704e-07, "epoch": 0.4708171206225681 }, { "current_steps": 242, "loss": 1.8884, "learning_rate": 9.437246963562753e-07, "epoch": 0.4727626459143969 }, { "current_steps": 243, "loss": 2.3278, "learning_rate": 9.433198380566802e-07, "epoch": 0.47470817120622566 }, { "current_steps": 244, "loss": 2.447, "learning_rate": 9.42914979757085e-07, "epoch": 0.4766536964980545 }, { "current_steps": 245, "loss": 2.6211, "learning_rate": 9.425101214574899e-07, "epoch": 0.4785992217898833 }, { "current_steps": 246, "loss": 2.0072, "learning_rate": 9.421052631578948e-07, "epoch": 0.4805447470817121 }, { "current_steps": 247, "loss": 1.8567, "learning_rate": 9.417004048582995e-07, "epoch": 0.48249027237354086 }, { "current_steps": 248, "loss": 2.2566, "learning_rate": 9.412955465587043e-07, "epoch": 0.48443579766536965 }, { "current_steps": 249, "loss": 2.6597, "learning_rate": 9.408906882591092e-07, "epoch": 0.48638132295719844 }, { "current_steps": 250, "loss": 1.9404, "learning_rate": 9.40485829959514e-07, "epoch": 0.4883268482490272 }, { "current_steps": 251, "loss": 2.1916, "learning_rate": 9.400809716599189e-07, "epoch": 0.490272373540856 }, { "current_steps": 252, "loss": 2.2207, "learning_rate": 9.396761133603238e-07, "epoch": 0.49221789883268485 }, { "current_steps": 253, "loss": 2.3397, "learning_rate": 9.392712550607287e-07, "epoch": 0.49416342412451364 }, { "current_steps": 254, "loss": 2.0964, "learning_rate": 9.388663967611335e-07, "epoch": 0.4961089494163424 }, { "current_steps": 255, "loss": 2.1752, "learning_rate": 9.384615384615384e-07, "epoch": 0.4980544747081712 }, { "current_steps": 256, "loss": 2.0168, "learning_rate": 9.380566801619433e-07, "epoch": 0.5 }, { "current_steps": 257, "loss": 2.0272, "learning_rate": 9.376518218623481e-07, "epoch": 0.5019455252918288 }, { "current_steps": 258, "loss": 2.4508, "learning_rate": 9.37246963562753e-07, "epoch": 0.5038910505836576 }, { "current_steps": 259, "loss": 2.1871, "learning_rate": 9.368421052631579e-07, "epoch": 0.5058365758754864 }, { "current_steps": 260, "loss": 2.4577, "learning_rate": 9.364372469635627e-07, "epoch": 0.5077821011673151 }, { "current_steps": 261, "loss": 2.4585, "learning_rate": 9.360323886639676e-07, "epoch": 0.5097276264591439 }, { "current_steps": 262, "loss": 2.1612, "learning_rate": 9.356275303643725e-07, "epoch": 0.5116731517509727 }, { "current_steps": 263, "loss": 2.2017, "learning_rate": 9.352226720647773e-07, "epoch": 0.5136186770428015 }, { "current_steps": 264, "loss": 2.2888, "learning_rate": 9.348178137651821e-07, "epoch": 0.5155642023346303 }, { "current_steps": 265, "loss": 1.5021, "learning_rate": 9.34412955465587e-07, "epoch": 0.5175097276264592 }, { "current_steps": 266, "loss": 1.6685, "learning_rate": 9.340080971659918e-07, "epoch": 0.519455252918288 }, { "current_steps": 267, "loss": 2.3715, "learning_rate": 9.336032388663967e-07, "epoch": 0.5214007782101168 }, { "current_steps": 268, "loss": 2.299, "learning_rate": 9.331983805668016e-07, "epoch": 0.5233463035019456 }, { "current_steps": 269, "loss": 2.1602, "learning_rate": 9.327935222672064e-07, "epoch": 0.5252918287937743 }, { "current_steps": 270, "loss": 2.3834, "learning_rate": 9.323886639676113e-07, "epoch": 0.5272373540856031 }, { "current_steps": 271, "loss": 2.2494, "learning_rate": 9.319838056680162e-07, "epoch": 0.5291828793774319 }, { "current_steps": 272, "loss": 2.1672, "learning_rate": 9.31578947368421e-07, "epoch": 0.5311284046692607 }, { "current_steps": 273, "loss": 2.4842, "learning_rate": 9.311740890688259e-07, "epoch": 0.5330739299610895 }, { "current_steps": 274, "loss": 2.6121, "learning_rate": 9.307692307692308e-07, "epoch": 0.5350194552529183 }, { "current_steps": 275, "loss": 2.54, "learning_rate": 9.303643724696356e-07, "epoch": 0.5369649805447471 }, { "current_steps": 276, "loss": 2.4968, "learning_rate": 9.299595141700405e-07, "epoch": 0.5389105058365758 }, { "current_steps": 277, "loss": 2.089, "learning_rate": 9.295546558704454e-07, "epoch": 0.5408560311284046 }, { "current_steps": 278, "loss": 2.2037, "learning_rate": 9.291497975708502e-07, "epoch": 0.5428015564202334 }, { "current_steps": 279, "loss": 2.2682, "learning_rate": 9.28744939271255e-07, "epoch": 0.5447470817120622 }, { "current_steps": 280, "loss": 2.2168, "learning_rate": 9.283400809716599e-07, "epoch": 0.546692607003891 }, { "current_steps": 281, "loss": 1.9233, "learning_rate": 9.279352226720646e-07, "epoch": 0.5486381322957199 }, { "current_steps": 282, "loss": 2.2471, "learning_rate": 9.275303643724695e-07, "epoch": 0.5505836575875487 }, { "current_steps": 283, "loss": 2.3821, "learning_rate": 9.271255060728744e-07, "epoch": 0.5525291828793775 }, { "current_steps": 284, "loss": 2.2343, "learning_rate": 9.267206477732793e-07, "epoch": 0.5544747081712063 }, { "current_steps": 285, "loss": 2.4054, "learning_rate": 9.263157894736841e-07, "epoch": 0.556420233463035 }, { "current_steps": 286, "loss": 2.2476, "learning_rate": 9.25910931174089e-07, "epoch": 0.5583657587548638 }, { "current_steps": 287, "loss": 2.407, "learning_rate": 9.255060728744939e-07, "epoch": 0.5603112840466926 }, { "current_steps": 288, "loss": 2.4795, "learning_rate": 9.251012145748987e-07, "epoch": 0.5622568093385214 }, { "current_steps": 289, "loss": 2.2916, "learning_rate": 9.246963562753036e-07, "epoch": 0.5642023346303502 }, { "current_steps": 290, "loss": 2.0932, "learning_rate": 9.242914979757085e-07, "epoch": 0.566147859922179 }, { "current_steps": 291, "loss": 2.1759, "learning_rate": 9.238866396761133e-07, "epoch": 0.5680933852140078 }, { "current_steps": 292, "loss": 2.3665, "learning_rate": 9.234817813765182e-07, "epoch": 0.5700389105058365 }, { "current_steps": 293, "loss": 2.1188, "learning_rate": 9.230769230769231e-07, "epoch": 0.5719844357976653 }, { "current_steps": 294, "loss": 2.3752, "learning_rate": 9.226720647773278e-07, "epoch": 0.5739299610894941 }, { "current_steps": 295, "loss": 2.6102, "learning_rate": 9.222672064777327e-07, "epoch": 0.5758754863813229 }, { "current_steps": 296, "loss": 2.4774, "learning_rate": 9.218623481781376e-07, "epoch": 0.5778210116731517 }, { "current_steps": 297, "loss": 2.424, "learning_rate": 9.214574898785424e-07, "epoch": 0.5797665369649806 }, { "current_steps": 298, "loss": 2.395, "learning_rate": 9.210526315789473e-07, "epoch": 0.5817120622568094 }, { "current_steps": 299, "loss": 2.2527, "learning_rate": 9.206477732793522e-07, "epoch": 0.5836575875486382 }, { "current_steps": 300, "loss": 1.7941, "learning_rate": 9.20242914979757e-07, "epoch": 0.585603112840467 }, { "current_steps": 301, "loss": 2.4797, "learning_rate": 9.198380566801619e-07, "epoch": 0.5875486381322957 }, { "current_steps": 302, "loss": 2.3082, "learning_rate": 9.194331983805668e-07, "epoch": 0.5894941634241245 }, { "current_steps": 303, "loss": 1.769, "learning_rate": 9.190283400809716e-07, "epoch": 0.5914396887159533 }, { "current_steps": 304, "loss": 2.1947, "learning_rate": 9.186234817813765e-07, "epoch": 0.5933852140077821 }, { "current_steps": 305, "loss": 2.491, "learning_rate": 9.182186234817814e-07, "epoch": 0.5953307392996109 }, { "current_steps": 306, "loss": 2.5071, "learning_rate": 9.178137651821862e-07, "epoch": 0.5972762645914397 }, { "current_steps": 307, "loss": 1.9217, "learning_rate": 9.174089068825911e-07, "epoch": 0.5992217898832685 }, { "current_steps": 308, "loss": 2.5751, "learning_rate": 9.17004048582996e-07, "epoch": 0.6011673151750972 }, { "current_steps": 309, "loss": 2.1265, "learning_rate": 9.165991902834007e-07, "epoch": 0.603112840466926 }, { "current_steps": 310, "loss": 2.3467, "learning_rate": 9.161943319838056e-07, "epoch": 0.6050583657587548 }, { "current_steps": 311, "loss": 2.182, "learning_rate": 9.157894736842105e-07, "epoch": 0.6070038910505836 }, { "current_steps": 312, "loss": 2.4211, "learning_rate": 9.153846153846153e-07, "epoch": 0.6089494163424124 }, { "current_steps": 313, "loss": 2.0385, "learning_rate": 9.149797570850202e-07, "epoch": 0.6108949416342413 }, { "current_steps": 314, "loss": 2.3833, "learning_rate": 9.145748987854251e-07, "epoch": 0.6128404669260701 }, { "current_steps": 315, "loss": 2.3259, "learning_rate": 9.1417004048583e-07, "epoch": 0.6147859922178989 }, { "current_steps": 316, "loss": 2.493, "learning_rate": 9.137651821862348e-07, "epoch": 0.6167315175097277 }, { "current_steps": 317, "loss": 2.3863, "learning_rate": 9.133603238866397e-07, "epoch": 0.6186770428015564 }, { "current_steps": 318, "loss": 2.6036, "learning_rate": 9.129554655870446e-07, "epoch": 0.6206225680933852 }, { "current_steps": 319, "loss": 2.2975, "learning_rate": 9.125506072874493e-07, "epoch": 0.622568093385214 }, { "current_steps": 320, "loss": 2.4009, "learning_rate": 9.121457489878542e-07, "epoch": 0.6245136186770428 }, { "current_steps": 321, "loss": 1.7818, "learning_rate": 9.117408906882591e-07, "epoch": 0.6264591439688716 }, { "current_steps": 322, "loss": 2.2009, "learning_rate": 9.113360323886639e-07, "epoch": 0.6284046692607004 }, { "current_steps": 323, "loss": 2.105, "learning_rate": 9.109311740890688e-07, "epoch": 0.6303501945525292 }, { "current_steps": 324, "loss": 2.2834, "learning_rate": 9.105263157894737e-07, "epoch": 0.632295719844358 }, { "current_steps": 325, "loss": 2.1864, "learning_rate": 9.101214574898784e-07, "epoch": 0.6342412451361867 }, { "current_steps": 326, "loss": 2.4228, "learning_rate": 9.097165991902833e-07, "epoch": 0.6361867704280155 }, { "current_steps": 327, "loss": 2.2421, "learning_rate": 9.093117408906882e-07, "epoch": 0.6381322957198443 }, { "current_steps": 328, "loss": 2.4945, "learning_rate": 9.08906882591093e-07, "epoch": 0.6400778210116731 }, { "current_steps": 329, "loss": 2.2722, "learning_rate": 9.085020242914979e-07, "epoch": 0.642023346303502 }, { "current_steps": 330, "loss": 1.8805, "learning_rate": 9.080971659919028e-07, "epoch": 0.6439688715953308 }, { "current_steps": 331, "loss": 2.3346, "learning_rate": 9.076923076923076e-07, "epoch": 0.6459143968871596 }, { "current_steps": 332, "loss": 2.2883, "learning_rate": 9.072874493927125e-07, "epoch": 0.6478599221789884 }, { "current_steps": 333, "loss": 2.142, "learning_rate": 9.068825910931174e-07, "epoch": 0.6498054474708171 }, { "current_steps": 334, "loss": 2.5224, "learning_rate": 9.064777327935222e-07, "epoch": 0.6517509727626459 }, { "current_steps": 335, "loss": 2.3542, "learning_rate": 9.060728744939271e-07, "epoch": 0.6536964980544747 }, { "current_steps": 336, "loss": 2.7632, "learning_rate": 9.05668016194332e-07, "epoch": 0.6556420233463035 }, { "current_steps": 337, "loss": 2.1694, "learning_rate": 9.052631578947368e-07, "epoch": 0.6575875486381323 }, { "current_steps": 338, "loss": 2.3805, "learning_rate": 9.048582995951417e-07, "epoch": 0.6595330739299611 }, { "current_steps": 339, "loss": 2.3513, "learning_rate": 9.044534412955466e-07, "epoch": 0.6614785992217899 }, { "current_steps": 340, "loss": 2.3352, "learning_rate": 9.040485829959513e-07, "epoch": 0.6634241245136187 }, { "current_steps": 341, "loss": 2.253, "learning_rate": 9.036437246963562e-07, "epoch": 0.6653696498054474 }, { "current_steps": 342, "loss": 2.3027, "learning_rate": 9.032388663967611e-07, "epoch": 0.6673151750972762 }, { "current_steps": 343, "loss": 2.3441, "learning_rate": 9.028340080971659e-07, "epoch": 0.669260700389105 }, { "current_steps": 344, "loss": 2.1845, "learning_rate": 9.024291497975708e-07, "epoch": 0.6712062256809338 }, { "current_steps": 345, "loss": 2.538, "learning_rate": 9.020242914979757e-07, "epoch": 0.6731517509727627 }, { "current_steps": 346, "loss": 2.3014, "learning_rate": 9.016194331983805e-07, "epoch": 0.6750972762645915 }, { "current_steps": 347, "loss": 2.1807, "learning_rate": 9.012145748987854e-07, "epoch": 0.6770428015564203 }, { "current_steps": 348, "loss": 1.7949, "learning_rate": 9.008097165991903e-07, "epoch": 0.6789883268482491 }, { "current_steps": 349, "loss": 2.3092, "learning_rate": 9.004048582995952e-07, "epoch": 0.6809338521400778 }, { "current_steps": 350, "loss": 2.5602, "learning_rate": 9e-07, "epoch": 0.6828793774319066 }, { "current_steps": 351, "loss": 2.6082, "learning_rate": 8.995951417004049e-07, "epoch": 0.6848249027237354 }, { "current_steps": 352, "loss": 2.3473, "learning_rate": 8.991902834008098e-07, "epoch": 0.6867704280155642 }, { "current_steps": 353, "loss": 2.3422, "learning_rate": 8.987854251012146e-07, "epoch": 0.688715953307393 }, { "current_steps": 354, "loss": 2.3167, "learning_rate": 8.983805668016195e-07, "epoch": 0.6906614785992218 }, { "current_steps": 355, "loss": 2.1183, "learning_rate": 8.979757085020242e-07, "epoch": 0.6926070038910506 }, { "current_steps": 356, "loss": 2.0557, "learning_rate": 8.97570850202429e-07, "epoch": 0.6945525291828794 }, { "current_steps": 357, "loss": 2.3471, "learning_rate": 8.971659919028339e-07, "epoch": 0.6964980544747081 }, { "current_steps": 358, "loss": 2.1335, "learning_rate": 8.967611336032388e-07, "epoch": 0.6984435797665369 }, { "current_steps": 359, "loss": 2.233, "learning_rate": 8.963562753036436e-07, "epoch": 0.7003891050583657 }, { "current_steps": 360, "loss": 2.142, "learning_rate": 8.959514170040485e-07, "epoch": 0.7023346303501945 }, { "current_steps": 361, "loss": 2.4992, "learning_rate": 8.955465587044534e-07, "epoch": 0.7042801556420234 }, { "current_steps": 362, "loss": 2.1617, "learning_rate": 8.951417004048582e-07, "epoch": 0.7062256809338522 }, { "current_steps": 363, "loss": 2.2386, "learning_rate": 8.947368421052631e-07, "epoch": 0.708171206225681 }, { "current_steps": 364, "loss": 2.1248, "learning_rate": 8.94331983805668e-07, "epoch": 0.7101167315175098 }, { "current_steps": 365, "loss": 2.276, "learning_rate": 8.939271255060728e-07, "epoch": 0.7120622568093385 }, { "current_steps": 366, "loss": 2.3365, "learning_rate": 8.935222672064777e-07, "epoch": 0.7140077821011673 }, { "current_steps": 367, "loss": 2.2751, "learning_rate": 8.931174089068826e-07, "epoch": 0.7159533073929961 }, { "current_steps": 368, "loss": 2.4098, "learning_rate": 8.927125506072874e-07, "epoch": 0.7178988326848249 }, { "current_steps": 369, "loss": 2.1223, "learning_rate": 8.923076923076923e-07, "epoch": 0.7198443579766537 }, { "current_steps": 370, "loss": 2.2619, "learning_rate": 8.919028340080971e-07, "epoch": 0.7217898832684825 }, { "current_steps": 371, "loss": 2.4654, "learning_rate": 8.914979757085019e-07, "epoch": 0.7237354085603113 }, { "current_steps": 372, "loss": 1.9212, "learning_rate": 8.910931174089068e-07, "epoch": 0.72568093385214 }, { "current_steps": 373, "loss": 2.2462, "learning_rate": 8.906882591093117e-07, "epoch": 0.7276264591439688 }, { "current_steps": 374, "loss": 2.3605, "learning_rate": 8.902834008097165e-07, "epoch": 0.7295719844357976 }, { "current_steps": 375, "loss": 2.2557, "learning_rate": 8.898785425101214e-07, "epoch": 0.7315175097276264 }, { "current_steps": 376, "loss": 2.4447, "learning_rate": 8.894736842105263e-07, "epoch": 0.7334630350194552 }, { "current_steps": 377, "loss": 1.7057, "learning_rate": 8.890688259109311e-07, "epoch": 0.7354085603112841 }, { "current_steps": 378, "loss": 2.2001, "learning_rate": 8.88663967611336e-07, "epoch": 0.7373540856031129 }, { "current_steps": 379, "loss": 2.4257, "learning_rate": 8.882591093117409e-07, "epoch": 0.7392996108949417 }, { "current_steps": 380, "loss": 2.376, "learning_rate": 8.878542510121458e-07, "epoch": 0.7412451361867705 }, { "current_steps": 381, "loss": 2.1511, "learning_rate": 8.874493927125506e-07, "epoch": 0.7431906614785992 }, { "current_steps": 382, "loss": 2.4626, "learning_rate": 8.870445344129555e-07, "epoch": 0.745136186770428 }, { "current_steps": 383, "loss": 2.2466, "learning_rate": 8.866396761133604e-07, "epoch": 0.7470817120622568 }, { "current_steps": 384, "loss": 1.8923, "learning_rate": 8.862348178137652e-07, "epoch": 0.7490272373540856 }, { "current_steps": 385, "loss": 2.4609, "learning_rate": 8.858299595141701e-07, "epoch": 0.7509727626459144 }, { "current_steps": 386, "loss": 2.3103, "learning_rate": 8.854251012145749e-07, "epoch": 0.7529182879377432 }, { "current_steps": 387, "loss": 2.3556, "learning_rate": 8.850202429149797e-07, "epoch": 0.754863813229572 }, { "current_steps": 388, "loss": 1.8511, "learning_rate": 8.846153846153846e-07, "epoch": 0.7568093385214008 }, { "current_steps": 389, "loss": 2.0744, "learning_rate": 8.842105263157895e-07, "epoch": 0.7587548638132295 }, { "current_steps": 390, "loss": 1.9669, "learning_rate": 8.838056680161942e-07, "epoch": 0.7607003891050583 }, { "current_steps": 391, "loss": 2.3148, "learning_rate": 8.834008097165991e-07, "epoch": 0.7626459143968871 }, { "current_steps": 392, "loss": 2.3064, "learning_rate": 8.82995951417004e-07, "epoch": 0.7645914396887159 }, { "current_steps": 393, "loss": 2.2556, "learning_rate": 8.825910931174088e-07, "epoch": 0.7665369649805448 }, { "current_steps": 394, "loss": 1.7605, "learning_rate": 8.821862348178137e-07, "epoch": 0.7684824902723736 }, { "current_steps": 395, "loss": 2.3886, "learning_rate": 8.817813765182186e-07, "epoch": 0.7704280155642024 }, { "current_steps": 396, "loss": 2.2929, "learning_rate": 8.813765182186234e-07, "epoch": 0.7723735408560312 }, { "current_steps": 397, "loss": 1.9885, "learning_rate": 8.809716599190283e-07, "epoch": 0.77431906614786 }, { "current_steps": 398, "loss": 2.3393, "learning_rate": 8.805668016194332e-07, "epoch": 0.7762645914396887 }, { "current_steps": 399, "loss": 2.3985, "learning_rate": 8.80161943319838e-07, "epoch": 0.7782101167315175 }, { "current_steps": 400, "loss": 2.2485, "learning_rate": 8.797570850202429e-07, "epoch": 0.7801556420233463 }, { "current_steps": 401, "loss": 2.5363, "learning_rate": 8.793522267206477e-07, "epoch": 0.7821011673151751 }, { "current_steps": 402, "loss": 2.3025, "learning_rate": 8.789473684210525e-07, "epoch": 0.7840466926070039 }, { "current_steps": 403, "loss": 2.3757, "learning_rate": 8.785425101214574e-07, "epoch": 0.7859922178988327 }, { "current_steps": 404, "loss": 2.1859, "learning_rate": 8.781376518218623e-07, "epoch": 0.7879377431906615 }, { "current_steps": 405, "loss": 2.3152, "learning_rate": 8.777327935222671e-07, "epoch": 0.7898832684824902 }, { "current_steps": 406, "loss": 1.7867, "learning_rate": 8.77327935222672e-07, "epoch": 0.791828793774319 }, { "current_steps": 407, "loss": 2.3342, "learning_rate": 8.769230769230769e-07, "epoch": 0.7937743190661478 }, { "current_steps": 408, "loss": 2.4702, "learning_rate": 8.765182186234817e-07, "epoch": 0.7957198443579766 }, { "current_steps": 409, "loss": 2.4636, "learning_rate": 8.761133603238866e-07, "epoch": 0.7976653696498055 }, { "current_steps": 410, "loss": 2.4802, "learning_rate": 8.757085020242915e-07, "epoch": 0.7996108949416343 }, { "current_steps": 411, "loss": 2.6187, "learning_rate": 8.753036437246964e-07, "epoch": 0.8015564202334631 }, { "current_steps": 412, "loss": 1.7969, "learning_rate": 8.748987854251012e-07, "epoch": 0.8035019455252919 }, { "current_steps": 413, "loss": 2.3779, "learning_rate": 8.744939271255061e-07, "epoch": 0.8054474708171206 }, { "current_steps": 414, "loss": 1.9829, "learning_rate": 8.74089068825911e-07, "epoch": 0.8073929961089494 }, { "current_steps": 415, "loss": 2.0526, "learning_rate": 8.736842105263158e-07, "epoch": 0.8093385214007782 }, { "current_steps": 416, "loss": 2.3931, "learning_rate": 8.732793522267206e-07, "epoch": 0.811284046692607 }, { "current_steps": 417, "loss": 2.226, "learning_rate": 8.728744939271255e-07, "epoch": 0.8132295719844358 }, { "current_steps": 418, "loss": 2.2915, "learning_rate": 8.724696356275303e-07, "epoch": 0.8151750972762646 }, { "current_steps": 419, "loss": 2.3993, "learning_rate": 8.720647773279352e-07, "epoch": 0.8171206225680934 }, { "current_steps": 420, "loss": 2.4315, "learning_rate": 8.716599190283401e-07, "epoch": 0.8190661478599222 }, { "current_steps": 421, "loss": 2.3831, "learning_rate": 8.712550607287449e-07, "epoch": 0.8210116731517509 }, { "current_steps": 422, "loss": 2.3439, "learning_rate": 8.708502024291498e-07, "epoch": 0.8229571984435797 }, { "current_steps": 423, "loss": 1.47, "learning_rate": 8.704453441295547e-07, "epoch": 0.8249027237354085 }, { "current_steps": 424, "loss": 1.7815, "learning_rate": 8.700404858299594e-07, "epoch": 0.8268482490272373 }, { "current_steps": 425, "loss": 1.8567, "learning_rate": 8.700404858299594e-07, "epoch": 0.8287937743190662 }, { "current_steps": 426, "loss": 2.2315, "learning_rate": 8.696356275303643e-07, "epoch": 0.830739299610895 }, { "current_steps": 427, "loss": 1.8933, "learning_rate": 8.692307692307692e-07, "epoch": 0.8326848249027238 }, { "current_steps": 428, "loss": 2.3877, "learning_rate": 8.68825910931174e-07, "epoch": 0.8346303501945526 }, { "current_steps": 429, "loss": 2.2213, "learning_rate": 8.684210526315789e-07, "epoch": 0.8365758754863813 }, { "current_steps": 430, "loss": 2.3486, "learning_rate": 8.680161943319838e-07, "epoch": 0.8385214007782101 }, { "current_steps": 431, "loss": 2.0789, "learning_rate": 8.676113360323886e-07, "epoch": 0.8404669260700389 }, { "current_steps": 432, "loss": 2.1254, "learning_rate": 8.672064777327934e-07, "epoch": 0.8424124513618677 }, { "current_steps": 433, "loss": 2.3883, "learning_rate": 8.668016194331983e-07, "epoch": 0.8443579766536965 }, { "current_steps": 434, "loss": 2.3887, "learning_rate": 8.663967611336031e-07, "epoch": 0.8463035019455253 }, { "current_steps": 435, "loss": 2.0887, "learning_rate": 8.65991902834008e-07, "epoch": 0.8482490272373541 }, { "current_steps": 436, "loss": 1.9043, "learning_rate": 8.655870445344129e-07, "epoch": 0.8501945525291829 }, { "current_steps": 437, "loss": 1.8923, "learning_rate": 8.651821862348177e-07, "epoch": 0.8521400778210116 }, { "current_steps": 438, "loss": 2.505, "learning_rate": 8.647773279352226e-07, "epoch": 0.8540856031128404 }, { "current_steps": 439, "loss": 2.2354, "learning_rate": 8.643724696356275e-07, "epoch": 0.8560311284046692 }, { "current_steps": 440, "loss": 2.2925, "learning_rate": 8.639676113360323e-07, "epoch": 0.857976653696498 }, { "current_steps": 441, "loss": 2.1543, "learning_rate": 8.635627530364372e-07, "epoch": 0.8599221789883269 }, { "current_steps": 442, "loss": 2.3678, "learning_rate": 8.631578947368421e-07, "epoch": 0.8618677042801557 }, { "current_steps": 443, "loss": 2.2996, "learning_rate": 8.627530364372469e-07, "epoch": 0.8638132295719845 }, { "current_steps": 444, "loss": 2.1473, "learning_rate": 8.623481781376518e-07, "epoch": 0.8657587548638133 }, { "current_steps": 445, "loss": 2.4007, "learning_rate": 8.619433198380567e-07, "epoch": 0.867704280155642 }, { "current_steps": 446, "loss": 2.4281, "learning_rate": 8.615384615384616e-07, "epoch": 0.8696498054474708 }, { "current_steps": 447, "loss": 2.1206, "learning_rate": 8.611336032388664e-07, "epoch": 0.8715953307392996 }, { "current_steps": 448, "loss": 2.2754, "learning_rate": 8.607287449392712e-07, "epoch": 0.8735408560311284 }, { "current_steps": 449, "loss": 2.1754, "learning_rate": 8.603238866396761e-07, "epoch": 0.8754863813229572 }, { "current_steps": 450, "loss": 2.3147, "learning_rate": 8.599190283400809e-07, "epoch": 0.877431906614786 }, { "current_steps": 451, "loss": 1.9821, "learning_rate": 8.595141700404858e-07, "epoch": 0.8793774319066148 }, { "current_steps": 452, "loss": 2.459, "learning_rate": 8.591093117408907e-07, "epoch": 0.8813229571984436 }, { "current_steps": 453, "loss": 2.5062, "learning_rate": 8.587044534412955e-07, "epoch": 0.8832684824902723 }, { "current_steps": 454, "loss": 2.1994, "learning_rate": 8.582995951417004e-07, "epoch": 0.8852140077821011 }, { "current_steps": 455, "loss": 2.2841, "learning_rate": 8.578947368421053e-07, "epoch": 0.8871595330739299 }, { "current_steps": 456, "loss": 2.499, "learning_rate": 8.574898785425101e-07, "epoch": 0.8891050583657587 }, { "current_steps": 457, "loss": 2.3444, "learning_rate": 8.57085020242915e-07, "epoch": 0.8910505836575876 }, { "current_steps": 458, "loss": 2.5181, "learning_rate": 8.566801619433199e-07, "epoch": 0.8929961089494164 }, { "current_steps": 459, "loss": 2.6763, "learning_rate": 8.562753036437247e-07, "epoch": 0.8949416342412452 }, { "current_steps": 460, "loss": 1.7688, "learning_rate": 8.558704453441296e-07, "epoch": 0.896887159533074 }, { "current_steps": 461, "loss": 1.9225, "learning_rate": 8.554655870445345e-07, "epoch": 0.8988326848249028 }, { "current_steps": 462, "loss": 2.3067, "learning_rate": 8.550607287449392e-07, "epoch": 0.9007782101167315 }, { "current_steps": 463, "loss": 2.2563, "learning_rate": 8.54655870445344e-07, "epoch": 0.9027237354085603 }, { "current_steps": 464, "loss": 2.302, "learning_rate": 8.542510121457489e-07, "epoch": 0.9046692607003891 }, { "current_steps": 465, "loss": 2.1945, "learning_rate": 8.538461538461537e-07, "epoch": 0.9066147859922179 }, { "current_steps": 466, "loss": 2.4698, "learning_rate": 8.534412955465586e-07, "epoch": 0.9085603112840467 }, { "current_steps": 467, "loss": 2.1372, "learning_rate": 8.530364372469635e-07, "epoch": 0.9105058365758755 }, { "current_steps": 468, "loss": 2.1113, "learning_rate": 8.526315789473683e-07, "epoch": 0.9124513618677043 }, { "current_steps": 469, "loss": 2.2376, "learning_rate": 8.522267206477732e-07, "epoch": 0.914396887159533 }, { "current_steps": 470, "loss": 2.7027, "learning_rate": 8.518218623481781e-07, "epoch": 0.9163424124513618 }, { "current_steps": 471, "loss": 2.1185, "learning_rate": 8.514170040485829e-07, "epoch": 0.9182879377431906 }, { "current_steps": 472, "loss": 2.5189, "learning_rate": 8.510121457489878e-07, "epoch": 0.9202334630350194 }, { "current_steps": 473, "loss": 2.0806, "learning_rate": 8.506072874493927e-07, "epoch": 0.9221789883268483 }, { "current_steps": 474, "loss": 2.2486, "learning_rate": 8.502024291497975e-07, "epoch": 0.9241245136186771 }, { "current_steps": 475, "loss": 2.1336, "learning_rate": 8.497975708502024e-07, "epoch": 0.9260700389105059 }, { "current_steps": 476, "loss": 2.2803, "learning_rate": 8.493927125506073e-07, "epoch": 0.9280155642023347 }, { "current_steps": 477, "loss": 2.5206, "learning_rate": 8.489878542510122e-07, "epoch": 0.9299610894941635 }, { "current_steps": 478, "loss": 2.3767, "learning_rate": 8.485829959514169e-07, "epoch": 0.9319066147859922 }, { "current_steps": 479, "loss": 2.3634, "learning_rate": 8.481781376518218e-07, "epoch": 0.933852140077821 }, { "current_steps": 480, "loss": 2.2955, "learning_rate": 8.477732793522267e-07, "epoch": 0.9357976653696498 }, { "current_steps": 481, "loss": 2.4182, "learning_rate": 8.473684210526315e-07, "epoch": 0.9377431906614786 }, { "current_steps": 482, "loss": 2.2913, "learning_rate": 8.469635627530364e-07, "epoch": 0.9396887159533074 }, { "current_steps": 483, "loss": 2.2989, "learning_rate": 8.465587044534413e-07, "epoch": 0.9416342412451362 }, { "current_steps": 484, "loss": 2.3259, "learning_rate": 8.461538461538461e-07, "epoch": 0.943579766536965 }, { "current_steps": 485, "loss": 2.3182, "learning_rate": 8.45748987854251e-07, "epoch": 0.9455252918287937 }, { "current_steps": 486, "loss": 2.4366, "learning_rate": 8.453441295546559e-07, "epoch": 0.9474708171206225 }, { "current_steps": 487, "loss": 2.1621, "learning_rate": 8.449392712550607e-07, "epoch": 0.9494163424124513 }, { "current_steps": 488, "loss": 2.2769, "learning_rate": 8.445344129554656e-07, "epoch": 0.9513618677042801 }, { "current_steps": 489, "loss": 1.8244, "learning_rate": 8.441295546558705e-07, "epoch": 0.953307392996109 }, { "current_steps": 490, "loss": 2.3513, "learning_rate": 8.437246963562753e-07, "epoch": 0.9552529182879378 }, { "current_steps": 491, "loss": 2.1297, "learning_rate": 8.433198380566802e-07, "epoch": 0.9571984435797666 }, { "current_steps": 492, "loss": 2.187, "learning_rate": 8.429149797570851e-07, "epoch": 0.9591439688715954 }, { "current_steps": 493, "loss": 1.9724, "learning_rate": 8.425101214574898e-07, "epoch": 0.9610894941634242 }, { "current_steps": 494, "loss": 1.9239, "learning_rate": 8.421052631578947e-07, "epoch": 0.9630350194552529 }, { "current_steps": 495, "loss": 2.4268, "learning_rate": 8.417004048582996e-07, "epoch": 0.9649805447470817 }, { "current_steps": 496, "loss": 2.3946, "learning_rate": 8.412955465587043e-07, "epoch": 0.9669260700389105 }, { "current_steps": 497, "loss": 2.158, "learning_rate": 8.412955465587043e-07, "epoch": 0.9688715953307393 }, { "current_steps": 498, "loss": 2.2837, "learning_rate": 8.408906882591092e-07, "epoch": 0.9708171206225681 }, { "current_steps": 499, "loss": 2.4792, "learning_rate": 8.404858299595141e-07, "epoch": 0.9727626459143969 }, { "current_steps": 500, "loss": 2.0649, "learning_rate": 8.400809716599189e-07, "epoch": 0.9747081712062257 }, { "current_steps": 501, "loss": 2.0812, "learning_rate": 8.396761133603238e-07, "epoch": 0.9766536964980544 }, { "current_steps": 502, "loss": 1.9445, "learning_rate": 8.392712550607287e-07, "epoch": 0.9785992217898832 }, { "current_steps": 503, "loss": 2.2599, "learning_rate": 8.388663967611335e-07, "epoch": 0.980544747081712 }, { "current_steps": 504, "loss": 1.922, "learning_rate": 8.384615384615384e-07, "epoch": 0.9824902723735408 }, { "current_steps": 505, "loss": 2.5421, "learning_rate": 8.380566801619433e-07, "epoch": 0.9844357976653697 }, { "current_steps": 506, "loss": 2.0052, "learning_rate": 8.376518218623481e-07, "epoch": 0.9863813229571985 }, { "current_steps": 507, "loss": 2.3185, "learning_rate": 8.37246963562753e-07, "epoch": 0.9883268482490273 }, { "current_steps": 508, "loss": 2.5602, "learning_rate": 8.368421052631579e-07, "epoch": 0.9902723735408561 }, { "current_steps": 509, "loss": 2.295, "learning_rate": 8.364372469635628e-07, "epoch": 0.9922178988326849 }, { "current_steps": 510, "loss": 2.6195, "learning_rate": 8.360323886639675e-07, "epoch": 0.9941634241245136 }, { "current_steps": 511, "loss": 2.4747, "learning_rate": 8.356275303643724e-07, "epoch": 0.9961089494163424 }, { "current_steps": 512, "loss": 2.5255, "learning_rate": 8.352226720647772e-07, "epoch": 0.9980544747081712 }, { "current_steps": 513, "loss": 2.4686, "learning_rate": 8.348178137651821e-07, "epoch": 1.0 }, { "current_steps": 514, "loss": 2.2953, "learning_rate": 8.34412955465587e-07, "epoch": 1.0019455252918288 }, { "current_steps": 515, "loss": 2.2319, "learning_rate": 8.340080971659919e-07, "epoch": 1.0038910505836576 }, { "current_steps": 516, "loss": 2.2095, "learning_rate": 8.336032388663967e-07, "epoch": 1.0058365758754864 }, { "current_steps": 517, "loss": 2.4779, "learning_rate": 8.331983805668016e-07, "epoch": 1.0077821011673151 }, { "current_steps": 518, "loss": 2.2771, "learning_rate": 8.327935222672065e-07, "epoch": 1.009727626459144 }, { "current_steps": 519, "loss": 2.5114, "learning_rate": 8.323886639676113e-07, "epoch": 1.0116731517509727 }, { "current_steps": 520, "loss": 2.1811, "learning_rate": 8.319838056680162e-07, "epoch": 1.0136186770428015 }, { "current_steps": 521, "loss": 2.1813, "learning_rate": 8.315789473684211e-07, "epoch": 1.0155642023346303 }, { "current_steps": 522, "loss": 1.7464, "learning_rate": 8.311740890688259e-07, "epoch": 1.017509727626459 }, { "current_steps": 523, "loss": 2.4461, "learning_rate": 8.307692307692308e-07, "epoch": 1.0194552529182879 }, { "current_steps": 524, "loss": 2.1103, "learning_rate": 8.303643724696357e-07, "epoch": 1.0214007782101167 }, { "current_steps": 525, "loss": 2.0825, "learning_rate": 8.299595141700404e-07, "epoch": 1.0233463035019454 }, { "current_steps": 526, "loss": 2.2807, "learning_rate": 8.295546558704453e-07, "epoch": 1.0252918287937742 }, { "current_steps": 527, "loss": 2.272, "learning_rate": 8.291497975708502e-07, "epoch": 1.027237354085603 }, { "current_steps": 528, "loss": 2.3818, "learning_rate": 8.28744939271255e-07, "epoch": 1.0291828793774318 }, { "current_steps": 529, "loss": 2.4406, "learning_rate": 8.283400809716599e-07, "epoch": 1.0311284046692606 }, { "current_steps": 530, "loss": 2.3246, "learning_rate": 8.279352226720648e-07, "epoch": 1.0330739299610896 }, { "current_steps": 531, "loss": 1.9945, "learning_rate": 8.275303643724696e-07, "epoch": 1.0350194552529184 }, { "current_steps": 532, "loss": 2.4508, "learning_rate": 8.271255060728745e-07, "epoch": 1.0369649805447472 }, { "current_steps": 533, "loss": 2.1623, "learning_rate": 8.267206477732794e-07, "epoch": 1.038910505836576 }, { "current_steps": 534, "loss": 2.4021, "learning_rate": 8.263157894736841e-07, "epoch": 1.0408560311284047 }, { "current_steps": 535, "loss": 2.1649, "learning_rate": 8.25910931174089e-07, "epoch": 1.0428015564202335 }, { "current_steps": 536, "loss": 2.0952, "learning_rate": 8.255060728744939e-07, "epoch": 1.0447470817120623 }, { "current_steps": 537, "loss": 2.1999, "learning_rate": 8.251012145748987e-07, "epoch": 1.046692607003891 }, { "current_steps": 538, "loss": 2.3888, "learning_rate": 8.246963562753036e-07, "epoch": 1.04863813229572 }, { "current_steps": 539, "loss": 1.9581, "learning_rate": 8.242914979757085e-07, "epoch": 1.0505836575875487 }, { "current_steps": 540, "loss": 2.367, "learning_rate": 8.238866396761132e-07, "epoch": 1.0525291828793775 }, { "current_steps": 541, "loss": 2.4181, "learning_rate": 8.234817813765181e-07, "epoch": 1.0544747081712063 }, { "current_steps": 542, "loss": 2.4163, "learning_rate": 8.23076923076923e-07, "epoch": 1.056420233463035 }, { "current_steps": 543, "loss": 1.8646, "learning_rate": 8.226720647773278e-07, "epoch": 1.0583657587548638 }, { "current_steps": 544, "loss": 1.7765, "learning_rate": 8.222672064777327e-07, "epoch": 1.0603112840466926 }, { "current_steps": 545, "loss": 1.7839, "learning_rate": 8.218623481781376e-07, "epoch": 1.0622568093385214 }, { "current_steps": 546, "loss": 2.3206, "learning_rate": 8.214574898785425e-07, "epoch": 1.0642023346303502 }, { "current_steps": 547, "loss": 2.685, "learning_rate": 8.210526315789473e-07, "epoch": 1.066147859922179 }, { "current_steps": 548, "loss": 2.0982, "learning_rate": 8.206477732793522e-07, "epoch": 1.0680933852140078 }, { "current_steps": 549, "loss": 2.4379, "learning_rate": 8.202429149797571e-07, "epoch": 1.0700389105058365 }, { "current_steps": 550, "loss": 2.0505, "learning_rate": 8.198380566801619e-07, "epoch": 1.0719844357976653 }, { "current_steps": 551, "loss": 2.2621, "learning_rate": 8.194331983805668e-07, "epoch": 1.0739299610894941 }, { "current_steps": 552, "loss": 1.397, "learning_rate": 8.190283400809717e-07, "epoch": 1.075875486381323 }, { "current_steps": 553, "loss": 2.5035, "learning_rate": 8.186234817813765e-07, "epoch": 1.0778210116731517 }, { "current_steps": 554, "loss": 1.9758, "learning_rate": 8.182186234817814e-07, "epoch": 1.0797665369649805 }, { "current_steps": 555, "loss": 2.3523, "learning_rate": 8.178137651821862e-07, "epoch": 1.0817120622568093 }, { "current_steps": 556, "loss": 2.267, "learning_rate": 8.17408906882591e-07, "epoch": 1.083657587548638 }, { "current_steps": 557, "loss": 2.205, "learning_rate": 8.170040485829959e-07, "epoch": 1.0856031128404668 }, { "current_steps": 558, "loss": 2.1797, "learning_rate": 8.165991902834008e-07, "epoch": 1.0875486381322956 }, { "current_steps": 559, "loss": 2.436, "learning_rate": 8.161943319838056e-07, "epoch": 1.0894941634241244 }, { "current_steps": 560, "loss": 2.2896, "learning_rate": 8.157894736842105e-07, "epoch": 1.0914396887159532 }, { "current_steps": 561, "loss": 2.3633, "learning_rate": 8.153846153846154e-07, "epoch": 1.0933852140077822 }, { "current_steps": 562, "loss": 2.345, "learning_rate": 8.149797570850202e-07, "epoch": 1.095330739299611 }, { "current_steps": 563, "loss": 2.2637, "learning_rate": 8.145748987854251e-07, "epoch": 1.0972762645914398 }, { "current_steps": 564, "loss": 2.3981, "learning_rate": 8.1417004048583e-07, "epoch": 1.0992217898832686 }, { "current_steps": 565, "loss": 1.7515, "learning_rate": 8.137651821862348e-07, "epoch": 1.1011673151750974 }, { "current_steps": 566, "loss": 1.7928, "learning_rate": 8.133603238866397e-07, "epoch": 1.1031128404669261 }, { "current_steps": 567, "loss": 2.2605, "learning_rate": 8.129554655870446e-07, "epoch": 1.105058365758755 }, { "current_steps": 568, "loss": 2.2249, "learning_rate": 8.125506072874494e-07, "epoch": 1.1070038910505837 }, { "current_steps": 569, "loss": 2.3711, "learning_rate": 8.121457489878543e-07, "epoch": 1.1089494163424125 }, { "current_steps": 570, "loss": 2.0845, "learning_rate": 8.117408906882591e-07, "epoch": 1.1108949416342413 }, { "current_steps": 571, "loss": 2.1194, "learning_rate": 8.113360323886638e-07, "epoch": 1.11284046692607 }, { "current_steps": 572, "loss": 1.9531, "learning_rate": 8.109311740890687e-07, "epoch": 1.1147859922178989 }, { "current_steps": 573, "loss": 2.095, "learning_rate": 8.105263157894736e-07, "epoch": 1.1167315175097277 }, { "current_steps": 574, "loss": 2.043, "learning_rate": 8.101214574898784e-07, "epoch": 1.1186770428015564 }, { "current_steps": 575, "loss": 2.4287, "learning_rate": 8.097165991902833e-07, "epoch": 1.1206225680933852 }, { "current_steps": 576, "loss": 2.3367, "learning_rate": 8.093117408906882e-07, "epoch": 1.122568093385214 }, { "current_steps": 577, "loss": 2.2825, "learning_rate": 8.089068825910931e-07, "epoch": 1.1245136186770428 }, { "current_steps": 578, "loss": 2.1683, "learning_rate": 8.085020242914979e-07, "epoch": 1.1264591439688716 }, { "current_steps": 579, "loss": 2.522, "learning_rate": 8.080971659919028e-07, "epoch": 1.1284046692607004 }, { "current_steps": 580, "loss": 1.8963, "learning_rate": 8.076923076923077e-07, "epoch": 1.1303501945525292 }, { "current_steps": 581, "loss": 2.1656, "learning_rate": 8.072874493927125e-07, "epoch": 1.132295719844358 }, { "current_steps": 582, "loss": 2.5389, "learning_rate": 8.068825910931174e-07, "epoch": 1.1342412451361867 }, { "current_steps": 583, "loss": 2.4169, "learning_rate": 8.064777327935223e-07, "epoch": 1.1361867704280155 }, { "current_steps": 584, "loss": 2.1156, "learning_rate": 8.060728744939271e-07, "epoch": 1.1381322957198443 }, { "current_steps": 585, "loss": 2.5203, "learning_rate": 8.05668016194332e-07, "epoch": 1.140077821011673 }, { "current_steps": 586, "loss": 2.3258, "learning_rate": 8.052631578947368e-07, "epoch": 1.1420233463035019 }, { "current_steps": 587, "loss": 2.0503, "learning_rate": 8.048582995951416e-07, "epoch": 1.1439688715953307 }, { "current_steps": 588, "loss": 1.918, "learning_rate": 8.044534412955465e-07, "epoch": 1.1459143968871595 }, { "current_steps": 589, "loss": 2.1414, "learning_rate": 8.040485829959514e-07, "epoch": 1.1478599221789882 }, { "current_steps": 590, "loss": 2.1983, "learning_rate": 8.036437246963562e-07, "epoch": 1.149805447470817 }, { "current_steps": 591, "loss": 2.1886, "learning_rate": 8.032388663967611e-07, "epoch": 1.1517509727626458 }, { "current_steps": 592, "loss": 2.1374, "learning_rate": 8.02834008097166e-07, "epoch": 1.1536964980544746 }, { "current_steps": 593, "loss": 2.0714, "learning_rate": 8.024291497975708e-07, "epoch": 1.1556420233463034 }, { "current_steps": 594, "loss": 2.2326, "learning_rate": 8.020242914979757e-07, "epoch": 1.1575875486381322 }, { "current_steps": 595, "loss": 1.7879, "learning_rate": 8.016194331983806e-07, "epoch": 1.159533073929961 }, { "current_steps": 596, "loss": 2.2906, "learning_rate": 8.012145748987854e-07, "epoch": 1.16147859922179 }, { "current_steps": 597, "loss": 2.3476, "learning_rate": 8.008097165991903e-07, "epoch": 1.1634241245136188 }, { "current_steps": 598, "loss": 2.2962, "learning_rate": 8.004048582995952e-07, "epoch": 1.1653696498054475 }, { "current_steps": 599, "loss": 2.1733, "learning_rate": 8e-07, "epoch": 1.1673151750972763 }, { "current_steps": 600, "loss": 2.0962, "learning_rate": 7.995951417004049e-07, "epoch": 1.1692607003891051 }, { "current_steps": 601, "loss": 2.392, "learning_rate": 7.991902834008097e-07, "epoch": 1.171206225680934 }, { "current_steps": 602, "loss": 2.3579, "learning_rate": 7.987854251012145e-07, "epoch": 1.1731517509727627 }, { "current_steps": 603, "loss": 2.5342, "learning_rate": 7.983805668016193e-07, "epoch": 1.1750972762645915 }, { "current_steps": 604, "loss": 2.4073, "learning_rate": 7.979757085020242e-07, "epoch": 1.1770428015564203 }, { "current_steps": 605, "loss": 2.1764, "learning_rate": 7.97570850202429e-07, "epoch": 1.178988326848249 }, { "current_steps": 606, "loss": 2.6448, "learning_rate": 7.971659919028339e-07, "epoch": 1.1809338521400778 }, { "current_steps": 607, "loss": 2.3214, "learning_rate": 7.967611336032388e-07, "epoch": 1.1828793774319066 }, { "current_steps": 608, "loss": 1.7521, "learning_rate": 7.963562753036436e-07, "epoch": 1.1848249027237354 }, { "current_steps": 609, "loss": 1.8379, "learning_rate": 7.959514170040485e-07, "epoch": 1.1867704280155642 }, { "current_steps": 610, "loss": 2.1674, "learning_rate": 7.955465587044534e-07, "epoch": 1.188715953307393 }, { "current_steps": 611, "loss": 2.2569, "learning_rate": 7.951417004048583e-07, "epoch": 1.1906614785992218 }, { "current_steps": 612, "loss": 2.1972, "learning_rate": 7.947368421052631e-07, "epoch": 1.1926070038910506 }, { "current_steps": 613, "loss": 2.2186, "learning_rate": 7.94331983805668e-07, "epoch": 1.1945525291828794 }, { "current_steps": 614, "loss": 1.8013, "learning_rate": 7.939271255060729e-07, "epoch": 1.1964980544747081 }, { "current_steps": 615, "loss": 2.3978, "learning_rate": 7.935222672064777e-07, "epoch": 1.198443579766537 }, { "current_steps": 616, "loss": 2.1403, "learning_rate": 7.931174089068825e-07, "epoch": 1.2003891050583657 }, { "current_steps": 617, "loss": 2.1448, "learning_rate": 7.927125506072874e-07, "epoch": 1.2023346303501945 }, { "current_steps": 618, "loss": 2.3399, "learning_rate": 7.923076923076922e-07, "epoch": 1.2042801556420233 }, { "current_steps": 619, "loss": 2.1799, "learning_rate": 7.919028340080971e-07, "epoch": 1.206225680933852 }, { "current_steps": 620, "loss": 2.2685, "learning_rate": 7.91497975708502e-07, "epoch": 1.2081712062256809 }, { "current_steps": 621, "loss": 2.4101, "learning_rate": 7.910931174089068e-07, "epoch": 1.2101167315175096 }, { "current_steps": 622, "loss": 2.5062, "learning_rate": 7.906882591093117e-07, "epoch": 1.2120622568093384 }, { "current_steps": 623, "loss": 2.1767, "learning_rate": 7.902834008097166e-07, "epoch": 1.2140077821011672 }, { "current_steps": 624, "loss": 1.9123, "learning_rate": 7.898785425101214e-07, "epoch": 1.2159533073929962 }, { "current_steps": 625, "loss": 2.0536, "learning_rate": 7.894736842105263e-07, "epoch": 1.217898832684825 }, { "current_steps": 626, "loss": 2.2734, "learning_rate": 7.890688259109312e-07, "epoch": 1.2198443579766538 }, { "current_steps": 627, "loss": 2.3391, "learning_rate": 7.88663967611336e-07, "epoch": 1.2217898832684826 }, { "current_steps": 628, "loss": 2.0884, "learning_rate": 7.882591093117409e-07, "epoch": 1.2237354085603114 }, { "current_steps": 629, "loss": 1.9204, "learning_rate": 7.878542510121458e-07, "epoch": 1.2256809338521402 }, { "current_steps": 630, "loss": 2.3399, "learning_rate": 7.874493927125506e-07, "epoch": 1.227626459143969 }, { "current_steps": 631, "loss": 2.2867, "learning_rate": 7.870445344129555e-07, "epoch": 1.2295719844357977 }, { "current_steps": 632, "loss": 2.1299, "learning_rate": 7.866396761133603e-07, "epoch": 1.2315175097276265 }, { "current_steps": 633, "loss": 2.3307, "learning_rate": 7.862348178137651e-07, "epoch": 1.2334630350194553 }, { "current_steps": 634, "loss": 2.3734, "learning_rate": 7.8582995951417e-07, "epoch": 1.235408560311284 }, { "current_steps": 635, "loss": 2.0798, "learning_rate": 7.854251012145749e-07, "epoch": 1.2373540856031129 }, { "current_steps": 636, "loss": 2.1519, "learning_rate": 7.850202429149797e-07, "epoch": 1.2392996108949417 }, { "current_steps": 637, "loss": 2.225, "learning_rate": 7.846153846153846e-07, "epoch": 1.2412451361867705 }, { "current_steps": 638, "loss": 1.9023, "learning_rate": 7.842105263157895e-07, "epoch": 1.2431906614785992 }, { "current_steps": 639, "loss": 2.3249, "learning_rate": 7.838056680161942e-07, "epoch": 1.245136186770428 }, { "current_steps": 640, "loss": 2.389, "learning_rate": 7.834008097165991e-07, "epoch": 1.2470817120622568 }, { "current_steps": 641, "loss": 2.3516, "learning_rate": 7.82995951417004e-07, "epoch": 1.2490272373540856 }, { "current_steps": 642, "loss": 2.3485, "learning_rate": 7.825910931174089e-07, "epoch": 1.2509727626459144 }, { "current_steps": 643, "loss": 2.2315, "learning_rate": 7.821862348178137e-07, "epoch": 1.2529182879377432 }, { "current_steps": 644, "loss": 2.2829, "learning_rate": 7.817813765182186e-07, "epoch": 1.254863813229572 }, { "current_steps": 645, "loss": 2.1083, "learning_rate": 7.813765182186235e-07, "epoch": 1.2568093385214008 }, { "current_steps": 646, "loss": 1.941, "learning_rate": 7.809716599190283e-07, "epoch": 1.2587548638132295 }, { "current_steps": 647, "loss": 2.3287, "learning_rate": 7.805668016194331e-07, "epoch": 1.2607003891050583 }, { "current_steps": 648, "loss": 1.7516, "learning_rate": 7.80161943319838e-07, "epoch": 1.2626459143968871 }, { "current_steps": 649, "loss": 2.1235, "learning_rate": 7.797570850202428e-07, "epoch": 1.264591439688716 }, { "current_steps": 650, "loss": 2.0973, "learning_rate": 7.793522267206477e-07, "epoch": 1.2665369649805447 }, { "current_steps": 651, "loss": 2.3948, "learning_rate": 7.789473684210526e-07, "epoch": 1.2684824902723735 }, { "current_steps": 652, "loss": 2.2144, "learning_rate": 7.785425101214574e-07, "epoch": 1.2704280155642023 }, { "current_steps": 653, "loss": 2.3117, "learning_rate": 7.781376518218623e-07, "epoch": 1.272373540856031 }, { "current_steps": 654, "loss": 2.364, "learning_rate": 7.777327935222672e-07, "epoch": 1.2743190661478598 }, { "current_steps": 655, "loss": 2.2772, "learning_rate": 7.77327935222672e-07, "epoch": 1.2762645914396886 }, { "current_steps": 656, "loss": 2.1386, "learning_rate": 7.769230769230769e-07, "epoch": 1.2782101167315174 }, { "current_steps": 657, "loss": 2.3708, "learning_rate": 7.765182186234818e-07, "epoch": 1.2801556420233462 }, { "current_steps": 658, "loss": 2.3769, "learning_rate": 7.761133603238866e-07, "epoch": 1.282101167315175 }, { "current_steps": 659, "loss": 2.3622, "learning_rate": 7.757085020242915e-07, "epoch": 1.2840466926070038 }, { "current_steps": 660, "loss": 2.4079, "learning_rate": 7.753036437246964e-07, "epoch": 1.2859922178988326 }, { "current_steps": 661, "loss": 2.1683, "learning_rate": 7.748987854251012e-07, "epoch": 1.2879377431906613 }, { "current_steps": 662, "loss": 2.3006, "learning_rate": 7.74493927125506e-07, "epoch": 1.2898832684824904 }, { "current_steps": 663, "loss": 2.3171, "learning_rate": 7.740890688259109e-07, "epoch": 1.2918287937743191 }, { "current_steps": 664, "loss": 2.4681, "learning_rate": 7.736842105263157e-07, "epoch": 1.293774319066148 }, { "current_steps": 665, "loss": 2.2137, "learning_rate": 7.732793522267206e-07, "epoch": 1.2957198443579767 }, { "current_steps": 666, "loss": 1.906, "learning_rate": 7.728744939271255e-07, "epoch": 1.2976653696498055 }, { "current_steps": 667, "loss": 2.4044, "learning_rate": 7.724696356275303e-07, "epoch": 1.2996108949416343 }, { "current_steps": 668, "loss": 2.1868, "learning_rate": 7.720647773279352e-07, "epoch": 1.301556420233463 }, { "current_steps": 669, "loss": 2.3689, "learning_rate": 7.716599190283401e-07, "epoch": 1.3035019455252919 }, { "current_steps": 670, "loss": 1.7723, "learning_rate": 7.712550607287449e-07, "epoch": 1.3054474708171206 }, { "current_steps": 671, "loss": 2.1138, "learning_rate": 7.708502024291498e-07, "epoch": 1.3073929961089494 }, { "current_steps": 672, "loss": 1.9508, "learning_rate": 7.704453441295547e-07, "epoch": 1.3093385214007782 }, { "current_steps": 673, "loss": 1.8743, "learning_rate": 7.700404858299596e-07, "epoch": 1.311284046692607 }, { "current_steps": 674, "loss": 2.3129, "learning_rate": 7.696356275303644e-07, "epoch": 1.3132295719844358 }, { "current_steps": 675, "loss": 2.2051, "learning_rate": 7.692307692307693e-07, "epoch": 1.3151750972762646 }, { "current_steps": 676, "loss": 2.2398, "learning_rate": 7.688259109311742e-07, "epoch": 1.3171206225680934 }, { "current_steps": 677, "loss": 2.204, "learning_rate": 7.684210526315788e-07, "epoch": 1.3190661478599222 }, { "current_steps": 678, "loss": 2.1505, "learning_rate": 7.680161943319837e-07, "epoch": 1.321011673151751 }, { "current_steps": 679, "loss": 2.222, "learning_rate": 7.676113360323886e-07, "epoch": 1.3229571984435797 }, { "current_steps": 680, "loss": 2.3927, "learning_rate": 7.672064777327934e-07, "epoch": 1.3249027237354085 }, { "current_steps": 681, "loss": 2.5649, "learning_rate": 7.668016194331983e-07, "epoch": 1.3268482490272373 }, { "current_steps": 682, "loss": 1.9926, "learning_rate": 7.663967611336032e-07, "epoch": 1.328793774319066 }, { "current_steps": 683, "loss": 2.098, "learning_rate": 7.65991902834008e-07, "epoch": 1.3307392996108949 }, { "current_steps": 684, "loss": 2.3445, "learning_rate": 7.655870445344129e-07, "epoch": 1.3326848249027237 }, { "current_steps": 685, "loss": 2.3078, "learning_rate": 7.651821862348178e-07, "epoch": 1.3346303501945525 }, { "current_steps": 686, "loss": 1.7818, "learning_rate": 7.647773279352226e-07, "epoch": 1.3365758754863815 }, { "current_steps": 687, "loss": 2.1806, "learning_rate": 7.643724696356275e-07, "epoch": 1.3385214007782102 }, { "current_steps": 688, "loss": 2.1617, "learning_rate": 7.639676113360324e-07, "epoch": 1.340466926070039 }, { "current_steps": 689, "loss": 2.2642, "learning_rate": 7.635627530364372e-07, "epoch": 1.3424124513618678 }, { "current_steps": 690, "loss": 2.2141, "learning_rate": 7.631578947368421e-07, "epoch": 1.3443579766536966 }, { "current_steps": 691, "loss": 2.301, "learning_rate": 7.62753036437247e-07, "epoch": 1.3463035019455254 }, { "current_steps": 692, "loss": 2.4206, "learning_rate": 7.623481781376518e-07, "epoch": 1.3482490272373542 }, { "current_steps": 693, "loss": 2.1699, "learning_rate": 7.619433198380566e-07, "epoch": 1.350194552529183 }, { "current_steps": 694, "loss": 2.4613, "learning_rate": 7.615384615384615e-07, "epoch": 1.3521400778210118 }, { "current_steps": 695, "loss": 2.2389, "learning_rate": 7.611336032388663e-07, "epoch": 1.3540856031128405 }, { "current_steps": 696, "loss": 2.0768, "learning_rate": 7.607287449392712e-07, "epoch": 1.3560311284046693 }, { "current_steps": 697, "loss": 1.84, "learning_rate": 7.603238866396761e-07, "epoch": 1.3579766536964981 }, { "current_steps": 698, "loss": 2.2201, "learning_rate": 7.599190283400809e-07, "epoch": 1.359922178988327 }, { "current_steps": 699, "loss": 2.2122, "learning_rate": 7.595141700404858e-07, "epoch": 1.3618677042801557 }, { "current_steps": 700, "loss": 2.3832, "learning_rate": 7.591093117408907e-07, "epoch": 1.3638132295719845 }, { "current_steps": 701, "loss": 2.3184, "learning_rate": 7.587044534412955e-07, "epoch": 1.3657587548638133 }, { "current_steps": 702, "loss": 1.6276, "learning_rate": 7.582995951417004e-07, "epoch": 1.367704280155642 }, { "current_steps": 703, "loss": 2.2341, "learning_rate": 7.578947368421053e-07, "epoch": 1.3696498054474708 }, { "current_steps": 704, "loss": 2.335, "learning_rate": 7.574898785425101e-07, "epoch": 1.3715953307392996 }, { "current_steps": 705, "loss": 2.0687, "learning_rate": 7.57085020242915e-07, "epoch": 1.3735408560311284 }, { "current_steps": 706, "loss": 2.36, "learning_rate": 7.566801619433199e-07, "epoch": 1.3754863813229572 }, { "current_steps": 707, "loss": 2.135, "learning_rate": 7.562753036437248e-07, "epoch": 1.377431906614786 }, { "current_steps": 708, "loss": 2.319, "learning_rate": 7.558704453441295e-07, "epoch": 1.3793774319066148 }, { "current_steps": 709, "loss": 2.099, "learning_rate": 7.554655870445344e-07, "epoch": 1.3813229571984436 }, { "current_steps": 710, "loss": 2.0804, "learning_rate": 7.550607287449393e-07, "epoch": 1.3832684824902723 }, { "current_steps": 711, "loss": 1.9943, "learning_rate": 7.54655870445344e-07, "epoch": 1.3852140077821011 }, { "current_steps": 712, "loss": 2.0355, "learning_rate": 7.542510121457489e-07, "epoch": 1.38715953307393 }, { "current_steps": 713, "loss": 2.3279, "learning_rate": 7.538461538461538e-07, "epoch": 1.3891050583657587 }, { "current_steps": 714, "loss": 2.2399, "learning_rate": 7.534412955465586e-07, "epoch": 1.3910505836575875 }, { "current_steps": 715, "loss": 2.1394, "learning_rate": 7.530364372469635e-07, "epoch": 1.3929961089494163 }, { "current_steps": 716, "loss": 2.123, "learning_rate": 7.526315789473684e-07, "epoch": 1.394941634241245 }, { "current_steps": 717, "loss": 2.0222, "learning_rate": 7.522267206477732e-07, "epoch": 1.3968871595330739 }, { "current_steps": 718, "loss": 2.233, "learning_rate": 7.518218623481781e-07, "epoch": 1.3988326848249026 }, { "current_steps": 719, "loss": 2.1192, "learning_rate": 7.51417004048583e-07, "epoch": 1.4007782101167314 }, { "current_steps": 720, "loss": 1.9871, "learning_rate": 7.510121457489878e-07, "epoch": 1.4027237354085602 }, { "current_steps": 721, "loss": 2.1269, "learning_rate": 7.506072874493927e-07, "epoch": 1.404669260700389 }, { "current_steps": 722, "loss": 2.1992, "learning_rate": 7.502024291497976e-07, "epoch": 1.4066147859922178 }, { "current_steps": 723, "loss": 2.0437, "learning_rate": 7.497975708502023e-07, "epoch": 1.4085603112840466 }, { "current_steps": 724, "loss": 1.7738, "learning_rate": 7.493927125506072e-07, "epoch": 1.4105058365758754 }, { "current_steps": 725, "loss": 2.0905, "learning_rate": 7.489878542510121e-07, "epoch": 1.4124513618677041 }, { "current_steps": 726, "loss": 2.095, "learning_rate": 7.485829959514169e-07, "epoch": 1.4143968871595332 }, { "current_steps": 727, "loss": 2.0718, "learning_rate": 7.481781376518218e-07, "epoch": 1.416342412451362 }, { "current_steps": 728, "loss": 2.2802, "learning_rate": 7.477732793522267e-07, "epoch": 1.4182879377431907 }, { "current_steps": 729, "loss": 2.1514, "learning_rate": 7.473684210526315e-07, "epoch": 1.4202334630350195 }, { "current_steps": 730, "loss": 2.2864, "learning_rate": 7.469635627530364e-07, "epoch": 1.4221789883268483 }, { "current_steps": 731, "loss": 1.9398, "learning_rate": 7.465587044534413e-07, "epoch": 1.424124513618677 }, { "current_steps": 732, "loss": 2.194, "learning_rate": 7.461538461538461e-07, "epoch": 1.4260700389105059 }, { "current_steps": 733, "loss": 2.2878, "learning_rate": 7.45748987854251e-07, "epoch": 1.4280155642023347 }, { "current_steps": 734, "loss": 1.9485, "learning_rate": 7.453441295546559e-07, "epoch": 1.4299610894941635 }, { "current_steps": 735, "loss": 2.1926, "learning_rate": 7.449392712550607e-07, "epoch": 1.4319066147859922 }, { "current_steps": 736, "loss": 2.4356, "learning_rate": 7.445344129554656e-07, "epoch": 1.433852140077821 }, { "current_steps": 737, "loss": 2.3228, "learning_rate": 7.441295546558705e-07, "epoch": 1.4357976653696498 }, { "current_steps": 738, "loss": 1.8354, "learning_rate": 7.437246963562752e-07, "epoch": 1.4377431906614786 }, { "current_steps": 739, "loss": 2.0439, "learning_rate": 7.433198380566801e-07, "epoch": 1.4396887159533074 }, { "current_steps": 740, "loss": 2.4113, "learning_rate": 7.42914979757085e-07, "epoch": 1.4416342412451362 }, { "current_steps": 741, "loss": 2.0752, "learning_rate": 7.425101214574899e-07, "epoch": 1.443579766536965 }, { "current_steps": 742, "loss": 1.9469, "learning_rate": 7.421052631578947e-07, "epoch": 1.4455252918287937 }, { "current_steps": 743, "loss": 2.091, "learning_rate": 7.417004048582996e-07, "epoch": 1.4474708171206225 }, { "current_steps": 744, "loss": 2.5164, "learning_rate": 7.412955465587045e-07, "epoch": 1.4494163424124513 }, { "current_steps": 745, "loss": 2.2156, "learning_rate": 7.408906882591093e-07, "epoch": 1.45136186770428 }, { "current_steps": 746, "loss": 2.2121, "learning_rate": 7.404858299595142e-07, "epoch": 1.453307392996109 }, { "current_steps": 747, "loss": 2.1681, "learning_rate": 7.40080971659919e-07, "epoch": 1.4552529182879377 }, { "current_steps": 748, "loss": 1.8178, "learning_rate": 7.396761133603238e-07, "epoch": 1.4571984435797665 }, { "current_steps": 749, "loss": 2.2571, "learning_rate": 7.392712550607287e-07, "epoch": 1.4591439688715953 }, { "current_steps": 750, "loss": 2.2856, "learning_rate": 7.388663967611336e-07, "epoch": 1.4610894941634243 }, { "current_steps": 751, "loss": 2.3552, "learning_rate": 7.384615384615384e-07, "epoch": 1.463035019455253 }, { "current_steps": 752, "loss": 1.7045, "learning_rate": 7.380566801619433e-07, "epoch": 1.4649805447470818 }, { "current_steps": 753, "loss": 2.1487, "learning_rate": 7.376518218623481e-07, "epoch": 1.4669260700389106 }, { "current_steps": 754, "loss": 1.9507, "learning_rate": 7.372469635627529e-07, "epoch": 1.4688715953307394 }, { "current_steps": 755, "loss": 2.6883, "learning_rate": 7.368421052631578e-07, "epoch": 1.4708171206225682 }, { "current_steps": 756, "loss": 2.0654, "learning_rate": 7.364372469635627e-07, "epoch": 1.472762645914397 }, { "current_steps": 757, "loss": 2.1876, "learning_rate": 7.360323886639675e-07, "epoch": 1.4747081712062258 }, { "current_steps": 758, "loss": 2.2609, "learning_rate": 7.356275303643724e-07, "epoch": 1.4766536964980546 }, { "current_steps": 759, "loss": 1.875, "learning_rate": 7.352226720647773e-07, "epoch": 1.4785992217898833 }, { "current_steps": 760, "loss": 1.9554, "learning_rate": 7.348178137651821e-07, "epoch": 1.4805447470817121 }, { "current_steps": 761, "loss": 2.2388, "learning_rate": 7.34412955465587e-07, "epoch": 1.482490272373541 }, { "current_steps": 762, "loss": 2.0944, "learning_rate": 7.340080971659919e-07, "epoch": 1.4844357976653697 }, { "current_steps": 763, "loss": 2.246, "learning_rate": 7.336032388663967e-07, "epoch": 1.4863813229571985 }, { "current_steps": 764, "loss": 2.3021, "learning_rate": 7.331983805668016e-07, "epoch": 1.4883268482490273 }, { "current_steps": 765, "loss": 2.3816, "learning_rate": 7.327935222672065e-07, "epoch": 1.490272373540856 }, { "current_steps": 766, "loss": 2.0286, "learning_rate": 7.323886639676113e-07, "epoch": 1.4922178988326849 }, { "current_steps": 767, "loss": 2.3329, "learning_rate": 7.319838056680162e-07, "epoch": 1.4941634241245136 }, { "current_steps": 768, "loss": 2.3775, "learning_rate": 7.315789473684211e-07, "epoch": 1.4961089494163424 }, { "current_steps": 769, "loss": 2.2878, "learning_rate": 7.311740890688258e-07, "epoch": 1.4980544747081712 }, { "current_steps": 770, "loss": 1.8248, "learning_rate": 7.307692307692307e-07, "epoch": 1.5 }, { "current_steps": 771, "loss": 2.0054, "learning_rate": 7.303643724696356e-07, "epoch": 1.5019455252918288 }, { "current_steps": 772, "loss": 2.2772, "learning_rate": 7.299595141700404e-07, "epoch": 1.5038910505836576 }, { "current_steps": 773, "loss": 2.3327, "learning_rate": 7.295546558704453e-07, "epoch": 1.5058365758754864 }, { "current_steps": 774, "loss": 2.2797, "learning_rate": 7.291497975708502e-07, "epoch": 1.5077821011673151 }, { "current_steps": 775, "loss": 1.8291, "learning_rate": 7.287449392712551e-07, "epoch": 1.509727626459144 }, { "current_steps": 776, "loss": 2.2051, "learning_rate": 7.283400809716599e-07, "epoch": 1.5116731517509727 }, { "current_steps": 777, "loss": 2.2001, "learning_rate": 7.279352226720648e-07, "epoch": 1.5136186770428015 }, { "current_steps": 778, "loss": 2.0597, "learning_rate": 7.275303643724697e-07, "epoch": 1.5155642023346303 }, { "current_steps": 779, "loss": 1.7549, "learning_rate": 7.271255060728745e-07, "epoch": 1.517509727626459 }, { "current_steps": 780, "loss": 1.9528, "learning_rate": 7.267206477732794e-07, "epoch": 1.5194552529182879 }, { "current_steps": 781, "loss": 2.1243, "learning_rate": 7.263157894736843e-07, "epoch": 1.5214007782101167 }, { "current_steps": 782, "loss": 2.0314, "learning_rate": 7.25910931174089e-07, "epoch": 1.5233463035019454 }, { "current_steps": 783, "loss": 2.2411, "learning_rate": 7.25506072874494e-07, "epoch": 1.5252918287937742 }, { "current_steps": 784, "loss": 2.3171, "learning_rate": 7.251012145748987e-07, "epoch": 1.527237354085603 }, { "current_steps": 785, "loss": 2.0432, "learning_rate": 7.246963562753035e-07, "epoch": 1.5291828793774318 }, { "current_steps": 786, "loss": 2.4074, "learning_rate": 7.242914979757084e-07, "epoch": 1.5311284046692606 }, { "current_steps": 787, "loss": 1.6248, "learning_rate": 7.238866396761133e-07, "epoch": 1.5330739299610894 }, { "current_steps": 788, "loss": 2.113, "learning_rate": 7.234817813765181e-07, "epoch": 1.5350194552529182 }, { "current_steps": 789, "loss": 1.9868, "learning_rate": 7.23076923076923e-07, "epoch": 1.536964980544747 }, { "current_steps": 790, "loss": 2.0545, "learning_rate": 7.226720647773279e-07, "epoch": 1.5389105058365757 }, { "current_steps": 791, "loss": 2.3115, "learning_rate": 7.222672064777327e-07, "epoch": 1.5408560311284045 }, { "current_steps": 792, "loss": 2.2364, "learning_rate": 7.218623481781376e-07, "epoch": 1.5428015564202333 }, { "current_steps": 793, "loss": 2.2656, "learning_rate": 7.214574898785425e-07, "epoch": 1.544747081712062 }, { "current_steps": 794, "loss": 2.049, "learning_rate": 7.210526315789473e-07, "epoch": 1.5466926070038909 }, { "current_steps": 795, "loss": 2.1477, "learning_rate": 7.206477732793522e-07, "epoch": 1.54863813229572 }, { "current_steps": 796, "loss": 2.3144, "learning_rate": 7.202429149797571e-07, "epoch": 1.5505836575875487 }, { "current_steps": 797, "loss": 2.2816, "learning_rate": 7.198380566801619e-07, "epoch": 1.5525291828793775 }, { "current_steps": 798, "loss": 1.8448, "learning_rate": 7.194331983805668e-07, "epoch": 1.5544747081712063 }, { "current_steps": 799, "loss": 2.354, "learning_rate": 7.190283400809716e-07, "epoch": 1.556420233463035 }, { "current_steps": 800, "loss": 2.0528, "learning_rate": 7.186234817813764e-07, "epoch": 1.5583657587548638 }, { "current_steps": 801, "loss": 2.349, "learning_rate": 7.182186234817813e-07, "epoch": 1.5603112840466926 }, { "current_steps": 802, "loss": 2.3199, "learning_rate": 7.178137651821862e-07, "epoch": 1.5622568093385214 }, { "current_steps": 803, "loss": 2.3816, "learning_rate": 7.17408906882591e-07, "epoch": 1.5642023346303502 }, { "current_steps": 804, "loss": 2.4202, "learning_rate": 7.170040485829959e-07, "epoch": 1.566147859922179 }, { "current_steps": 805, "loss": 2.2357, "learning_rate": 7.165991902834008e-07, "epoch": 1.5680933852140078 }, { "current_steps": 806, "loss": 2.6106, "learning_rate": 7.161943319838057e-07, "epoch": 1.5700389105058365 }, { "current_steps": 807, "loss": 2.0489, "learning_rate": 7.157894736842105e-07, "epoch": 1.5719844357976653 }, { "current_steps": 808, "loss": 2.2561, "learning_rate": 7.153846153846154e-07, "epoch": 1.5739299610894941 }, { "current_steps": 809, "loss": 2.3593, "learning_rate": 7.149797570850203e-07, "epoch": 1.575875486381323 }, { "current_steps": 810, "loss": 2.0486, "learning_rate": 7.145748987854251e-07, "epoch": 1.5778210116731517 }, { "current_steps": 811, "loss": 2.0996, "learning_rate": 7.1417004048583e-07, "epoch": 1.5797665369649807 }, { "current_steps": 812, "loss": 1.5311, "learning_rate": 7.137651821862349e-07, "epoch": 1.5817120622568095 }, { "current_steps": 813, "loss": 2.1538, "learning_rate": 7.133603238866397e-07, "epoch": 1.5836575875486383 }, { "current_steps": 814, "loss": 1.9634, "learning_rate": 7.129554655870445e-07, "epoch": 1.585603112840467 }, { "current_steps": 815, "loss": 1.9854, "learning_rate": 7.125506072874494e-07, "epoch": 1.5875486381322959 }, { "current_steps": 816, "loss": 2.2877, "learning_rate": 7.121457489878541e-07, "epoch": 1.5894941634241246 }, { "current_steps": 817, "loss": 2.1919, "learning_rate": 7.11740890688259e-07, "epoch": 1.5914396887159534 }, { "current_steps": 818, "loss": 2.1689, "learning_rate": 7.11336032388664e-07, "epoch": 1.5933852140077822 }, { "current_steps": 819, "loss": 2.3176, "learning_rate": 7.109311740890687e-07, "epoch": 1.595330739299611 }, { "current_steps": 820, "loss": 2.3419, "learning_rate": 7.105263157894736e-07, "epoch": 1.5972762645914398 }, { "current_steps": 821, "loss": 2.349, "learning_rate": 7.101214574898785e-07, "epoch": 1.5992217898832686 }, { "current_steps": 822, "loss": 2.2798, "learning_rate": 7.097165991902833e-07, "epoch": 1.6011673151750974 }, { "current_steps": 823, "loss": 2.34, "learning_rate": 7.093117408906882e-07, "epoch": 1.6031128404669261 }, { "current_steps": 824, "loss": 2.2742, "learning_rate": 7.089068825910931e-07, "epoch": 1.605058365758755 }, { "current_steps": 825, "loss": 2.2644, "learning_rate": 7.085020242914979e-07, "epoch": 1.6070038910505837 }, { "current_steps": 826, "loss": 2.2865, "learning_rate": 7.080971659919028e-07, "epoch": 1.6089494163424125 }, { "current_steps": 827, "loss": 2.3643, "learning_rate": 7.076923076923077e-07, "epoch": 1.6108949416342413 }, { "current_steps": 828, "loss": 2.1298, "learning_rate": 7.072874493927125e-07, "epoch": 1.61284046692607 }, { "current_steps": 829, "loss": 2.2215, "learning_rate": 7.068825910931174e-07, "epoch": 1.6147859922178989 }, { "current_steps": 830, "loss": 2.0733, "learning_rate": 7.064777327935222e-07, "epoch": 1.6167315175097277 }, { "current_steps": 831, "loss": 2.2091, "learning_rate": 7.06072874493927e-07, "epoch": 1.6186770428015564 }, { "current_steps": 832, "loss": 2.0576, "learning_rate": 7.056680161943319e-07, "epoch": 1.6206225680933852 }, { "current_steps": 833, "loss": 2.1306, "learning_rate": 7.052631578947368e-07, "epoch": 1.622568093385214 }, { "current_steps": 834, "loss": 2.1506, "learning_rate": 7.048582995951416e-07, "epoch": 1.6245136186770428 }, { "current_steps": 835, "loss": 2.054, "learning_rate": 7.044534412955465e-07, "epoch": 1.6264591439688716 }, { "current_steps": 836, "loss": 2.1397, "learning_rate": 7.040485829959514e-07, "epoch": 1.6284046692607004 }, { "current_steps": 837, "loss": 2.3836, "learning_rate": 7.036437246963563e-07, "epoch": 1.6303501945525292 }, { "current_steps": 838, "loss": 2.4045, "learning_rate": 7.032388663967611e-07, "epoch": 1.632295719844358 }, { "current_steps": 839, "loss": 2.1783, "learning_rate": 7.02834008097166e-07, "epoch": 1.6342412451361867 }, { "current_steps": 840, "loss": 2.4588, "learning_rate": 7.024291497975709e-07, "epoch": 1.6361867704280155 }, { "current_steps": 841, "loss": 2.0684, "learning_rate": 7.020242914979757e-07, "epoch": 1.6381322957198443 }, { "current_steps": 842, "loss": 2.0153, "learning_rate": 7.016194331983806e-07, "epoch": 1.640077821011673 }, { "current_steps": 843, "loss": 2.3721, "learning_rate": 7.012145748987855e-07, "epoch": 1.6420233463035019 }, { "current_steps": 844, "loss": 1.6322, "learning_rate": 7.008097165991903e-07, "epoch": 1.6439688715953307 }, { "current_steps": 845, "loss": 2.0128, "learning_rate": 7.004048582995951e-07, "epoch": 1.6459143968871595 }, { "current_steps": 846, "loss": 2.3425, "learning_rate": 7e-07, "epoch": 1.6478599221789882 }, { "current_steps": 847, "loss": 2.1655, "learning_rate": 6.995951417004048e-07, "epoch": 1.649805447470817 }, { "current_steps": 848, "loss": 2.4362, "learning_rate": 6.991902834008097e-07, "epoch": 1.6517509727626458 }, { "current_steps": 849, "loss": 2.1289, "learning_rate": 6.987854251012146e-07, "epoch": 1.6536964980544746 }, { "current_steps": 850, "loss": 1.8241, "learning_rate": 6.983805668016194e-07, "epoch": 1.6556420233463034 }, { "current_steps": 851, "loss": 1.8507, "learning_rate": 6.979757085020243e-07, "epoch": 1.6575875486381322 }, { "current_steps": 852, "loss": 2.3184, "learning_rate": 6.975708502024292e-07, "epoch": 1.659533073929961 }, { "current_steps": 853, "loss": 1.7699, "learning_rate": 6.971659919028339e-07, "epoch": 1.6614785992217898 }, { "current_steps": 854, "loss": 2.1509, "learning_rate": 6.967611336032388e-07, "epoch": 1.6634241245136185 }, { "current_steps": 855, "loss": 2.1414, "learning_rate": 6.963562753036437e-07, "epoch": 1.6653696498054473 }, { "current_steps": 856, "loss": 1.8552, "learning_rate": 6.959514170040485e-07, "epoch": 1.6673151750972761 }, { "current_steps": 857, "loss": 2.4825, "learning_rate": 6.955465587044534e-07, "epoch": 1.669260700389105 }, { "current_steps": 858, "loss": 1.9613, "learning_rate": 6.951417004048583e-07, "epoch": 1.6712062256809337 }, { "current_steps": 859, "loss": 2.3556, "learning_rate": 6.947368421052631e-07, "epoch": 1.6731517509727627 }, { "current_steps": 860, "loss": 2.6808, "learning_rate": 6.943319838056679e-07, "epoch": 1.6750972762645915 }, { "current_steps": 861, "loss": 2.3349, "learning_rate": 6.939271255060728e-07, "epoch": 1.6770428015564203 }, { "current_steps": 862, "loss": 2.1397, "learning_rate": 6.935222672064776e-07, "epoch": 1.678988326848249 }, { "current_steps": 863, "loss": 2.0612, "learning_rate": 6.931174089068825e-07, "epoch": 1.6809338521400778 }, { "current_steps": 864, "loss": 2.116, "learning_rate": 6.927125506072874e-07, "epoch": 1.6828793774319066 }, { "current_steps": 865, "loss": 2.2047, "learning_rate": 6.923076923076922e-07, "epoch": 1.6848249027237354 }, { "current_steps": 866, "loss": 2.232, "learning_rate": 6.919028340080971e-07, "epoch": 1.6867704280155642 }, { "current_steps": 867, "loss": 2.0991, "learning_rate": 6.91497975708502e-07, "epoch": 1.688715953307393 }, { "current_steps": 868, "loss": 2.0277, "learning_rate": 6.910931174089068e-07, "epoch": 1.6906614785992218 }, { "current_steps": 869, "loss": 1.6164, "learning_rate": 6.906882591093117e-07, "epoch": 1.6926070038910506 }, { "current_steps": 870, "loss": 2.1732, "learning_rate": 6.902834008097166e-07, "epoch": 1.6945525291828794 }, { "current_steps": 871, "loss": 2.1521, "learning_rate": 6.898785425101215e-07, "epoch": 1.6964980544747081 }, { "current_steps": 872, "loss": 2.2203, "learning_rate": 6.894736842105263e-07, "epoch": 1.698443579766537 }, { "current_steps": 873, "loss": 2.3508, "learning_rate": 6.890688259109312e-07, "epoch": 1.7003891050583657 }, { "current_steps": 874, "loss": 2.0042, "learning_rate": 6.886639676113361e-07, "epoch": 1.7023346303501945 }, { "current_steps": 875, "loss": 2.2213, "learning_rate": 6.882591093117408e-07, "epoch": 1.7042801556420235 }, { "current_steps": 876, "loss": 1.9559, "learning_rate": 6.878542510121457e-07, "epoch": 1.7062256809338523 }, { "current_steps": 877, "loss": 1.6859, "learning_rate": 6.874493927125506e-07, "epoch": 1.708171206225681 }, { "current_steps": 878, "loss": 2.2684, "learning_rate": 6.870445344129554e-07, "epoch": 1.7101167315175099 }, { "current_steps": 879, "loss": 2.1417, "learning_rate": 6.866396761133603e-07, "epoch": 1.7120622568093387 }, { "current_steps": 880, "loss": 2.2387, "learning_rate": 6.862348178137652e-07, "epoch": 1.7140077821011674 }, { "current_steps": 881, "loss": 1.9528, "learning_rate": 6.8582995951417e-07, "epoch": 1.7159533073929962 }, { "current_steps": 882, "loss": 2.2697, "learning_rate": 6.854251012145749e-07, "epoch": 1.717898832684825 }, { "current_steps": 883, "loss": 2.2855, "learning_rate": 6.850202429149798e-07, "epoch": 1.7198443579766538 }, { "current_steps": 884, "loss": 2.2746, "learning_rate": 6.846153846153846e-07, "epoch": 1.7217898832684826 }, { "current_steps": 885, "loss": 2.0496, "learning_rate": 6.842105263157895e-07, "epoch": 1.7237354085603114 }, { "current_steps": 886, "loss": 1.9592, "learning_rate": 6.838056680161944e-07, "epoch": 1.7256809338521402 }, { "current_steps": 887, "loss": 2.2464, "learning_rate": 6.834008097165992e-07, "epoch": 1.727626459143969 }, { "current_steps": 888, "loss": 2.2467, "learning_rate": 6.82995951417004e-07, "epoch": 1.7295719844357977 }, { "current_steps": 889, "loss": 1.9037, "learning_rate": 6.82591093117409e-07, "epoch": 1.7315175097276265 }, { "current_steps": 890, "loss": 1.4651, "learning_rate": 6.821862348178137e-07, "epoch": 1.7334630350194553 }, { "current_steps": 891, "loss": 2.1082, "learning_rate": 6.817813765182185e-07, "epoch": 1.735408560311284 }, { "current_steps": 892, "loss": 2.3547, "learning_rate": 6.813765182186234e-07, "epoch": 1.7373540856031129 }, { "current_steps": 893, "loss": 2.189, "learning_rate": 6.809716599190282e-07, "epoch": 1.7392996108949417 }, { "current_steps": 894, "loss": 2.1957, "learning_rate": 6.805668016194331e-07, "epoch": 1.7412451361867705 }, { "current_steps": 895, "loss": 1.7558, "learning_rate": 6.80161943319838e-07, "epoch": 1.7431906614785992 }, { "current_steps": 896, "loss": 1.9964, "learning_rate": 6.797570850202428e-07, "epoch": 1.745136186770428 }, { "current_steps": 897, "loss": 1.9201, "learning_rate": 6.793522267206477e-07, "epoch": 1.7470817120622568 }, { "current_steps": 898, "loss": 2.5081, "learning_rate": 6.789473684210526e-07, "epoch": 1.7490272373540856 }, { "current_steps": 899, "loss": 2.2549, "learning_rate": 6.785425101214574e-07, "epoch": 1.7509727626459144 }, { "current_steps": 900, "loss": 2.2163, "learning_rate": 6.781376518218623e-07, "epoch": 1.7529182879377432 }, { "current_steps": 901, "loss": 2.2599, "learning_rate": 6.777327935222672e-07, "epoch": 1.754863813229572 }, { "current_steps": 902, "loss": 1.9352, "learning_rate": 6.773279352226721e-07, "epoch": 1.7568093385214008 }, { "current_steps": 903, "loss": 2.4098, "learning_rate": 6.769230769230769e-07, "epoch": 1.7587548638132295 }, { "current_steps": 904, "loss": 2.2639, "learning_rate": 6.765182186234818e-07, "epoch": 1.7607003891050583 }, { "current_steps": 905, "loss": 1.7665, "learning_rate": 6.761133603238867e-07, "epoch": 1.7626459143968871 }, { "current_steps": 906, "loss": 2.0507, "learning_rate": 6.757085020242914e-07, "epoch": 1.764591439688716 }, { "current_steps": 907, "loss": 2.0684, "learning_rate": 6.753036437246963e-07, "epoch": 1.7665369649805447 }, { "current_steps": 908, "loss": 2.6082, "learning_rate": 6.748987854251012e-07, "epoch": 1.7684824902723735 }, { "current_steps": 909, "loss": 1.6983, "learning_rate": 6.74493927125506e-07, "epoch": 1.7704280155642023 }, { "current_steps": 910, "loss": 2.2762, "learning_rate": 6.740890688259109e-07, "epoch": 1.772373540856031 }, { "current_steps": 911, "loss": 2.1885, "learning_rate": 6.736842105263158e-07, "epoch": 1.7743190661478598 }, { "current_steps": 912, "loss": 2.2782, "learning_rate": 6.732793522267206e-07, "epoch": 1.7762645914396886 }, { "current_steps": 913, "loss": 2.0888, "learning_rate": 6.728744939271255e-07, "epoch": 1.7782101167315174 }, { "current_steps": 914, "loss": 2.1194, "learning_rate": 6.724696356275304e-07, "epoch": 1.7801556420233462 }, { "current_steps": 915, "loss": 2.1619, "learning_rate": 6.720647773279352e-07, "epoch": 1.782101167315175 }, { "current_steps": 916, "loss": 2.3935, "learning_rate": 6.716599190283401e-07, "epoch": 1.7840466926070038 }, { "current_steps": 917, "loss": 2.1575, "learning_rate": 6.71255060728745e-07, "epoch": 1.7859922178988326 }, { "current_steps": 918, "loss": 2.1956, "learning_rate": 6.708502024291498e-07, "epoch": 1.7879377431906613 }, { "current_steps": 919, "loss": 2.246, "learning_rate": 6.704453441295547e-07, "epoch": 1.7898832684824901 }, { "current_steps": 920, "loss": 2.2959, "learning_rate": 6.700404858299596e-07, "epoch": 1.791828793774319 }, { "current_steps": 921, "loss": 1.8563, "learning_rate": 6.696356275303643e-07, "epoch": 1.7937743190661477 }, { "current_steps": 922, "loss": 2.4743, "learning_rate": 6.692307692307692e-07, "epoch": 1.7957198443579765 }, { "current_steps": 923, "loss": 2.2841, "learning_rate": 6.68825910931174e-07, "epoch": 1.7976653696498055 }, { "current_steps": 924, "loss": 2.2562, "learning_rate": 6.684210526315788e-07, "epoch": 1.7996108949416343 }, { "current_steps": 925, "loss": 1.9672, "learning_rate": 6.680161943319837e-07, "epoch": 1.801556420233463 }, { "current_steps": 926, "loss": 1.4681, "learning_rate": 6.676113360323886e-07, "epoch": 1.8035019455252919 }, { "current_steps": 927, "loss": 2.1061, "learning_rate": 6.672064777327934e-07, "epoch": 1.8054474708171206 }, { "current_steps": 928, "loss": 2.2873, "learning_rate": 6.668016194331983e-07, "epoch": 1.8073929961089494 }, { "current_steps": 929, "loss": 1.9433, "learning_rate": 6.663967611336032e-07, "epoch": 1.8093385214007782 }, { "current_steps": 930, "loss": 2.1612, "learning_rate": 6.65991902834008e-07, "epoch": 1.811284046692607 }, { "current_steps": 931, "loss": 2.4879, "learning_rate": 6.655870445344129e-07, "epoch": 1.8132295719844358 }, { "current_steps": 932, "loss": 2.5283, "learning_rate": 6.651821862348178e-07, "epoch": 1.8151750972762646 }, { "current_steps": 933, "loss": 2.2038, "learning_rate": 6.647773279352227e-07, "epoch": 1.8171206225680934 }, { "current_steps": 934, "loss": 2.1743, "learning_rate": 6.643724696356275e-07, "epoch": 1.8190661478599222 }, { "current_steps": 935, "loss": 2.2922, "learning_rate": 6.639676113360324e-07, "epoch": 1.821011673151751 }, { "current_steps": 936, "loss": 1.8023, "learning_rate": 6.635627530364371e-07, "epoch": 1.8229571984435797 }, { "current_steps": 937, "loss": 2.0862, "learning_rate": 6.63157894736842e-07, "epoch": 1.8249027237354085 }, { "current_steps": 938, "loss": 2.448, "learning_rate": 6.627530364372469e-07, "epoch": 1.8268482490272373 }, { "current_steps": 939, "loss": 2.2369, "learning_rate": 6.623481781376518e-07, "epoch": 1.8287937743190663 }, { "current_steps": 940, "loss": 1.8588, "learning_rate": 6.619433198380566e-07, "epoch": 1.830739299610895 }, { "current_steps": 941, "loss": 2.3545, "learning_rate": 6.615384615384615e-07, "epoch": 1.8326848249027239 }, { "current_steps": 942, "loss": 2.1595, "learning_rate": 6.611336032388664e-07, "epoch": 1.8346303501945527 }, { "current_steps": 943, "loss": 1.918, "learning_rate": 6.607287449392712e-07, "epoch": 1.8365758754863815 }, { "current_steps": 944, "loss": 2.3025, "learning_rate": 6.603238866396761e-07, "epoch": 1.8385214007782102 }, { "current_steps": 945, "loss": 2.166, "learning_rate": 6.59919028340081e-07, "epoch": 1.840466926070039 }, { "current_steps": 946, "loss": 2.3401, "learning_rate": 6.595141700404858e-07, "epoch": 1.8424124513618678 }, { "current_steps": 947, "loss": 2.3197, "learning_rate": 6.591093117408907e-07, "epoch": 1.8443579766536966 }, { "current_steps": 948, "loss": 2.2506, "learning_rate": 6.587044534412956e-07, "epoch": 1.8463035019455254 }, { "current_steps": 949, "loss": 1.9714, "learning_rate": 6.582995951417004e-07, "epoch": 1.8482490272373542 }, { "current_steps": 950, "loss": 2.2811, "learning_rate": 6.578947368421053e-07, "epoch": 1.850194552529183 }, { "current_steps": 951, "loss": 2.1276, "learning_rate": 6.574898785425102e-07, "epoch": 1.8521400778210118 }, { "current_steps": 952, "loss": 2.1436, "learning_rate": 6.570850202429149e-07, "epoch": 1.8540856031128405 }, { "current_steps": 953, "loss": 2.2332, "learning_rate": 6.566801619433198e-07, "epoch": 1.8560311284046693 }, { "current_steps": 954, "loss": 2.174, "learning_rate": 6.562753036437247e-07, "epoch": 1.8579766536964981 }, { "current_steps": 955, "loss": 2.2034, "learning_rate": 6.558704453441295e-07, "epoch": 1.859922178988327 }, { "current_steps": 956, "loss": 2.0853, "learning_rate": 6.554655870445344e-07, "epoch": 1.8618677042801557 }, { "current_steps": 957, "loss": 2.0299, "learning_rate": 6.550607287449393e-07, "epoch": 1.8638132295719845 }, { "current_steps": 958, "loss": 2.3678, "learning_rate": 6.54655870445344e-07, "epoch": 1.8657587548638133 }, { "current_steps": 959, "loss": 2.1474, "learning_rate": 6.54251012145749e-07, "epoch": 1.867704280155642 }, { "current_steps": 960, "loss": 1.6199, "learning_rate": 6.538461538461538e-07, "epoch": 1.8696498054474708 }, { "current_steps": 961, "loss": 2.3431, "learning_rate": 6.534412955465586e-07, "epoch": 1.8715953307392996 }, { "current_steps": 962, "loss": 2.2348, "learning_rate": 6.530364372469635e-07, "epoch": 1.8735408560311284 }, { "current_steps": 963, "loss": 2.2322, "learning_rate": 6.526315789473684e-07, "epoch": 1.8754863813229572 }, { "current_steps": 964, "loss": 2.1894, "learning_rate": 6.522267206477732e-07, "epoch": 1.877431906614786 }, { "current_steps": 965, "loss": 2.1138, "learning_rate": 6.518218623481781e-07, "epoch": 1.8793774319066148 }, { "current_steps": 966, "loss": 1.8524, "learning_rate": 6.51417004048583e-07, "epoch": 1.8813229571984436 }, { "current_steps": 967, "loss": 2.2913, "learning_rate": 6.510121457489877e-07, "epoch": 1.8832684824902723 }, { "current_steps": 968, "loss": 1.8844, "learning_rate": 6.506072874493926e-07, "epoch": 1.8852140077821011 }, { "current_steps": 969, "loss": 2.1253, "learning_rate": 6.502024291497975e-07, "epoch": 1.88715953307393 }, { "current_steps": 970, "loss": 2.0752, "learning_rate": 6.497975708502024e-07, "epoch": 1.8891050583657587 }, { "current_steps": 971, "loss": 2.2119, "learning_rate": 6.493927125506072e-07, "epoch": 1.8910505836575875 }, { "current_steps": 972, "loss": 2.1737, "learning_rate": 6.489878542510121e-07, "epoch": 1.8929961089494163 }, { "current_steps": 973, "loss": 2.1406, "learning_rate": 6.48582995951417e-07, "epoch": 1.894941634241245 }, { "current_steps": 974, "loss": 2.2259, "learning_rate": 6.481781376518218e-07, "epoch": 1.8968871595330739 }, { "current_steps": 975, "loss": 2.2293, "learning_rate": 6.477732793522267e-07, "epoch": 1.8988326848249026 }, { "current_steps": 976, "loss": 2.2172, "learning_rate": 6.473684210526316e-07, "epoch": 1.9007782101167314 }, { "current_steps": 977, "loss": 1.7384, "learning_rate": 6.469635627530364e-07, "epoch": 1.9027237354085602 }, { "current_steps": 978, "loss": 1.6871, "learning_rate": 6.465587044534413e-07, "epoch": 1.904669260700389 }, { "current_steps": 979, "loss": 2.2959, "learning_rate": 6.461538461538462e-07, "epoch": 1.9066147859922178 }, { "current_steps": 980, "loss": 2.0458, "learning_rate": 6.45748987854251e-07, "epoch": 1.9085603112840466 }, { "current_steps": 981, "loss": 2.0137, "learning_rate": 6.453441295546559e-07, "epoch": 1.9105058365758754 }, { "current_steps": 982, "loss": 2.3689, "learning_rate": 6.449392712550607e-07, "epoch": 1.9124513618677041 }, { "current_steps": 983, "loss": 2.4528, "learning_rate": 6.445344129554655e-07, "epoch": 1.914396887159533 }, { "current_steps": 984, "loss": 2.1354, "learning_rate": 6.441295546558704e-07, "epoch": 1.9163424124513617 }, { "current_steps": 985, "loss": 2.1249, "learning_rate": 6.437246963562753e-07, "epoch": 1.9182879377431905 }, { "current_steps": 986, "loss": 2.4498, "learning_rate": 6.433198380566801e-07, "epoch": 1.9202334630350193 }, { "current_steps": 987, "loss": 2.1292, "learning_rate": 6.42914979757085e-07, "epoch": 1.9221789883268483 }, { "current_steps": 988, "loss": 2.4469, "learning_rate": 6.425101214574899e-07, "epoch": 1.924124513618677 }, { "current_steps": 989, "loss": 2.0244, "learning_rate": 6.421052631578947e-07, "epoch": 1.9260700389105059 }, { "current_steps": 990, "loss": 2.1983, "learning_rate": 6.417004048582996e-07, "epoch": 1.9280155642023347 }, { "current_steps": 991, "loss": 2.3918, "learning_rate": 6.412955465587045e-07, "epoch": 1.9299610894941635 }, { "current_steps": 992, "loss": 2.1551, "learning_rate": 6.408906882591093e-07, "epoch": 1.9319066147859922 }, { "current_steps": 993, "loss": 1.5777, "learning_rate": 6.404858299595142e-07, "epoch": 1.933852140077821 }, { "current_steps": 994, "loss": 2.0895, "learning_rate": 6.400809716599191e-07, "epoch": 1.9357976653696498 }, { "current_steps": 995, "loss": 1.6723, "learning_rate": 6.396761133603238e-07, "epoch": 1.9377431906614786 }, { "current_steps": 996, "loss": 2.2698, "learning_rate": 6.392712550607287e-07, "epoch": 1.9396887159533074 }, { "current_steps": 997, "loss": 2.1898, "learning_rate": 6.388663967611335e-07, "epoch": 1.9416342412451362 }, { "current_steps": 998, "loss": 1.8646, "learning_rate": 6.384615384615383e-07, "epoch": 1.943579766536965 }, { "current_steps": 999, "loss": 2.2302, "learning_rate": 6.380566801619432e-07, "epoch": 1.9455252918287937 }, { "current_steps": 1000, "loss": 2.1008, "learning_rate": 6.376518218623481e-07, "epoch": 1.9474708171206225 }, { "current_steps": 1001, "loss": 1.7757, "learning_rate": 6.37246963562753e-07, "epoch": 1.9494163424124513 }, { "current_steps": 1002, "loss": 2.3053, "learning_rate": 6.368421052631578e-07, "epoch": 1.95136186770428 }, { "current_steps": 1003, "loss": 1.8822, "learning_rate": 6.364372469635627e-07, "epoch": 1.9533073929961091 }, { "current_steps": 1004, "loss": 2.2743, "learning_rate": 6.360323886639676e-07, "epoch": 1.955252918287938 }, { "current_steps": 1005, "loss": 2.2723, "learning_rate": 6.356275303643724e-07, "epoch": 1.9571984435797667 }, { "current_steps": 1006, "loss": 2.1287, "learning_rate": 6.352226720647773e-07, "epoch": 1.9591439688715955 }, { "current_steps": 1007, "loss": 2.1775, "learning_rate": 6.348178137651822e-07, "epoch": 1.9610894941634243 }, { "current_steps": 1008, "loss": 2.2843, "learning_rate": 6.34412955465587e-07, "epoch": 1.963035019455253 }, { "current_steps": 1009, "loss": 1.738, "learning_rate": 6.340080971659919e-07, "epoch": 1.9649805447470818 }, { "current_steps": 1010, "loss": 2.3333, "learning_rate": 6.336032388663968e-07, "epoch": 1.9669260700389106 }, { "current_steps": 1011, "loss": 1.7123, "learning_rate": 6.331983805668016e-07, "epoch": 1.9688715953307394 }, { "current_steps": 1012, "loss": 2.2074, "learning_rate": 6.327935222672065e-07, "epoch": 1.9708171206225682 }, { "current_steps": 1013, "loss": 2.266, "learning_rate": 6.323886639676113e-07, "epoch": 1.972762645914397 }, { "current_steps": 1014, "loss": 2.198, "learning_rate": 6.319838056680161e-07, "epoch": 1.9747081712062258 }, { "current_steps": 1015, "loss": 2.0522, "learning_rate": 6.31578947368421e-07, "epoch": 1.9766536964980546 }, { "current_steps": 1016, "loss": 2.1179, "learning_rate": 6.311740890688259e-07, "epoch": 1.9785992217898833 }, { "current_steps": 1017, "loss": 1.9921, "learning_rate": 6.307692307692307e-07, "epoch": 1.9805447470817121 }, { "current_steps": 1018, "loss": 2.0678, "learning_rate": 6.303643724696356e-07, "epoch": 1.982490272373541 }, { "current_steps": 1019, "loss": 1.9707, "learning_rate": 6.299595141700405e-07, "epoch": 1.9844357976653697 }, { "current_steps": 1020, "loss": 1.9077, "learning_rate": 6.295546558704453e-07, "epoch": 1.9863813229571985 }, { "current_steps": 1021, "loss": 2.0573, "learning_rate": 6.291497975708502e-07, "epoch": 1.9883268482490273 }, { "current_steps": 1022, "loss": 1.9802, "learning_rate": 6.287449392712551e-07, "epoch": 1.990272373540856 }, { "current_steps": 1023, "loss": 1.0103, "learning_rate": 6.283400809716599e-07, "epoch": 1.9922178988326849 }, { "current_steps": 1024, "loss": 1.9342, "learning_rate": 6.279352226720648e-07, "epoch": 1.9941634241245136 }, { "current_steps": 1025, "loss": 2.3557, "learning_rate": 6.275303643724697e-07, "epoch": 1.9961089494163424 }, { "current_steps": 1026, "loss": 2.2177, "learning_rate": 6.271255060728745e-07, "epoch": 1.9980544747081712 }, { "current_steps": 1027, "loss": 1.5673, "learning_rate": 6.267206477732794e-07, "epoch": 2.0 }, { "current_steps": 1028, "loss": 2.2539, "learning_rate": 6.263157894736842e-07, "epoch": 2.001945525291829 }, { "current_steps": 1029, "loss": 2.2417, "learning_rate": 6.25910931174089e-07, "epoch": 2.0038910505836576 }, { "current_steps": 1030, "loss": 2.3209, "learning_rate": 6.255060728744938e-07, "epoch": 2.0058365758754864 }, { "current_steps": 1031, "loss": 2.1119, "learning_rate": 6.251012145748987e-07, "epoch": 2.007782101167315 }, { "current_steps": 1032, "loss": 2.2721, "learning_rate": 6.246963562753035e-07, "epoch": 2.009727626459144 }, { "current_steps": 1033, "loss": 2.2338, "learning_rate": 6.242914979757084e-07, "epoch": 2.0116731517509727 }, { "current_steps": 1034, "loss": 2.1364, "learning_rate": 6.238866396761133e-07, "epoch": 2.0136186770428015 }, { "current_steps": 1035, "loss": 2.3221, "learning_rate": 6.234817813765182e-07, "epoch": 2.0155642023346303 }, { "current_steps": 1036, "loss": 1.8971, "learning_rate": 6.23076923076923e-07, "epoch": 2.017509727626459 }, { "current_steps": 1037, "loss": 2.419, "learning_rate": 6.226720647773279e-07, "epoch": 2.019455252918288 }, { "current_steps": 1038, "loss": 2.2517, "learning_rate": 6.222672064777328e-07, "epoch": 2.0214007782101167 }, { "current_steps": 1039, "loss": 1.9879, "learning_rate": 6.218623481781376e-07, "epoch": 2.0233463035019454 }, { "current_steps": 1040, "loss": 1.6718, "learning_rate": 6.214574898785425e-07, "epoch": 2.0252918287937742 }, { "current_steps": 1041, "loss": 2.1453, "learning_rate": 6.210526315789474e-07, "epoch": 2.027237354085603 }, { "current_steps": 1042, "loss": 2.2453, "learning_rate": 6.206477732793522e-07, "epoch": 2.029182879377432 }, { "current_steps": 1043, "loss": 2.1581, "learning_rate": 6.20242914979757e-07, "epoch": 2.0311284046692606 }, { "current_steps": 1044, "loss": 2.0303, "learning_rate": 6.198380566801619e-07, "epoch": 2.0330739299610894 }, { "current_steps": 1045, "loss": 2.2907, "learning_rate": 6.194331983805667e-07, "epoch": 2.035019455252918 }, { "current_steps": 1046, "loss": 2.2664, "learning_rate": 6.190283400809716e-07, "epoch": 2.036964980544747 }, { "current_steps": 1047, "loss": 2.2142, "learning_rate": 6.186234817813765e-07, "epoch": 2.0389105058365757 }, { "current_steps": 1048, "loss": 2.0748, "learning_rate": 6.182186234817813e-07, "epoch": 2.0408560311284045 }, { "current_steps": 1049, "loss": 1.9869, "learning_rate": 6.178137651821862e-07, "epoch": 2.0428015564202333 }, { "current_steps": 1050, "loss": 2.0667, "learning_rate": 6.174089068825911e-07, "epoch": 2.044747081712062 }, { "current_steps": 1051, "loss": 2.2627, "learning_rate": 6.170040485829959e-07, "epoch": 2.046692607003891 }, { "current_steps": 1052, "loss": 2.1681, "learning_rate": 6.165991902834008e-07, "epoch": 2.0486381322957197 }, { "current_steps": 1053, "loss": 2.0929, "learning_rate": 6.161943319838057e-07, "epoch": 2.0505836575875485 }, { "current_steps": 1054, "loss": 2.2611, "learning_rate": 6.157894736842105e-07, "epoch": 2.0525291828793772 }, { "current_steps": 1055, "loss": 1.9106, "learning_rate": 6.153846153846154e-07, "epoch": 2.054474708171206 }, { "current_steps": 1056, "loss": 2.0726, "learning_rate": 6.149797570850203e-07, "epoch": 2.056420233463035 }, { "current_steps": 1057, "loss": 1.8697, "learning_rate": 6.145748987854251e-07, "epoch": 2.0583657587548636 }, { "current_steps": 1058, "loss": 2.0629, "learning_rate": 6.141700404858299e-07, "epoch": 2.0603112840466924 }, { "current_steps": 1059, "loss": 2.3941, "learning_rate": 6.137651821862348e-07, "epoch": 2.062256809338521 }, { "current_steps": 1060, "loss": 2.1068, "learning_rate": 6.133603238866396e-07, "epoch": 2.06420233463035 }, { "current_steps": 1061, "loss": 2.341, "learning_rate": 6.129554655870445e-07, "epoch": 2.066147859922179 }, { "current_steps": 1062, "loss": 2.3541, "learning_rate": 6.125506072874494e-07, "epoch": 2.068093385214008 }, { "current_steps": 1063, "loss": 2.3433, "learning_rate": 6.121457489878542e-07, "epoch": 2.0700389105058368 }, { "current_steps": 1064, "loss": 2.2093, "learning_rate": 6.11740890688259e-07, "epoch": 2.0719844357976656 }, { "current_steps": 1065, "loss": 2.4843, "learning_rate": 6.11336032388664e-07, "epoch": 2.0739299610894943 }, { "current_steps": 1066, "loss": 2.2517, "learning_rate": 6.109311740890689e-07, "epoch": 2.075875486381323 }, { "current_steps": 1067, "loss": 2.2333, "learning_rate": 6.105263157894736e-07, "epoch": 2.077821011673152 }, { "current_steps": 1068, "loss": 2.1923, "learning_rate": 6.101214574898785e-07, "epoch": 2.0797665369649807 }, { "current_steps": 1069, "loss": 2.2327, "learning_rate": 6.097165991902834e-07, "epoch": 2.0817120622568095 }, { "current_steps": 1070, "loss": 2.2926, "learning_rate": 6.093117408906882e-07, "epoch": 2.0836575875486383 }, { "current_steps": 1071, "loss": 2.0572, "learning_rate": 6.089068825910931e-07, "epoch": 2.085603112840467 }, { "current_steps": 1072, "loss": 2.2682, "learning_rate": 6.08502024291498e-07, "epoch": 2.087548638132296 }, { "current_steps": 1073, "loss": 2.1995, "learning_rate": 6.080971659919028e-07, "epoch": 2.0894941634241246 }, { "current_steps": 1074, "loss": 1.9126, "learning_rate": 6.076923076923076e-07, "epoch": 2.0914396887159534 }, { "current_steps": 1075, "loss": 1.693, "learning_rate": 6.072874493927125e-07, "epoch": 2.093385214007782 }, { "current_steps": 1076, "loss": 2.3834, "learning_rate": 6.068825910931173e-07, "epoch": 2.095330739299611 }, { "current_steps": 1077, "loss": 1.7341, "learning_rate": 6.064777327935222e-07, "epoch": 2.09727626459144 }, { "current_steps": 1078, "loss": 2.0287, "learning_rate": 6.060728744939271e-07, "epoch": 2.0992217898832686 }, { "current_steps": 1079, "loss": 2.0829, "learning_rate": 6.056680161943319e-07, "epoch": 2.1011673151750974 }, { "current_steps": 1080, "loss": 2.2202, "learning_rate": 6.052631578947368e-07, "epoch": 2.103112840466926 }, { "current_steps": 1081, "loss": 1.7932, "learning_rate": 6.048582995951417e-07, "epoch": 2.105058365758755 }, { "current_steps": 1082, "loss": 2.2145, "learning_rate": 6.044534412955465e-07, "epoch": 2.1070038910505837 }, { "current_steps": 1083, "loss": 2.0782, "learning_rate": 6.040485829959514e-07, "epoch": 2.1089494163424125 }, { "current_steps": 1084, "loss": 2.259, "learning_rate": 6.036437246963563e-07, "epoch": 2.1108949416342413 }, { "current_steps": 1085, "loss": 1.9896, "learning_rate": 6.032388663967611e-07, "epoch": 2.11284046692607 }, { "current_steps": 1086, "loss": 2.0998, "learning_rate": 6.02834008097166e-07, "epoch": 2.114785992217899 }, { "current_steps": 1087, "loss": 2.319, "learning_rate": 6.024291497975709e-07, "epoch": 2.1167315175097277 }, { "current_steps": 1088, "loss": 2.1565, "learning_rate": 6.020242914979757e-07, "epoch": 2.1186770428015564 }, { "current_steps": 1089, "loss": 1.8162, "learning_rate": 6.016194331983805e-07, "epoch": 2.1206225680933852 }, { "current_steps": 1090, "loss": 2.0901, "learning_rate": 6.012145748987854e-07, "epoch": 2.122568093385214 }, { "current_steps": 1091, "loss": 2.156, "learning_rate": 6.008097165991902e-07, "epoch": 2.124513618677043 }, { "current_steps": 1092, "loss": 2.4021, "learning_rate": 6.004048582995951e-07, "epoch": 2.1264591439688716 }, { "current_steps": 1093, "loss": 2.1235, "learning_rate": 6e-07, "epoch": 2.1284046692607004 }, { "current_steps": 1094, "loss": 2.2998, "learning_rate": 5.995951417004048e-07, "epoch": 2.130350194552529 }, { "current_steps": 1095, "loss": 2.4521, "learning_rate": 5.991902834008097e-07, "epoch": 2.132295719844358 }, { "current_steps": 1096, "loss": 1.9548, "learning_rate": 5.987854251012146e-07, "epoch": 2.1342412451361867 }, { "current_steps": 1097, "loss": 2.3032, "learning_rate": 5.983805668016195e-07, "epoch": 2.1361867704280155 }, { "current_steps": 1098, "loss": 2.2136, "learning_rate": 5.979757085020243e-07, "epoch": 2.1381322957198443 }, { "current_steps": 1099, "loss": 2.2424, "learning_rate": 5.975708502024292e-07, "epoch": 2.140077821011673 }, { "current_steps": 1100, "loss": 1.8524, "learning_rate": 5.971659919028341e-07, "epoch": 2.142023346303502 }, { "current_steps": 1101, "loss": 1.648, "learning_rate": 5.967611336032389e-07, "epoch": 2.1439688715953307 }, { "current_steps": 1102, "loss": 2.1178, "learning_rate": 5.963562753036438e-07, "epoch": 2.1459143968871595 }, { "current_steps": 1103, "loss": 2.2784, "learning_rate": 5.959514170040486e-07, "epoch": 2.1478599221789882 }, { "current_steps": 1104, "loss": 2.0001, "learning_rate": 5.955465587044533e-07, "epoch": 2.149805447470817 }, { "current_steps": 1105, "loss": 2.0541, "learning_rate": 5.951417004048582e-07, "epoch": 2.151750972762646 }, { "current_steps": 1106, "loss": 2.3298, "learning_rate": 5.947368421052631e-07, "epoch": 2.1536964980544746 }, { "current_steps": 1107, "loss": 2.3313, "learning_rate": 5.943319838056679e-07, "epoch": 2.1556420233463034 }, { "current_steps": 1108, "loss": 2.2263, "learning_rate": 5.939271255060728e-07, "epoch": 2.157587548638132 }, { "current_steps": 1109, "loss": 2.1405, "learning_rate": 5.935222672064777e-07, "epoch": 2.159533073929961 }, { "current_steps": 1110, "loss": 2.1838, "learning_rate": 5.931174089068825e-07, "epoch": 2.1614785992217898 }, { "current_steps": 1111, "loss": 2.2917, "learning_rate": 5.927125506072874e-07, "epoch": 2.1634241245136185 }, { "current_steps": 1112, "loss": 2.2211, "learning_rate": 5.923076923076923e-07, "epoch": 2.1653696498054473 }, { "current_steps": 1113, "loss": 2.0071, "learning_rate": 5.919028340080971e-07, "epoch": 2.167315175097276 }, { "current_steps": 1114, "loss": 2.1164, "learning_rate": 5.91497975708502e-07, "epoch": 2.169260700389105 }, { "current_steps": 1115, "loss": 2.0838, "learning_rate": 5.910931174089069e-07, "epoch": 2.1712062256809337 }, { "current_steps": 1116, "loss": 2.2119, "learning_rate": 5.906882591093117e-07, "epoch": 2.1731517509727625 }, { "current_steps": 1117, "loss": 2.1911, "learning_rate": 5.902834008097166e-07, "epoch": 2.1750972762645913 }, { "current_steps": 1118, "loss": 2.2673, "learning_rate": 5.898785425101215e-07, "epoch": 2.17704280155642 }, { "current_steps": 1119, "loss": 2.0319, "learning_rate": 5.894736842105262e-07, "epoch": 2.178988326848249 }, { "current_steps": 1120, "loss": 2.0183, "learning_rate": 5.890688259109311e-07, "epoch": 2.1809338521400776 }, { "current_steps": 1121, "loss": 2.2102, "learning_rate": 5.88663967611336e-07, "epoch": 2.1828793774319064 }, { "current_steps": 1122, "loss": 2.1534, "learning_rate": 5.882591093117408e-07, "epoch": 2.1848249027237356 }, { "current_steps": 1123, "loss": 2.1357, "learning_rate": 5.878542510121457e-07, "epoch": 2.1867704280155644 }, { "current_steps": 1124, "loss": 2.2346, "learning_rate": 5.874493927125506e-07, "epoch": 2.188715953307393 }, { "current_steps": 1125, "loss": 2.1646, "learning_rate": 5.870445344129554e-07, "epoch": 2.190661478599222 }, { "current_steps": 1126, "loss": 2.2764, "learning_rate": 5.866396761133603e-07, "epoch": 2.192607003891051 }, { "current_steps": 1127, "loss": 2.1115, "learning_rate": 5.862348178137652e-07, "epoch": 2.1945525291828796 }, { "current_steps": 1128, "loss": 1.9998, "learning_rate": 5.8582995951417e-07, "epoch": 2.1964980544747084 }, { "current_steps": 1129, "loss": 2.3608, "learning_rate": 5.854251012145749e-07, "epoch": 2.198443579766537 }, { "current_steps": 1130, "loss": 2.3284, "learning_rate": 5.850202429149798e-07, "epoch": 2.200389105058366 }, { "current_steps": 1131, "loss": 1.9998, "learning_rate": 5.846153846153847e-07, "epoch": 2.2023346303501947 }, { "current_steps": 1132, "loss": 2.1586, "learning_rate": 5.842105263157895e-07, "epoch": 2.2042801556420235 }, { "current_steps": 1133, "loss": 2.1237, "learning_rate": 5.838056680161944e-07, "epoch": 2.2062256809338523 }, { "current_steps": 1134, "loss": 2.1983, "learning_rate": 5.834008097165993e-07, "epoch": 2.208171206225681 }, { "current_steps": 1135, "loss": 2.0564, "learning_rate": 5.82995951417004e-07, "epoch": 2.21011673151751 }, { "current_steps": 1136, "loss": 2.3481, "learning_rate": 5.825910931174089e-07, "epoch": 2.2120622568093387 }, { "current_steps": 1137, "loss": 2.4831, "learning_rate": 5.821862348178137e-07, "epoch": 2.2140077821011674 }, { "current_steps": 1138, "loss": 2.1214, "learning_rate": 5.817813765182185e-07, "epoch": 2.2159533073929962 }, { "current_steps": 1139, "loss": 1.9326, "learning_rate": 5.813765182186234e-07, "epoch": 2.217898832684825 }, { "current_steps": 1140, "loss": 1.9875, "learning_rate": 5.809716599190283e-07, "epoch": 2.219844357976654 }, { "current_steps": 1141, "loss": 2.0973, "learning_rate": 5.805668016194331e-07, "epoch": 2.2217898832684826 }, { "current_steps": 1142, "loss": 1.7724, "learning_rate": 5.80161943319838e-07, "epoch": 2.2237354085603114 }, { "current_steps": 1143, "loss": 1.8333, "learning_rate": 5.797570850202429e-07, "epoch": 2.22568093385214 }, { "current_steps": 1144, "loss": 2.147, "learning_rate": 5.793522267206477e-07, "epoch": 2.227626459143969 }, { "current_steps": 1145, "loss": 1.7203, "learning_rate": 5.789473684210526e-07, "epoch": 2.2295719844357977 }, { "current_steps": 1146, "loss": 2.1252, "learning_rate": 5.785425101214575e-07, "epoch": 2.2315175097276265 }, { "current_steps": 1147, "loss": 2.0983, "learning_rate": 5.781376518218623e-07, "epoch": 2.2334630350194553 }, { "current_steps": 1148, "loss": 2.1005, "learning_rate": 5.777327935222672e-07, "epoch": 2.235408560311284 }, { "current_steps": 1149, "loss": 1.8834, "learning_rate": 5.773279352226721e-07, "epoch": 2.237354085603113 }, { "current_steps": 1150, "loss": 1.7282, "learning_rate": 5.769230769230768e-07, "epoch": 2.2392996108949417 }, { "current_steps": 1151, "loss": 1.9915, "learning_rate": 5.765182186234817e-07, "epoch": 2.2412451361867705 }, { "current_steps": 1152, "loss": 1.9982, "learning_rate": 5.761133603238866e-07, "epoch": 2.2431906614785992 }, { "current_steps": 1153, "loss": 2.2858, "learning_rate": 5.757085020242914e-07, "epoch": 2.245136186770428 }, { "current_steps": 1154, "loss": 2.3025, "learning_rate": 5.753036437246963e-07, "epoch": 2.247081712062257 }, { "current_steps": 1155, "loss": 2.2761, "learning_rate": 5.748987854251012e-07, "epoch": 2.2490272373540856 }, { "current_steps": 1156, "loss": 2.2282, "learning_rate": 5.74493927125506e-07, "epoch": 2.2509727626459144 }, { "current_steps": 1157, "loss": 2.0037, "learning_rate": 5.740890688259109e-07, "epoch": 2.252918287937743 }, { "current_steps": 1158, "loss": 2.1332, "learning_rate": 5.736842105263158e-07, "epoch": 2.254863813229572 }, { "current_steps": 1159, "loss": 2.0577, "learning_rate": 5.732793522267206e-07, "epoch": 2.2568093385214008 }, { "current_steps": 1160, "loss": 1.4825, "learning_rate": 5.728744939271255e-07, "epoch": 2.2587548638132295 }, { "current_steps": 1161, "loss": 1.923, "learning_rate": 5.724696356275304e-07, "epoch": 2.2607003891050583 }, { "current_steps": 1162, "loss": 1.6041, "learning_rate": 5.720647773279353e-07, "epoch": 2.262645914396887 }, { "current_steps": 1163, "loss": 2.3999, "learning_rate": 5.716599190283401e-07, "epoch": 2.264591439688716 }, { "current_steps": 1164, "loss": 2.0982, "learning_rate": 5.71255060728745e-07, "epoch": 2.2665369649805447 }, { "current_steps": 1165, "loss": 1.8588, "learning_rate": 5.708502024291498e-07, "epoch": 2.2684824902723735 }, { "current_steps": 1166, "loss": 2.0859, "learning_rate": 5.704453441295546e-07, "epoch": 2.2704280155642023 }, { "current_steps": 1167, "loss": 1.8949, "learning_rate": 5.700404858299595e-07, "epoch": 2.272373540856031 }, { "current_steps": 1168, "loss": 2.2134, "learning_rate": 5.696356275303644e-07, "epoch": 2.27431906614786 }, { "current_steps": 1169, "loss": 2.2925, "learning_rate": 5.692307692307692e-07, "epoch": 2.2762645914396886 }, { "current_steps": 1170, "loss": 2.3243, "learning_rate": 5.688259109311741e-07, "epoch": 2.2782101167315174 }, { "current_steps": 1171, "loss": 2.433, "learning_rate": 5.68421052631579e-07, "epoch": 2.280155642023346 }, { "current_steps": 1172, "loss": 2.1877, "learning_rate": 5.680161943319837e-07, "epoch": 2.282101167315175 }, { "current_steps": 1173, "loss": 2.1938, "learning_rate": 5.676113360323886e-07, "epoch": 2.2840466926070038 }, { "current_steps": 1174, "loss": 2.0752, "learning_rate": 5.672064777327935e-07, "epoch": 2.2859922178988326 }, { "current_steps": 1175, "loss": 2.1431, "learning_rate": 5.668016194331983e-07, "epoch": 2.2879377431906613 }, { "current_steps": 1176, "loss": 2.1396, "learning_rate": 5.663967611336032e-07, "epoch": 2.28988326848249 }, { "current_steps": 1177, "loss": 2.1474, "learning_rate": 5.659919028340081e-07, "epoch": 2.291828793774319 }, { "current_steps": 1178, "loss": 1.97, "learning_rate": 5.655870445344129e-07, "epoch": 2.2937743190661477 }, { "current_steps": 1179, "loss": 2.2989, "learning_rate": 5.651821862348178e-07, "epoch": 2.2957198443579765 }, { "current_steps": 1180, "loss": 2.679, "learning_rate": 5.647773279352226e-07, "epoch": 2.2976653696498053 }, { "current_steps": 1181, "loss": 2.2135, "learning_rate": 5.643724696356274e-07, "epoch": 2.299610894941634 }, { "current_steps": 1182, "loss": 2.1922, "learning_rate": 5.639676113360323e-07, "epoch": 2.301556420233463 }, { "current_steps": 1183, "loss": 1.6651, "learning_rate": 5.635627530364372e-07, "epoch": 2.3035019455252916 }, { "current_steps": 1184, "loss": 2.2459, "learning_rate": 5.63157894736842e-07, "epoch": 2.3054474708171204 }, { "current_steps": 1185, "loss": 1.553, "learning_rate": 5.627530364372469e-07, "epoch": 2.307392996108949 }, { "current_steps": 1186, "loss": 2.2349, "learning_rate": 5.623481781376518e-07, "epoch": 2.309338521400778 }, { "current_steps": 1187, "loss": 2.0194, "learning_rate": 5.619433198380566e-07, "epoch": 2.311284046692607 }, { "current_steps": 1188, "loss": 2.3264, "learning_rate": 5.615384615384615e-07, "epoch": 2.3132295719844356 }, { "current_steps": 1189, "loss": 1.4863, "learning_rate": 5.611336032388664e-07, "epoch": 2.3151750972762644 }, { "current_steps": 1190, "loss": 2.2276, "learning_rate": 5.607287449392712e-07, "epoch": 2.317120622568093 }, { "current_steps": 1191, "loss": 2.2852, "learning_rate": 5.603238866396761e-07, "epoch": 2.319066147859922 }, { "current_steps": 1192, "loss": 2.2729, "learning_rate": 5.59919028340081e-07, "epoch": 2.321011673151751 }, { "current_steps": 1193, "loss": 2.1116, "learning_rate": 5.595141700404859e-07, "epoch": 2.32295719844358 }, { "current_steps": 1194, "loss": 1.7806, "learning_rate": 5.591093117408907e-07, "epoch": 2.3249027237354087 }, { "current_steps": 1195, "loss": 2.1672, "learning_rate": 5.587044534412956e-07, "epoch": 2.3268482490272375 }, { "current_steps": 1196, "loss": 1.7839, "learning_rate": 5.582995951417004e-07, "epoch": 2.3287937743190663 }, { "current_steps": 1197, "loss": 2.0575, "learning_rate": 5.578947368421052e-07, "epoch": 2.330739299610895 }, { "current_steps": 1198, "loss": 1.8851, "learning_rate": 5.574898785425101e-07, "epoch": 2.332684824902724 }, { "current_steps": 1199, "loss": 2.2176, "learning_rate": 5.57085020242915e-07, "epoch": 2.3346303501945527 }, { "current_steps": 1200, "loss": 1.9838, "learning_rate": 5.566801619433198e-07, "epoch": 2.3365758754863815 }, { "current_steps": 1201, "loss": 2.1483, "learning_rate": 5.562753036437247e-07, "epoch": 2.3385214007782102 }, { "current_steps": 1202, "loss": 2.0405, "learning_rate": 5.558704453441296e-07, "epoch": 2.340466926070039 }, { "current_steps": 1203, "loss": 1.6789, "learning_rate": 5.554655870445344e-07, "epoch": 2.342412451361868 }, { "current_steps": 1204, "loss": 2.107, "learning_rate": 5.550607287449393e-07, "epoch": 2.3443579766536966 }, { "current_steps": 1205, "loss": 2.1725, "learning_rate": 5.546558704453442e-07, "epoch": 2.3463035019455254 }, { "current_steps": 1206, "loss": 1.654, "learning_rate": 5.54251012145749e-07, "epoch": 2.348249027237354 }, { "current_steps": 1207, "loss": 2.3965, "learning_rate": 5.538461538461539e-07, "epoch": 2.350194552529183 }, { "current_steps": 1208, "loss": 2.0567, "learning_rate": 5.534412955465588e-07, "epoch": 2.3521400778210118 }, { "current_steps": 1209, "loss": 1.8693, "learning_rate": 5.530364372469635e-07, "epoch": 2.3540856031128405 }, { "current_steps": 1210, "loss": 2.2347, "learning_rate": 5.526315789473684e-07, "epoch": 2.3560311284046693 }, { "current_steps": 1211, "loss": 2.3452, "learning_rate": 5.522267206477732e-07, "epoch": 2.357976653696498 }, { "current_steps": 1212, "loss": 2.145, "learning_rate": 5.51821862348178e-07, "epoch": 2.359922178988327 }, { "current_steps": 1213, "loss": 2.1702, "learning_rate": 5.514170040485829e-07, "epoch": 2.3618677042801557 }, { "current_steps": 1214, "loss": 2.2127, "learning_rate": 5.510121457489878e-07, "epoch": 2.3638132295719845 }, { "current_steps": 1215, "loss": 1.9574, "learning_rate": 5.506072874493926e-07, "epoch": 2.3657587548638133 }, { "current_steps": 1216, "loss": 1.8465, "learning_rate": 5.502024291497975e-07, "epoch": 2.367704280155642 }, { "current_steps": 1217, "loss": 2.2272, "learning_rate": 5.497975708502024e-07, "epoch": 2.369649805447471 }, { "current_steps": 1218, "loss": 2.2423, "learning_rate": 5.493927125506072e-07, "epoch": 2.3715953307392996 }, { "current_steps": 1219, "loss": 1.8412, "learning_rate": 5.489878542510121e-07, "epoch": 2.3735408560311284 }, { "current_steps": 1220, "loss": 1.8803, "learning_rate": 5.48582995951417e-07, "epoch": 2.375486381322957 }, { "current_steps": 1221, "loss": 2.2457, "learning_rate": 5.481781376518218e-07, "epoch": 2.377431906614786 }, { "current_steps": 1222, "loss": 2.3056, "learning_rate": 5.477732793522267e-07, "epoch": 2.3793774319066148 }, { "current_steps": 1223, "loss": 2.2342, "learning_rate": 5.473684210526316e-07, "epoch": 2.3813229571984436 }, { "current_steps": 1224, "loss": 2.0727, "learning_rate": 5.469635627530364e-07, "epoch": 2.3832684824902723 }, { "current_steps": 1225, "loss": 2.1526, "learning_rate": 5.465587044534413e-07, "epoch": 2.385214007782101 }, { "current_steps": 1226, "loss": 1.8585, "learning_rate": 5.461538461538461e-07, "epoch": 2.38715953307393 }, { "current_steps": 1227, "loss": 1.9822, "learning_rate": 5.457489878542509e-07, "epoch": 2.3891050583657587 }, { "current_steps": 1228, "loss": 2.3916, "learning_rate": 5.453441295546558e-07, "epoch": 2.3910505836575875 }, { "current_steps": 1229, "loss": 1.9773, "learning_rate": 5.449392712550607e-07, "epoch": 2.3929961089494163 }, { "current_steps": 1230, "loss": 2.0343, "learning_rate": 5.445344129554656e-07, "epoch": 2.394941634241245 }, { "current_steps": 1231, "loss": 2.1889, "learning_rate": 5.441295546558704e-07, "epoch": 2.396887159533074 }, { "current_steps": 1232, "loss": 2.295, "learning_rate": 5.437246963562753e-07, "epoch": 2.3988326848249026 }, { "current_steps": 1233, "loss": 2.1391, "learning_rate": 5.433198380566802e-07, "epoch": 2.4007782101167314 }, { "current_steps": 1234, "loss": 2.3434, "learning_rate": 5.42914979757085e-07, "epoch": 2.40272373540856 }, { "current_steps": 1235, "loss": 2.2972, "learning_rate": 5.425101214574899e-07, "epoch": 2.404669260700389 }, { "current_steps": 1236, "loss": 2.3311, "learning_rate": 5.421052631578948e-07, "epoch": 2.406614785992218 }, { "current_steps": 1237, "loss": 2.1704, "learning_rate": 5.417004048582996e-07, "epoch": 2.4085603112840466 }, { "current_steps": 1238, "loss": 1.8418, "learning_rate": 5.412955465587045e-07, "epoch": 2.4105058365758754 }, { "current_steps": 1239, "loss": 2.1497, "learning_rate": 5.408906882591094e-07, "epoch": 2.412451361867704 }, { "current_steps": 1240, "loss": 1.7333, "learning_rate": 5.404858299595142e-07, "epoch": 2.414396887159533 }, { "current_steps": 1241, "loss": 1.7713, "learning_rate": 5.40080971659919e-07, "epoch": 2.4163424124513617 }, { "current_steps": 1242, "loss": 2.2353, "learning_rate": 5.396761133603239e-07, "epoch": 2.4182879377431905 }, { "current_steps": 1243, "loss": 2.0546, "learning_rate": 5.392712550607286e-07, "epoch": 2.4202334630350193 }, { "current_steps": 1244, "loss": 2.1859, "learning_rate": 5.388663967611335e-07, "epoch": 2.422178988326848 }, { "current_steps": 1245, "loss": 2.1308, "learning_rate": 5.384615384615384e-07, "epoch": 2.424124513618677 }, { "current_steps": 1246, "loss": 1.7735, "learning_rate": 5.380566801619432e-07, "epoch": 2.4260700389105057 }, { "current_steps": 1247, "loss": 2.0188, "learning_rate": 5.376518218623481e-07, "epoch": 2.4280155642023344 }, { "current_steps": 1248, "loss": 2.2957, "learning_rate": 5.37246963562753e-07, "epoch": 2.4299610894941637 }, { "current_steps": 1249, "loss": 2.2799, "learning_rate": 5.368421052631578e-07, "epoch": 2.4319066147859925 }, { "current_steps": 1250, "loss": 2.3201, "learning_rate": 5.364372469635627e-07, "epoch": 2.4338521400778212 }, { "current_steps": 1251, "loss": 1.9403, "learning_rate": 5.360323886639676e-07, "epoch": 2.43579766536965 }, { "current_steps": 1252, "loss": 2.2707, "learning_rate": 5.356275303643724e-07, "epoch": 2.437743190661479 }, { "current_steps": 1253, "loss": 1.9973, "learning_rate": 5.352226720647773e-07, "epoch": 2.4396887159533076 }, { "current_steps": 1254, "loss": 1.7221, "learning_rate": 5.348178137651822e-07, "epoch": 2.4416342412451364 }, { "current_steps": 1255, "loss": 2.3152, "learning_rate": 5.34412955465587e-07, "epoch": 2.443579766536965 }, { "current_steps": 1256, "loss": 2.2076, "learning_rate": 5.340080971659919e-07, "epoch": 2.445525291828794 }, { "current_steps": 1257, "loss": 2.3573, "learning_rate": 5.336032388663967e-07, "epoch": 2.4474708171206228 }, { "current_steps": 1258, "loss": 2.3294, "learning_rate": 5.331983805668015e-07, "epoch": 2.4494163424124515 }, { "current_steps": 1259, "loss": 2.4309, "learning_rate": 5.327935222672064e-07, "epoch": 2.4513618677042803 }, { "current_steps": 1260, "loss": 2.485, "learning_rate": 5.323886639676113e-07, "epoch": 2.453307392996109 }, { "current_steps": 1261, "loss": 2.338, "learning_rate": 5.319838056680162e-07, "epoch": 2.455252918287938 }, { "current_steps": 1262, "loss": 1.926, "learning_rate": 5.31578947368421e-07, "epoch": 2.4571984435797667 }, { "current_steps": 1263, "loss": 1.9253, "learning_rate": 5.311740890688259e-07, "epoch": 2.4591439688715955 }, { "current_steps": 1264, "loss": 2.2718, "learning_rate": 5.307692307692308e-07, "epoch": 2.4610894941634243 }, { "current_steps": 1265, "loss": 2.1952, "learning_rate": 5.303643724696356e-07, "epoch": 2.463035019455253 }, { "current_steps": 1266, "loss": 2.0586, "learning_rate": 5.299595141700405e-07, "epoch": 2.464980544747082 }, { "current_steps": 1267, "loss": 1.8643, "learning_rate": 5.295546558704454e-07, "epoch": 2.4669260700389106 }, { "current_steps": 1268, "loss": 2.26, "learning_rate": 5.291497975708502e-07, "epoch": 2.4688715953307394 }, { "current_steps": 1269, "loss": 2.0511, "learning_rate": 5.287449392712551e-07, "epoch": 2.470817120622568 }, { "current_steps": 1270, "loss": 2.3081, "learning_rate": 5.2834008097166e-07, "epoch": 2.472762645914397 }, { "current_steps": 1271, "loss": 2.1084, "learning_rate": 5.279352226720648e-07, "epoch": 2.4747081712062258 }, { "current_steps": 1272, "loss": 2.2567, "learning_rate": 5.275303643724696e-07, "epoch": 2.4766536964980546 }, { "current_steps": 1273, "loss": 2.1763, "learning_rate": 5.271255060728745e-07, "epoch": 2.4785992217898833 }, { "current_steps": 1274, "loss": 2.4158, "learning_rate": 5.267206477732793e-07, "epoch": 2.480544747081712 }, { "current_steps": 1275, "loss": 2.3051, "learning_rate": 5.263157894736842e-07, "epoch": 2.482490272373541 }, { "current_steps": 1276, "loss": 2.0592, "learning_rate": 5.259109311740891e-07, "epoch": 2.4844357976653697 }, { "current_steps": 1277, "loss": 2.1145, "learning_rate": 5.255060728744939e-07, "epoch": 2.4863813229571985 }, { "current_steps": 1278, "loss": 2.226, "learning_rate": 5.251012145748988e-07, "epoch": 2.4883268482490273 }, { "current_steps": 1279, "loss": 2.2325, "learning_rate": 5.246963562753037e-07, "epoch": 2.490272373540856 }, { "current_steps": 1280, "loss": 2.2251, "learning_rate": 5.242914979757084e-07, "epoch": 2.492217898832685 }, { "current_steps": 1281, "loss": 1.6811, "learning_rate": 5.238866396761133e-07, "epoch": 2.4941634241245136 }, { "current_steps": 1282, "loss": 1.6737, "learning_rate": 5.234817813765182e-07, "epoch": 2.4961089494163424 }, { "current_steps": 1283, "loss": 2.1409, "learning_rate": 5.23076923076923e-07, "epoch": 2.498054474708171 }, { "current_steps": 1284, "loss": 2.3759, "learning_rate": 5.226720647773279e-07, "epoch": 2.5 }, { "current_steps": 1285, "loss": 2.2151, "learning_rate": 5.222672064777328e-07, "epoch": 2.501945525291829 }, { "current_steps": 1286, "loss": 2.1287, "learning_rate": 5.218623481781376e-07, "epoch": 2.5038910505836576 }, { "current_steps": 1287, "loss": 2.1425, "learning_rate": 5.214574898785424e-07, "epoch": 2.5058365758754864 }, { "current_steps": 1288, "loss": 2.2199, "learning_rate": 5.210526315789473e-07, "epoch": 2.507782101167315 }, { "current_steps": 1289, "loss": 1.8668, "learning_rate": 5.206477732793521e-07, "epoch": 2.509727626459144 }, { "current_steps": 1290, "loss": 2.0994, "learning_rate": 5.20242914979757e-07, "epoch": 2.5116731517509727 }, { "current_steps": 1291, "loss": 2.1378, "learning_rate": 5.198380566801619e-07, "epoch": 2.5136186770428015 }, { "current_steps": 1292, "loss": 2.3491, "learning_rate": 5.194331983805668e-07, "epoch": 2.5155642023346303 }, { "current_steps": 1293, "loss": 1.8481, "learning_rate": 5.190283400809716e-07, "epoch": 2.517509727626459 }, { "current_steps": 1294, "loss": 2.1897, "learning_rate": 5.186234817813765e-07, "epoch": 2.519455252918288 }, { "current_steps": 1295, "loss": 1.5223, "learning_rate": 5.182186234817814e-07, "epoch": 2.5214007782101167 }, { "current_steps": 1296, "loss": 2.2289, "learning_rate": 5.178137651821862e-07, "epoch": 2.5233463035019454 }, { "current_steps": 1297, "loss": 2.1315, "learning_rate": 5.174089068825911e-07, "epoch": 2.5252918287937742 }, { "current_steps": 1298, "loss": 1.8147, "learning_rate": 5.17004048582996e-07, "epoch": 2.527237354085603 }, { "current_steps": 1299, "loss": 2.1897, "learning_rate": 5.165991902834008e-07, "epoch": 2.529182879377432 }, { "current_steps": 1300, "loss": 1.693, "learning_rate": 5.161943319838057e-07, "epoch": 2.5311284046692606 }, { "current_steps": 1301, "loss": 1.9529, "learning_rate": 5.157894736842106e-07, "epoch": 2.5330739299610894 }, { "current_steps": 1302, "loss": 2.2761, "learning_rate": 5.153846153846153e-07, "epoch": 2.535019455252918 }, { "current_steps": 1303, "loss": 2.0791, "learning_rate": 5.149797570850202e-07, "epoch": 2.536964980544747 }, { "current_steps": 1304, "loss": 2.364, "learning_rate": 5.145748987854251e-07, "epoch": 2.5389105058365757 }, { "current_steps": 1305, "loss": 2.2329, "learning_rate": 5.141700404858299e-07, "epoch": 2.5408560311284045 }, { "current_steps": 1306, "loss": 1.8507, "learning_rate": 5.137651821862348e-07, "epoch": 2.5428015564202333 }, { "current_steps": 1307, "loss": 2.3448, "learning_rate": 5.133603238866397e-07, "epoch": 2.544747081712062 }, { "current_steps": 1308, "loss": 1.6867, "learning_rate": 5.129554655870445e-07, "epoch": 2.546692607003891 }, { "current_steps": 1309, "loss": 2.0414, "learning_rate": 5.125506072874494e-07, "epoch": 2.5486381322957197 }, { "current_steps": 1310, "loss": 1.6716, "learning_rate": 5.121457489878543e-07, "epoch": 2.5505836575875485 }, { "current_steps": 1311, "loss": 1.5192, "learning_rate": 5.117408906882591e-07, "epoch": 2.5525291828793772 }, { "current_steps": 1312, "loss": 2.279, "learning_rate": 5.11336032388664e-07, "epoch": 2.554474708171206 }, { "current_steps": 1313, "loss": 2.1601, "learning_rate": 5.109311740890689e-07, "epoch": 2.556420233463035 }, { "current_steps": 1314, "loss": 2.1662, "learning_rate": 5.105263157894736e-07, "epoch": 2.5583657587548636 }, { "current_steps": 1315, "loss": 2.2426, "learning_rate": 5.101214574898785e-07, "epoch": 2.5603112840466924 }, { "current_steps": 1316, "loss": 1.7935, "learning_rate": 5.097165991902834e-07, "epoch": 2.562256809338521 }, { "current_steps": 1317, "loss": 2.2986, "learning_rate": 5.093117408906882e-07, "epoch": 2.56420233463035 }, { "current_steps": 1318, "loss": 2.0752, "learning_rate": 5.08906882591093e-07, "epoch": 2.5661478599221788 }, { "current_steps": 1319, "loss": 1.7757, "learning_rate": 5.085020242914979e-07, "epoch": 2.5680933852140075 }, { "current_steps": 1320, "loss": 1.8338, "learning_rate": 5.080971659919027e-07, "epoch": 2.5700389105058363 }, { "current_steps": 1321, "loss": 2.1749, "learning_rate": 5.076923076923076e-07, "epoch": 2.571984435797665 }, { "current_steps": 1322, "loss": 2.139, "learning_rate": 5.072874493927125e-07, "epoch": 2.573929961089494 }, { "current_steps": 1323, "loss": 2.1869, "learning_rate": 5.068825910931173e-07, "epoch": 2.5758754863813227 }, { "current_steps": 1324, "loss": 2.4801, "learning_rate": 5.064777327935222e-07, "epoch": 2.5778210116731515 }, { "current_steps": 1325, "loss": 2.1382, "learning_rate": 5.060728744939271e-07, "epoch": 2.5797665369649807 }, { "current_steps": 1326, "loss": 2.0563, "learning_rate": 5.05668016194332e-07, "epoch": 2.5817120622568095 }, { "current_steps": 1327, "loss": 2.2115, "learning_rate": 5.052631578947368e-07, "epoch": 2.5836575875486383 }, { "current_steps": 1328, "loss": 1.7426, "learning_rate": 5.048582995951417e-07, "epoch": 2.585603112840467 }, { "current_steps": 1329, "loss": 2.2641, "learning_rate": 5.044534412955466e-07, "epoch": 2.587548638132296 }, { "current_steps": 1330, "loss": 2.3158, "learning_rate": 5.040485829959514e-07, "epoch": 2.5894941634241246 }, { "current_steps": 1331, "loss": 2.0381, "learning_rate": 5.036437246963563e-07, "epoch": 2.5914396887159534 }, { "current_steps": 1332, "loss": 2.0794, "learning_rate": 5.032388663967612e-07, "epoch": 2.593385214007782 }, { "current_steps": 1333, "loss": 2.2221, "learning_rate": 5.028340080971659e-07, "epoch": 2.595330739299611 }, { "current_steps": 1334, "loss": 2.1406, "learning_rate": 5.024291497975708e-07, "epoch": 2.59727626459144 }, { "current_steps": 1335, "loss": 2.336, "learning_rate": 5.020242914979757e-07, "epoch": 2.5992217898832686 }, { "current_steps": 1336, "loss": 2.0554, "learning_rate": 5.016194331983805e-07, "epoch": 2.6011673151750974 }, { "current_steps": 1337, "loss": 2.3704, "learning_rate": 5.012145748987854e-07, "epoch": 2.603112840466926 }, { "current_steps": 1338, "loss": 2.2337, "learning_rate": 5.008097165991903e-07, "epoch": 2.605058365758755 }, { "current_steps": 1339, "loss": 1.8702, "learning_rate": 5.004048582995951e-07, "epoch": 2.6070038910505837 }, { "current_steps": 1340, "loss": 2.2808, "learning_rate": 5e-07, "epoch": 2.6089494163424125 }, { "current_steps": 1341, "loss": 2.2097, "learning_rate": 4.995951417004049e-07, "epoch": 2.6108949416342413 }, { "current_steps": 1342, "loss": 2.0908, "learning_rate": 4.991902834008097e-07, "epoch": 2.61284046692607 }, { "current_steps": 1343, "loss": 2.1161, "learning_rate": 4.987854251012146e-07, "epoch": 2.614785992217899 }, { "current_steps": 1344, "loss": 2.5197, "learning_rate": 4.983805668016194e-07, "epoch": 2.6167315175097277 }, { "current_steps": 1345, "loss": 2.2796, "learning_rate": 4.979757085020243e-07, "epoch": 2.6186770428015564 }, { "current_steps": 1346, "loss": 1.7859, "learning_rate": 4.975708502024292e-07, "epoch": 2.6206225680933852 }, { "current_steps": 1347, "loss": 2.4135, "learning_rate": 4.97165991902834e-07, "epoch": 2.622568093385214 }, { "current_steps": 1348, "loss": 2.1001, "learning_rate": 4.967611336032389e-07, "epoch": 2.624513618677043 }, { "current_steps": 1349, "loss": 2.4889, "learning_rate": 4.963562753036438e-07, "epoch": 2.6264591439688716 }, { "current_steps": 1350, "loss": 2.1469, "learning_rate": 4.959514170040485e-07, "epoch": 2.6284046692607004 }, { "current_steps": 1351, "loss": 1.9349, "learning_rate": 4.955465587044534e-07, "epoch": 2.630350194552529 }, { "current_steps": 1352, "loss": 1.7257, "learning_rate": 4.951417004048582e-07, "epoch": 2.632295719844358 }, { "current_steps": 1353, "loss": 2.1899, "learning_rate": 4.947368421052631e-07, "epoch": 2.6342412451361867 }, { "current_steps": 1354, "loss": 2.3953, "learning_rate": 4.943319838056679e-07, "epoch": 2.6361867704280155 }, { "current_steps": 1355, "loss": 1.5225, "learning_rate": 4.939271255060728e-07, "epoch": 2.6381322957198443 }, { "current_steps": 1356, "loss": 2.2236, "learning_rate": 4.935222672064777e-07, "epoch": 2.640077821011673 }, { "current_steps": 1357, "loss": 1.7383, "learning_rate": 4.931174089068826e-07, "epoch": 2.642023346303502 }, { "current_steps": 1358, "loss": 2.3173, "learning_rate": 4.927125506072874e-07, "epoch": 2.6439688715953307 }, { "current_steps": 1359, "loss": 1.6171, "learning_rate": 4.923076923076923e-07, "epoch": 2.6459143968871595 }, { "current_steps": 1360, "loss": 2.0492, "learning_rate": 4.919028340080971e-07, "epoch": 2.6478599221789882 }, { "current_steps": 1361, "loss": 1.9713, "learning_rate": 4.91497975708502e-07, "epoch": 2.649805447470817 }, { "current_steps": 1362, "loss": 2.3032, "learning_rate": 4.910931174089068e-07, "epoch": 2.651750972762646 }, { "current_steps": 1363, "loss": 2.1426, "learning_rate": 4.906882591093117e-07, "epoch": 2.6536964980544746 }, { "current_steps": 1364, "loss": 2.4758, "learning_rate": 4.902834008097166e-07, "epoch": 2.6556420233463034 }, { "current_steps": 1365, "loss": 2.3815, "learning_rate": 4.898785425101214e-07, "epoch": 2.657587548638132 }, { "current_steps": 1366, "loss": 2.0462, "learning_rate": 4.894736842105263e-07, "epoch": 2.659533073929961 }, { "current_steps": 1367, "loss": 2.4181, "learning_rate": 4.890688259109311e-07, "epoch": 2.6614785992217898 }, { "current_steps": 1368, "loss": 1.9326, "learning_rate": 4.88663967611336e-07, "epoch": 2.6634241245136185 }, { "current_steps": 1369, "loss": 1.7266, "learning_rate": 4.882591093117409e-07, "epoch": 2.6653696498054473 }, { "current_steps": 1370, "loss": 2.2718, "learning_rate": 4.878542510121457e-07, "epoch": 2.667315175097276 }, { "current_steps": 1371, "loss": 2.1894, "learning_rate": 4.874493927125506e-07, "epoch": 2.669260700389105 }, { "current_steps": 1372, "loss": 1.6369, "learning_rate": 4.870445344129555e-07, "epoch": 2.6712062256809337 }, { "current_steps": 1373, "loss": 1.8155, "learning_rate": 4.866396761133603e-07, "epoch": 2.673151750972763 }, { "current_steps": 1374, "loss": 1.9073, "learning_rate": 4.862348178137652e-07, "epoch": 2.6750972762645917 }, { "current_steps": 1375, "loss": 2.2366, "learning_rate": 4.8582995951417e-07, "epoch": 2.6770428015564205 }, { "current_steps": 1376, "loss": 1.7776, "learning_rate": 4.854251012145749e-07, "epoch": 2.6789883268482493 }, { "current_steps": 1377, "loss": 1.3905, "learning_rate": 4.850202429149798e-07, "epoch": 2.680933852140078 }, { "current_steps": 1378, "loss": 2.0245, "learning_rate": 4.846153846153846e-07, "epoch": 2.682879377431907 }, { "current_steps": 1379, "loss": 2.0779, "learning_rate": 4.842105263157895e-07, "epoch": 2.6848249027237356 }, { "current_steps": 1380, "loss": 2.0148, "learning_rate": 4.838056680161944e-07, "epoch": 2.6867704280155644 }, { "current_steps": 1381, "loss": 2.3624, "learning_rate": 4.834008097165992e-07, "epoch": 2.688715953307393 }, { "current_steps": 1382, "loss": 2.1182, "learning_rate": 4.82995951417004e-07, "epoch": 2.690661478599222 }, { "current_steps": 1383, "loss": 2.4772, "learning_rate": 4.825910931174089e-07, "epoch": 2.692607003891051 }, { "current_steps": 1384, "loss": 1.7722, "learning_rate": 4.821862348178138e-07, "epoch": 2.6945525291828796 }, { "current_steps": 1385, "loss": 2.3741, "learning_rate": 4.817813765182185e-07, "epoch": 2.6964980544747084 }, { "current_steps": 1386, "loss": 2.1349, "learning_rate": 4.813765182186234e-07, "epoch": 2.698443579766537 }, { "current_steps": 1387, "loss": 1.8868, "learning_rate": 4.809716599190283e-07, "epoch": 2.700389105058366 }, { "current_steps": 1388, "loss": 2.2089, "learning_rate": 4.805668016194331e-07, "epoch": 2.7023346303501947 }, { "current_steps": 1389, "loss": 1.8125, "learning_rate": 4.80161943319838e-07, "epoch": 2.7042801556420235 }, { "current_steps": 1390, "loss": 1.8044, "learning_rate": 4.797570850202428e-07, "epoch": 2.7062256809338523 }, { "current_steps": 1391, "loss": 2.1764, "learning_rate": 4.793522267206477e-07, "epoch": 2.708171206225681 }, { "current_steps": 1392, "loss": 2.1193, "learning_rate": 4.789473684210526e-07, "epoch": 2.71011673151751 }, { "current_steps": 1393, "loss": 2.1413, "learning_rate": 4.785425101214574e-07, "epoch": 2.7120622568093387 }, { "current_steps": 1394, "loss": 2.2653, "learning_rate": 4.781376518218623e-07, "epoch": 2.7140077821011674 }, { "current_steps": 1395, "loss": 2.2167, "learning_rate": 4.777327935222672e-07, "epoch": 2.7159533073929962 }, { "current_steps": 1396, "loss": 2.2137, "learning_rate": 4.77327935222672e-07, "epoch": 2.717898832684825 }, { "current_steps": 1397, "loss": 2.2491, "learning_rate": 4.769230769230769e-07, "epoch": 2.719844357976654 }, { "current_steps": 1398, "loss": 2.0452, "learning_rate": 4.7651821862348176e-07, "epoch": 2.7217898832684826 }, { "current_steps": 1399, "loss": 2.2373, "learning_rate": 4.761133603238866e-07, "epoch": 2.7237354085603114 }, { "current_steps": 1400, "loss": 1.7664, "learning_rate": 4.7570850202429146e-07, "epoch": 2.72568093385214 }, { "current_steps": 1401, "loss": 2.0009, "learning_rate": 4.7530364372469636e-07, "epoch": 2.727626459143969 }, { "current_steps": 1402, "loss": 1.9341, "learning_rate": 4.748987854251012e-07, "epoch": 2.7295719844357977 }, { "current_steps": 1403, "loss": 2.0996, "learning_rate": 4.7449392712550605e-07, "epoch": 2.7315175097276265 }, { "current_steps": 1404, "loss": 2.261, "learning_rate": 4.7408906882591095e-07, "epoch": 2.7334630350194553 }, { "current_steps": 1405, "loss": 2.0365, "learning_rate": 4.7368421052631574e-07, "epoch": 2.735408560311284 }, { "current_steps": 1406, "loss": 1.9149, "learning_rate": 4.732793522267206e-07, "epoch": 2.737354085603113 }, { "current_steps": 1407, "loss": 2.1135, "learning_rate": 4.728744939271255e-07, "epoch": 2.7392996108949417 }, { "current_steps": 1408, "loss": 2.3776, "learning_rate": 4.7246963562753034e-07, "epoch": 2.7412451361867705 }, { "current_steps": 1409, "loss": 2.1309, "learning_rate": 4.720647773279352e-07, "epoch": 2.7431906614785992 }, { "current_steps": 1410, "loss": 2.3206, "learning_rate": 4.716599190283401e-07, "epoch": 2.745136186770428 }, { "current_steps": 1411, "loss": 2.1253, "learning_rate": 4.7125506072874493e-07, "epoch": 2.747081712062257 }, { "current_steps": 1412, "loss": 2.011, "learning_rate": 4.708502024291498e-07, "epoch": 2.7490272373540856 }, { "current_steps": 1413, "loss": 2.3409, "learning_rate": 4.704453441295546e-07, "epoch": 2.7509727626459144 }, { "current_steps": 1414, "loss": 2.267, "learning_rate": 4.7004048582995947e-07, "epoch": 2.752918287937743 }, { "current_steps": 1415, "loss": 2.1173, "learning_rate": 4.6963562753036437e-07, "epoch": 2.754863813229572 }, { "current_steps": 1416, "loss": 2.03, "learning_rate": 4.692307692307692e-07, "epoch": 2.7568093385214008 }, { "current_steps": 1417, "loss": 2.1614, "learning_rate": 4.6882591093117406e-07, "epoch": 2.7587548638132295 }, { "current_steps": 1418, "loss": 2.0364, "learning_rate": 4.6842105263157896e-07, "epoch": 2.7607003891050583 }, { "current_steps": 1419, "loss": 2.0013, "learning_rate": 4.680161943319838e-07, "epoch": 2.762645914396887 }, { "current_steps": 1420, "loss": 2.1427, "learning_rate": 4.6761133603238865e-07, "epoch": 2.764591439688716 }, { "current_steps": 1421, "loss": 2.1357, "learning_rate": 4.672064777327935e-07, "epoch": 2.7665369649805447 }, { "current_steps": 1422, "loss": 2.1179, "learning_rate": 4.6680161943319835e-07, "epoch": 2.7684824902723735 }, { "current_steps": 1423, "loss": 2.3089, "learning_rate": 4.663967611336032e-07, "epoch": 2.7704280155642023 }, { "current_steps": 1424, "loss": 1.6561, "learning_rate": 4.659919028340081e-07, "epoch": 2.772373540856031 }, { "current_steps": 1425, "loss": 1.9276, "learning_rate": 4.6558704453441294e-07, "epoch": 2.77431906614786 }, { "current_steps": 1426, "loss": 1.9467, "learning_rate": 4.651821862348178e-07, "epoch": 2.7762645914396886 }, { "current_steps": 1427, "loss": 2.291, "learning_rate": 4.647773279352227e-07, "epoch": 2.7782101167315174 }, { "current_steps": 1428, "loss": 2.1091, "learning_rate": 4.643724696356275e-07, "epoch": 2.780155642023346 }, { "current_steps": 1429, "loss": 2.2434, "learning_rate": 4.639676113360323e-07, "epoch": 2.782101167315175 }, { "current_steps": 1430, "loss": 2.1248, "learning_rate": 4.635627530364372e-07, "epoch": 2.7840466926070038 }, { "current_steps": 1431, "loss": 1.9181, "learning_rate": 4.6315789473684207e-07, "epoch": 2.7859922178988326 }, { "current_steps": 1432, "loss": 2.0458, "learning_rate": 4.6275303643724697e-07, "epoch": 2.7879377431906613 }, { "current_steps": 1433, "loss": 1.5493, "learning_rate": 4.623481781376518e-07, "epoch": 2.78988326848249 }, { "current_steps": 1434, "loss": 1.888, "learning_rate": 4.6194331983805666e-07, "epoch": 2.791828793774319 }, { "current_steps": 1435, "loss": 2.0262, "learning_rate": 4.6153846153846156e-07, "epoch": 2.7937743190661477 }, { "current_steps": 1436, "loss": 1.9428, "learning_rate": 4.6113360323886636e-07, "epoch": 2.7957198443579765 }, { "current_steps": 1437, "loss": 2.1436, "learning_rate": 4.607287449392712e-07, "epoch": 2.7976653696498053 }, { "current_steps": 1438, "loss": 2.2862, "learning_rate": 4.603238866396761e-07, "epoch": 2.799610894941634 }, { "current_steps": 1439, "loss": 2.0991, "learning_rate": 4.5991902834008095e-07, "epoch": 2.801556420233463 }, { "current_steps": 1440, "loss": 2.1481, "learning_rate": 4.595141700404858e-07, "epoch": 2.8035019455252916 }, { "current_steps": 1441, "loss": 1.8829, "learning_rate": 4.591093117408907e-07, "epoch": 2.8054474708171204 }, { "current_steps": 1442, "loss": 1.9018, "learning_rate": 4.5870445344129554e-07, "epoch": 2.807392996108949 }, { "current_steps": 1443, "loss": 1.7361, "learning_rate": 4.5829959514170033e-07, "epoch": 2.809338521400778 }, { "current_steps": 1444, "loss": 2.1623, "learning_rate": 4.5789473684210523e-07, "epoch": 2.811284046692607 }, { "current_steps": 1445, "loss": 2.3796, "learning_rate": 4.574898785425101e-07, "epoch": 2.8132295719844356 }, { "current_steps": 1446, "loss": 1.9783, "learning_rate": 4.57085020242915e-07, "epoch": 2.8151750972762644 }, { "current_steps": 1447, "loss": 1.5225, "learning_rate": 4.5668016194331983e-07, "epoch": 2.817120622568093 }, { "current_steps": 1448, "loss": 2.0546, "learning_rate": 4.5627530364372467e-07, "epoch": 2.819066147859922 }, { "current_steps": 1449, "loss": 1.7635, "learning_rate": 4.5587044534412957e-07, "epoch": 2.8210116731517507 }, { "current_steps": 1450, "loss": 2.374, "learning_rate": 4.554655870445344e-07, "epoch": 2.8229571984435795 }, { "current_steps": 1451, "loss": 2.1584, "learning_rate": 4.550607287449392e-07, "epoch": 2.8249027237354083 }, { "current_steps": 1452, "loss": 2.035, "learning_rate": 4.546558704453441e-07, "epoch": 2.826848249027237 }, { "current_steps": 1453, "loss": 2.1067, "learning_rate": 4.5425101214574896e-07, "epoch": 2.8287937743190663 }, { "current_steps": 1454, "loss": 2.2225, "learning_rate": 4.538461538461538e-07, "epoch": 2.830739299610895 }, { "current_steps": 1455, "loss": 1.951, "learning_rate": 4.534412955465587e-07, "epoch": 2.832684824902724 }, { "current_steps": 1456, "loss": 1.8527, "learning_rate": 4.5303643724696355e-07, "epoch": 2.8346303501945527 }, { "current_steps": 1457, "loss": 2.3361, "learning_rate": 4.526315789473684e-07, "epoch": 2.8365758754863815 }, { "current_steps": 1458, "loss": 2.1578, "learning_rate": 4.522267206477733e-07, "epoch": 2.8385214007782102 }, { "current_steps": 1459, "loss": 2.299, "learning_rate": 4.518218623481781e-07, "epoch": 2.840466926070039 }, { "current_steps": 1460, "loss": 1.7795, "learning_rate": 4.5141700404858294e-07, "epoch": 2.842412451361868 }, { "current_steps": 1461, "loss": 1.6571, "learning_rate": 4.5101214574898784e-07, "epoch": 2.8443579766536966 }, { "current_steps": 1462, "loss": 2.4182, "learning_rate": 4.506072874493927e-07, "epoch": 2.8463035019455254 }, { "current_steps": 1463, "loss": 2.2352, "learning_rate": 4.502024291497976e-07, "epoch": 2.848249027237354 }, { "current_steps": 1464, "loss": 2.2146, "learning_rate": 4.4979757085020243e-07, "epoch": 2.850194552529183 }, { "current_steps": 1465, "loss": 2.1497, "learning_rate": 4.493927125506073e-07, "epoch": 2.8521400778210118 }, { "current_steps": 1466, "loss": 2.1939, "learning_rate": 4.489878542510121e-07, "epoch": 2.8540856031128405 }, { "current_steps": 1467, "loss": 2.4697, "learning_rate": 4.4858299595141697e-07, "epoch": 2.8560311284046693 }, { "current_steps": 1468, "loss": 2.1986, "learning_rate": 4.481781376518218e-07, "epoch": 2.857976653696498 }, { "current_steps": 1469, "loss": 2.1687, "learning_rate": 4.477732793522267e-07, "epoch": 2.859922178988327 }, { "current_steps": 1470, "loss": 2.0708, "learning_rate": 4.4736842105263156e-07, "epoch": 2.8618677042801557 }, { "current_steps": 1471, "loss": 2.0674, "learning_rate": 4.469635627530364e-07, "epoch": 2.8638132295719845 }, { "current_steps": 1472, "loss": 2.2015, "learning_rate": 4.465587044534413e-07, "epoch": 2.8657587548638133 }, { "current_steps": 1473, "loss": 2.107, "learning_rate": 4.4615384615384615e-07, "epoch": 2.867704280155642 }, { "current_steps": 1474, "loss": 2.0808, "learning_rate": 4.4574898785425095e-07, "epoch": 2.869649805447471 }, { "current_steps": 1475, "loss": 1.6606, "learning_rate": 4.4534412955465585e-07, "epoch": 2.8715953307392996 }, { "current_steps": 1476, "loss": 1.7873, "learning_rate": 4.449392712550607e-07, "epoch": 2.8735408560311284 }, { "current_steps": 1477, "loss": 2.5233, "learning_rate": 4.4453441295546554e-07, "epoch": 2.875486381322957 }, { "current_steps": 1478, "loss": 2.1614, "learning_rate": 4.4412955465587044e-07, "epoch": 2.877431906614786 }, { "current_steps": 1479, "loss": 2.2667, "learning_rate": 4.437246963562753e-07, "epoch": 2.8793774319066148 }, { "current_steps": 1480, "loss": 2.0556, "learning_rate": 4.433198380566802e-07, "epoch": 2.8813229571984436 }, { "current_steps": 1481, "loss": 1.9755, "learning_rate": 4.4291497975708503e-07, "epoch": 2.8832684824902723 }, { "current_steps": 1482, "loss": 1.9783, "learning_rate": 4.425101214574898e-07, "epoch": 2.885214007782101 }, { "current_steps": 1483, "loss": 2.1617, "learning_rate": 4.421052631578947e-07, "epoch": 2.88715953307393 }, { "current_steps": 1484, "loss": 2.1555, "learning_rate": 4.4170040485829957e-07, "epoch": 2.8891050583657587 }, { "current_steps": 1485, "loss": 2.2298, "learning_rate": 4.412955465587044e-07, "epoch": 2.8910505836575875 }, { "current_steps": 1486, "loss": 1.904, "learning_rate": 4.408906882591093e-07, "epoch": 2.8929961089494163 }, { "current_steps": 1487, "loss": 2.1657, "learning_rate": 4.4048582995951416e-07, "epoch": 2.894941634241245 }, { "current_steps": 1488, "loss": 2.1436, "learning_rate": 4.40080971659919e-07, "epoch": 2.896887159533074 }, { "current_steps": 1489, "loss": 2.3766, "learning_rate": 4.3967611336032386e-07, "epoch": 2.8988326848249026 }, { "current_steps": 1490, "loss": 2.3259, "learning_rate": 4.392712550607287e-07, "epoch": 2.9007782101167314 }, { "current_steps": 1491, "loss": 2.2441, "learning_rate": 4.3886639676113355e-07, "epoch": 2.90272373540856 }, { "current_steps": 1492, "loss": 2.0067, "learning_rate": 4.3846153846153845e-07, "epoch": 2.904669260700389 }, { "current_steps": 1493, "loss": 1.8808, "learning_rate": 4.380566801619433e-07, "epoch": 2.906614785992218 }, { "current_steps": 1494, "loss": 1.7123, "learning_rate": 4.376518218623482e-07, "epoch": 2.9085603112840466 }, { "current_steps": 1495, "loss": 2.2826, "learning_rate": 4.3724696356275304e-07, "epoch": 2.9105058365758754 }, { "current_steps": 1496, "loss": 1.9753, "learning_rate": 4.368421052631579e-07, "epoch": 2.912451361867704 }, { "current_steps": 1497, "loss": 2.102, "learning_rate": 4.3643724696356274e-07, "epoch": 2.914396887159533 }, { "current_steps": 1498, "loss": 1.9221, "learning_rate": 4.360323886639676e-07, "epoch": 2.9163424124513617 }, { "current_steps": 1499, "loss": 2.1505, "learning_rate": 4.3562753036437243e-07, "epoch": 2.9182879377431905 }, { "current_steps": 1500, "loss": 1.9216, "learning_rate": 4.3522267206477733e-07, "epoch": 2.9202334630350193 }, { "current_steps": 1501, "loss": 1.8766, "learning_rate": 4.348178137651822e-07, "epoch": 2.9221789883268485 }, { "current_steps": 1502, "loss": 2.1231, "learning_rate": 4.34412955465587e-07, "epoch": 2.9241245136186773 }, { "current_steps": 1503, "loss": 1.949, "learning_rate": 4.340080971659919e-07, "epoch": 2.926070038910506 }, { "current_steps": 1504, "loss": 2.1932, "learning_rate": 4.336032388663967e-07, "epoch": 2.928015564202335 }, { "current_steps": 1505, "loss": 1.9993, "learning_rate": 4.3319838056680156e-07, "epoch": 2.9299610894941637 }, { "current_steps": 1506, "loss": 2.226, "learning_rate": 4.3279352226720646e-07, "epoch": 2.9319066147859925 }, { "current_steps": 1507, "loss": 2.0602, "learning_rate": 4.323886639676113e-07, "epoch": 2.9338521400778212 }, { "current_steps": 1508, "loss": 2.2975, "learning_rate": 4.3198380566801615e-07, "epoch": 2.93579766536965 }, { "current_steps": 1509, "loss": 2.427, "learning_rate": 4.3157894736842105e-07, "epoch": 2.937743190661479 }, { "current_steps": 1510, "loss": 2.0163, "learning_rate": 4.311740890688259e-07, "epoch": 2.9396887159533076 }, { "current_steps": 1511, "loss": 2.0355, "learning_rate": 4.307692307692308e-07, "epoch": 2.9416342412451364 }, { "current_steps": 1512, "loss": 2.3814, "learning_rate": 4.303643724696356e-07, "epoch": 2.943579766536965 }, { "current_steps": 1513, "loss": 2.0255, "learning_rate": 4.2995951417004044e-07, "epoch": 2.945525291828794 }, { "current_steps": 1514, "loss": 2.0214, "learning_rate": 4.2955465587044534e-07, "epoch": 2.9474708171206228 }, { "current_steps": 1515, "loss": 2.2085, "learning_rate": 4.291497975708502e-07, "epoch": 2.9494163424124515 }, { "current_steps": 1516, "loss": 2.0253, "learning_rate": 4.2874493927125503e-07, "epoch": 2.9513618677042803 }, { "current_steps": 1517, "loss": 1.6386, "learning_rate": 4.2834008097165993e-07, "epoch": 2.953307392996109 }, { "current_steps": 1518, "loss": 2.0867, "learning_rate": 4.279352226720648e-07, "epoch": 2.955252918287938 }, { "current_steps": 1519, "loss": 2.2051, "learning_rate": 4.275303643724696e-07, "epoch": 2.9571984435797667 }, { "current_steps": 1520, "loss": 2.243, "learning_rate": 4.2712550607287447e-07, "epoch": 2.9591439688715955 }, { "current_steps": 1521, "loss": 2.1193, "learning_rate": 4.267206477732793e-07, "epoch": 2.9610894941634243 }, { "current_steps": 1522, "loss": 1.6687, "learning_rate": 4.2631578947368416e-07, "epoch": 2.963035019455253 }, { "current_steps": 1523, "loss": 1.8363, "learning_rate": 4.2591093117408906e-07, "epoch": 2.964980544747082 }, { "current_steps": 1524, "loss": 2.1775, "learning_rate": 4.255060728744939e-07, "epoch": 2.9669260700389106 }, { "current_steps": 1525, "loss": 1.8443, "learning_rate": 4.2510121457489876e-07, "epoch": 2.9688715953307394 }, { "current_steps": 1526, "loss": 2.2936, "learning_rate": 4.2469635627530366e-07, "epoch": 2.970817120622568 }, { "current_steps": 1527, "loss": 2.121, "learning_rate": 4.2429149797570845e-07, "epoch": 2.972762645914397 }, { "current_steps": 1528, "loss": 2.1626, "learning_rate": 4.2388663967611335e-07, "epoch": 2.9747081712062258 }, { "current_steps": 1529, "loss": 2.3196, "learning_rate": 4.234817813765182e-07, "epoch": 2.9766536964980546 }, { "current_steps": 1530, "loss": 2.1216, "learning_rate": 4.2307692307692304e-07, "epoch": 2.9785992217898833 }, { "current_steps": 1531, "loss": 2.4063, "learning_rate": 4.2267206477732794e-07, "epoch": 2.980544747081712 }, { "current_steps": 1532, "loss": 2.2169, "learning_rate": 4.222672064777328e-07, "epoch": 2.982490272373541 }, { "current_steps": 1533, "loss": 2.0368, "learning_rate": 4.2186234817813763e-07, "epoch": 2.9844357976653697 }, { "current_steps": 1534, "loss": 2.1142, "learning_rate": 4.2145748987854253e-07, "epoch": 2.9863813229571985 }, { "current_steps": 1535, "loss": 2.2297, "learning_rate": 4.2105263157894733e-07, "epoch": 2.9883268482490273 }, { "current_steps": 1536, "loss": 2.2349, "learning_rate": 4.206477732793522e-07, "epoch": 2.990272373540856 }, { "current_steps": 1537, "loss": 2.142, "learning_rate": 4.2024291497975707e-07, "epoch": 2.992217898832685 }, { "current_steps": 1538, "loss": 2.3191, "learning_rate": 4.198380566801619e-07, "epoch": 2.9941634241245136 }, { "current_steps": 1539, "loss": 2.0837, "learning_rate": 4.1943319838056677e-07, "epoch": 2.9961089494163424 }, { "current_steps": 1540, "loss": 2.3467, "learning_rate": 4.1902834008097167e-07, "epoch": 2.998054474708171 }, { "current_steps": 1541, "loss": 1.7567, "learning_rate": 4.186234817813765e-07, "epoch": 3.0 }, { "current_steps": 1542, "loss": 2.2319, "learning_rate": 4.182186234817814e-07, "epoch": 3.001945525291829 }, { "current_steps": 1543, "loss": 2.1602, "learning_rate": 4.178137651821862e-07, "epoch": 3.0038910505836576 }, { "current_steps": 1544, "loss": 1.8431, "learning_rate": 4.1740890688259105e-07, "epoch": 3.0058365758754864 }, { "current_steps": 1545, "loss": 1.9224, "learning_rate": 4.1700404858299595e-07, "epoch": 3.007782101167315 }, { "current_steps": 1546, "loss": 2.0696, "learning_rate": 4.165991902834008e-07, "epoch": 3.009727626459144 }, { "current_steps": 1547, "loss": 1.9947, "learning_rate": 4.1619433198380564e-07, "epoch": 3.0116731517509727 }, { "current_steps": 1548, "loss": 1.7802, "learning_rate": 4.1578947368421054e-07, "epoch": 3.0136186770428015 }, { "current_steps": 1549, "loss": 2.0079, "learning_rate": 4.153846153846154e-07, "epoch": 3.0155642023346303 }, { "current_steps": 1550, "loss": 2.4013, "learning_rate": 4.149797570850202e-07, "epoch": 3.017509727626459 }, { "current_steps": 1551, "loss": 2.1907, "learning_rate": 4.145748987854251e-07, "epoch": 3.019455252918288 }, { "current_steps": 1552, "loss": 1.793, "learning_rate": 4.1417004048582993e-07, "epoch": 3.0214007782101167 }, { "current_steps": 1553, "loss": 1.9295, "learning_rate": 4.137651821862348e-07, "epoch": 3.0233463035019454 }, { "current_steps": 1554, "loss": 2.313, "learning_rate": 4.133603238866397e-07, "epoch": 3.0252918287937742 }, { "current_steps": 1555, "loss": 2.0111, "learning_rate": 4.129554655870445e-07, "epoch": 3.027237354085603 }, { "current_steps": 1556, "loss": 1.8635, "learning_rate": 4.1255060728744937e-07, "epoch": 3.029182879377432 }, { "current_steps": 1557, "loss": 2.1171, "learning_rate": 4.1214574898785427e-07, "epoch": 3.0311284046692606 }, { "current_steps": 1558, "loss": 1.8703, "learning_rate": 4.1174089068825906e-07, "epoch": 3.0330739299610894 }, { "current_steps": 1559, "loss": 2.2713, "learning_rate": 4.113360323886639e-07, "epoch": 3.035019455252918 }, { "current_steps": 1560, "loss": 2.1322, "learning_rate": 4.109311740890688e-07, "epoch": 3.036964980544747 }, { "current_steps": 1561, "loss": 2.1725, "learning_rate": 4.1052631578947365e-07, "epoch": 3.0389105058365757 }, { "current_steps": 1562, "loss": 2.3998, "learning_rate": 4.1012145748987855e-07, "epoch": 3.0408560311284045 }, { "current_steps": 1563, "loss": 2.1841, "learning_rate": 4.097165991902834e-07, "epoch": 3.0428015564202333 }, { "current_steps": 1564, "loss": 2.2576, "learning_rate": 4.0931174089068825e-07, "epoch": 3.044747081712062 }, { "current_steps": 1565, "loss": 2.2423, "learning_rate": 4.089068825910931e-07, "epoch": 3.046692607003891 }, { "current_steps": 1566, "loss": 1.7175, "learning_rate": 4.0850202429149794e-07, "epoch": 3.0486381322957197 }, { "current_steps": 1567, "loss": 1.9029, "learning_rate": 4.080971659919028e-07, "epoch": 3.0505836575875485 }, { "current_steps": 1568, "loss": 2.1516, "learning_rate": 4.076923076923077e-07, "epoch": 3.0525291828793772 }, { "current_steps": 1569, "loss": 2.1053, "learning_rate": 4.0728744939271253e-07, "epoch": 3.054474708171206 }, { "current_steps": 1570, "loss": 1.9283, "learning_rate": 4.068825910931174e-07, "epoch": 3.056420233463035 }, { "current_steps": 1571, "loss": 1.9633, "learning_rate": 4.064777327935223e-07, "epoch": 3.0583657587548636 }, { "current_steps": 1572, "loss": 1.9429, "learning_rate": 4.060728744939271e-07, "epoch": 3.0603112840466924 }, { "current_steps": 1573, "loss": 2.1711, "learning_rate": 4.056680161943319e-07, "epoch": 3.062256809338521 }, { "current_steps": 1574, "loss": 2.1348, "learning_rate": 4.052631578947368e-07, "epoch": 3.06420233463035 }, { "current_steps": 1575, "loss": 2.2915, "learning_rate": 4.0485829959514166e-07, "epoch": 3.066147859922179 }, { "current_steps": 1576, "loss": 2.1579, "learning_rate": 4.0445344129554656e-07, "epoch": 3.068093385214008 }, { "current_steps": 1577, "loss": 2.4956, "learning_rate": 4.040485829959514e-07, "epoch": 3.0700389105058368 }, { "current_steps": 1578, "loss": 2.3889, "learning_rate": 4.0364372469635626e-07, "epoch": 3.0719844357976656 }, { "current_steps": 1579, "loss": 1.7386, "learning_rate": 4.0323886639676116e-07, "epoch": 3.0739299610894943 }, { "current_steps": 1580, "loss": 2.1956, "learning_rate": 4.02834008097166e-07, "epoch": 3.075875486381323 }, { "current_steps": 1581, "loss": 2.2158, "learning_rate": 4.024291497975708e-07, "epoch": 3.077821011673152 }, { "current_steps": 1582, "loss": 1.9738, "learning_rate": 4.020242914979757e-07, "epoch": 3.0797665369649807 }, { "current_steps": 1583, "loss": 2.4733, "learning_rate": 4.0161943319838054e-07, "epoch": 3.0817120622568095 }, { "current_steps": 1584, "loss": 1.9448, "learning_rate": 4.012145748987854e-07, "epoch": 3.0836575875486383 }, { "current_steps": 1585, "loss": 2.3981, "learning_rate": 4.008097165991903e-07, "epoch": 3.085603112840467 }, { "current_steps": 1586, "loss": 1.9953, "learning_rate": 4.0040485829959514e-07, "epoch": 3.087548638132296 }, { "current_steps": 1587, "loss": 1.9482, "learning_rate": 4e-07, "epoch": 3.0894941634241246 }, { "current_steps": 1588, "loss": 2.1542, "learning_rate": 3.9959514170040483e-07, "epoch": 3.0914396887159534 }, { "current_steps": 1589, "loss": 1.7892, "learning_rate": 3.991902834008097e-07, "epoch": 3.093385214007782 }, { "current_steps": 1590, "loss": 1.977, "learning_rate": 3.987854251012145e-07, "epoch": 3.095330739299611 }, { "current_steps": 1591, "loss": 1.7704, "learning_rate": 3.983805668016194e-07, "epoch": 3.09727626459144 }, { "current_steps": 1592, "loss": 2.0456, "learning_rate": 3.9797570850202427e-07, "epoch": 3.0992217898832686 }, { "current_steps": 1593, "loss": 1.7214, "learning_rate": 3.9757085020242917e-07, "epoch": 3.1011673151750974 }, { "current_steps": 1594, "loss": 2.0978, "learning_rate": 3.97165991902834e-07, "epoch": 3.103112840466926 }, { "current_steps": 1595, "loss": 2.1972, "learning_rate": 3.9676113360323886e-07, "epoch": 3.105058365758755 }, { "current_steps": 1596, "loss": 1.9078, "learning_rate": 3.963562753036437e-07, "epoch": 3.1070038910505837 }, { "current_steps": 1597, "loss": 2.4299, "learning_rate": 3.9595141700404855e-07, "epoch": 3.1089494163424125 }, { "current_steps": 1598, "loss": 2.2185, "learning_rate": 3.955465587044534e-07, "epoch": 3.1108949416342413 }, { "current_steps": 1599, "loss": 2.3106, "learning_rate": 3.951417004048583e-07, "epoch": 3.11284046692607 }, { "current_steps": 1600, "loss": 2.3387, "learning_rate": 3.9473684210526315e-07, "epoch": 3.114785992217899 }, { "current_steps": 1601, "loss": 2.1111, "learning_rate": 3.94331983805668e-07, "epoch": 3.1167315175097277 }, { "current_steps": 1602, "loss": 1.9732, "learning_rate": 3.939271255060729e-07, "epoch": 3.1186770428015564 }, { "current_steps": 1603, "loss": 2.2984, "learning_rate": 3.9352226720647774e-07, "epoch": 3.1206225680933852 }, { "current_steps": 1604, "loss": 2.2047, "learning_rate": 3.9311740890688253e-07, "epoch": 3.122568093385214 }, { "current_steps": 1605, "loss": 2.201, "learning_rate": 3.9271255060728743e-07, "epoch": 3.124513618677043 }, { "current_steps": 1606, "loss": 2.2841, "learning_rate": 3.923076923076923e-07, "epoch": 3.1264591439688716 }, { "current_steps": 1607, "loss": 2.3292, "learning_rate": 3.919028340080971e-07, "epoch": 3.1284046692607004 }, { "current_steps": 1608, "loss": 2.0112, "learning_rate": 3.91497975708502e-07, "epoch": 3.130350194552529 }, { "current_steps": 1609, "loss": 2.1722, "learning_rate": 3.9109311740890687e-07, "epoch": 3.132295719844358 }, { "current_steps": 1610, "loss": 2.1423, "learning_rate": 3.9068825910931177e-07, "epoch": 3.1342412451361867 }, { "current_steps": 1611, "loss": 2.2379, "learning_rate": 3.9028340080971656e-07, "epoch": 3.1361867704280155 }, { "current_steps": 1612, "loss": 2.239, "learning_rate": 3.898785425101214e-07, "epoch": 3.1381322957198443 }, { "current_steps": 1613, "loss": 1.9535, "learning_rate": 3.894736842105263e-07, "epoch": 3.140077821011673 }, { "current_steps": 1614, "loss": 2.0712, "learning_rate": 3.8906882591093116e-07, "epoch": 3.142023346303502 }, { "current_steps": 1615, "loss": 2.1671, "learning_rate": 3.88663967611336e-07, "epoch": 3.1439688715953307 }, { "current_steps": 1616, "loss": 1.8017, "learning_rate": 3.882591093117409e-07, "epoch": 3.1459143968871595 }, { "current_steps": 1617, "loss": 2.0496, "learning_rate": 3.8785425101214575e-07, "epoch": 3.1478599221789882 }, { "current_steps": 1618, "loss": 2.1625, "learning_rate": 3.874493927125506e-07, "epoch": 3.149805447470817 }, { "current_steps": 1619, "loss": 1.9613, "learning_rate": 3.8704453441295544e-07, "epoch": 3.151750972762646 }, { "current_steps": 1620, "loss": 2.2658, "learning_rate": 3.866396761133603e-07, "epoch": 3.1536964980544746 }, { "current_steps": 1621, "loss": 2.1928, "learning_rate": 3.8623481781376513e-07, "epoch": 3.1556420233463034 }, { "current_steps": 1622, "loss": 2.1299, "learning_rate": 3.8582995951417003e-07, "epoch": 3.157587548638132 }, { "current_steps": 1623, "loss": 2.1428, "learning_rate": 3.854251012145749e-07, "epoch": 3.159533073929961 }, { "current_steps": 1624, "loss": 1.7445, "learning_rate": 3.850202429149798e-07, "epoch": 3.1614785992217898 }, { "current_steps": 1625, "loss": 1.9814, "learning_rate": 3.8461538461538463e-07, "epoch": 3.1634241245136185 }, { "current_steps": 1626, "loss": 2.0744, "learning_rate": 3.842105263157894e-07, "epoch": 3.1653696498054473 }, { "current_steps": 1627, "loss": 2.1054, "learning_rate": 3.838056680161943e-07, "epoch": 3.167315175097276 }, { "current_steps": 1628, "loss": 2.1878, "learning_rate": 3.8340080971659917e-07, "epoch": 3.169260700389105 }, { "current_steps": 1629, "loss": 1.7769, "learning_rate": 3.82995951417004e-07, "epoch": 3.1712062256809337 }, { "current_steps": 1630, "loss": 1.8938, "learning_rate": 3.825910931174089e-07, "epoch": 3.1731517509727625 }, { "current_steps": 1631, "loss": 2.1817, "learning_rate": 3.8218623481781376e-07, "epoch": 3.1750972762645913 }, { "current_steps": 1632, "loss": 2.2215, "learning_rate": 3.817813765182186e-07, "epoch": 3.17704280155642 }, { "current_steps": 1633, "loss": 2.2449, "learning_rate": 3.813765182186235e-07, "epoch": 3.178988326848249 }, { "current_steps": 1634, "loss": 2.2822, "learning_rate": 3.809716599190283e-07, "epoch": 3.1809338521400776 }, { "current_steps": 1635, "loss": 1.9981, "learning_rate": 3.8056680161943314e-07, "epoch": 3.1828793774319064 }, { "current_steps": 1636, "loss": 1.9112, "learning_rate": 3.8016194331983804e-07, "epoch": 3.1848249027237356 }, { "current_steps": 1637, "loss": 2.1939, "learning_rate": 3.797570850202429e-07, "epoch": 3.1867704280155644 }, { "current_steps": 1638, "loss": 2.2013, "learning_rate": 3.7935222672064774e-07, "epoch": 3.188715953307393 }, { "current_steps": 1639, "loss": 2.3349, "learning_rate": 3.7894736842105264e-07, "epoch": 3.190661478599222 }, { "current_steps": 1640, "loss": 2.0548, "learning_rate": 3.785425101214575e-07, "epoch": 3.192607003891051 }, { "current_steps": 1641, "loss": 2.0821, "learning_rate": 3.781376518218624e-07, "epoch": 3.1945525291828796 }, { "current_steps": 1642, "loss": 2.0051, "learning_rate": 3.777327935222672e-07, "epoch": 3.1964980544747084 }, { "current_steps": 1643, "loss": 2.1287, "learning_rate": 3.77327935222672e-07, "epoch": 3.198443579766537 }, { "current_steps": 1644, "loss": 2.1017, "learning_rate": 3.769230769230769e-07, "epoch": 3.200389105058366 }, { "current_steps": 1645, "loss": 2.0769, "learning_rate": 3.7651821862348177e-07, "epoch": 3.2023346303501947 }, { "current_steps": 1646, "loss": 2.174, "learning_rate": 3.761133603238866e-07, "epoch": 3.2042801556420235 }, { "current_steps": 1647, "loss": 1.9028, "learning_rate": 3.757085020242915e-07, "epoch": 3.2062256809338523 }, { "current_steps": 1648, "loss": 2.1866, "learning_rate": 3.7530364372469636e-07, "epoch": 3.208171206225681 }, { "current_steps": 1649, "loss": 2.2154, "learning_rate": 3.7489878542510115e-07, "epoch": 3.21011673151751 }, { "current_steps": 1650, "loss": 2.1467, "learning_rate": 3.7449392712550605e-07, "epoch": 3.2120622568093387 }, { "current_steps": 1651, "loss": 2.2275, "learning_rate": 3.740890688259109e-07, "epoch": 3.2140077821011674 }, { "current_steps": 1652, "loss": 2.2132, "learning_rate": 3.7368421052631575e-07, "epoch": 3.2159533073929962 }, { "current_steps": 1653, "loss": 2.1072, "learning_rate": 3.7327935222672065e-07, "epoch": 3.217898832684825 }, { "current_steps": 1654, "loss": 2.2598, "learning_rate": 3.728744939271255e-07, "epoch": 3.219844357976654 }, { "current_steps": 1655, "loss": 2.0094, "learning_rate": 3.7246963562753034e-07, "epoch": 3.2217898832684826 }, { "current_steps": 1656, "loss": 1.7012, "learning_rate": 3.7206477732793524e-07, "epoch": 3.2237354085603114 }, { "current_steps": 1657, "loss": 2.229, "learning_rate": 3.7165991902834003e-07, "epoch": 3.22568093385214 }, { "current_steps": 1658, "loss": 1.7522, "learning_rate": 3.7125506072874493e-07, "epoch": 3.227626459143969 }, { "current_steps": 1659, "loss": 2.2227, "learning_rate": 3.708502024291498e-07, "epoch": 3.2295719844357977 }, { "current_steps": 1660, "loss": 1.911, "learning_rate": 3.704453441295546e-07, "epoch": 3.2315175097276265 }, { "current_steps": 1661, "loss": 1.6856, "learning_rate": 3.700404858299595e-07, "epoch": 3.2334630350194553 }, { "current_steps": 1662, "loss": 2.24, "learning_rate": 3.6963562753036437e-07, "epoch": 3.235408560311284 }, { "current_steps": 1663, "loss": 1.9047, "learning_rate": 3.692307692307692e-07, "epoch": 3.237354085603113 }, { "current_steps": 1664, "loss": 2.0995, "learning_rate": 3.6882591093117406e-07, "epoch": 3.2392996108949417 }, { "current_steps": 1665, "loss": 2.0818, "learning_rate": 3.684210526315789e-07, "epoch": 3.2412451361867705 }, { "current_steps": 1666, "loss": 2.3419, "learning_rate": 3.6801619433198376e-07, "epoch": 3.2431906614785992 }, { "current_steps": 1667, "loss": 2.0597, "learning_rate": 3.6761133603238866e-07, "epoch": 3.245136186770428 }, { "current_steps": 1668, "loss": 1.6223, "learning_rate": 3.672064777327935e-07, "epoch": 3.247081712062257 }, { "current_steps": 1669, "loss": 2.3295, "learning_rate": 3.6680161943319835e-07, "epoch": 3.2490272373540856 }, { "current_steps": 1670, "loss": 2.1102, "learning_rate": 3.6639676113360325e-07, "epoch": 3.2509727626459144 }, { "current_steps": 1671, "loss": 2.0061, "learning_rate": 3.659919028340081e-07, "epoch": 3.252918287937743 }, { "current_steps": 1672, "loss": 1.9643, "learning_rate": 3.655870445344129e-07, "epoch": 3.254863813229572 }, { "current_steps": 1673, "loss": 2.0348, "learning_rate": 3.651821862348178e-07, "epoch": 3.2568093385214008 }, { "current_steps": 1674, "loss": 2.0626, "learning_rate": 3.6477732793522264e-07, "epoch": 3.2587548638132295 }, { "current_steps": 1675, "loss": 2.0896, "learning_rate": 3.6437246963562754e-07, "epoch": 3.2607003891050583 }, { "current_steps": 1676, "loss": 1.8469, "learning_rate": 3.639676113360324e-07, "epoch": 3.262645914396887 }, { "current_steps": 1677, "loss": 2.0704, "learning_rate": 3.6356275303643723e-07, "epoch": 3.264591439688716 }, { "current_steps": 1678, "loss": 1.537, "learning_rate": 3.6315789473684213e-07, "epoch": 3.2665369649805447 }, { "current_steps": 1679, "loss": 2.0385, "learning_rate": 3.62753036437247e-07, "epoch": 3.2684824902723735 }, { "current_steps": 1680, "loss": 2.0776, "learning_rate": 3.6234817813765177e-07, "epoch": 3.2704280155642023 }, { "current_steps": 1681, "loss": 2.2791, "learning_rate": 3.6194331983805667e-07, "epoch": 3.272373540856031 }, { "current_steps": 1682, "loss": 1.9512, "learning_rate": 3.615384615384615e-07, "epoch": 3.27431906614786 }, { "current_steps": 1683, "loss": 1.906, "learning_rate": 3.6113360323886636e-07, "epoch": 3.2762645914396886 }, { "current_steps": 1684, "loss": 2.0956, "learning_rate": 3.6072874493927126e-07, "epoch": 3.2782101167315174 }, { "current_steps": 1685, "loss": 2.0287, "learning_rate": 3.603238866396761e-07, "epoch": 3.280155642023346 }, { "current_steps": 1686, "loss": 1.9642, "learning_rate": 3.5991902834008095e-07, "epoch": 3.282101167315175 }, { "current_steps": 1687, "loss": 2.0348, "learning_rate": 3.595141700404858e-07, "epoch": 3.2840466926070038 }, { "current_steps": 1688, "loss": 1.6719, "learning_rate": 3.5910931174089065e-07, "epoch": 3.2859922178988326 }, { "current_steps": 1689, "loss": 1.9094, "learning_rate": 3.587044534412955e-07, "epoch": 3.2879377431906613 }, { "current_steps": 1690, "loss": 2.1159, "learning_rate": 3.582995951417004e-07, "epoch": 3.28988326848249 }, { "current_steps": 1691, "loss": 1.938, "learning_rate": 3.5789473684210524e-07, "epoch": 3.291828793774319 }, { "current_steps": 1692, "loss": 2.0171, "learning_rate": 3.5748987854251014e-07, "epoch": 3.2937743190661477 }, { "current_steps": 1693, "loss": 2.1951, "learning_rate": 3.57085020242915e-07, "epoch": 3.2957198443579765 }, { "current_steps": 1694, "loss": 2.2227, "learning_rate": 3.5668016194331983e-07, "epoch": 3.2976653696498053 }, { "current_steps": 1695, "loss": 2.2318, "learning_rate": 3.562753036437247e-07, "epoch": 3.299610894941634 }, { "current_steps": 1696, "loss": 2.543, "learning_rate": 3.558704453441295e-07, "epoch": 3.301556420233463 }, { "current_steps": 1697, "loss": 2.1286, "learning_rate": 3.5546558704453437e-07, "epoch": 3.3035019455252916 }, { "current_steps": 1698, "loss": 2.0553, "learning_rate": 3.5506072874493927e-07, "epoch": 3.3054474708171204 }, { "current_steps": 1699, "loss": 2.0395, "learning_rate": 3.546558704453441e-07, "epoch": 3.307392996108949 }, { "current_steps": 1700, "loss": 2.2424, "learning_rate": 3.5425101214574896e-07, "epoch": 3.309338521400778 }, { "current_steps": 1701, "loss": 2.0627, "learning_rate": 3.5384615384615386e-07, "epoch": 3.311284046692607 }, { "current_steps": 1702, "loss": 1.5976, "learning_rate": 3.534412955465587e-07, "epoch": 3.3132295719844356 }, { "current_steps": 1703, "loss": 2.4913, "learning_rate": 3.530364372469635e-07, "epoch": 3.3151750972762644 }, { "current_steps": 1704, "loss": 2.0588, "learning_rate": 3.526315789473684e-07, "epoch": 3.317120622568093 }, { "current_steps": 1705, "loss": 2.1868, "learning_rate": 3.5222672064777325e-07, "epoch": 3.319066147859922 }, { "current_steps": 1706, "loss": 1.815, "learning_rate": 3.5182186234817815e-07, "epoch": 3.321011673151751 }, { "current_steps": 1707, "loss": 2.1276, "learning_rate": 3.51417004048583e-07, "epoch": 3.32295719844358 }, { "current_steps": 1708, "loss": 1.5913, "learning_rate": 3.5101214574898784e-07, "epoch": 3.3249027237354087 }, { "current_steps": 1709, "loss": 2.2019, "learning_rate": 3.5060728744939274e-07, "epoch": 3.3268482490272375 }, { "current_steps": 1710, "loss": 1.7611, "learning_rate": 3.5020242914979753e-07, "epoch": 3.3287937743190663 }, { "current_steps": 1711, "loss": 2.0623, "learning_rate": 3.497975708502024e-07, "epoch": 3.330739299610895 }, { "current_steps": 1712, "loss": 2.3241, "learning_rate": 3.493927125506073e-07, "epoch": 3.332684824902724 }, { "current_steps": 1713, "loss": 2.2417, "learning_rate": 3.4898785425101213e-07, "epoch": 3.3346303501945527 }, { "current_steps": 1714, "loss": 2.0361, "learning_rate": 3.4858299595141697e-07, "epoch": 3.3365758754863815 }, { "current_steps": 1715, "loss": 1.9331, "learning_rate": 3.4817813765182187e-07, "epoch": 3.3385214007782102 }, { "current_steps": 1716, "loss": 1.9732, "learning_rate": 3.477732793522267e-07, "epoch": 3.340466926070039 }, { "current_steps": 1717, "loss": 2.0524, "learning_rate": 3.4736842105263157e-07, "epoch": 3.342412451361868 }, { "current_steps": 1718, "loss": 2.4532, "learning_rate": 3.469635627530364e-07, "epoch": 3.3443579766536966 }, { "current_steps": 1719, "loss": 2.1033, "learning_rate": 3.4655870445344126e-07, "epoch": 3.3463035019455254 }, { "current_steps": 1720, "loss": 1.7016, "learning_rate": 3.461538461538461e-07, "epoch": 3.348249027237354 }, { "current_steps": 1721, "loss": 2.1314, "learning_rate": 3.45748987854251e-07, "epoch": 3.350194552529183 }, { "current_steps": 1722, "loss": 1.5296, "learning_rate": 3.4534412955465585e-07, "epoch": 3.3521400778210118 }, { "current_steps": 1723, "loss": 1.819, "learning_rate": 3.4493927125506075e-07, "epoch": 3.3540856031128405 }, { "current_steps": 1724, "loss": 2.003, "learning_rate": 3.445344129554656e-07, "epoch": 3.3560311284046693 }, { "current_steps": 1725, "loss": 1.8787, "learning_rate": 3.441295546558704e-07, "epoch": 3.357976653696498 }, { "current_steps": 1726, "loss": 2.0925, "learning_rate": 3.437246963562753e-07, "epoch": 3.359922178988327 }, { "current_steps": 1727, "loss": 1.6908, "learning_rate": 3.4331983805668014e-07, "epoch": 3.3618677042801557 }, { "current_steps": 1728, "loss": 1.7465, "learning_rate": 3.42914979757085e-07, "epoch": 3.3638132295719845 }, { "current_steps": 1729, "loss": 2.0273, "learning_rate": 3.425101214574899e-07, "epoch": 3.3657587548638133 }, { "current_steps": 1730, "loss": 2.2703, "learning_rate": 3.4210526315789473e-07, "epoch": 3.367704280155642 }, { "current_steps": 1731, "loss": 2.2659, "learning_rate": 3.417004048582996e-07, "epoch": 3.369649805447471 }, { "current_steps": 1732, "loss": 2.1173, "learning_rate": 3.412955465587045e-07, "epoch": 3.3715953307392996 }, { "current_steps": 1733, "loss": 1.9927, "learning_rate": 3.4089068825910927e-07, "epoch": 3.3735408560311284 }, { "current_steps": 1734, "loss": 1.9867, "learning_rate": 3.404858299595141e-07, "epoch": 3.375486381322957 }, { "current_steps": 1735, "loss": 2.2229, "learning_rate": 3.40080971659919e-07, "epoch": 3.377431906614786 }, { "current_steps": 1736, "loss": 2.0997, "learning_rate": 3.3967611336032386e-07, "epoch": 3.3793774319066148 }, { "current_steps": 1737, "loss": 2.0035, "learning_rate": 3.392712550607287e-07, "epoch": 3.3813229571984436 }, { "current_steps": 1738, "loss": 2.344, "learning_rate": 3.388663967611336e-07, "epoch": 3.3832684824902723 }, { "current_steps": 1739, "loss": 2.2633, "learning_rate": 3.3846153846153845e-07, "epoch": 3.385214007782101 }, { "current_steps": 1740, "loss": 2.1621, "learning_rate": 3.3805668016194335e-07, "epoch": 3.38715953307393 }, { "current_steps": 1741, "loss": 1.9397, "learning_rate": 3.3765182186234815e-07, "epoch": 3.3891050583657587 }, { "current_steps": 1742, "loss": 2.1708, "learning_rate": 3.37246963562753e-07, "epoch": 3.3910505836575875 }, { "current_steps": 1743, "loss": 1.7386, "learning_rate": 3.368421052631579e-07, "epoch": 3.3929961089494163 }, { "current_steps": 1744, "loss": 1.7601, "learning_rate": 3.3643724696356274e-07, "epoch": 3.394941634241245 }, { "current_steps": 1745, "loss": 2.0625, "learning_rate": 3.360323886639676e-07, "epoch": 3.396887159533074 }, { "current_steps": 1746, "loss": 2.188, "learning_rate": 3.356275303643725e-07, "epoch": 3.3988326848249026 }, { "current_steps": 1747, "loss": 2.3313, "learning_rate": 3.3522267206477733e-07, "epoch": 3.4007782101167314 }, { "current_steps": 1748, "loss": 2.2734, "learning_rate": 3.348178137651821e-07, "epoch": 3.40272373540856 }, { "current_steps": 1749, "loss": 2.2079, "learning_rate": 3.34412955465587e-07, "epoch": 3.404669260700389 }, { "current_steps": 1750, "loss": 2.0182, "learning_rate": 3.3400809716599187e-07, "epoch": 3.406614785992218 }, { "current_steps": 1751, "loss": 2.0508, "learning_rate": 3.336032388663967e-07, "epoch": 3.4085603112840466 }, { "current_steps": 1752, "loss": 1.9852, "learning_rate": 3.331983805668016e-07, "epoch": 3.4105058365758754 }, { "current_steps": 1753, "loss": 2.0253, "learning_rate": 3.3279352226720646e-07, "epoch": 3.412451361867704 }, { "current_steps": 1754, "loss": 1.7831, "learning_rate": 3.3238866396761136e-07, "epoch": 3.414396887159533 }, { "current_steps": 1755, "loss": 2.259, "learning_rate": 3.319838056680162e-07, "epoch": 3.4163424124513617 }, { "current_steps": 1756, "loss": 2.2524, "learning_rate": 3.31578947368421e-07, "epoch": 3.4182879377431905 }, { "current_steps": 1757, "loss": 2.039, "learning_rate": 3.311740890688259e-07, "epoch": 3.4202334630350193 }, { "current_steps": 1758, "loss": 1.9546, "learning_rate": 3.3076923076923075e-07, "epoch": 3.422178988326848 }, { "current_steps": 1759, "loss": 2.1471, "learning_rate": 3.303643724696356e-07, "epoch": 3.424124513618677 }, { "current_steps": 1760, "loss": 2.1325, "learning_rate": 3.299595141700405e-07, "epoch": 3.4260700389105057 }, { "current_steps": 1761, "loss": 2.1329, "learning_rate": 3.2955465587044534e-07, "epoch": 3.4280155642023344 }, { "current_steps": 1762, "loss": 2.0561, "learning_rate": 3.291497975708502e-07, "epoch": 3.4299610894941637 }, { "current_steps": 1763, "loss": 1.891, "learning_rate": 3.287449392712551e-07, "epoch": 3.4319066147859925 }, { "current_steps": 1764, "loss": 2.2212, "learning_rate": 3.283400809716599e-07, "epoch": 3.4338521400778212 }, { "current_steps": 1765, "loss": 2.6647, "learning_rate": 3.2793522267206473e-07, "epoch": 3.43579766536965 }, { "current_steps": 1766, "loss": 2.1809, "learning_rate": 3.2753036437246963e-07, "epoch": 3.437743190661479 }, { "current_steps": 1767, "loss": 2.267, "learning_rate": 3.271255060728745e-07, "epoch": 3.4396887159533076 }, { "current_steps": 1768, "loss": 2.4944, "learning_rate": 3.267206477732793e-07, "epoch": 3.4416342412451364 }, { "current_steps": 1769, "loss": 2.3302, "learning_rate": 3.263157894736842e-07, "epoch": 3.443579766536965 }, { "current_steps": 1770, "loss": 2.1232, "learning_rate": 3.2591093117408907e-07, "epoch": 3.445525291828794 }, { "current_steps": 1771, "loss": 1.6991, "learning_rate": 3.2550607287449386e-07, "epoch": 3.4474708171206228 }, { "current_steps": 1772, "loss": 1.9896, "learning_rate": 3.2510121457489876e-07, "epoch": 3.4494163424124515 }, { "current_steps": 1773, "loss": 1.9265, "learning_rate": 3.246963562753036e-07, "epoch": 3.4513618677042803 }, { "current_steps": 1774, "loss": 1.9367, "learning_rate": 3.242914979757085e-07, "epoch": 3.453307392996109 }, { "current_steps": 1775, "loss": 1.7504, "learning_rate": 3.2388663967611335e-07, "epoch": 3.455252918287938 }, { "current_steps": 1776, "loss": 2.1079, "learning_rate": 3.234817813765182e-07, "epoch": 3.4571984435797667 }, { "current_steps": 1777, "loss": 2.2187, "learning_rate": 3.230769230769231e-07, "epoch": 3.4591439688715955 }, { "current_steps": 1778, "loss": 2.2214, "learning_rate": 3.2267206477732795e-07, "epoch": 3.4610894941634243 }, { "current_steps": 1779, "loss": 2.0146, "learning_rate": 3.2226720647773274e-07, "epoch": 3.463035019455253 }, { "current_steps": 1780, "loss": 2.2812, "learning_rate": 3.2186234817813764e-07, "epoch": 3.464980544747082 }, { "current_steps": 1781, "loss": 2.0012, "learning_rate": 3.214574898785425e-07, "epoch": 3.4669260700389106 }, { "current_steps": 1782, "loss": 2.355, "learning_rate": 3.2105263157894733e-07, "epoch": 3.4688715953307394 }, { "current_steps": 1783, "loss": 2.0773, "learning_rate": 3.2064777327935223e-07, "epoch": 3.470817120622568 }, { "current_steps": 1784, "loss": 2.2175, "learning_rate": 3.202429149797571e-07, "epoch": 3.472762645914397 }, { "current_steps": 1785, "loss": 2.2855, "learning_rate": 3.198380566801619e-07, "epoch": 3.4747081712062258 }, { "current_steps": 1786, "loss": 2.2109, "learning_rate": 3.1943319838056677e-07, "epoch": 3.4766536964980546 }, { "current_steps": 1787, "loss": 1.8802, "learning_rate": 3.190283400809716e-07, "epoch": 3.4785992217898833 }, { "current_steps": 1788, "loss": 2.1308, "learning_rate": 3.186234817813765e-07, "epoch": 3.480544747081712 }, { "current_steps": 1789, "loss": 2.172, "learning_rate": 3.1821862348178136e-07, "epoch": 3.482490272373541 }, { "current_steps": 1790, "loss": 1.9439, "learning_rate": 3.178137651821862e-07, "epoch": 3.4844357976653697 }, { "current_steps": 1791, "loss": 1.8057, "learning_rate": 3.174089068825911e-07, "epoch": 3.4863813229571985 }, { "current_steps": 1792, "loss": 1.8114, "learning_rate": 3.1700404858299596e-07, "epoch": 3.4883268482490273 }, { "current_steps": 1793, "loss": 1.5294, "learning_rate": 3.165991902834008e-07, "epoch": 3.490272373540856 }, { "current_steps": 1794, "loss": 1.5109, "learning_rate": 3.1619433198380565e-07, "epoch": 3.492217898832685 }, { "current_steps": 1795, "loss": 1.9669, "learning_rate": 3.157894736842105e-07, "epoch": 3.4941634241245136 }, { "current_steps": 1796, "loss": 2.0011, "learning_rate": 3.1538461538461534e-07, "epoch": 3.4961089494163424 }, { "current_steps": 1797, "loss": 2.1295, "learning_rate": 3.1497975708502024e-07, "epoch": 3.498054474708171 }, { "current_steps": 1798, "loss": 2.3906, "learning_rate": 3.145748987854251e-07, "epoch": 3.5 }, { "current_steps": 1799, "loss": 1.9042, "learning_rate": 3.1417004048582993e-07, "epoch": 3.501945525291829 }, { "current_steps": 1800, "loss": 2.2078, "learning_rate": 3.1376518218623483e-07, "epoch": 3.5038910505836576 }, { "current_steps": 1801, "loss": 2.126, "learning_rate": 3.133603238866397e-07, "epoch": 3.5058365758754864 }, { "current_steps": 1802, "loss": 2.285, "learning_rate": 3.129554655870445e-07, "epoch": 3.507782101167315 }, { "current_steps": 1803, "loss": 1.9514, "learning_rate": 3.1255060728744937e-07, "epoch": 3.509727626459144 }, { "current_steps": 1804, "loss": 2.2703, "learning_rate": 3.121457489878542e-07, "epoch": 3.5116731517509727 }, { "current_steps": 1805, "loss": 2.2596, "learning_rate": 3.117408906882591e-07, "epoch": 3.5136186770428015 }, { "current_steps": 1806, "loss": 1.8186, "learning_rate": 3.1133603238866397e-07, "epoch": 3.5155642023346303 }, { "current_steps": 1807, "loss": 2.2598, "learning_rate": 3.109311740890688e-07, "epoch": 3.517509727626459 }, { "current_steps": 1808, "loss": 1.8276, "learning_rate": 3.105263157894737e-07, "epoch": 3.519455252918288 }, { "current_steps": 1809, "loss": 1.7466, "learning_rate": 3.101214574898785e-07, "epoch": 3.5214007782101167 }, { "current_steps": 1810, "loss": 2.1578, "learning_rate": 3.0971659919028335e-07, "epoch": 3.5233463035019454 }, { "current_steps": 1811, "loss": 2.165, "learning_rate": 3.0931174089068825e-07, "epoch": 3.5252918287937742 }, { "current_steps": 1812, "loss": 2.0838, "learning_rate": 3.089068825910931e-07, "epoch": 3.527237354085603 }, { "current_steps": 1813, "loss": 2.016, "learning_rate": 3.0850202429149794e-07, "epoch": 3.529182879377432 }, { "current_steps": 1814, "loss": 2.1191, "learning_rate": 3.0809716599190284e-07, "epoch": 3.5311284046692606 }, { "current_steps": 1815, "loss": 2.1597, "learning_rate": 3.076923076923077e-07, "epoch": 3.5330739299610894 }, { "current_steps": 1816, "loss": 2.3065, "learning_rate": 3.0728744939271254e-07, "epoch": 3.535019455252918 }, { "current_steps": 1817, "loss": 2.3364, "learning_rate": 3.068825910931174e-07, "epoch": 3.536964980544747 }, { "current_steps": 1818, "loss": 2.3264, "learning_rate": 3.0647773279352223e-07, "epoch": 3.5389105058365757 }, { "current_steps": 1819, "loss": 2.2807, "learning_rate": 3.060728744939271e-07, "epoch": 3.5408560311284045 }, { "current_steps": 1820, "loss": 2.2653, "learning_rate": 3.05668016194332e-07, "epoch": 3.5428015564202333 }, { "current_steps": 1821, "loss": 2.3083, "learning_rate": 3.052631578947368e-07, "epoch": 3.544747081712062 }, { "current_steps": 1822, "loss": 2.2143, "learning_rate": 3.048582995951417e-07, "epoch": 3.546692607003891 }, { "current_steps": 1823, "loss": 2.1005, "learning_rate": 3.0445344129554657e-07, "epoch": 3.5486381322957197 }, { "current_steps": 1824, "loss": 2.2758, "learning_rate": 3.040485829959514e-07, "epoch": 3.5505836575875485 }, { "current_steps": 1825, "loss": 2.1989, "learning_rate": 3.0364372469635626e-07, "epoch": 3.5525291828793772 }, { "current_steps": 1826, "loss": 2.0952, "learning_rate": 3.032388663967611e-07, "epoch": 3.554474708171206 }, { "current_steps": 1827, "loss": 2.1574, "learning_rate": 3.0283400809716595e-07, "epoch": 3.556420233463035 }, { "current_steps": 1828, "loss": 2.0732, "learning_rate": 3.0242914979757085e-07, "epoch": 3.5583657587548636 }, { "current_steps": 1829, "loss": 2.201, "learning_rate": 3.020242914979757e-07, "epoch": 3.5603112840466924 }, { "current_steps": 1830, "loss": 2.0036, "learning_rate": 3.0161943319838055e-07, "epoch": 3.562256809338521 }, { "current_steps": 1831, "loss": 2.039, "learning_rate": 3.0121457489878545e-07, "epoch": 3.56420233463035 }, { "current_steps": 1832, "loss": 2.0423, "learning_rate": 3.0080971659919024e-07, "epoch": 3.5661478599221788 }, { "current_steps": 1833, "loss": 2.1498, "learning_rate": 3.004048582995951e-07, "epoch": 3.5680933852140075 }, { "current_steps": 1834, "loss": 1.973, "learning_rate": 3e-07, "epoch": 3.5700389105058363 }, { "current_steps": 1835, "loss": 1.9355, "learning_rate": 2.9959514170040483e-07, "epoch": 3.571984435797665 }, { "current_steps": 1836, "loss": 2.1897, "learning_rate": 2.9919028340080973e-07, "epoch": 3.573929961089494 }, { "current_steps": 1837, "loss": 2.1749, "learning_rate": 2.987854251012146e-07, "epoch": 3.5758754863813227 }, { "current_steps": 1838, "loss": 1.8794, "learning_rate": 2.983805668016194e-07, "epoch": 3.5778210116731515 }, { "current_steps": 1839, "loss": 2.0771, "learning_rate": 2.979757085020243e-07, "epoch": 3.5797665369649807 }, { "current_steps": 1840, "loss": 2.082, "learning_rate": 2.975708502024291e-07, "epoch": 3.5817120622568095 }, { "current_steps": 1841, "loss": 1.6815, "learning_rate": 2.9716599190283397e-07, "epoch": 3.5836575875486383 }, { "current_steps": 1842, "loss": 2.0361, "learning_rate": 2.9676113360323886e-07, "epoch": 3.585603112840467 }, { "current_steps": 1843, "loss": 2.1064, "learning_rate": 2.963562753036437e-07, "epoch": 3.587548638132296 }, { "current_steps": 1844, "loss": 2.2981, "learning_rate": 2.9595141700404856e-07, "epoch": 3.5894941634241246 }, { "current_steps": 1845, "loss": 1.3514, "learning_rate": 2.9554655870445346e-07, "epoch": 3.5914396887159534 }, { "current_steps": 1846, "loss": 1.987, "learning_rate": 2.951417004048583e-07, "epoch": 3.593385214007782 }, { "current_steps": 1847, "loss": 1.9677, "learning_rate": 2.947368421052631e-07, "epoch": 3.595330739299611 }, { "current_steps": 1848, "loss": 2.1571, "learning_rate": 2.94331983805668e-07, "epoch": 3.59727626459144 }, { "current_steps": 1849, "loss": 2.2728, "learning_rate": 2.9392712550607284e-07, "epoch": 3.5992217898832686 }, { "current_steps": 1850, "loss": 2.3073, "learning_rate": 2.935222672064777e-07, "epoch": 3.6011673151750974 }, { "current_steps": 1851, "loss": 2.2999, "learning_rate": 2.931174089068826e-07, "epoch": 3.603112840466926 }, { "current_steps": 1852, "loss": 1.6395, "learning_rate": 2.9271255060728744e-07, "epoch": 3.605058365758755 }, { "current_steps": 1853, "loss": 1.8307, "learning_rate": 2.9230769230769234e-07, "epoch": 3.6070038910505837 }, { "current_steps": 1854, "loss": 2.2028, "learning_rate": 2.919028340080972e-07, "epoch": 3.6089494163424125 }, { "current_steps": 1855, "loss": 2.293, "learning_rate": 2.91497975708502e-07, "epoch": 3.6108949416342413 }, { "current_steps": 1856, "loss": 2.0104, "learning_rate": 2.910931174089069e-07, "epoch": 3.61284046692607 }, { "current_steps": 1857, "loss": 2.0461, "learning_rate": 2.906882591093117e-07, "epoch": 3.614785992217899 }, { "current_steps": 1858, "loss": 2.263, "learning_rate": 2.9028340080971657e-07, "epoch": 3.6167315175097277 }, { "current_steps": 1859, "loss": 2.062, "learning_rate": 2.8987854251012147e-07, "epoch": 3.6186770428015564 }, { "current_steps": 1860, "loss": 2.1124, "learning_rate": 2.894736842105263e-07, "epoch": 3.6206225680933852 }, { "current_steps": 1861, "loss": 1.7972, "learning_rate": 2.8906882591093116e-07, "epoch": 3.622568093385214 }, { "current_steps": 1862, "loss": 2.1843, "learning_rate": 2.8866396761133606e-07, "epoch": 3.624513618677043 }, { "current_steps": 1863, "loss": 2.2348, "learning_rate": 2.8825910931174085e-07, "epoch": 3.6264591439688716 }, { "current_steps": 1864, "loss": 2.2823, "learning_rate": 2.878542510121457e-07, "epoch": 3.6284046692607004 }, { "current_steps": 1865, "loss": 2.1447, "learning_rate": 2.874493927125506e-07, "epoch": 3.630350194552529 }, { "current_steps": 1866, "loss": 1.747, "learning_rate": 2.8704453441295545e-07, "epoch": 3.632295719844358 }, { "current_steps": 1867, "loss": 2.1976, "learning_rate": 2.866396761133603e-07, "epoch": 3.6342412451361867 }, { "current_steps": 1868, "loss": 1.8353, "learning_rate": 2.862348178137652e-07, "epoch": 3.6361867704280155 }, { "current_steps": 1869, "loss": 1.974, "learning_rate": 2.8582995951417004e-07, "epoch": 3.6381322957198443 }, { "current_steps": 1870, "loss": 1.8591, "learning_rate": 2.854251012145749e-07, "epoch": 3.640077821011673 }, { "current_steps": 1871, "loss": 2.2346, "learning_rate": 2.8502024291497973e-07, "epoch": 3.642023346303502 }, { "current_steps": 1872, "loss": 2.2975, "learning_rate": 2.846153846153846e-07, "epoch": 3.6439688715953307 }, { "current_steps": 1873, "loss": 2.0656, "learning_rate": 2.842105263157895e-07, "epoch": 3.6459143968871595 }, { "current_steps": 1874, "loss": 2.1498, "learning_rate": 2.838056680161943e-07, "epoch": 3.6478599221789882 }, { "current_steps": 1875, "loss": 2.0367, "learning_rate": 2.8340080971659917e-07, "epoch": 3.649805447470817 }, { "current_steps": 1876, "loss": 1.6306, "learning_rate": 2.8299595141700407e-07, "epoch": 3.651750972762646 }, { "current_steps": 1877, "loss": 2.152, "learning_rate": 2.825910931174089e-07, "epoch": 3.6536964980544746 }, { "current_steps": 1878, "loss": 2.2294, "learning_rate": 2.821862348178137e-07, "epoch": 3.6556420233463034 }, { "current_steps": 1879, "loss": 2.106, "learning_rate": 2.817813765182186e-07, "epoch": 3.657587548638132 }, { "current_steps": 1880, "loss": 2.2014, "learning_rate": 2.8137651821862346e-07, "epoch": 3.659533073929961 }, { "current_steps": 1881, "loss": 2.2003, "learning_rate": 2.809716599190283e-07, "epoch": 3.6614785992217898 }, { "current_steps": 1882, "loss": 1.764, "learning_rate": 2.805668016194332e-07, "epoch": 3.6634241245136185 }, { "current_steps": 1883, "loss": 1.9891, "learning_rate": 2.8016194331983805e-07, "epoch": 3.6653696498054473 }, { "current_steps": 1884, "loss": 2.0754, "learning_rate": 2.7975708502024295e-07, "epoch": 3.667315175097276 }, { "current_steps": 1885, "loss": 2.2527, "learning_rate": 2.793522267206478e-07, "epoch": 3.669260700389105 }, { "current_steps": 1886, "loss": 2.2514, "learning_rate": 2.789473684210526e-07, "epoch": 3.6712062256809337 }, { "current_steps": 1887, "loss": 2.2152, "learning_rate": 2.785425101214575e-07, "epoch": 3.673151750972763 }, { "current_steps": 1888, "loss": 1.984, "learning_rate": 2.7813765182186233e-07, "epoch": 3.6750972762645917 }, { "current_steps": 1889, "loss": 1.8465, "learning_rate": 2.777327935222672e-07, "epoch": 3.6770428015564205 }, { "current_steps": 1890, "loss": 2.0497, "learning_rate": 2.773279352226721e-07, "epoch": 3.6789883268482493 }, { "current_steps": 1891, "loss": 2.1556, "learning_rate": 2.7692307692307693e-07, "epoch": 3.680933852140078 }, { "current_steps": 1892, "loss": 2.2374, "learning_rate": 2.7651821862348177e-07, "epoch": 3.682879377431907 }, { "current_steps": 1893, "loss": 2.22, "learning_rate": 2.761133603238866e-07, "epoch": 3.6848249027237356 }, { "current_steps": 1894, "loss": 2.4045, "learning_rate": 2.7570850202429147e-07, "epoch": 3.6867704280155644 }, { "current_steps": 1895, "loss": 1.996, "learning_rate": 2.753036437246963e-07, "epoch": 3.688715953307393 }, { "current_steps": 1896, "loss": 2.2124, "learning_rate": 2.748987854251012e-07, "epoch": 3.690661478599222 }, { "current_steps": 1897, "loss": 2.2453, "learning_rate": 2.7449392712550606e-07, "epoch": 3.692607003891051 }, { "current_steps": 1898, "loss": 1.7859, "learning_rate": 2.740890688259109e-07, "epoch": 3.6945525291828796 }, { "current_steps": 1899, "loss": 2.3805, "learning_rate": 2.736842105263158e-07, "epoch": 3.6964980544747084 }, { "current_steps": 1900, "loss": 2.1449, "learning_rate": 2.7327935222672065e-07, "epoch": 3.698443579766537 }, { "current_steps": 1901, "loss": 2.2548, "learning_rate": 2.7287449392712545e-07, "epoch": 3.700389105058366 }, { "current_steps": 1902, "loss": 1.6784, "learning_rate": 2.7246963562753034e-07, "epoch": 3.7023346303501947 }, { "current_steps": 1903, "loss": 1.7988, "learning_rate": 2.720647773279352e-07, "epoch": 3.7042801556420235 }, { "current_steps": 1904, "loss": 2.3208, "learning_rate": 2.716599190283401e-07, "epoch": 3.7062256809338523 }, { "current_steps": 1905, "loss": 1.9801, "learning_rate": 2.7125506072874494e-07, "epoch": 3.708171206225681 }, { "current_steps": 1906, "loss": 1.8155, "learning_rate": 2.708502024291498e-07, "epoch": 3.71011673151751 }, { "current_steps": 1907, "loss": 2.0877, "learning_rate": 2.704453441295547e-07, "epoch": 3.7120622568093387 }, { "current_steps": 1908, "loss": 2.2108, "learning_rate": 2.700404858299595e-07, "epoch": 3.7140077821011674 }, { "current_steps": 1909, "loss": 2.1292, "learning_rate": 2.696356275303643e-07, "epoch": 3.7159533073929962 }, { "current_steps": 1910, "loss": 2.2344, "learning_rate": 2.692307692307692e-07, "epoch": 3.717898832684825 }, { "current_steps": 1911, "loss": 2.2667, "learning_rate": 2.6882591093117407e-07, "epoch": 3.719844357976654 }, { "current_steps": 1912, "loss": 2.1605, "learning_rate": 2.684210526315789e-07, "epoch": 3.7217898832684826 }, { "current_steps": 1913, "loss": 2.2421, "learning_rate": 2.680161943319838e-07, "epoch": 3.7237354085603114 }, { "current_steps": 1914, "loss": 2.1899, "learning_rate": 2.6761133603238866e-07, "epoch": 3.72568093385214 }, { "current_steps": 1915, "loss": 2.2178, "learning_rate": 2.672064777327935e-07, "epoch": 3.727626459143969 }, { "current_steps": 1916, "loss": 2.2881, "learning_rate": 2.6680161943319835e-07, "epoch": 3.7295719844357977 }, { "current_steps": 1917, "loss": 2.0969, "learning_rate": 2.663967611336032e-07, "epoch": 3.7315175097276265 }, { "current_steps": 1918, "loss": 2.1822, "learning_rate": 2.659919028340081e-07, "epoch": 3.7334630350194553 }, { "current_steps": 1919, "loss": 1.8511, "learning_rate": 2.6558704453441295e-07, "epoch": 3.735408560311284 }, { "current_steps": 1920, "loss": 2.176, "learning_rate": 2.651821862348178e-07, "epoch": 3.737354085603113 }, { "current_steps": 1921, "loss": 2.1628, "learning_rate": 2.647773279352227e-07, "epoch": 3.7392996108949417 }, { "current_steps": 1922, "loss": 2.0384, "learning_rate": 2.6437246963562754e-07, "epoch": 3.7412451361867705 }, { "current_steps": 1923, "loss": 1.6278, "learning_rate": 2.639676113360324e-07, "epoch": 3.7431906614785992 }, { "current_steps": 1924, "loss": 2.1075, "learning_rate": 2.6356275303643723e-07, "epoch": 3.745136186770428 }, { "current_steps": 1925, "loss": 1.9405, "learning_rate": 2.631578947368421e-07, "epoch": 3.747081712062257 }, { "current_steps": 1926, "loss": 2.1497, "learning_rate": 2.627530364372469e-07, "epoch": 3.7490272373540856 }, { "current_steps": 1927, "loss": 1.997, "learning_rate": 2.623481781376518e-07, "epoch": 3.7509727626459144 }, { "current_steps": 1928, "loss": 2.2703, "learning_rate": 2.6194331983805667e-07, "epoch": 3.752918287937743 }, { "current_steps": 1929, "loss": 1.7385, "learning_rate": 2.615384615384615e-07, "epoch": 3.754863813229572 }, { "current_steps": 1930, "loss": 2.1869, "learning_rate": 2.611336032388664e-07, "epoch": 3.7568093385214008 }, { "current_steps": 1931, "loss": 2.2542, "learning_rate": 2.607287449392712e-07, "epoch": 3.7587548638132295 }, { "current_steps": 1932, "loss": 2.0754, "learning_rate": 2.6032388663967606e-07, "epoch": 3.7607003891050583 }, { "current_steps": 1933, "loss": 2.1951, "learning_rate": 2.5991902834008096e-07, "epoch": 3.762645914396887 }, { "current_steps": 1934, "loss": 2.2114, "learning_rate": 2.595141700404858e-07, "epoch": 3.764591439688716 }, { "current_steps": 1935, "loss": 2.0215, "learning_rate": 2.591093117408907e-07, "epoch": 3.7665369649805447 }, { "current_steps": 1936, "loss": 2.2283, "learning_rate": 2.5870445344129555e-07, "epoch": 3.7684824902723735 }, { "current_steps": 1937, "loss": 1.899, "learning_rate": 2.582995951417004e-07, "epoch": 3.7704280155642023 }, { "current_steps": 1938, "loss": 1.6843, "learning_rate": 2.578947368421053e-07, "epoch": 3.772373540856031 }, { "current_steps": 1939, "loss": 1.7526, "learning_rate": 2.574898785425101e-07, "epoch": 3.77431906614786 }, { "current_steps": 1940, "loss": 1.9869, "learning_rate": 2.5708502024291494e-07, "epoch": 3.7762645914396886 }, { "current_steps": 1941, "loss": 2.4109, "learning_rate": 2.5668016194331984e-07, "epoch": 3.7782101167315174 }, { "current_steps": 1942, "loss": 2.104, "learning_rate": 2.562753036437247e-07, "epoch": 3.780155642023346 }, { "current_steps": 1943, "loss": 1.7913, "learning_rate": 2.5587044534412953e-07, "epoch": 3.782101167315175 }, { "current_steps": 1944, "loss": 1.7814, "learning_rate": 2.5546558704453443e-07, "epoch": 3.7840466926070038 }, { "current_steps": 1945, "loss": 1.8563, "learning_rate": 2.550607287449393e-07, "epoch": 3.7859922178988326 }, { "current_steps": 1946, "loss": 2.0697, "learning_rate": 2.546558704453441e-07, "epoch": 3.7879377431906613 }, { "current_steps": 1947, "loss": 2.1857, "learning_rate": 2.5425101214574897e-07, "epoch": 3.78988326848249 }, { "current_steps": 1948, "loss": 2.0338, "learning_rate": 2.538461538461538e-07, "epoch": 3.791828793774319 }, { "current_steps": 1949, "loss": 2.0635, "learning_rate": 2.5344129554655866e-07, "epoch": 3.7937743190661477 }, { "current_steps": 1950, "loss": 2.2217, "learning_rate": 2.5303643724696356e-07, "epoch": 3.7957198443579765 }, { "current_steps": 1951, "loss": 2.183, "learning_rate": 2.526315789473684e-07, "epoch": 3.7976653696498053 }, { "current_steps": 1952, "loss": 2.1804, "learning_rate": 2.522267206477733e-07, "epoch": 3.799610894941634 }, { "current_steps": 1953, "loss": 2.1923, "learning_rate": 2.5182186234817815e-07, "epoch": 3.801556420233463 }, { "current_steps": 1954, "loss": 2.3577, "learning_rate": 2.5141700404858295e-07, "epoch": 3.8035019455252916 }, { "current_steps": 1955, "loss": 2.2205, "learning_rate": 2.5101214574898785e-07, "epoch": 3.8054474708171204 }, { "current_steps": 1956, "loss": 1.9184, "learning_rate": 2.506072874493927e-07, "epoch": 3.807392996108949 }, { "current_steps": 1957, "loss": 2.1101, "learning_rate": 2.5020242914979754e-07, "epoch": 3.809338521400778 }, { "current_steps": 1958, "loss": 2.3204, "learning_rate": 2.4979757085020244e-07, "epoch": 3.811284046692607 }, { "current_steps": 1959, "loss": 2.1745, "learning_rate": 2.493927125506073e-07, "epoch": 3.8132295719844356 }, { "current_steps": 1960, "loss": 2.4869, "learning_rate": 2.4898785425101213e-07, "epoch": 3.8151750972762644 }, { "current_steps": 1961, "loss": 2.2861, "learning_rate": 2.48582995951417e-07, "epoch": 3.817120622568093 }, { "current_steps": 1962, "loss": 2.4314, "learning_rate": 2.481781376518219e-07, "epoch": 3.819066147859922 }, { "current_steps": 1963, "loss": 1.9495, "learning_rate": 2.477732793522267e-07, "epoch": 3.8210116731517507 }, { "current_steps": 1964, "loss": 1.815, "learning_rate": 2.4736842105263157e-07, "epoch": 3.8229571984435795 }, { "current_steps": 1965, "loss": 1.626, "learning_rate": 2.469635627530364e-07, "epoch": 3.8249027237354083 }, { "current_steps": 1966, "loss": 1.8571, "learning_rate": 2.465587044534413e-07, "epoch": 3.826848249027237 }, { "current_steps": 1967, "loss": 2.1213, "learning_rate": 2.4615384615384616e-07, "epoch": 3.8287937743190663 }, { "current_steps": 1968, "loss": 2.2706, "learning_rate": 2.45748987854251e-07, "epoch": 3.830739299610895 }, { "current_steps": 1969, "loss": 1.889, "learning_rate": 2.4534412955465586e-07, "epoch": 3.832684824902724 }, { "current_steps": 1970, "loss": 1.7838, "learning_rate": 2.449392712550607e-07, "epoch": 3.8346303501945527 }, { "current_steps": 1971, "loss": 2.2624, "learning_rate": 2.4453441295546555e-07, "epoch": 3.8365758754863815 }, { "current_steps": 1972, "loss": 1.6221, "learning_rate": 2.4412955465587045e-07, "epoch": 3.8385214007782102 }, { "current_steps": 1973, "loss": 2.0807, "learning_rate": 2.437246963562753e-07, "epoch": 3.840466926070039 }, { "current_steps": 1974, "loss": 2.0037, "learning_rate": 2.4331983805668014e-07, "epoch": 3.842412451361868 }, { "current_steps": 1975, "loss": 2.0076, "learning_rate": 2.42914979757085e-07, "epoch": 3.8443579766536966 }, { "current_steps": 1976, "loss": 1.9701, "learning_rate": 2.425101214574899e-07, "epoch": 3.8463035019455254 }, { "current_steps": 1977, "loss": 2.2026, "learning_rate": 2.4210526315789473e-07, "epoch": 3.848249027237354 }, { "current_steps": 1978, "loss": 2.3318, "learning_rate": 2.417004048582996e-07, "epoch": 3.850194552529183 }, { "current_steps": 1979, "loss": 2.2724, "learning_rate": 2.4129554655870443e-07, "epoch": 3.8521400778210118 }, { "current_steps": 1980, "loss": 2.2075, "learning_rate": 2.408906882591093e-07, "epoch": 3.8540856031128405 }, { "current_steps": 1981, "loss": 2.2593, "learning_rate": 2.4048582995951417e-07, "epoch": 3.8560311284046693 }, { "current_steps": 1982, "loss": 1.7021, "learning_rate": 2.40080971659919e-07, "epoch": 3.857976653696498 }, { "current_steps": 1983, "loss": 1.9743, "learning_rate": 2.3967611336032387e-07, "epoch": 3.859922178988327 }, { "current_steps": 1984, "loss": 2.037, "learning_rate": 2.392712550607287e-07, "epoch": 3.8618677042801557 }, { "current_steps": 1985, "loss": 2.2152, "learning_rate": 2.388663967611336e-07, "epoch": 3.8638132295719845 }, { "current_steps": 1986, "loss": 2.0673, "learning_rate": 2.3846153846153846e-07, "epoch": 3.8657587548638133 }, { "current_steps": 1987, "loss": 2.0433, "learning_rate": 2.380566801619433e-07, "epoch": 3.867704280155642 }, { "current_steps": 1988, "loss": 2.2908, "learning_rate": 2.3765182186234818e-07, "epoch": 3.869649805447471 }, { "current_steps": 1989, "loss": 2.2596, "learning_rate": 2.3724696356275302e-07, "epoch": 3.8715953307392996 }, { "current_steps": 1990, "loss": 2.1832, "learning_rate": 2.3684210526315787e-07, "epoch": 3.8735408560311284 }, { "current_steps": 1991, "loss": 2.0211, "learning_rate": 2.3643724696356274e-07, "epoch": 3.875486381322957 }, { "current_steps": 1992, "loss": 1.8196, "learning_rate": 2.360323886639676e-07, "epoch": 3.877431906614786 }, { "current_steps": 1993, "loss": 2.1275, "learning_rate": 2.3562753036437246e-07, "epoch": 3.8793774319066148 }, { "current_steps": 1994, "loss": 2.0257, "learning_rate": 2.352226720647773e-07, "epoch": 3.8813229571984436 }, { "current_steps": 1995, "loss": 1.9692, "learning_rate": 2.3481781376518218e-07, "epoch": 3.8832684824902723 }, { "current_steps": 1996, "loss": 2.2015, "learning_rate": 2.3441295546558703e-07, "epoch": 3.885214007782101 }, { "current_steps": 1997, "loss": 1.5446, "learning_rate": 2.340080971659919e-07, "epoch": 3.88715953307393 }, { "current_steps": 1998, "loss": 2.0543, "learning_rate": 2.3360323886639675e-07, "epoch": 3.8891050583657587 }, { "current_steps": 1999, "loss": 1.7719, "learning_rate": 2.331983805668016e-07, "epoch": 3.8910505836575875 }, { "current_steps": 2000, "loss": 1.239, "learning_rate": 2.3279352226720647e-07, "epoch": 3.8929961089494163 }, { "current_steps": 2001, "loss": 2.0245, "learning_rate": 2.3238866396761134e-07, "epoch": 3.894941634241245 }, { "current_steps": 2002, "loss": 2.0481, "learning_rate": 2.3198380566801616e-07, "epoch": 3.896887159533074 }, { "current_steps": 2003, "loss": 2.253, "learning_rate": 2.3157894736842104e-07, "epoch": 3.8988326848249026 }, { "current_steps": 2004, "loss": 2.4497, "learning_rate": 2.311740890688259e-07, "epoch": 3.9007782101167314 }, { "current_steps": 2005, "loss": 2.1789, "learning_rate": 2.3076923076923078e-07, "epoch": 3.90272373540856 }, { "current_steps": 2006, "loss": 2.3731, "learning_rate": 2.303643724696356e-07, "epoch": 3.904669260700389 }, { "current_steps": 2007, "loss": 2.2582, "learning_rate": 2.2995951417004047e-07, "epoch": 3.906614785992218 }, { "current_steps": 2008, "loss": 2.2242, "learning_rate": 2.2955465587044535e-07, "epoch": 3.9085603112840466 }, { "current_steps": 2009, "loss": 2.2444, "learning_rate": 2.2914979757085017e-07, "epoch": 3.9105058365758754 }, { "current_steps": 2010, "loss": 2.157, "learning_rate": 2.2874493927125504e-07, "epoch": 3.912451361867704 }, { "current_steps": 2011, "loss": 1.9799, "learning_rate": 2.2834008097165991e-07, "epoch": 3.914396887159533 }, { "current_steps": 2012, "loss": 2.555, "learning_rate": 2.2793522267206479e-07, "epoch": 3.9163424124513617 }, { "current_steps": 2013, "loss": 2.1442, "learning_rate": 2.275303643724696e-07, "epoch": 3.9182879377431905 }, { "current_steps": 2014, "loss": 1.9816, "learning_rate": 2.2712550607287448e-07, "epoch": 3.9202334630350193 }, { "current_steps": 2015, "loss": 1.7759, "learning_rate": 2.2672064777327935e-07, "epoch": 3.9221789883268485 }, { "current_steps": 2016, "loss": 1.6671, "learning_rate": 2.263157894736842e-07, "epoch": 3.9241245136186773 }, { "current_steps": 2017, "loss": 2.2673, "learning_rate": 2.2591093117408905e-07, "epoch": 3.926070038910506 }, { "current_steps": 2018, "loss": 2.2412, "learning_rate": 2.2550607287449392e-07, "epoch": 3.928015564202335 }, { "current_steps": 2019, "loss": 2.1908, "learning_rate": 2.251012145748988e-07, "epoch": 3.9299610894941637 }, { "current_steps": 2020, "loss": 2.0819, "learning_rate": 2.2469635627530364e-07, "epoch": 3.9319066147859925 }, { "current_steps": 2021, "loss": 1.8162, "learning_rate": 2.2429149797570848e-07, "epoch": 3.9338521400778212 }, { "current_steps": 2022, "loss": 2.2864, "learning_rate": 2.2388663967611336e-07, "epoch": 3.93579766536965 }, { "current_steps": 2023, "loss": 2.0097, "learning_rate": 2.234817813765182e-07, "epoch": 3.937743190661479 }, { "current_steps": 2024, "loss": 1.8354, "learning_rate": 2.2307692307692308e-07, "epoch": 3.9396887159533076 }, { "current_steps": 2025, "loss": 2.1364, "learning_rate": 2.2267206477732792e-07, "epoch": 3.9416342412451364 }, { "current_steps": 2026, "loss": 2.0766, "learning_rate": 2.2226720647773277e-07, "epoch": 3.943579766536965 }, { "current_steps": 2027, "loss": 2.0179, "learning_rate": 2.2186234817813764e-07, "epoch": 3.945525291828794 }, { "current_steps": 2028, "loss": 2.3047, "learning_rate": 2.2145748987854252e-07, "epoch": 3.9474708171206228 }, { "current_steps": 2029, "loss": 2.0024, "learning_rate": 2.2105263157894736e-07, "epoch": 3.9494163424124515 }, { "current_steps": 2030, "loss": 1.9624, "learning_rate": 2.206477732793522e-07, "epoch": 3.9513618677042803 }, { "current_steps": 2031, "loss": 2.0209, "learning_rate": 2.2024291497975708e-07, "epoch": 3.953307392996109 }, { "current_steps": 2032, "loss": 2.1008, "learning_rate": 2.1983805668016193e-07, "epoch": 3.955252918287938 }, { "current_steps": 2033, "loss": 2.0893, "learning_rate": 2.1943319838056678e-07, "epoch": 3.9571984435797667 }, { "current_steps": 2034, "loss": 1.8607, "learning_rate": 2.1902834008097165e-07, "epoch": 3.9591439688715955 }, { "current_steps": 2035, "loss": 2.4197, "learning_rate": 2.1862348178137652e-07, "epoch": 3.9610894941634243 }, { "current_steps": 2036, "loss": 2.0482, "learning_rate": 2.1821862348178137e-07, "epoch": 3.963035019455253 }, { "current_steps": 2037, "loss": 2.0158, "learning_rate": 2.1781376518218621e-07, "epoch": 3.964980544747082 }, { "current_steps": 2038, "loss": 2.0037, "learning_rate": 2.174089068825911e-07, "epoch": 3.9669260700389106 }, { "current_steps": 2039, "loss": 2.2538, "learning_rate": 2.1700404858299596e-07, "epoch": 3.9688715953307394 }, { "current_steps": 2040, "loss": 2.098, "learning_rate": 2.1659919028340078e-07, "epoch": 3.970817120622568 }, { "current_steps": 2041, "loss": 2.0266, "learning_rate": 2.1619433198380565e-07, "epoch": 3.972762645914397 }, { "current_steps": 2042, "loss": 1.7409, "learning_rate": 2.1578947368421053e-07, "epoch": 3.9747081712062258 }, { "current_steps": 2043, "loss": 2.0556, "learning_rate": 2.153846153846154e-07, "epoch": 3.9766536964980546 }, { "current_steps": 2044, "loss": 1.8364, "learning_rate": 2.1497975708502022e-07, "epoch": 3.9785992217898833 }, { "current_steps": 2045, "loss": 2.2779, "learning_rate": 2.145748987854251e-07, "epoch": 3.980544747081712 }, { "current_steps": 2046, "loss": 2.0399, "learning_rate": 2.1417004048582997e-07, "epoch": 3.982490272373541 }, { "current_steps": 2047, "loss": 2.3603, "learning_rate": 2.137651821862348e-07, "epoch": 3.9844357976653697 }, { "current_steps": 2048, "loss": 2.1301, "learning_rate": 2.1336032388663966e-07, "epoch": 3.9863813229571985 }, { "current_steps": 2049, "loss": 2.0112, "learning_rate": 2.1295546558704453e-07, "epoch": 3.9883268482490273 }, { "current_steps": 2050, "loss": 1.7564, "learning_rate": 2.1255060728744938e-07, "epoch": 3.990272373540856 }, { "current_steps": 2051, "loss": 2.2234, "learning_rate": 2.1214574898785422e-07, "epoch": 3.992217898832685 }, { "current_steps": 2052, "loss": 2.2786, "learning_rate": 2.117408906882591e-07, "epoch": 3.9941634241245136 }, { "current_steps": 2053, "loss": 1.6758, "learning_rate": 2.1133603238866397e-07, "epoch": 3.9961089494163424 }, { "current_steps": 2054, "loss": 1.8247, "learning_rate": 2.1093117408906882e-07, "epoch": 3.998054474708171 }, { "current_steps": 2055, "loss": 2.2507, "learning_rate": 2.1052631578947366e-07, "epoch": 4.0 }, { "current_steps": 2056, "loss": 2.2801, "learning_rate": 2.1012145748987854e-07, "epoch": 4.001945525291829 }, { "current_steps": 2057, "loss": 2.1666, "learning_rate": 2.0971659919028338e-07, "epoch": 4.003891050583658 }, { "current_steps": 2058, "loss": 2.2155, "learning_rate": 2.0931174089068826e-07, "epoch": 4.005836575875486 }, { "current_steps": 2059, "loss": 2.2205, "learning_rate": 2.089068825910931e-07, "epoch": 4.007782101167315 }, { "current_steps": 2060, "loss": 2.0365, "learning_rate": 2.0850202429149798e-07, "epoch": 4.009727626459144 }, { "current_steps": 2061, "loss": 2.1038, "learning_rate": 2.0809716599190282e-07, "epoch": 4.011673151750973 }, { "current_steps": 2062, "loss": 2.1324, "learning_rate": 2.076923076923077e-07, "epoch": 4.0136186770428015 }, { "current_steps": 2063, "loss": 2.0657, "learning_rate": 2.0728744939271254e-07, "epoch": 4.01556420233463 }, { "current_steps": 2064, "loss": 1.9209, "learning_rate": 2.068825910931174e-07, "epoch": 4.017509727626459 }, { "current_steps": 2065, "loss": 2.0754, "learning_rate": 2.0647773279352226e-07, "epoch": 4.019455252918288 }, { "current_steps": 2066, "loss": 2.3121, "learning_rate": 2.0607287449392713e-07, "epoch": 4.021400778210117 }, { "current_steps": 2067, "loss": 2.1639, "learning_rate": 2.0566801619433195e-07, "epoch": 4.023346303501945 }, { "current_steps": 2068, "loss": 2.2584, "learning_rate": 2.0526315789473683e-07, "epoch": 4.025291828793774 }, { "current_steps": 2069, "loss": 2.0545, "learning_rate": 2.048582995951417e-07, "epoch": 4.027237354085603 }, { "current_steps": 2070, "loss": 2.2452, "learning_rate": 2.0445344129554655e-07, "epoch": 4.029182879377432 }, { "current_steps": 2071, "loss": 2.0836, "learning_rate": 2.040485829959514e-07, "epoch": 4.031128404669261 }, { "current_steps": 2072, "loss": 2.259, "learning_rate": 2.0364372469635627e-07, "epoch": 4.033073929961089 }, { "current_steps": 2073, "loss": 1.8664, "learning_rate": 2.0323886639676114e-07, "epoch": 4.035019455252918 }, { "current_steps": 2074, "loss": 2.106, "learning_rate": 2.0283400809716596e-07, "epoch": 4.036964980544747 }, { "current_steps": 2075, "loss": 1.8801, "learning_rate": 2.0242914979757083e-07, "epoch": 4.038910505836576 }, { "current_steps": 2076, "loss": 2.0641, "learning_rate": 2.020242914979757e-07, "epoch": 4.0408560311284045 }, { "current_steps": 2077, "loss": 2.0995, "learning_rate": 2.0161943319838058e-07, "epoch": 4.042801556420233 }, { "current_steps": 2078, "loss": 1.9942, "learning_rate": 2.012145748987854e-07, "epoch": 4.044747081712062 }, { "current_steps": 2079, "loss": 2.1732, "learning_rate": 2.0080971659919027e-07, "epoch": 4.046692607003891 }, { "current_steps": 2080, "loss": 2.3169, "learning_rate": 2.0040485829959514e-07, "epoch": 4.04863813229572 }, { "current_steps": 2081, "loss": 2.0414, "learning_rate": 2e-07, "epoch": 4.0505836575875485 }, { "current_steps": 2082, "loss": 1.8323, "learning_rate": 1.9959514170040484e-07, "epoch": 4.052529182879377 }, { "current_steps": 2083, "loss": 1.8879, "learning_rate": 1.991902834008097e-07, "epoch": 4.054474708171206 }, { "current_steps": 2084, "loss": 2.1105, "learning_rate": 1.9878542510121458e-07, "epoch": 4.056420233463035 }, { "current_steps": 2085, "loss": 2.0964, "learning_rate": 1.9838056680161943e-07, "epoch": 4.058365758754864 }, { "current_steps": 2086, "loss": 2.0048, "learning_rate": 1.9797570850202428e-07, "epoch": 4.060311284046692 }, { "current_steps": 2087, "loss": 2.039, "learning_rate": 1.9757085020242915e-07, "epoch": 4.062256809338521 }, { "current_steps": 2088, "loss": 2.2594, "learning_rate": 1.97165991902834e-07, "epoch": 4.06420233463035 }, { "current_steps": 2089, "loss": 2.0367, "learning_rate": 1.9676113360323887e-07, "epoch": 4.066147859922179 }, { "current_steps": 2090, "loss": 2.1754, "learning_rate": 1.9635627530364372e-07, "epoch": 4.0680933852140075 }, { "current_steps": 2091, "loss": 2.1008, "learning_rate": 1.9595141700404856e-07, "epoch": 4.070038910505836 }, { "current_steps": 2092, "loss": 2.0437, "learning_rate": 1.9554655870445344e-07, "epoch": 4.071984435797665 }, { "current_steps": 2093, "loss": 1.5795, "learning_rate": 1.9514170040485828e-07, "epoch": 4.073929961089494 }, { "current_steps": 2094, "loss": 2.12, "learning_rate": 1.9473684210526315e-07, "epoch": 4.075875486381323 }, { "current_steps": 2095, "loss": 1.7826, "learning_rate": 1.94331983805668e-07, "epoch": 4.0778210116731515 }, { "current_steps": 2096, "loss": 1.9761, "learning_rate": 1.9392712550607287e-07, "epoch": 4.07976653696498 }, { "current_steps": 2097, "loss": 1.9374, "learning_rate": 1.9352226720647772e-07, "epoch": 4.081712062256809 }, { "current_steps": 2098, "loss": 2.1646, "learning_rate": 1.9311740890688257e-07, "epoch": 4.083657587548638 }, { "current_steps": 2099, "loss": 1.9916, "learning_rate": 1.9271255060728744e-07, "epoch": 4.085603112840467 }, { "current_steps": 2100, "loss": 2.0146, "learning_rate": 1.9230769230769231e-07, "epoch": 4.087548638132295 }, { "current_steps": 2101, "loss": 2.2736, "learning_rate": 1.9190283400809716e-07, "epoch": 4.089494163424124 }, { "current_steps": 2102, "loss": 2.2432, "learning_rate": 1.91497975708502e-07, "epoch": 4.091439688715953 }, { "current_steps": 2103, "loss": 1.997, "learning_rate": 1.9109311740890688e-07, "epoch": 4.093385214007782 }, { "current_steps": 2104, "loss": 2.3176, "learning_rate": 1.9068825910931175e-07, "epoch": 4.095330739299611 }, { "current_steps": 2105, "loss": 1.5088, "learning_rate": 1.9028340080971657e-07, "epoch": 4.097276264591439 }, { "current_steps": 2106, "loss": 1.7421, "learning_rate": 1.8987854251012145e-07, "epoch": 4.099221789883268 }, { "current_steps": 2107, "loss": 1.8006, "learning_rate": 1.8947368421052632e-07, "epoch": 4.101167315175097 }, { "current_steps": 2108, "loss": 2.271, "learning_rate": 1.890688259109312e-07, "epoch": 4.103112840466926 }, { "current_steps": 2109, "loss": 1.8175, "learning_rate": 1.88663967611336e-07, "epoch": 4.1050583657587545 }, { "current_steps": 2110, "loss": 2.0376, "learning_rate": 1.8825910931174088e-07, "epoch": 4.107003891050583 }, { "current_steps": 2111, "loss": 2.4856, "learning_rate": 1.8785425101214576e-07, "epoch": 4.108949416342412 }, { "current_steps": 2112, "loss": 1.6771, "learning_rate": 1.8744939271255058e-07, "epoch": 4.110894941634241 }, { "current_steps": 2113, "loss": 2.2438, "learning_rate": 1.8704453441295545e-07, "epoch": 4.11284046692607 }, { "current_steps": 2114, "loss": 1.8899, "learning_rate": 1.8663967611336032e-07, "epoch": 4.114785992217898 }, { "current_steps": 2115, "loss": 1.9346, "learning_rate": 1.8623481781376517e-07, "epoch": 4.116731517509727 }, { "current_steps": 2116, "loss": 2.0663, "learning_rate": 1.8582995951417002e-07, "epoch": 4.118677042801556 }, { "current_steps": 2117, "loss": 1.7195, "learning_rate": 1.854251012145749e-07, "epoch": 4.120622568093385 }, { "current_steps": 2118, "loss": 2.2104, "learning_rate": 1.8502024291497976e-07, "epoch": 4.122568093385214 }, { "current_steps": 2119, "loss": 1.9975, "learning_rate": 1.846153846153846e-07, "epoch": 4.124513618677042 }, { "current_steps": 2120, "loss": 1.8669, "learning_rate": 1.8421052631578946e-07, "epoch": 4.126459143968871 }, { "current_steps": 2121, "loss": 2.1131, "learning_rate": 1.8380566801619433e-07, "epoch": 4.1284046692607 }, { "current_steps": 2122, "loss": 2.1442, "learning_rate": 1.8340080971659918e-07, "epoch": 4.13035019455253 }, { "current_steps": 2123, "loss": 2.2333, "learning_rate": 1.8299595141700405e-07, "epoch": 4.132295719844358 }, { "current_steps": 2124, "loss": 2.4257, "learning_rate": 1.825910931174089e-07, "epoch": 4.134241245136187 }, { "current_steps": 2125, "loss": 1.4615, "learning_rate": 1.8218623481781377e-07, "epoch": 4.136186770428016 }, { "current_steps": 2126, "loss": 2.1378, "learning_rate": 1.8178137651821861e-07, "epoch": 4.138132295719845 }, { "current_steps": 2127, "loss": 2.2105, "learning_rate": 1.813765182186235e-07, "epoch": 4.1400778210116735 }, { "current_steps": 2128, "loss": 2.2636, "learning_rate": 1.8097165991902833e-07, "epoch": 4.142023346303502 }, { "current_steps": 2129, "loss": 2.1858, "learning_rate": 1.8056680161943318e-07, "epoch": 4.143968871595331 }, { "current_steps": 2130, "loss": 2.1167, "learning_rate": 1.8016194331983805e-07, "epoch": 4.14591439688716 }, { "current_steps": 2131, "loss": 2.212, "learning_rate": 1.797570850202429e-07, "epoch": 4.147859922178989 }, { "current_steps": 2132, "loss": 2.1358, "learning_rate": 1.7935222672064775e-07, "epoch": 4.1498054474708175 }, { "current_steps": 2133, "loss": 1.9865, "learning_rate": 1.7894736842105262e-07, "epoch": 4.151750972762646 }, { "current_steps": 2134, "loss": 2.1604, "learning_rate": 1.785425101214575e-07, "epoch": 4.153696498054475 }, { "current_steps": 2135, "loss": 1.7259, "learning_rate": 1.7813765182186234e-07, "epoch": 4.155642023346304 }, { "current_steps": 2136, "loss": 1.5215, "learning_rate": 1.7773279352226719e-07, "epoch": 4.157587548638133 }, { "current_steps": 2137, "loss": 2.1393, "learning_rate": 1.7732793522267206e-07, "epoch": 4.159533073929961 }, { "current_steps": 2138, "loss": 1.6864, "learning_rate": 1.7692307692307693e-07, "epoch": 4.16147859922179 }, { "current_steps": 2139, "loss": 1.9951, "learning_rate": 1.7651821862348175e-07, "epoch": 4.163424124513619 }, { "current_steps": 2140, "loss": 2.189, "learning_rate": 1.7611336032388662e-07, "epoch": 4.165369649805448 }, { "current_steps": 2141, "loss": 2.1099, "learning_rate": 1.757085020242915e-07, "epoch": 4.167315175097277 }, { "current_steps": 2142, "loss": 2.3393, "learning_rate": 1.7530364372469637e-07, "epoch": 4.169260700389105 }, { "current_steps": 2143, "loss": 1.6908, "learning_rate": 1.748987854251012e-07, "epoch": 4.171206225680934 }, { "current_steps": 2144, "loss": 1.8379, "learning_rate": 1.7449392712550606e-07, "epoch": 4.173151750972763 }, { "current_steps": 2145, "loss": 2.1007, "learning_rate": 1.7408906882591094e-07, "epoch": 4.175097276264592 }, { "current_steps": 2146, "loss": 1.9688, "learning_rate": 1.7368421052631578e-07, "epoch": 4.1770428015564205 }, { "current_steps": 2147, "loss": 1.8008, "learning_rate": 1.7327935222672063e-07, "epoch": 4.178988326848249 }, { "current_steps": 2148, "loss": 2.0692, "learning_rate": 1.728744939271255e-07, "epoch": 4.180933852140078 }, { "current_steps": 2149, "loss": 2.0123, "learning_rate": 1.7246963562753038e-07, "epoch": 4.182879377431907 }, { "current_steps": 2150, "loss": 2.0901, "learning_rate": 1.720647773279352e-07, "epoch": 4.184824902723736 }, { "current_steps": 2151, "loss": 1.7542, "learning_rate": 1.7165991902834007e-07, "epoch": 4.186770428015564 }, { "current_steps": 2152, "loss": 2.037, "learning_rate": 1.7125506072874494e-07, "epoch": 4.188715953307393 }, { "current_steps": 2153, "loss": 2.1336, "learning_rate": 1.708502024291498e-07, "epoch": 4.190661478599222 }, { "current_steps": 2154, "loss": 2.1952, "learning_rate": 1.7044534412955463e-07, "epoch": 4.192607003891051 }, { "current_steps": 2155, "loss": 2.2231, "learning_rate": 1.700404858299595e-07, "epoch": 4.19455252918288 }, { "current_steps": 2156, "loss": 1.9166, "learning_rate": 1.6963562753036435e-07, "epoch": 4.196498054474708 }, { "current_steps": 2157, "loss": 1.9807, "learning_rate": 1.6923076923076923e-07, "epoch": 4.198443579766537 }, { "current_steps": 2158, "loss": 2.0386, "learning_rate": 1.6882591093117407e-07, "epoch": 4.200389105058366 }, { "current_steps": 2159, "loss": 1.871, "learning_rate": 1.6842105263157895e-07, "epoch": 4.202334630350195 }, { "current_steps": 2160, "loss": 2.1297, "learning_rate": 1.680161943319838e-07, "epoch": 4.2042801556420235 }, { "current_steps": 2161, "loss": 1.9919, "learning_rate": 1.6761133603238867e-07, "epoch": 4.206225680933852 }, { "current_steps": 2162, "loss": 2.3496, "learning_rate": 1.672064777327935e-07, "epoch": 4.208171206225681 }, { "current_steps": 2163, "loss": 1.7649, "learning_rate": 1.6680161943319836e-07, "epoch": 4.21011673151751 }, { "current_steps": 2164, "loss": 2.2766, "learning_rate": 1.6639676113360323e-07, "epoch": 4.212062256809339 }, { "current_steps": 2165, "loss": 2.1629, "learning_rate": 1.659919028340081e-07, "epoch": 4.214007782101167 }, { "current_steps": 2166, "loss": 2.1708, "learning_rate": 1.6558704453441295e-07, "epoch": 4.215953307392996 }, { "current_steps": 2167, "loss": 2.1055, "learning_rate": 1.651821862348178e-07, "epoch": 4.217898832684825 }, { "current_steps": 2168, "loss": 2.1434, "learning_rate": 1.6477732793522267e-07, "epoch": 4.219844357976654 }, { "current_steps": 2169, "loss": 2.2184, "learning_rate": 1.6437246963562754e-07, "epoch": 4.221789883268483 }, { "current_steps": 2170, "loss": 2.2784, "learning_rate": 1.6396761133603236e-07, "epoch": 4.223735408560311 }, { "current_steps": 2171, "loss": 1.9678, "learning_rate": 1.6356275303643724e-07, "epoch": 4.22568093385214 }, { "current_steps": 2172, "loss": 2.1288, "learning_rate": 1.631578947368421e-07, "epoch": 4.227626459143969 }, { "current_steps": 2173, "loss": 2.1103, "learning_rate": 1.6275303643724693e-07, "epoch": 4.229571984435798 }, { "current_steps": 2174, "loss": 2.0156, "learning_rate": 1.623481781376518e-07, "epoch": 4.2315175097276265 }, { "current_steps": 2175, "loss": 2.0004, "learning_rate": 1.6194331983805668e-07, "epoch": 4.233463035019455 }, { "current_steps": 2176, "loss": 1.9689, "learning_rate": 1.6153846153846155e-07, "epoch": 4.235408560311284 }, { "current_steps": 2177, "loss": 2.0472, "learning_rate": 1.6113360323886637e-07, "epoch": 4.237354085603113 }, { "current_steps": 2178, "loss": 1.7555, "learning_rate": 1.6072874493927124e-07, "epoch": 4.239299610894942 }, { "current_steps": 2179, "loss": 2.1869, "learning_rate": 1.6032388663967612e-07, "epoch": 4.2412451361867705 }, { "current_steps": 2180, "loss": 2.0864, "learning_rate": 1.5991902834008096e-07, "epoch": 4.243190661478599 }, { "current_steps": 2181, "loss": 1.9434, "learning_rate": 1.595141700404858e-07, "epoch": 4.245136186770428 }, { "current_steps": 2182, "loss": 2.0088, "learning_rate": 1.5910931174089068e-07, "epoch": 4.247081712062257 }, { "current_steps": 2183, "loss": 2.0585, "learning_rate": 1.5870445344129555e-07, "epoch": 4.249027237354086 }, { "current_steps": 2184, "loss": 2.2203, "learning_rate": 1.582995951417004e-07, "epoch": 4.250972762645914 }, { "current_steps": 2185, "loss": 1.5884, "learning_rate": 1.5789473684210525e-07, "epoch": 4.252918287937743 }, { "current_steps": 2186, "loss": 2.2633, "learning_rate": 1.5748987854251012e-07, "epoch": 4.254863813229572 }, { "current_steps": 2187, "loss": 2.1737, "learning_rate": 1.5708502024291497e-07, "epoch": 4.256809338521401 }, { "current_steps": 2188, "loss": 1.8464, "learning_rate": 1.5668016194331984e-07, "epoch": 4.2587548638132295 }, { "current_steps": 2189, "loss": 2.3197, "learning_rate": 1.5627530364372469e-07, "epoch": 4.260700389105058 }, { "current_steps": 2190, "loss": 2.2304, "learning_rate": 1.5587044534412956e-07, "epoch": 4.262645914396887 }, { "current_steps": 2191, "loss": 2.0863, "learning_rate": 1.554655870445344e-07, "epoch": 4.264591439688716 }, { "current_steps": 2192, "loss": 2.0632, "learning_rate": 1.5506072874493925e-07, "epoch": 4.266536964980545 }, { "current_steps": 2193, "loss": 2.019, "learning_rate": 1.5465587044534413e-07, "epoch": 4.2684824902723735 }, { "current_steps": 2194, "loss": 2.2063, "learning_rate": 1.5425101214574897e-07, "epoch": 4.270428015564202 }, { "current_steps": 2195, "loss": 2.1838, "learning_rate": 1.5384615384615385e-07, "epoch": 4.272373540856031 }, { "current_steps": 2196, "loss": 1.968, "learning_rate": 1.534412955465587e-07, "epoch": 4.27431906614786 }, { "current_steps": 2197, "loss": 1.9648, "learning_rate": 1.5303643724696354e-07, "epoch": 4.276264591439689 }, { "current_steps": 2198, "loss": 2.2263, "learning_rate": 1.526315789473684e-07, "epoch": 4.278210116731517 }, { "current_steps": 2199, "loss": 1.8421, "learning_rate": 1.5222672064777328e-07, "epoch": 4.280155642023346 }, { "current_steps": 2200, "loss": 1.934, "learning_rate": 1.5182186234817813e-07, "epoch": 4.282101167315175 }, { "current_steps": 2201, "loss": 1.9548, "learning_rate": 1.5141700404858298e-07, "epoch": 4.284046692607004 }, { "current_steps": 2202, "loss": 2.3981, "learning_rate": 1.5101214574898785e-07, "epoch": 4.285992217898833 }, { "current_steps": 2203, "loss": 2.061, "learning_rate": 1.5060728744939272e-07, "epoch": 4.287937743190661 }, { "current_steps": 2204, "loss": 2.1762, "learning_rate": 1.5020242914979754e-07, "epoch": 4.28988326848249 }, { "current_steps": 2205, "loss": 1.9414, "learning_rate": 1.4979757085020242e-07, "epoch": 4.291828793774319 }, { "current_steps": 2206, "loss": 1.9773, "learning_rate": 1.493927125506073e-07, "epoch": 4.293774319066148 }, { "current_steps": 2207, "loss": 2.3973, "learning_rate": 1.4898785425101216e-07, "epoch": 4.2957198443579765 }, { "current_steps": 2208, "loss": 2.1575, "learning_rate": 1.4858299595141698e-07, "epoch": 4.297665369649805 }, { "current_steps": 2209, "loss": 2.0198, "learning_rate": 1.4817813765182186e-07, "epoch": 4.299610894941634 }, { "current_steps": 2210, "loss": 1.5642, "learning_rate": 1.4777327935222673e-07, "epoch": 4.301556420233463 }, { "current_steps": 2211, "loss": 2.2285, "learning_rate": 1.4736842105263155e-07, "epoch": 4.303501945525292 }, { "current_steps": 2212, "loss": 1.3883, "learning_rate": 1.4696356275303642e-07, "epoch": 4.30544747081712 }, { "current_steps": 2213, "loss": 2.2872, "learning_rate": 1.465587044534413e-07, "epoch": 4.307392996108949 }, { "current_steps": 2214, "loss": 1.8314, "learning_rate": 1.4615384615384617e-07, "epoch": 4.309338521400778 }, { "current_steps": 2215, "loss": 2.1769, "learning_rate": 1.45748987854251e-07, "epoch": 4.311284046692607 }, { "current_steps": 2216, "loss": 2.287, "learning_rate": 1.4534412955465586e-07, "epoch": 4.313229571984436 }, { "current_steps": 2217, "loss": 1.8511, "learning_rate": 1.4493927125506073e-07, "epoch": 4.315175097276264 }, { "current_steps": 2218, "loss": 1.2768, "learning_rate": 1.4453441295546558e-07, "epoch": 4.317120622568093 }, { "current_steps": 2219, "loss": 2.1187, "learning_rate": 1.4412955465587043e-07, "epoch": 4.319066147859922 }, { "current_steps": 2220, "loss": 2.0963, "learning_rate": 1.437246963562753e-07, "epoch": 4.321011673151751 }, { "current_steps": 2221, "loss": 1.9288, "learning_rate": 1.4331983805668015e-07, "epoch": 4.3229571984435795 }, { "current_steps": 2222, "loss": 1.6563, "learning_rate": 1.4291497975708502e-07, "epoch": 4.324902723735408 }, { "current_steps": 2223, "loss": 1.7655, "learning_rate": 1.4251012145748987e-07, "epoch": 4.326848249027237 }, { "current_steps": 2224, "loss": 2.2284, "learning_rate": 1.4210526315789474e-07, "epoch": 4.328793774319066 }, { "current_steps": 2225, "loss": 2.2161, "learning_rate": 1.4170040485829959e-07, "epoch": 4.330739299610895 }, { "current_steps": 2226, "loss": 2.1991, "learning_rate": 1.4129554655870446e-07, "epoch": 4.332684824902723 }, { "current_steps": 2227, "loss": 1.4725, "learning_rate": 1.408906882591093e-07, "epoch": 4.334630350194552 }, { "current_steps": 2228, "loss": 2.2272, "learning_rate": 1.4048582995951415e-07, "epoch": 4.336575875486381 }, { "current_steps": 2229, "loss": 2.4142, "learning_rate": 1.4008097165991902e-07, "epoch": 4.33852140077821 }, { "current_steps": 2230, "loss": 1.9919, "learning_rate": 1.396761133603239e-07, "epoch": 4.340466926070039 }, { "current_steps": 2231, "loss": 2.1093, "learning_rate": 1.3927125506072874e-07, "epoch": 4.342412451361867 }, { "current_steps": 2232, "loss": 1.6427, "learning_rate": 1.388663967611336e-07, "epoch": 4.344357976653696 }, { "current_steps": 2233, "loss": 2.0405, "learning_rate": 1.3846153846153846e-07, "epoch": 4.346303501945525 }, { "current_steps": 2234, "loss": 1.904, "learning_rate": 1.380566801619433e-07, "epoch": 4.348249027237354 }, { "current_steps": 2235, "loss": 2.0557, "learning_rate": 1.3765182186234816e-07, "epoch": 4.3501945525291825 }, { "current_steps": 2236, "loss": 2.171, "learning_rate": 1.3724696356275303e-07, "epoch": 4.352140077821011 }, { "current_steps": 2237, "loss": 2.2408, "learning_rate": 1.368421052631579e-07, "epoch": 4.35408560311284 }, { "current_steps": 2238, "loss": 1.8975, "learning_rate": 1.3643724696356272e-07, "epoch": 4.356031128404669 }, { "current_steps": 2239, "loss": 2.2377, "learning_rate": 1.360323886639676e-07, "epoch": 4.357976653696498 }, { "current_steps": 2240, "loss": 1.908, "learning_rate": 1.3562753036437247e-07, "epoch": 4.3599221789883265 }, { "current_steps": 2241, "loss": 1.8576, "learning_rate": 1.3522267206477734e-07, "epoch": 4.361867704280155 }, { "current_steps": 2242, "loss": 2.2841, "learning_rate": 1.3481781376518216e-07, "epoch": 4.363813229571984 }, { "current_steps": 2243, "loss": 2.1605, "learning_rate": 1.3441295546558703e-07, "epoch": 4.365758754863813 }, { "current_steps": 2244, "loss": 1.8862, "learning_rate": 1.340080971659919e-07, "epoch": 4.3677042801556425 }, { "current_steps": 2245, "loss": 2.1633, "learning_rate": 1.3360323886639675e-07, "epoch": 4.369649805447471 }, { "current_steps": 2246, "loss": 1.868, "learning_rate": 1.331983805668016e-07, "epoch": 4.3715953307393 }, { "current_steps": 2247, "loss": 2.5609, "learning_rate": 1.3279352226720647e-07, "epoch": 4.373540856031129 }, { "current_steps": 2248, "loss": 2.1511, "learning_rate": 1.3238866396761135e-07, "epoch": 4.375486381322958 }, { "current_steps": 2249, "loss": 1.7924, "learning_rate": 1.319838056680162e-07, "epoch": 4.377431906614786 }, { "current_steps": 2250, "loss": 1.9348, "learning_rate": 1.3157894736842104e-07, "epoch": 4.379377431906615 }, { "current_steps": 2251, "loss": 2.0349, "learning_rate": 1.311740890688259e-07, "epoch": 4.381322957198444 }, { "current_steps": 2252, "loss": 1.9375, "learning_rate": 1.3076923076923076e-07, "epoch": 4.383268482490273 }, { "current_steps": 2253, "loss": 2.215, "learning_rate": 1.303643724696356e-07, "epoch": 4.385214007782102 }, { "current_steps": 2254, "loss": 2.2971, "learning_rate": 1.2995951417004048e-07, "epoch": 4.38715953307393 }, { "current_steps": 2255, "loss": 2.1454, "learning_rate": 1.2955465587044535e-07, "epoch": 4.389105058365759 }, { "current_steps": 2256, "loss": 2.1493, "learning_rate": 1.291497975708502e-07, "epoch": 4.391050583657588 }, { "current_steps": 2257, "loss": 1.9156, "learning_rate": 1.2874493927125504e-07, "epoch": 4.392996108949417 }, { "current_steps": 2258, "loss": 1.9683, "learning_rate": 1.2834008097165992e-07, "epoch": 4.3949416342412455 }, { "current_steps": 2259, "loss": 2.1256, "learning_rate": 1.2793522267206476e-07, "epoch": 4.396887159533074 }, { "current_steps": 2260, "loss": 2.4656, "learning_rate": 1.2753036437246964e-07, "epoch": 4.398832684824903 }, { "current_steps": 2261, "loss": 1.8926, "learning_rate": 1.2712550607287448e-07, "epoch": 4.400778210116732 }, { "current_steps": 2262, "loss": 2.3056, "learning_rate": 1.2672064777327933e-07, "epoch": 4.402723735408561 }, { "current_steps": 2263, "loss": 2.2891, "learning_rate": 1.263157894736842e-07, "epoch": 4.404669260700389 }, { "current_steps": 2264, "loss": 1.953, "learning_rate": 1.2591093117408908e-07, "epoch": 4.406614785992218 }, { "current_steps": 2265, "loss": 1.9508, "learning_rate": 1.2550607287449392e-07, "epoch": 4.408560311284047 }, { "current_steps": 2266, "loss": 2.1041, "learning_rate": 1.2510121457489877e-07, "epoch": 4.410505836575876 }, { "current_steps": 2267, "loss": 2.2486, "learning_rate": 1.2469635627530364e-07, "epoch": 4.412451361867705 }, { "current_steps": 2268, "loss": 2.0444, "learning_rate": 1.242914979757085e-07, "epoch": 4.414396887159533 }, { "current_steps": 2269, "loss": 2.2233, "learning_rate": 1.2388663967611336e-07, "epoch": 4.416342412451362 }, { "current_steps": 2270, "loss": 2.3264, "learning_rate": 1.234817813765182e-07, "epoch": 4.418287937743191 }, { "current_steps": 2271, "loss": 2.1396, "learning_rate": 1.2307692307692308e-07, "epoch": 4.42023346303502 }, { "current_steps": 2272, "loss": 2.1273, "learning_rate": 1.2267206477732793e-07, "epoch": 4.4221789883268485 }, { "current_steps": 2273, "loss": 2.317, "learning_rate": 1.2226720647773277e-07, "epoch": 4.424124513618677 }, { "current_steps": 2274, "loss": 1.9418, "learning_rate": 1.2186234817813765e-07, "epoch": 4.426070038910506 }, { "current_steps": 2275, "loss": 1.7564, "learning_rate": 1.214574898785425e-07, "epoch": 4.428015564202335 }, { "current_steps": 2276, "loss": 2.2866, "learning_rate": 1.2105263157894737e-07, "epoch": 4.429961089494164 }, { "current_steps": 2277, "loss": 2.0419, "learning_rate": 1.2064777327935221e-07, "epoch": 4.4319066147859925 }, { "current_steps": 2278, "loss": 2.0106, "learning_rate": 1.2024291497975709e-07, "epoch": 4.433852140077821 }, { "current_steps": 2279, "loss": 1.5103, "learning_rate": 1.1983805668016193e-07, "epoch": 4.43579766536965 }, { "current_steps": 2280, "loss": 1.7018, "learning_rate": 1.194331983805668e-07, "epoch": 4.437743190661479 }, { "current_steps": 2281, "loss": 2.2058, "learning_rate": 1.1902834008097165e-07, "epoch": 4.439688715953308 }, { "current_steps": 2282, "loss": 2.0363, "learning_rate": 1.1862348178137651e-07, "epoch": 4.441634241245136 }, { "current_steps": 2283, "loss": 2.1559, "learning_rate": 1.1821862348178137e-07, "epoch": 4.443579766536965 }, { "current_steps": 2284, "loss": 2.382, "learning_rate": 1.1781376518218623e-07, "epoch": 4.445525291828794 }, { "current_steps": 2285, "loss": 2.2512, "learning_rate": 1.1740890688259109e-07, "epoch": 4.447470817120623 }, { "current_steps": 2286, "loss": 2.2323, "learning_rate": 1.1700404858299595e-07, "epoch": 4.4494163424124515 }, { "current_steps": 2287, "loss": 2.0394, "learning_rate": 1.165991902834008e-07, "epoch": 4.45136186770428 }, { "current_steps": 2288, "loss": 2.3223, "learning_rate": 1.1619433198380567e-07, "epoch": 4.453307392996109 }, { "current_steps": 2289, "loss": 1.7242, "learning_rate": 1.1578947368421052e-07, "epoch": 4.455252918287938 }, { "current_steps": 2290, "loss": 1.8005, "learning_rate": 1.1538461538461539e-07, "epoch": 4.457198443579767 }, { "current_steps": 2291, "loss": 2.0514, "learning_rate": 1.1497975708502024e-07, "epoch": 4.4591439688715955 }, { "current_steps": 2292, "loss": 2.0549, "learning_rate": 1.1457489878542508e-07, "epoch": 4.461089494163424 }, { "current_steps": 2293, "loss": 1.9331, "learning_rate": 1.1417004048582996e-07, "epoch": 4.463035019455253 }, { "current_steps": 2294, "loss": 2.1634, "learning_rate": 1.137651821862348e-07, "epoch": 4.464980544747082 }, { "current_steps": 2295, "loss": 2.1563, "learning_rate": 1.1336032388663968e-07, "epoch": 4.466926070038911 }, { "current_steps": 2296, "loss": 2.1333, "learning_rate": 1.1295546558704452e-07, "epoch": 4.468871595330739 }, { "current_steps": 2297, "loss": 1.9183, "learning_rate": 1.125506072874494e-07, "epoch": 4.470817120622568 }, { "current_steps": 2298, "loss": 2.001, "learning_rate": 1.1214574898785424e-07, "epoch": 4.472762645914397 }, { "current_steps": 2299, "loss": 1.8586, "learning_rate": 1.117408906882591e-07, "epoch": 4.474708171206226 }, { "current_steps": 2300, "loss": 1.9263, "learning_rate": 1.1133603238866396e-07, "epoch": 4.476653696498055 }, { "current_steps": 2301, "loss": 2.0239, "learning_rate": 1.1093117408906882e-07, "epoch": 4.478599221789883 }, { "current_steps": 2302, "loss": 2.0846, "learning_rate": 1.1052631578947368e-07, "epoch": 4.480544747081712 }, { "current_steps": 2303, "loss": 2.166, "learning_rate": 1.1012145748987854e-07, "epoch": 4.482490272373541 }, { "current_steps": 2304, "loss": 1.4213, "learning_rate": 1.0971659919028339e-07, "epoch": 4.48443579766537 }, { "current_steps": 2305, "loss": 2.0831, "learning_rate": 1.0931174089068826e-07, "epoch": 4.4863813229571985 }, { "current_steps": 2306, "loss": 2.066, "learning_rate": 1.0890688259109311e-07, "epoch": 4.488326848249027 }, { "current_steps": 2307, "loss": 2.1141, "learning_rate": 1.0850202429149798e-07, "epoch": 4.490272373540856 }, { "current_steps": 2308, "loss": 1.8931, "learning_rate": 1.0809716599190283e-07, "epoch": 4.492217898832685 }, { "current_steps": 2309, "loss": 2.2313, "learning_rate": 1.076923076923077e-07, "epoch": 4.494163424124514 }, { "current_steps": 2310, "loss": 2.2087, "learning_rate": 1.0728744939271255e-07, "epoch": 4.496108949416342 }, { "current_steps": 2311, "loss": 1.9752, "learning_rate": 1.068825910931174e-07, "epoch": 4.498054474708171 }, { "current_steps": 2312, "loss": 2.282, "learning_rate": 1.0647773279352227e-07, "epoch": 4.5 }, { "current_steps": 2313, "loss": 2.3254, "learning_rate": 1.0607287449392711e-07, "epoch": 4.501945525291829 }, { "current_steps": 2314, "loss": 2.2425, "learning_rate": 1.0566801619433199e-07, "epoch": 4.503891050583658 }, { "current_steps": 2315, "loss": 2.0981, "learning_rate": 1.0526315789473683e-07, "epoch": 4.505836575875486 }, { "current_steps": 2316, "loss": 1.9853, "learning_rate": 1.0485829959514169e-07, "epoch": 4.507782101167315 }, { "current_steps": 2317, "loss": 2.2734, "learning_rate": 1.0445344129554655e-07, "epoch": 4.509727626459144 }, { "current_steps": 2318, "loss": 1.931, "learning_rate": 1.0404858299595141e-07, "epoch": 4.511673151750973 }, { "current_steps": 2319, "loss": 2.0312, "learning_rate": 1.0364372469635627e-07, "epoch": 4.5136186770428015 }, { "current_steps": 2320, "loss": 2.0899, "learning_rate": 1.0323886639676113e-07, "epoch": 4.51556420233463 }, { "current_steps": 2321, "loss": 1.9136, "learning_rate": 1.0283400809716598e-07, "epoch": 4.517509727626459 }, { "current_steps": 2322, "loss": 2.1124, "learning_rate": 1.0242914979757085e-07, "epoch": 4.519455252918288 }, { "current_steps": 2323, "loss": 2.0532, "learning_rate": 1.020242914979757e-07, "epoch": 4.521400778210117 }, { "current_steps": 2324, "loss": 2.3959, "learning_rate": 1.0161943319838057e-07, "epoch": 4.523346303501945 }, { "current_steps": 2325, "loss": 2.1902, "learning_rate": 1.0121457489878542e-07, "epoch": 4.525291828793774 }, { "current_steps": 2326, "loss": 1.7862, "learning_rate": 1.0080971659919029e-07, "epoch": 4.527237354085603 }, { "current_steps": 2327, "loss": 2.2113, "learning_rate": 1.0040485829959514e-07, "epoch": 4.529182879377432 }, { "current_steps": 2328, "loss": 2.2121, "learning_rate": 1e-07, "epoch": 4.531128404669261 }, { "current_steps": 2329, "loss": 2.1231, "learning_rate": 9.959514170040486e-08, "epoch": 4.533073929961089 }, { "current_steps": 2330, "loss": 2.0081, "learning_rate": 9.919028340080972e-08, "epoch": 4.535019455252918 }, { "current_steps": 2331, "loss": 1.9696, "learning_rate": 9.878542510121457e-08, "epoch": 4.536964980544747 }, { "current_steps": 2332, "loss": 2.1453, "learning_rate": 9.838056680161943e-08, "epoch": 4.538910505836576 }, { "current_steps": 2333, "loss": 1.7669, "learning_rate": 9.797570850202428e-08, "epoch": 4.5408560311284045 }, { "current_steps": 2334, "loss": 2.0061, "learning_rate": 9.757085020242914e-08, "epoch": 4.542801556420233 }, { "current_steps": 2335, "loss": 1.8673, "learning_rate": 9.7165991902834e-08, "epoch": 4.544747081712062 }, { "current_steps": 2336, "loss": 2.2979, "learning_rate": 9.676113360323886e-08, "epoch": 4.546692607003891 }, { "current_steps": 2337, "loss": 2.1828, "learning_rate": 9.635627530364372e-08, "epoch": 4.54863813229572 }, { "current_steps": 2338, "loss": 1.7778, "learning_rate": 9.595141700404858e-08, "epoch": 4.5505836575875485 }, { "current_steps": 2339, "loss": 2.282, "learning_rate": 9.554655870445344e-08, "epoch": 4.552529182879377 }, { "current_steps": 2340, "loss": 2.1357, "learning_rate": 9.514170040485829e-08, "epoch": 4.554474708171206 }, { "current_steps": 2341, "loss": 1.7649, "learning_rate": 9.473684210526316e-08, "epoch": 4.556420233463035 }, { "current_steps": 2342, "loss": 2.0922, "learning_rate": 9.4331983805668e-08, "epoch": 4.558365758754864 }, { "current_steps": 2343, "loss": 2.2921, "learning_rate": 9.392712550607288e-08, "epoch": 4.560311284046692 }, { "current_steps": 2344, "loss": 2.0862, "learning_rate": 9.352226720647773e-08, "epoch": 4.562256809338521 }, { "current_steps": 2345, "loss": 2.3573, "learning_rate": 9.311740890688259e-08, "epoch": 4.56420233463035 }, { "current_steps": 2346, "loss": 2.2289, "learning_rate": 9.271255060728744e-08, "epoch": 4.566147859922179 }, { "current_steps": 2347, "loss": 2.3357, "learning_rate": 9.23076923076923e-08, "epoch": 4.5680933852140075 }, { "current_steps": 2348, "loss": 2.1622, "learning_rate": 9.190283400809716e-08, "epoch": 4.570038910505836 }, { "current_steps": 2349, "loss": 2.3764, "learning_rate": 9.149797570850202e-08, "epoch": 4.571984435797665 }, { "current_steps": 2350, "loss": 2.2479, "learning_rate": 9.109311740890688e-08, "epoch": 4.573929961089494 }, { "current_steps": 2351, "loss": 2.012, "learning_rate": 9.068825910931174e-08, "epoch": 4.575875486381323 }, { "current_steps": 2352, "loss": 2.3346, "learning_rate": 9.028340080971659e-08, "epoch": 4.5778210116731515 }, { "current_steps": 2353, "loss": 2.393, "learning_rate": 8.987854251012145e-08, "epoch": 4.57976653696498 }, { "current_steps": 2354, "loss": 2.1819, "learning_rate": 8.947368421052631e-08, "epoch": 4.581712062256809 }, { "current_steps": 2355, "loss": 2.2099, "learning_rate": 8.906882591093117e-08, "epoch": 4.583657587548638 }, { "current_steps": 2356, "loss": 2.2421, "learning_rate": 8.866396761133603e-08, "epoch": 4.585603112840467 }, { "current_steps": 2357, "loss": 2.0111, "learning_rate": 8.825910931174088e-08, "epoch": 4.587548638132295 }, { "current_steps": 2358, "loss": 2.1729, "learning_rate": 8.785425101214575e-08, "epoch": 4.589494163424124 }, { "current_steps": 2359, "loss": 2.4654, "learning_rate": 8.74493927125506e-08, "epoch": 4.591439688715953 }, { "current_steps": 2360, "loss": 1.9915, "learning_rate": 8.704453441295547e-08, "epoch": 4.593385214007782 }, { "current_steps": 2361, "loss": 2.026, "learning_rate": 8.663967611336031e-08, "epoch": 4.595330739299611 }, { "current_steps": 2362, "loss": 2.2874, "learning_rate": 8.623481781376519e-08, "epoch": 4.597276264591439 }, { "current_steps": 2363, "loss": 2.0729, "learning_rate": 8.582995951417003e-08, "epoch": 4.599221789883268 }, { "current_steps": 2364, "loss": 2.0545, "learning_rate": 8.54251012145749e-08, "epoch": 4.601167315175097 }, { "current_steps": 2365, "loss": 2.0677, "learning_rate": 8.502024291497975e-08, "epoch": 4.603112840466926 }, { "current_steps": 2366, "loss": 2.1988, "learning_rate": 8.461538461538461e-08, "epoch": 4.6050583657587545 }, { "current_steps": 2367, "loss": 1.4236, "learning_rate": 8.421052631578947e-08, "epoch": 4.607003891050583 }, { "current_steps": 2368, "loss": 2.3124, "learning_rate": 8.380566801619433e-08, "epoch": 4.608949416342412 }, { "current_steps": 2369, "loss": 2.0258, "learning_rate": 8.340080971659918e-08, "epoch": 4.610894941634241 }, { "current_steps": 2370, "loss": 2.331, "learning_rate": 8.299595141700405e-08, "epoch": 4.61284046692607 }, { "current_steps": 2371, "loss": 2.0216, "learning_rate": 8.25910931174089e-08, "epoch": 4.614785992217898 }, { "current_steps": 2372, "loss": 2.0977, "learning_rate": 8.218623481781377e-08, "epoch": 4.616731517509727 }, { "current_steps": 2373, "loss": 1.9623, "learning_rate": 8.178137651821862e-08, "epoch": 4.618677042801556 }, { "current_steps": 2374, "loss": 2.0379, "learning_rate": 8.137651821862347e-08, "epoch": 4.620622568093385 }, { "current_steps": 2375, "loss": 1.8241, "learning_rate": 8.097165991902834e-08, "epoch": 4.622568093385214 }, { "current_steps": 2376, "loss": 2.4894, "learning_rate": 8.056680161943318e-08, "epoch": 4.624513618677042 }, { "current_steps": 2377, "loss": 2.3934, "learning_rate": 8.016194331983806e-08, "epoch": 4.626459143968871 }, { "current_steps": 2378, "loss": 1.8943, "learning_rate": 7.97570850202429e-08, "epoch": 4.6284046692607 }, { "current_steps": 2379, "loss": 2.0764, "learning_rate": 7.935222672064778e-08, "epoch": 4.630350194552529 }, { "current_steps": 2380, "loss": 2.1584, "learning_rate": 7.894736842105262e-08, "epoch": 4.6322957198443575 }, { "current_steps": 2381, "loss": 2.1429, "learning_rate": 7.854251012145748e-08, "epoch": 4.634241245136186 }, { "current_steps": 2382, "loss": 1.9538, "learning_rate": 7.813765182186234e-08, "epoch": 4.636186770428015 }, { "current_steps": 2383, "loss": 2.4872, "learning_rate": 7.77327935222672e-08, "epoch": 4.638132295719844 }, { "current_steps": 2384, "loss": 2.4161, "learning_rate": 7.732793522267206e-08, "epoch": 4.640077821011673 }, { "current_steps": 2385, "loss": 2.2889, "learning_rate": 7.692307692307692e-08, "epoch": 4.642023346303502 }, { "current_steps": 2386, "loss": 2.1131, "learning_rate": 7.651821862348177e-08, "epoch": 4.643968871595331 }, { "current_steps": 2387, "loss": 1.9476, "learning_rate": 7.611336032388664e-08, "epoch": 4.64591439688716 }, { "current_steps": 2388, "loss": 2.0714, "learning_rate": 7.570850202429149e-08, "epoch": 4.647859922178989 }, { "current_steps": 2389, "loss": 2.2547, "learning_rate": 7.530364372469636e-08, "epoch": 4.6498054474708175 }, { "current_steps": 2390, "loss": 2.1798, "learning_rate": 7.489878542510121e-08, "epoch": 4.651750972762646 }, { "current_steps": 2391, "loss": 1.7496, "learning_rate": 7.449392712550608e-08, "epoch": 4.653696498054475 }, { "current_steps": 2392, "loss": 1.954, "learning_rate": 7.408906882591093e-08, "epoch": 4.655642023346304 }, { "current_steps": 2393, "loss": 2.2402, "learning_rate": 7.368421052631577e-08, "epoch": 4.657587548638133 }, { "current_steps": 2394, "loss": 2.1416, "learning_rate": 7.327935222672065e-08, "epoch": 4.659533073929961 }, { "current_steps": 2395, "loss": 2.4773, "learning_rate": 7.28744939271255e-08, "epoch": 4.66147859922179 }, { "current_steps": 2396, "loss": 2.2116, "learning_rate": 7.246963562753037e-08, "epoch": 4.663424124513619 }, { "current_steps": 2397, "loss": 2.1067, "learning_rate": 7.206477732793521e-08, "epoch": 4.665369649805448 }, { "current_steps": 2398, "loss": 2.1138, "learning_rate": 7.165991902834007e-08, "epoch": 4.667315175097277 }, { "current_steps": 2399, "loss": 1.9094, "learning_rate": 7.125506072874493e-08, "epoch": 4.669260700389105 }, { "current_steps": 2400, "loss": 2.0819, "learning_rate": 7.085020242914979e-08, "epoch": 4.671206225680934 }, { "current_steps": 2401, "loss": 2.3377, "learning_rate": 7.044534412955465e-08, "epoch": 4.673151750972763 }, { "current_steps": 2402, "loss": 1.9866, "learning_rate": 7.004048582995951e-08, "epoch": 4.675097276264592 }, { "current_steps": 2403, "loss": 1.9613, "learning_rate": 6.963562753036437e-08, "epoch": 4.6770428015564205 }, { "current_steps": 2404, "loss": 2.1234, "learning_rate": 6.923076923076923e-08, "epoch": 4.678988326848249 }, { "current_steps": 2405, "loss": 1.4866, "learning_rate": 6.882591093117408e-08, "epoch": 4.680933852140078 }, { "current_steps": 2406, "loss": 1.721, "learning_rate": 6.842105263157895e-08, "epoch": 4.682879377431907 }, { "current_steps": 2407, "loss": 1.8424, "learning_rate": 6.80161943319838e-08, "epoch": 4.684824902723736 }, { "current_steps": 2408, "loss": 2.1204, "learning_rate": 6.761133603238867e-08, "epoch": 4.686770428015564 }, { "current_steps": 2409, "loss": 2.0779, "learning_rate": 6.720647773279352e-08, "epoch": 4.688715953307393 }, { "current_steps": 2410, "loss": 2.0557, "learning_rate": 6.680161943319838e-08, "epoch": 4.690661478599222 }, { "current_steps": 2411, "loss": 1.1398, "learning_rate": 6.639676113360324e-08, "epoch": 4.692607003891051 }, { "current_steps": 2412, "loss": 2.2897, "learning_rate": 6.59919028340081e-08, "epoch": 4.69455252918288 }, { "current_steps": 2413, "loss": 1.9499, "learning_rate": 6.558704453441296e-08, "epoch": 4.696498054474708 }, { "current_steps": 2414, "loss": 2.153, "learning_rate": 6.51821862348178e-08, "epoch": 4.698443579766537 }, { "current_steps": 2415, "loss": 1.6132, "learning_rate": 6.477732793522268e-08, "epoch": 4.700389105058366 }, { "current_steps": 2416, "loss": 1.872, "learning_rate": 6.437246963562752e-08, "epoch": 4.702334630350195 }, { "current_steps": 2417, "loss": 2.0976, "learning_rate": 6.396761133603238e-08, "epoch": 4.7042801556420235 }, { "current_steps": 2418, "loss": 1.9191, "learning_rate": 6.356275303643724e-08, "epoch": 4.706225680933852 }, { "current_steps": 2419, "loss": 2.3675, "learning_rate": 6.31578947368421e-08, "epoch": 4.708171206225681 }, { "current_steps": 2420, "loss": 2.253, "learning_rate": 6.275303643724696e-08, "epoch": 4.71011673151751 }, { "current_steps": 2421, "loss": 1.9531, "learning_rate": 6.234817813765182e-08, "epoch": 4.712062256809339 }, { "current_steps": 2422, "loss": 2.2032, "learning_rate": 6.194331983805668e-08, "epoch": 4.714007782101167 }, { "current_steps": 2423, "loss": 1.9182, "learning_rate": 6.153846153846154e-08, "epoch": 4.715953307392996 }, { "current_steps": 2424, "loss": 1.8969, "learning_rate": 6.113360323886639e-08, "epoch": 4.717898832684825 }, { "current_steps": 2425, "loss": 2.0649, "learning_rate": 6.072874493927125e-08, "epoch": 4.719844357976654 }, { "current_steps": 2426, "loss": 1.9386, "learning_rate": 6.032388663967611e-08, "epoch": 4.721789883268483 }, { "current_steps": 2427, "loss": 2.2879, "learning_rate": 5.991902834008097e-08, "epoch": 4.723735408560311 }, { "current_steps": 2428, "loss": 1.7753, "learning_rate": 5.9514170040485826e-08, "epoch": 4.72568093385214 }, { "current_steps": 2429, "loss": 2.067, "learning_rate": 5.9109311740890686e-08, "epoch": 4.727626459143969 }, { "current_steps": 2430, "loss": 2.003, "learning_rate": 5.8704453441295546e-08, "epoch": 4.729571984435798 }, { "current_steps": 2431, "loss": 1.7233, "learning_rate": 5.82995951417004e-08, "epoch": 4.7315175097276265 }, { "current_steps": 2432, "loss": 2.2434, "learning_rate": 5.789473684210526e-08, "epoch": 4.733463035019455 }, { "current_steps": 2433, "loss": 1.8563, "learning_rate": 5.748987854251012e-08, "epoch": 4.735408560311284 }, { "current_steps": 2434, "loss": 2.2889, "learning_rate": 5.708502024291498e-08, "epoch": 4.737354085603113 }, { "current_steps": 2435, "loss": 2.1407, "learning_rate": 5.668016194331984e-08, "epoch": 4.739299610894942 }, { "current_steps": 2436, "loss": 2.3897, "learning_rate": 5.62753036437247e-08, "epoch": 4.7412451361867705 }, { "current_steps": 2437, "loss": 2.3365, "learning_rate": 5.587044534412955e-08, "epoch": 4.743190661478599 }, { "current_steps": 2438, "loss": 2.2225, "learning_rate": 5.546558704453441e-08, "epoch": 4.745136186770428 }, { "current_steps": 2439, "loss": 2.1271, "learning_rate": 5.506072874493927e-08, "epoch": 4.747081712062257 }, { "current_steps": 2440, "loss": 1.5352, "learning_rate": 5.465587044534413e-08, "epoch": 4.749027237354086 }, { "current_steps": 2441, "loss": 2.0332, "learning_rate": 5.425101214574899e-08, "epoch": 4.750972762645914 }, { "current_steps": 2442, "loss": 2.1578, "learning_rate": 5.384615384615385e-08, "epoch": 4.752918287937743 }, { "current_steps": 2443, "loss": 2.0057, "learning_rate": 5.34412955465587e-08, "epoch": 4.754863813229572 }, { "current_steps": 2444, "loss": 1.3528, "learning_rate": 5.3036437246963556e-08, "epoch": 4.756809338521401 }, { "current_steps": 2445, "loss": 2.1213, "learning_rate": 5.2631578947368416e-08, "epoch": 4.7587548638132295 }, { "current_steps": 2446, "loss": 2.0805, "learning_rate": 5.2226720647773276e-08, "epoch": 4.760700389105058 }, { "current_steps": 2447, "loss": 1.3012, "learning_rate": 5.1821862348178135e-08, "epoch": 4.762645914396887 }, { "current_steps": 2448, "loss": 1.6506, "learning_rate": 5.141700404858299e-08, "epoch": 4.764591439688716 }, { "current_steps": 2449, "loss": 2.0657, "learning_rate": 5.101214574898785e-08, "epoch": 4.766536964980545 }, { "current_steps": 2450, "loss": 2.2499, "learning_rate": 5.060728744939271e-08, "epoch": 4.7684824902723735 }, { "current_steps": 2451, "loss": 2.0278, "learning_rate": 5.020242914979757e-08, "epoch": 4.770428015564202 }, { "current_steps": 2452, "loss": 2.3026, "learning_rate": 4.979757085020243e-08, "epoch": 4.772373540856031 }, { "current_steps": 2453, "loss": 2.3466, "learning_rate": 4.939271255060729e-08, "epoch": 4.77431906614786 }, { "current_steps": 2454, "loss": 1.9283, "learning_rate": 4.898785425101214e-08, "epoch": 4.776264591439689 }, { "current_steps": 2455, "loss": 2.1553, "learning_rate": 4.8582995951417e-08, "epoch": 4.778210116731517 }, { "current_steps": 2456, "loss": 1.9607, "learning_rate": 4.817813765182186e-08, "epoch": 4.780155642023346 }, { "current_steps": 2457, "loss": 2.0503, "learning_rate": 4.777327935222672e-08, "epoch": 4.782101167315175 }, { "current_steps": 2458, "loss": 2.2186, "learning_rate": 4.736842105263158e-08, "epoch": 4.784046692607004 }, { "current_steps": 2459, "loss": 2.1438, "learning_rate": 4.696356275303644e-08, "epoch": 4.785992217898833 }, { "current_steps": 2460, "loss": 2.3328, "learning_rate": 4.655870445344129e-08, "epoch": 4.787937743190661 }, { "current_steps": 2461, "loss": 2.2212, "learning_rate": 4.615384615384615e-08, "epoch": 4.78988326848249 }, { "current_steps": 2462, "loss": 1.5952, "learning_rate": 4.574898785425101e-08, "epoch": 4.791828793774319 }, { "current_steps": 2463, "loss": 2.3791, "learning_rate": 4.534412955465587e-08, "epoch": 4.793774319066148 }, { "current_steps": 2464, "loss": 2.1287, "learning_rate": 4.4939271255060725e-08, "epoch": 4.7957198443579765 }, { "current_steps": 2465, "loss": 2.1007, "learning_rate": 4.4534412955465585e-08, "epoch": 4.797665369649805 }, { "current_steps": 2466, "loss": 2.1427, "learning_rate": 4.412955465587044e-08, "epoch": 4.799610894941634 }, { "current_steps": 2467, "loss": 2.0278, "learning_rate": 4.37246963562753e-08, "epoch": 4.801556420233463 }, { "current_steps": 2468, "loss": 2.062, "learning_rate": 4.331983805668016e-08, "epoch": 4.803501945525292 }, { "current_steps": 2469, "loss": 1.9172, "learning_rate": 4.291497975708502e-08, "epoch": 4.80544747081712 }, { "current_steps": 2470, "loss": 2.1843, "learning_rate": 4.251012145748988e-08, "epoch": 4.807392996108949 }, { "current_steps": 2471, "loss": 2.1835, "learning_rate": 4.2105263157894737e-08, "epoch": 4.809338521400778 }, { "current_steps": 2472, "loss": 2.2789, "learning_rate": 4.170040485829959e-08, "epoch": 4.811284046692607 }, { "current_steps": 2473, "loss": 2.1742, "learning_rate": 4.129554655870445e-08, "epoch": 4.813229571984436 }, { "current_steps": 2474, "loss": 1.7106, "learning_rate": 4.089068825910931e-08, "epoch": 4.815175097276264 }, { "current_steps": 2475, "loss": 2.2121, "learning_rate": 4.048582995951417e-08, "epoch": 4.817120622568093 }, { "current_steps": 2476, "loss": 1.8937, "learning_rate": 4.008097165991903e-08, "epoch": 4.819066147859922 }, { "current_steps": 2477, "loss": 2.3253, "learning_rate": 3.967611336032389e-08, "epoch": 4.821011673151751 }, { "current_steps": 2478, "loss": 1.7619, "learning_rate": 3.927125506072874e-08, "epoch": 4.8229571984435795 }, { "current_steps": 2479, "loss": 1.9291, "learning_rate": 3.88663967611336e-08, "epoch": 4.824902723735408 }, { "current_steps": 2480, "loss": 2.0991, "learning_rate": 3.846153846153846e-08, "epoch": 4.826848249027237 }, { "current_steps": 2481, "loss": 1.6111, "learning_rate": 3.805668016194332e-08, "epoch": 4.828793774319066 }, { "current_steps": 2482, "loss": 1.9646, "learning_rate": 3.765182186234818e-08, "epoch": 4.830739299610895 }, { "current_steps": 2483, "loss": 2.1193, "learning_rate": 3.724696356275304e-08, "epoch": 4.832684824902723 }, { "current_steps": 2484, "loss": 1.8468, "learning_rate": 3.684210526315789e-08, "epoch": 4.834630350194552 }, { "current_steps": 2485, "loss": 1.9254, "learning_rate": 3.643724696356275e-08, "epoch": 4.836575875486381 }, { "current_steps": 2486, "loss": 2.1202, "learning_rate": 3.6032388663967607e-08, "epoch": 4.83852140077821 }, { "current_steps": 2487, "loss": 1.8462, "learning_rate": 3.5627530364372466e-08, "epoch": 4.840466926070039 }, { "current_steps": 2488, "loss": 2.2435, "learning_rate": 3.5222672064777326e-08, "epoch": 4.842412451361867 }, { "current_steps": 2489, "loss": 2.296, "learning_rate": 3.4817813765182186e-08, "epoch": 4.844357976653696 }, { "current_steps": 2490, "loss": 2.0567, "learning_rate": 3.441295546558704e-08, "epoch": 4.846303501945525 }, { "current_steps": 2491, "loss": 1.4528, "learning_rate": 3.40080971659919e-08, "epoch": 4.848249027237354 }, { "current_steps": 2492, "loss": 2.1352, "learning_rate": 3.360323886639676e-08, "epoch": 4.8501945525291825 }, { "current_steps": 2493, "loss": 2.283, "learning_rate": 3.319838056680162e-08, "epoch": 4.852140077821011 }, { "current_steps": 2494, "loss": 2.2751, "learning_rate": 3.279352226720648e-08, "epoch": 4.85408560311284 }, { "current_steps": 2495, "loss": 2.1067, "learning_rate": 3.238866396761134e-08, "epoch": 4.856031128404669 }, { "current_steps": 2496, "loss": 2.3014, "learning_rate": 3.198380566801619e-08, "epoch": 4.857976653696498 }, { "current_steps": 2497, "loss": 2.4211, "learning_rate": 3.157894736842105e-08, "epoch": 4.859922178988327 }, { "current_steps": 2498, "loss": 2.0251, "learning_rate": 3.117408906882591e-08, "epoch": 4.861867704280156 }, { "current_steps": 2499, "loss": 2.1741, "learning_rate": 3.076923076923077e-08, "epoch": 4.863813229571985 }, { "current_steps": 2500, "loss": 2.4396, "learning_rate": 3.0364372469635624e-08, "epoch": 4.865758754863814 }, { "current_steps": 2501, "loss": 2.1873, "learning_rate": 2.995951417004048e-08, "epoch": 4.8677042801556425 }, { "current_steps": 2502, "loss": 2.0155, "learning_rate": 2.9554655870445343e-08, "epoch": 4.869649805447471 }, { "current_steps": 2503, "loss": 1.7113, "learning_rate": 2.91497975708502e-08, "epoch": 4.8715953307393 }, { "current_steps": 2504, "loss": 1.9568, "learning_rate": 2.874493927125506e-08, "epoch": 4.873540856031129 }, { "current_steps": 2505, "loss": 2.168, "learning_rate": 2.834008097165992e-08, "epoch": 4.875486381322958 }, { "current_steps": 2506, "loss": 1.931, "learning_rate": 2.7935222672064776e-08, "epoch": 4.877431906614786 }, { "current_steps": 2507, "loss": 1.7827, "learning_rate": 2.7530364372469635e-08, "epoch": 4.879377431906615 }, { "current_steps": 2508, "loss": 2.0304, "learning_rate": 2.7125506072874495e-08, "epoch": 4.881322957198444 }, { "current_steps": 2509, "loss": 2.088, "learning_rate": 2.672064777327935e-08, "epoch": 4.883268482490273 }, { "current_steps": 2510, "loss": 1.9037, "learning_rate": 2.6315789473684208e-08, "epoch": 4.885214007782102 }, { "current_steps": 2511, "loss": 1.7314, "learning_rate": 2.5910931174089068e-08, "epoch": 4.88715953307393 }, { "current_steps": 2512, "loss": 1.6674, "learning_rate": 2.5506072874493924e-08, "epoch": 4.889105058365759 }, { "current_steps": 2513, "loss": 2.2153, "learning_rate": 2.5101214574898784e-08, "epoch": 4.891050583657588 }, { "current_steps": 2514, "loss": 2.1141, "learning_rate": 2.4696356275303644e-08, "epoch": 4.892996108949417 }, { "current_steps": 2515, "loss": 2.0337, "learning_rate": 2.42914979757085e-08, "epoch": 4.8949416342412455 }, { "current_steps": 2516, "loss": 2.2302, "learning_rate": 2.388663967611336e-08, "epoch": 4.896887159533074 }, { "current_steps": 2517, "loss": 2.3364, "learning_rate": 2.348178137651822e-08, "epoch": 4.898832684824903 }, { "current_steps": 2518, "loss": 1.9478, "learning_rate": 2.3076923076923076e-08, "epoch": 4.900778210116732 }, { "current_steps": 2519, "loss": 1.9851, "learning_rate": 2.2672064777327936e-08, "epoch": 4.902723735408561 }, { "current_steps": 2520, "loss": 2.1556, "learning_rate": 2.2267206477732792e-08, "epoch": 4.904669260700389 }, { "current_steps": 2521, "loss": 1.8649, "learning_rate": 2.186234817813765e-08, "epoch": 4.906614785992218 }, { "current_steps": 2522, "loss": 1.7711, "learning_rate": 2.145748987854251e-08, "epoch": 4.908560311284047 }, { "current_steps": 2523, "loss": 2.2542, "learning_rate": 2.1052631578947368e-08, "epoch": 4.910505836575876 }, { "current_steps": 2524, "loss": 1.8253, "learning_rate": 2.0647773279352225e-08, "epoch": 4.912451361867705 }, { "current_steps": 2525, "loss": 2.1628, "learning_rate": 2.0242914979757085e-08, "epoch": 4.914396887159533 }, { "current_steps": 2526, "loss": 2.2028, "learning_rate": 1.9838056680161944e-08, "epoch": 4.916342412451362 }, { "current_steps": 2527, "loss": 2.0165, "learning_rate": 1.94331983805668e-08, "epoch": 4.918287937743191 }, { "current_steps": 2528, "loss": 2.1796, "learning_rate": 1.902834008097166e-08, "epoch": 4.92023346303502 }, { "current_steps": 2529, "loss": 2.3287, "learning_rate": 1.862348178137652e-08, "epoch": 4.9221789883268485 }, { "current_steps": 2530, "loss": 2.318, "learning_rate": 1.8218623481781373e-08, "epoch": 4.924124513618677 }, { "current_steps": 2531, "loss": 2.2234, "learning_rate": 1.7813765182186233e-08, "epoch": 4.926070038910506 }, { "current_steps": 2532, "loss": 2.1367, "learning_rate": 1.7408906882591093e-08, "epoch": 4.928015564202335 }, { "current_steps": 2533, "loss": 2.1121, "learning_rate": 1.700404858299595e-08, "epoch": 4.929961089494164 }, { "current_steps": 2534, "loss": 2.1045, "learning_rate": 1.659919028340081e-08, "epoch": 4.9319066147859925 }, { "current_steps": 2535, "loss": 2.0635, "learning_rate": 1.619433198380567e-08, "epoch": 4.933852140077821 }, { "current_steps": 2536, "loss": 2.267, "learning_rate": 1.5789473684210525e-08, "epoch": 4.93579766536965 }, { "current_steps": 2537, "loss": 2.0892, "learning_rate": 1.5384615384615385e-08, "epoch": 4.937743190661479 }, { "current_steps": 2538, "loss": 2.0939, "learning_rate": 1.497975708502024e-08, "epoch": 4.939688715953308 }, { "current_steps": 2539, "loss": 2.3492, "learning_rate": 1.45748987854251e-08, "epoch": 4.941634241245136 }, { "current_steps": 2540, "loss": 2.1248, "learning_rate": 1.417004048582996e-08, "epoch": 4.943579766536965 }, { "current_steps": 2541, "loss": 2.403, "learning_rate": 1.3765182186234818e-08, "epoch": 4.945525291828794 }, { "current_steps": 2542, "loss": 2.2336, "learning_rate": 1.3360323886639676e-08, "epoch": 4.947470817120623 }, { "current_steps": 2543, "loss": 2.0172, "learning_rate": 1.2955465587044534e-08, "epoch": 4.9494163424124515 }, { "current_steps": 2544, "loss": 2.0328, "learning_rate": 1.2550607287449392e-08, "epoch": 4.95136186770428 }, { "current_steps": 2545, "loss": 1.9879, "learning_rate": 1.214574898785425e-08, "epoch": 4.953307392996109 }, { "current_steps": 2546, "loss": 2.0842, "learning_rate": 1.174089068825911e-08, "epoch": 4.955252918287938 }, { "current_steps": 2547, "loss": 2.1042, "learning_rate": 1.1336032388663968e-08, "epoch": 4.957198443579767 }, { "current_steps": 2548, "loss": 2.212, "learning_rate": 1.0931174089068824e-08, "epoch": 4.9591439688715955 }, { "current_steps": 2549, "loss": 1.893, "learning_rate": 1.0526315789473684e-08, "epoch": 4.961089494163424 }, { "current_steps": 2550, "loss": 2.0078, "learning_rate": 1.0121457489878542e-08, "epoch": 4.963035019455253 }, { "current_steps": 2551, "loss": 2.1154, "learning_rate": 9.7165991902834e-09, "epoch": 4.964980544747082 }, { "current_steps": 2552, "loss": 2.1022, "learning_rate": 9.31174089068826e-09, "epoch": 4.966926070038911 }, { "current_steps": 2553, "loss": 1.9507, "learning_rate": 8.906882591093117e-09, "epoch": 4.968871595330739 }, { "current_steps": 2554, "loss": 2.1817, "learning_rate": 8.502024291497975e-09, "epoch": 4.970817120622568 }, { "current_steps": 2555, "loss": 2.0657, "learning_rate": 8.097165991902834e-09, "epoch": 4.972762645914397 }, { "current_steps": 2556, "loss": 2.0101, "learning_rate": 7.692307692307693e-09, "epoch": 4.974708171206226 }, { "current_steps": 2557, "loss": 2.3433, "learning_rate": 7.28744939271255e-09, "epoch": 4.976653696498055 }, { "current_steps": 2558, "loss": 1.5427, "learning_rate": 6.882591093117409e-09, "epoch": 4.978599221789883 }, { "current_steps": 2559, "loss": 2.3658, "learning_rate": 6.477732793522267e-09, "epoch": 4.980544747081712 }, { "current_steps": 2560, "loss": 1.6504, "learning_rate": 6.072874493927125e-09, "epoch": 4.982490272373541 }, { "current_steps": 2561, "loss": 2.074, "learning_rate": 5.668016194331984e-09, "epoch": 4.98443579766537 }, { "current_steps": 2562, "loss": 2.2611, "learning_rate": 5.263157894736842e-09, "epoch": 4.9863813229571985 }, { "current_steps": 2563, "loss": 2.1333, "learning_rate": 4.8582995951417e-09, "epoch": 4.988326848249027 }, { "current_steps": 2564, "loss": 2.1105, "learning_rate": 4.453441295546558e-09, "epoch": 4.990272373540856 }, { "current_steps": 2565, "loss": 2.0117, "learning_rate": 4.048582995951417e-09, "epoch": 4.992217898832685 }, { "current_steps": 2566, "loss": 1.7513, "learning_rate": 3.643724696356275e-09, "epoch": 4.994163424124514 }, { "current_steps": 2567, "loss": 1.9519, "learning_rate": 3.2388663967611335e-09, "epoch": 4.996108949416342 }, { "current_steps": 2568, "loss": 2.0413, "learning_rate": 2.834008097165992e-09, "epoch": 4.998054474708171 }, { "current_steps": 2569, "loss": 2.2669, "learning_rate": 2.42914979757085e-09, "epoch": 5.0 }, { "current_steps": 2569, "loss": 2.2669, "learning_rate": 2.42914979757085e-09, "epoch": 5.0 } ]