[ { "current_steps": 0, "loss": 2.4728, "learning_rate": 1e-08, "epoch": 0.0012330456226880395 }, { "current_steps": 1, "loss": 2.6026, "learning_rate": 2e-08, "epoch": 0.002466091245376079 }, { "current_steps": 2, "loss": 2.2399, "learning_rate": 3e-08, "epoch": 0.0036991368680641184 }, { "current_steps": 3, "loss": 1.5704, "learning_rate": 4e-08, "epoch": 0.004932182490752158 }, { "current_steps": 4, "loss": 2.5034, "learning_rate": 5e-08, "epoch": 0.006165228113440197 }, { "current_steps": 5, "loss": 2.3134, "learning_rate": 6e-08, "epoch": 0.007398273736128237 }, { "current_steps": 6, "loss": 2.3583, "learning_rate": 7e-08, "epoch": 0.008631319358816275 }, { "current_steps": 7, "loss": 1.8529, "learning_rate": 8e-08, "epoch": 0.009864364981504316 }, { "current_steps": 8, "loss": 1.9016, "learning_rate": 9e-08, "epoch": 0.011097410604192354 }, { "current_steps": 9, "loss": 2.4655, "learning_rate": 1e-07, "epoch": 0.012330456226880395 }, { "current_steps": 10, "loss": 2.5761, "learning_rate": 1.0999999999999999e-07, "epoch": 0.013563501849568433 }, { "current_steps": 11, "loss": 2.6907, "learning_rate": 1.2e-07, "epoch": 0.014796547472256474 }, { "current_steps": 12, "loss": 2.1643, "learning_rate": 1.3e-07, "epoch": 0.016029593094944512 }, { "current_steps": 13, "loss": 2.5927, "learning_rate": 1.4e-07, "epoch": 0.01726263871763255 }, { "current_steps": 14, "loss": 2.567, "learning_rate": 1.5e-07, "epoch": 0.018495684340320593 }, { "current_steps": 15, "loss": 2.0362, "learning_rate": 1.6e-07, "epoch": 0.01972872996300863 }, { "current_steps": 16, "loss": 2.2704, "learning_rate": 1.7000000000000001e-07, "epoch": 0.02096177558569667 }, { "current_steps": 17, "loss": 2.07, "learning_rate": 1.8e-07, "epoch": 0.02219482120838471 }, { "current_steps": 18, "loss": 2.3112, "learning_rate": 1.8999999999999998e-07, "epoch": 0.02342786683107275 }, { "current_steps": 19, "loss": 2.5459, "learning_rate": 2e-07, "epoch": 0.02466091245376079 }, { "current_steps": 20, "loss": 2.4732, "learning_rate": 2.0999999999999997e-07, "epoch": 0.025893958076448828 }, { "current_steps": 21, "loss": 1.9689, "learning_rate": 2.1999999999999998e-07, "epoch": 0.027127003699136867 }, { "current_steps": 22, "loss": 2.4828, "learning_rate": 2.3e-07, "epoch": 0.02836004932182491 }, { "current_steps": 23, "loss": 2.8733, "learning_rate": 2.4e-07, "epoch": 0.029593094944512947 }, { "current_steps": 24, "loss": 1.9689, "learning_rate": 2.5e-07, "epoch": 0.030826140567200986 }, { "current_steps": 25, "loss": 2.0206, "learning_rate": 2.6e-07, "epoch": 0.032059186189889025 }, { "current_steps": 26, "loss": 2.3598, "learning_rate": 2.7e-07, "epoch": 0.03329223181257707 }, { "current_steps": 27, "loss": 2.3877, "learning_rate": 2.8e-07, "epoch": 0.0345252774352651 }, { "current_steps": 28, "loss": 2.0529, "learning_rate": 2.9e-07, "epoch": 0.035758323057953144 }, { "current_steps": 29, "loss": 2.2823, "learning_rate": 3e-07, "epoch": 0.036991368680641186 }, { "current_steps": 30, "loss": 2.3265, "learning_rate": 3.1e-07, "epoch": 0.03822441430332922 }, { "current_steps": 31, "loss": 2.4427, "learning_rate": 3.2e-07, "epoch": 0.03945745992601726 }, { "current_steps": 32, "loss": 2.4659, "learning_rate": 3.3e-07, "epoch": 0.040690505548705305 }, { "current_steps": 33, "loss": 2.3061, "learning_rate": 3.4000000000000003e-07, "epoch": 0.04192355117139334 }, { "current_steps": 34, "loss": 2.4883, "learning_rate": 3.5e-07, "epoch": 0.04315659679408138 }, { "current_steps": 35, "loss": 2.747, "learning_rate": 3.6e-07, "epoch": 0.04438964241676942 }, { "current_steps": 36, "loss": 2.6578, "learning_rate": 3.7e-07, "epoch": 0.04562268803945746 }, { "current_steps": 37, "loss": 2.4525, "learning_rate": 3.7999999999999996e-07, "epoch": 0.0468557336621455 }, { "current_steps": 38, "loss": 2.3157, "learning_rate": 3.8999999999999997e-07, "epoch": 0.04808877928483354 }, { "current_steps": 39, "loss": 2.3949, "learning_rate": 4e-07, "epoch": 0.04932182490752158 }, { "current_steps": 40, "loss": 2.2718, "learning_rate": 4.0999999999999994e-07, "epoch": 0.05055487053020962 }, { "current_steps": 41, "loss": 2.6077, "learning_rate": 4.1999999999999995e-07, "epoch": 0.051787916152897656 }, { "current_steps": 42, "loss": 2.4276, "learning_rate": 4.2999999999999996e-07, "epoch": 0.0530209617755857 }, { "current_steps": 43, "loss": 2.5256, "learning_rate": 4.3999999999999997e-07, "epoch": 0.05425400739827373 }, { "current_steps": 44, "loss": 2.3817, "learning_rate": 4.5e-07, "epoch": 0.055487053020961775 }, { "current_steps": 45, "loss": 2.6506, "learning_rate": 4.6e-07, "epoch": 0.05672009864364982 }, { "current_steps": 46, "loss": 2.0753, "learning_rate": 4.6999999999999995e-07, "epoch": 0.05795314426633785 }, { "current_steps": 47, "loss": 2.1276, "learning_rate": 4.8e-07, "epoch": 0.059186189889025895 }, { "current_steps": 48, "loss": 2.2117, "learning_rate": 4.8e-07, "epoch": 0.06041923551171394 }, { "current_steps": 49, "loss": 2.141, "learning_rate": 4.9e-07, "epoch": 0.06165228113440197 }, { "current_steps": 50, "loss": 2.4854, "learning_rate": 5e-07, "epoch": 0.06288532675709001 }, { "current_steps": 51, "loss": 2.2603, "learning_rate": 5.1e-07, "epoch": 0.06411837237977805 }, { "current_steps": 52, "loss": 2.7403, "learning_rate": 5.2e-07, "epoch": 0.06535141800246609 }, { "current_steps": 53, "loss": 2.191, "learning_rate": 5.2e-07, "epoch": 0.06658446362515413 }, { "current_steps": 54, "loss": 2.1462, "learning_rate": 5.3e-07, "epoch": 0.06781750924784218 }, { "current_steps": 55, "loss": 2.5038, "learning_rate": 5.4e-07, "epoch": 0.0690505548705302 }, { "current_steps": 56, "loss": 2.2442, "learning_rate": 5.5e-07, "epoch": 0.07028360049321825 }, { "current_steps": 57, "loss": 1.8827, "learning_rate": 5.6e-07, "epoch": 0.07151664611590629 }, { "current_steps": 58, "loss": 2.5311, "learning_rate": 5.699999999999999e-07, "epoch": 0.07274969173859433 }, { "current_steps": 59, "loss": 2.5931, "learning_rate": 5.8e-07, "epoch": 0.07398273736128237 }, { "current_steps": 60, "loss": 2.2628, "learning_rate": 5.9e-07, "epoch": 0.0752157829839704 }, { "current_steps": 61, "loss": 2.2149, "learning_rate": 6e-07, "epoch": 0.07644882860665844 }, { "current_steps": 62, "loss": 2.3989, "learning_rate": 6.1e-07, "epoch": 0.07768187422934648 }, { "current_steps": 63, "loss": 2.0978, "learning_rate": 6.2e-07, "epoch": 0.07891491985203453 }, { "current_steps": 64, "loss": 1.9621, "learning_rate": 6.3e-07, "epoch": 0.08014796547472257 }, { "current_steps": 65, "loss": 1.775, "learning_rate": 6.4e-07, "epoch": 0.08138101109741061 }, { "current_steps": 66, "loss": 2.6093, "learning_rate": 6.5e-07, "epoch": 0.08261405672009864 }, { "current_steps": 67, "loss": 2.6, "learning_rate": 6.6e-07, "epoch": 0.08384710234278668 }, { "current_steps": 68, "loss": 2.1964, "learning_rate": 6.7e-07, "epoch": 0.08508014796547472 }, { "current_steps": 69, "loss": 1.9477, "learning_rate": 6.800000000000001e-07, "epoch": 0.08631319358816276 }, { "current_steps": 70, "loss": 2.2726, "learning_rate": 6.9e-07, "epoch": 0.08754623921085081 }, { "current_steps": 71, "loss": 2.8606, "learning_rate": 7e-07, "epoch": 0.08877928483353884 }, { "current_steps": 72, "loss": 2.1482, "learning_rate": 7.1e-07, "epoch": 0.09001233045622688 }, { "current_steps": 73, "loss": 2.0779, "learning_rate": 7.2e-07, "epoch": 0.09124537607891492 }, { "current_steps": 74, "loss": 2.1104, "learning_rate": 7.3e-07, "epoch": 0.09247842170160296 }, { "current_steps": 75, "loss": 2.2784, "learning_rate": 7.4e-07, "epoch": 0.093711467324291 }, { "current_steps": 76, "loss": 2.0205, "learning_rate": 7.5e-07, "epoch": 0.09494451294697903 }, { "current_steps": 77, "loss": 2.4314, "learning_rate": 7.599999999999999e-07, "epoch": 0.09617755856966707 }, { "current_steps": 78, "loss": 2.206, "learning_rate": 7.699999999999999e-07, "epoch": 0.09741060419235512 }, { "current_steps": 79, "loss": 2.4954, "learning_rate": 7.799999999999999e-07, "epoch": 0.09864364981504316 }, { "current_steps": 80, "loss": 2.3073, "learning_rate": 7.9e-07, "epoch": 0.0998766954377312 }, { "current_steps": 81, "loss": 2.3731, "learning_rate": 8e-07, "epoch": 0.10110974106041924 }, { "current_steps": 82, "loss": 2.6875, "learning_rate": 8.1e-07, "epoch": 0.10234278668310727 }, { "current_steps": 83, "loss": 2.3726, "learning_rate": 8.199999999999999e-07, "epoch": 0.10357583230579531 }, { "current_steps": 84, "loss": 2.368, "learning_rate": 8.299999999999999e-07, "epoch": 0.10480887792848335 }, { "current_steps": 85, "loss": 2.8464, "learning_rate": 8.399999999999999e-07, "epoch": 0.1060419235511714 }, { "current_steps": 86, "loss": 2.5616, "learning_rate": 8.499999999999999e-07, "epoch": 0.10727496917385944 }, { "current_steps": 87, "loss": 2.4791, "learning_rate": 8.599999999999999e-07, "epoch": 0.10850801479654747 }, { "current_steps": 88, "loss": 2.6427, "learning_rate": 8.699999999999999e-07, "epoch": 0.10974106041923551 }, { "current_steps": 89, "loss": 2.3288, "learning_rate": 8.799999999999999e-07, "epoch": 0.11097410604192355 }, { "current_steps": 90, "loss": 2.2888, "learning_rate": 8.9e-07, "epoch": 0.11220715166461159 }, { "current_steps": 91, "loss": 2.1456, "learning_rate": 9e-07, "epoch": 0.11344019728729964 }, { "current_steps": 92, "loss": 2.173, "learning_rate": 9.1e-07, "epoch": 0.11467324290998766 }, { "current_steps": 93, "loss": 2.0691, "learning_rate": 9.2e-07, "epoch": 0.1159062885326757 }, { "current_steps": 94, "loss": 2.2972, "learning_rate": 9.3e-07, "epoch": 0.11713933415536375 }, { "current_steps": 95, "loss": 2.4864, "learning_rate": 9.399999999999999e-07, "epoch": 0.11837237977805179 }, { "current_steps": 96, "loss": 2.5763, "learning_rate": 9.499999999999999e-07, "epoch": 0.11960542540073983 }, { "current_steps": 97, "loss": 2.5676, "learning_rate": 9.6e-07, "epoch": 0.12083847102342787 }, { "current_steps": 98, "loss": 2.0428, "learning_rate": 9.7e-07, "epoch": 0.1220715166461159 }, { "current_steps": 99, "loss": 2.2045, "learning_rate": 9.8e-07, "epoch": 0.12330456226880394 }, { "current_steps": 100, "loss": 2.2652, "learning_rate": 9.9e-07, "epoch": 0.12453760789149199 }, { "current_steps": 101, "loss": 2.1752, "learning_rate": 1e-06, "epoch": 0.12577065351418001 }, { "current_steps": 102, "loss": 2.2061, "learning_rate": 9.997471554993678e-07, "epoch": 0.12700369913686807 }, { "current_steps": 103, "loss": 2.1564, "learning_rate": 9.994943109987357e-07, "epoch": 0.1282367447595561 }, { "current_steps": 104, "loss": 2.2756, "learning_rate": 9.992414664981036e-07, "epoch": 0.12946979038224415 }, { "current_steps": 105, "loss": 2.6245, "learning_rate": 9.989886219974715e-07, "epoch": 0.13070283600493218 }, { "current_steps": 106, "loss": 2.2232, "learning_rate": 9.987357774968394e-07, "epoch": 0.1319358816276202 }, { "current_steps": 107, "loss": 2.7268, "learning_rate": 9.984829329962073e-07, "epoch": 0.13316892725030827 }, { "current_steps": 108, "loss": 2.3726, "learning_rate": 9.982300884955752e-07, "epoch": 0.1344019728729963 }, { "current_steps": 109, "loss": 2.8164, "learning_rate": 9.97977243994943e-07, "epoch": 0.13563501849568435 }, { "current_steps": 110, "loss": 2.2156, "learning_rate": 9.97724399494311e-07, "epoch": 0.13686806411837238 }, { "current_steps": 111, "loss": 2.428, "learning_rate": 9.974715549936788e-07, "epoch": 0.1381011097410604 }, { "current_steps": 112, "loss": 2.4493, "learning_rate": 9.972187104930467e-07, "epoch": 0.13933415536374846 }, { "current_steps": 113, "loss": 2.1442, "learning_rate": 9.969658659924146e-07, "epoch": 0.1405672009864365 }, { "current_steps": 114, "loss": 2.4871, "learning_rate": 9.967130214917825e-07, "epoch": 0.14180024660912455 }, { "current_steps": 115, "loss": 2.3341, "learning_rate": 9.964601769911504e-07, "epoch": 0.14303329223181258 }, { "current_steps": 116, "loss": 2.2965, "learning_rate": 9.962073324905183e-07, "epoch": 0.1442663378545006 }, { "current_steps": 117, "loss": 2.7655, "learning_rate": 9.959544879898862e-07, "epoch": 0.14549938347718866 }, { "current_steps": 118, "loss": 2.1662, "learning_rate": 9.95701643489254e-07, "epoch": 0.1467324290998767 }, { "current_steps": 119, "loss": 2.6571, "learning_rate": 9.95448798988622e-07, "epoch": 0.14796547472256474 }, { "current_steps": 120, "loss": 2.3126, "learning_rate": 9.951959544879898e-07, "epoch": 0.14919852034525277 }, { "current_steps": 121, "loss": 2.1292, "learning_rate": 9.949431099873577e-07, "epoch": 0.1504315659679408 }, { "current_steps": 122, "loss": 2.0266, "learning_rate": 9.946902654867256e-07, "epoch": 0.15166461159062886 }, { "current_steps": 123, "loss": 2.2914, "learning_rate": 9.944374209860935e-07, "epoch": 0.15289765721331688 }, { "current_steps": 124, "loss": 2.0119, "learning_rate": 9.941845764854614e-07, "epoch": 0.15413070283600494 }, { "current_steps": 125, "loss": 2.4431, "learning_rate": 9.939317319848293e-07, "epoch": 0.15536374845869297 }, { "current_steps": 126, "loss": 2.7428, "learning_rate": 9.936788874841971e-07, "epoch": 0.15659679408138102 }, { "current_steps": 127, "loss": 2.2577, "learning_rate": 9.93426042983565e-07, "epoch": 0.15782983970406905 }, { "current_steps": 128, "loss": 2.5318, "learning_rate": 9.93173198482933e-07, "epoch": 0.15906288532675708 }, { "current_steps": 129, "loss": 2.3165, "learning_rate": 9.929203539823008e-07, "epoch": 0.16029593094944514 }, { "current_steps": 130, "loss": 2.273, "learning_rate": 9.926675094816687e-07, "epoch": 0.16152897657213316 }, { "current_steps": 131, "loss": 2.2688, "learning_rate": 9.924146649810366e-07, "epoch": 0.16276202219482122 }, { "current_steps": 132, "loss": 2.5029, "learning_rate": 9.921618204804045e-07, "epoch": 0.16399506781750925 }, { "current_steps": 133, "loss": 2.3351, "learning_rate": 9.919089759797724e-07, "epoch": 0.16522811344019728 }, { "current_steps": 134, "loss": 2.3306, "learning_rate": 9.916561314791402e-07, "epoch": 0.16646115906288533 }, { "current_steps": 135, "loss": 2.1771, "learning_rate": 9.914032869785081e-07, "epoch": 0.16769420468557336 }, { "current_steps": 136, "loss": 2.4638, "learning_rate": 9.91150442477876e-07, "epoch": 0.16892725030826142 }, { "current_steps": 137, "loss": 2.634, "learning_rate": 9.90897597977244e-07, "epoch": 0.17016029593094945 }, { "current_steps": 138, "loss": 1.9946, "learning_rate": 9.906447534766118e-07, "epoch": 0.17139334155363747 }, { "current_steps": 139, "loss": 2.2979, "learning_rate": 9.903919089759797e-07, "epoch": 0.17262638717632553 }, { "current_steps": 140, "loss": 1.5132, "learning_rate": 9.901390644753476e-07, "epoch": 0.17385943279901356 }, { "current_steps": 141, "loss": 2.1438, "learning_rate": 9.898862199747155e-07, "epoch": 0.17509247842170161 }, { "current_steps": 142, "loss": 2.1895, "learning_rate": 9.896333754740833e-07, "epoch": 0.17632552404438964 }, { "current_steps": 143, "loss": 2.3998, "learning_rate": 9.893805309734512e-07, "epoch": 0.17755856966707767 }, { "current_steps": 144, "loss": 2.2283, "learning_rate": 9.891276864728191e-07, "epoch": 0.17879161528976573 }, { "current_steps": 145, "loss": 2.7197, "learning_rate": 9.891276864728191e-07, "epoch": 0.18002466091245375 }, { "current_steps": 146, "loss": 2.4331, "learning_rate": 9.88874841972187e-07, "epoch": 0.1812577065351418 }, { "current_steps": 147, "loss": 2.0481, "learning_rate": 9.88621997471555e-07, "epoch": 0.18249075215782984 }, { "current_steps": 148, "loss": 2.5388, "learning_rate": 9.883691529709228e-07, "epoch": 0.18372379778051787 }, { "current_steps": 149, "loss": 2.6644, "learning_rate": 9.881163084702907e-07, "epoch": 0.18495684340320592 }, { "current_steps": 150, "loss": 2.787, "learning_rate": 9.878634639696586e-07, "epoch": 0.18618988902589395 }, { "current_steps": 151, "loss": 2.0133, "learning_rate": 9.876106194690265e-07, "epoch": 0.187422934648582 }, { "current_steps": 152, "loss": 2.38, "learning_rate": 9.873577749683943e-07, "epoch": 0.18865598027127004 }, { "current_steps": 153, "loss": 2.3236, "learning_rate": 9.871049304677622e-07, "epoch": 0.18988902589395806 }, { "current_steps": 154, "loss": 1.2257, "learning_rate": 9.868520859671301e-07, "epoch": 0.19112207151664612 }, { "current_steps": 155, "loss": 1.9105, "learning_rate": 9.86599241466498e-07, "epoch": 0.19235511713933415 }, { "current_steps": 156, "loss": 2.7715, "learning_rate": 9.863463969658659e-07, "epoch": 0.1935881627620222 }, { "current_steps": 157, "loss": 2.2885, "learning_rate": 9.860935524652338e-07, "epoch": 0.19482120838471023 }, { "current_steps": 158, "loss": 2.383, "learning_rate": 9.858407079646017e-07, "epoch": 0.1960542540073983 }, { "current_steps": 159, "loss": 2.3045, "learning_rate": 9.855878634639696e-07, "epoch": 0.19728729963008632 }, { "current_steps": 160, "loss": 2.3893, "learning_rate": 9.853350189633374e-07, "epoch": 0.19852034525277434 }, { "current_steps": 161, "loss": 2.4892, "learning_rate": 9.850821744627053e-07, "epoch": 0.1997533908754624 }, { "current_steps": 162, "loss": 1.7079, "learning_rate": 9.848293299620732e-07, "epoch": 0.20098643649815043 }, { "current_steps": 163, "loss": 1.9064, "learning_rate": 9.84576485461441e-07, "epoch": 0.20221948212083848 }, { "current_steps": 164, "loss": 2.5225, "learning_rate": 9.843236409608092e-07, "epoch": 0.2034525277435265 }, { "current_steps": 165, "loss": 2.0326, "learning_rate": 9.84070796460177e-07, "epoch": 0.20468557336621454 }, { "current_steps": 166, "loss": 2.5441, "learning_rate": 9.838179519595448e-07, "epoch": 0.2059186189889026 }, { "current_steps": 167, "loss": 2.2858, "learning_rate": 9.835651074589127e-07, "epoch": 0.20715166461159062 }, { "current_steps": 168, "loss": 2.5683, "learning_rate": 9.833122629582805e-07, "epoch": 0.20838471023427868 }, { "current_steps": 169, "loss": 1.9911, "learning_rate": 9.830594184576484e-07, "epoch": 0.2096177558569667 }, { "current_steps": 170, "loss": 2.409, "learning_rate": 9.828065739570163e-07, "epoch": 0.21085080147965474 }, { "current_steps": 171, "loss": 2.4484, "learning_rate": 9.825537294563842e-07, "epoch": 0.2120838471023428 }, { "current_steps": 172, "loss": 2.7544, "learning_rate": 9.82300884955752e-07, "epoch": 0.21331689272503082 }, { "current_steps": 173, "loss": 2.4521, "learning_rate": 9.8204804045512e-07, "epoch": 0.21454993834771888 }, { "current_steps": 174, "loss": 2.1578, "learning_rate": 9.817951959544879e-07, "epoch": 0.2157829839704069 }, { "current_steps": 175, "loss": 2.2096, "learning_rate": 9.81542351453856e-07, "epoch": 0.21701602959309493 }, { "current_steps": 176, "loss": 2.0764, "learning_rate": 9.812895069532239e-07, "epoch": 0.218249075215783 }, { "current_steps": 177, "loss": 2.3992, "learning_rate": 9.810366624525917e-07, "epoch": 0.21948212083847102 }, { "current_steps": 178, "loss": 2.6554, "learning_rate": 9.807838179519594e-07, "epoch": 0.22071516646115907 }, { "current_steps": 179, "loss": 2.2881, "learning_rate": 9.805309734513273e-07, "epoch": 0.2219482120838471 }, { "current_steps": 180, "loss": 2.1019, "learning_rate": 9.802781289506952e-07, "epoch": 0.22318125770653513 }, { "current_steps": 181, "loss": 2.5905, "learning_rate": 9.80025284450063e-07, "epoch": 0.22441430332922319 }, { "current_steps": 182, "loss": 2.3833, "learning_rate": 9.79772439949431e-07, "epoch": 0.22564734895191121 }, { "current_steps": 183, "loss": 2.0468, "learning_rate": 9.795195954487989e-07, "epoch": 0.22688039457459927 }, { "current_steps": 184, "loss": 2.0682, "learning_rate": 9.792667509481667e-07, "epoch": 0.2281134401972873 }, { "current_steps": 185, "loss": 2.4567, "learning_rate": 9.790139064475346e-07, "epoch": 0.22934648581997533 }, { "current_steps": 186, "loss": 2.0734, "learning_rate": 9.787610619469025e-07, "epoch": 0.23057953144266338 }, { "current_steps": 187, "loss": 2.1011, "learning_rate": 9.785082174462706e-07, "epoch": 0.2318125770653514 }, { "current_steps": 188, "loss": 1.7606, "learning_rate": 9.782553729456385e-07, "epoch": 0.23304562268803947 }, { "current_steps": 189, "loss": 2.159, "learning_rate": 9.780025284450064e-07, "epoch": 0.2342786683107275 }, { "current_steps": 190, "loss": 2.6561, "learning_rate": 9.777496839443743e-07, "epoch": 0.23551171393341552 }, { "current_steps": 191, "loss": 2.4713, "learning_rate": 9.77496839443742e-07, "epoch": 0.23674475955610358 }, { "current_steps": 192, "loss": 2.7225, "learning_rate": 9.772439949431098e-07, "epoch": 0.2379778051787916 }, { "current_steps": 193, "loss": 1.7532, "learning_rate": 9.769911504424777e-07, "epoch": 0.23921085080147966 }, { "current_steps": 194, "loss": 2.2671, "learning_rate": 9.767383059418456e-07, "epoch": 0.2404438964241677 }, { "current_steps": 195, "loss": 2.1696, "learning_rate": 9.764854614412135e-07, "epoch": 0.24167694204685575 }, { "current_steps": 196, "loss": 2.236, "learning_rate": 9.762326169405814e-07, "epoch": 0.24290998766954378 }, { "current_steps": 197, "loss": 2.3724, "learning_rate": 9.759797724399493e-07, "epoch": 0.2441430332922318 }, { "current_steps": 198, "loss": 2.4374, "learning_rate": 9.757269279393174e-07, "epoch": 0.24537607891491986 }, { "current_steps": 199, "loss": 2.2181, "learning_rate": 9.754740834386853e-07, "epoch": 0.2466091245376079 }, { "current_steps": 200, "loss": 2.8533, "learning_rate": 9.752212389380532e-07, "epoch": 0.24784217016029594 }, { "current_steps": 201, "loss": 2.6601, "learning_rate": 9.74968394437421e-07, "epoch": 0.24907521578298397 }, { "current_steps": 202, "loss": 2.1262, "learning_rate": 9.74715549936789e-07, "epoch": 0.25030826140567203 }, { "current_steps": 203, "loss": 2.6572, "learning_rate": 9.744627054361568e-07, "epoch": 0.25154130702836003 }, { "current_steps": 204, "loss": 2.2906, "learning_rate": 9.742098609355245e-07, "epoch": 0.2527743526510481 }, { "current_steps": 205, "loss": 2.5079, "learning_rate": 9.739570164348924e-07, "epoch": 0.25400739827373614 }, { "current_steps": 206, "loss": 2.5017, "learning_rate": 9.737041719342603e-07, "epoch": 0.25524044389642414 }, { "current_steps": 207, "loss": 2.6337, "learning_rate": 9.734513274336282e-07, "epoch": 0.2564734895191122 }, { "current_steps": 208, "loss": 2.4189, "learning_rate": 9.73198482932996e-07, "epoch": 0.25770653514180025 }, { "current_steps": 209, "loss": 2.9359, "learning_rate": 9.729456384323641e-07, "epoch": 0.2589395807644883 }, { "current_steps": 210, "loss": 2.2731, "learning_rate": 9.72692793931732e-07, "epoch": 0.2601726263871763 }, { "current_steps": 211, "loss": 2.1633, "learning_rate": 9.724399494311e-07, "epoch": 0.26140567200986436 }, { "current_steps": 212, "loss": 2.6957, "learning_rate": 9.721871049304678e-07, "epoch": 0.2626387176325524 }, { "current_steps": 213, "loss": 2.2706, "learning_rate": 9.719342604298357e-07, "epoch": 0.2638717632552404 }, { "current_steps": 214, "loss": 2.4269, "learning_rate": 9.716814159292036e-07, "epoch": 0.2651048088779285 }, { "current_steps": 215, "loss": 1.8479, "learning_rate": 9.714285714285715e-07, "epoch": 0.26633785450061653 }, { "current_steps": 216, "loss": 2.238, "learning_rate": 9.711757269279392e-07, "epoch": 0.2675709001233046 }, { "current_steps": 217, "loss": 1.9367, "learning_rate": 9.70922882427307e-07, "epoch": 0.2688039457459926 }, { "current_steps": 218, "loss": 2.4265, "learning_rate": 9.70670037926675e-07, "epoch": 0.27003699136868065 }, { "current_steps": 219, "loss": 1.9244, "learning_rate": 9.704171934260428e-07, "epoch": 0.2712700369913687 }, { "current_steps": 220, "loss": 2.201, "learning_rate": 9.70164348925411e-07, "epoch": 0.2725030826140567 }, { "current_steps": 221, "loss": 2.2387, "learning_rate": 9.699115044247788e-07, "epoch": 0.27373612823674476 }, { "current_steps": 222, "loss": 2.4545, "learning_rate": 9.696586599241467e-07, "epoch": 0.2749691738594328 }, { "current_steps": 223, "loss": 2.3753, "learning_rate": 9.694058154235146e-07, "epoch": 0.2762022194821208 }, { "current_steps": 224, "loss": 2.4218, "learning_rate": 9.691529709228825e-07, "epoch": 0.27743526510480887 }, { "current_steps": 225, "loss": 1.8953, "learning_rate": 9.689001264222504e-07, "epoch": 0.2786683107274969 }, { "current_steps": 226, "loss": 2.7977, "learning_rate": 9.686472819216182e-07, "epoch": 0.279901356350185 }, { "current_steps": 227, "loss": 2.077, "learning_rate": 9.683944374209861e-07, "epoch": 0.281134401972873 }, { "current_steps": 228, "loss": 2.4823, "learning_rate": 9.68141592920354e-07, "epoch": 0.28236744759556104 }, { "current_steps": 229, "loss": 2.4623, "learning_rate": 9.678887484197217e-07, "epoch": 0.2836004932182491 }, { "current_steps": 230, "loss": 2.0842, "learning_rate": 9.676359039190896e-07, "epoch": 0.2848335388409371 }, { "current_steps": 231, "loss": 2.0893, "learning_rate": 9.673830594184577e-07, "epoch": 0.28606658446362515 }, { "current_steps": 232, "loss": 2.154, "learning_rate": 9.671302149178256e-07, "epoch": 0.2872996300863132 }, { "current_steps": 233, "loss": 2.5034, "learning_rate": 9.668773704171935e-07, "epoch": 0.2885326757090012 }, { "current_steps": 234, "loss": 1.8443, "learning_rate": 9.666245259165613e-07, "epoch": 0.28976572133168926 }, { "current_steps": 235, "loss": 2.2518, "learning_rate": 9.663716814159292e-07, "epoch": 0.2909987669543773 }, { "current_steps": 236, "loss": 2.2615, "learning_rate": 9.661188369152971e-07, "epoch": 0.2922318125770654 }, { "current_steps": 237, "loss": 2.0934, "learning_rate": 9.65865992414665e-07, "epoch": 0.2934648581997534 }, { "current_steps": 238, "loss": 1.8453, "learning_rate": 9.656131479140329e-07, "epoch": 0.29469790382244143 }, { "current_steps": 239, "loss": 2.3842, "learning_rate": 9.653603034134008e-07, "epoch": 0.2959309494451295 }, { "current_steps": 240, "loss": 2.2355, "learning_rate": 9.651074589127687e-07, "epoch": 0.2971639950678175 }, { "current_steps": 241, "loss": 2.2289, "learning_rate": 9.648546144121363e-07, "epoch": 0.29839704069050554 }, { "current_steps": 242, "loss": 2.6883, "learning_rate": 9.646017699115042e-07, "epoch": 0.2996300863131936 }, { "current_steps": 243, "loss": 2.6352, "learning_rate": 9.643489254108723e-07, "epoch": 0.3008631319358816 }, { "current_steps": 244, "loss": 2.136, "learning_rate": 9.640960809102402e-07, "epoch": 0.30209617755856966 }, { "current_steps": 245, "loss": 2.1396, "learning_rate": 9.638432364096081e-07, "epoch": 0.3033292231812577 }, { "current_steps": 246, "loss": 2.4079, "learning_rate": 9.63590391908976e-07, "epoch": 0.30456226880394577 }, { "current_steps": 247, "loss": 2.6435, "learning_rate": 9.633375474083439e-07, "epoch": 0.30579531442663377 }, { "current_steps": 248, "loss": 2.2238, "learning_rate": 9.630847029077118e-07, "epoch": 0.3070283600493218 }, { "current_steps": 249, "loss": 2.8386, "learning_rate": 9.628318584070797e-07, "epoch": 0.3082614056720099 }, { "current_steps": 250, "loss": 2.4775, "learning_rate": 9.625790139064475e-07, "epoch": 0.3094944512946979 }, { "current_steps": 251, "loss": 2.4936, "learning_rate": 9.623261694058154e-07, "epoch": 0.31072749691738594 }, { "current_steps": 252, "loss": 2.0831, "learning_rate": 9.620733249051833e-07, "epoch": 0.311960542540074 }, { "current_steps": 253, "loss": 2.0647, "learning_rate": 9.618204804045512e-07, "epoch": 0.31319358816276205 }, { "current_steps": 254, "loss": 1.9711, "learning_rate": 9.61567635903919e-07, "epoch": 0.31442663378545005 }, { "current_steps": 255, "loss": 2.2024, "learning_rate": 9.61314791403287e-07, "epoch": 0.3156596794081381 }, { "current_steps": 256, "loss": 2.2001, "learning_rate": 9.610619469026549e-07, "epoch": 0.31689272503082616 }, { "current_steps": 257, "loss": 2.4807, "learning_rate": 9.608091024020228e-07, "epoch": 0.31812577065351416 }, { "current_steps": 258, "loss": 2.4564, "learning_rate": 9.605562579013906e-07, "epoch": 0.3193588162762022 }, { "current_steps": 259, "loss": 2.5815, "learning_rate": 9.603034134007585e-07, "epoch": 0.3205918618988903 }, { "current_steps": 260, "loss": 2.2209, "learning_rate": 9.600505689001264e-07, "epoch": 0.3218249075215783 }, { "current_steps": 261, "loss": 2.7264, "learning_rate": 9.597977243994943e-07, "epoch": 0.32305795314426633 }, { "current_steps": 262, "loss": 2.4135, "learning_rate": 9.595448798988622e-07, "epoch": 0.3242909987669544 }, { "current_steps": 263, "loss": 2.2147, "learning_rate": 9.5929203539823e-07, "epoch": 0.32552404438964244 }, { "current_steps": 264, "loss": 1.7948, "learning_rate": 9.59039190897598e-07, "epoch": 0.32675709001233044 }, { "current_steps": 265, "loss": 2.3197, "learning_rate": 9.587863463969659e-07, "epoch": 0.3279901356350185 }, { "current_steps": 266, "loss": 2.3894, "learning_rate": 9.585335018963337e-07, "epoch": 0.32922318125770655 }, { "current_steps": 267, "loss": 3.0251, "learning_rate": 9.585335018963337e-07, "epoch": 0.33045622688039455 }, { "current_steps": 268, "loss": 2.1421, "learning_rate": 9.582806573957016e-07, "epoch": 0.3316892725030826 }, { "current_steps": 269, "loss": 2.1777, "learning_rate": 9.580278128950695e-07, "epoch": 0.33292231812577067 }, { "current_steps": 270, "loss": 2.2141, "learning_rate": 9.577749683944374e-07, "epoch": 0.33415536374845867 }, { "current_steps": 271, "loss": 2.3415, "learning_rate": 9.575221238938053e-07, "epoch": 0.3353884093711467 }, { "current_steps": 272, "loss": 2.4078, "learning_rate": 9.572692793931732e-07, "epoch": 0.3366214549938348 }, { "current_steps": 273, "loss": 2.3456, "learning_rate": 9.57016434892541e-07, "epoch": 0.33785450061652283 }, { "current_steps": 274, "loss": 2.2706, "learning_rate": 9.56763590391909e-07, "epoch": 0.33908754623921084 }, { "current_steps": 275, "loss": 2.1707, "learning_rate": 9.565107458912768e-07, "epoch": 0.3403205918618989 }, { "current_steps": 276, "loss": 2.3252, "learning_rate": 9.562579013906447e-07, "epoch": 0.34155363748458695 }, { "current_steps": 277, "loss": 2.3785, "learning_rate": 9.560050568900126e-07, "epoch": 0.34278668310727495 }, { "current_steps": 278, "loss": 1.5646, "learning_rate": 9.557522123893805e-07, "epoch": 0.344019728729963 }, { "current_steps": 279, "loss": 2.6544, "learning_rate": 9.554993678887484e-07, "epoch": 0.34525277435265106 }, { "current_steps": 280, "loss": 2.4265, "learning_rate": 9.552465233881163e-07, "epoch": 0.34648581997533906 }, { "current_steps": 281, "loss": 2.5073, "learning_rate": 9.549936788874842e-07, "epoch": 0.3477188655980271 }, { "current_steps": 282, "loss": 2.3678, "learning_rate": 9.54740834386852e-07, "epoch": 0.34895191122071517 }, { "current_steps": 283, "loss": 2.0234, "learning_rate": 9.5448798988622e-07, "epoch": 0.35018495684340323 }, { "current_steps": 284, "loss": 2.153, "learning_rate": 9.542351453855878e-07, "epoch": 0.35141800246609123 }, { "current_steps": 285, "loss": 2.3276, "learning_rate": 9.539823008849557e-07, "epoch": 0.3526510480887793 }, { "current_steps": 286, "loss": 1.9875, "learning_rate": 9.537294563843236e-07, "epoch": 0.35388409371146734 }, { "current_steps": 287, "loss": 2.0281, "learning_rate": 9.534766118836914e-07, "epoch": 0.35511713933415534 }, { "current_steps": 288, "loss": 2.3789, "learning_rate": 9.532237673830594e-07, "epoch": 0.3563501849568434 }, { "current_steps": 289, "loss": 2.021, "learning_rate": 9.529709228824273e-07, "epoch": 0.35758323057953145 }, { "current_steps": 290, "loss": 2.1254, "learning_rate": 9.527180783817952e-07, "epoch": 0.3588162762022195 }, { "current_steps": 291, "loss": 1.9059, "learning_rate": 9.52465233881163e-07, "epoch": 0.3600493218249075 }, { "current_steps": 292, "loss": 2.0932, "learning_rate": 9.522123893805309e-07, "epoch": 0.36128236744759556 }, { "current_steps": 293, "loss": 2.1044, "learning_rate": 9.519595448798988e-07, "epoch": 0.3625154130702836 }, { "current_steps": 294, "loss": 2.2376, "learning_rate": 9.517067003792667e-07, "epoch": 0.3637484586929716 }, { "current_steps": 295, "loss": 2.2645, "learning_rate": 9.514538558786346e-07, "epoch": 0.3649815043156597 }, { "current_steps": 296, "loss": 2.4771, "learning_rate": 9.512010113780025e-07, "epoch": 0.36621454993834773 }, { "current_steps": 297, "loss": 2.7204, "learning_rate": 9.509481668773704e-07, "epoch": 0.36744759556103573 }, { "current_steps": 298, "loss": 2.1946, "learning_rate": 9.506953223767383e-07, "epoch": 0.3686806411837238 }, { "current_steps": 299, "loss": 1.893, "learning_rate": 9.50442477876106e-07, "epoch": 0.36991368680641185 }, { "current_steps": 300, "loss": 2.3613, "learning_rate": 9.50189633375474e-07, "epoch": 0.3711467324290999 }, { "current_steps": 301, "loss": 2.2339, "learning_rate": 9.499367888748419e-07, "epoch": 0.3723797780517879 }, { "current_steps": 302, "loss": 2.6089, "learning_rate": 9.496839443742098e-07, "epoch": 0.37361282367447596 }, { "current_steps": 303, "loss": 1.8003, "learning_rate": 9.494310998735777e-07, "epoch": 0.374845869297164 }, { "current_steps": 304, "loss": 2.3363, "learning_rate": 9.491782553729456e-07, "epoch": 0.376078914919852 }, { "current_steps": 305, "loss": 2.39, "learning_rate": 9.489254108723135e-07, "epoch": 0.37731196054254007 }, { "current_steps": 306, "loss": 2.1855, "learning_rate": 9.486725663716814e-07, "epoch": 0.3785450061652281 }, { "current_steps": 307, "loss": 2.1988, "learning_rate": 9.484197218710493e-07, "epoch": 0.3797780517879161 }, { "current_steps": 308, "loss": 2.2336, "learning_rate": 9.481668773704171e-07, "epoch": 0.3810110974106042 }, { "current_steps": 309, "loss": 2.2841, "learning_rate": 9.47914032869785e-07, "epoch": 0.38224414303329224 }, { "current_steps": 310, "loss": 1.9701, "learning_rate": 9.476611883691529e-07, "epoch": 0.3834771886559803 }, { "current_steps": 311, "loss": 2.3607, "learning_rate": 9.474083438685209e-07, "epoch": 0.3847102342786683 }, { "current_steps": 312, "loss": 2.4619, "learning_rate": 9.471554993678887e-07, "epoch": 0.38594327990135635 }, { "current_steps": 313, "loss": 2.5611, "learning_rate": 9.469026548672566e-07, "epoch": 0.3871763255240444 }, { "current_steps": 314, "loss": 2.8078, "learning_rate": 9.466498103666245e-07, "epoch": 0.3884093711467324 }, { "current_steps": 315, "loss": 2.3492, "learning_rate": 9.463969658659924e-07, "epoch": 0.38964241676942046 }, { "current_steps": 316, "loss": 2.2175, "learning_rate": 9.461441213653602e-07, "epoch": 0.3908754623921085 }, { "current_steps": 317, "loss": 2.5227, "learning_rate": 9.458912768647281e-07, "epoch": 0.3921085080147966 }, { "current_steps": 318, "loss": 2.7856, "learning_rate": 9.45638432364096e-07, "epoch": 0.3933415536374846 }, { "current_steps": 319, "loss": 2.2818, "learning_rate": 9.453855878634639e-07, "epoch": 0.39457459926017263 }, { "current_steps": 320, "loss": 2.3538, "learning_rate": 9.451327433628318e-07, "epoch": 0.3958076448828607 }, { "current_steps": 321, "loss": 2.1171, "learning_rate": 9.448798988621997e-07, "epoch": 0.3970406905055487 }, { "current_steps": 322, "loss": 2.7694, "learning_rate": 9.446270543615677e-07, "epoch": 0.39827373612823674 }, { "current_steps": 323, "loss": 2.2313, "learning_rate": 9.443742098609356e-07, "epoch": 0.3995067817509248 }, { "current_steps": 324, "loss": 2.7043, "learning_rate": 9.441213653603035e-07, "epoch": 0.4007398273736128 }, { "current_steps": 325, "loss": 2.0613, "learning_rate": 9.438685208596712e-07, "epoch": 0.40197287299630086 }, { "current_steps": 326, "loss": 2.275, "learning_rate": 9.436156763590391e-07, "epoch": 0.4032059186189889 }, { "current_steps": 327, "loss": 2.7574, "learning_rate": 9.43362831858407e-07, "epoch": 0.40443896424167697 }, { "current_steps": 328, "loss": 2.2324, "learning_rate": 9.431099873577749e-07, "epoch": 0.40567200986436497 }, { "current_steps": 329, "loss": 2.1636, "learning_rate": 9.428571428571428e-07, "epoch": 0.406905055487053 }, { "current_steps": 330, "loss": 1.8051, "learning_rate": 9.426042983565107e-07, "epoch": 0.4081381011097411 }, { "current_steps": 331, "loss": 2.5478, "learning_rate": 9.423514538558786e-07, "epoch": 0.4093711467324291 }, { "current_steps": 332, "loss": 2.3849, "learning_rate": 9.420986093552464e-07, "epoch": 0.41060419235511714 }, { "current_steps": 333, "loss": 2.0857, "learning_rate": 9.418457648546144e-07, "epoch": 0.4118372379778052 }, { "current_steps": 334, "loss": 2.2967, "learning_rate": 9.415929203539823e-07, "epoch": 0.4130702836004932 }, { "current_steps": 335, "loss": 2.3323, "learning_rate": 9.413400758533502e-07, "epoch": 0.41430332922318125 }, { "current_steps": 336, "loss": 2.2711, "learning_rate": 9.410872313527181e-07, "epoch": 0.4155363748458693 }, { "current_steps": 337, "loss": 2.4079, "learning_rate": 9.408343868520859e-07, "epoch": 0.41676942046855736 }, { "current_steps": 338, "loss": 2.329, "learning_rate": 9.405815423514538e-07, "epoch": 0.41800246609124536 }, { "current_steps": 339, "loss": 2.2051, "learning_rate": 9.403286978508217e-07, "epoch": 0.4192355117139334 }, { "current_steps": 340, "loss": 2.3369, "learning_rate": 9.400758533501895e-07, "epoch": 0.4204685573366215 }, { "current_steps": 341, "loss": 2.3005, "learning_rate": 9.398230088495574e-07, "epoch": 0.4217016029593095 }, { "current_steps": 342, "loss": 2.5354, "learning_rate": 9.395701643489253e-07, "epoch": 0.42293464858199753 }, { "current_steps": 343, "loss": 1.5064, "learning_rate": 9.393173198482932e-07, "epoch": 0.4241676942046856 }, { "current_steps": 344, "loss": 1.7727, "learning_rate": 9.390644753476612e-07, "epoch": 0.4254007398273736 }, { "current_steps": 345, "loss": 2.2714, "learning_rate": 9.388116308470291e-07, "epoch": 0.42663378545006164 }, { "current_steps": 346, "loss": 2.1001, "learning_rate": 9.38558786346397e-07, "epoch": 0.4278668310727497 }, { "current_steps": 347, "loss": 1.8009, "learning_rate": 9.383059418457649e-07, "epoch": 0.42909987669543775 }, { "current_steps": 348, "loss": 2.538, "learning_rate": 9.380530973451328e-07, "epoch": 0.43033292231812575 }, { "current_steps": 349, "loss": 2.6135, "learning_rate": 9.378002528445006e-07, "epoch": 0.4315659679408138 }, { "current_steps": 350, "loss": 2.7319, "learning_rate": 9.375474083438684e-07, "epoch": 0.43279901356350187 }, { "current_steps": 351, "loss": 1.7998, "learning_rate": 9.372945638432363e-07, "epoch": 0.43403205918618987 }, { "current_steps": 352, "loss": 2.2468, "learning_rate": 9.370417193426042e-07, "epoch": 0.4352651048088779 }, { "current_steps": 353, "loss": 2.6597, "learning_rate": 9.367888748419721e-07, "epoch": 0.436498150431566 }, { "current_steps": 354, "loss": 2.1562, "learning_rate": 9.3653603034134e-07, "epoch": 0.43773119605425403 }, { "current_steps": 355, "loss": 1.8127, "learning_rate": 9.362831858407079e-07, "epoch": 0.43896424167694204 }, { "current_steps": 356, "loss": 2.2487, "learning_rate": 9.360303413400759e-07, "epoch": 0.4401972872996301 }, { "current_steps": 357, "loss": 2.332, "learning_rate": 9.357774968394437e-07, "epoch": 0.44143033292231815 }, { "current_steps": 358, "loss": 2.3694, "learning_rate": 9.355246523388116e-07, "epoch": 0.44266337854500615 }, { "current_steps": 359, "loss": 2.5623, "learning_rate": 9.352718078381795e-07, "epoch": 0.4438964241676942 }, { "current_steps": 360, "loss": 2.1655, "learning_rate": 9.350189633375474e-07, "epoch": 0.44512946979038226 }, { "current_steps": 361, "loss": 2.2915, "learning_rate": 9.347661188369153e-07, "epoch": 0.44636251541307026 }, { "current_steps": 362, "loss": 2.5272, "learning_rate": 9.345132743362831e-07, "epoch": 0.4475955610357583 }, { "current_steps": 363, "loss": 2.4146, "learning_rate": 9.34260429835651e-07, "epoch": 0.44882860665844637 }, { "current_steps": 364, "loss": 2.5728, "learning_rate": 9.340075853350189e-07, "epoch": 0.45006165228113443 }, { "current_steps": 365, "loss": 1.9795, "learning_rate": 9.337547408343867e-07, "epoch": 0.45129469790382243 }, { "current_steps": 366, "loss": 2.5236, "learning_rate": 9.335018963337546e-07, "epoch": 0.4525277435265105 }, { "current_steps": 367, "loss": 2.0682, "learning_rate": 9.332490518331226e-07, "epoch": 0.45376078914919854 }, { "current_steps": 368, "loss": 2.6695, "learning_rate": 9.329962073324905e-07, "epoch": 0.45499383477188654 }, { "current_steps": 369, "loss": 2.4756, "learning_rate": 9.327433628318584e-07, "epoch": 0.4562268803945746 }, { "current_steps": 370, "loss": 2.2454, "learning_rate": 9.324905183312263e-07, "epoch": 0.45745992601726265 }, { "current_steps": 371, "loss": 2.1364, "learning_rate": 9.322376738305942e-07, "epoch": 0.45869297163995065 }, { "current_steps": 372, "loss": 2.5393, "learning_rate": 9.319848293299621e-07, "epoch": 0.4599260172626387 }, { "current_steps": 373, "loss": 2.7397, "learning_rate": 9.3173198482933e-07, "epoch": 0.46115906288532676 }, { "current_steps": 374, "loss": 2.2007, "learning_rate": 9.314791403286978e-07, "epoch": 0.4623921085080148 }, { "current_steps": 375, "loss": 2.1814, "learning_rate": 9.312262958280656e-07, "epoch": 0.4636251541307028 }, { "current_steps": 376, "loss": 2.0149, "learning_rate": 9.309734513274335e-07, "epoch": 0.4648581997533909 }, { "current_steps": 377, "loss": 2.2889, "learning_rate": 9.307206068268014e-07, "epoch": 0.46609124537607893 }, { "current_steps": 378, "loss": 2.2615, "learning_rate": 9.304677623261694e-07, "epoch": 0.46732429099876693 }, { "current_steps": 379, "loss": 1.9836, "learning_rate": 9.302149178255373e-07, "epoch": 0.468557336621455 }, { "current_steps": 380, "loss": 2.1572, "learning_rate": 9.299620733249052e-07, "epoch": 0.46979038224414305 }, { "current_steps": 381, "loss": 1.9217, "learning_rate": 9.29709228824273e-07, "epoch": 0.47102342786683105 }, { "current_steps": 382, "loss": 2.3144, "learning_rate": 9.294563843236409e-07, "epoch": 0.4722564734895191 }, { "current_steps": 383, "loss": 2.7753, "learning_rate": 9.292035398230088e-07, "epoch": 0.47348951911220716 }, { "current_steps": 384, "loss": 2.3561, "learning_rate": 9.289506953223767e-07, "epoch": 0.4747225647348952 }, { "current_steps": 385, "loss": 2.3396, "learning_rate": 9.286978508217446e-07, "epoch": 0.4759556103575832 }, { "current_steps": 386, "loss": 1.7809, "learning_rate": 9.284450063211125e-07, "epoch": 0.47718865598027127 }, { "current_steps": 387, "loss": 2.2047, "learning_rate": 9.281921618204804e-07, "epoch": 0.4784217016029593 }, { "current_steps": 388, "loss": 2.4686, "learning_rate": 9.279393173198482e-07, "epoch": 0.4796547472256473 }, { "current_steps": 389, "loss": 2.3979, "learning_rate": 9.276864728192162e-07, "epoch": 0.4808877928483354 }, { "current_steps": 390, "loss": 2.6611, "learning_rate": 9.27433628318584e-07, "epoch": 0.48212083847102344 }, { "current_steps": 391, "loss": 2.1443, "learning_rate": 9.271807838179519e-07, "epoch": 0.4833538840937115 }, { "current_steps": 392, "loss": 2.1548, "learning_rate": 9.269279393173198e-07, "epoch": 0.4845869297163995 }, { "current_steps": 393, "loss": 2.3218, "learning_rate": 9.266750948166877e-07, "epoch": 0.48581997533908755 }, { "current_steps": 394, "loss": 2.6553, "learning_rate": 9.264222503160556e-07, "epoch": 0.4870530209617756 }, { "current_steps": 395, "loss": 1.874, "learning_rate": 9.261694058154235e-07, "epoch": 0.4882860665844636 }, { "current_steps": 396, "loss": 2.6436, "learning_rate": 9.259165613147914e-07, "epoch": 0.48951911220715166 }, { "current_steps": 397, "loss": 2.201, "learning_rate": 9.256637168141593e-07, "epoch": 0.4907521578298397 }, { "current_steps": 398, "loss": 2.2414, "learning_rate": 9.254108723135271e-07, "epoch": 0.4919852034525277 }, { "current_steps": 399, "loss": 2.1858, "learning_rate": 9.25158027812895e-07, "epoch": 0.4932182490752158 }, { "current_steps": 400, "loss": 2.2444, "learning_rate": 9.24905183312263e-07, "epoch": 0.49445129469790383 }, { "current_steps": 401, "loss": 2.0255, "learning_rate": 9.246523388116308e-07, "epoch": 0.4956843403205919 }, { "current_steps": 402, "loss": 2.515, "learning_rate": 9.243994943109987e-07, "epoch": 0.4969173859432799 }, { "current_steps": 403, "loss": 2.556, "learning_rate": 9.241466498103666e-07, "epoch": 0.49815043156596794 }, { "current_steps": 404, "loss": 2.3963, "learning_rate": 9.238938053097345e-07, "epoch": 0.499383477188656 }, { "current_steps": 405, "loss": 2.543, "learning_rate": 9.236409608091024e-07, "epoch": 0.5006165228113441 }, { "current_steps": 406, "loss": 2.3638, "learning_rate": 9.233881163084702e-07, "epoch": 0.5018495684340321 }, { "current_steps": 407, "loss": 2.4068, "learning_rate": 9.231352718078381e-07, "epoch": 0.5030826140567201 }, { "current_steps": 408, "loss": 2.4725, "learning_rate": 9.22882427307206e-07, "epoch": 0.5043156596794082 }, { "current_steps": 409, "loss": 1.8032, "learning_rate": 9.226295828065739e-07, "epoch": 0.5055487053020962 }, { "current_steps": 410, "loss": 1.9112, "learning_rate": 9.223767383059418e-07, "epoch": 0.5067817509247842 }, { "current_steps": 411, "loss": 2.3782, "learning_rate": 9.221238938053097e-07, "epoch": 0.5080147965474723 }, { "current_steps": 412, "loss": 2.7222, "learning_rate": 9.218710493046777e-07, "epoch": 0.5092478421701603 }, { "current_steps": 413, "loss": 2.758, "learning_rate": 9.216182048040455e-07, "epoch": 0.5104808877928483 }, { "current_steps": 414, "loss": 1.8418, "learning_rate": 9.213653603034133e-07, "epoch": 0.5117139334155364 }, { "current_steps": 415, "loss": 2.5211, "learning_rate": 9.211125158027812e-07, "epoch": 0.5129469790382244 }, { "current_steps": 416, "loss": 2.3967, "learning_rate": 9.208596713021491e-07, "epoch": 0.5141800246609125 }, { "current_steps": 417, "loss": 2.4835, "learning_rate": 9.20606826801517e-07, "epoch": 0.5154130702836005 }, { "current_steps": 418, "loss": 1.9793, "learning_rate": 9.203539823008849e-07, "epoch": 0.5166461159062885 }, { "current_steps": 419, "loss": 2.0815, "learning_rate": 9.201011378002528e-07, "epoch": 0.5178791615289766 }, { "current_steps": 420, "loss": 2.265, "learning_rate": 9.198482932996207e-07, "epoch": 0.5191122071516646 }, { "current_steps": 421, "loss": 2.1743, "learning_rate": 9.195954487989886e-07, "epoch": 0.5203452527743526 }, { "current_steps": 422, "loss": 2.4033, "learning_rate": 9.193426042983564e-07, "epoch": 0.5215782983970407 }, { "current_steps": 423, "loss": 1.6231, "learning_rate": 9.190897597977244e-07, "epoch": 0.5228113440197287 }, { "current_steps": 424, "loss": 2.2485, "learning_rate": 9.188369152970923e-07, "epoch": 0.5240443896424167 }, { "current_steps": 425, "loss": 2.4478, "learning_rate": 9.185840707964602e-07, "epoch": 0.5252774352651048 }, { "current_steps": 426, "loss": 2.2444, "learning_rate": 9.18331226295828e-07, "epoch": 0.5265104808877928 }, { "current_steps": 427, "loss": 2.2926, "learning_rate": 9.180783817951959e-07, "epoch": 0.5277435265104808 }, { "current_steps": 428, "loss": 2.0063, "learning_rate": 9.178255372945638e-07, "epoch": 0.528976572133169 }, { "current_steps": 429, "loss": 2.7166, "learning_rate": 9.175726927939317e-07, "epoch": 0.530209617755857 }, { "current_steps": 430, "loss": 2.5073, "learning_rate": 9.173198482932995e-07, "epoch": 0.531442663378545 }, { "current_steps": 431, "loss": 1.9613, "learning_rate": 9.170670037926674e-07, "epoch": 0.5326757090012331 }, { "current_steps": 432, "loss": 2.2571, "learning_rate": 9.168141592920353e-07, "epoch": 0.5339087546239211 }, { "current_steps": 433, "loss": 2.1215, "learning_rate": 9.165613147914032e-07, "epoch": 0.5351418002466092 }, { "current_steps": 434, "loss": 2.2062, "learning_rate": 9.163084702907712e-07, "epoch": 0.5363748458692972 }, { "current_steps": 435, "loss": 2.0765, "learning_rate": 9.160556257901391e-07, "epoch": 0.5376078914919852 }, { "current_steps": 436, "loss": 1.9445, "learning_rate": 9.15802781289507e-07, "epoch": 0.5388409371146733 }, { "current_steps": 437, "loss": 1.6937, "learning_rate": 9.155499367888749e-07, "epoch": 0.5400739827373613 }, { "current_steps": 438, "loss": 2.2443, "learning_rate": 9.152970922882426e-07, "epoch": 0.5413070283600493 }, { "current_steps": 439, "loss": 1.8612, "learning_rate": 9.150442477876105e-07, "epoch": 0.5425400739827374 }, { "current_steps": 440, "loss": 2.0255, "learning_rate": 9.147914032869784e-07, "epoch": 0.5437731196054254 }, { "current_steps": 441, "loss": 2.6127, "learning_rate": 9.145385587863463e-07, "epoch": 0.5450061652281134 }, { "current_steps": 442, "loss": 2.4162, "learning_rate": 9.142857142857142e-07, "epoch": 0.5462392108508015 }, { "current_steps": 443, "loss": 2.0383, "learning_rate": 9.140328697850821e-07, "epoch": 0.5474722564734895 }, { "current_steps": 444, "loss": 2.5415, "learning_rate": 9.1378002528445e-07, "epoch": 0.5487053020961775 }, { "current_steps": 445, "loss": 2.4035, "learning_rate": 9.13527180783818e-07, "epoch": 0.5499383477188656 }, { "current_steps": 446, "loss": 2.555, "learning_rate": 9.132743362831859e-07, "epoch": 0.5511713933415536 }, { "current_steps": 447, "loss": 2.5426, "learning_rate": 9.130214917825537e-07, "epoch": 0.5524044389642416 }, { "current_steps": 448, "loss": 1.9203, "learning_rate": 9.127686472819216e-07, "epoch": 0.5536374845869297 }, { "current_steps": 449, "loss": 2.8394, "learning_rate": 9.125158027812895e-07, "epoch": 0.5548705302096177 }, { "current_steps": 450, "loss": 2.4241, "learning_rate": 9.122629582806574e-07, "epoch": 0.5561035758323057 }, { "current_steps": 451, "loss": 2.2886, "learning_rate": 9.120101137800252e-07, "epoch": 0.5573366214549939 }, { "current_steps": 452, "loss": 2.5552, "learning_rate": 9.117572692793931e-07, "epoch": 0.5585696670776819 }, { "current_steps": 453, "loss": 2.0821, "learning_rate": 9.11504424778761e-07, "epoch": 0.55980271270037 }, { "current_steps": 454, "loss": 2.1095, "learning_rate": 9.112515802781289e-07, "epoch": 0.561035758323058 }, { "current_steps": 455, "loss": 2.4583, "learning_rate": 9.109987357774967e-07, "epoch": 0.562268803945746 }, { "current_steps": 456, "loss": 2.3318, "learning_rate": 9.107458912768647e-07, "epoch": 0.5635018495684341 }, { "current_steps": 457, "loss": 1.9452, "learning_rate": 9.104930467762326e-07, "epoch": 0.5647348951911221 }, { "current_steps": 458, "loss": 2.4944, "learning_rate": 9.102402022756005e-07, "epoch": 0.5659679408138101 }, { "current_steps": 459, "loss": 2.3821, "learning_rate": 9.099873577749684e-07, "epoch": 0.5672009864364982 }, { "current_steps": 460, "loss": 2.8173, "learning_rate": 9.097345132743363e-07, "epoch": 0.5684340320591862 }, { "current_steps": 461, "loss": 2.4555, "learning_rate": 9.094816687737042e-07, "epoch": 0.5696670776818742 }, { "current_steps": 462, "loss": 2.2182, "learning_rate": 9.092288242730721e-07, "epoch": 0.5709001233045623 }, { "current_steps": 463, "loss": 2.0896, "learning_rate": 9.089759797724399e-07, "epoch": 0.5721331689272503 }, { "current_steps": 464, "loss": 2.1178, "learning_rate": 9.087231352718077e-07, "epoch": 0.5733662145499383 }, { "current_steps": 465, "loss": 2.4725, "learning_rate": 9.084702907711756e-07, "epoch": 0.5745992601726264 }, { "current_steps": 466, "loss": 2.4493, "learning_rate": 9.082174462705435e-07, "epoch": 0.5758323057953144 }, { "current_steps": 467, "loss": 2.2465, "learning_rate": 9.079646017699114e-07, "epoch": 0.5770653514180024 }, { "current_steps": 468, "loss": 2.2245, "learning_rate": 9.077117572692794e-07, "epoch": 0.5782983970406905 }, { "current_steps": 469, "loss": 1.5658, "learning_rate": 9.074589127686473e-07, "epoch": 0.5795314426633785 }, { "current_steps": 470, "loss": 1.8463, "learning_rate": 9.072060682680152e-07, "epoch": 0.5807644882860666 }, { "current_steps": 471, "loss": 2.5105, "learning_rate": 9.06953223767383e-07, "epoch": 0.5819975339087546 }, { "current_steps": 472, "loss": 2.0256, "learning_rate": 9.067003792667509e-07, "epoch": 0.5832305795314426 }, { "current_steps": 473, "loss": 2.6363, "learning_rate": 9.064475347661188e-07, "epoch": 0.5844636251541308 }, { "current_steps": 474, "loss": 2.1253, "learning_rate": 9.061946902654867e-07, "epoch": 0.5856966707768188 }, { "current_steps": 475, "loss": 1.8946, "learning_rate": 9.059418457648546e-07, "epoch": 0.5869297163995068 }, { "current_steps": 476, "loss": 2.3164, "learning_rate": 9.056890012642224e-07, "epoch": 0.5881627620221949 }, { "current_steps": 477, "loss": 1.9677, "learning_rate": 9.054361567635903e-07, "epoch": 0.5893958076448829 }, { "current_steps": 478, "loss": 2.3427, "learning_rate": 9.051833122629582e-07, "epoch": 0.5906288532675709 }, { "current_steps": 479, "loss": 2.0907, "learning_rate": 9.049304677623261e-07, "epoch": 0.591861898890259 }, { "current_steps": 480, "loss": 2.6826, "learning_rate": 9.04677623261694e-07, "epoch": 0.593094944512947 }, { "current_steps": 481, "loss": 2.7021, "learning_rate": 9.044247787610619e-07, "epoch": 0.594327990135635 }, { "current_steps": 482, "loss": 2.3491, "learning_rate": 9.041719342604298e-07, "epoch": 0.5955610357583231 }, { "current_steps": 483, "loss": 2.0486, "learning_rate": 9.039190897597977e-07, "epoch": 0.5967940813810111 }, { "current_steps": 484, "loss": 2.2962, "learning_rate": 9.036662452591656e-07, "epoch": 0.5980271270036991 }, { "current_steps": 485, "loss": 2.5774, "learning_rate": 9.034134007585335e-07, "epoch": 0.5992601726263872 }, { "current_steps": 486, "loss": 2.3961, "learning_rate": 9.031605562579014e-07, "epoch": 0.6004932182490752 }, { "current_steps": 487, "loss": 1.8802, "learning_rate": 9.029077117572693e-07, "epoch": 0.6017262638717632 }, { "current_steps": 488, "loss": 2.4356, "learning_rate": 9.026548672566371e-07, "epoch": 0.6029593094944513 }, { "current_steps": 489, "loss": 2.2854, "learning_rate": 9.024020227560049e-07, "epoch": 0.6041923551171393 }, { "current_steps": 490, "loss": 2.4693, "learning_rate": 9.021491782553729e-07, "epoch": 0.6054254007398274 }, { "current_steps": 491, "loss": 2.7859, "learning_rate": 9.018963337547408e-07, "epoch": 0.6066584463625154 }, { "current_steps": 492, "loss": 2.3638, "learning_rate": 9.016434892541087e-07, "epoch": 0.6078914919852034 }, { "current_steps": 493, "loss": 2.5073, "learning_rate": 9.013906447534766e-07, "epoch": 0.6091245376078915 }, { "current_steps": 494, "loss": 2.3959, "learning_rate": 9.011378002528445e-07, "epoch": 0.6103575832305795 }, { "current_steps": 495, "loss": 2.2911, "learning_rate": 9.008849557522124e-07, "epoch": 0.6115906288532675 }, { "current_steps": 496, "loss": 2.4639, "learning_rate": 9.006321112515802e-07, "epoch": 0.6128236744759556 }, { "current_steps": 497, "loss": 2.521, "learning_rate": 9.003792667509481e-07, "epoch": 0.6140567200986436 }, { "current_steps": 498, "loss": 2.3229, "learning_rate": 9.00126422250316e-07, "epoch": 0.6152897657213316 }, { "current_steps": 499, "loss": 1.6935, "learning_rate": 8.998735777496839e-07, "epoch": 0.6165228113440198 }, { "current_steps": 500, "loss": 2.4466, "learning_rate": 8.996207332490518e-07, "epoch": 0.6177558569667078 }, { "current_steps": 501, "loss": 2.4442, "learning_rate": 8.993678887484198e-07, "epoch": 0.6189889025893958 }, { "current_steps": 502, "loss": 2.5683, "learning_rate": 8.991150442477876e-07, "epoch": 0.6202219482120839 }, { "current_steps": 503, "loss": 2.4833, "learning_rate": 8.988621997471555e-07, "epoch": 0.6214549938347719 }, { "current_steps": 504, "loss": 2.4812, "learning_rate": 8.986093552465233e-07, "epoch": 0.6226880394574599 }, { "current_steps": 505, "loss": 1.991, "learning_rate": 8.983565107458912e-07, "epoch": 0.623921085080148 }, { "current_steps": 506, "loss": 2.4268, "learning_rate": 8.981036662452591e-07, "epoch": 0.625154130702836 }, { "current_steps": 507, "loss": 2.4349, "learning_rate": 8.97850821744627e-07, "epoch": 0.6263871763255241 }, { "current_steps": 508, "loss": 1.7836, "learning_rate": 8.975979772439949e-07, "epoch": 0.6276202219482121 }, { "current_steps": 509, "loss": 2.5157, "learning_rate": 8.973451327433628e-07, "epoch": 0.6288532675709001 }, { "current_steps": 510, "loss": 2.2919, "learning_rate": 8.970922882427307e-07, "epoch": 0.6300863131935882 }, { "current_steps": 511, "loss": 2.3611, "learning_rate": 8.968394437420986e-07, "epoch": 0.6313193588162762 }, { "current_steps": 512, "loss": 2.1333, "learning_rate": 8.965865992414665e-07, "epoch": 0.6325524044389642 }, { "current_steps": 513, "loss": 2.5036, "learning_rate": 8.963337547408344e-07, "epoch": 0.6337854500616523 }, { "current_steps": 514, "loss": 2.5236, "learning_rate": 8.960809102402022e-07, "epoch": 0.6350184956843403 }, { "current_steps": 515, "loss": 2.4812, "learning_rate": 8.958280657395701e-07, "epoch": 0.6362515413070283 }, { "current_steps": 516, "loss": 2.4675, "learning_rate": 8.95575221238938e-07, "epoch": 0.6374845869297164 }, { "current_steps": 517, "loss": 2.3742, "learning_rate": 8.953223767383059e-07, "epoch": 0.6387176325524044 }, { "current_steps": 518, "loss": 2.5053, "learning_rate": 8.950695322376738e-07, "epoch": 0.6399506781750924 }, { "current_steps": 519, "loss": 2.2722, "learning_rate": 8.948166877370417e-07, "epoch": 0.6411837237977805 }, { "current_steps": 520, "loss": 2.5343, "learning_rate": 8.945638432364095e-07, "epoch": 0.6424167694204685 }, { "current_steps": 521, "loss": 2.4107, "learning_rate": 8.943109987357774e-07, "epoch": 0.6436498150431565 }, { "current_steps": 522, "loss": 2.1502, "learning_rate": 8.940581542351453e-07, "epoch": 0.6448828606658447 }, { "current_steps": 523, "loss": 2.4172, "learning_rate": 8.938053097345132e-07, "epoch": 0.6461159062885327 }, { "current_steps": 524, "loss": 2.1924, "learning_rate": 8.935524652338812e-07, "epoch": 0.6473489519112207 }, { "current_steps": 525, "loss": 2.3346, "learning_rate": 8.932996207332491e-07, "epoch": 0.6485819975339088 }, { "current_steps": 526, "loss": 2.3331, "learning_rate": 8.93046776232617e-07, "epoch": 0.6498150431565968 }, { "current_steps": 527, "loss": 2.4975, "learning_rate": 8.927939317319848e-07, "epoch": 0.6510480887792849 }, { "current_steps": 528, "loss": 1.7049, "learning_rate": 8.925410872313526e-07, "epoch": 0.6522811344019729 }, { "current_steps": 529, "loss": 2.2224, "learning_rate": 8.922882427307205e-07, "epoch": 0.6535141800246609 }, { "current_steps": 530, "loss": 2.0792, "learning_rate": 8.920353982300884e-07, "epoch": 0.654747225647349 }, { "current_steps": 531, "loss": 2.0314, "learning_rate": 8.917825537294563e-07, "epoch": 0.655980271270037 }, { "current_steps": 532, "loss": 1.9269, "learning_rate": 8.915297092288242e-07, "epoch": 0.657213316892725 }, { "current_steps": 533, "loss": 2.0558, "learning_rate": 8.912768647281921e-07, "epoch": 0.6584463625154131 }, { "current_steps": 534, "loss": 1.7351, "learning_rate": 8.9102402022756e-07, "epoch": 0.6596794081381011 }, { "current_steps": 535, "loss": 2.185, "learning_rate": 8.90771175726928e-07, "epoch": 0.6609124537607891 }, { "current_steps": 536, "loss": 2.4034, "learning_rate": 8.905183312262959e-07, "epoch": 0.6621454993834772 }, { "current_steps": 537, "loss": 1.9719, "learning_rate": 8.902654867256637e-07, "epoch": 0.6633785450061652 }, { "current_steps": 538, "loss": 2.5571, "learning_rate": 8.900126422250316e-07, "epoch": 0.6646115906288532 }, { "current_steps": 539, "loss": 1.7914, "learning_rate": 8.897597977243994e-07, "epoch": 0.6658446362515413 }, { "current_steps": 540, "loss": 2.2456, "learning_rate": 8.895069532237673e-07, "epoch": 0.6670776818742293 }, { "current_steps": 541, "loss": 2.1826, "learning_rate": 8.892541087231352e-07, "epoch": 0.6683107274969173 }, { "current_steps": 542, "loss": 2.2003, "learning_rate": 8.890012642225031e-07, "epoch": 0.6695437731196054 }, { "current_steps": 543, "loss": 2.2176, "learning_rate": 8.88748419721871e-07, "epoch": 0.6707768187422934 }, { "current_steps": 544, "loss": 2.3611, "learning_rate": 8.884955752212388e-07, "epoch": 0.6720098643649816 }, { "current_steps": 545, "loss": 2.5022, "learning_rate": 8.882427307206067e-07, "epoch": 0.6732429099876696 }, { "current_steps": 546, "loss": 2.1843, "learning_rate": 8.879898862199747e-07, "epoch": 0.6744759556103576 }, { "current_steps": 547, "loss": 2.5071, "learning_rate": 8.877370417193426e-07, "epoch": 0.6757090012330457 }, { "current_steps": 548, "loss": 1.8036, "learning_rate": 8.874841972187105e-07, "epoch": 0.6769420468557337 }, { "current_steps": 549, "loss": 2.4599, "learning_rate": 8.872313527180784e-07, "epoch": 0.6781750924784217 }, { "current_steps": 550, "loss": 2.6436, "learning_rate": 8.869785082174463e-07, "epoch": 0.6794081381011098 }, { "current_steps": 551, "loss": 2.0334, "learning_rate": 8.867256637168142e-07, "epoch": 0.6806411837237978 }, { "current_steps": 552, "loss": 2.5083, "learning_rate": 8.86472819216182e-07, "epoch": 0.6818742293464858 }, { "current_steps": 553, "loss": 1.8982, "learning_rate": 8.862199747155498e-07, "epoch": 0.6831072749691739 }, { "current_steps": 554, "loss": 2.2161, "learning_rate": 8.859671302149177e-07, "epoch": 0.6843403205918619 }, { "current_steps": 555, "loss": 2.1756, "learning_rate": 8.857142857142856e-07, "epoch": 0.6855733662145499 }, { "current_steps": 556, "loss": 2.3905, "learning_rate": 8.854614412136535e-07, "epoch": 0.686806411837238 }, { "current_steps": 557, "loss": 2.4446, "learning_rate": 8.852085967130215e-07, "epoch": 0.688039457459926 }, { "current_steps": 558, "loss": 1.9184, "learning_rate": 8.849557522123894e-07, "epoch": 0.689272503082614 }, { "current_steps": 559, "loss": 2.3313, "learning_rate": 8.847029077117573e-07, "epoch": 0.6905055487053021 }, { "current_steps": 560, "loss": 2.4423, "learning_rate": 8.844500632111252e-07, "epoch": 0.6917385943279901 }, { "current_steps": 561, "loss": 1.8181, "learning_rate": 8.84197218710493e-07, "epoch": 0.6929716399506781 }, { "current_steps": 562, "loss": 1.9047, "learning_rate": 8.839443742098609e-07, "epoch": 0.6942046855733662 }, { "current_steps": 563, "loss": 2.2613, "learning_rate": 8.836915297092288e-07, "epoch": 0.6954377311960542 }, { "current_steps": 564, "loss": 2.3782, "learning_rate": 8.834386852085967e-07, "epoch": 0.6966707768187423 }, { "current_steps": 565, "loss": 2.494, "learning_rate": 8.831858407079645e-07, "epoch": 0.6979038224414303 }, { "current_steps": 566, "loss": 2.0779, "learning_rate": 8.829329962073324e-07, "epoch": 0.6991368680641183 }, { "current_steps": 567, "loss": 2.2494, "learning_rate": 8.826801517067003e-07, "epoch": 0.7003699136868065 }, { "current_steps": 568, "loss": 2.4136, "learning_rate": 8.824273072060683e-07, "epoch": 0.7016029593094945 }, { "current_steps": 569, "loss": 2.3703, "learning_rate": 8.821744627054361e-07, "epoch": 0.7028360049321825 }, { "current_steps": 570, "loss": 2.6053, "learning_rate": 8.81921618204804e-07, "epoch": 0.7040690505548706 }, { "current_steps": 571, "loss": 2.2282, "learning_rate": 8.816687737041719e-07, "epoch": 0.7053020961775586 }, { "current_steps": 572, "loss": 1.8158, "learning_rate": 8.814159292035398e-07, "epoch": 0.7065351418002466 }, { "current_steps": 573, "loss": 2.1208, "learning_rate": 8.811630847029077e-07, "epoch": 0.7077681874229347 }, { "current_steps": 574, "loss": 2.0483, "learning_rate": 8.809102402022756e-07, "epoch": 0.7090012330456227 }, { "current_steps": 575, "loss": 2.2031, "learning_rate": 8.806573957016435e-07, "epoch": 0.7102342786683107 }, { "current_steps": 576, "loss": 2.2805, "learning_rate": 8.804045512010114e-07, "epoch": 0.7114673242909988 }, { "current_steps": 577, "loss": 2.1773, "learning_rate": 8.801517067003791e-07, "epoch": 0.7127003699136868 }, { "current_steps": 578, "loss": 2.2886, "learning_rate": 8.79898862199747e-07, "epoch": 0.7139334155363748 }, { "current_steps": 579, "loss": 1.3089, "learning_rate": 8.796460176991149e-07, "epoch": 0.7151664611590629 }, { "current_steps": 580, "loss": 2.5229, "learning_rate": 8.793931731984829e-07, "epoch": 0.7163995067817509 }, { "current_steps": 581, "loss": 2.1789, "learning_rate": 8.791403286978508e-07, "epoch": 0.717632552404439 }, { "current_steps": 582, "loss": 2.4471, "learning_rate": 8.788874841972187e-07, "epoch": 0.718865598027127 }, { "current_steps": 583, "loss": 2.4622, "learning_rate": 8.786346396965866e-07, "epoch": 0.720098643649815 }, { "current_steps": 584, "loss": 2.2438, "learning_rate": 8.783817951959545e-07, "epoch": 0.7213316892725031 }, { "current_steps": 585, "loss": 2.0152, "learning_rate": 8.781289506953224e-07, "epoch": 0.7225647348951911 }, { "current_steps": 586, "loss": 2.1345, "learning_rate": 8.778761061946902e-07, "epoch": 0.7237977805178791 }, { "current_steps": 587, "loss": 2.4258, "learning_rate": 8.776232616940581e-07, "epoch": 0.7250308261405672 }, { "current_steps": 588, "loss": 2.5192, "learning_rate": 8.77370417193426e-07, "epoch": 0.7262638717632552 }, { "current_steps": 589, "loss": 2.2613, "learning_rate": 8.771175726927939e-07, "epoch": 0.7274969173859432 }, { "current_steps": 590, "loss": 2.4079, "learning_rate": 8.768647281921617e-07, "epoch": 0.7287299630086314 }, { "current_steps": 591, "loss": 2.4555, "learning_rate": 8.766118836915297e-07, "epoch": 0.7299630086313194 }, { "current_steps": 592, "loss": 2.1649, "learning_rate": 8.763590391908976e-07, "epoch": 0.7311960542540074 }, { "current_steps": 593, "loss": 2.3463, "learning_rate": 8.761061946902655e-07, "epoch": 0.7324290998766955 }, { "current_steps": 594, "loss": 2.024, "learning_rate": 8.758533501896333e-07, "epoch": 0.7336621454993835 }, { "current_steps": 595, "loss": 2.2353, "learning_rate": 8.756005056890012e-07, "epoch": 0.7348951911220715 }, { "current_steps": 596, "loss": 2.0522, "learning_rate": 8.753476611883691e-07, "epoch": 0.7361282367447596 }, { "current_steps": 597, "loss": 2.5836, "learning_rate": 8.75094816687737e-07, "epoch": 0.7373612823674476 }, { "current_steps": 598, "loss": 2.2671, "learning_rate": 8.748419721871049e-07, "epoch": 0.7385943279901356 }, { "current_steps": 599, "loss": 2.4221, "learning_rate": 8.745891276864728e-07, "epoch": 0.7398273736128237 }, { "current_steps": 600, "loss": 2.2419, "learning_rate": 8.743362831858407e-07, "epoch": 0.7410604192355117 }, { "current_steps": 601, "loss": 1.7886, "learning_rate": 8.740834386852086e-07, "epoch": 0.7422934648581998 }, { "current_steps": 602, "loss": 2.4457, "learning_rate": 8.738305941845765e-07, "epoch": 0.7435265104808878 }, { "current_steps": 603, "loss": 2.2089, "learning_rate": 8.735777496839443e-07, "epoch": 0.7447595561035758 }, { "current_steps": 604, "loss": 1.8413, "learning_rate": 8.733249051833122e-07, "epoch": 0.7459926017262639 }, { "current_steps": 605, "loss": 2.603, "learning_rate": 8.730720606826801e-07, "epoch": 0.7472256473489519 }, { "current_steps": 606, "loss": 2.3233, "learning_rate": 8.72819216182048e-07, "epoch": 0.7484586929716399 }, { "current_steps": 607, "loss": 2.6709, "learning_rate": 8.725663716814159e-07, "epoch": 0.749691738594328 }, { "current_steps": 608, "loss": 2.036, "learning_rate": 8.723135271807838e-07, "epoch": 0.750924784217016 }, { "current_steps": 609, "loss": 1.7972, "learning_rate": 8.720606826801517e-07, "epoch": 0.752157829839704 }, { "current_steps": 610, "loss": 1.8951, "learning_rate": 8.718078381795195e-07, "epoch": 0.7533908754623921 }, { "current_steps": 611, "loss": 2.2895, "learning_rate": 8.715549936788874e-07, "epoch": 0.7546239210850801 }, { "current_steps": 612, "loss": 2.6682, "learning_rate": 8.713021491782553e-07, "epoch": 0.7558569667077681 }, { "current_steps": 613, "loss": 2.3206, "learning_rate": 8.710493046776233e-07, "epoch": 0.7570900123304563 }, { "current_steps": 614, "loss": 2.3391, "learning_rate": 8.707964601769912e-07, "epoch": 0.7583230579531443 }, { "current_steps": 615, "loss": 2.0005, "learning_rate": 8.70543615676359e-07, "epoch": 0.7595561035758323 }, { "current_steps": 616, "loss": 2.4637, "learning_rate": 8.702907711757269e-07, "epoch": 0.7607891491985204 }, { "current_steps": 617, "loss": 2.5225, "learning_rate": 8.700379266750948e-07, "epoch": 0.7620221948212084 }, { "current_steps": 618, "loss": 2.1401, "learning_rate": 8.697850821744626e-07, "epoch": 0.7632552404438965 }, { "current_steps": 619, "loss": 2.4934, "learning_rate": 8.695322376738305e-07, "epoch": 0.7644882860665845 }, { "current_steps": 620, "loss": 2.4653, "learning_rate": 8.692793931731984e-07, "epoch": 0.7657213316892725 }, { "current_steps": 621, "loss": 2.6559, "learning_rate": 8.690265486725663e-07, "epoch": 0.7669543773119606 }, { "current_steps": 622, "loss": 2.1612, "learning_rate": 8.687737041719342e-07, "epoch": 0.7681874229346486 }, { "current_steps": 623, "loss": 2.4333, "learning_rate": 8.685208596713021e-07, "epoch": 0.7694204685573366 }, { "current_steps": 624, "loss": 2.4996, "learning_rate": 8.682680151706701e-07, "epoch": 0.7706535141800247 }, { "current_steps": 625, "loss": 1.6237, "learning_rate": 8.68015170670038e-07, "epoch": 0.7718865598027127 }, { "current_steps": 626, "loss": 2.0037, "learning_rate": 8.677623261694059e-07, "epoch": 0.7731196054254007 }, { "current_steps": 627, "loss": 2.2815, "learning_rate": 8.675094816687737e-07, "epoch": 0.7743526510480888 }, { "current_steps": 628, "loss": 2.5922, "learning_rate": 8.672566371681415e-07, "epoch": 0.7755856966707768 }, { "current_steps": 629, "loss": 1.9239, "learning_rate": 8.670037926675094e-07, "epoch": 0.7768187422934648 }, { "current_steps": 630, "loss": 2.3001, "learning_rate": 8.667509481668773e-07, "epoch": 0.7780517879161529 }, { "current_steps": 631, "loss": 2.6081, "learning_rate": 8.664981036662452e-07, "epoch": 0.7792848335388409 }, { "current_steps": 632, "loss": 2.0093, "learning_rate": 8.662452591656131e-07, "epoch": 0.7805178791615289 }, { "current_steps": 633, "loss": 2.4955, "learning_rate": 8.65992414664981e-07, "epoch": 0.781750924784217 }, { "current_steps": 634, "loss": 2.343, "learning_rate": 8.657395701643488e-07, "epoch": 0.782983970406905 }, { "current_steps": 635, "loss": 2.1485, "learning_rate": 8.654867256637167e-07, "epoch": 0.7842170160295932 }, { "current_steps": 636, "loss": 1.7794, "learning_rate": 8.652338811630847e-07, "epoch": 0.7854500616522812 }, { "current_steps": 637, "loss": 2.4132, "learning_rate": 8.649810366624526e-07, "epoch": 0.7866831072749692 }, { "current_steps": 638, "loss": 2.1479, "learning_rate": 8.647281921618205e-07, "epoch": 0.7879161528976573 }, { "current_steps": 639, "loss": 2.2478, "learning_rate": 8.644753476611884e-07, "epoch": 0.7891491985203453 }, { "current_steps": 640, "loss": 1.9944, "learning_rate": 8.642225031605563e-07, "epoch": 0.7903822441430333 }, { "current_steps": 641, "loss": 2.2376, "learning_rate": 8.639696586599241e-07, "epoch": 0.7916152897657214 }, { "current_steps": 642, "loss": 2.0245, "learning_rate": 8.63716814159292e-07, "epoch": 0.7928483353884094 }, { "current_steps": 643, "loss": 2.1555, "learning_rate": 8.634639696586598e-07, "epoch": 0.7940813810110974 }, { "current_steps": 644, "loss": 1.7777, "learning_rate": 8.632111251580277e-07, "epoch": 0.7953144266337855 }, { "current_steps": 645, "loss": 2.2233, "learning_rate": 8.629582806573956e-07, "epoch": 0.7965474722564735 }, { "current_steps": 646, "loss": 2.2846, "learning_rate": 8.627054361567635e-07, "epoch": 0.7977805178791615 }, { "current_steps": 647, "loss": 2.3133, "learning_rate": 8.624525916561315e-07, "epoch": 0.7990135635018496 }, { "current_steps": 648, "loss": 1.7868, "learning_rate": 8.621997471554994e-07, "epoch": 0.8002466091245376 }, { "current_steps": 649, "loss": 2.7134, "learning_rate": 8.619469026548673e-07, "epoch": 0.8014796547472256 }, { "current_steps": 650, "loss": 2.4183, "learning_rate": 8.616940581542352e-07, "epoch": 0.8027127003699137 }, { "current_steps": 651, "loss": 2.399, "learning_rate": 8.61441213653603e-07, "epoch": 0.8039457459926017 }, { "current_steps": 652, "loss": 2.1158, "learning_rate": 8.611883691529709e-07, "epoch": 0.8051787916152897 }, { "current_steps": 653, "loss": 2.9294, "learning_rate": 8.609355246523387e-07, "epoch": 0.8064118372379778 }, { "current_steps": 654, "loss": 2.5074, "learning_rate": 8.606826801517066e-07, "epoch": 0.8076448828606658 }, { "current_steps": 655, "loss": 1.7353, "learning_rate": 8.604298356510745e-07, "epoch": 0.8088779284833539 }, { "current_steps": 656, "loss": 1.9906, "learning_rate": 8.601769911504424e-07, "epoch": 0.8101109741060419 }, { "current_steps": 657, "loss": 2.2188, "learning_rate": 8.599241466498103e-07, "epoch": 0.8113440197287299 }, { "current_steps": 658, "loss": 2.2691, "learning_rate": 8.596713021491783e-07, "epoch": 0.812577065351418 }, { "current_steps": 659, "loss": 2.0375, "learning_rate": 8.594184576485461e-07, "epoch": 0.813810110974106 }, { "current_steps": 660, "loss": 1.5384, "learning_rate": 8.59165613147914e-07, "epoch": 0.815043156596794 }, { "current_steps": 661, "loss": 1.642, "learning_rate": 8.589127686472819e-07, "epoch": 0.8162762022194822 }, { "current_steps": 662, "loss": 2.0306, "learning_rate": 8.586599241466498e-07, "epoch": 0.8175092478421702 }, { "current_steps": 663, "loss": 2.4643, "learning_rate": 8.584070796460177e-07, "epoch": 0.8187422934648582 }, { "current_steps": 664, "loss": 2.4434, "learning_rate": 8.581542351453856e-07, "epoch": 0.8199753390875463 }, { "current_steps": 665, "loss": 2.3324, "learning_rate": 8.579013906447535e-07, "epoch": 0.8212083847102343 }, { "current_steps": 666, "loss": 2.3161, "learning_rate": 8.576485461441213e-07, "epoch": 0.8224414303329223 }, { "current_steps": 667, "loss": 2.3132, "learning_rate": 8.573957016434891e-07, "epoch": 0.8236744759556104 }, { "current_steps": 668, "loss": 2.1124, "learning_rate": 8.57142857142857e-07, "epoch": 0.8249075215782984 }, { "current_steps": 669, "loss": 2.2149, "learning_rate": 8.56890012642225e-07, "epoch": 0.8261405672009864 }, { "current_steps": 670, "loss": 2.4372, "learning_rate": 8.566371681415929e-07, "epoch": 0.8273736128236745 }, { "current_steps": 671, "loss": 2.2186, "learning_rate": 8.563843236409608e-07, "epoch": 0.8286066584463625 }, { "current_steps": 672, "loss": 2.2762, "learning_rate": 8.561314791403287e-07, "epoch": 0.8298397040690506 }, { "current_steps": 673, "loss": 2.7169, "learning_rate": 8.558786346396966e-07, "epoch": 0.8310727496917386 }, { "current_steps": 674, "loss": 1.8965, "learning_rate": 8.556257901390645e-07, "epoch": 0.8323057953144266 }, { "current_steps": 675, "loss": 2.096, "learning_rate": 8.553729456384323e-07, "epoch": 0.8335388409371147 }, { "current_steps": 676, "loss": 1.9761, "learning_rate": 8.551201011378002e-07, "epoch": 0.8347718865598027 }, { "current_steps": 677, "loss": 2.2771, "learning_rate": 8.548672566371681e-07, "epoch": 0.8360049321824907 }, { "current_steps": 678, "loss": 2.2994, "learning_rate": 8.546144121365359e-07, "epoch": 0.8372379778051788 }, { "current_steps": 679, "loss": 2.5609, "learning_rate": 8.543615676359038e-07, "epoch": 0.8384710234278668 }, { "current_steps": 680, "loss": 2.091, "learning_rate": 8.541087231352717e-07, "epoch": 0.8397040690505548 }, { "current_steps": 681, "loss": 2.1868, "learning_rate": 8.538558786346397e-07, "epoch": 0.840937114673243 }, { "current_steps": 682, "loss": 2.2906, "learning_rate": 8.536030341340076e-07, "epoch": 0.842170160295931 }, { "current_steps": 683, "loss": 2.2978, "learning_rate": 8.533501896333754e-07, "epoch": 0.843403205918619 }, { "current_steps": 684, "loss": 2.3053, "learning_rate": 8.530973451327433e-07, "epoch": 0.8446362515413071 }, { "current_steps": 685, "loss": 2.4469, "learning_rate": 8.528445006321112e-07, "epoch": 0.8458692971639951 }, { "current_steps": 686, "loss": 1.8973, "learning_rate": 8.525916561314791e-07, "epoch": 0.8471023427866831 }, { "current_steps": 687, "loss": 1.9687, "learning_rate": 8.52338811630847e-07, "epoch": 0.8483353884093712 }, { "current_steps": 688, "loss": 2.2078, "learning_rate": 8.520859671302149e-07, "epoch": 0.8495684340320592 }, { "current_steps": 689, "loss": 2.754, "learning_rate": 8.518331226295828e-07, "epoch": 0.8508014796547472 }, { "current_steps": 690, "loss": 2.0939, "learning_rate": 8.515802781289507e-07, "epoch": 0.8520345252774353 }, { "current_steps": 691, "loss": 2.518, "learning_rate": 8.513274336283184e-07, "epoch": 0.8532675709001233 }, { "current_steps": 692, "loss": 2.0569, "learning_rate": 8.510745891276864e-07, "epoch": 0.8545006165228114 }, { "current_steps": 693, "loss": 2.2935, "learning_rate": 8.508217446270543e-07, "epoch": 0.8557336621454994 }, { "current_steps": 694, "loss": 2.1045, "learning_rate": 8.505689001264222e-07, "epoch": 0.8569667077681874 }, { "current_steps": 695, "loss": 1.9838, "learning_rate": 8.503160556257901e-07, "epoch": 0.8581997533908755 }, { "current_steps": 696, "loss": 2.2604, "learning_rate": 8.50063211125158e-07, "epoch": 0.8594327990135635 }, { "current_steps": 697, "loss": 1.8575, "learning_rate": 8.498103666245259e-07, "epoch": 0.8606658446362515 }, { "current_steps": 698, "loss": 1.9998, "learning_rate": 8.495575221238938e-07, "epoch": 0.8618988902589396 }, { "current_steps": 699, "loss": 2.1887, "learning_rate": 8.493046776232617e-07, "epoch": 0.8631319358816276 }, { "current_steps": 700, "loss": 2.0563, "learning_rate": 8.490518331226295e-07, "epoch": 0.8643649815043156 }, { "current_steps": 701, "loss": 2.2139, "learning_rate": 8.487989886219974e-07, "epoch": 0.8655980271270037 }, { "current_steps": 702, "loss": 2.4064, "learning_rate": 8.485461441213653e-07, "epoch": 0.8668310727496917 }, { "current_steps": 703, "loss": 2.2002, "learning_rate": 8.482932996207333e-07, "epoch": 0.8680641183723797 }, { "current_steps": 704, "loss": 1.9289, "learning_rate": 8.480404551201011e-07, "epoch": 0.8692971639950678 }, { "current_steps": 705, "loss": 2.4842, "learning_rate": 8.47787610619469e-07, "epoch": 0.8705302096177558 }, { "current_steps": 706, "loss": 2.3016, "learning_rate": 8.475347661188369e-07, "epoch": 0.8717632552404438 }, { "current_steps": 707, "loss": 1.7546, "learning_rate": 8.472819216182048e-07, "epoch": 0.872996300863132 }, { "current_steps": 708, "loss": 2.0051, "learning_rate": 8.470290771175726e-07, "epoch": 0.87422934648582 }, { "current_steps": 709, "loss": 2.0351, "learning_rate": 8.467762326169405e-07, "epoch": 0.8754623921085081 }, { "current_steps": 710, "loss": 2.1082, "learning_rate": 8.465233881163084e-07, "epoch": 0.8766954377311961 }, { "current_steps": 711, "loss": 2.2689, "learning_rate": 8.462705436156763e-07, "epoch": 0.8779284833538841 }, { "current_steps": 712, "loss": 2.4801, "learning_rate": 8.460176991150442e-07, "epoch": 0.8791615289765722 }, { "current_steps": 713, "loss": 2.0556, "learning_rate": 8.457648546144121e-07, "epoch": 0.8803945745992602 }, { "current_steps": 714, "loss": 2.305, "learning_rate": 8.455120101137801e-07, "epoch": 0.8816276202219482 }, { "current_steps": 715, "loss": 1.657, "learning_rate": 8.45259165613148e-07, "epoch": 0.8828606658446363 }, { "current_steps": 716, "loss": 1.9426, "learning_rate": 8.450063211125158e-07, "epoch": 0.8840937114673243 }, { "current_steps": 717, "loss": 2.2507, "learning_rate": 8.447534766118836e-07, "epoch": 0.8853267570900123 }, { "current_steps": 718, "loss": 2.0051, "learning_rate": 8.445006321112515e-07, "epoch": 0.8865598027127004 }, { "current_steps": 719, "loss": 2.3606, "learning_rate": 8.442477876106194e-07, "epoch": 0.8877928483353884 }, { "current_steps": 720, "loss": 2.0273, "learning_rate": 8.439949431099873e-07, "epoch": 0.8890258939580764 }, { "current_steps": 721, "loss": 2.1686, "learning_rate": 8.437420986093552e-07, "epoch": 0.8902589395807645 }, { "current_steps": 722, "loss": 2.6145, "learning_rate": 8.434892541087231e-07, "epoch": 0.8914919852034525 }, { "current_steps": 723, "loss": 2.1211, "learning_rate": 8.43236409608091e-07, "epoch": 0.8927250308261405 }, { "current_steps": 724, "loss": 1.9869, "learning_rate": 8.429835651074588e-07, "epoch": 0.8939580764488286 }, { "current_steps": 725, "loss": 2.5479, "learning_rate": 8.427307206068268e-07, "epoch": 0.8951911220715166 }, { "current_steps": 726, "loss": 2.1826, "learning_rate": 8.424778761061947e-07, "epoch": 0.8964241676942046 }, { "current_steps": 727, "loss": 2.2196, "learning_rate": 8.422250316055626e-07, "epoch": 0.8976572133168927 }, { "current_steps": 728, "loss": 2.3683, "learning_rate": 8.419721871049305e-07, "epoch": 0.8988902589395807 }, { "current_steps": 729, "loss": 1.9611, "learning_rate": 8.417193426042983e-07, "epoch": 0.9001233045622689 }, { "current_steps": 730, "loss": 2.3018, "learning_rate": 8.414664981036662e-07, "epoch": 0.9013563501849569 }, { "current_steps": 731, "loss": 2.4584, "learning_rate": 8.412136536030341e-07, "epoch": 0.9025893958076449 }, { "current_steps": 732, "loss": 2.4268, "learning_rate": 8.409608091024019e-07, "epoch": 0.903822441430333 }, { "current_steps": 733, "loss": 2.4709, "learning_rate": 8.407079646017698e-07, "epoch": 0.905055487053021 }, { "current_steps": 734, "loss": 2.5134, "learning_rate": 8.404551201011377e-07, "epoch": 0.906288532675709 }, { "current_steps": 735, "loss": 2.4824, "learning_rate": 8.402022756005056e-07, "epoch": 0.9075215782983971 }, { "current_steps": 736, "loss": 1.9354, "learning_rate": 8.399494310998735e-07, "epoch": 0.9087546239210851 }, { "current_steps": 737, "loss": 2.418, "learning_rate": 8.396965865992415e-07, "epoch": 0.9099876695437731 }, { "current_steps": 738, "loss": 2.2711, "learning_rate": 8.394437420986094e-07, "epoch": 0.9112207151664612 }, { "current_steps": 739, "loss": 2.3862, "learning_rate": 8.391908975979773e-07, "epoch": 0.9124537607891492 }, { "current_steps": 740, "loss": 2.0147, "learning_rate": 8.389380530973452e-07, "epoch": 0.9136868064118372 }, { "current_steps": 741, "loss": 2.3207, "learning_rate": 8.38685208596713e-07, "epoch": 0.9149198520345253 }, { "current_steps": 742, "loss": 2.0275, "learning_rate": 8.384323640960808e-07, "epoch": 0.9161528976572133 }, { "current_steps": 743, "loss": 2.42, "learning_rate": 8.381795195954487e-07, "epoch": 0.9173859432799013 }, { "current_steps": 744, "loss": 2.2791, "learning_rate": 8.379266750948166e-07, "epoch": 0.9186189889025894 }, { "current_steps": 745, "loss": 2.4587, "learning_rate": 8.376738305941845e-07, "epoch": 0.9198520345252774 }, { "current_steps": 746, "loss": 2.1701, "learning_rate": 8.374209860935524e-07, "epoch": 0.9210850801479655 }, { "current_steps": 747, "loss": 2.5819, "learning_rate": 8.371681415929203e-07, "epoch": 0.9223181257706535 }, { "current_steps": 748, "loss": 1.547, "learning_rate": 8.369152970922883e-07, "epoch": 0.9235511713933415 }, { "current_steps": 749, "loss": 2.6646, "learning_rate": 8.366624525916561e-07, "epoch": 0.9247842170160296 }, { "current_steps": 750, "loss": 1.6957, "learning_rate": 8.36409608091024e-07, "epoch": 0.9260172626387176 }, { "current_steps": 751, "loss": 2.3362, "learning_rate": 8.361567635903919e-07, "epoch": 0.9272503082614056 }, { "current_steps": 752, "loss": 2.2564, "learning_rate": 8.359039190897598e-07, "epoch": 0.9284833538840938 }, { "current_steps": 753, "loss": 2.2379, "learning_rate": 8.356510745891277e-07, "epoch": 0.9297163995067818 }, { "current_steps": 754, "loss": 2.0148, "learning_rate": 8.353982300884955e-07, "epoch": 0.9309494451294698 }, { "current_steps": 755, "loss": 2.6317, "learning_rate": 8.351453855878634e-07, "epoch": 0.9321824907521579 }, { "current_steps": 756, "loss": 2.201, "learning_rate": 8.348925410872313e-07, "epoch": 0.9334155363748459 }, { "current_steps": 757, "loss": 1.8944, "learning_rate": 8.346396965865991e-07, "epoch": 0.9346485819975339 }, { "current_steps": 758, "loss": 2.3253, "learning_rate": 8.34386852085967e-07, "epoch": 0.935881627620222 }, { "current_steps": 759, "loss": 2.3755, "learning_rate": 8.34134007585335e-07, "epoch": 0.93711467324291 }, { "current_steps": 760, "loss": 2.0786, "learning_rate": 8.338811630847029e-07, "epoch": 0.938347718865598 }, { "current_steps": 761, "loss": 1.8448, "learning_rate": 8.336283185840708e-07, "epoch": 0.9395807644882861 }, { "current_steps": 762, "loss": 2.1048, "learning_rate": 8.333754740834387e-07, "epoch": 0.9408138101109741 }, { "current_steps": 763, "loss": 1.9669, "learning_rate": 8.331226295828066e-07, "epoch": 0.9420468557336621 }, { "current_steps": 764, "loss": 2.019, "learning_rate": 8.328697850821745e-07, "epoch": 0.9432799013563502 }, { "current_steps": 765, "loss": 2.0553, "learning_rate": 8.326169405815423e-07, "epoch": 0.9445129469790382 }, { "current_steps": 766, "loss": 2.1567, "learning_rate": 8.323640960809102e-07, "epoch": 0.9457459926017263 }, { "current_steps": 767, "loss": 2.1477, "learning_rate": 8.32111251580278e-07, "epoch": 0.9469790382244143 }, { "current_steps": 768, "loss": 2.6458, "learning_rate": 8.318584070796459e-07, "epoch": 0.9482120838471023 }, { "current_steps": 769, "loss": 1.9325, "learning_rate": 8.316055625790138e-07, "epoch": 0.9494451294697904 }, { "current_steps": 770, "loss": 2.2892, "learning_rate": 8.313527180783818e-07, "epoch": 0.9506781750924784 }, { "current_steps": 771, "loss": 2.7826, "learning_rate": 8.310998735777497e-07, "epoch": 0.9519112207151664 }, { "current_steps": 772, "loss": 1.8195, "learning_rate": 8.308470290771176e-07, "epoch": 0.9531442663378545 }, { "current_steps": 773, "loss": 2.5274, "learning_rate": 8.305941845764854e-07, "epoch": 0.9543773119605425 }, { "current_steps": 774, "loss": 1.9181, "learning_rate": 8.303413400758533e-07, "epoch": 0.9556103575832305 }, { "current_steps": 775, "loss": 2.4266, "learning_rate": 8.300884955752212e-07, "epoch": 0.9568434032059187 }, { "current_steps": 776, "loss": 2.1789, "learning_rate": 8.298356510745891e-07, "epoch": 0.9580764488286067 }, { "current_steps": 777, "loss": 2.1692, "learning_rate": 8.29582806573957e-07, "epoch": 0.9593094944512947 }, { "current_steps": 778, "loss": 2.2067, "learning_rate": 8.293299620733249e-07, "epoch": 0.9605425400739828 }, { "current_steps": 779, "loss": 1.8932, "learning_rate": 8.290771175726927e-07, "epoch": 0.9617755856966708 }, { "current_steps": 780, "loss": 2.3289, "learning_rate": 8.288242730720606e-07, "epoch": 0.9630086313193588 }, { "current_steps": 781, "loss": 2.22, "learning_rate": 8.285714285714285e-07, "epoch": 0.9642416769420469 }, { "current_steps": 782, "loss": 2.1083, "learning_rate": 8.283185840707964e-07, "epoch": 0.9654747225647349 }, { "current_steps": 783, "loss": 1.799, "learning_rate": 8.280657395701643e-07, "epoch": 0.966707768187423 }, { "current_steps": 784, "loss": 2.3038, "learning_rate": 8.278128950695322e-07, "epoch": 0.967940813810111 }, { "current_steps": 785, "loss": 2.1657, "learning_rate": 8.275600505689001e-07, "epoch": 0.969173859432799 }, { "current_steps": 786, "loss": 2.7266, "learning_rate": 8.27307206068268e-07, "epoch": 0.9704069050554871 }, { "current_steps": 787, "loss": 2.1185, "learning_rate": 8.270543615676359e-07, "epoch": 0.9716399506781751 }, { "current_steps": 788, "loss": 1.8495, "learning_rate": 8.268015170670038e-07, "epoch": 0.9728729963008631 }, { "current_steps": 789, "loss": 2.5684, "learning_rate": 8.265486725663717e-07, "epoch": 0.9741060419235512 }, { "current_steps": 790, "loss": 2.3318, "learning_rate": 8.262958280657395e-07, "epoch": 0.9753390875462392 }, { "current_steps": 791, "loss": 2.4453, "learning_rate": 8.260429835651074e-07, "epoch": 0.9765721331689272 }, { "current_steps": 792, "loss": 2.2818, "learning_rate": 8.257901390644752e-07, "epoch": 0.9778051787916153 }, { "current_steps": 793, "loss": 2.3321, "learning_rate": 8.255372945638432e-07, "epoch": 0.9790382244143033 }, { "current_steps": 794, "loss": 1.8672, "learning_rate": 8.252844500632111e-07, "epoch": 0.9802712700369913 }, { "current_steps": 795, "loss": 2.2124, "learning_rate": 8.25031605562579e-07, "epoch": 0.9815043156596794 }, { "current_steps": 796, "loss": 1.5517, "learning_rate": 8.247787610619469e-07, "epoch": 0.9827373612823674 }, { "current_steps": 797, "loss": 2.2969, "learning_rate": 8.245259165613148e-07, "epoch": 0.9839704069050554 }, { "current_steps": 798, "loss": 2.016, "learning_rate": 8.242730720606826e-07, "epoch": 0.9852034525277436 }, { "current_steps": 799, "loss": 2.1329, "learning_rate": 8.240202275600505e-07, "epoch": 0.9864364981504316 }, { "current_steps": 800, "loss": 2.2202, "learning_rate": 8.237673830594184e-07, "epoch": 0.9876695437731196 }, { "current_steps": 801, "loss": 2.5662, "learning_rate": 8.235145385587863e-07, "epoch": 0.9889025893958077 }, { "current_steps": 802, "loss": 2.7176, "learning_rate": 8.232616940581542e-07, "epoch": 0.9901356350184957 }, { "current_steps": 803, "loss": 2.2727, "learning_rate": 8.230088495575221e-07, "epoch": 0.9913686806411838 }, { "current_steps": 804, "loss": 2.5595, "learning_rate": 8.227560050568901e-07, "epoch": 0.9926017262638718 }, { "current_steps": 805, "loss": 1.8179, "learning_rate": 8.225031605562579e-07, "epoch": 0.9938347718865598 }, { "current_steps": 806, "loss": 2.3552, "learning_rate": 8.222503160556257e-07, "epoch": 0.9950678175092479 }, { "current_steps": 807, "loss": 2.3041, "learning_rate": 8.219974715549936e-07, "epoch": 0.9963008631319359 }, { "current_steps": 808, "loss": 2.723, "learning_rate": 8.217446270543615e-07, "epoch": 0.9975339087546239 }, { "current_steps": 809, "loss": 2.1154, "learning_rate": 8.214917825537294e-07, "epoch": 0.998766954377312 }, { "current_steps": 810, "loss": 3.0939, "learning_rate": 8.212389380530973e-07, "epoch": 1.0 }, { "current_steps": 811, "loss": 2.0732, "learning_rate": 8.209860935524652e-07, "epoch": 1.0012330456226881 }, { "current_steps": 812, "loss": 2.1605, "learning_rate": 8.207332490518331e-07, "epoch": 1.002466091245376 }, { "current_steps": 813, "loss": 2.3647, "learning_rate": 8.20480404551201e-07, "epoch": 1.0036991368680641 }, { "current_steps": 814, "loss": 2.4371, "learning_rate": 8.202275600505688e-07, "epoch": 1.0049321824907522 }, { "current_steps": 815, "loss": 1.4726, "learning_rate": 8.199747155499368e-07, "epoch": 1.0061652281134401 }, { "current_steps": 816, "loss": 2.478, "learning_rate": 8.197218710493047e-07, "epoch": 1.0073982737361282 }, { "current_steps": 817, "loss": 2.4486, "learning_rate": 8.194690265486726e-07, "epoch": 1.0086313193588163 }, { "current_steps": 818, "loss": 2.301, "learning_rate": 8.192161820480404e-07, "epoch": 1.0098643649815042 }, { "current_steps": 819, "loss": 2.166, "learning_rate": 8.189633375474083e-07, "epoch": 1.0110974106041923 }, { "current_steps": 820, "loss": 2.0401, "learning_rate": 8.187104930467762e-07, "epoch": 1.0123304562268804 }, { "current_steps": 821, "loss": 2.4067, "learning_rate": 8.184576485461441e-07, "epoch": 1.0135635018495683 }, { "current_steps": 822, "loss": 2.0039, "learning_rate": 8.182048040455119e-07, "epoch": 1.0147965474722564 }, { "current_steps": 823, "loss": 2.2673, "learning_rate": 8.179519595448798e-07, "epoch": 1.0160295930949446 }, { "current_steps": 824, "loss": 2.3143, "learning_rate": 8.176991150442477e-07, "epoch": 1.0172626387176325 }, { "current_steps": 825, "loss": 2.3984, "learning_rate": 8.174462705436156e-07, "epoch": 1.0184956843403206 }, { "current_steps": 826, "loss": 2.3462, "learning_rate": 8.171934260429836e-07, "epoch": 1.0197287299630087 }, { "current_steps": 827, "loss": 2.2897, "learning_rate": 8.169405815423515e-07, "epoch": 1.0209617755856966 }, { "current_steps": 828, "loss": 2.155, "learning_rate": 8.166877370417194e-07, "epoch": 1.0221948212083847 }, { "current_steps": 829, "loss": 2.2846, "learning_rate": 8.164348925410873e-07, "epoch": 1.0234278668310728 }, { "current_steps": 830, "loss": 2.196, "learning_rate": 8.16182048040455e-07, "epoch": 1.024660912453761 }, { "current_steps": 831, "loss": 2.4415, "learning_rate": 8.159292035398229e-07, "epoch": 1.0258939580764488 }, { "current_steps": 832, "loss": 2.277, "learning_rate": 8.156763590391908e-07, "epoch": 1.027127003699137 }, { "current_steps": 833, "loss": 2.0643, "learning_rate": 8.154235145385587e-07, "epoch": 1.028360049321825 }, { "current_steps": 834, "loss": 2.2633, "learning_rate": 8.151706700379266e-07, "epoch": 1.029593094944513 }, { "current_steps": 835, "loss": 2.286, "learning_rate": 8.149178255372945e-07, "epoch": 1.030826140567201 }, { "current_steps": 836, "loss": 1.6787, "learning_rate": 8.146649810366624e-07, "epoch": 1.0320591861898891 }, { "current_steps": 837, "loss": 2.538, "learning_rate": 8.144121365360304e-07, "epoch": 1.033292231812577 }, { "current_steps": 838, "loss": 1.6554, "learning_rate": 8.141592920353983e-07, "epoch": 1.0345252774352651 }, { "current_steps": 839, "loss": 2.608, "learning_rate": 8.139064475347661e-07, "epoch": 1.0357583230579532 }, { "current_steps": 840, "loss": 2.101, "learning_rate": 8.13653603034134e-07, "epoch": 1.0369913686806411 }, { "current_steps": 841, "loss": 1.8837, "learning_rate": 8.134007585335019e-07, "epoch": 1.0382244143033292 }, { "current_steps": 842, "loss": 2.292, "learning_rate": 8.131479140328698e-07, "epoch": 1.0394574599260173 }, { "current_steps": 843, "loss": 2.0765, "learning_rate": 8.128950695322376e-07, "epoch": 1.0406905055487052 }, { "current_steps": 844, "loss": 2.0706, "learning_rate": 8.126422250316055e-07, "epoch": 1.0419235511713933 }, { "current_steps": 845, "loss": 2.0763, "learning_rate": 8.123893805309734e-07, "epoch": 1.0431565967940815 }, { "current_steps": 846, "loss": 2.0723, "learning_rate": 8.121365360303412e-07, "epoch": 1.0443896424167693 }, { "current_steps": 847, "loss": 1.6571, "learning_rate": 8.118836915297091e-07, "epoch": 1.0456226880394575 }, { "current_steps": 848, "loss": 2.0325, "learning_rate": 8.11630847029077e-07, "epoch": 1.0468557336621456 }, { "current_steps": 849, "loss": 2.3562, "learning_rate": 8.11378002528445e-07, "epoch": 1.0480887792848335 }, { "current_steps": 850, "loss": 2.1812, "learning_rate": 8.111251580278129e-07, "epoch": 1.0493218249075216 }, { "current_steps": 851, "loss": 2.3169, "learning_rate": 8.108723135271808e-07, "epoch": 1.0505548705302097 }, { "current_steps": 852, "loss": 2.2498, "learning_rate": 8.106194690265487e-07, "epoch": 1.0517879161528976 }, { "current_steps": 853, "loss": 2.2728, "learning_rate": 8.103666245259166e-07, "epoch": 1.0530209617755857 }, { "current_steps": 854, "loss": 2.2253, "learning_rate": 8.101137800252845e-07, "epoch": 1.0542540073982738 }, { "current_steps": 855, "loss": 2.3187, "learning_rate": 8.098609355246522e-07, "epoch": 1.0554870530209617 }, { "current_steps": 856, "loss": 2.1154, "learning_rate": 8.096080910240201e-07, "epoch": 1.0567200986436498 }, { "current_steps": 857, "loss": 2.0866, "learning_rate": 8.09355246523388e-07, "epoch": 1.057953144266338 }, { "current_steps": 858, "loss": 2.3064, "learning_rate": 8.091024020227559e-07, "epoch": 1.0591861898890258 }, { "current_steps": 859, "loss": 1.7755, "learning_rate": 8.088495575221238e-07, "epoch": 1.060419235511714 }, { "current_steps": 860, "loss": 1.8455, "learning_rate": 8.085967130214918e-07, "epoch": 1.061652281134402 }, { "current_steps": 861, "loss": 2.3371, "learning_rate": 8.083438685208597e-07, "epoch": 1.06288532675709 }, { "current_steps": 862, "loss": 2.3596, "learning_rate": 8.080910240202276e-07, "epoch": 1.064118372379778 }, { "current_steps": 863, "loss": 2.1116, "learning_rate": 8.078381795195954e-07, "epoch": 1.0653514180024661 }, { "current_steps": 864, "loss": 1.9959, "learning_rate": 8.075853350189633e-07, "epoch": 1.066584463625154 }, { "current_steps": 865, "loss": 2.4093, "learning_rate": 8.073324905183312e-07, "epoch": 1.0678175092478421 }, { "current_steps": 866, "loss": 2.5969, "learning_rate": 8.070796460176991e-07, "epoch": 1.0690505548705302 }, { "current_steps": 867, "loss": 1.7934, "learning_rate": 8.06826801517067e-07, "epoch": 1.0702836004932181 }, { "current_steps": 868, "loss": 1.6353, "learning_rate": 8.065739570164348e-07, "epoch": 1.0715166461159062 }, { "current_steps": 869, "loss": 2.0435, "learning_rate": 8.063211125158027e-07, "epoch": 1.0727496917385944 }, { "current_steps": 870, "loss": 1.9564, "learning_rate": 8.060682680151706e-07, "epoch": 1.0739827373612825 }, { "current_steps": 871, "loss": 2.4362, "learning_rate": 8.058154235145385e-07, "epoch": 1.0752157829839704 }, { "current_steps": 872, "loss": 2.1093, "learning_rate": 8.055625790139064e-07, "epoch": 1.0764488286066585 }, { "current_steps": 873, "loss": 1.8293, "learning_rate": 8.053097345132743e-07, "epoch": 1.0776818742293466 }, { "current_steps": 874, "loss": 2.1961, "learning_rate": 8.050568900126422e-07, "epoch": 1.0789149198520345 }, { "current_steps": 875, "loss": 1.9291, "learning_rate": 8.048040455120101e-07, "epoch": 1.0801479654747226 }, { "current_steps": 876, "loss": 1.9118, "learning_rate": 8.04551201011378e-07, "epoch": 1.0813810110974107 }, { "current_steps": 877, "loss": 1.8863, "learning_rate": 8.042983565107459e-07, "epoch": 1.0826140567200986 }, { "current_steps": 878, "loss": 2.349, "learning_rate": 8.040455120101138e-07, "epoch": 1.0838471023427867 }, { "current_steps": 879, "loss": 2.1284, "learning_rate": 8.037926675094816e-07, "epoch": 1.0850801479654748 }, { "current_steps": 880, "loss": 2.0192, "learning_rate": 8.035398230088494e-07, "epoch": 1.0863131935881627 }, { "current_steps": 881, "loss": 1.9507, "learning_rate": 8.032869785082173e-07, "epoch": 1.0875462392108508 }, { "current_steps": 882, "loss": 2.7366, "learning_rate": 8.030341340075853e-07, "epoch": 1.088779284833539 }, { "current_steps": 883, "loss": 2.3407, "learning_rate": 8.027812895069532e-07, "epoch": 1.0900123304562268 }, { "current_steps": 884, "loss": 1.9969, "learning_rate": 8.025284450063211e-07, "epoch": 1.091245376078915 }, { "current_steps": 885, "loss": 2.3622, "learning_rate": 8.02275600505689e-07, "epoch": 1.092478421701603 }, { "current_steps": 886, "loss": 1.7532, "learning_rate": 8.020227560050569e-07, "epoch": 1.093711467324291 }, { "current_steps": 887, "loss": 2.2172, "learning_rate": 8.017699115044247e-07, "epoch": 1.094944512946979 }, { "current_steps": 888, "loss": 2.3606, "learning_rate": 8.015170670037926e-07, "epoch": 1.0961775585696671 }, { "current_steps": 889, "loss": 2.3918, "learning_rate": 8.012642225031605e-07, "epoch": 1.097410604192355 }, { "current_steps": 890, "loss": 2.4464, "learning_rate": 8.010113780025284e-07, "epoch": 1.0986436498150431 }, { "current_steps": 891, "loss": 1.8535, "learning_rate": 8.007585335018963e-07, "epoch": 1.0998766954377313 }, { "current_steps": 892, "loss": 1.886, "learning_rate": 8.005056890012642e-07, "epoch": 1.1011097410604191 }, { "current_steps": 893, "loss": 2.1037, "learning_rate": 8.002528445006322e-07, "epoch": 1.1023427866831073 }, { "current_steps": 894, "loss": 2.1018, "learning_rate": 8e-07, "epoch": 1.1035758323057954 }, { "current_steps": 895, "loss": 2.2487, "learning_rate": 7.997471554993679e-07, "epoch": 1.1048088779284833 }, { "current_steps": 896, "loss": 2.1537, "learning_rate": 7.994943109987357e-07, "epoch": 1.1060419235511714 }, { "current_steps": 897, "loss": 2.0752, "learning_rate": 7.992414664981036e-07, "epoch": 1.1072749691738595 }, { "current_steps": 898, "loss": 2.4705, "learning_rate": 7.989886219974715e-07, "epoch": 1.1085080147965474 }, { "current_steps": 899, "loss": 2.4388, "learning_rate": 7.987357774968394e-07, "epoch": 1.1097410604192355 }, { "current_steps": 900, "loss": 2.2571, "learning_rate": 7.984829329962073e-07, "epoch": 1.1109741060419236 }, { "current_steps": 901, "loss": 2.0337, "learning_rate": 7.982300884955752e-07, "epoch": 1.1122071516646117 }, { "current_steps": 902, "loss": 2.0208, "learning_rate": 7.979772439949431e-07, "epoch": 1.1134401972872996 }, { "current_steps": 903, "loss": 2.4637, "learning_rate": 7.97724399494311e-07, "epoch": 1.1146732429099877 }, { "current_steps": 904, "loss": 2.4718, "learning_rate": 7.974715549936788e-07, "epoch": 1.1159062885326758 }, { "current_steps": 905, "loss": 2.4534, "learning_rate": 7.972187104930468e-07, "epoch": 1.1171393341553637 }, { "current_steps": 906, "loss": 2.2623, "learning_rate": 7.969658659924146e-07, "epoch": 1.1183723797780518 }, { "current_steps": 907, "loss": 1.9764, "learning_rate": 7.967130214917825e-07, "epoch": 1.11960542540074 }, { "current_steps": 908, "loss": 2.4706, "learning_rate": 7.964601769911504e-07, "epoch": 1.1208384710234278 }, { "current_steps": 909, "loss": 2.199, "learning_rate": 7.962073324905183e-07, "epoch": 1.122071516646116 }, { "current_steps": 910, "loss": 2.3486, "learning_rate": 7.959544879898862e-07, "epoch": 1.123304562268804 }, { "current_steps": 911, "loss": 2.1522, "learning_rate": 7.957016434892541e-07, "epoch": 1.124537607891492 }, { "current_steps": 912, "loss": 2.424, "learning_rate": 7.954487989886219e-07, "epoch": 1.12577065351418 }, { "current_steps": 913, "loss": 1.8255, "learning_rate": 7.951959544879898e-07, "epoch": 1.1270036991368682 }, { "current_steps": 914, "loss": 2.0172, "learning_rate": 7.949431099873577e-07, "epoch": 1.128236744759556 }, { "current_steps": 915, "loss": 2.4164, "learning_rate": 7.946902654867256e-07, "epoch": 1.1294697903822442 }, { "current_steps": 916, "loss": 1.7638, "learning_rate": 7.944374209860936e-07, "epoch": 1.1307028360049323 }, { "current_steps": 917, "loss": 2.1057, "learning_rate": 7.941845764854615e-07, "epoch": 1.1319358816276202 }, { "current_steps": 918, "loss": 2.4875, "learning_rate": 7.939317319848294e-07, "epoch": 1.1331689272503083 }, { "current_steps": 919, "loss": 2.3816, "learning_rate": 7.936788874841972e-07, "epoch": 1.1344019728729964 }, { "current_steps": 920, "loss": 2.3146, "learning_rate": 7.93426042983565e-07, "epoch": 1.1356350184956843 }, { "current_steps": 921, "loss": 1.9847, "learning_rate": 7.931731984829329e-07, "epoch": 1.1368680641183724 }, { "current_steps": 922, "loss": 2.3637, "learning_rate": 7.929203539823008e-07, "epoch": 1.1381011097410605 }, { "current_steps": 923, "loss": 2.1051, "learning_rate": 7.926675094816687e-07, "epoch": 1.1393341553637484 }, { "current_steps": 924, "loss": 2.3846, "learning_rate": 7.924146649810366e-07, "epoch": 1.1405672009864365 }, { "current_steps": 925, "loss": 2.542, "learning_rate": 7.921618204804045e-07, "epoch": 1.1418002466091246 }, { "current_steps": 926, "loss": 1.6541, "learning_rate": 7.919089759797724e-07, "epoch": 1.1430332922318125 }, { "current_steps": 927, "loss": 1.7115, "learning_rate": 7.916561314791404e-07, "epoch": 1.1442663378545006 }, { "current_steps": 928, "loss": 2.5993, "learning_rate": 7.914032869785083e-07, "epoch": 1.1454993834771887 }, { "current_steps": 929, "loss": 2.2029, "learning_rate": 7.911504424778761e-07, "epoch": 1.1467324290998766 }, { "current_steps": 930, "loss": 2.4574, "learning_rate": 7.90897597977244e-07, "epoch": 1.1479654747225647 }, { "current_steps": 931, "loss": 2.4275, "learning_rate": 7.906447534766118e-07, "epoch": 1.1491985203452528 }, { "current_steps": 932, "loss": 2.5263, "learning_rate": 7.903919089759797e-07, "epoch": 1.1504315659679407 }, { "current_steps": 933, "loss": 1.9305, "learning_rate": 7.901390644753476e-07, "epoch": 1.1516646115906288 }, { "current_steps": 934, "loss": 1.9902, "learning_rate": 7.898862199747155e-07, "epoch": 1.152897657213317 }, { "current_steps": 935, "loss": 2.0966, "learning_rate": 7.896333754740834e-07, "epoch": 1.1541307028360048 }, { "current_steps": 936, "loss": 1.8587, "learning_rate": 7.893805309734512e-07, "epoch": 1.155363748458693 }, { "current_steps": 937, "loss": 1.9452, "learning_rate": 7.891276864728191e-07, "epoch": 1.156596794081381 }, { "current_steps": 938, "loss": 1.9366, "learning_rate": 7.888748419721871e-07, "epoch": 1.157829839704069 }, { "current_steps": 939, "loss": 2.1783, "learning_rate": 7.88621997471555e-07, "epoch": 1.159062885326757 }, { "current_steps": 940, "loss": 1.8953, "learning_rate": 7.883691529709229e-07, "epoch": 1.1602959309494452 }, { "current_steps": 941, "loss": 2.2669, "learning_rate": 7.881163084702908e-07, "epoch": 1.161528976572133 }, { "current_steps": 942, "loss": 2.1251, "learning_rate": 7.878634639696587e-07, "epoch": 1.1627620221948212 }, { "current_steps": 943, "loss": 2.1935, "learning_rate": 7.876106194690266e-07, "epoch": 1.1639950678175093 }, { "current_steps": 944, "loss": 1.7076, "learning_rate": 7.873577749683943e-07, "epoch": 1.1652281134401972 }, { "current_steps": 945, "loss": 2.5775, "learning_rate": 7.871049304677622e-07, "epoch": 1.1664611590628853 }, { "current_steps": 946, "loss": 1.8766, "learning_rate": 7.868520859671301e-07, "epoch": 1.1676942046855734 }, { "current_steps": 947, "loss": 2.3509, "learning_rate": 7.86599241466498e-07, "epoch": 1.1689272503082615 }, { "current_steps": 948, "loss": 2.0635, "learning_rate": 7.863463969658659e-07, "epoch": 1.1701602959309494 }, { "current_steps": 949, "loss": 2.0568, "learning_rate": 7.860935524652339e-07, "epoch": 1.1713933415536375 }, { "current_steps": 950, "loss": 2.2312, "learning_rate": 7.858407079646018e-07, "epoch": 1.1726263871763256 }, { "current_steps": 951, "loss": 2.3402, "learning_rate": 7.855878634639697e-07, "epoch": 1.1738594327990135 }, { "current_steps": 952, "loss": 1.8872, "learning_rate": 7.853350189633376e-07, "epoch": 1.1750924784217016 }, { "current_steps": 953, "loss": 2.1627, "learning_rate": 7.850821744627054e-07, "epoch": 1.1763255240443897 }, { "current_steps": 954, "loss": 2.0604, "learning_rate": 7.848293299620733e-07, "epoch": 1.1775585696670776 }, { "current_steps": 955, "loss": 2.1067, "learning_rate": 7.845764854614412e-07, "epoch": 1.1787916152897657 }, { "current_steps": 956, "loss": 2.1318, "learning_rate": 7.84323640960809e-07, "epoch": 1.1800246609124538 }, { "current_steps": 957, "loss": 2.5237, "learning_rate": 7.840707964601769e-07, "epoch": 1.1812577065351417 }, { "current_steps": 958, "loss": 2.2966, "learning_rate": 7.838179519595448e-07, "epoch": 1.1824907521578298 }, { "current_steps": 959, "loss": 1.9746, "learning_rate": 7.835651074589127e-07, "epoch": 1.183723797780518 }, { "current_steps": 960, "loss": 2.2066, "learning_rate": 7.833122629582806e-07, "epoch": 1.1849568434032058 }, { "current_steps": 961, "loss": 2.0128, "learning_rate": 7.830594184576485e-07, "epoch": 1.186189889025894 }, { "current_steps": 962, "loss": 2.3681, "learning_rate": 7.828065739570164e-07, "epoch": 1.187422934648582 }, { "current_steps": 963, "loss": 2.3418, "learning_rate": 7.825537294563843e-07, "epoch": 1.18865598027127 }, { "current_steps": 964, "loss": 2.3596, "learning_rate": 7.823008849557522e-07, "epoch": 1.189889025893958 }, { "current_steps": 965, "loss": 2.4667, "learning_rate": 7.820480404551201e-07, "epoch": 1.1911220715166462 }, { "current_steps": 966, "loss": 2.1095, "learning_rate": 7.81795195954488e-07, "epoch": 1.192355117139334 }, { "current_steps": 967, "loss": 2.3561, "learning_rate": 7.815423514538559e-07, "epoch": 1.1935881627620222 }, { "current_steps": 968, "loss": 2.4449, "learning_rate": 7.812895069532238e-07, "epoch": 1.1948212083847103 }, { "current_steps": 969, "loss": 2.0928, "learning_rate": 7.810366624525915e-07, "epoch": 1.1960542540073984 }, { "current_steps": 970, "loss": 2.1513, "learning_rate": 7.807838179519594e-07, "epoch": 1.1972872996300863 }, { "current_steps": 971, "loss": 2.3896, "learning_rate": 7.805309734513273e-07, "epoch": 1.1985203452527744 }, { "current_steps": 972, "loss": 2.2765, "learning_rate": 7.802781289506953e-07, "epoch": 1.1997533908754625 }, { "current_steps": 973, "loss": 1.3908, "learning_rate": 7.800252844500632e-07, "epoch": 1.2009864364981504 }, { "current_steps": 974, "loss": 2.0824, "learning_rate": 7.797724399494311e-07, "epoch": 1.2022194821208385 }, { "current_steps": 975, "loss": 2.0754, "learning_rate": 7.79519595448799e-07, "epoch": 1.2034525277435266 }, { "current_steps": 976, "loss": 2.2917, "learning_rate": 7.792667509481669e-07, "epoch": 1.2046855733662145 }, { "current_steps": 977, "loss": 2.0465, "learning_rate": 7.790139064475347e-07, "epoch": 1.2059186189889026 }, { "current_steps": 978, "loss": 2.7964, "learning_rate": 7.787610619469026e-07, "epoch": 1.2071516646115907 }, { "current_steps": 979, "loss": 1.551, "learning_rate": 7.785082174462705e-07, "epoch": 1.2083847102342786 }, { "current_steps": 980, "loss": 2.4436, "learning_rate": 7.782553729456384e-07, "epoch": 1.2096177558569667 }, { "current_steps": 981, "loss": 2.0556, "learning_rate": 7.780025284450062e-07, "epoch": 1.2108508014796548 }, { "current_steps": 982, "loss": 2.2134, "learning_rate": 7.777496839443741e-07, "epoch": 1.2120838471023427 }, { "current_steps": 983, "loss": 2.0523, "learning_rate": 7.774968394437421e-07, "epoch": 1.2133168927250308 }, { "current_steps": 984, "loss": 2.0605, "learning_rate": 7.7724399494311e-07, "epoch": 1.214549938347719 }, { "current_steps": 985, "loss": 1.88, "learning_rate": 7.769911504424778e-07, "epoch": 1.2157829839704068 }, { "current_steps": 986, "loss": 2.1515, "learning_rate": 7.767383059418457e-07, "epoch": 1.217016029593095 }, { "current_steps": 987, "loss": 2.5427, "learning_rate": 7.764854614412136e-07, "epoch": 1.218249075215783 }, { "current_steps": 988, "loss": 1.7729, "learning_rate": 7.762326169405815e-07, "epoch": 1.219482120838471 }, { "current_steps": 989, "loss": 1.7454, "learning_rate": 7.759797724399494e-07, "epoch": 1.220715166461159 }, { "current_steps": 990, "loss": 1.7557, "learning_rate": 7.757269279393173e-07, "epoch": 1.2219482120838472 }, { "current_steps": 991, "loss": 2.0233, "learning_rate": 7.754740834386852e-07, "epoch": 1.223181257706535 }, { "current_steps": 992, "loss": 2.4986, "learning_rate": 7.752212389380531e-07, "epoch": 1.2244143033292232 }, { "current_steps": 993, "loss": 2.5922, "learning_rate": 7.74968394437421e-07, "epoch": 1.2256473489519113 }, { "current_steps": 994, "loss": 2.5846, "learning_rate": 7.747155499367889e-07, "epoch": 1.2268803945745992 }, { "current_steps": 995, "loss": 2.3886, "learning_rate": 7.744627054361567e-07, "epoch": 1.2281134401972873 }, { "current_steps": 996, "loss": 1.8899, "learning_rate": 7.742098609355246e-07, "epoch": 1.2293464858199754 }, { "current_steps": 997, "loss": 2.1614, "learning_rate": 7.739570164348925e-07, "epoch": 1.2305795314426633 }, { "current_steps": 998, "loss": 2.2337, "learning_rate": 7.737041719342604e-07, "epoch": 1.2318125770653514 }, { "current_steps": 999, "loss": 2.5519, "learning_rate": 7.734513274336283e-07, "epoch": 1.2330456226880395 }, { "current_steps": 1000, "loss": 2.4568, "learning_rate": 7.731984829329962e-07, "epoch": 1.2342786683107274 }, { "current_steps": 1001, "loss": 2.6572, "learning_rate": 7.72945638432364e-07, "epoch": 1.2355117139334155 }, { "current_steps": 1002, "loss": 2.2555, "learning_rate": 7.726927939317319e-07, "epoch": 1.2367447595561036 }, { "current_steps": 1003, "loss": 2.6538, "learning_rate": 7.724399494310998e-07, "epoch": 1.2379778051787915 }, { "current_steps": 1004, "loss": 2.5598, "learning_rate": 7.721871049304677e-07, "epoch": 1.2392108508014796 }, { "current_steps": 1005, "loss": 2.4472, "learning_rate": 7.719342604298357e-07, "epoch": 1.2404438964241677 }, { "current_steps": 1006, "loss": 2.2146, "learning_rate": 7.716814159292036e-07, "epoch": 1.2416769420468556 }, { "current_steps": 1007, "loss": 1.804, "learning_rate": 7.714285714285714e-07, "epoch": 1.2429099876695437 }, { "current_steps": 1008, "loss": 2.3046, "learning_rate": 7.711757269279393e-07, "epoch": 1.2441430332922319 }, { "current_steps": 1009, "loss": 1.7193, "learning_rate": 7.709228824273072e-07, "epoch": 1.2453760789149197 }, { "current_steps": 1010, "loss": 1.3726, "learning_rate": 7.70670037926675e-07, "epoch": 1.2466091245376079 }, { "current_steps": 1011, "loss": 2.5942, "learning_rate": 7.704171934260429e-07, "epoch": 1.247842170160296 }, { "current_steps": 1012, "loss": 2.0422, "learning_rate": 7.701643489254108e-07, "epoch": 1.2490752157829839 }, { "current_steps": 1013, "loss": 2.0468, "learning_rate": 7.699115044247787e-07, "epoch": 1.250308261405672 }, { "current_steps": 1014, "loss": 2.427, "learning_rate": 7.696586599241466e-07, "epoch": 1.25154130702836 }, { "current_steps": 1015, "loss": 1.9125, "learning_rate": 7.694058154235145e-07, "epoch": 1.252774352651048 }, { "current_steps": 1016, "loss": 2.2372, "learning_rate": 7.691529709228824e-07, "epoch": 1.254007398273736 }, { "current_steps": 1017, "loss": 1.2812, "learning_rate": 7.689001264222504e-07, "epoch": 1.2552404438964242 }, { "current_steps": 1018, "loss": 1.6945, "learning_rate": 7.686472819216182e-07, "epoch": 1.256473489519112 }, { "current_steps": 1019, "loss": 2.0559, "learning_rate": 7.683944374209861e-07, "epoch": 1.2577065351418002 }, { "current_steps": 1020, "loss": 1.7192, "learning_rate": 7.681415929203539e-07, "epoch": 1.2589395807644883 }, { "current_steps": 1021, "loss": 1.7939, "learning_rate": 7.678887484197218e-07, "epoch": 1.2601726263871762 }, { "current_steps": 1022, "loss": 2.2582, "learning_rate": 7.676359039190897e-07, "epoch": 1.2614056720098643 }, { "current_steps": 1023, "loss": 2.3631, "learning_rate": 7.673830594184576e-07, "epoch": 1.2626387176325524 }, { "current_steps": 1024, "loss": 1.7351, "learning_rate": 7.671302149178255e-07, "epoch": 1.2638717632552403 }, { "current_steps": 1025, "loss": 2.3819, "learning_rate": 7.668773704171934e-07, "epoch": 1.2651048088779284 }, { "current_steps": 1026, "loss": 1.7408, "learning_rate": 7.666245259165612e-07, "epoch": 1.2663378545006165 }, { "current_steps": 1027, "loss": 2.1288, "learning_rate": 7.663716814159291e-07, "epoch": 1.2675709001233046 }, { "current_steps": 1028, "loss": 2.0447, "learning_rate": 7.661188369152971e-07, "epoch": 1.2688039457459925 }, { "current_steps": 1029, "loss": 2.3597, "learning_rate": 7.65865992414665e-07, "epoch": 1.2700369913686806 }, { "current_steps": 1030, "loss": 1.8986, "learning_rate": 7.656131479140329e-07, "epoch": 1.2712700369913688 }, { "current_steps": 1031, "loss": 2.1878, "learning_rate": 7.653603034134008e-07, "epoch": 1.2725030826140566 }, { "current_steps": 1032, "loss": 2.4578, "learning_rate": 7.651074589127686e-07, "epoch": 1.2737361282367448 }, { "current_steps": 1033, "loss": 2.3049, "learning_rate": 7.648546144121365e-07, "epoch": 1.2749691738594329 }, { "current_steps": 1034, "loss": 1.8046, "learning_rate": 7.646017699115043e-07, "epoch": 1.2762022194821208 }, { "current_steps": 1035, "loss": 2.2182, "learning_rate": 7.643489254108722e-07, "epoch": 1.2774352651048089 }, { "current_steps": 1036, "loss": 2.3767, "learning_rate": 7.640960809102401e-07, "epoch": 1.278668310727497 }, { "current_steps": 1037, "loss": 2.3575, "learning_rate": 7.63843236409608e-07, "epoch": 1.279901356350185 }, { "current_steps": 1038, "loss": 2.2475, "learning_rate": 7.635903919089759e-07, "epoch": 1.281134401972873 }, { "current_steps": 1039, "loss": 1.9286, "learning_rate": 7.633375474083439e-07, "epoch": 1.282367447595561 }, { "current_steps": 1040, "loss": 2.2316, "learning_rate": 7.630847029077118e-07, "epoch": 1.2836004932182492 }, { "current_steps": 1041, "loss": 2.9561, "learning_rate": 7.628318584070797e-07, "epoch": 1.284833538840937 }, { "current_steps": 1042, "loss": 2.2732, "learning_rate": 7.625790139064476e-07, "epoch": 1.2860665844636252 }, { "current_steps": 1043, "loss": 2.1162, "learning_rate": 7.623261694058154e-07, "epoch": 1.2872996300863133 }, { "current_steps": 1044, "loss": 2.1757, "learning_rate": 7.620733249051833e-07, "epoch": 1.2885326757090012 }, { "current_steps": 1045, "loss": 2.3465, "learning_rate": 7.618204804045511e-07, "epoch": 1.2897657213316893 }, { "current_steps": 1046, "loss": 2.1416, "learning_rate": 7.61567635903919e-07, "epoch": 1.2909987669543774 }, { "current_steps": 1047, "loss": 2.3153, "learning_rate": 7.613147914032869e-07, "epoch": 1.2922318125770653 }, { "current_steps": 1048, "loss": 2.0138, "learning_rate": 7.610619469026548e-07, "epoch": 1.2934648581997534 }, { "current_steps": 1049, "loss": 2.0917, "learning_rate": 7.608091024020227e-07, "epoch": 1.2946979038224415 }, { "current_steps": 1050, "loss": 2.225, "learning_rate": 7.605562579013907e-07, "epoch": 1.2959309494451294 }, { "current_steps": 1051, "loss": 2.5362, "learning_rate": 7.603034134007585e-07, "epoch": 1.2971639950678175 }, { "current_steps": 1052, "loss": 1.5376, "learning_rate": 7.600505689001264e-07, "epoch": 1.2983970406905057 }, { "current_steps": 1053, "loss": 1.7783, "learning_rate": 7.597977243994943e-07, "epoch": 1.2996300863131935 }, { "current_steps": 1054, "loss": 1.375, "learning_rate": 7.595448798988622e-07, "epoch": 1.3008631319358817 }, { "current_steps": 1055, "loss": 1.9354, "learning_rate": 7.592920353982301e-07, "epoch": 1.3020961775585698 }, { "current_steps": 1056, "loss": 2.1084, "learning_rate": 7.59039190897598e-07, "epoch": 1.3033292231812577 }, { "current_steps": 1057, "loss": 1.5556, "learning_rate": 7.587863463969658e-07, "epoch": 1.3045622688039458 }, { "current_steps": 1058, "loss": 1.7025, "learning_rate": 7.585335018963337e-07, "epoch": 1.3057953144266339 }, { "current_steps": 1059, "loss": 2.6113, "learning_rate": 7.582806573957015e-07, "epoch": 1.3070283600493218 }, { "current_steps": 1060, "loss": 2.3667, "learning_rate": 7.580278128950694e-07, "epoch": 1.3082614056720099 }, { "current_steps": 1061, "loss": 2.3546, "learning_rate": 7.577749683944374e-07, "epoch": 1.309494451294698 }, { "current_steps": 1062, "loss": 2.4563, "learning_rate": 7.575221238938053e-07, "epoch": 1.3107274969173859 }, { "current_steps": 1063, "loss": 1.7618, "learning_rate": 7.572692793931732e-07, "epoch": 1.311960542540074 }, { "current_steps": 1064, "loss": 2.1374, "learning_rate": 7.570164348925411e-07, "epoch": 1.313193588162762 }, { "current_steps": 1065, "loss": 1.8623, "learning_rate": 7.56763590391909e-07, "epoch": 1.31442663378545 }, { "current_steps": 1066, "loss": 1.8387, "learning_rate": 7.565107458912769e-07, "epoch": 1.315659679408138 }, { "current_steps": 1067, "loss": 1.9862, "learning_rate": 7.562579013906447e-07, "epoch": 1.3168927250308262 }, { "current_steps": 1068, "loss": 2.3574, "learning_rate": 7.560050568900126e-07, "epoch": 1.318125770653514 }, { "current_steps": 1069, "loss": 1.6243, "learning_rate": 7.557522123893805e-07, "epoch": 1.3193588162762022 }, { "current_steps": 1070, "loss": 2.111, "learning_rate": 7.554993678887483e-07, "epoch": 1.3205918618988903 }, { "current_steps": 1071, "loss": 1.7334, "learning_rate": 7.552465233881162e-07, "epoch": 1.3218249075215782 }, { "current_steps": 1072, "loss": 2.1577, "learning_rate": 7.549936788874841e-07, "epoch": 1.3230579531442663 }, { "current_steps": 1073, "loss": 2.4905, "learning_rate": 7.547408343868521e-07, "epoch": 1.3242909987669544 }, { "current_steps": 1074, "loss": 1.7223, "learning_rate": 7.5448798988622e-07, "epoch": 1.3255240443896423 }, { "current_steps": 1075, "loss": 2.0887, "learning_rate": 7.542351453855878e-07, "epoch": 1.3267570900123304 }, { "current_steps": 1076, "loss": 2.5339, "learning_rate": 7.539823008849557e-07, "epoch": 1.3279901356350186 }, { "current_steps": 1077, "loss": 2.2615, "learning_rate": 7.537294563843236e-07, "epoch": 1.3292231812577064 }, { "current_steps": 1078, "loss": 2.4096, "learning_rate": 7.534766118836915e-07, "epoch": 1.3304562268803946 }, { "current_steps": 1079, "loss": 1.6777, "learning_rate": 7.532237673830594e-07, "epoch": 1.3316892725030827 }, { "current_steps": 1080, "loss": 1.9791, "learning_rate": 7.529709228824273e-07, "epoch": 1.3329223181257706 }, { "current_steps": 1081, "loss": 2.0074, "learning_rate": 7.527180783817952e-07, "epoch": 1.3341553637484587 }, { "current_steps": 1082, "loss": 2.6001, "learning_rate": 7.524652338811631e-07, "epoch": 1.3353884093711468 }, { "current_steps": 1083, "loss": 2.6193, "learning_rate": 7.522123893805308e-07, "epoch": 1.3366214549938347 }, { "current_steps": 1084, "loss": 2.375, "learning_rate": 7.519595448798988e-07, "epoch": 1.3378545006165228 }, { "current_steps": 1085, "loss": 2.0149, "learning_rate": 7.517067003792667e-07, "epoch": 1.339087546239211 }, { "current_steps": 1086, "loss": 2.3791, "learning_rate": 7.514538558786346e-07, "epoch": 1.3403205918618988 }, { "current_steps": 1087, "loss": 2.3237, "learning_rate": 7.512010113780025e-07, "epoch": 1.341553637484587 }, { "current_steps": 1088, "loss": 2.0836, "learning_rate": 7.509481668773704e-07, "epoch": 1.342786683107275 }, { "current_steps": 1089, "loss": 1.7621, "learning_rate": 7.506953223767383e-07, "epoch": 1.344019728729963 }, { "current_steps": 1090, "loss": 1.9621, "learning_rate": 7.504424778761062e-07, "epoch": 1.345252774352651 }, { "current_steps": 1091, "loss": 2.2935, "learning_rate": 7.50189633375474e-07, "epoch": 1.3464858199753391 }, { "current_steps": 1092, "loss": 2.0011, "learning_rate": 7.499367888748419e-07, "epoch": 1.347718865598027 }, { "current_steps": 1093, "loss": 1.957, "learning_rate": 7.496839443742098e-07, "epoch": 1.3489519112207151 }, { "current_steps": 1094, "loss": 1.9948, "learning_rate": 7.494310998735777e-07, "epoch": 1.3501849568434032 }, { "current_steps": 1095, "loss": 2.2873, "learning_rate": 7.491782553729457e-07, "epoch": 1.3514180024660911 }, { "current_steps": 1096, "loss": 1.9609, "learning_rate": 7.489254108723135e-07, "epoch": 1.3526510480887792 }, { "current_steps": 1097, "loss": 2.2311, "learning_rate": 7.486725663716814e-07, "epoch": 1.3538840937114673 }, { "current_steps": 1098, "loss": 2.3757, "learning_rate": 7.484197218710493e-07, "epoch": 1.3551171393341552 }, { "current_steps": 1099, "loss": 2.1845, "learning_rate": 7.481668773704172e-07, "epoch": 1.3563501849568433 }, { "current_steps": 1100, "loss": 1.7845, "learning_rate": 7.47914032869785e-07, "epoch": 1.3575832305795315 }, { "current_steps": 1101, "loss": 2.5153, "learning_rate": 7.476611883691529e-07, "epoch": 1.3588162762022196 }, { "current_steps": 1102, "loss": 2.1714, "learning_rate": 7.474083438685208e-07, "epoch": 1.3600493218249075 }, { "current_steps": 1103, "loss": 2.7141, "learning_rate": 7.471554993678887e-07, "epoch": 1.3612823674475956 }, { "current_steps": 1104, "loss": 1.6362, "learning_rate": 7.469026548672566e-07, "epoch": 1.3625154130702837 }, { "current_steps": 1105, "loss": 2.3815, "learning_rate": 7.466498103666245e-07, "epoch": 1.3637484586929716 }, { "current_steps": 1106, "loss": 2.2599, "learning_rate": 7.463969658659925e-07, "epoch": 1.3649815043156597 }, { "current_steps": 1107, "loss": 1.9394, "learning_rate": 7.461441213653604e-07, "epoch": 1.3662145499383478 }, { "current_steps": 1108, "loss": 1.3381, "learning_rate": 7.458912768647281e-07, "epoch": 1.3674475955610357 }, { "current_steps": 1109, "loss": 2.1033, "learning_rate": 7.45638432364096e-07, "epoch": 1.3686806411837238 }, { "current_steps": 1110, "loss": 2.1555, "learning_rate": 7.453855878634639e-07, "epoch": 1.369913686806412 }, { "current_steps": 1111, "loss": 2.199, "learning_rate": 7.451327433628318e-07, "epoch": 1.3711467324291 }, { "current_steps": 1112, "loss": 2.1335, "learning_rate": 7.448798988621997e-07, "epoch": 1.372379778051788 }, { "current_steps": 1113, "loss": 2.2319, "learning_rate": 7.446270543615676e-07, "epoch": 1.373612823674476 }, { "current_steps": 1114, "loss": 2.1243, "learning_rate": 7.443742098609355e-07, "epoch": 1.3748458692971641 }, { "current_steps": 1115, "loss": 1.7553, "learning_rate": 7.441213653603034e-07, "epoch": 1.376078914919852 }, { "current_steps": 1116, "loss": 2.0212, "learning_rate": 7.438685208596712e-07, "epoch": 1.3773119605425401 }, { "current_steps": 1117, "loss": 2.3791, "learning_rate": 7.436156763590392e-07, "epoch": 1.3785450061652282 }, { "current_steps": 1118, "loss": 2.1338, "learning_rate": 7.433628318584071e-07, "epoch": 1.3797780517879161 }, { "current_steps": 1119, "loss": 2.5525, "learning_rate": 7.43109987357775e-07, "epoch": 1.3810110974106042 }, { "current_steps": 1120, "loss": 2.2189, "learning_rate": 7.428571428571429e-07, "epoch": 1.3822441430332923 }, { "current_steps": 1121, "loss": 2.3437, "learning_rate": 7.426042983565107e-07, "epoch": 1.3834771886559802 }, { "current_steps": 1122, "loss": 2.0043, "learning_rate": 7.423514538558786e-07, "epoch": 1.3847102342786684 }, { "current_steps": 1123, "loss": 2.3172, "learning_rate": 7.420986093552465e-07, "epoch": 1.3859432799013565 }, { "current_steps": 1124, "loss": 2.3278, "learning_rate": 7.418457648546143e-07, "epoch": 1.3871763255240444 }, { "current_steps": 1125, "loss": 2.418, "learning_rate": 7.415929203539822e-07, "epoch": 1.3884093711467325 }, { "current_steps": 1126, "loss": 2.2681, "learning_rate": 7.413400758533501e-07, "epoch": 1.3896424167694206 }, { "current_steps": 1127, "loss": 1.9669, "learning_rate": 7.41087231352718e-07, "epoch": 1.3908754623921085 }, { "current_steps": 1128, "loss": 2.052, "learning_rate": 7.408343868520859e-07, "epoch": 1.3921085080147966 }, { "current_steps": 1129, "loss": 2.1372, "learning_rate": 7.405815423514539e-07, "epoch": 1.3933415536374847 }, { "current_steps": 1130, "loss": 2.1138, "learning_rate": 7.403286978508218e-07, "epoch": 1.3945745992601726 }, { "current_steps": 1131, "loss": 2.0794, "learning_rate": 7.400758533501897e-07, "epoch": 1.3958076448828607 }, { "current_steps": 1132, "loss": 1.851, "learning_rate": 7.398230088495576e-07, "epoch": 1.3970406905055488 }, { "current_steps": 1133, "loss": 1.9632, "learning_rate": 7.395701643489253e-07, "epoch": 1.3982737361282367 }, { "current_steps": 1134, "loss": 2.0056, "learning_rate": 7.393173198482932e-07, "epoch": 1.3995067817509248 }, { "current_steps": 1135, "loss": 2.2714, "learning_rate": 7.390644753476611e-07, "epoch": 1.400739827373613 }, { "current_steps": 1136, "loss": 2.0032, "learning_rate": 7.38811630847029e-07, "epoch": 1.4019728729963008 }, { "current_steps": 1137, "loss": 1.796, "learning_rate": 7.385587863463969e-07, "epoch": 1.403205918618989 }, { "current_steps": 1138, "loss": 2.1444, "learning_rate": 7.383059418457648e-07, "epoch": 1.404438964241677 }, { "current_steps": 1139, "loss": 1.965, "learning_rate": 7.380530973451327e-07, "epoch": 1.405672009864365 }, { "current_steps": 1140, "loss": 2.1856, "learning_rate": 7.378002528445007e-07, "epoch": 1.406905055487053 }, { "current_steps": 1141, "loss": 1.7779, "learning_rate": 7.375474083438685e-07, "epoch": 1.4081381011097411 }, { "current_steps": 1142, "loss": 2.5575, "learning_rate": 7.372945638432364e-07, "epoch": 1.409371146732429 }, { "current_steps": 1143, "loss": 1.9737, "learning_rate": 7.370417193426043e-07, "epoch": 1.4106041923551171 }, { "current_steps": 1144, "loss": 2.1451, "learning_rate": 7.367888748419722e-07, "epoch": 1.4118372379778052 }, { "current_steps": 1145, "loss": 2.0502, "learning_rate": 7.365360303413401e-07, "epoch": 1.4130702836004931 }, { "current_steps": 1146, "loss": 2.115, "learning_rate": 7.362831858407079e-07, "epoch": 1.4143033292231812 }, { "current_steps": 1147, "loss": 2.0162, "learning_rate": 7.360303413400758e-07, "epoch": 1.4155363748458694 }, { "current_steps": 1148, "loss": 1.8077, "learning_rate": 7.357774968394436e-07, "epoch": 1.4167694204685573 }, { "current_steps": 1149, "loss": 2.2097, "learning_rate": 7.355246523388115e-07, "epoch": 1.4180024660912454 }, { "current_steps": 1150, "loss": 2.0927, "learning_rate": 7.352718078381794e-07, "epoch": 1.4192355117139335 }, { "current_steps": 1151, "loss": 2.019, "learning_rate": 7.350189633375474e-07, "epoch": 1.4204685573366214 }, { "current_steps": 1152, "loss": 1.7018, "learning_rate": 7.347661188369153e-07, "epoch": 1.4217016029593095 }, { "current_steps": 1153, "loss": 2.2789, "learning_rate": 7.345132743362832e-07, "epoch": 1.4229346485819976 }, { "current_steps": 1154, "loss": 2.3638, "learning_rate": 7.342604298356511e-07, "epoch": 1.4241676942046855 }, { "current_steps": 1155, "loss": 2.3093, "learning_rate": 7.34007585335019e-07, "epoch": 1.4254007398273736 }, { "current_steps": 1156, "loss": 2.2738, "learning_rate": 7.337547408343869e-07, "epoch": 1.4266337854500617 }, { "current_steps": 1157, "loss": 1.8068, "learning_rate": 7.335018963337547e-07, "epoch": 1.4278668310727496 }, { "current_steps": 1158, "loss": 2.2967, "learning_rate": 7.332490518331225e-07, "epoch": 1.4290998766954377 }, { "current_steps": 1159, "loss": 1.7325, "learning_rate": 7.329962073324904e-07, "epoch": 1.4303329223181258 }, { "current_steps": 1160, "loss": 1.9444, "learning_rate": 7.327433628318583e-07, "epoch": 1.4315659679408137 }, { "current_steps": 1161, "loss": 2.0711, "learning_rate": 7.324905183312262e-07, "epoch": 1.4327990135635018 }, { "current_steps": 1162, "loss": 2.2206, "learning_rate": 7.322376738305942e-07, "epoch": 1.43403205918619 }, { "current_steps": 1163, "loss": 2.2495, "learning_rate": 7.319848293299621e-07, "epoch": 1.4352651048088778 }, { "current_steps": 1164, "loss": 2.277, "learning_rate": 7.3173198482933e-07, "epoch": 1.436498150431566 }, { "current_steps": 1165, "loss": 1.8755, "learning_rate": 7.314791403286978e-07, "epoch": 1.437731196054254 }, { "current_steps": 1166, "loss": 1.7104, "learning_rate": 7.312262958280657e-07, "epoch": 1.438964241676942 }, { "current_steps": 1167, "loss": 2.2664, "learning_rate": 7.309734513274336e-07, "epoch": 1.44019728729963 }, { "current_steps": 1168, "loss": 2.0012, "learning_rate": 7.307206068268015e-07, "epoch": 1.4414303329223181 }, { "current_steps": 1169, "loss": 2.7157, "learning_rate": 7.304677623261694e-07, "epoch": 1.442663378545006 }, { "current_steps": 1170, "loss": 2.0271, "learning_rate": 7.302149178255373e-07, "epoch": 1.4438964241676941 }, { "current_steps": 1171, "loss": 2.3203, "learning_rate": 7.299620733249051e-07, "epoch": 1.4451294697903823 }, { "current_steps": 1172, "loss": 2.0559, "learning_rate": 7.29709228824273e-07, "epoch": 1.4463625154130701 }, { "current_steps": 1173, "loss": 2.2276, "learning_rate": 7.294563843236409e-07, "epoch": 1.4475955610357583 }, { "current_steps": 1174, "loss": 2.4417, "learning_rate": 7.292035398230088e-07, "epoch": 1.4488286066584464 }, { "current_steps": 1175, "loss": 1.8568, "learning_rate": 7.289506953223767e-07, "epoch": 1.4500616522811345 }, { "current_steps": 1176, "loss": 2.5128, "learning_rate": 7.286978508217446e-07, "epoch": 1.4512946979038224 }, { "current_steps": 1177, "loss": 1.279, "learning_rate": 7.284450063211125e-07, "epoch": 1.4525277435265105 }, { "current_steps": 1178, "loss": 1.8658, "learning_rate": 7.281921618204804e-07, "epoch": 1.4537607891491986 }, { "current_steps": 1179, "loss": 2.1246, "learning_rate": 7.279393173198483e-07, "epoch": 1.4549938347718865 }, { "current_steps": 1180, "loss": 2.4425, "learning_rate": 7.276864728192162e-07, "epoch": 1.4562268803945746 }, { "current_steps": 1181, "loss": 2.3133, "learning_rate": 7.27433628318584e-07, "epoch": 1.4574599260172627 }, { "current_steps": 1182, "loss": 2.5263, "learning_rate": 7.271807838179519e-07, "epoch": 1.4586929716399506 }, { "current_steps": 1183, "loss": 2.3051, "learning_rate": 7.269279393173198e-07, "epoch": 1.4599260172626387 }, { "current_steps": 1184, "loss": 2.2606, "learning_rate": 7.266750948166876e-07, "epoch": 1.4611590628853268 }, { "current_steps": 1185, "loss": 2.1186, "learning_rate": 7.264222503160556e-07, "epoch": 1.462392108508015 }, { "current_steps": 1186, "loss": 1.751, "learning_rate": 7.261694058154235e-07, "epoch": 1.4636251541307028 }, { "current_steps": 1187, "loss": 1.9432, "learning_rate": 7.259165613147914e-07, "epoch": 1.464858199753391 }, { "current_steps": 1188, "loss": 1.6322, "learning_rate": 7.256637168141593e-07, "epoch": 1.466091245376079 }, { "current_steps": 1189, "loss": 2.4228, "learning_rate": 7.254108723135271e-07, "epoch": 1.467324290998767 }, { "current_steps": 1190, "loss": 2.3414, "learning_rate": 7.25158027812895e-07, "epoch": 1.468557336621455 }, { "current_steps": 1191, "loss": 2.0731, "learning_rate": 7.249051833122629e-07, "epoch": 1.4697903822441432 }, { "current_steps": 1192, "loss": 2.2516, "learning_rate": 7.246523388116308e-07, "epoch": 1.471023427866831 }, { "current_steps": 1193, "loss": 2.2323, "learning_rate": 7.243994943109987e-07, "epoch": 1.4722564734895192 }, { "current_steps": 1194, "loss": 1.9278, "learning_rate": 7.241466498103666e-07, "epoch": 1.4734895191122073 }, { "current_steps": 1195, "loss": 2.3965, "learning_rate": 7.238938053097345e-07, "epoch": 1.4747225647348952 }, { "current_steps": 1196, "loss": 1.6417, "learning_rate": 7.236409608091025e-07, "epoch": 1.4759556103575833 }, { "current_steps": 1197, "loss": 2.3715, "learning_rate": 7.233881163084703e-07, "epoch": 1.4771886559802714 }, { "current_steps": 1198, "loss": 1.8652, "learning_rate": 7.231352718078381e-07, "epoch": 1.4784217016029593 }, { "current_steps": 1199, "loss": 2.106, "learning_rate": 7.22882427307206e-07, "epoch": 1.4796547472256474 }, { "current_steps": 1200, "loss": 2.3679, "learning_rate": 7.226295828065739e-07, "epoch": 1.4808877928483355 }, { "current_steps": 1201, "loss": 2.1009, "learning_rate": 7.223767383059418e-07, "epoch": 1.4821208384710234 }, { "current_steps": 1202, "loss": 1.8341, "learning_rate": 7.221238938053097e-07, "epoch": 1.4833538840937115 }, { "current_steps": 1203, "loss": 2.3227, "learning_rate": 7.218710493046776e-07, "epoch": 1.4845869297163996 }, { "current_steps": 1204, "loss": 2.1275, "learning_rate": 7.216182048040455e-07, "epoch": 1.4858199753390875 }, { "current_steps": 1205, "loss": 1.8227, "learning_rate": 7.213653603034134e-07, "epoch": 1.4870530209617756 }, { "current_steps": 1206, "loss": 2.3987, "learning_rate": 7.211125158027812e-07, "epoch": 1.4882860665844637 }, { "current_steps": 1207, "loss": 2.1028, "learning_rate": 7.208596713021492e-07, "epoch": 1.4895191122071516 }, { "current_steps": 1208, "loss": 2.5598, "learning_rate": 7.206068268015171e-07, "epoch": 1.4907521578298397 }, { "current_steps": 1209, "loss": 2.2907, "learning_rate": 7.203539823008849e-07, "epoch": 1.4919852034525278 }, { "current_steps": 1210, "loss": 2.1456, "learning_rate": 7.201011378002528e-07, "epoch": 1.4932182490752157 }, { "current_steps": 1211, "loss": 2.6021, "learning_rate": 7.198482932996207e-07, "epoch": 1.4944512946979038 }, { "current_steps": 1212, "loss": 2.2161, "learning_rate": 7.195954487989886e-07, "epoch": 1.495684340320592 }, { "current_steps": 1213, "loss": 1.8107, "learning_rate": 7.193426042983565e-07, "epoch": 1.4969173859432798 }, { "current_steps": 1214, "loss": 1.7602, "learning_rate": 7.190897597977243e-07, "epoch": 1.498150431565968 }, { "current_steps": 1215, "loss": 2.7177, "learning_rate": 7.188369152970922e-07, "epoch": 1.499383477188656 }, { "current_steps": 1216, "loss": 1.9132, "learning_rate": 7.185840707964601e-07, "epoch": 1.500616522811344 }, { "current_steps": 1217, "loss": 2.2019, "learning_rate": 7.18331226295828e-07, "epoch": 1.501849568434032 }, { "current_steps": 1218, "loss": 2.5031, "learning_rate": 7.18078381795196e-07, "epoch": 1.5030826140567202 }, { "current_steps": 1219, "loss": 2.3293, "learning_rate": 7.178255372945639e-07, "epoch": 1.504315659679408 }, { "current_steps": 1220, "loss": 2.1997, "learning_rate": 7.175726927939318e-07, "epoch": 1.5055487053020962 }, { "current_steps": 1221, "loss": 2.1229, "learning_rate": 7.173198482932997e-07, "epoch": 1.5067817509247843 }, { "current_steps": 1222, "loss": 2.0399, "learning_rate": 7.170670037926674e-07, "epoch": 1.5080147965474722 }, { "current_steps": 1223, "loss": 1.7436, "learning_rate": 7.168141592920353e-07, "epoch": 1.5092478421701603 }, { "current_steps": 1224, "loss": 1.6404, "learning_rate": 7.165613147914032e-07, "epoch": 1.5104808877928484 }, { "current_steps": 1225, "loss": 2.3152, "learning_rate": 7.163084702907711e-07, "epoch": 1.5117139334155363 }, { "current_steps": 1226, "loss": 2.5617, "learning_rate": 7.16055625790139e-07, "epoch": 1.5129469790382244 }, { "current_steps": 1227, "loss": 2.6331, "learning_rate": 7.158027812895069e-07, "epoch": 1.5141800246609125 }, { "current_steps": 1228, "loss": 2.1443, "learning_rate": 7.155499367888748e-07, "epoch": 1.5154130702836004 }, { "current_steps": 1229, "loss": 1.7598, "learning_rate": 7.152970922882428e-07, "epoch": 1.5166461159062885 }, { "current_steps": 1230, "loss": 2.2106, "learning_rate": 7.150442477876107e-07, "epoch": 1.5178791615289766 }, { "current_steps": 1231, "loss": 2.4234, "learning_rate": 7.147914032869785e-07, "epoch": 1.5191122071516645 }, { "current_steps": 1232, "loss": 2.3466, "learning_rate": 7.145385587863464e-07, "epoch": 1.5203452527743526 }, { "current_steps": 1233, "loss": 2.0359, "learning_rate": 7.142857142857143e-07, "epoch": 1.5215782983970407 }, { "current_steps": 1234, "loss": 1.9787, "learning_rate": 7.140328697850821e-07, "epoch": 1.5228113440197286 }, { "current_steps": 1235, "loss": 2.2009, "learning_rate": 7.1378002528445e-07, "epoch": 1.5240443896424167 }, { "current_steps": 1236, "loss": 2.2308, "learning_rate": 7.135271807838179e-07, "epoch": 1.5252774352651048 }, { "current_steps": 1237, "loss": 2.3265, "learning_rate": 7.132743362831858e-07, "epoch": 1.5265104808877927 }, { "current_steps": 1238, "loss": 2.3499, "learning_rate": 7.130214917825536e-07, "epoch": 1.5277435265104808 }, { "current_steps": 1239, "loss": 2.2876, "learning_rate": 7.127686472819215e-07, "epoch": 1.528976572133169 }, { "current_steps": 1240, "loss": 1.7809, "learning_rate": 7.125158027812894e-07, "epoch": 1.5302096177558568 }, { "current_steps": 1241, "loss": 1.8124, "learning_rate": 7.122629582806574e-07, "epoch": 1.531442663378545 }, { "current_steps": 1242, "loss": 1.7088, "learning_rate": 7.120101137800253e-07, "epoch": 1.532675709001233 }, { "current_steps": 1243, "loss": 2.4678, "learning_rate": 7.117572692793932e-07, "epoch": 1.533908754623921 }, { "current_steps": 1244, "loss": 2.2685, "learning_rate": 7.115044247787611e-07, "epoch": 1.5351418002466093 }, { "current_steps": 1245, "loss": 2.3172, "learning_rate": 7.11251580278129e-07, "epoch": 1.5363748458692972 }, { "current_steps": 1246, "loss": 2.1468, "learning_rate": 7.109987357774969e-07, "epoch": 1.537607891491985 }, { "current_steps": 1247, "loss": 2.1755, "learning_rate": 7.107458912768646e-07, "epoch": 1.5388409371146734 }, { "current_steps": 1248, "loss": 2.0865, "learning_rate": 7.104930467762325e-07, "epoch": 1.5400739827373613 }, { "current_steps": 1249, "loss": 2.405, "learning_rate": 7.102402022756004e-07, "epoch": 1.5413070283600492 }, { "current_steps": 1250, "loss": 2.1534, "learning_rate": 7.099873577749683e-07, "epoch": 1.5425400739827375 }, { "current_steps": 1251, "loss": 2.0158, "learning_rate": 7.097345132743362e-07, "epoch": 1.5437731196054254 }, { "current_steps": 1252, "loss": 1.5967, "learning_rate": 7.094816687737042e-07, "epoch": 1.5450061652281133 }, { "current_steps": 1253, "loss": 1.9648, "learning_rate": 7.092288242730721e-07, "epoch": 1.5462392108508016 }, { "current_steps": 1254, "loss": 2.0899, "learning_rate": 7.0897597977244e-07, "epoch": 1.5474722564734895 }, { "current_steps": 1255, "loss": 2.1494, "learning_rate": 7.087231352718078e-07, "epoch": 1.5487053020961774 }, { "current_steps": 1256, "loss": 1.9271, "learning_rate": 7.084702907711757e-07, "epoch": 1.5499383477188657 }, { "current_steps": 1257, "loss": 1.8169, "learning_rate": 7.082174462705436e-07, "epoch": 1.5511713933415536 }, { "current_steps": 1258, "loss": 2.1138, "learning_rate": 7.079646017699115e-07, "epoch": 1.5524044389642415 }, { "current_steps": 1259, "loss": 2.5976, "learning_rate": 7.077117572692794e-07, "epoch": 1.5536374845869299 }, { "current_steps": 1260, "loss": 2.0087, "learning_rate": 7.074589127686472e-07, "epoch": 1.5548705302096177 }, { "current_steps": 1261, "loss": 2.301, "learning_rate": 7.072060682680151e-07, "epoch": 1.5561035758323056 }, { "current_steps": 1262, "loss": 2.3515, "learning_rate": 7.06953223767383e-07, "epoch": 1.557336621454994 }, { "current_steps": 1263, "loss": 1.9943, "learning_rate": 7.067003792667509e-07, "epoch": 1.5585696670776819 }, { "current_steps": 1264, "loss": 2.1613, "learning_rate": 7.064475347661188e-07, "epoch": 1.55980271270037 }, { "current_steps": 1265, "loss": 2.2966, "learning_rate": 7.061946902654867e-07, "epoch": 1.561035758323058 }, { "current_steps": 1266, "loss": 2.1086, "learning_rate": 7.059418457648546e-07, "epoch": 1.562268803945746 }, { "current_steps": 1267, "loss": 1.9292, "learning_rate": 7.056890012642225e-07, "epoch": 1.563501849568434 }, { "current_steps": 1268, "loss": 2.333, "learning_rate": 7.054361567635904e-07, "epoch": 1.5647348951911222 }, { "current_steps": 1269, "loss": 2.2655, "learning_rate": 7.051833122629583e-07, "epoch": 1.56596794081381 }, { "current_steps": 1270, "loss": 2.4284, "learning_rate": 7.049304677623262e-07, "epoch": 1.5672009864364982 }, { "current_steps": 1271, "loss": 1.9691, "learning_rate": 7.04677623261694e-07, "epoch": 1.5684340320591863 }, { "current_steps": 1272, "loss": 1.6334, "learning_rate": 7.044247787610618e-07, "epoch": 1.5696670776818742 }, { "current_steps": 1273, "loss": 2.3688, "learning_rate": 7.041719342604297e-07, "epoch": 1.5709001233045623 }, { "current_steps": 1274, "loss": 2.3398, "learning_rate": 7.039190897597977e-07, "epoch": 1.5721331689272504 }, { "current_steps": 1275, "loss": 2.4134, "learning_rate": 7.036662452591656e-07, "epoch": 1.5733662145499383 }, { "current_steps": 1276, "loss": 2.3181, "learning_rate": 7.034134007585335e-07, "epoch": 1.5745992601726264 }, { "current_steps": 1277, "loss": 2.1849, "learning_rate": 7.031605562579014e-07, "epoch": 1.5758323057953145 }, { "current_steps": 1278, "loss": 1.602, "learning_rate": 7.029077117572693e-07, "epoch": 1.5770653514180024 }, { "current_steps": 1279, "loss": 2.4189, "learning_rate": 7.026548672566371e-07, "epoch": 1.5782983970406905 }, { "current_steps": 1280, "loss": 2.4844, "learning_rate": 7.02402022756005e-07, "epoch": 1.5795314426633786 }, { "current_steps": 1281, "loss": 2.212, "learning_rate": 7.021491782553729e-07, "epoch": 1.5807644882860665 }, { "current_steps": 1282, "loss": 2.1017, "learning_rate": 7.018963337547408e-07, "epoch": 1.5819975339087546 }, { "current_steps": 1283, "loss": 2.2122, "learning_rate": 7.016434892541087e-07, "epoch": 1.5832305795314427 }, { "current_steps": 1284, "loss": 2.326, "learning_rate": 7.013906447534766e-07, "epoch": 1.5844636251541306 }, { "current_steps": 1285, "loss": 1.9768, "learning_rate": 7.011378002528444e-07, "epoch": 1.5856966707768188 }, { "current_steps": 1286, "loss": 2.253, "learning_rate": 7.008849557522124e-07, "epoch": 1.5869297163995069 }, { "current_steps": 1287, "loss": 1.9565, "learning_rate": 7.006321112515802e-07, "epoch": 1.5881627620221948 }, { "current_steps": 1288, "loss": 1.9488, "learning_rate": 7.003792667509481e-07, "epoch": 1.5893958076448829 }, { "current_steps": 1289, "loss": 2.245, "learning_rate": 7.00126422250316e-07, "epoch": 1.590628853267571 }, { "current_steps": 1290, "loss": 1.716, "learning_rate": 6.998735777496839e-07, "epoch": 1.5918618988902589 }, { "current_steps": 1291, "loss": 1.7115, "learning_rate": 6.996207332490518e-07, "epoch": 1.593094944512947 }, { "current_steps": 1292, "loss": 1.7597, "learning_rate": 6.993678887484197e-07, "epoch": 1.594327990135635 }, { "current_steps": 1293, "loss": 1.6991, "learning_rate": 6.991150442477876e-07, "epoch": 1.595561035758323 }, { "current_steps": 1294, "loss": 2.3567, "learning_rate": 6.988621997471555e-07, "epoch": 1.596794081381011 }, { "current_steps": 1295, "loss": 1.7528, "learning_rate": 6.986093552465234e-07, "epoch": 1.5980271270036992 }, { "current_steps": 1296, "loss": 2.0128, "learning_rate": 6.983565107458912e-07, "epoch": 1.599260172626387 }, { "current_steps": 1297, "loss": 2.2105, "learning_rate": 6.981036662452592e-07, "epoch": 1.6004932182490752 }, { "current_steps": 1298, "loss": 1.5283, "learning_rate": 6.97850821744627e-07, "epoch": 1.6017262638717633 }, { "current_steps": 1299, "loss": 2.1, "learning_rate": 6.975979772439949e-07, "epoch": 1.6029593094944512 }, { "current_steps": 1300, "loss": 1.8392, "learning_rate": 6.973451327433628e-07, "epoch": 1.6041923551171393 }, { "current_steps": 1301, "loss": 1.9138, "learning_rate": 6.970922882427307e-07, "epoch": 1.6054254007398274 }, { "current_steps": 1302, "loss": 2.4691, "learning_rate": 6.968394437420986e-07, "epoch": 1.6066584463625153 }, { "current_steps": 1303, "loss": 2.4594, "learning_rate": 6.965865992414665e-07, "epoch": 1.6078914919852034 }, { "current_steps": 1304, "loss": 1.5494, "learning_rate": 6.963337547408343e-07, "epoch": 1.6091245376078915 }, { "current_steps": 1305, "loss": 2.2314, "learning_rate": 6.960809102402022e-07, "epoch": 1.6103575832305794 }, { "current_steps": 1306, "loss": 2.4251, "learning_rate": 6.958280657395701e-07, "epoch": 1.6115906288532675 }, { "current_steps": 1307, "loss": 2.0512, "learning_rate": 6.95575221238938e-07, "epoch": 1.6128236744759556 }, { "current_steps": 1308, "loss": 2.0841, "learning_rate": 6.95322376738306e-07, "epoch": 1.6140567200986435 }, { "current_steps": 1309, "loss": 2.0267, "learning_rate": 6.950695322376739e-07, "epoch": 1.6152897657213316 }, { "current_steps": 1310, "loss": 2.2607, "learning_rate": 6.948166877370417e-07, "epoch": 1.6165228113440198 }, { "current_steps": 1311, "loss": 2.2643, "learning_rate": 6.945638432364096e-07, "epoch": 1.6177558569667077 }, { "current_steps": 1312, "loss": 2.0921, "learning_rate": 6.943109987357774e-07, "epoch": 1.6189889025893958 }, { "current_steps": 1313, "loss": 1.642, "learning_rate": 6.940581542351453e-07, "epoch": 1.6202219482120839 }, { "current_steps": 1314, "loss": 1.9847, "learning_rate": 6.938053097345132e-07, "epoch": 1.6214549938347718 }, { "current_steps": 1315, "loss": 1.925, "learning_rate": 6.935524652338811e-07, "epoch": 1.6226880394574599 }, { "current_steps": 1316, "loss": 2.2076, "learning_rate": 6.93299620733249e-07, "epoch": 1.623921085080148 }, { "current_steps": 1317, "loss": 2.2897, "learning_rate": 6.930467762326169e-07, "epoch": 1.6251541307028359 }, { "current_steps": 1318, "loss": 2.2925, "learning_rate": 6.927939317319848e-07, "epoch": 1.6263871763255242 }, { "current_steps": 1319, "loss": 2.4977, "learning_rate": 6.925410872313528e-07, "epoch": 1.627620221948212 }, { "current_steps": 1320, "loss": 2.4749, "learning_rate": 6.922882427307206e-07, "epoch": 1.6288532675709 }, { "current_steps": 1321, "loss": 2.2735, "learning_rate": 6.920353982300885e-07, "epoch": 1.6300863131935883 }, { "current_steps": 1322, "loss": 2.1997, "learning_rate": 6.917825537294564e-07, "epoch": 1.6313193588162762 }, { "current_steps": 1323, "loss": 2.5001, "learning_rate": 6.915297092288242e-07, "epoch": 1.632552404438964 }, { "current_steps": 1324, "loss": 2.4425, "learning_rate": 6.912768647281921e-07, "epoch": 1.6337854500616524 }, { "current_steps": 1325, "loss": 1.8612, "learning_rate": 6.9102402022756e-07, "epoch": 1.6350184956843403 }, { "current_steps": 1326, "loss": 2.1166, "learning_rate": 6.907711757269279e-07, "epoch": 1.6362515413070282 }, { "current_steps": 1327, "loss": 2.4818, "learning_rate": 6.905183312262958e-07, "epoch": 1.6374845869297165 }, { "current_steps": 1328, "loss": 1.98, "learning_rate": 6.902654867256636e-07, "epoch": 1.6387176325524044 }, { "current_steps": 1329, "loss": 1.9891, "learning_rate": 6.900126422250315e-07, "epoch": 1.6399506781750923 }, { "current_steps": 1330, "loss": 2.3601, "learning_rate": 6.897597977243995e-07, "epoch": 1.6411837237977807 }, { "current_steps": 1331, "loss": 2.3367, "learning_rate": 6.895069532237674e-07, "epoch": 1.6424167694204685 }, { "current_steps": 1332, "loss": 2.276, "learning_rate": 6.892541087231353e-07, "epoch": 1.6436498150431564 }, { "current_steps": 1333, "loss": 2.139, "learning_rate": 6.890012642225032e-07, "epoch": 1.6448828606658448 }, { "current_steps": 1334, "loss": 1.9277, "learning_rate": 6.887484197218711e-07, "epoch": 1.6461159062885327 }, { "current_steps": 1335, "loss": 1.9766, "learning_rate": 6.88495575221239e-07, "epoch": 1.6473489519112205 }, { "current_steps": 1336, "loss": 2.3092, "learning_rate": 6.882427307206067e-07, "epoch": 1.6485819975339089 }, { "current_steps": 1337, "loss": 2.0491, "learning_rate": 6.879898862199746e-07, "epoch": 1.6498150431565968 }, { "current_steps": 1338, "loss": 2.0409, "learning_rate": 6.877370417193425e-07, "epoch": 1.6510480887792849 }, { "current_steps": 1339, "loss": 1.6072, "learning_rate": 6.874841972187104e-07, "epoch": 1.652281134401973 }, { "current_steps": 1340, "loss": 1.9862, "learning_rate": 6.872313527180783e-07, "epoch": 1.6535141800246609 }, { "current_steps": 1341, "loss": 2.436, "learning_rate": 6.869785082174462e-07, "epoch": 1.654747225647349 }, { "current_steps": 1342, "loss": 2.1815, "learning_rate": 6.867256637168142e-07, "epoch": 1.655980271270037 }, { "current_steps": 1343, "loss": 2.1158, "learning_rate": 6.864728192161821e-07, "epoch": 1.657213316892725 }, { "current_steps": 1344, "loss": 2.4003, "learning_rate": 6.8621997471555e-07, "epoch": 1.658446362515413 }, { "current_steps": 1345, "loss": 2.2183, "learning_rate": 6.859671302149178e-07, "epoch": 1.6596794081381012 }, { "current_steps": 1346, "loss": 2.5306, "learning_rate": 6.857142857142857e-07, "epoch": 1.660912453760789 }, { "current_steps": 1347, "loss": 2.4982, "learning_rate": 6.854614412136536e-07, "epoch": 1.6621454993834772 }, { "current_steps": 1348, "loss": 1.6468, "learning_rate": 6.852085967130214e-07, "epoch": 1.6633785450061653 }, { "current_steps": 1349, "loss": 2.2785, "learning_rate": 6.849557522123893e-07, "epoch": 1.6646115906288532 }, { "current_steps": 1350, "loss": 1.86, "learning_rate": 6.847029077117572e-07, "epoch": 1.6658446362515413 }, { "current_steps": 1351, "loss": 1.8637, "learning_rate": 6.844500632111251e-07, "epoch": 1.6670776818742294 }, { "current_steps": 1352, "loss": 2.4353, "learning_rate": 6.84197218710493e-07, "epoch": 1.6683107274969173 }, { "current_steps": 1353, "loss": 2.1855, "learning_rate": 6.839443742098609e-07, "epoch": 1.6695437731196054 }, { "current_steps": 1354, "loss": 2.2788, "learning_rate": 6.836915297092288e-07, "epoch": 1.6707768187422936 }, { "current_steps": 1355, "loss": 2.3886, "learning_rate": 6.834386852085967e-07, "epoch": 1.6720098643649814 }, { "current_steps": 1356, "loss": 2.2957, "learning_rate": 6.831858407079646e-07, "epoch": 1.6732429099876696 }, { "current_steps": 1357, "loss": 2.1449, "learning_rate": 6.829329962073325e-07, "epoch": 1.6744759556103577 }, { "current_steps": 1358, "loss": 2.0584, "learning_rate": 6.826801517067004e-07, "epoch": 1.6757090012330456 }, { "current_steps": 1359, "loss": 2.2083, "learning_rate": 6.824273072060683e-07, "epoch": 1.6769420468557337 }, { "current_steps": 1360, "loss": 1.8647, "learning_rate": 6.821744627054362e-07, "epoch": 1.6781750924784218 }, { "current_steps": 1361, "loss": 1.6852, "learning_rate": 6.819216182048039e-07, "epoch": 1.6794081381011097 }, { "current_steps": 1362, "loss": 2.4709, "learning_rate": 6.816687737041718e-07, "epoch": 1.6806411837237978 }, { "current_steps": 1363, "loss": 1.6157, "learning_rate": 6.814159292035397e-07, "epoch": 1.681874229346486 }, { "current_steps": 1364, "loss": 2.3439, "learning_rate": 6.811630847029077e-07, "epoch": 1.6831072749691738 }, { "current_steps": 1365, "loss": 2.1863, "learning_rate": 6.809102402022756e-07, "epoch": 1.684340320591862 }, { "current_steps": 1366, "loss": 1.8713, "learning_rate": 6.806573957016435e-07, "epoch": 1.68557336621455 }, { "current_steps": 1367, "loss": 2.366, "learning_rate": 6.804045512010114e-07, "epoch": 1.686806411837238 }, { "current_steps": 1368, "loss": 2.2507, "learning_rate": 6.801517067003793e-07, "epoch": 1.688039457459926 }, { "current_steps": 1369, "loss": 1.8357, "learning_rate": 6.798988621997471e-07, "epoch": 1.6892725030826141 }, { "current_steps": 1370, "loss": 1.9831, "learning_rate": 6.79646017699115e-07, "epoch": 1.690505548705302 }, { "current_steps": 1371, "loss": 1.7247, "learning_rate": 6.793931731984829e-07, "epoch": 1.6917385943279901 }, { "current_steps": 1372, "loss": 1.7706, "learning_rate": 6.791403286978508e-07, "epoch": 1.6929716399506782 }, { "current_steps": 1373, "loss": 1.685, "learning_rate": 6.788874841972186e-07, "epoch": 1.6942046855733661 }, { "current_steps": 1374, "loss": 1.8229, "learning_rate": 6.786346396965865e-07, "epoch": 1.6954377311960542 }, { "current_steps": 1375, "loss": 2.2308, "learning_rate": 6.783817951959545e-07, "epoch": 1.6966707768187423 }, { "current_steps": 1376, "loss": 2.755, "learning_rate": 6.781289506953224e-07, "epoch": 1.6979038224414302 }, { "current_steps": 1377, "loss": 1.7678, "learning_rate": 6.778761061946902e-07, "epoch": 1.6991368680641183 }, { "current_steps": 1378, "loss": 2.049, "learning_rate": 6.776232616940581e-07, "epoch": 1.7003699136868065 }, { "current_steps": 1379, "loss": 2.0337, "learning_rate": 6.77370417193426e-07, "epoch": 1.7016029593094943 }, { "current_steps": 1380, "loss": 1.9732, "learning_rate": 6.771175726927939e-07, "epoch": 1.7028360049321825 }, { "current_steps": 1381, "loss": 2.1827, "learning_rate": 6.768647281921618e-07, "epoch": 1.7040690505548706 }, { "current_steps": 1382, "loss": 2.2387, "learning_rate": 6.766118836915297e-07, "epoch": 1.7053020961775585 }, { "current_steps": 1383, "loss": 2.2775, "learning_rate": 6.763590391908976e-07, "epoch": 1.7065351418002466 }, { "current_steps": 1384, "loss": 2.4115, "learning_rate": 6.761061946902655e-07, "epoch": 1.7077681874229347 }, { "current_steps": 1385, "loss": 2.0823, "learning_rate": 6.758533501896333e-07, "epoch": 1.7090012330456226 }, { "current_steps": 1386, "loss": 1.2707, "learning_rate": 6.756005056890012e-07, "epoch": 1.7102342786683107 }, { "current_steps": 1387, "loss": 1.8145, "learning_rate": 6.753476611883691e-07, "epoch": 1.7114673242909988 }, { "current_steps": 1388, "loss": 1.6481, "learning_rate": 6.75094816687737e-07, "epoch": 1.7127003699136867 }, { "current_steps": 1389, "loss": 1.4407, "learning_rate": 6.748419721871049e-07, "epoch": 1.7139334155363748 }, { "current_steps": 1390, "loss": 2.2321, "learning_rate": 6.745891276864728e-07, "epoch": 1.715166461159063 }, { "current_steps": 1391, "loss": 1.8446, "learning_rate": 6.743362831858407e-07, "epoch": 1.7163995067817508 }, { "current_steps": 1392, "loss": 1.6879, "learning_rate": 6.740834386852086e-07, "epoch": 1.7176325524044391 }, { "current_steps": 1393, "loss": 1.7916, "learning_rate": 6.738305941845765e-07, "epoch": 1.718865598027127 }, { "current_steps": 1394, "loss": 2.1051, "learning_rate": 6.735777496839443e-07, "epoch": 1.720098643649815 }, { "current_steps": 1395, "loss": 2.2642, "learning_rate": 6.733249051833122e-07, "epoch": 1.7213316892725032 }, { "current_steps": 1396, "loss": 2.1317, "learning_rate": 6.730720606826801e-07, "epoch": 1.7225647348951911 }, { "current_steps": 1397, "loss": 2.3847, "learning_rate": 6.72819216182048e-07, "epoch": 1.723797780517879 }, { "current_steps": 1398, "loss": 2.1906, "learning_rate": 6.72566371681416e-07, "epoch": 1.7250308261405674 }, { "current_steps": 1399, "loss": 1.703, "learning_rate": 6.723135271807838e-07, "epoch": 1.7262638717632552 }, { "current_steps": 1400, "loss": 1.7756, "learning_rate": 6.720606826801517e-07, "epoch": 1.7274969173859431 }, { "current_steps": 1401, "loss": 2.2757, "learning_rate": 6.718078381795196e-07, "epoch": 1.7287299630086315 }, { "current_steps": 1402, "loss": 2.1181, "learning_rate": 6.715549936788874e-07, "epoch": 1.7299630086313194 }, { "current_steps": 1403, "loss": 2.2272, "learning_rate": 6.713021491782553e-07, "epoch": 1.7311960542540072 }, { "current_steps": 1404, "loss": 1.8281, "learning_rate": 6.710493046776232e-07, "epoch": 1.7324290998766956 }, { "current_steps": 1405, "loss": 2.216, "learning_rate": 6.707964601769911e-07, "epoch": 1.7336621454993835 }, { "current_steps": 1406, "loss": 2.2437, "learning_rate": 6.70543615676359e-07, "epoch": 1.7348951911220714 }, { "current_steps": 1407, "loss": 2.1799, "learning_rate": 6.702907711757269e-07, "epoch": 1.7361282367447597 }, { "current_steps": 1408, "loss": 2.4149, "learning_rate": 6.700379266750948e-07, "epoch": 1.7373612823674476 }, { "current_steps": 1409, "loss": 2.3615, "learning_rate": 6.697850821744628e-07, "epoch": 1.7385943279901355 }, { "current_steps": 1410, "loss": 1.6322, "learning_rate": 6.695322376738306e-07, "epoch": 1.7398273736128238 }, { "current_steps": 1411, "loss": 1.8041, "learning_rate": 6.692793931731984e-07, "epoch": 1.7410604192355117 }, { "current_steps": 1412, "loss": 2.1373, "learning_rate": 6.690265486725663e-07, "epoch": 1.7422934648581998 }, { "current_steps": 1413, "loss": 2.5679, "learning_rate": 6.687737041719342e-07, "epoch": 1.743526510480888 }, { "current_steps": 1414, "loss": 1.9438, "learning_rate": 6.685208596713021e-07, "epoch": 1.7447595561035758 }, { "current_steps": 1415, "loss": 1.9608, "learning_rate": 6.6826801517067e-07, "epoch": 1.745992601726264 }, { "current_steps": 1416, "loss": 2.0904, "learning_rate": 6.680151706700379e-07, "epoch": 1.747225647348952 }, { "current_steps": 1417, "loss": 2.1313, "learning_rate": 6.677623261694058e-07, "epoch": 1.74845869297164 }, { "current_steps": 1418, "loss": 1.6947, "learning_rate": 6.675094816687736e-07, "epoch": 1.749691738594328 }, { "current_steps": 1419, "loss": 2.4186, "learning_rate": 6.672566371681415e-07, "epoch": 1.7509247842170161 }, { "current_steps": 1420, "loss": 2.3097, "learning_rate": 6.670037926675095e-07, "epoch": 1.752157829839704 }, { "current_steps": 1421, "loss": 2.0202, "learning_rate": 6.667509481668774e-07, "epoch": 1.7533908754623921 }, { "current_steps": 1422, "loss": 1.8597, "learning_rate": 6.664981036662453e-07, "epoch": 1.7546239210850803 }, { "current_steps": 1423, "loss": 2.2015, "learning_rate": 6.662452591656132e-07, "epoch": 1.7558569667077681 }, { "current_steps": 1424, "loss": 2.3836, "learning_rate": 6.65992414664981e-07, "epoch": 1.7570900123304563 }, { "current_steps": 1425, "loss": 1.8132, "learning_rate": 6.657395701643489e-07, "epoch": 1.7583230579531444 }, { "current_steps": 1426, "loss": 1.7223, "learning_rate": 6.654867256637167e-07, "epoch": 1.7595561035758323 }, { "current_steps": 1427, "loss": 2.3341, "learning_rate": 6.652338811630846e-07, "epoch": 1.7607891491985204 }, { "current_steps": 1428, "loss": 1.6517, "learning_rate": 6.649810366624525e-07, "epoch": 1.7620221948212085 }, { "current_steps": 1429, "loss": 2.2171, "learning_rate": 6.647281921618204e-07, "epoch": 1.7632552404438964 }, { "current_steps": 1430, "loss": 2.1791, "learning_rate": 6.644753476611883e-07, "epoch": 1.7644882860665845 }, { "current_steps": 1431, "loss": 2.2809, "learning_rate": 6.642225031605563e-07, "epoch": 1.7657213316892726 }, { "current_steps": 1432, "loss": 2.1187, "learning_rate": 6.639696586599242e-07, "epoch": 1.7669543773119605 }, { "current_steps": 1433, "loss": 2.1663, "learning_rate": 6.637168141592921e-07, "epoch": 1.7681874229346486 }, { "current_steps": 1434, "loss": 2.2227, "learning_rate": 6.6346396965866e-07, "epoch": 1.7694204685573367 }, { "current_steps": 1435, "loss": 2.3222, "learning_rate": 6.632111251580278e-07, "epoch": 1.7706535141800246 }, { "current_steps": 1436, "loss": 2.2421, "learning_rate": 6.629582806573957e-07, "epoch": 1.7718865598027127 }, { "current_steps": 1437, "loss": 1.9961, "learning_rate": 6.627054361567635e-07, "epoch": 1.7731196054254008 }, { "current_steps": 1438, "loss": 1.9526, "learning_rate": 6.624525916561314e-07, "epoch": 1.7743526510480887 }, { "current_steps": 1439, "loss": 2.2803, "learning_rate": 6.621997471554993e-07, "epoch": 1.7755856966707768 }, { "current_steps": 1440, "loss": 2.16, "learning_rate": 6.619469026548672e-07, "epoch": 1.776818742293465 }, { "current_steps": 1441, "loss": 2.3867, "learning_rate": 6.616940581542351e-07, "epoch": 1.7780517879161528 }, { "current_steps": 1442, "loss": 1.9998, "learning_rate": 6.61441213653603e-07, "epoch": 1.779284833538841 }, { "current_steps": 1443, "loss": 2.2761, "learning_rate": 6.611883691529709e-07, "epoch": 1.780517879161529 }, { "current_steps": 1444, "loss": 2.3223, "learning_rate": 6.609355246523388e-07, "epoch": 1.781750924784217 }, { "current_steps": 1445, "loss": 2.4985, "learning_rate": 6.606826801517067e-07, "epoch": 1.782983970406905 }, { "current_steps": 1446, "loss": 2.4723, "learning_rate": 6.604298356510746e-07, "epoch": 1.7842170160295932 }, { "current_steps": 1447, "loss": 2.2496, "learning_rate": 6.601769911504425e-07, "epoch": 1.785450061652281 }, { "current_steps": 1448, "loss": 2.0271, "learning_rate": 6.599241466498104e-07, "epoch": 1.7866831072749692 }, { "current_steps": 1449, "loss": 1.9015, "learning_rate": 6.596713021491782e-07, "epoch": 1.7879161528976573 }, { "current_steps": 1450, "loss": 2.4606, "learning_rate": 6.59418457648546e-07, "epoch": 1.7891491985203452 }, { "current_steps": 1451, "loss": 2.068, "learning_rate": 6.591656131479139e-07, "epoch": 1.7903822441430333 }, { "current_steps": 1452, "loss": 2.0033, "learning_rate": 6.589127686472818e-07, "epoch": 1.7916152897657214 }, { "current_steps": 1453, "loss": 2.3039, "learning_rate": 6.586599241466497e-07, "epoch": 1.7928483353884093 }, { "current_steps": 1454, "loss": 2.3853, "learning_rate": 6.584070796460177e-07, "epoch": 1.7940813810110974 }, { "current_steps": 1455, "loss": 2.195, "learning_rate": 6.581542351453856e-07, "epoch": 1.7953144266337855 }, { "current_steps": 1456, "loss": 2.1822, "learning_rate": 6.579013906447535e-07, "epoch": 1.7965474722564734 }, { "current_steps": 1457, "loss": 2.3711, "learning_rate": 6.576485461441214e-07, "epoch": 1.7977805178791615 }, { "current_steps": 1458, "loss": 1.9698, "learning_rate": 6.573957016434893e-07, "epoch": 1.7990135635018496 }, { "current_steps": 1459, "loss": 2.4656, "learning_rate": 6.571428571428571e-07, "epoch": 1.8002466091245375 }, { "current_steps": 1460, "loss": 2.1716, "learning_rate": 6.56890012642225e-07, "epoch": 1.8014796547472256 }, { "current_steps": 1461, "loss": 1.6282, "learning_rate": 6.566371681415929e-07, "epoch": 1.8027127003699137 }, { "current_steps": 1462, "loss": 2.2023, "learning_rate": 6.563843236409607e-07, "epoch": 1.8039457459926016 }, { "current_steps": 1463, "loss": 2.3715, "learning_rate": 6.561314791403286e-07, "epoch": 1.8051787916152897 }, { "current_steps": 1464, "loss": 2.0779, "learning_rate": 6.558786346396965e-07, "epoch": 1.8064118372379778 }, { "current_steps": 1465, "loss": 1.8376, "learning_rate": 6.556257901390645e-07, "epoch": 1.8076448828606657 }, { "current_steps": 1466, "loss": 2.2253, "learning_rate": 6.553729456384324e-07, "epoch": 1.808877928483354 }, { "current_steps": 1467, "loss": 2.3438, "learning_rate": 6.551201011378002e-07, "epoch": 1.810110974106042 }, { "current_steps": 1468, "loss": 2.0974, "learning_rate": 6.548672566371681e-07, "epoch": 1.8113440197287298 }, { "current_steps": 1469, "loss": 1.6272, "learning_rate": 6.54614412136536e-07, "epoch": 1.8125770653514182 }, { "current_steps": 1470, "loss": 1.8159, "learning_rate": 6.543615676359039e-07, "epoch": 1.813810110974106 }, { "current_steps": 1471, "loss": 2.0442, "learning_rate": 6.541087231352718e-07, "epoch": 1.815043156596794 }, { "current_steps": 1472, "loss": 2.0841, "learning_rate": 6.538558786346397e-07, "epoch": 1.8162762022194823 }, { "current_steps": 1473, "loss": 1.808, "learning_rate": 6.536030341340076e-07, "epoch": 1.8175092478421702 }, { "current_steps": 1474, "loss": 2.2535, "learning_rate": 6.533501896333754e-07, "epoch": 1.818742293464858 }, { "current_steps": 1475, "loss": 1.9033, "learning_rate": 6.530973451327432e-07, "epoch": 1.8199753390875464 }, { "current_steps": 1476, "loss": 2.0661, "learning_rate": 6.528445006321112e-07, "epoch": 1.8212083847102343 }, { "current_steps": 1477, "loss": 2.0733, "learning_rate": 6.525916561314791e-07, "epoch": 1.8224414303329222 }, { "current_steps": 1478, "loss": 1.8143, "learning_rate": 6.52338811630847e-07, "epoch": 1.8236744759556105 }, { "current_steps": 1479, "loss": 1.5323, "learning_rate": 6.520859671302149e-07, "epoch": 1.8249075215782984 }, { "current_steps": 1480, "loss": 2.2398, "learning_rate": 6.518331226295828e-07, "epoch": 1.8261405672009863 }, { "current_steps": 1481, "loss": 2.2546, "learning_rate": 6.515802781289507e-07, "epoch": 1.8273736128236746 }, { "current_steps": 1482, "loss": 2.1421, "learning_rate": 6.513274336283186e-07, "epoch": 1.8286066584463625 }, { "current_steps": 1483, "loss": 2.3988, "learning_rate": 6.510745891276864e-07, "epoch": 1.8298397040690506 }, { "current_steps": 1484, "loss": 1.6965, "learning_rate": 6.508217446270543e-07, "epoch": 1.8310727496917387 }, { "current_steps": 1485, "loss": 2.4989, "learning_rate": 6.505689001264222e-07, "epoch": 1.8323057953144266 }, { "current_steps": 1486, "loss": 2.4313, "learning_rate": 6.503160556257901e-07, "epoch": 1.8335388409371147 }, { "current_steps": 1487, "loss": 2.0232, "learning_rate": 6.50063211125158e-07, "epoch": 1.8347718865598028 }, { "current_steps": 1488, "loss": 2.5576, "learning_rate": 6.498103666245259e-07, "epoch": 1.8360049321824907 }, { "current_steps": 1489, "loss": 2.0385, "learning_rate": 6.495575221238938e-07, "epoch": 1.8372379778051788 }, { "current_steps": 1490, "loss": 1.9234, "learning_rate": 6.493046776232617e-07, "epoch": 1.838471023427867 }, { "current_steps": 1491, "loss": 1.9061, "learning_rate": 6.490518331226295e-07, "epoch": 1.8397040690505548 }, { "current_steps": 1492, "loss": 2.3394, "learning_rate": 6.487989886219974e-07, "epoch": 1.840937114673243 }, { "current_steps": 1493, "loss": 1.9404, "learning_rate": 6.485461441213653e-07, "epoch": 1.842170160295931 }, { "current_steps": 1494, "loss": 2.1535, "learning_rate": 6.482932996207332e-07, "epoch": 1.843403205918619 }, { "current_steps": 1495, "loss": 2.1521, "learning_rate": 6.480404551201011e-07, "epoch": 1.844636251541307 }, { "current_steps": 1496, "loss": 2.0507, "learning_rate": 6.47787610619469e-07, "epoch": 1.8458692971639952 }, { "current_steps": 1497, "loss": 2.2416, "learning_rate": 6.475347661188369e-07, "epoch": 1.847102342786683 }, { "current_steps": 1498, "loss": 2.1244, "learning_rate": 6.472819216182049e-07, "epoch": 1.8483353884093712 }, { "current_steps": 1499, "loss": 2.2918, "learning_rate": 6.470290771175728e-07, "epoch": 1.8495684340320593 }, { "current_steps": 1500, "loss": 1.9759, "learning_rate": 6.467762326169405e-07, "epoch": 1.8508014796547472 }, { "current_steps": 1501, "loss": 2.2134, "learning_rate": 6.465233881163084e-07, "epoch": 1.8520345252774353 }, { "current_steps": 1502, "loss": 2.582, "learning_rate": 6.462705436156763e-07, "epoch": 1.8532675709001234 }, { "current_steps": 1503, "loss": 2.0217, "learning_rate": 6.460176991150442e-07, "epoch": 1.8545006165228113 }, { "current_steps": 1504, "loss": 2.0022, "learning_rate": 6.457648546144121e-07, "epoch": 1.8557336621454994 }, { "current_steps": 1505, "loss": 2.0176, "learning_rate": 6.4551201011378e-07, "epoch": 1.8569667077681875 }, { "current_steps": 1506, "loss": 2.0742, "learning_rate": 6.452591656131479e-07, "epoch": 1.8581997533908754 }, { "current_steps": 1507, "loss": 2.2694, "learning_rate": 6.450063211125158e-07, "epoch": 1.8594327990135635 }, { "current_steps": 1508, "loss": 2.1398, "learning_rate": 6.447534766118836e-07, "epoch": 1.8606658446362516 }, { "current_steps": 1509, "loss": 2.5432, "learning_rate": 6.445006321112515e-07, "epoch": 1.8618988902589395 }, { "current_steps": 1510, "loss": 2.1181, "learning_rate": 6.442477876106195e-07, "epoch": 1.8631319358816276 }, { "current_steps": 1511, "loss": 2.4965, "learning_rate": 6.439949431099874e-07, "epoch": 1.8643649815043157 }, { "current_steps": 1512, "loss": 2.1456, "learning_rate": 6.437420986093553e-07, "epoch": 1.8655980271270036 }, { "current_steps": 1513, "loss": 2.3779, "learning_rate": 6.434892541087231e-07, "epoch": 1.8668310727496917 }, { "current_steps": 1514, "loss": 1.8807, "learning_rate": 6.43236409608091e-07, "epoch": 1.8680641183723798 }, { "current_steps": 1515, "loss": 2.0694, "learning_rate": 6.429835651074589e-07, "epoch": 1.8692971639950677 }, { "current_steps": 1516, "loss": 2.2625, "learning_rate": 6.427307206068267e-07, "epoch": 1.8705302096177558 }, { "current_steps": 1517, "loss": 1.7827, "learning_rate": 6.424778761061946e-07, "epoch": 1.871763255240444 }, { "current_steps": 1518, "loss": 2.4359, "learning_rate": 6.422250316055625e-07, "epoch": 1.8729963008631318 }, { "current_steps": 1519, "loss": 2.1652, "learning_rate": 6.419721871049304e-07, "epoch": 1.87422934648582 }, { "current_steps": 1520, "loss": 2.2926, "learning_rate": 6.417193426042983e-07, "epoch": 1.875462392108508 }, { "current_steps": 1521, "loss": 1.5696, "learning_rate": 6.414664981036663e-07, "epoch": 1.876695437731196 }, { "current_steps": 1522, "loss": 2.3431, "learning_rate": 6.412136536030342e-07, "epoch": 1.877928483353884 }, { "current_steps": 1523, "loss": 2.2387, "learning_rate": 6.409608091024021e-07, "epoch": 1.8791615289765722 }, { "current_steps": 1524, "loss": 2.3041, "learning_rate": 6.4070796460177e-07, "epoch": 1.88039457459926 }, { "current_steps": 1525, "loss": 2.1034, "learning_rate": 6.404551201011377e-07, "epoch": 1.8816276202219482 }, { "current_steps": 1526, "loss": 1.8234, "learning_rate": 6.402022756005056e-07, "epoch": 1.8828606658446363 }, { "current_steps": 1527, "loss": 2.0179, "learning_rate": 6.399494310998735e-07, "epoch": 1.8840937114673242 }, { "current_steps": 1528, "loss": 1.6566, "learning_rate": 6.396965865992414e-07, "epoch": 1.8853267570900123 }, { "current_steps": 1529, "loss": 2.2221, "learning_rate": 6.394437420986093e-07, "epoch": 1.8865598027127004 }, { "current_steps": 1530, "loss": 2.5029, "learning_rate": 6.391908975979772e-07, "epoch": 1.8877928483353883 }, { "current_steps": 1531, "loss": 1.8757, "learning_rate": 6.389380530973451e-07, "epoch": 1.8890258939580764 }, { "current_steps": 1532, "loss": 1.8536, "learning_rate": 6.38685208596713e-07, "epoch": 1.8902589395807645 }, { "current_steps": 1533, "loss": 2.3046, "learning_rate": 6.384323640960809e-07, "epoch": 1.8914919852034524 }, { "current_steps": 1534, "loss": 1.8979, "learning_rate": 6.381795195954488e-07, "epoch": 1.8927250308261405 }, { "current_steps": 1535, "loss": 2.7792, "learning_rate": 6.379266750948167e-07, "epoch": 1.8939580764488286 }, { "current_steps": 1536, "loss": 1.6986, "learning_rate": 6.376738305941846e-07, "epoch": 1.8951911220715165 }, { "current_steps": 1537, "loss": 1.9078, "learning_rate": 6.374209860935525e-07, "epoch": 1.8964241676942046 }, { "current_steps": 1538, "loss": 2.2003, "learning_rate": 6.371681415929203e-07, "epoch": 1.8976572133168927 }, { "current_steps": 1539, "loss": 1.7025, "learning_rate": 6.369152970922882e-07, "epoch": 1.8988902589395806 }, { "current_steps": 1540, "loss": 1.9687, "learning_rate": 6.36662452591656e-07, "epoch": 1.900123304562269 }, { "current_steps": 1541, "loss": 2.2578, "learning_rate": 6.364096080910239e-07, "epoch": 1.9013563501849569 }, { "current_steps": 1542, "loss": 2.428, "learning_rate": 6.361567635903918e-07, "epoch": 1.9025893958076447 }, { "current_steps": 1543, "loss": 1.9973, "learning_rate": 6.359039190897598e-07, "epoch": 1.903822441430333 }, { "current_steps": 1544, "loss": 2.6179, "learning_rate": 6.356510745891277e-07, "epoch": 1.905055487053021 }, { "current_steps": 1545, "loss": 2.1068, "learning_rate": 6.353982300884956e-07, "epoch": 1.9062885326757089 }, { "current_steps": 1546, "loss": 1.9569, "learning_rate": 6.351453855878635e-07, "epoch": 1.9075215782983972 }, { "current_steps": 1547, "loss": 2.3457, "learning_rate": 6.348925410872314e-07, "epoch": 1.908754623921085 }, { "current_steps": 1548, "loss": 2.133, "learning_rate": 6.346396965865993e-07, "epoch": 1.909987669543773 }, { "current_steps": 1549, "loss": 1.6077, "learning_rate": 6.343868520859671e-07, "epoch": 1.9112207151664613 }, { "current_steps": 1550, "loss": 1.6328, "learning_rate": 6.341340075853349e-07, "epoch": 1.9124537607891492 }, { "current_steps": 1551, "loss": 2.882, "learning_rate": 6.338811630847028e-07, "epoch": 1.913686806411837 }, { "current_steps": 1552, "loss": 2.3654, "learning_rate": 6.336283185840707e-07, "epoch": 1.9149198520345254 }, { "current_steps": 1553, "loss": 1.7348, "learning_rate": 6.333754740834386e-07, "epoch": 1.9161528976572133 }, { "current_steps": 1554, "loss": 1.5942, "learning_rate": 6.331226295828066e-07, "epoch": 1.9173859432799012 }, { "current_steps": 1555, "loss": 1.5437, "learning_rate": 6.328697850821745e-07, "epoch": 1.9186189889025895 }, { "current_steps": 1556, "loss": 2.2133, "learning_rate": 6.326169405815424e-07, "epoch": 1.9198520345252774 }, { "current_steps": 1557, "loss": 2.3184, "learning_rate": 6.323640960809102e-07, "epoch": 1.9210850801479655 }, { "current_steps": 1558, "loss": 2.7126, "learning_rate": 6.321112515802781e-07, "epoch": 1.9223181257706536 }, { "current_steps": 1559, "loss": 2.2606, "learning_rate": 6.31858407079646e-07, "epoch": 1.9235511713933415 }, { "current_steps": 1560, "loss": 2.0281, "learning_rate": 6.316055625790139e-07, "epoch": 1.9247842170160296 }, { "current_steps": 1561, "loss": 2.2306, "learning_rate": 6.313527180783818e-07, "epoch": 1.9260172626387178 }, { "current_steps": 1562, "loss": 1.8654, "learning_rate": 6.310998735777497e-07, "epoch": 1.9272503082614056 }, { "current_steps": 1563, "loss": 2.2147, "learning_rate": 6.308470290771175e-07, "epoch": 1.9284833538840938 }, { "current_steps": 1564, "loss": 2.0369, "learning_rate": 6.305941845764854e-07, "epoch": 1.9297163995067819 }, { "current_steps": 1565, "loss": 2.3783, "learning_rate": 6.303413400758532e-07, "epoch": 1.9309494451294698 }, { "current_steps": 1566, "loss": 1.5646, "learning_rate": 6.300884955752212e-07, "epoch": 1.9321824907521579 }, { "current_steps": 1567, "loss": 1.9196, "learning_rate": 6.298356510745891e-07, "epoch": 1.933415536374846 }, { "current_steps": 1568, "loss": 2.381, "learning_rate": 6.29582806573957e-07, "epoch": 1.9346485819975339 }, { "current_steps": 1569, "loss": 2.1859, "learning_rate": 6.293299620733249e-07, "epoch": 1.935881627620222 }, { "current_steps": 1570, "loss": 2.4835, "learning_rate": 6.290771175726928e-07, "epoch": 1.93711467324291 }, { "current_steps": 1571, "loss": 2.2811, "learning_rate": 6.288242730720607e-07, "epoch": 1.938347718865598 }, { "current_steps": 1572, "loss": 2.2898, "learning_rate": 6.285714285714286e-07, "epoch": 1.939580764488286 }, { "current_steps": 1573, "loss": 2.2081, "learning_rate": 6.283185840707964e-07, "epoch": 1.9408138101109742 }, { "current_steps": 1574, "loss": 2.3223, "learning_rate": 6.280657395701643e-07, "epoch": 1.942046855733662 }, { "current_steps": 1575, "loss": 2.0349, "learning_rate": 6.278128950695321e-07, "epoch": 1.9432799013563502 }, { "current_steps": 1576, "loss": 2.5177, "learning_rate": 6.275600505689e-07, "epoch": 1.9445129469790383 }, { "current_steps": 1577, "loss": 2.1973, "learning_rate": 6.27307206068268e-07, "epoch": 1.9457459926017262 }, { "current_steps": 1578, "loss": 2.162, "learning_rate": 6.270543615676359e-07, "epoch": 1.9469790382244143 }, { "current_steps": 1579, "loss": 1.6221, "learning_rate": 6.268015170670038e-07, "epoch": 1.9482120838471024 }, { "current_steps": 1580, "loss": 1.9978, "learning_rate": 6.265486725663717e-07, "epoch": 1.9494451294697903 }, { "current_steps": 1581, "loss": 2.2909, "learning_rate": 6.262958280657395e-07, "epoch": 1.9506781750924784 }, { "current_steps": 1582, "loss": 2.4064, "learning_rate": 6.260429835651074e-07, "epoch": 1.9519112207151665 }, { "current_steps": 1583, "loss": 2.3266, "learning_rate": 6.257901390644753e-07, "epoch": 1.9531442663378544 }, { "current_steps": 1584, "loss": 2.1972, "learning_rate": 6.255372945638432e-07, "epoch": 1.9543773119605425 }, { "current_steps": 1585, "loss": 1.8524, "learning_rate": 6.252844500632111e-07, "epoch": 1.9556103575832307 }, { "current_steps": 1586, "loss": 1.9775, "learning_rate": 6.25031605562579e-07, "epoch": 1.9568434032059185 }, { "current_steps": 1587, "loss": 2.3677, "learning_rate": 6.247787610619469e-07, "epoch": 1.9580764488286067 }, { "current_steps": 1588, "loss": 1.6178, "learning_rate": 6.245259165613148e-07, "epoch": 1.9593094944512948 }, { "current_steps": 1589, "loss": 1.7002, "learning_rate": 6.242730720606826e-07, "epoch": 1.9605425400739827 }, { "current_steps": 1590, "loss": 2.0748, "learning_rate": 6.240202275600505e-07, "epoch": 1.9617755856966708 }, { "current_steps": 1591, "loss": 1.7085, "learning_rate": 6.237673830594184e-07, "epoch": 1.9630086313193589 }, { "current_steps": 1592, "loss": 1.482, "learning_rate": 6.235145385587863e-07, "epoch": 1.9642416769420468 }, { "current_steps": 1593, "loss": 1.7039, "learning_rate": 6.232616940581542e-07, "epoch": 1.9654747225647349 }, { "current_steps": 1594, "loss": 1.9193, "learning_rate": 6.230088495575221e-07, "epoch": 1.966707768187423 }, { "current_steps": 1595, "loss": 1.7196, "learning_rate": 6.2275600505689e-07, "epoch": 1.9679408138101109 }, { "current_steps": 1596, "loss": 2.4012, "learning_rate": 6.225031605562579e-07, "epoch": 1.969173859432799 }, { "current_steps": 1597, "loss": 1.9035, "learning_rate": 6.222503160556258e-07, "epoch": 1.970406905055487 }, { "current_steps": 1598, "loss": 2.1209, "learning_rate": 6.219974715549936e-07, "epoch": 1.971639950678175 }, { "current_steps": 1599, "loss": 1.8616, "learning_rate": 6.217446270543616e-07, "epoch": 1.972872996300863 }, { "current_steps": 1600, "loss": 1.9463, "learning_rate": 6.214917825537295e-07, "epoch": 1.9741060419235512 }, { "current_steps": 1601, "loss": 2.3527, "learning_rate": 6.212389380530973e-07, "epoch": 1.975339087546239 }, { "current_steps": 1602, "loss": 2.026, "learning_rate": 6.209860935524652e-07, "epoch": 1.9765721331689272 }, { "current_steps": 1603, "loss": 1.5881, "learning_rate": 6.207332490518331e-07, "epoch": 1.9778051787916153 }, { "current_steps": 1604, "loss": 2.1983, "learning_rate": 6.20480404551201e-07, "epoch": 1.9790382244143032 }, { "current_steps": 1605, "loss": 2.344, "learning_rate": 6.202275600505689e-07, "epoch": 1.9802712700369913 }, { "current_steps": 1606, "loss": 1.9296, "learning_rate": 6.199747155499367e-07, "epoch": 1.9815043156596794 }, { "current_steps": 1607, "loss": 2.1191, "learning_rate": 6.197218710493046e-07, "epoch": 1.9827373612823673 }, { "current_steps": 1608, "loss": 2.1913, "learning_rate": 6.194690265486725e-07, "epoch": 1.9839704069050554 }, { "current_steps": 1609, "loss": 2.1518, "learning_rate": 6.192161820480404e-07, "epoch": 1.9852034525277436 }, { "current_steps": 1610, "loss": 2.3907, "learning_rate": 6.189633375474084e-07, "epoch": 1.9864364981504314 }, { "current_steps": 1611, "loss": 2.4299, "learning_rate": 6.187104930467763e-07, "epoch": 1.9876695437731196 }, { "current_steps": 1612, "loss": 1.8654, "learning_rate": 6.184576485461442e-07, "epoch": 1.9889025893958077 }, { "current_steps": 1613, "loss": 2.0141, "learning_rate": 6.182048040455121e-07, "epoch": 1.9901356350184956 }, { "current_steps": 1614, "loss": 1.9654, "learning_rate": 6.179519595448798e-07, "epoch": 1.9913686806411839 }, { "current_steps": 1615, "loss": 2.2752, "learning_rate": 6.176991150442477e-07, "epoch": 1.9926017262638718 }, { "current_steps": 1616, "loss": 2.5543, "learning_rate": 6.174462705436156e-07, "epoch": 1.9938347718865597 }, { "current_steps": 1617, "loss": 1.9885, "learning_rate": 6.171934260429835e-07, "epoch": 1.995067817509248 }, { "current_steps": 1618, "loss": 1.961, "learning_rate": 6.169405815423514e-07, "epoch": 1.9963008631319359 }, { "current_steps": 1619, "loss": 2.1751, "learning_rate": 6.166877370417193e-07, "epoch": 1.9975339087546238 }, { "current_steps": 1620, "loss": 2.1305, "learning_rate": 6.164348925410872e-07, "epoch": 1.998766954377312 }, { "current_steps": 1621, "loss": 1.7882, "learning_rate": 6.161820480404551e-07, "epoch": 2.0 }, { "current_steps": 1622, "loss": 2.5962, "learning_rate": 6.15929203539823e-07, "epoch": 2.001233045622688 }, { "current_steps": 1623, "loss": 1.9411, "learning_rate": 6.156763590391909e-07, "epoch": 2.0024660912453762 }, { "current_steps": 1624, "loss": 1.7947, "learning_rate": 6.154235145385588e-07, "epoch": 2.003699136868064 }, { "current_steps": 1625, "loss": 2.2096, "learning_rate": 6.151706700379267e-07, "epoch": 2.004932182490752 }, { "current_steps": 1626, "loss": 2.2047, "learning_rate": 6.149178255372945e-07, "epoch": 2.0061652281134403 }, { "current_steps": 1627, "loss": 2.4501, "learning_rate": 6.146649810366624e-07, "epoch": 2.0073982737361282 }, { "current_steps": 1628, "loss": 1.9055, "learning_rate": 6.144121365360303e-07, "epoch": 2.008631319358816 }, { "current_steps": 1629, "loss": 2.5551, "learning_rate": 6.141592920353982e-07, "epoch": 2.0098643649815044 }, { "current_steps": 1630, "loss": 1.9802, "learning_rate": 6.13906447534766e-07, "epoch": 2.0110974106041923 }, { "current_steps": 1631, "loss": 1.7837, "learning_rate": 6.136536030341339e-07, "epoch": 2.0123304562268802 }, { "current_steps": 1632, "loss": 2.1175, "learning_rate": 6.134007585335018e-07, "epoch": 2.0135635018495686 }, { "current_steps": 1633, "loss": 2.1366, "learning_rate": 6.131479140328698e-07, "epoch": 2.0147965474722564 }, { "current_steps": 1634, "loss": 2.3745, "learning_rate": 6.128950695322377e-07, "epoch": 2.0160295930949443 }, { "current_steps": 1635, "loss": 1.7385, "learning_rate": 6.126422250316056e-07, "epoch": 2.0172626387176327 }, { "current_steps": 1636, "loss": 1.1622, "learning_rate": 6.123893805309735e-07, "epoch": 2.0184956843403206 }, { "current_steps": 1637, "loss": 2.3392, "learning_rate": 6.121365360303414e-07, "epoch": 2.0197287299630085 }, { "current_steps": 1638, "loss": 2.2071, "learning_rate": 6.118836915297093e-07, "epoch": 2.020961775585697 }, { "current_steps": 1639, "loss": 2.0372, "learning_rate": 6.11630847029077e-07, "epoch": 2.0221948212083847 }, { "current_steps": 1640, "loss": 2.2466, "learning_rate": 6.113780025284449e-07, "epoch": 2.0234278668310726 }, { "current_steps": 1641, "loss": 2.3818, "learning_rate": 6.111251580278128e-07, "epoch": 2.024660912453761 }, { "current_steps": 1642, "loss": 1.9808, "learning_rate": 6.108723135271807e-07, "epoch": 2.025893958076449 }, { "current_steps": 1643, "loss": 1.9486, "learning_rate": 6.106194690265486e-07, "epoch": 2.0271270036991367 }, { "current_steps": 1644, "loss": 2.1815, "learning_rate": 6.103666245259166e-07, "epoch": 2.028360049321825 }, { "current_steps": 1645, "loss": 2.3438, "learning_rate": 6.101137800252845e-07, "epoch": 2.029593094944513 }, { "current_steps": 1646, "loss": 1.873, "learning_rate": 6.098609355246524e-07, "epoch": 2.030826140567201 }, { "current_steps": 1647, "loss": 2.1266, "learning_rate": 6.096080910240202e-07, "epoch": 2.032059186189889 }, { "current_steps": 1648, "loss": 2.0303, "learning_rate": 6.093552465233881e-07, "epoch": 2.033292231812577 }, { "current_steps": 1649, "loss": 1.7774, "learning_rate": 6.09102402022756e-07, "epoch": 2.034525277435265 }, { "current_steps": 1650, "loss": 1.7078, "learning_rate": 6.088495575221239e-07, "epoch": 2.0357583230579532 }, { "current_steps": 1651, "loss": 2.1692, "learning_rate": 6.085967130214917e-07, "epoch": 2.036991368680641 }, { "current_steps": 1652, "loss": 2.3063, "learning_rate": 6.083438685208596e-07, "epoch": 2.038224414303329 }, { "current_steps": 1653, "loss": 1.7903, "learning_rate": 6.080910240202275e-07, "epoch": 2.0394574599260173 }, { "current_steps": 1654, "loss": 2.3063, "learning_rate": 6.078381795195953e-07, "epoch": 2.0406905055487052 }, { "current_steps": 1655, "loss": 2.1247, "learning_rate": 6.075853350189633e-07, "epoch": 2.041923551171393 }, { "current_steps": 1656, "loss": 2.1108, "learning_rate": 6.073324905183312e-07, "epoch": 2.0431565967940815 }, { "current_steps": 1657, "loss": 1.9585, "learning_rate": 6.070796460176991e-07, "epoch": 2.0443896424167693 }, { "current_steps": 1658, "loss": 2.0225, "learning_rate": 6.06826801517067e-07, "epoch": 2.0456226880394572 }, { "current_steps": 1659, "loss": 1.8648, "learning_rate": 6.065739570164349e-07, "epoch": 2.0468557336621456 }, { "current_steps": 1660, "loss": 2.0584, "learning_rate": 6.063211125158028e-07, "epoch": 2.0480887792848335 }, { "current_steps": 1661, "loss": 2.2502, "learning_rate": 6.060682680151707e-07, "epoch": 2.049321824907522 }, { "current_steps": 1662, "loss": 2.0851, "learning_rate": 6.058154235145386e-07, "epoch": 2.0505548705302097 }, { "current_steps": 1663, "loss": 2.1534, "learning_rate": 6.055625790139064e-07, "epoch": 2.0517879161528976 }, { "current_steps": 1664, "loss": 1.8156, "learning_rate": 6.053097345132742e-07, "epoch": 2.053020961775586 }, { "current_steps": 1665, "loss": 2.1658, "learning_rate": 6.050568900126421e-07, "epoch": 2.054254007398274 }, { "current_steps": 1666, "loss": 2.3917, "learning_rate": 6.048040455120101e-07, "epoch": 2.0554870530209617 }, { "current_steps": 1667, "loss": 2.5773, "learning_rate": 6.04551201011378e-07, "epoch": 2.05672009864365 }, { "current_steps": 1668, "loss": 2.6493, "learning_rate": 6.042983565107459e-07, "epoch": 2.057953144266338 }, { "current_steps": 1669, "loss": 2.1378, "learning_rate": 6.040455120101138e-07, "epoch": 2.059186189889026 }, { "current_steps": 1670, "loss": 1.5719, "learning_rate": 6.037926675094817e-07, "epoch": 2.060419235511714 }, { "current_steps": 1671, "loss": 2.2564, "learning_rate": 6.035398230088495e-07, "epoch": 2.061652281134402 }, { "current_steps": 1672, "loss": 1.8674, "learning_rate": 6.032869785082174e-07, "epoch": 2.06288532675709 }, { "current_steps": 1673, "loss": 1.9759, "learning_rate": 6.030341340075853e-07, "epoch": 2.0641183723797782 }, { "current_steps": 1674, "loss": 2.6081, "learning_rate": 6.027812895069532e-07, "epoch": 2.065351418002466 }, { "current_steps": 1675, "loss": 1.6438, "learning_rate": 6.025284450063211e-07, "epoch": 2.066584463625154 }, { "current_steps": 1676, "loss": 1.8943, "learning_rate": 6.022756005056889e-07, "epoch": 2.0678175092478424 }, { "current_steps": 1677, "loss": 2.1449, "learning_rate": 6.020227560050568e-07, "epoch": 2.0690505548705302 }, { "current_steps": 1678, "loss": 1.7557, "learning_rate": 6.017699115044248e-07, "epoch": 2.070283600493218 }, { "current_steps": 1679, "loss": 2.0144, "learning_rate": 6.015170670037926e-07, "epoch": 2.0715166461159065 }, { "current_steps": 1680, "loss": 2.3653, "learning_rate": 6.012642225031605e-07, "epoch": 2.0727496917385944 }, { "current_steps": 1681, "loss": 1.8393, "learning_rate": 6.010113780025284e-07, "epoch": 2.0739827373612822 }, { "current_steps": 1682, "loss": 1.6968, "learning_rate": 6.007585335018963e-07, "epoch": 2.0752157829839706 }, { "current_steps": 1683, "loss": 2.0262, "learning_rate": 6.005056890012642e-07, "epoch": 2.0764488286066585 }, { "current_steps": 1684, "loss": 2.5561, "learning_rate": 6.002528445006321e-07, "epoch": 2.0776818742293464 }, { "current_steps": 1685, "loss": 1.7388, "learning_rate": 6e-07, "epoch": 2.0789149198520347 }, { "current_steps": 1686, "loss": 2.5387, "learning_rate": 5.997471554993679e-07, "epoch": 2.0801479654747226 }, { "current_steps": 1687, "loss": 1.7434, "learning_rate": 5.994943109987357e-07, "epoch": 2.0813810110974105 }, { "current_steps": 1688, "loss": 2.3208, "learning_rate": 5.992414664981036e-07, "epoch": 2.082614056720099 }, { "current_steps": 1689, "loss": 2.0421, "learning_rate": 5.989886219974716e-07, "epoch": 2.0838471023427867 }, { "current_steps": 1690, "loss": 2.0771, "learning_rate": 5.987357774968394e-07, "epoch": 2.0850801479654746 }, { "current_steps": 1691, "loss": 1.9324, "learning_rate": 5.984829329962073e-07, "epoch": 2.086313193588163 }, { "current_steps": 1692, "loss": 2.0185, "learning_rate": 5.982300884955752e-07, "epoch": 2.087546239210851 }, { "current_steps": 1693, "loss": 2.4963, "learning_rate": 5.979772439949431e-07, "epoch": 2.0887792848335387 }, { "current_steps": 1694, "loss": 2.4522, "learning_rate": 5.97724399494311e-07, "epoch": 2.090012330456227 }, { "current_steps": 1695, "loss": 2.7067, "learning_rate": 5.974715549936788e-07, "epoch": 2.091245376078915 }, { "current_steps": 1696, "loss": 2.3525, "learning_rate": 5.972187104930467e-07, "epoch": 2.092478421701603 }, { "current_steps": 1697, "loss": 2.24, "learning_rate": 5.969658659924146e-07, "epoch": 2.093711467324291 }, { "current_steps": 1698, "loss": 1.9437, "learning_rate": 5.967130214917825e-07, "epoch": 2.094944512946979 }, { "current_steps": 1699, "loss": 1.9598, "learning_rate": 5.964601769911504e-07, "epoch": 2.096177558569667 }, { "current_steps": 1700, "loss": 2.122, "learning_rate": 5.962073324905184e-07, "epoch": 2.0974106041923553 }, { "current_steps": 1701, "loss": 2.1337, "learning_rate": 5.959544879898863e-07, "epoch": 2.098643649815043 }, { "current_steps": 1702, "loss": 2.555, "learning_rate": 5.957016434892541e-07, "epoch": 2.099876695437731 }, { "current_steps": 1703, "loss": 2.2288, "learning_rate": 5.95448798988622e-07, "epoch": 2.1011097410604194 }, { "current_steps": 1704, "loss": 2.0033, "learning_rate": 5.951959544879898e-07, "epoch": 2.1023427866831073 }, { "current_steps": 1705, "loss": 1.6264, "learning_rate": 5.949431099873577e-07, "epoch": 2.103575832305795 }, { "current_steps": 1706, "loss": 2.3905, "learning_rate": 5.946902654867256e-07, "epoch": 2.1048088779284835 }, { "current_steps": 1707, "loss": 2.241, "learning_rate": 5.944374209860935e-07, "epoch": 2.1060419235511714 }, { "current_steps": 1708, "loss": 1.8467, "learning_rate": 5.941845764854614e-07, "epoch": 2.1072749691738593 }, { "current_steps": 1709, "loss": 2.6743, "learning_rate": 5.939317319848293e-07, "epoch": 2.1085080147965476 }, { "current_steps": 1710, "loss": 2.2291, "learning_rate": 5.936788874841972e-07, "epoch": 2.1097410604192355 }, { "current_steps": 1711, "loss": 2.5819, "learning_rate": 5.934260429835652e-07, "epoch": 2.1109741060419234 }, { "current_steps": 1712, "loss": 2.0416, "learning_rate": 5.93173198482933e-07, "epoch": 2.1122071516646117 }, { "current_steps": 1713, "loss": 2.2805, "learning_rate": 5.929203539823009e-07, "epoch": 2.1134401972872996 }, { "current_steps": 1714, "loss": 2.4214, "learning_rate": 5.926675094816688e-07, "epoch": 2.1146732429099875 }, { "current_steps": 1715, "loss": 2.2559, "learning_rate": 5.924146649810366e-07, "epoch": 2.115906288532676 }, { "current_steps": 1716, "loss": 1.7746, "learning_rate": 5.921618204804045e-07, "epoch": 2.1171393341553637 }, { "current_steps": 1717, "loss": 2.3074, "learning_rate": 5.919089759797724e-07, "epoch": 2.1183723797780516 }, { "current_steps": 1718, "loss": 1.703, "learning_rate": 5.916561314791403e-07, "epoch": 2.11960542540074 }, { "current_steps": 1719, "loss": 2.3266, "learning_rate": 5.914032869785082e-07, "epoch": 2.120838471023428 }, { "current_steps": 1720, "loss": 1.9368, "learning_rate": 5.91150442477876e-07, "epoch": 2.1220715166461157 }, { "current_steps": 1721, "loss": 2.0343, "learning_rate": 5.908975979772439e-07, "epoch": 2.123304562268804 }, { "current_steps": 1722, "loss": 2.1774, "learning_rate": 5.906447534766119e-07, "epoch": 2.124537607891492 }, { "current_steps": 1723, "loss": 2.007, "learning_rate": 5.903919089759798e-07, "epoch": 2.12577065351418 }, { "current_steps": 1724, "loss": 1.4654, "learning_rate": 5.901390644753477e-07, "epoch": 2.127003699136868 }, { "current_steps": 1725, "loss": 2.0761, "learning_rate": 5.898862199747156e-07, "epoch": 2.128236744759556 }, { "current_steps": 1726, "loss": 2.2634, "learning_rate": 5.896333754740835e-07, "epoch": 2.1294697903822444 }, { "current_steps": 1727, "loss": 2.1695, "learning_rate": 5.893805309734513e-07, "epoch": 2.1307028360049323 }, { "current_steps": 1728, "loss": 2.1014, "learning_rate": 5.891276864728191e-07, "epoch": 2.13193588162762 }, { "current_steps": 1729, "loss": 2.1071, "learning_rate": 5.88874841972187e-07, "epoch": 2.133168927250308 }, { "current_steps": 1730, "loss": 1.8328, "learning_rate": 5.886219974715549e-07, "epoch": 2.1344019728729964 }, { "current_steps": 1731, "loss": 2.3443, "learning_rate": 5.883691529709228e-07, "epoch": 2.1356350184956843 }, { "current_steps": 1732, "loss": 1.9774, "learning_rate": 5.881163084702907e-07, "epoch": 2.1368680641183726 }, { "current_steps": 1733, "loss": 2.3512, "learning_rate": 5.878634639696586e-07, "epoch": 2.1381011097410605 }, { "current_steps": 1734, "loss": 1.8761, "learning_rate": 5.876106194690266e-07, "epoch": 2.1393341553637484 }, { "current_steps": 1735, "loss": 2.273, "learning_rate": 5.873577749683945e-07, "epoch": 2.1405672009864363 }, { "current_steps": 1736, "loss": 2.2784, "learning_rate": 5.871049304677624e-07, "epoch": 2.1418002466091246 }, { "current_steps": 1737, "loss": 1.4417, "learning_rate": 5.868520859671302e-07, "epoch": 2.1430332922318125 }, { "current_steps": 1738, "loss": 1.9027, "learning_rate": 5.865992414664981e-07, "epoch": 2.144266337854501 }, { "current_steps": 1739, "loss": 2.3631, "learning_rate": 5.86346396965866e-07, "epoch": 2.1454993834771887 }, { "current_steps": 1740, "loss": 2.2996, "learning_rate": 5.860935524652338e-07, "epoch": 2.1467324290998766 }, { "current_steps": 1741, "loss": 2.2973, "learning_rate": 5.858407079646017e-07, "epoch": 2.147965474722565 }, { "current_steps": 1742, "loss": 1.4981, "learning_rate": 5.855878634639696e-07, "epoch": 2.149198520345253 }, { "current_steps": 1743, "loss": 2.2354, "learning_rate": 5.853350189633375e-07, "epoch": 2.1504315659679407 }, { "current_steps": 1744, "loss": 1.9112, "learning_rate": 5.850821744627053e-07, "epoch": 2.151664611590629 }, { "current_steps": 1745, "loss": 2.2017, "learning_rate": 5.848293299620733e-07, "epoch": 2.152897657213317 }, { "current_steps": 1746, "loss": 1.2737, "learning_rate": 5.845764854614412e-07, "epoch": 2.154130702836005 }, { "current_steps": 1747, "loss": 2.3572, "learning_rate": 5.843236409608091e-07, "epoch": 2.155363748458693 }, { "current_steps": 1748, "loss": 2.3017, "learning_rate": 5.84070796460177e-07, "epoch": 2.156596794081381 }, { "current_steps": 1749, "loss": 1.6382, "learning_rate": 5.838179519595449e-07, "epoch": 2.157829839704069 }, { "current_steps": 1750, "loss": 2.2708, "learning_rate": 5.835651074589128e-07, "epoch": 2.1590628853267573 }, { "current_steps": 1751, "loss": 1.5938, "learning_rate": 5.833122629582807e-07, "epoch": 2.160295930949445 }, { "current_steps": 1752, "loss": 1.9964, "learning_rate": 5.830594184576484e-07, "epoch": 2.161528976572133 }, { "current_steps": 1753, "loss": 1.82, "learning_rate": 5.828065739570163e-07, "epoch": 2.1627620221948214 }, { "current_steps": 1754, "loss": 2.5812, "learning_rate": 5.825537294563842e-07, "epoch": 2.1639950678175093 }, { "current_steps": 1755, "loss": 2.3607, "learning_rate": 5.823008849557521e-07, "epoch": 2.165228113440197 }, { "current_steps": 1756, "loss": 2.2783, "learning_rate": 5.820480404551201e-07, "epoch": 2.1664611590628855 }, { "current_steps": 1757, "loss": 1.8535, "learning_rate": 5.81795195954488e-07, "epoch": 2.1676942046855734 }, { "current_steps": 1758, "loss": 2.1177, "learning_rate": 5.815423514538559e-07, "epoch": 2.1689272503082613 }, { "current_steps": 1759, "loss": 2.2665, "learning_rate": 5.812895069532238e-07, "epoch": 2.1701602959309496 }, { "current_steps": 1760, "loss": 1.9395, "learning_rate": 5.810366624525917e-07, "epoch": 2.1713933415536375 }, { "current_steps": 1761, "loss": 2.3546, "learning_rate": 5.807838179519595e-07, "epoch": 2.1726263871763254 }, { "current_steps": 1762, "loss": 2.2308, "learning_rate": 5.805309734513274e-07, "epoch": 2.1738594327990137 }, { "current_steps": 1763, "loss": 2.2687, "learning_rate": 5.802781289506953e-07, "epoch": 2.1750924784217016 }, { "current_steps": 1764, "loss": 2.2091, "learning_rate": 5.800252844500632e-07, "epoch": 2.1763255240443895 }, { "current_steps": 1765, "loss": 1.7675, "learning_rate": 5.79772439949431e-07, "epoch": 2.177558569667078 }, { "current_steps": 1766, "loss": 2.0935, "learning_rate": 5.795195954487989e-07, "epoch": 2.1787916152897657 }, { "current_steps": 1767, "loss": 2.1606, "learning_rate": 5.792667509481669e-07, "epoch": 2.1800246609124536 }, { "current_steps": 1768, "loss": 1.754, "learning_rate": 5.790139064475348e-07, "epoch": 2.181257706535142 }, { "current_steps": 1769, "loss": 1.5752, "learning_rate": 5.787610619469026e-07, "epoch": 2.18249075215783 }, { "current_steps": 1770, "loss": 2.1546, "learning_rate": 5.785082174462705e-07, "epoch": 2.1837237977805177 }, { "current_steps": 1771, "loss": 1.8428, "learning_rate": 5.782553729456384e-07, "epoch": 2.184956843403206 }, { "current_steps": 1772, "loss": 1.9607, "learning_rate": 5.780025284450063e-07, "epoch": 2.186189889025894 }, { "current_steps": 1773, "loss": 2.2507, "learning_rate": 5.777496839443742e-07, "epoch": 2.187422934648582 }, { "current_steps": 1774, "loss": 2.1011, "learning_rate": 5.774968394437421e-07, "epoch": 2.18865598027127 }, { "current_steps": 1775, "loss": 1.9598, "learning_rate": 5.7724399494311e-07, "epoch": 2.189889025893958 }, { "current_steps": 1776, "loss": 1.9073, "learning_rate": 5.769911504424779e-07, "epoch": 2.191122071516646 }, { "current_steps": 1777, "loss": 2.25, "learning_rate": 5.767383059418456e-07, "epoch": 2.1923551171393343 }, { "current_steps": 1778, "loss": 2.1309, "learning_rate": 5.764854614412136e-07, "epoch": 2.193588162762022 }, { "current_steps": 1779, "loss": 2.2818, "learning_rate": 5.762326169405815e-07, "epoch": 2.19482120838471 }, { "current_steps": 1780, "loss": 2.1635, "learning_rate": 5.759797724399494e-07, "epoch": 2.1960542540073984 }, { "current_steps": 1781, "loss": 2.0539, "learning_rate": 5.757269279393173e-07, "epoch": 2.1972872996300863 }, { "current_steps": 1782, "loss": 2.2571, "learning_rate": 5.754740834386852e-07, "epoch": 2.198520345252774 }, { "current_steps": 1783, "loss": 1.9067, "learning_rate": 5.752212389380531e-07, "epoch": 2.1997533908754625 }, { "current_steps": 1784, "loss": 2.3388, "learning_rate": 5.74968394437421e-07, "epoch": 2.2009864364981504 }, { "current_steps": 1785, "loss": 2.2956, "learning_rate": 5.747155499367888e-07, "epoch": 2.2022194821208383 }, { "current_steps": 1786, "loss": 2.0376, "learning_rate": 5.744627054361567e-07, "epoch": 2.2034525277435266 }, { "current_steps": 1787, "loss": 2.0372, "learning_rate": 5.742098609355246e-07, "epoch": 2.2046855733662145 }, { "current_steps": 1788, "loss": 2.4555, "learning_rate": 5.739570164348925e-07, "epoch": 2.2059186189889024 }, { "current_steps": 1789, "loss": 2.0261, "learning_rate": 5.737041719342604e-07, "epoch": 2.2071516646115907 }, { "current_steps": 1790, "loss": 2.0806, "learning_rate": 5.734513274336284e-07, "epoch": 2.2083847102342786 }, { "current_steps": 1791, "loss": 2.5194, "learning_rate": 5.731984829329962e-07, "epoch": 2.2096177558569665 }, { "current_steps": 1792, "loss": 1.6799, "learning_rate": 5.729456384323641e-07, "epoch": 2.210850801479655 }, { "current_steps": 1793, "loss": 1.496, "learning_rate": 5.72692793931732e-07, "epoch": 2.2120838471023427 }, { "current_steps": 1794, "loss": 2.4192, "learning_rate": 5.724399494310998e-07, "epoch": 2.2133168927250306 }, { "current_steps": 1795, "loss": 2.1032, "learning_rate": 5.721871049304677e-07, "epoch": 2.214549938347719 }, { "current_steps": 1796, "loss": 2.3853, "learning_rate": 5.719342604298356e-07, "epoch": 2.215782983970407 }, { "current_steps": 1797, "loss": 2.0475, "learning_rate": 5.716814159292035e-07, "epoch": 2.2170160295930947 }, { "current_steps": 1798, "loss": 2.0995, "learning_rate": 5.714285714285714e-07, "epoch": 2.218249075215783 }, { "current_steps": 1799, "loss": 1.8376, "learning_rate": 5.711757269279393e-07, "epoch": 2.219482120838471 }, { "current_steps": 1800, "loss": 1.9991, "learning_rate": 5.709228824273072e-07, "epoch": 2.220715166461159 }, { "current_steps": 1801, "loss": 2.1687, "learning_rate": 5.706700379266752e-07, "epoch": 2.221948212083847 }, { "current_steps": 1802, "loss": 1.8675, "learning_rate": 5.70417193426043e-07, "epoch": 2.223181257706535 }, { "current_steps": 1803, "loss": 2.4008, "learning_rate": 5.701643489254108e-07, "epoch": 2.2244143033292234 }, { "current_steps": 1804, "loss": 2.1782, "learning_rate": 5.699115044247787e-07, "epoch": 2.2256473489519113 }, { "current_steps": 1805, "loss": 2.0755, "learning_rate": 5.696586599241466e-07, "epoch": 2.226880394574599 }, { "current_steps": 1806, "loss": 2.261, "learning_rate": 5.694058154235145e-07, "epoch": 2.228113440197287 }, { "current_steps": 1807, "loss": 2.2832, "learning_rate": 5.691529709228824e-07, "epoch": 2.2293464858199754 }, { "current_steps": 1808, "loss": 1.8825, "learning_rate": 5.689001264222503e-07, "epoch": 2.2305795314426633 }, { "current_steps": 1809, "loss": 1.7573, "learning_rate": 5.686472819216182e-07, "epoch": 2.2318125770653516 }, { "current_steps": 1810, "loss": 2.1204, "learning_rate": 5.68394437420986e-07, "epoch": 2.2330456226880395 }, { "current_steps": 1811, "loss": 2.0086, "learning_rate": 5.681415929203539e-07, "epoch": 2.2342786683107274 }, { "current_steps": 1812, "loss": 1.9978, "learning_rate": 5.678887484197219e-07, "epoch": 2.2355117139334153 }, { "current_steps": 1813, "loss": 1.276, "learning_rate": 5.676359039190898e-07, "epoch": 2.2367447595561036 }, { "current_steps": 1814, "loss": 2.107, "learning_rate": 5.673830594184577e-07, "epoch": 2.2379778051787915 }, { "current_steps": 1815, "loss": 1.7145, "learning_rate": 5.671302149178256e-07, "epoch": 2.23921085080148 }, { "current_steps": 1816, "loss": 2.3391, "learning_rate": 5.668773704171934e-07, "epoch": 2.2404438964241677 }, { "current_steps": 1817, "loss": 1.6311, "learning_rate": 5.666245259165613e-07, "epoch": 2.2416769420468556 }, { "current_steps": 1818, "loss": 2.3436, "learning_rate": 5.663716814159291e-07, "epoch": 2.242909987669544 }, { "current_steps": 1819, "loss": 2.3432, "learning_rate": 5.66118836915297e-07, "epoch": 2.244143033292232 }, { "current_steps": 1820, "loss": 2.2033, "learning_rate": 5.658659924146649e-07, "epoch": 2.2453760789149197 }, { "current_steps": 1821, "loss": 2.0928, "learning_rate": 5.656131479140328e-07, "epoch": 2.246609124537608 }, { "current_steps": 1822, "loss": 2.4771, "learning_rate": 5.653603034134007e-07, "epoch": 2.247842170160296 }, { "current_steps": 1823, "loss": 2.3514, "learning_rate": 5.651074589127687e-07, "epoch": 2.249075215782984 }, { "current_steps": 1824, "loss": 2.4286, "learning_rate": 5.648546144121366e-07, "epoch": 2.250308261405672 }, { "current_steps": 1825, "loss": 1.9348, "learning_rate": 5.646017699115045e-07, "epoch": 2.25154130702836 }, { "current_steps": 1826, "loss": 1.63, "learning_rate": 5.643489254108723e-07, "epoch": 2.252774352651048 }, { "current_steps": 1827, "loss": 2.1674, "learning_rate": 5.640960809102402e-07, "epoch": 2.2540073982737363 }, { "current_steps": 1828, "loss": 1.9186, "learning_rate": 5.63843236409608e-07, "epoch": 2.255240443896424 }, { "current_steps": 1829, "loss": 2.397, "learning_rate": 5.635903919089759e-07, "epoch": 2.256473489519112 }, { "current_steps": 1830, "loss": 1.9018, "learning_rate": 5.633375474083438e-07, "epoch": 2.2577065351418004 }, { "current_steps": 1831, "loss": 2.3673, "learning_rate": 5.630847029077117e-07, "epoch": 2.2589395807644883 }, { "current_steps": 1832, "loss": 2.2433, "learning_rate": 5.628318584070796e-07, "epoch": 2.260172626387176 }, { "current_steps": 1833, "loss": 2.1391, "learning_rate": 5.625790139064475e-07, "epoch": 2.2614056720098645 }, { "current_steps": 1834, "loss": 1.854, "learning_rate": 5.623261694058153e-07, "epoch": 2.2626387176325524 }, { "current_steps": 1835, "loss": 1.7953, "learning_rate": 5.620733249051833e-07, "epoch": 2.2638717632552403 }, { "current_steps": 1836, "loss": 1.8381, "learning_rate": 5.618204804045512e-07, "epoch": 2.2651048088779286 }, { "current_steps": 1837, "loss": 2.0565, "learning_rate": 5.615676359039191e-07, "epoch": 2.2663378545006165 }, { "current_steps": 1838, "loss": 2.2528, "learning_rate": 5.61314791403287e-07, "epoch": 2.2675709001233044 }, { "current_steps": 1839, "loss": 2.1712, "learning_rate": 5.610619469026549e-07, "epoch": 2.2688039457459928 }, { "current_steps": 1840, "loss": 1.6019, "learning_rate": 5.608091024020228e-07, "epoch": 2.2700369913686806 }, { "current_steps": 1841, "loss": 2.0733, "learning_rate": 5.605562579013906e-07, "epoch": 2.2712700369913685 }, { "current_steps": 1842, "loss": 2.1259, "learning_rate": 5.603034134007584e-07, "epoch": 2.272503082614057 }, { "current_steps": 1843, "loss": 2.3048, "learning_rate": 5.600505689001263e-07, "epoch": 2.2737361282367448 }, { "current_steps": 1844, "loss": 1.8739, "learning_rate": 5.597977243994942e-07, "epoch": 2.2749691738594326 }, { "current_steps": 1845, "loss": 2.3759, "learning_rate": 5.595448798988621e-07, "epoch": 2.276202219482121 }, { "current_steps": 1846, "loss": 1.4303, "learning_rate": 5.592920353982301e-07, "epoch": 2.277435265104809 }, { "current_steps": 1847, "loss": 1.6787, "learning_rate": 5.59039190897598e-07, "epoch": 2.2786683107274968 }, { "current_steps": 1848, "loss": 2.3904, "learning_rate": 5.587863463969659e-07, "epoch": 2.279901356350185 }, { "current_steps": 1849, "loss": 2.2769, "learning_rate": 5.585335018963338e-07, "epoch": 2.281134401972873 }, { "current_steps": 1850, "loss": 2.0341, "learning_rate": 5.582806573957017e-07, "epoch": 2.282367447595561 }, { "current_steps": 1851, "loss": 2.1921, "learning_rate": 5.580278128950695e-07, "epoch": 2.283600493218249 }, { "current_steps": 1852, "loss": 2.0239, "learning_rate": 5.577749683944374e-07, "epoch": 2.284833538840937 }, { "current_steps": 1853, "loss": 2.0366, "learning_rate": 5.575221238938052e-07, "epoch": 2.286066584463625 }, { "current_steps": 1854, "loss": 2.1367, "learning_rate": 5.572692793931731e-07, "epoch": 2.2872996300863133 }, { "current_steps": 1855, "loss": 1.677, "learning_rate": 5.57016434892541e-07, "epoch": 2.288532675709001 }, { "current_steps": 1856, "loss": 2.1252, "learning_rate": 5.567635903919089e-07, "epoch": 2.289765721331689 }, { "current_steps": 1857, "loss": 1.8492, "learning_rate": 5.565107458912769e-07, "epoch": 2.2909987669543774 }, { "current_steps": 1858, "loss": 2.0364, "learning_rate": 5.562579013906448e-07, "epoch": 2.2922318125770653 }, { "current_steps": 1859, "loss": 2.0223, "learning_rate": 5.560050568900126e-07, "epoch": 2.293464858199753 }, { "current_steps": 1860, "loss": 2.1621, "learning_rate": 5.557522123893805e-07, "epoch": 2.2946979038224415 }, { "current_steps": 1861, "loss": 2.2167, "learning_rate": 5.554993678887484e-07, "epoch": 2.2959309494451294 }, { "current_steps": 1862, "loss": 1.9497, "learning_rate": 5.552465233881163e-07, "epoch": 2.2971639950678173 }, { "current_steps": 1863, "loss": 2.4056, "learning_rate": 5.549936788874842e-07, "epoch": 2.2983970406905057 }, { "current_steps": 1864, "loss": 1.9048, "learning_rate": 5.547408343868521e-07, "epoch": 2.2996300863131935 }, { "current_steps": 1865, "loss": 2.2819, "learning_rate": 5.5448798988622e-07, "epoch": 2.3008631319358814 }, { "current_steps": 1866, "loss": 2.2229, "learning_rate": 5.542351453855878e-07, "epoch": 2.3020961775585698 }, { "current_steps": 1867, "loss": 2.0165, "learning_rate": 5.539823008849556e-07, "epoch": 2.3033292231812577 }, { "current_steps": 1868, "loss": 2.3895, "learning_rate": 5.537294563843236e-07, "epoch": 2.304562268803946 }, { "current_steps": 1869, "loss": 2.1156, "learning_rate": 5.534766118836915e-07, "epoch": 2.305795314426634 }, { "current_steps": 1870, "loss": 2.4549, "learning_rate": 5.532237673830594e-07, "epoch": 2.3070283600493218 }, { "current_steps": 1871, "loss": 1.9371, "learning_rate": 5.529709228824273e-07, "epoch": 2.3082614056720097 }, { "current_steps": 1872, "loss": 1.8105, "learning_rate": 5.527180783817952e-07, "epoch": 2.309494451294698 }, { "current_steps": 1873, "loss": 1.9059, "learning_rate": 5.524652338811631e-07, "epoch": 2.310727496917386 }, { "current_steps": 1874, "loss": 2.0484, "learning_rate": 5.52212389380531e-07, "epoch": 2.311960542540074 }, { "current_steps": 1875, "loss": 2.1142, "learning_rate": 5.519595448798988e-07, "epoch": 2.313193588162762 }, { "current_steps": 1876, "loss": 2.0635, "learning_rate": 5.517067003792667e-07, "epoch": 2.31442663378545 }, { "current_steps": 1877, "loss": 2.3797, "learning_rate": 5.514538558786346e-07, "epoch": 2.315659679408138 }, { "current_steps": 1878, "loss": 2.2408, "learning_rate": 5.512010113780025e-07, "epoch": 2.316892725030826 }, { "current_steps": 1879, "loss": 1.9054, "learning_rate": 5.509481668773704e-07, "epoch": 2.318125770653514 }, { "current_steps": 1880, "loss": 1.1232, "learning_rate": 5.506953223767383e-07, "epoch": 2.3193588162762024 }, { "current_steps": 1881, "loss": 2.1564, "learning_rate": 5.504424778761062e-07, "epoch": 2.3205918618988903 }, { "current_steps": 1882, "loss": 2.1887, "learning_rate": 5.501896333754741e-07, "epoch": 2.321824907521578 }, { "current_steps": 1883, "loss": 2.4624, "learning_rate": 5.499367888748419e-07, "epoch": 2.323057953144266 }, { "current_steps": 1884, "loss": 2.1465, "learning_rate": 5.496839443742098e-07, "epoch": 2.3242909987669544 }, { "current_steps": 1885, "loss": 2.2654, "learning_rate": 5.494310998735777e-07, "epoch": 2.3255240443896423 }, { "current_steps": 1886, "loss": 2.1386, "learning_rate": 5.491782553729456e-07, "epoch": 2.3267570900123307 }, { "current_steps": 1887, "loss": 2.3304, "learning_rate": 5.489254108723135e-07, "epoch": 2.3279901356350186 }, { "current_steps": 1888, "loss": 1.8698, "learning_rate": 5.486725663716814e-07, "epoch": 2.3292231812577064 }, { "current_steps": 1889, "loss": 1.8811, "learning_rate": 5.484197218710493e-07, "epoch": 2.3304562268803943 }, { "current_steps": 1890, "loss": 2.1879, "learning_rate": 5.481668773704172e-07, "epoch": 2.3316892725030827 }, { "current_steps": 1891, "loss": 1.5258, "learning_rate": 5.479140328697852e-07, "epoch": 2.3329223181257706 }, { "current_steps": 1892, "loss": 2.2799, "learning_rate": 5.476611883691529e-07, "epoch": 2.334155363748459 }, { "current_steps": 1893, "loss": 2.1345, "learning_rate": 5.474083438685208e-07, "epoch": 2.335388409371147 }, { "current_steps": 1894, "loss": 2.0767, "learning_rate": 5.471554993678887e-07, "epoch": 2.3366214549938347 }, { "current_steps": 1895, "loss": 2.0601, "learning_rate": 5.469026548672566e-07, "epoch": 2.337854500616523 }, { "current_steps": 1896, "loss": 2.1897, "learning_rate": 5.466498103666245e-07, "epoch": 2.339087546239211 }, { "current_steps": 1897, "loss": 2.1738, "learning_rate": 5.463969658659924e-07, "epoch": 2.340320591861899 }, { "current_steps": 1898, "loss": 2.4638, "learning_rate": 5.461441213653603e-07, "epoch": 2.341553637484587 }, { "current_steps": 1899, "loss": 1.8911, "learning_rate": 5.458912768647282e-07, "epoch": 2.342786683107275 }, { "current_steps": 1900, "loss": 2.6031, "learning_rate": 5.45638432364096e-07, "epoch": 2.344019728729963 }, { "current_steps": 1901, "loss": 2.0204, "learning_rate": 5.453855878634639e-07, "epoch": 2.3452527743526512 }, { "current_steps": 1902, "loss": 1.879, "learning_rate": 5.451327433628319e-07, "epoch": 2.346485819975339 }, { "current_steps": 1903, "loss": 2.1376, "learning_rate": 5.448798988621998e-07, "epoch": 2.347718865598027 }, { "current_steps": 1904, "loss": 1.6895, "learning_rate": 5.446270543615676e-07, "epoch": 2.3489519112207153 }, { "current_steps": 1905, "loss": 2.1918, "learning_rate": 5.443742098609355e-07, "epoch": 2.3501849568434032 }, { "current_steps": 1906, "loss": 1.6838, "learning_rate": 5.441213653603034e-07, "epoch": 2.351418002466091 }, { "current_steps": 1907, "loss": 2.5302, "learning_rate": 5.438685208596713e-07, "epoch": 2.3526510480887795 }, { "current_steps": 1908, "loss": 2.1425, "learning_rate": 5.436156763590391e-07, "epoch": 2.3538840937114673 }, { "current_steps": 1909, "loss": 2.0898, "learning_rate": 5.43362831858407e-07, "epoch": 2.3551171393341552 }, { "current_steps": 1910, "loss": 2.5221, "learning_rate": 5.431099873577749e-07, "epoch": 2.3563501849568436 }, { "current_steps": 1911, "loss": 2.202, "learning_rate": 5.428571428571428e-07, "epoch": 2.3575832305795315 }, { "current_steps": 1912, "loss": 1.8586, "learning_rate": 5.426042983565107e-07, "epoch": 2.3588162762022193 }, { "current_steps": 1913, "loss": 1.7676, "learning_rate": 5.423514538558787e-07, "epoch": 2.3600493218249077 }, { "current_steps": 1914, "loss": 1.9155, "learning_rate": 5.420986093552466e-07, "epoch": 2.3612823674475956 }, { "current_steps": 1915, "loss": 1.7532, "learning_rate": 5.418457648546145e-07, "epoch": 2.3625154130702835 }, { "current_steps": 1916, "loss": 2.2353, "learning_rate": 5.415929203539823e-07, "epoch": 2.363748458692972 }, { "current_steps": 1917, "loss": 2.3574, "learning_rate": 5.413400758533501e-07, "epoch": 2.3649815043156597 }, { "current_steps": 1918, "loss": 1.5782, "learning_rate": 5.41087231352718e-07, "epoch": 2.3662145499383476 }, { "current_steps": 1919, "loss": 1.6307, "learning_rate": 5.408343868520859e-07, "epoch": 2.367447595561036 }, { "current_steps": 1920, "loss": 1.6337, "learning_rate": 5.405815423514538e-07, "epoch": 2.368680641183724 }, { "current_steps": 1921, "loss": 2.0917, "learning_rate": 5.403286978508217e-07, "epoch": 2.3699136868064117 }, { "current_steps": 1922, "loss": 1.9608, "learning_rate": 5.400758533501896e-07, "epoch": 2.3711467324291 }, { "current_steps": 1923, "loss": 2.195, "learning_rate": 5.398230088495575e-07, "epoch": 2.372379778051788 }, { "current_steps": 1924, "loss": 1.5768, "learning_rate": 5.395701643489254e-07, "epoch": 2.373612823674476 }, { "current_steps": 1925, "loss": 1.7032, "learning_rate": 5.393173198482933e-07, "epoch": 2.374845869297164 }, { "current_steps": 1926, "loss": 2.3284, "learning_rate": 5.390644753476612e-07, "epoch": 2.376078914919852 }, { "current_steps": 1927, "loss": 1.843, "learning_rate": 5.388116308470291e-07, "epoch": 2.37731196054254 }, { "current_steps": 1928, "loss": 2.1787, "learning_rate": 5.38558786346397e-07, "epoch": 2.3785450061652282 }, { "current_steps": 1929, "loss": 2.1512, "learning_rate": 5.383059418457648e-07, "epoch": 2.379778051787916 }, { "current_steps": 1930, "loss": 1.6717, "learning_rate": 5.380530973451327e-07, "epoch": 2.381011097410604 }, { "current_steps": 1931, "loss": 2.6849, "learning_rate": 5.378002528445006e-07, "epoch": 2.3822441430332923 }, { "current_steps": 1932, "loss": 2.3394, "learning_rate": 5.375474083438684e-07, "epoch": 2.3834771886559802 }, { "current_steps": 1933, "loss": 2.2101, "learning_rate": 5.372945638432363e-07, "epoch": 2.384710234278668 }, { "current_steps": 1934, "loss": 1.5214, "learning_rate": 5.370417193426042e-07, "epoch": 2.3859432799013565 }, { "current_steps": 1935, "loss": 2.167, "learning_rate": 5.367888748419722e-07, "epoch": 2.3871763255240444 }, { "current_steps": 1936, "loss": 2.0329, "learning_rate": 5.365360303413401e-07, "epoch": 2.3884093711467322 }, { "current_steps": 1937, "loss": 1.8807, "learning_rate": 5.36283185840708e-07, "epoch": 2.3896424167694206 }, { "current_steps": 1938, "loss": 1.8081, "learning_rate": 5.360303413400759e-07, "epoch": 2.3908754623921085 }, { "current_steps": 1939, "loss": 2.464, "learning_rate": 5.357774968394438e-07, "epoch": 2.392108508014797 }, { "current_steps": 1940, "loss": 2.207, "learning_rate": 5.355246523388117e-07, "epoch": 2.3933415536374847 }, { "current_steps": 1941, "loss": 2.1187, "learning_rate": 5.352718078381795e-07, "epoch": 2.3945745992601726 }, { "current_steps": 1942, "loss": 1.8605, "learning_rate": 5.350189633375473e-07, "epoch": 2.3958076448828605 }, { "current_steps": 1943, "loss": 1.6229, "learning_rate": 5.347661188369152e-07, "epoch": 2.397040690505549 }, { "current_steps": 1944, "loss": 1.9935, "learning_rate": 5.345132743362831e-07, "epoch": 2.3982737361282367 }, { "current_steps": 1945, "loss": 2.1059, "learning_rate": 5.34260429835651e-07, "epoch": 2.399506781750925 }, { "current_steps": 1946, "loss": 1.776, "learning_rate": 5.340075853350189e-07, "epoch": 2.400739827373613 }, { "current_steps": 1947, "loss": 2.442, "learning_rate": 5.337547408343869e-07, "epoch": 2.401972872996301 }, { "current_steps": 1948, "loss": 1.7228, "learning_rate": 5.335018963337548e-07, "epoch": 2.4032059186189887 }, { "current_steps": 1949, "loss": 1.9118, "learning_rate": 5.332490518331226e-07, "epoch": 2.404438964241677 }, { "current_steps": 1950, "loss": 2.2523, "learning_rate": 5.329962073324905e-07, "epoch": 2.405672009864365 }, { "current_steps": 1951, "loss": 2.4058, "learning_rate": 5.327433628318584e-07, "epoch": 2.4069050554870532 }, { "current_steps": 1952, "loss": 1.6808, "learning_rate": 5.324905183312263e-07, "epoch": 2.408138101109741 }, { "current_steps": 1953, "loss": 2.1219, "learning_rate": 5.322376738305942e-07, "epoch": 2.409371146732429 }, { "current_steps": 1954, "loss": 1.9899, "learning_rate": 5.319848293299621e-07, "epoch": 2.410604192355117 }, { "current_steps": 1955, "loss": 2.0692, "learning_rate": 5.317319848293299e-07, "epoch": 2.4118372379778052 }, { "current_steps": 1956, "loss": 2.3735, "learning_rate": 5.314791403286977e-07, "epoch": 2.413070283600493 }, { "current_steps": 1957, "loss": 2.0816, "learning_rate": 5.312262958280656e-07, "epoch": 2.4143033292231815 }, { "current_steps": 1958, "loss": 2.2902, "learning_rate": 5.309734513274336e-07, "epoch": 2.4155363748458694 }, { "current_steps": 1959, "loss": 2.1123, "learning_rate": 5.307206068268015e-07, "epoch": 2.4167694204685573 }, { "current_steps": 1960, "loss": 1.8449, "learning_rate": 5.304677623261694e-07, "epoch": 2.418002466091245 }, { "current_steps": 1961, "loss": 2.372, "learning_rate": 5.302149178255373e-07, "epoch": 2.4192355117139335 }, { "current_steps": 1962, "loss": 1.8454, "learning_rate": 5.299620733249052e-07, "epoch": 2.4204685573366214 }, { "current_steps": 1963, "loss": 2.064, "learning_rate": 5.297092288242731e-07, "epoch": 2.4217016029593097 }, { "current_steps": 1964, "loss": 1.9405, "learning_rate": 5.29456384323641e-07, "epoch": 2.4229346485819976 }, { "current_steps": 1965, "loss": 2.3305, "learning_rate": 5.292035398230088e-07, "epoch": 2.4241676942046855 }, { "current_steps": 1966, "loss": 1.5394, "learning_rate": 5.289506953223767e-07, "epoch": 2.4254007398273734 }, { "current_steps": 1967, "loss": 2.1274, "learning_rate": 5.286978508217445e-07, "epoch": 2.4266337854500617 }, { "current_steps": 1968, "loss": 1.5298, "learning_rate": 5.284450063211124e-07, "epoch": 2.4278668310727496 }, { "current_steps": 1969, "loss": 2.1553, "learning_rate": 5.281921618204804e-07, "epoch": 2.429099876695438 }, { "current_steps": 1970, "loss": 2.1567, "learning_rate": 5.279393173198483e-07, "epoch": 2.430332922318126 }, { "current_steps": 1971, "loss": 2.3023, "learning_rate": 5.276864728192162e-07, "epoch": 2.4315659679408137 }, { "current_steps": 1972, "loss": 1.4114, "learning_rate": 5.274336283185841e-07, "epoch": 2.432799013563502 }, { "current_steps": 1973, "loss": 2.2481, "learning_rate": 5.271807838179519e-07, "epoch": 2.43403205918619 }, { "current_steps": 1974, "loss": 1.9738, "learning_rate": 5.269279393173198e-07, "epoch": 2.435265104808878 }, { "current_steps": 1975, "loss": 2.1533, "learning_rate": 5.266750948166877e-07, "epoch": 2.436498150431566 }, { "current_steps": 1976, "loss": 2.3185, "learning_rate": 5.264222503160556e-07, "epoch": 2.437731196054254 }, { "current_steps": 1977, "loss": 1.7884, "learning_rate": 5.261694058154235e-07, "epoch": 2.438964241676942 }, { "current_steps": 1978, "loss": 2.3809, "learning_rate": 5.259165613147914e-07, "epoch": 2.4401972872996303 }, { "current_steps": 1979, "loss": 2.3135, "learning_rate": 5.256637168141593e-07, "epoch": 2.441430332922318 }, { "current_steps": 1980, "loss": 2.2796, "learning_rate": 5.254108723135272e-07, "epoch": 2.442663378545006 }, { "current_steps": 1981, "loss": 2.3535, "learning_rate": 5.25158027812895e-07, "epoch": 2.4438964241676944 }, { "current_steps": 1982, "loss": 2.1039, "learning_rate": 5.249051833122629e-07, "epoch": 2.4451294697903823 }, { "current_steps": 1983, "loss": 2.3974, "learning_rate": 5.246523388116308e-07, "epoch": 2.44636251541307 }, { "current_steps": 1984, "loss": 2.1386, "learning_rate": 5.243994943109987e-07, "epoch": 2.4475955610357585 }, { "current_steps": 1985, "loss": 2.0043, "learning_rate": 5.241466498103666e-07, "epoch": 2.4488286066584464 }, { "current_steps": 1986, "loss": 2.6987, "learning_rate": 5.238938053097345e-07, "epoch": 2.4500616522811343 }, { "current_steps": 1987, "loss": 1.9259, "learning_rate": 5.236409608091024e-07, "epoch": 2.4512946979038226 }, { "current_steps": 1988, "loss": 2.4087, "learning_rate": 5.233881163084703e-07, "epoch": 2.4525277435265105 }, { "current_steps": 1989, "loss": 1.8459, "learning_rate": 5.231352718078381e-07, "epoch": 2.4537607891491984 }, { "current_steps": 1990, "loss": 2.379, "learning_rate": 5.22882427307206e-07, "epoch": 2.4549938347718867 }, { "current_steps": 1991, "loss": 1.9884, "learning_rate": 5.22629582806574e-07, "epoch": 2.4562268803945746 }, { "current_steps": 1992, "loss": 1.716, "learning_rate": 5.223767383059419e-07, "epoch": 2.4574599260172625 }, { "current_steps": 1993, "loss": 2.005, "learning_rate": 5.221238938053097e-07, "epoch": 2.458692971639951 }, { "current_steps": 1994, "loss": 1.9038, "learning_rate": 5.218710493046776e-07, "epoch": 2.4599260172626387 }, { "current_steps": 1995, "loss": 2.2732, "learning_rate": 5.216182048040455e-07, "epoch": 2.4611590628853266 }, { "current_steps": 1996, "loss": 1.9992, "learning_rate": 5.213653603034134e-07, "epoch": 2.462392108508015 }, { "current_steps": 1997, "loss": 2.1844, "learning_rate": 5.211125158027812e-07, "epoch": 2.463625154130703 }, { "current_steps": 1998, "loss": 1.9383, "learning_rate": 5.208596713021491e-07, "epoch": 2.4648581997533907 }, { "current_steps": 1999, "loss": 2.0025, "learning_rate": 5.20606826801517e-07, "epoch": 2.466091245376079 }, { "current_steps": 2000, "loss": 2.2383, "learning_rate": 5.203539823008849e-07, "epoch": 2.467324290998767 }, { "current_steps": 2001, "loss": 1.6922, "learning_rate": 5.201011378002528e-07, "epoch": 2.468557336621455 }, { "current_steps": 2002, "loss": 2.0749, "learning_rate": 5.198482932996207e-07, "epoch": 2.469790382244143 }, { "current_steps": 2003, "loss": 1.7476, "learning_rate": 5.195954487989887e-07, "epoch": 2.471023427866831 }, { "current_steps": 2004, "loss": 2.606, "learning_rate": 5.193426042983566e-07, "epoch": 2.472256473489519 }, { "current_steps": 2005, "loss": 2.0341, "learning_rate": 5.190897597977244e-07, "epoch": 2.4734895191122073 }, { "current_steps": 2006, "loss": 1.8913, "learning_rate": 5.188369152970922e-07, "epoch": 2.474722564734895 }, { "current_steps": 2007, "loss": 2.1204, "learning_rate": 5.185840707964601e-07, "epoch": 2.475955610357583 }, { "current_steps": 2008, "loss": 1.8273, "learning_rate": 5.18331226295828e-07, "epoch": 2.4771886559802714 }, { "current_steps": 2009, "loss": 2.3065, "learning_rate": 5.180783817951959e-07, "epoch": 2.4784217016029593 }, { "current_steps": 2010, "loss": 2.3006, "learning_rate": 5.178255372945638e-07, "epoch": 2.479654747225647 }, { "current_steps": 2011, "loss": 1.9093, "learning_rate": 5.175726927939317e-07, "epoch": 2.4808877928483355 }, { "current_steps": 2012, "loss": 2.0809, "learning_rate": 5.173198482932996e-07, "epoch": 2.4821208384710234 }, { "current_steps": 2013, "loss": 2.5947, "learning_rate": 5.170670037926675e-07, "epoch": 2.4833538840937113 }, { "current_steps": 2014, "loss": 1.7512, "learning_rate": 5.168141592920354e-07, "epoch": 2.4845869297163996 }, { "current_steps": 2015, "loss": 2.1908, "learning_rate": 5.165613147914033e-07, "epoch": 2.4858199753390875 }, { "current_steps": 2016, "loss": 1.8389, "learning_rate": 5.163084702907712e-07, "epoch": 2.487053020961776 }, { "current_steps": 2017, "loss": 1.9734, "learning_rate": 5.160556257901391e-07, "epoch": 2.4882860665844637 }, { "current_steps": 2018, "loss": 2.0126, "learning_rate": 5.158027812895069e-07, "epoch": 2.4895191122071516 }, { "current_steps": 2019, "loss": 2.1779, "learning_rate": 5.155499367888748e-07, "epoch": 2.4907521578298395 }, { "current_steps": 2020, "loss": 2.0832, "learning_rate": 5.152970922882427e-07, "epoch": 2.491985203452528 }, { "current_steps": 2021, "loss": 2.1883, "learning_rate": 5.150442477876106e-07, "epoch": 2.4932182490752157 }, { "current_steps": 2022, "loss": 2.0659, "learning_rate": 5.147914032869784e-07, "epoch": 2.494451294697904 }, { "current_steps": 2023, "loss": 2.3714, "learning_rate": 5.145385587863463e-07, "epoch": 2.495684340320592 }, { "current_steps": 2024, "loss": 2.2169, "learning_rate": 5.142857142857142e-07, "epoch": 2.49691738594328 }, { "current_steps": 2025, "loss": 1.9577, "learning_rate": 5.140328697850822e-07, "epoch": 2.4981504315659677 }, { "current_steps": 2026, "loss": 1.9975, "learning_rate": 5.137800252844501e-07, "epoch": 2.499383477188656 }, { "current_steps": 2027, "loss": 2.0456, "learning_rate": 5.13527180783818e-07, "epoch": 2.500616522811344 }, { "current_steps": 2028, "loss": 1.4224, "learning_rate": 5.132743362831859e-07, "epoch": 2.5018495684340323 }, { "current_steps": 2029, "loss": 1.4711, "learning_rate": 5.130214917825538e-07, "epoch": 2.50308261405672 }, { "current_steps": 2030, "loss": 1.719, "learning_rate": 5.127686472819215e-07, "epoch": 2.504315659679408 }, { "current_steps": 2031, "loss": 1.721, "learning_rate": 5.125158027812894e-07, "epoch": 2.505548705302096 }, { "current_steps": 2032, "loss": 2.1662, "learning_rate": 5.122629582806573e-07, "epoch": 2.5067817509247843 }, { "current_steps": 2033, "loss": 2.0037, "learning_rate": 5.120101137800252e-07, "epoch": 2.508014796547472 }, { "current_steps": 2034, "loss": 1.9854, "learning_rate": 5.117572692793931e-07, "epoch": 2.5092478421701605 }, { "current_steps": 2035, "loss": 1.987, "learning_rate": 5.11504424778761e-07, "epoch": 2.5104808877928484 }, { "current_steps": 2036, "loss": 2.3984, "learning_rate": 5.11251580278129e-07, "epoch": 2.5117139334155363 }, { "current_steps": 2037, "loss": 2.1548, "learning_rate": 5.109987357774969e-07, "epoch": 2.512946979038224 }, { "current_steps": 2038, "loss": 2.2404, "learning_rate": 5.107458912768648e-07, "epoch": 2.5141800246609125 }, { "current_steps": 2039, "loss": 2.6486, "learning_rate": 5.104930467762326e-07, "epoch": 2.5154130702836004 }, { "current_steps": 2040, "loss": 2.206, "learning_rate": 5.102402022756005e-07, "epoch": 2.5166461159062887 }, { "current_steps": 2041, "loss": 1.775, "learning_rate": 5.099873577749684e-07, "epoch": 2.5178791615289766 }, { "current_steps": 2042, "loss": 1.4774, "learning_rate": 5.097345132743363e-07, "epoch": 2.5191122071516645 }, { "current_steps": 2043, "loss": 2.2591, "learning_rate": 5.094816687737041e-07, "epoch": 2.5203452527743524 }, { "current_steps": 2044, "loss": 1.5831, "learning_rate": 5.09228824273072e-07, "epoch": 2.5215782983970407 }, { "current_steps": 2045, "loss": 1.9318, "learning_rate": 5.089759797724399e-07, "epoch": 2.5228113440197286 }, { "current_steps": 2046, "loss": 2.0027, "learning_rate": 5.087231352718077e-07, "epoch": 2.524044389642417 }, { "current_steps": 2047, "loss": 2.1909, "learning_rate": 5.084702907711757e-07, "epoch": 2.525277435265105 }, { "current_steps": 2048, "loss": 1.9033, "learning_rate": 5.082174462705436e-07, "epoch": 2.5265104808877927 }, { "current_steps": 2049, "loss": 2.3561, "learning_rate": 5.079646017699115e-07, "epoch": 2.5277435265104806 }, { "current_steps": 2050, "loss": 1.8842, "learning_rate": 5.077117572692794e-07, "epoch": 2.528976572133169 }, { "current_steps": 2051, "loss": 2.094, "learning_rate": 5.074589127686473e-07, "epoch": 2.530209617755857 }, { "current_steps": 2052, "loss": 2.4883, "learning_rate": 5.072060682680152e-07, "epoch": 2.531442663378545 }, { "current_steps": 2053, "loss": 1.7497, "learning_rate": 5.069532237673831e-07, "epoch": 2.532675709001233 }, { "current_steps": 2054, "loss": 2.3659, "learning_rate": 5.06700379266751e-07, "epoch": 2.533908754623921 }, { "current_steps": 2055, "loss": 1.5777, "learning_rate": 5.064475347661188e-07, "epoch": 2.5351418002466093 }, { "current_steps": 2056, "loss": 1.9843, "learning_rate": 5.061946902654866e-07, "epoch": 2.536374845869297 }, { "current_steps": 2057, "loss": 1.7762, "learning_rate": 5.059418457648545e-07, "epoch": 2.537607891491985 }, { "current_steps": 2058, "loss": 1.7336, "learning_rate": 5.056890012642224e-07, "epoch": 2.5388409371146734 }, { "current_steps": 2059, "loss": 2.0604, "learning_rate": 5.054361567635904e-07, "epoch": 2.5400739827373613 }, { "current_steps": 2060, "loss": 2.3813, "learning_rate": 5.051833122629583e-07, "epoch": 2.541307028360049 }, { "current_steps": 2061, "loss": 1.9804, "learning_rate": 5.049304677623262e-07, "epoch": 2.5425400739827375 }, { "current_steps": 2062, "loss": 2.1699, "learning_rate": 5.046776232616941e-07, "epoch": 2.5437731196054254 }, { "current_steps": 2063, "loss": 2.0554, "learning_rate": 5.044247787610619e-07, "epoch": 2.5450061652281133 }, { "current_steps": 2064, "loss": 1.8685, "learning_rate": 5.041719342604298e-07, "epoch": 2.5462392108508016 }, { "current_steps": 2065, "loss": 2.2202, "learning_rate": 5.039190897597977e-07, "epoch": 2.5474722564734895 }, { "current_steps": 2066, "loss": 1.5926, "learning_rate": 5.036662452591656e-07, "epoch": 2.5487053020961774 }, { "current_steps": 2067, "loss": 2.0562, "learning_rate": 5.034134007585335e-07, "epoch": 2.5499383477188657 }, { "current_steps": 2068, "loss": 2.1018, "learning_rate": 5.031605562579013e-07, "epoch": 2.5511713933415536 }, { "current_steps": 2069, "loss": 2.1275, "learning_rate": 5.029077117572692e-07, "epoch": 2.5524044389642415 }, { "current_steps": 2070, "loss": 2.0643, "learning_rate": 5.026548672566372e-07, "epoch": 2.55363748458693 }, { "current_steps": 2071, "loss": 2.0036, "learning_rate": 5.02402022756005e-07, "epoch": 2.5548705302096177 }, { "current_steps": 2072, "loss": 2.201, "learning_rate": 5.021491782553729e-07, "epoch": 2.5561035758323056 }, { "current_steps": 2073, "loss": 2.0972, "learning_rate": 5.018963337547408e-07, "epoch": 2.557336621454994 }, { "current_steps": 2074, "loss": 2.3432, "learning_rate": 5.016434892541087e-07, "epoch": 2.558569667077682 }, { "current_steps": 2075, "loss": 2.4124, "learning_rate": 5.013906447534766e-07, "epoch": 2.55980271270037 }, { "current_steps": 2076, "loss": 1.3698, "learning_rate": 5.011378002528445e-07, "epoch": 2.561035758323058 }, { "current_steps": 2077, "loss": 2.1611, "learning_rate": 5.008849557522124e-07, "epoch": 2.562268803945746 }, { "current_steps": 2078, "loss": 1.9342, "learning_rate": 5.006321112515803e-07, "epoch": 2.563501849568434 }, { "current_steps": 2079, "loss": 2.2614, "learning_rate": 5.003792667509481e-07, "epoch": 2.564734895191122 }, { "current_steps": 2080, "loss": 2.4443, "learning_rate": 5.00126422250316e-07, "epoch": 2.56596794081381 }, { "current_steps": 2081, "loss": 2.2623, "learning_rate": 4.998735777496839e-07, "epoch": 2.5672009864364984 }, { "current_steps": 2082, "loss": 2.1004, "learning_rate": 4.996207332490518e-07, "epoch": 2.5684340320591863 }, { "current_steps": 2083, "loss": 2.1207, "learning_rate": 4.993678887484197e-07, "epoch": 2.569667077681874 }, { "current_steps": 2084, "loss": 1.8455, "learning_rate": 4.991150442477876e-07, "epoch": 2.570900123304562 }, { "current_steps": 2085, "loss": 2.3096, "learning_rate": 4.988621997471555e-07, "epoch": 2.5721331689272504 }, { "current_steps": 2086, "loss": 2.3304, "learning_rate": 4.986093552465234e-07, "epoch": 2.5733662145499383 }, { "current_steps": 2087, "loss": 1.791, "learning_rate": 4.983565107458912e-07, "epoch": 2.5745992601726266 }, { "current_steps": 2088, "loss": 1.9399, "learning_rate": 4.981036662452591e-07, "epoch": 2.5758323057953145 }, { "current_steps": 2089, "loss": 1.5982, "learning_rate": 4.97850821744627e-07, "epoch": 2.5770653514180024 }, { "current_steps": 2090, "loss": 2.3686, "learning_rate": 4.975979772439949e-07, "epoch": 2.5782983970406903 }, { "current_steps": 2091, "loss": 2.1912, "learning_rate": 4.973451327433628e-07, "epoch": 2.5795314426633786 }, { "current_steps": 2092, "loss": 2.0189, "learning_rate": 4.970922882427307e-07, "epoch": 2.5807644882860665 }, { "current_steps": 2093, "loss": 1.9292, "learning_rate": 4.968394437420986e-07, "epoch": 2.581997533908755 }, { "current_steps": 2094, "loss": 2.3027, "learning_rate": 4.965865992414665e-07, "epoch": 2.5832305795314427 }, { "current_steps": 2095, "loss": 1.2611, "learning_rate": 4.963337547408343e-07, "epoch": 2.5844636251541306 }, { "current_steps": 2096, "loss": 2.1373, "learning_rate": 4.960809102402022e-07, "epoch": 2.5856966707768185 }, { "current_steps": 2097, "loss": 1.8223, "learning_rate": 4.958280657395701e-07, "epoch": 2.586929716399507 }, { "current_steps": 2098, "loss": 2.2464, "learning_rate": 4.95575221238938e-07, "epoch": 2.5881627620221948 }, { "current_steps": 2099, "loss": 1.6519, "learning_rate": 4.953223767383059e-07, "epoch": 2.589395807644883 }, { "current_steps": 2100, "loss": 2.1999, "learning_rate": 4.950695322376738e-07, "epoch": 2.590628853267571 }, { "current_steps": 2101, "loss": 1.8848, "learning_rate": 4.948166877370417e-07, "epoch": 2.591861898890259 }, { "current_steps": 2102, "loss": 2.3262, "learning_rate": 4.945638432364096e-07, "epoch": 2.5930949445129468 }, { "current_steps": 2103, "loss": 2.1122, "learning_rate": 4.943109987357775e-07, "epoch": 2.594327990135635 }, { "current_steps": 2104, "loss": 1.5732, "learning_rate": 4.940581542351453e-07, "epoch": 2.595561035758323 }, { "current_steps": 2105, "loss": 2.1255, "learning_rate": 4.938053097345132e-07, "epoch": 2.5967940813810113 }, { "current_steps": 2106, "loss": 2.514, "learning_rate": 4.935524652338811e-07, "epoch": 2.598027127003699 }, { "current_steps": 2107, "loss": 2.3301, "learning_rate": 4.93299620733249e-07, "epoch": 2.599260172626387 }, { "current_steps": 2108, "loss": 2.4071, "learning_rate": 4.930467762326169e-07, "epoch": 2.600493218249075 }, { "current_steps": 2109, "loss": 1.806, "learning_rate": 4.927939317319848e-07, "epoch": 2.6017262638717633 }, { "current_steps": 2110, "loss": 1.8715, "learning_rate": 4.925410872313527e-07, "epoch": 2.602959309494451 }, { "current_steps": 2111, "loss": 2.5184, "learning_rate": 4.922882427307206e-07, "epoch": 2.6041923551171395 }, { "current_steps": 2112, "loss": 2.2656, "learning_rate": 4.920353982300885e-07, "epoch": 2.6054254007398274 }, { "current_steps": 2113, "loss": 2.3511, "learning_rate": 4.917825537294563e-07, "epoch": 2.6066584463625153 }, { "current_steps": 2114, "loss": 1.3379, "learning_rate": 4.915297092288242e-07, "epoch": 2.607891491985203 }, { "current_steps": 2115, "loss": 1.9445, "learning_rate": 4.912768647281921e-07, "epoch": 2.6091245376078915 }, { "current_steps": 2116, "loss": 1.9542, "learning_rate": 4.9102402022756e-07, "epoch": 2.6103575832305794 }, { "current_steps": 2117, "loss": 2.077, "learning_rate": 4.90771175726928e-07, "epoch": 2.6115906288532678 }, { "current_steps": 2118, "loss": 1.9714, "learning_rate": 4.905183312262959e-07, "epoch": 2.6128236744759556 }, { "current_steps": 2119, "loss": 2.5052, "learning_rate": 4.902654867256637e-07, "epoch": 2.6140567200986435 }, { "current_steps": 2120, "loss": 2.2148, "learning_rate": 4.900126422250315e-07, "epoch": 2.6152897657213314 }, { "current_steps": 2121, "loss": 2.3291, "learning_rate": 4.897597977243994e-07, "epoch": 2.6165228113440198 }, { "current_steps": 2122, "loss": 1.5848, "learning_rate": 4.895069532237673e-07, "epoch": 2.6177558569667077 }, { "current_steps": 2123, "loss": 2.2092, "learning_rate": 4.892541087231353e-07, "epoch": 2.618988902589396 }, { "current_steps": 2124, "loss": 2.3153, "learning_rate": 4.890012642225032e-07, "epoch": 2.620221948212084 }, { "current_steps": 2125, "loss": 1.8346, "learning_rate": 4.88748419721871e-07, "epoch": 2.6214549938347718 }, { "current_steps": 2126, "loss": 2.2321, "learning_rate": 4.884955752212389e-07, "epoch": 2.6226880394574597 }, { "current_steps": 2127, "loss": 1.058, "learning_rate": 4.882427307206068e-07, "epoch": 2.623921085080148 }, { "current_steps": 2128, "loss": 2.0767, "learning_rate": 4.879898862199746e-07, "epoch": 2.625154130702836 }, { "current_steps": 2129, "loss": 1.6547, "learning_rate": 4.877370417193426e-07, "epoch": 2.626387176325524 }, { "current_steps": 2130, "loss": 2.0054, "learning_rate": 4.874841972187105e-07, "epoch": 2.627620221948212 }, { "current_steps": 2131, "loss": 2.2274, "learning_rate": 4.872313527180784e-07, "epoch": 2.6288532675709 }, { "current_steps": 2132, "loss": 2.0742, "learning_rate": 4.869785082174462e-07, "epoch": 2.6300863131935883 }, { "current_steps": 2133, "loss": 2.2762, "learning_rate": 4.867256637168141e-07, "epoch": 2.631319358816276 }, { "current_steps": 2134, "loss": 2.2019, "learning_rate": 4.864728192161821e-07, "epoch": 2.632552404438964 }, { "current_steps": 2135, "loss": 1.5852, "learning_rate": 4.8621997471555e-07, "epoch": 2.6337854500616524 }, { "current_steps": 2136, "loss": 1.8731, "learning_rate": 4.859671302149178e-07, "epoch": 2.6350184956843403 }, { "current_steps": 2137, "loss": 2.6003, "learning_rate": 4.857142857142857e-07, "epoch": 2.636251541307028 }, { "current_steps": 2138, "loss": 2.1031, "learning_rate": 4.854614412136535e-07, "epoch": 2.6374845869297165 }, { "current_steps": 2139, "loss": 1.673, "learning_rate": 4.852085967130214e-07, "epoch": 2.6387176325524044 }, { "current_steps": 2140, "loss": 2.3701, "learning_rate": 4.849557522123894e-07, "epoch": 2.6399506781750923 }, { "current_steps": 2141, "loss": 2.2592, "learning_rate": 4.847029077117573e-07, "epoch": 2.6411837237977807 }, { "current_steps": 2142, "loss": 1.8688, "learning_rate": 4.844500632111252e-07, "epoch": 2.6424167694204685 }, { "current_steps": 2143, "loss": 2.0071, "learning_rate": 4.841972187104931e-07, "epoch": 2.6436498150431564 }, { "current_steps": 2144, "loss": 2.4172, "learning_rate": 4.839443742098608e-07, "epoch": 2.6448828606658448 }, { "current_steps": 2145, "loss": 2.0956, "learning_rate": 4.836915297092288e-07, "epoch": 2.6461159062885327 }, { "current_steps": 2146, "loss": 2.2457, "learning_rate": 4.834386852085967e-07, "epoch": 2.6473489519112205 }, { "current_steps": 2147, "loss": 2.1983, "learning_rate": 4.831858407079646e-07, "epoch": 2.648581997533909 }, { "current_steps": 2148, "loss": 2.4775, "learning_rate": 4.829329962073325e-07, "epoch": 2.6498150431565968 }, { "current_steps": 2149, "loss": 2.0197, "learning_rate": 4.826801517067004e-07, "epoch": 2.6510480887792847 }, { "current_steps": 2150, "loss": 2.1035, "learning_rate": 4.824273072060682e-07, "epoch": 2.652281134401973 }, { "current_steps": 2151, "loss": 2.3899, "learning_rate": 4.821744627054362e-07, "epoch": 2.653514180024661 }, { "current_steps": 2152, "loss": 2.0278, "learning_rate": 4.819216182048041e-07, "epoch": 2.654747225647349 }, { "current_steps": 2153, "loss": 2.44, "learning_rate": 4.816687737041719e-07, "epoch": 2.655980271270037 }, { "current_steps": 2154, "loss": 1.9523, "learning_rate": 4.814159292035398e-07, "epoch": 2.657213316892725 }, { "current_steps": 2155, "loss": 1.6776, "learning_rate": 4.811630847029077e-07, "epoch": 2.658446362515413 }, { "current_steps": 2156, "loss": 1.9909, "learning_rate": 4.809102402022756e-07, "epoch": 2.659679408138101 }, { "current_steps": 2157, "loss": 2.0946, "learning_rate": 4.806573957016435e-07, "epoch": 2.660912453760789 }, { "current_steps": 2158, "loss": 2.433, "learning_rate": 4.804045512010114e-07, "epoch": 2.6621454993834774 }, { "current_steps": 2159, "loss": 1.7948, "learning_rate": 4.801517067003793e-07, "epoch": 2.6633785450061653 }, { "current_steps": 2160, "loss": 2.261, "learning_rate": 4.798988621997472e-07, "epoch": 2.664611590628853 }, { "current_steps": 2161, "loss": 1.9355, "learning_rate": 4.79646017699115e-07, "epoch": 2.665844636251541 }, { "current_steps": 2162, "loss": 1.9367, "learning_rate": 4.793931731984829e-07, "epoch": 2.6670776818742294 }, { "current_steps": 2163, "loss": 1.558, "learning_rate": 4.791403286978508e-07, "epoch": 2.6683107274969173 }, { "current_steps": 2164, "loss": 2.2885, "learning_rate": 4.788874841972187e-07, "epoch": 2.6695437731196057 }, { "current_steps": 2165, "loss": 2.2136, "learning_rate": 4.786346396965866e-07, "epoch": 2.6707768187422936 }, { "current_steps": 2166, "loss": 1.8361, "learning_rate": 4.783817951959545e-07, "epoch": 2.6720098643649814 }, { "current_steps": 2167, "loss": 2.0921, "learning_rate": 4.781289506953224e-07, "epoch": 2.6732429099876693 }, { "current_steps": 2168, "loss": 1.889, "learning_rate": 4.778761061946903e-07, "epoch": 2.6744759556103577 }, { "current_steps": 2169, "loss": 2.1278, "learning_rate": 4.776232616940581e-07, "epoch": 2.6757090012330456 }, { "current_steps": 2170, "loss": 2.0927, "learning_rate": 4.77370417193426e-07, "epoch": 2.676942046855734 }, { "current_steps": 2171, "loss": 2.3545, "learning_rate": 4.771175726927939e-07, "epoch": 2.678175092478422 }, { "current_steps": 2172, "loss": 1.9413, "learning_rate": 4.768647281921618e-07, "epoch": 2.6794081381011097 }, { "current_steps": 2173, "loss": 2.1438, "learning_rate": 4.766118836915297e-07, "epoch": 2.6806411837237976 }, { "current_steps": 2174, "loss": 1.7313, "learning_rate": 4.763590391908976e-07, "epoch": 2.681874229346486 }, { "current_steps": 2175, "loss": 2.1165, "learning_rate": 4.7610619469026547e-07, "epoch": 2.683107274969174 }, { "current_steps": 2176, "loss": 2.267, "learning_rate": 4.7585335018963336e-07, "epoch": 2.684340320591862 }, { "current_steps": 2177, "loss": 1.9676, "learning_rate": 4.7560050568900125e-07, "epoch": 2.68557336621455 }, { "current_steps": 2178, "loss": 2.1107, "learning_rate": 4.7534766118836913e-07, "epoch": 2.686806411837238 }, { "current_steps": 2179, "loss": 2.0872, "learning_rate": 4.75094816687737e-07, "epoch": 2.688039457459926 }, { "current_steps": 2180, "loss": 2.2047, "learning_rate": 4.748419721871049e-07, "epoch": 2.689272503082614 }, { "current_steps": 2181, "loss": 1.2049, "learning_rate": 4.745891276864728e-07, "epoch": 2.690505548705302 }, { "current_steps": 2182, "loss": 2.2073, "learning_rate": 4.743362831858407e-07, "epoch": 2.6917385943279903 }, { "current_steps": 2183, "loss": 2.215, "learning_rate": 4.7408343868520857e-07, "epoch": 2.6929716399506782 }, { "current_steps": 2184, "loss": 2.368, "learning_rate": 4.7383059418457646e-07, "epoch": 2.694204685573366 }, { "current_steps": 2185, "loss": 1.6362, "learning_rate": 4.7357774968394435e-07, "epoch": 2.695437731196054 }, { "current_steps": 2186, "loss": 1.9599, "learning_rate": 4.7332490518331223e-07, "epoch": 2.6966707768187423 }, { "current_steps": 2187, "loss": 2.0759, "learning_rate": 4.730720606826801e-07, "epoch": 2.6979038224414302 }, { "current_steps": 2188, "loss": 2.2861, "learning_rate": 4.72819216182048e-07, "epoch": 2.6991368680641186 }, { "current_steps": 2189, "loss": 1.9634, "learning_rate": 4.725663716814159e-07, "epoch": 2.7003699136868065 }, { "current_steps": 2190, "loss": 2.0047, "learning_rate": 4.7231352718078384e-07, "epoch": 2.7016029593094943 }, { "current_steps": 2191, "loss": 1.8407, "learning_rate": 4.720606826801517e-07, "epoch": 2.7028360049321822 }, { "current_steps": 2192, "loss": 1.8714, "learning_rate": 4.7180783817951956e-07, "epoch": 2.7040690505548706 }, { "current_steps": 2193, "loss": 1.841, "learning_rate": 4.7155499367888745e-07, "epoch": 2.7053020961775585 }, { "current_steps": 2194, "loss": 1.9716, "learning_rate": 4.7130214917825534e-07, "epoch": 2.706535141800247 }, { "current_steps": 2195, "loss": 2.2142, "learning_rate": 4.710493046776232e-07, "epoch": 2.7077681874229347 }, { "current_steps": 2196, "loss": 2.4132, "learning_rate": 4.7079646017699116e-07, "epoch": 2.7090012330456226 }, { "current_steps": 2197, "loss": 2.6463, "learning_rate": 4.7054361567635905e-07, "epoch": 2.7102342786683105 }, { "current_steps": 2198, "loss": 1.9007, "learning_rate": 4.702907711757269e-07, "epoch": 2.711467324290999 }, { "current_steps": 2199, "loss": 2.1146, "learning_rate": 4.700379266750948e-07, "epoch": 2.7127003699136867 }, { "current_steps": 2200, "loss": 1.9449, "learning_rate": 4.6978508217446266e-07, "epoch": 2.713933415536375 }, { "current_steps": 2201, "loss": 2.2999, "learning_rate": 4.695322376738306e-07, "epoch": 2.715166461159063 }, { "current_steps": 2202, "loss": 1.3104, "learning_rate": 4.692793931731985e-07, "epoch": 2.716399506781751 }, { "current_steps": 2203, "loss": 2.1832, "learning_rate": 4.690265486725664e-07, "epoch": 2.717632552404439 }, { "current_steps": 2204, "loss": 2.2937, "learning_rate": 4.687737041719342e-07, "epoch": 2.718865598027127 }, { "current_steps": 2205, "loss": 1.8232, "learning_rate": 4.685208596713021e-07, "epoch": 2.720098643649815 }, { "current_steps": 2206, "loss": 1.9086, "learning_rate": 4.6826801517067e-07, "epoch": 2.7213316892725032 }, { "current_steps": 2207, "loss": 2.1116, "learning_rate": 4.6801517067003793e-07, "epoch": 2.722564734895191 }, { "current_steps": 2208, "loss": 2.3043, "learning_rate": 4.677623261694058e-07, "epoch": 2.723797780517879 }, { "current_steps": 2209, "loss": 1.5918, "learning_rate": 4.675094816687737e-07, "epoch": 2.7250308261405674 }, { "current_steps": 2210, "loss": 2.0677, "learning_rate": 4.6725663716814154e-07, "epoch": 2.7262638717632552 }, { "current_steps": 2211, "loss": 1.9205, "learning_rate": 4.6700379266750943e-07, "epoch": 2.727496917385943 }, { "current_steps": 2212, "loss": 2.3364, "learning_rate": 4.667509481668773e-07, "epoch": 2.7287299630086315 }, { "current_steps": 2213, "loss": 2.2026, "learning_rate": 4.6649810366624525e-07, "epoch": 2.7299630086313194 }, { "current_steps": 2214, "loss": 1.9101, "learning_rate": 4.6624525916561314e-07, "epoch": 2.7311960542540072 }, { "current_steps": 2215, "loss": 2.2794, "learning_rate": 4.6599241466498103e-07, "epoch": 2.7324290998766956 }, { "current_steps": 2216, "loss": 1.8362, "learning_rate": 4.657395701643489e-07, "epoch": 2.7336621454993835 }, { "current_steps": 2217, "loss": 1.9773, "learning_rate": 4.6548672566371675e-07, "epoch": 2.7348951911220714 }, { "current_steps": 2218, "loss": 2.0857, "learning_rate": 4.652338811630847e-07, "epoch": 2.7361282367447597 }, { "current_steps": 2219, "loss": 2.2136, "learning_rate": 4.649810366624526e-07, "epoch": 2.7373612823674476 }, { "current_steps": 2220, "loss": 2.2202, "learning_rate": 4.6472819216182047e-07, "epoch": 2.7385943279901355 }, { "current_steps": 2221, "loss": 2.0668, "learning_rate": 4.6447534766118836e-07, "epoch": 2.739827373612824 }, { "current_steps": 2222, "loss": 2.1929, "learning_rate": 4.6422250316055624e-07, "epoch": 2.7410604192355117 }, { "current_steps": 2223, "loss": 1.703, "learning_rate": 4.639696586599241e-07, "epoch": 2.7422934648582 }, { "current_steps": 2224, "loss": 2.1709, "learning_rate": 4.63716814159292e-07, "epoch": 2.743526510480888 }, { "current_steps": 2225, "loss": 1.8694, "learning_rate": 4.634639696586599e-07, "epoch": 2.744759556103576 }, { "current_steps": 2226, "loss": 1.7752, "learning_rate": 4.632111251580278e-07, "epoch": 2.7459926017262637 }, { "current_steps": 2227, "loss": 2.164, "learning_rate": 4.629582806573957e-07, "epoch": 2.747225647348952 }, { "current_steps": 2228, "loss": 1.7009, "learning_rate": 4.6270543615676357e-07, "epoch": 2.74845869297164 }, { "current_steps": 2229, "loss": 2.2615, "learning_rate": 4.624525916561315e-07, "epoch": 2.7496917385943282 }, { "current_steps": 2230, "loss": 2.3253, "learning_rate": 4.6219974715549935e-07, "epoch": 2.750924784217016 }, { "current_steps": 2231, "loss": 1.8501, "learning_rate": 4.6194690265486723e-07, "epoch": 2.752157829839704 }, { "current_steps": 2232, "loss": 2.0748, "learning_rate": 4.616940581542351e-07, "epoch": 2.753390875462392 }, { "current_steps": 2233, "loss": 2.2822, "learning_rate": 4.61441213653603e-07, "epoch": 2.7546239210850803 }, { "current_steps": 2234, "loss": 2.6514, "learning_rate": 4.611883691529709e-07, "epoch": 2.755856966707768 }, { "current_steps": 2235, "loss": 2.2536, "learning_rate": 4.6093552465233884e-07, "epoch": 2.7570900123304565 }, { "current_steps": 2236, "loss": 2.6178, "learning_rate": 4.6068268015170667e-07, "epoch": 2.7583230579531444 }, { "current_steps": 2237, "loss": 1.9885, "learning_rate": 4.6042983565107456e-07, "epoch": 2.7595561035758323 }, { "current_steps": 2238, "loss": 2.0204, "learning_rate": 4.6017699115044245e-07, "epoch": 2.76078914919852 }, { "current_steps": 2239, "loss": 2.1594, "learning_rate": 4.5992414664981033e-07, "epoch": 2.7620221948212085 }, { "current_steps": 2240, "loss": 2.4342, "learning_rate": 4.596713021491782e-07, "epoch": 2.7632552404438964 }, { "current_steps": 2241, "loss": 2.4409, "learning_rate": 4.5941845764854616e-07, "epoch": 2.7644882860665847 }, { "current_steps": 2242, "loss": 2.1353, "learning_rate": 4.59165613147914e-07, "epoch": 2.7657213316892726 }, { "current_steps": 2243, "loss": 2.2837, "learning_rate": 4.589127686472819e-07, "epoch": 2.7669543773119605 }, { "current_steps": 2244, "loss": 2.1392, "learning_rate": 4.5865992414664977e-07, "epoch": 2.7681874229346484 }, { "current_steps": 2245, "loss": 2.2659, "learning_rate": 4.5840707964601766e-07, "epoch": 2.7694204685573367 }, { "current_steps": 2246, "loss": 2.2951, "learning_rate": 4.581542351453856e-07, "epoch": 2.7706535141800246 }, { "current_steps": 2247, "loss": 2.1492, "learning_rate": 4.579013906447535e-07, "epoch": 2.771886559802713 }, { "current_steps": 2248, "loss": 2.3296, "learning_rate": 4.576485461441213e-07, "epoch": 2.773119605425401 }, { "current_steps": 2249, "loss": 1.7481, "learning_rate": 4.573957016434892e-07, "epoch": 2.7743526510480887 }, { "current_steps": 2250, "loss": 2.1737, "learning_rate": 4.571428571428571e-07, "epoch": 2.7755856966707766 }, { "current_steps": 2251, "loss": 1.7522, "learning_rate": 4.56890012642225e-07, "epoch": 2.776818742293465 }, { "current_steps": 2252, "loss": 1.5661, "learning_rate": 4.5663716814159293e-07, "epoch": 2.778051787916153 }, { "current_steps": 2253, "loss": 2.5244, "learning_rate": 4.563843236409608e-07, "epoch": 2.779284833538841 }, { "current_steps": 2254, "loss": 2.1241, "learning_rate": 4.561314791403287e-07, "epoch": 2.780517879161529 }, { "current_steps": 2255, "loss": 1.6145, "learning_rate": 4.5587863463969654e-07, "epoch": 2.781750924784217 }, { "current_steps": 2256, "loss": 1.7959, "learning_rate": 4.556257901390644e-07, "epoch": 2.782983970406905 }, { "current_steps": 2257, "loss": 1.537, "learning_rate": 4.5537294563843237e-07, "epoch": 2.784217016029593 }, { "current_steps": 2258, "loss": 2.6259, "learning_rate": 4.5512010113780025e-07, "epoch": 2.785450061652281 }, { "current_steps": 2259, "loss": 2.2873, "learning_rate": 4.5486725663716814e-07, "epoch": 2.7866831072749694 }, { "current_steps": 2260, "loss": 2.1238, "learning_rate": 4.5461441213653603e-07, "epoch": 2.7879161528976573 }, { "current_steps": 2261, "loss": 2.5246, "learning_rate": 4.5436156763590386e-07, "epoch": 2.789149198520345 }, { "current_steps": 2262, "loss": 1.935, "learning_rate": 4.5410872313527175e-07, "epoch": 2.790382244143033 }, { "current_steps": 2263, "loss": 2.5791, "learning_rate": 4.538558786346397e-07, "epoch": 2.7916152897657214 }, { "current_steps": 2264, "loss": 2.3878, "learning_rate": 4.536030341340076e-07, "epoch": 2.7928483353884093 }, { "current_steps": 2265, "loss": 1.937, "learning_rate": 4.5335018963337547e-07, "epoch": 2.7940813810110976 }, { "current_steps": 2266, "loss": 2.0522, "learning_rate": 4.5309734513274336e-07, "epoch": 2.7953144266337855 }, { "current_steps": 2267, "loss": 2.4398, "learning_rate": 4.528445006321112e-07, "epoch": 2.7965474722564734 }, { "current_steps": 2268, "loss": 2.2414, "learning_rate": 4.525916561314791e-07, "epoch": 2.7977805178791613 }, { "current_steps": 2269, "loss": 1.863, "learning_rate": 4.52338811630847e-07, "epoch": 2.7990135635018496 }, { "current_steps": 2270, "loss": 1.7516, "learning_rate": 4.520859671302149e-07, "epoch": 2.8002466091245375 }, { "current_steps": 2271, "loss": 2.554, "learning_rate": 4.518331226295828e-07, "epoch": 2.801479654747226 }, { "current_steps": 2272, "loss": 1.8325, "learning_rate": 4.515802781289507e-07, "epoch": 2.8027127003699137 }, { "current_steps": 2273, "loss": 2.3295, "learning_rate": 4.5132743362831857e-07, "epoch": 2.8039457459926016 }, { "current_steps": 2274, "loss": 1.9746, "learning_rate": 4.5107458912768646e-07, "epoch": 2.8051787916152895 }, { "current_steps": 2275, "loss": 2.2623, "learning_rate": 4.5082174462705434e-07, "epoch": 2.806411837237978 }, { "current_steps": 2276, "loss": 1.4222, "learning_rate": 4.5056890012642223e-07, "epoch": 2.8076448828606657 }, { "current_steps": 2277, "loss": 2.2351, "learning_rate": 4.503160556257901e-07, "epoch": 2.808877928483354 }, { "current_steps": 2278, "loss": 2.6396, "learning_rate": 4.50063211125158e-07, "epoch": 2.810110974106042 }, { "current_steps": 2279, "loss": 1.8185, "learning_rate": 4.498103666245259e-07, "epoch": 2.81134401972873 }, { "current_steps": 2280, "loss": 2.4894, "learning_rate": 4.495575221238938e-07, "epoch": 2.812577065351418 }, { "current_steps": 2281, "loss": 2.5051, "learning_rate": 4.4930467762326167e-07, "epoch": 2.813810110974106 }, { "current_steps": 2282, "loss": 1.8727, "learning_rate": 4.4905183312262956e-07, "epoch": 2.815043156596794 }, { "current_steps": 2283, "loss": 2.1133, "learning_rate": 4.4879898862199745e-07, "epoch": 2.8162762022194823 }, { "current_steps": 2284, "loss": 1.8387, "learning_rate": 4.4854614412136533e-07, "epoch": 2.81750924784217 }, { "current_steps": 2285, "loss": 2.2285, "learning_rate": 4.482932996207333e-07, "epoch": 2.818742293464858 }, { "current_steps": 2286, "loss": 1.9309, "learning_rate": 4.480404551201011e-07, "epoch": 2.8199753390875464 }, { "current_steps": 2287, "loss": 2.0625, "learning_rate": 4.47787610619469e-07, "epoch": 2.8212083847102343 }, { "current_steps": 2288, "loss": 2.011, "learning_rate": 4.475347661188369e-07, "epoch": 2.822441430332922 }, { "current_steps": 2289, "loss": 1.7213, "learning_rate": 4.4728192161820477e-07, "epoch": 2.8236744759556105 }, { "current_steps": 2290, "loss": 1.7146, "learning_rate": 4.4702907711757266e-07, "epoch": 2.8249075215782984 }, { "current_steps": 2291, "loss": 2.0273, "learning_rate": 4.467762326169406e-07, "epoch": 2.8261405672009863 }, { "current_steps": 2292, "loss": 2.0784, "learning_rate": 4.465233881163085e-07, "epoch": 2.8273736128236746 }, { "current_steps": 2293, "loss": 2.0489, "learning_rate": 4.462705436156763e-07, "epoch": 2.8286066584463625 }, { "current_steps": 2294, "loss": 1.9406, "learning_rate": 4.460176991150442e-07, "epoch": 2.829839704069051 }, { "current_steps": 2295, "loss": 2.3647, "learning_rate": 4.457648546144121e-07, "epoch": 2.8310727496917387 }, { "current_steps": 2296, "loss": 2.2318, "learning_rate": 4.4551201011378e-07, "epoch": 2.8323057953144266 }, { "current_steps": 2297, "loss": 1.5348, "learning_rate": 4.4525916561314793e-07, "epoch": 2.8335388409371145 }, { "current_steps": 2298, "loss": 2.2229, "learning_rate": 4.450063211125158e-07, "epoch": 2.834771886559803 }, { "current_steps": 2299, "loss": 2.5568, "learning_rate": 4.4475347661188365e-07, "epoch": 2.8360049321824907 }, { "current_steps": 2300, "loss": 2.3941, "learning_rate": 4.4450063211125154e-07, "epoch": 2.837237977805179 }, { "current_steps": 2301, "loss": 2.1819, "learning_rate": 4.442477876106194e-07, "epoch": 2.838471023427867 }, { "current_steps": 2302, "loss": 2.0931, "learning_rate": 4.4399494310998737e-07, "epoch": 2.839704069050555 }, { "current_steps": 2303, "loss": 1.7282, "learning_rate": 4.4374209860935525e-07, "epoch": 2.8409371146732427 }, { "current_steps": 2304, "loss": 2.0147, "learning_rate": 4.4348925410872314e-07, "epoch": 2.842170160295931 }, { "current_steps": 2305, "loss": 1.6026, "learning_rate": 4.43236409608091e-07, "epoch": 2.843403205918619 }, { "current_steps": 2306, "loss": 2.2512, "learning_rate": 4.4298356510745886e-07, "epoch": 2.8446362515413073 }, { "current_steps": 2307, "loss": 2.243, "learning_rate": 4.4273072060682675e-07, "epoch": 2.845869297163995 }, { "current_steps": 2308, "loss": 2.1948, "learning_rate": 4.424778761061947e-07, "epoch": 2.847102342786683 }, { "current_steps": 2309, "loss": 2.1902, "learning_rate": 4.422250316055626e-07, "epoch": 2.848335388409371 }, { "current_steps": 2310, "loss": 2.2584, "learning_rate": 4.4197218710493047e-07, "epoch": 2.8495684340320593 }, { "current_steps": 2311, "loss": 2.0812, "learning_rate": 4.4171934260429835e-07, "epoch": 2.850801479654747 }, { "current_steps": 2312, "loss": 2.0427, "learning_rate": 4.414664981036662e-07, "epoch": 2.8520345252774355 }, { "current_steps": 2313, "loss": 1.63, "learning_rate": 4.4121365360303413e-07, "epoch": 2.8532675709001234 }, { "current_steps": 2314, "loss": 1.9677, "learning_rate": 4.40960809102402e-07, "epoch": 2.8545006165228113 }, { "current_steps": 2315, "loss": 2.507, "learning_rate": 4.407079646017699e-07, "epoch": 2.855733662145499 }, { "current_steps": 2316, "loss": 2.2913, "learning_rate": 4.404551201011378e-07, "epoch": 2.8569667077681875 }, { "current_steps": 2317, "loss": 2.3089, "learning_rate": 4.402022756005057e-07, "epoch": 2.8581997533908754 }, { "current_steps": 2318, "loss": 1.8636, "learning_rate": 4.399494310998735e-07, "epoch": 2.8594327990135637 }, { "current_steps": 2319, "loss": 1.8914, "learning_rate": 4.3969658659924146e-07, "epoch": 2.8606658446362516 }, { "current_steps": 2320, "loss": 1.8761, "learning_rate": 4.3944374209860934e-07, "epoch": 2.8618988902589395 }, { "current_steps": 2321, "loss": 2.2116, "learning_rate": 4.3919089759797723e-07, "epoch": 2.8631319358816274 }, { "current_steps": 2322, "loss": 2.5078, "learning_rate": 4.389380530973451e-07, "epoch": 2.8643649815043157 }, { "current_steps": 2323, "loss": 1.8345, "learning_rate": 4.38685208596713e-07, "epoch": 2.8655980271270036 }, { "current_steps": 2324, "loss": 1.8859, "learning_rate": 4.3843236409608084e-07, "epoch": 2.866831072749692 }, { "current_steps": 2325, "loss": 2.2963, "learning_rate": 4.381795195954488e-07, "epoch": 2.86806411837238 }, { "current_steps": 2326, "loss": 2.057, "learning_rate": 4.3792667509481667e-07, "epoch": 2.8692971639950677 }, { "current_steps": 2327, "loss": 2.0552, "learning_rate": 4.3767383059418456e-07, "epoch": 2.8705302096177556 }, { "current_steps": 2328, "loss": 2.4635, "learning_rate": 4.3742098609355245e-07, "epoch": 2.871763255240444 }, { "current_steps": 2329, "loss": 1.6813, "learning_rate": 4.3716814159292033e-07, "epoch": 2.872996300863132 }, { "current_steps": 2330, "loss": 2.0492, "learning_rate": 4.3691529709228827e-07, "epoch": 2.87422934648582 }, { "current_steps": 2331, "loss": 2.1219, "learning_rate": 4.366624525916561e-07, "epoch": 2.875462392108508 }, { "current_steps": 2332, "loss": 2.2678, "learning_rate": 4.36409608091024e-07, "epoch": 2.876695437731196 }, { "current_steps": 2333, "loss": 2.0747, "learning_rate": 4.361567635903919e-07, "epoch": 2.877928483353884 }, { "current_steps": 2334, "loss": 2.0989, "learning_rate": 4.3590391908975977e-07, "epoch": 2.879161528976572 }, { "current_steps": 2335, "loss": 1.9494, "learning_rate": 4.3565107458912766e-07, "epoch": 2.88039457459926 }, { "current_steps": 2336, "loss": 2.0613, "learning_rate": 4.353982300884956e-07, "epoch": 2.8816276202219484 }, { "current_steps": 2337, "loss": 1.7509, "learning_rate": 4.3514538558786343e-07, "epoch": 2.8828606658446363 }, { "current_steps": 2338, "loss": 1.8856, "learning_rate": 4.348925410872313e-07, "epoch": 2.884093711467324 }, { "current_steps": 2339, "loss": 1.4607, "learning_rate": 4.346396965865992e-07, "epoch": 2.885326757090012 }, { "current_steps": 2340, "loss": 2.1897, "learning_rate": 4.343868520859671e-07, "epoch": 2.8865598027127004 }, { "current_steps": 2341, "loss": 2.1911, "learning_rate": 4.3413400758533504e-07, "epoch": 2.8877928483353883 }, { "current_steps": 2342, "loss": 1.8828, "learning_rate": 4.338811630847029e-07, "epoch": 2.8890258939580766 }, { "current_steps": 2343, "loss": 2.529, "learning_rate": 4.3362831858407076e-07, "epoch": 2.8902589395807645 }, { "current_steps": 2344, "loss": 2.0428, "learning_rate": 4.3337547408343865e-07, "epoch": 2.8914919852034524 }, { "current_steps": 2345, "loss": 2.3195, "learning_rate": 4.3312262958280654e-07, "epoch": 2.8927250308261403 }, { "current_steps": 2346, "loss": 2.498, "learning_rate": 4.328697850821744e-07, "epoch": 2.8939580764488286 }, { "current_steps": 2347, "loss": 2.2121, "learning_rate": 4.3261694058154236e-07, "epoch": 2.8951911220715165 }, { "current_steps": 2348, "loss": 2.2603, "learning_rate": 4.3236409608091025e-07, "epoch": 2.896424167694205 }, { "current_steps": 2349, "loss": 2.4271, "learning_rate": 4.3211125158027814e-07, "epoch": 2.8976572133168927 }, { "current_steps": 2350, "loss": 1.6938, "learning_rate": 4.31858407079646e-07, "epoch": 2.8988902589395806 }, { "current_steps": 2351, "loss": 1.845, "learning_rate": 4.3160556257901386e-07, "epoch": 2.900123304562269 }, { "current_steps": 2352, "loss": 2.0265, "learning_rate": 4.3135271807838175e-07, "epoch": 2.901356350184957 }, { "current_steps": 2353, "loss": 2.1483, "learning_rate": 4.310998735777497e-07, "epoch": 2.9025893958076447 }, { "current_steps": 2354, "loss": 2.1743, "learning_rate": 4.308470290771176e-07, "epoch": 2.903822441430333 }, { "current_steps": 2355, "loss": 1.9097, "learning_rate": 4.3059418457648547e-07, "epoch": 2.905055487053021 }, { "current_steps": 2356, "loss": 2.2422, "learning_rate": 4.303413400758533e-07, "epoch": 2.906288532675709 }, { "current_steps": 2357, "loss": 2.045, "learning_rate": 4.300884955752212e-07, "epoch": 2.907521578298397 }, { "current_steps": 2358, "loss": 1.9179, "learning_rate": 4.2983565107458913e-07, "epoch": 2.908754623921085 }, { "current_steps": 2359, "loss": 2.4533, "learning_rate": 4.29582806573957e-07, "epoch": 2.909987669543773 }, { "current_steps": 2360, "loss": 2.2991, "learning_rate": 4.293299620733249e-07, "epoch": 2.9112207151664613 }, { "current_steps": 2361, "loss": 2.1191, "learning_rate": 4.290771175726928e-07, "epoch": 2.912453760789149 }, { "current_steps": 2362, "loss": 1.9996, "learning_rate": 4.2882427307206063e-07, "epoch": 2.913686806411837 }, { "current_steps": 2363, "loss": 2.0175, "learning_rate": 4.285714285714285e-07, "epoch": 2.9149198520345254 }, { "current_steps": 2364, "loss": 1.83, "learning_rate": 4.2831858407079645e-07, "epoch": 2.9161528976572133 }, { "current_steps": 2365, "loss": 2.132, "learning_rate": 4.2806573957016434e-07, "epoch": 2.917385943279901 }, { "current_steps": 2366, "loss": 2.0489, "learning_rate": 4.2781289506953223e-07, "epoch": 2.9186189889025895 }, { "current_steps": 2367, "loss": 1.6797, "learning_rate": 4.275600505689001e-07, "epoch": 2.9198520345252774 }, { "current_steps": 2368, "loss": 1.7422, "learning_rate": 4.2730720606826795e-07, "epoch": 2.9210850801479653 }, { "current_steps": 2369, "loss": 2.3271, "learning_rate": 4.2705436156763584e-07, "epoch": 2.9223181257706536 }, { "current_steps": 2370, "loss": 1.8297, "learning_rate": 4.268015170670038e-07, "epoch": 2.9235511713933415 }, { "current_steps": 2371, "loss": 1.8977, "learning_rate": 4.2654867256637167e-07, "epoch": 2.92478421701603 }, { "current_steps": 2372, "loss": 2.0385, "learning_rate": 4.2629582806573956e-07, "epoch": 2.9260172626387178 }, { "current_steps": 2373, "loss": 1.5209, "learning_rate": 4.2604298356510744e-07, "epoch": 2.9272503082614056 }, { "current_steps": 2374, "loss": 2.361, "learning_rate": 4.2579013906447533e-07, "epoch": 2.9284833538840935 }, { "current_steps": 2375, "loss": 1.8, "learning_rate": 4.255372945638432e-07, "epoch": 2.929716399506782 }, { "current_steps": 2376, "loss": 2.0037, "learning_rate": 4.252844500632111e-07, "epoch": 2.9309494451294698 }, { "current_steps": 2377, "loss": 2.0809, "learning_rate": 4.25031605562579e-07, "epoch": 2.932182490752158 }, { "current_steps": 2378, "loss": 2.3595, "learning_rate": 4.247787610619469e-07, "epoch": 2.933415536374846 }, { "current_steps": 2379, "loss": 1.8354, "learning_rate": 4.2452591656131477e-07, "epoch": 2.934648581997534 }, { "current_steps": 2380, "loss": 1.7115, "learning_rate": 4.2427307206068266e-07, "epoch": 2.9358816276202218 }, { "current_steps": 2381, "loss": 1.9443, "learning_rate": 4.2402022756005055e-07, "epoch": 2.93711467324291 }, { "current_steps": 2382, "loss": 2.1064, "learning_rate": 4.2376738305941843e-07, "epoch": 2.938347718865598 }, { "current_steps": 2383, "loss": 2.2029, "learning_rate": 4.235145385587863e-07, "epoch": 2.9395807644882863 }, { "current_steps": 2384, "loss": 1.851, "learning_rate": 4.232616940581542e-07, "epoch": 2.940813810110974 }, { "current_steps": 2385, "loss": 1.7687, "learning_rate": 4.230088495575221e-07, "epoch": 2.942046855733662 }, { "current_steps": 2386, "loss": 1.8122, "learning_rate": 4.2275600505689004e-07, "epoch": 2.94327990135635 }, { "current_steps": 2387, "loss": 2.11, "learning_rate": 4.2275600505689004e-07, "epoch": 2.9445129469790383 }, { "current_steps": 2388, "loss": 2.3288, "learning_rate": 4.225031605562579e-07, "epoch": 2.945745992601726 }, { "current_steps": 2389, "loss": 2.1005, "learning_rate": 4.2225031605562576e-07, "epoch": 2.9469790382244145 }, { "current_steps": 2390, "loss": 2.5912, "learning_rate": 4.2199747155499365e-07, "epoch": 2.9482120838471024 }, { "current_steps": 2391, "loss": 2.0295, "learning_rate": 4.2174462705436153e-07, "epoch": 2.9494451294697903 }, { "current_steps": 2392, "loss": 2.3918, "learning_rate": 4.214917825537294e-07, "epoch": 2.950678175092478 }, { "current_steps": 2393, "loss": 1.5616, "learning_rate": 4.2123893805309736e-07, "epoch": 2.9519112207151665 }, { "current_steps": 2394, "loss": 2.1236, "learning_rate": 4.2098609355246525e-07, "epoch": 2.9531442663378544 }, { "current_steps": 2395, "loss": 2.0464, "learning_rate": 4.207332490518331e-07, "epoch": 2.9543773119605428 }, { "current_steps": 2396, "loss": 2.102, "learning_rate": 4.2048040455120097e-07, "epoch": 2.9556103575832307 }, { "current_steps": 2397, "loss": 1.4132, "learning_rate": 4.2022756005056886e-07, "epoch": 2.9568434032059185 }, { "current_steps": 2398, "loss": 2.585, "learning_rate": 4.1997471554993675e-07, "epoch": 2.9580764488286064 }, { "current_steps": 2399, "loss": 1.9925, "learning_rate": 4.197218710493047e-07, "epoch": 2.9593094944512948 }, { "current_steps": 2400, "loss": 2.2081, "learning_rate": 4.194690265486726e-07, "epoch": 2.9605425400739827 }, { "current_steps": 2401, "loss": 1.5875, "learning_rate": 4.192161820480404e-07, "epoch": 2.961775585696671 }, { "current_steps": 2402, "loss": 1.6121, "learning_rate": 4.189633375474083e-07, "epoch": 2.963008631319359 }, { "current_steps": 2403, "loss": 2.0916, "learning_rate": 4.187104930467762e-07, "epoch": 2.9642416769420468 }, { "current_steps": 2404, "loss": 2.3755, "learning_rate": 4.1845764854614413e-07, "epoch": 2.9654747225647347 }, { "current_steps": 2405, "loss": 1.9182, "learning_rate": 4.18204804045512e-07, "epoch": 2.966707768187423 }, { "current_steps": 2406, "loss": 1.4949, "learning_rate": 4.179519595448799e-07, "epoch": 2.967940813810111 }, { "current_steps": 2407, "loss": 2.0304, "learning_rate": 4.1769911504424774e-07, "epoch": 2.969173859432799 }, { "current_steps": 2408, "loss": 2.4125, "learning_rate": 4.174462705436156e-07, "epoch": 2.970406905055487 }, { "current_steps": 2409, "loss": 2.124, "learning_rate": 4.171934260429835e-07, "epoch": 2.971639950678175 }, { "current_steps": 2410, "loss": 2.044, "learning_rate": 4.1694058154235145e-07, "epoch": 2.972872996300863 }, { "current_steps": 2411, "loss": 1.6673, "learning_rate": 4.1668773704171934e-07, "epoch": 2.974106041923551 }, { "current_steps": 2412, "loss": 1.5893, "learning_rate": 4.1643489254108723e-07, "epoch": 2.975339087546239 }, { "current_steps": 2413, "loss": 1.9926, "learning_rate": 4.161820480404551e-07, "epoch": 2.9765721331689274 }, { "current_steps": 2414, "loss": 1.8306, "learning_rate": 4.1592920353982295e-07, "epoch": 2.9778051787916153 }, { "current_steps": 2415, "loss": 2.2129, "learning_rate": 4.156763590391909e-07, "epoch": 2.979038224414303 }, { "current_steps": 2416, "loss": 1.7222, "learning_rate": 4.154235145385588e-07, "epoch": 2.980271270036991 }, { "current_steps": 2417, "loss": 1.9493, "learning_rate": 4.1517067003792667e-07, "epoch": 2.9815043156596794 }, { "current_steps": 2418, "loss": 2.5336, "learning_rate": 4.1491782553729456e-07, "epoch": 2.9827373612823673 }, { "current_steps": 2419, "loss": 1.7284, "learning_rate": 4.1466498103666244e-07, "epoch": 2.9839704069050557 }, { "current_steps": 2420, "loss": 2.4457, "learning_rate": 4.144121365360303e-07, "epoch": 2.9852034525277436 }, { "current_steps": 2421, "loss": 1.746, "learning_rate": 4.141592920353982e-07, "epoch": 2.9864364981504314 }, { "current_steps": 2422, "loss": 2.0847, "learning_rate": 4.139064475347661e-07, "epoch": 2.9876695437731193 }, { "current_steps": 2423, "loss": 1.5963, "learning_rate": 4.13653603034134e-07, "epoch": 2.9889025893958077 }, { "current_steps": 2424, "loss": 2.4599, "learning_rate": 4.134007585335019e-07, "epoch": 2.9901356350184956 }, { "current_steps": 2425, "loss": 2.2751, "learning_rate": 4.1314791403286977e-07, "epoch": 2.991368680641184 }, { "current_steps": 2426, "loss": 2.266, "learning_rate": 4.128950695322376e-07, "epoch": 2.9926017262638718 }, { "current_steps": 2427, "loss": 1.9371, "learning_rate": 4.1264222503160554e-07, "epoch": 2.9938347718865597 }, { "current_steps": 2428, "loss": 2.3721, "learning_rate": 4.1238938053097343e-07, "epoch": 2.995067817509248 }, { "current_steps": 2429, "loss": 2.341, "learning_rate": 4.121365360303413e-07, "epoch": 2.996300863131936 }, { "current_steps": 2430, "loss": 1.9231, "learning_rate": 4.118836915297092e-07, "epoch": 2.9975339087546238 }, { "current_steps": 2431, "loss": 2.4054, "learning_rate": 4.116308470290771e-07, "epoch": 2.998766954377312 }, { "current_steps": 2432, "loss": 2.531, "learning_rate": 4.1137800252844504e-07, "epoch": 3.0 }, { "current_steps": 2433, "loss": 2.3898, "learning_rate": 4.1112515802781287e-07, "epoch": 3.001233045622688 }, { "current_steps": 2434, "loss": 1.3822, "learning_rate": 4.1087231352718076e-07, "epoch": 3.0024660912453762 }, { "current_steps": 2435, "loss": 1.8103, "learning_rate": 4.1061946902654865e-07, "epoch": 3.003699136868064 }, { "current_steps": 2436, "loss": 1.6588, "learning_rate": 4.1036662452591653e-07, "epoch": 3.004932182490752 }, { "current_steps": 2437, "loss": 2.1936, "learning_rate": 4.101137800252844e-07, "epoch": 3.0061652281134403 }, { "current_steps": 2438, "loss": 2.168, "learning_rate": 4.0986093552465236e-07, "epoch": 3.0073982737361282 }, { "current_steps": 2439, "loss": 2.3581, "learning_rate": 4.096080910240202e-07, "epoch": 3.008631319358816 }, { "current_steps": 2440, "loss": 1.8929, "learning_rate": 4.093552465233881e-07, "epoch": 3.0098643649815044 }, { "current_steps": 2441, "loss": 1.9125, "learning_rate": 4.0910240202275597e-07, "epoch": 3.0110974106041923 }, { "current_steps": 2442, "loss": 1.9894, "learning_rate": 4.0884955752212386e-07, "epoch": 3.0123304562268802 }, { "current_steps": 2443, "loss": 1.9749, "learning_rate": 4.085967130214918e-07, "epoch": 3.0135635018495686 }, { "current_steps": 2444, "loss": 2.245, "learning_rate": 4.083438685208597e-07, "epoch": 3.0147965474722564 }, { "current_steps": 2445, "loss": 2.0729, "learning_rate": 4.080910240202275e-07, "epoch": 3.0160295930949443 }, { "current_steps": 2446, "loss": 2.0391, "learning_rate": 4.078381795195954e-07, "epoch": 3.0172626387176327 }, { "current_steps": 2447, "loss": 1.7305, "learning_rate": 4.075853350189633e-07, "epoch": 3.0184956843403206 }, { "current_steps": 2448, "loss": 2.0588, "learning_rate": 4.073324905183312e-07, "epoch": 3.0197287299630085 }, { "current_steps": 2449, "loss": 1.8481, "learning_rate": 4.0707964601769913e-07, "epoch": 3.020961775585697 }, { "current_steps": 2450, "loss": 1.6975, "learning_rate": 4.06826801517067e-07, "epoch": 3.0221948212083847 }, { "current_steps": 2451, "loss": 1.7939, "learning_rate": 4.065739570164349e-07, "epoch": 3.0234278668310726 }, { "current_steps": 2452, "loss": 2.2706, "learning_rate": 4.0632111251580274e-07, "epoch": 3.024660912453761 }, { "current_steps": 2453, "loss": 1.9771, "learning_rate": 4.060682680151706e-07, "epoch": 3.025893958076449 }, { "current_steps": 2454, "loss": 1.6256, "learning_rate": 4.058154235145385e-07, "epoch": 3.0271270036991367 }, { "current_steps": 2455, "loss": 2.2996, "learning_rate": 4.0556257901390645e-07, "epoch": 3.028360049321825 }, { "current_steps": 2456, "loss": 1.5363, "learning_rate": 4.0530973451327434e-07, "epoch": 3.029593094944513 }, { "current_steps": 2457, "loss": 1.9975, "learning_rate": 4.0505689001264223e-07, "epoch": 3.030826140567201 }, { "current_steps": 2458, "loss": 2.3652, "learning_rate": 4.0480404551201006e-07, "epoch": 3.032059186189889 }, { "current_steps": 2459, "loss": 1.9246, "learning_rate": 4.0455120101137795e-07, "epoch": 3.033292231812577 }, { "current_steps": 2460, "loss": 2.3199, "learning_rate": 4.042983565107459e-07, "epoch": 3.034525277435265 }, { "current_steps": 2461, "loss": 1.7357, "learning_rate": 4.040455120101138e-07, "epoch": 3.0357583230579532 }, { "current_steps": 2462, "loss": 1.9248, "learning_rate": 4.0379266750948167e-07, "epoch": 3.036991368680641 }, { "current_steps": 2463, "loss": 1.8101, "learning_rate": 4.0353982300884955e-07, "epoch": 3.038224414303329 }, { "current_steps": 2464, "loss": 1.6185, "learning_rate": 4.032869785082174e-07, "epoch": 3.0394574599260173 }, { "current_steps": 2465, "loss": 1.6065, "learning_rate": 4.030341340075853e-07, "epoch": 3.0406905055487052 }, { "current_steps": 2466, "loss": 2.1541, "learning_rate": 4.027812895069532e-07, "epoch": 3.041923551171393 }, { "current_steps": 2467, "loss": 1.8333, "learning_rate": 4.025284450063211e-07, "epoch": 3.0431565967940815 }, { "current_steps": 2468, "loss": 1.8548, "learning_rate": 4.02275600505689e-07, "epoch": 3.0443896424167693 }, { "current_steps": 2469, "loss": 2.2488, "learning_rate": 4.020227560050569e-07, "epoch": 3.0456226880394572 }, { "current_steps": 2470, "loss": 2.1857, "learning_rate": 4.017699115044247e-07, "epoch": 3.0468557336621456 }, { "current_steps": 2471, "loss": 1.8444, "learning_rate": 4.0151706700379266e-07, "epoch": 3.0480887792848335 }, { "current_steps": 2472, "loss": 2.2152, "learning_rate": 4.0126422250316054e-07, "epoch": 3.049321824907522 }, { "current_steps": 2473, "loss": 2.1747, "learning_rate": 4.0101137800252843e-07, "epoch": 3.0505548705302097 }, { "current_steps": 2474, "loss": 2.2946, "learning_rate": 4.007585335018963e-07, "epoch": 3.0517879161528976 }, { "current_steps": 2475, "loss": 1.7921, "learning_rate": 4.005056890012642e-07, "epoch": 3.053020961775586 }, { "current_steps": 2476, "loss": 2.3719, "learning_rate": 4.002528445006321e-07, "epoch": 3.054254007398274 }, { "current_steps": 2477, "loss": 2.4014, "learning_rate": 4e-07, "epoch": 3.0554870530209617 }, { "current_steps": 2478, "loss": 1.6005, "learning_rate": 3.9974715549936787e-07, "epoch": 3.05672009864365 }, { "current_steps": 2479, "loss": 1.9854, "learning_rate": 3.9949431099873576e-07, "epoch": 3.057953144266338 }, { "current_steps": 2480, "loss": 2.2898, "learning_rate": 3.9924146649810364e-07, "epoch": 3.059186189889026 }, { "current_steps": 2481, "loss": 2.1399, "learning_rate": 3.9898862199747153e-07, "epoch": 3.060419235511714 }, { "current_steps": 2482, "loss": 1.8878, "learning_rate": 3.987357774968394e-07, "epoch": 3.061652281134402 }, { "current_steps": 2483, "loss": 2.234, "learning_rate": 3.984829329962073e-07, "epoch": 3.06288532675709 }, { "current_steps": 2484, "loss": 2.187, "learning_rate": 3.982300884955752e-07, "epoch": 3.0641183723797782 }, { "current_steps": 2485, "loss": 2.0257, "learning_rate": 3.979772439949431e-07, "epoch": 3.065351418002466 }, { "current_steps": 2486, "loss": 2.1148, "learning_rate": 3.9772439949431097e-07, "epoch": 3.066584463625154 }, { "current_steps": 2487, "loss": 2.0154, "learning_rate": 3.9747155499367886e-07, "epoch": 3.0678175092478424 }, { "current_steps": 2488, "loss": 2.2527, "learning_rate": 3.972187104930468e-07, "epoch": 3.0690505548705302 }, { "current_steps": 2489, "loss": 1.9876, "learning_rate": 3.969658659924147e-07, "epoch": 3.070283600493218 }, { "current_steps": 2490, "loss": 2.1879, "learning_rate": 3.967130214917825e-07, "epoch": 3.0715166461159065 }, { "current_steps": 2491, "loss": 1.7618, "learning_rate": 3.964601769911504e-07, "epoch": 3.0727496917385944 }, { "current_steps": 2492, "loss": 2.4232, "learning_rate": 3.962073324905183e-07, "epoch": 3.0739827373612822 }, { "current_steps": 2493, "loss": 1.6855, "learning_rate": 3.959544879898862e-07, "epoch": 3.0752157829839706 }, { "current_steps": 2494, "loss": 2.2655, "learning_rate": 3.957016434892541e-07, "epoch": 3.0764488286066585 }, { "current_steps": 2495, "loss": 2.3284, "learning_rate": 3.95448798988622e-07, "epoch": 3.0776818742293464 }, { "current_steps": 2496, "loss": 2.0105, "learning_rate": 3.9519595448798985e-07, "epoch": 3.0789149198520347 }, { "current_steps": 2497, "loss": 2.138, "learning_rate": 3.9494310998735774e-07, "epoch": 3.0801479654747226 }, { "current_steps": 2498, "loss": 1.4818, "learning_rate": 3.946902654867256e-07, "epoch": 3.0813810110974105 }, { "current_steps": 2499, "loss": 1.9127, "learning_rate": 3.9443742098609356e-07, "epoch": 3.082614056720099 }, { "current_steps": 2500, "loss": 2.2847, "learning_rate": 3.9418457648546145e-07, "epoch": 3.0838471023427867 }, { "current_steps": 2501, "loss": 1.7146, "learning_rate": 3.9393173198482934e-07, "epoch": 3.0850801479654746 }, { "current_steps": 2502, "loss": 1.9371, "learning_rate": 3.936788874841972e-07, "epoch": 3.086313193588163 }, { "current_steps": 2503, "loss": 2.1486, "learning_rate": 3.9342604298356506e-07, "epoch": 3.087546239210851 }, { "current_steps": 2504, "loss": 1.9673, "learning_rate": 3.9317319848293295e-07, "epoch": 3.0887792848335387 }, { "current_steps": 2505, "loss": 1.7853, "learning_rate": 3.929203539823009e-07, "epoch": 3.090012330456227 }, { "current_steps": 2506, "loss": 2.5347, "learning_rate": 3.926675094816688e-07, "epoch": 3.091245376078915 }, { "current_steps": 2507, "loss": 2.1412, "learning_rate": 3.9241466498103667e-07, "epoch": 3.092478421701603 }, { "current_steps": 2508, "loss": 2.1886, "learning_rate": 3.921618204804045e-07, "epoch": 3.093711467324291 }, { "current_steps": 2509, "loss": 2.2033, "learning_rate": 3.919089759797724e-07, "epoch": 3.094944512946979 }, { "current_steps": 2510, "loss": 1.7771, "learning_rate": 3.916561314791403e-07, "epoch": 3.096177558569667 }, { "current_steps": 2511, "loss": 1.3855, "learning_rate": 3.914032869785082e-07, "epoch": 3.0974106041923553 }, { "current_steps": 2512, "loss": 2.3947, "learning_rate": 3.911504424778761e-07, "epoch": 3.098643649815043 }, { "current_steps": 2513, "loss": 2.4167, "learning_rate": 3.90897597977244e-07, "epoch": 3.099876695437731 }, { "current_steps": 2514, "loss": 1.926, "learning_rate": 3.906447534766119e-07, "epoch": 3.1011097410604194 }, { "current_steps": 2515, "loss": 1.9478, "learning_rate": 3.903919089759797e-07, "epoch": 3.1023427866831073 }, { "current_steps": 2516, "loss": 1.8007, "learning_rate": 3.9013906447534765e-07, "epoch": 3.103575832305795 }, { "current_steps": 2517, "loss": 2.3264, "learning_rate": 3.8988621997471554e-07, "epoch": 3.1048088779284835 }, { "current_steps": 2518, "loss": 2.2026, "learning_rate": 3.8963337547408343e-07, "epoch": 3.1060419235511714 }, { "current_steps": 2519, "loss": 2.1907, "learning_rate": 3.893805309734513e-07, "epoch": 3.1072749691738593 }, { "current_steps": 2520, "loss": 2.331, "learning_rate": 3.891276864728192e-07, "epoch": 3.1085080147965476 }, { "current_steps": 2521, "loss": 2.0997, "learning_rate": 3.8887484197218704e-07, "epoch": 3.1097410604192355 }, { "current_steps": 2522, "loss": 0.9494, "learning_rate": 3.88621997471555e-07, "epoch": 3.1109741060419234 }, { "current_steps": 2523, "loss": 1.9708, "learning_rate": 3.8836915297092287e-07, "epoch": 3.1122071516646117 }, { "current_steps": 2524, "loss": 2.134, "learning_rate": 3.8811630847029076e-07, "epoch": 3.1134401972872996 }, { "current_steps": 2525, "loss": 2.3198, "learning_rate": 3.8786346396965864e-07, "epoch": 3.1146732429099875 }, { "current_steps": 2526, "loss": 2.2074, "learning_rate": 3.8761061946902653e-07, "epoch": 3.115906288532676 }, { "current_steps": 2527, "loss": 1.6931, "learning_rate": 3.8735777496839447e-07, "epoch": 3.1171393341553637 }, { "current_steps": 2528, "loss": 2.2027, "learning_rate": 3.871049304677623e-07, "epoch": 3.1183723797780516 }, { "current_steps": 2529, "loss": 1.8111, "learning_rate": 3.868520859671302e-07, "epoch": 3.11960542540074 }, { "current_steps": 2530, "loss": 1.7858, "learning_rate": 3.865992414664981e-07, "epoch": 3.120838471023428 }, { "current_steps": 2531, "loss": 1.7514, "learning_rate": 3.8634639696586597e-07, "epoch": 3.1220715166461157 }, { "current_steps": 2532, "loss": 2.1679, "learning_rate": 3.8609355246523386e-07, "epoch": 3.123304562268804 }, { "current_steps": 2533, "loss": 2.3571, "learning_rate": 3.858407079646018e-07, "epoch": 3.124537607891492 }, { "current_steps": 2534, "loss": 2.2306, "learning_rate": 3.8558786346396963e-07, "epoch": 3.12577065351418 }, { "current_steps": 2535, "loss": 2.0812, "learning_rate": 3.853350189633375e-07, "epoch": 3.127003699136868 }, { "current_steps": 2536, "loss": 1.9346, "learning_rate": 3.850821744627054e-07, "epoch": 3.128236744759556 }, { "current_steps": 2537, "loss": 1.995, "learning_rate": 3.848293299620733e-07, "epoch": 3.1294697903822444 }, { "current_steps": 2538, "loss": 1.9013, "learning_rate": 3.845764854614412e-07, "epoch": 3.1307028360049323 }, { "current_steps": 2539, "loss": 2.0693, "learning_rate": 3.843236409608091e-07, "epoch": 3.13193588162762 }, { "current_steps": 2540, "loss": 1.8798, "learning_rate": 3.8407079646017696e-07, "epoch": 3.133168927250308 }, { "current_steps": 2541, "loss": 1.6473, "learning_rate": 3.8381795195954485e-07, "epoch": 3.1344019728729964 }, { "current_steps": 2542, "loss": 2.0363, "learning_rate": 3.8356510745891273e-07, "epoch": 3.1356350184956843 }, { "current_steps": 2543, "loss": 1.8011, "learning_rate": 3.833122629582806e-07, "epoch": 3.1368680641183726 }, { "current_steps": 2544, "loss": 2.4341, "learning_rate": 3.8305941845764856e-07, "epoch": 3.1381011097410605 }, { "current_steps": 2545, "loss": 2.0036, "learning_rate": 3.8280657395701645e-07, "epoch": 3.1393341553637484 }, { "current_steps": 2546, "loss": 2.0839, "learning_rate": 3.825537294563843e-07, "epoch": 3.1405672009864363 }, { "current_steps": 2547, "loss": 1.8405, "learning_rate": 3.8230088495575217e-07, "epoch": 3.1418002466091246 }, { "current_steps": 2548, "loss": 2.1529, "learning_rate": 3.8204804045512006e-07, "epoch": 3.1430332922318125 }, { "current_steps": 2549, "loss": 2.6302, "learning_rate": 3.8179519595448795e-07, "epoch": 3.144266337854501 }, { "current_steps": 2550, "loss": 1.4303, "learning_rate": 3.815423514538559e-07, "epoch": 3.1454993834771887 }, { "current_steps": 2551, "loss": 2.1363, "learning_rate": 3.812895069532238e-07, "epoch": 3.1467324290998766 }, { "current_steps": 2552, "loss": 1.6894, "learning_rate": 3.8103666245259166e-07, "epoch": 3.147965474722565 }, { "current_steps": 2553, "loss": 1.1924, "learning_rate": 3.807838179519595e-07, "epoch": 3.149198520345253 }, { "current_steps": 2554, "loss": 1.6179, "learning_rate": 3.805309734513274e-07, "epoch": 3.1504315659679407 }, { "current_steps": 2555, "loss": 2.5681, "learning_rate": 3.8027812895069533e-07, "epoch": 3.151664611590629 }, { "current_steps": 2556, "loss": 2.1007, "learning_rate": 3.800252844500632e-07, "epoch": 3.152897657213317 }, { "current_steps": 2557, "loss": 2.083, "learning_rate": 3.797724399494311e-07, "epoch": 3.154130702836005 }, { "current_steps": 2558, "loss": 2.1853, "learning_rate": 3.79519595448799e-07, "epoch": 3.155363748458693 }, { "current_steps": 2559, "loss": 1.7985, "learning_rate": 3.792667509481668e-07, "epoch": 3.156596794081381 }, { "current_steps": 2560, "loss": 2.0366, "learning_rate": 3.790139064475347e-07, "epoch": 3.157829839704069 }, { "current_steps": 2561, "loss": 2.2602, "learning_rate": 3.7876106194690265e-07, "epoch": 3.1590628853267573 }, { "current_steps": 2562, "loss": 1.9379, "learning_rate": 3.7850821744627054e-07, "epoch": 3.160295930949445 }, { "current_steps": 2563, "loss": 1.6904, "learning_rate": 3.7825537294563843e-07, "epoch": 3.161528976572133 }, { "current_steps": 2564, "loss": 1.9894, "learning_rate": 3.780025284450063e-07, "epoch": 3.1627620221948214 }, { "current_steps": 2565, "loss": 2.1477, "learning_rate": 3.7774968394437415e-07, "epoch": 3.1639950678175093 }, { "current_steps": 2566, "loss": 2.1947, "learning_rate": 3.7749683944374204e-07, "epoch": 3.165228113440197 }, { "current_steps": 2567, "loss": 1.8692, "learning_rate": 3.7724399494311e-07, "epoch": 3.1664611590628855 }, { "current_steps": 2568, "loss": 2.2465, "learning_rate": 3.7699115044247787e-07, "epoch": 3.1676942046855734 }, { "current_steps": 2569, "loss": 2.5186, "learning_rate": 3.7673830594184576e-07, "epoch": 3.1689272503082613 }, { "current_steps": 2570, "loss": 2.3371, "learning_rate": 3.7648546144121364e-07, "epoch": 3.1701602959309496 }, { "current_steps": 2571, "loss": 2.1734, "learning_rate": 3.7623261694058153e-07, "epoch": 3.1713933415536375 }, { "current_steps": 2572, "loss": 1.6863, "learning_rate": 3.759797724399494e-07, "epoch": 3.1726263871763254 }, { "current_steps": 2573, "loss": 1.9306, "learning_rate": 3.757269279393173e-07, "epoch": 3.1738594327990137 }, { "current_steps": 2574, "loss": 2.0006, "learning_rate": 3.754740834386852e-07, "epoch": 3.1750924784217016 }, { "current_steps": 2575, "loss": 2.189, "learning_rate": 3.752212389380531e-07, "epoch": 3.1763255240443895 }, { "current_steps": 2576, "loss": 2.1088, "learning_rate": 3.7496839443742097e-07, "epoch": 3.177558569667078 }, { "current_steps": 2577, "loss": 2.1016, "learning_rate": 3.7471554993678886e-07, "epoch": 3.1787916152897657 }, { "current_steps": 2578, "loss": 1.916, "learning_rate": 3.7446270543615674e-07, "epoch": 3.1800246609124536 }, { "current_steps": 2579, "loss": 2.2051, "learning_rate": 3.7420986093552463e-07, "epoch": 3.181257706535142 }, { "current_steps": 2580, "loss": 1.4645, "learning_rate": 3.739570164348925e-07, "epoch": 3.18249075215783 }, { "current_steps": 2581, "loss": 2.1288, "learning_rate": 3.737041719342604e-07, "epoch": 3.1837237977805177 }, { "current_steps": 2582, "loss": 1.7162, "learning_rate": 3.734513274336283e-07, "epoch": 3.184956843403206 }, { "current_steps": 2583, "loss": 2.5001, "learning_rate": 3.7319848293299624e-07, "epoch": 3.186189889025894 }, { "current_steps": 2584, "loss": 1.571, "learning_rate": 3.7294563843236407e-07, "epoch": 3.187422934648582 }, { "current_steps": 2585, "loss": 1.3095, "learning_rate": 3.7269279393173196e-07, "epoch": 3.18865598027127 }, { "current_steps": 2586, "loss": 1.8883, "learning_rate": 3.7243994943109985e-07, "epoch": 3.189889025893958 }, { "current_steps": 2587, "loss": 1.7347, "learning_rate": 3.7218710493046773e-07, "epoch": 3.191122071516646 }, { "current_steps": 2588, "loss": 1.9175, "learning_rate": 3.719342604298356e-07, "epoch": 3.1923551171393343 }, { "current_steps": 2589, "loss": 2.3485, "learning_rate": 3.7168141592920356e-07, "epoch": 3.193588162762022 }, { "current_steps": 2590, "loss": 2.1563, "learning_rate": 3.7142857142857145e-07, "epoch": 3.19482120838471 }, { "current_steps": 2591, "loss": 1.6923, "learning_rate": 3.711757269279393e-07, "epoch": 3.1960542540073984 }, { "current_steps": 2592, "loss": 2.0028, "learning_rate": 3.7092288242730717e-07, "epoch": 3.1972872996300863 }, { "current_steps": 2593, "loss": 2.0861, "learning_rate": 3.7067003792667506e-07, "epoch": 3.198520345252774 }, { "current_steps": 2594, "loss": 2.2992, "learning_rate": 3.7041719342604295e-07, "epoch": 3.1997533908754625 }, { "current_steps": 2595, "loss": 1.6044, "learning_rate": 3.701643489254109e-07, "epoch": 3.2009864364981504 }, { "current_steps": 2596, "loss": 2.309, "learning_rate": 3.699115044247788e-07, "epoch": 3.2022194821208383 }, { "current_steps": 2597, "loss": 1.5248, "learning_rate": 3.696586599241466e-07, "epoch": 3.2034525277435266 }, { "current_steps": 2598, "loss": 2.3524, "learning_rate": 3.694058154235145e-07, "epoch": 3.2046855733662145 }, { "current_steps": 2599, "loss": 2.4713, "learning_rate": 3.691529709228824e-07, "epoch": 3.2059186189889024 }, { "current_steps": 2600, "loss": 2.1202, "learning_rate": 3.689001264222503e-07, "epoch": 3.2071516646115907 }, { "current_steps": 2601, "loss": 2.4172, "learning_rate": 3.686472819216182e-07, "epoch": 3.2083847102342786 }, { "current_steps": 2602, "loss": 2.0737, "learning_rate": 3.683944374209861e-07, "epoch": 3.2096177558569665 }, { "current_steps": 2603, "loss": 1.8385, "learning_rate": 3.6814159292035394e-07, "epoch": 3.210850801479655 }, { "current_steps": 2604, "loss": 2.3264, "learning_rate": 3.678887484197218e-07, "epoch": 3.2120838471023427 }, { "current_steps": 2605, "loss": 2.1842, "learning_rate": 3.676359039190897e-07, "epoch": 3.2133168927250306 }, { "current_steps": 2606, "loss": 1.6319, "learning_rate": 3.6738305941845765e-07, "epoch": 3.214549938347719 }, { "current_steps": 2607, "loss": 1.9883, "learning_rate": 3.6713021491782554e-07, "epoch": 3.215782983970407 }, { "current_steps": 2608, "loss": 2.0786, "learning_rate": 3.6687737041719343e-07, "epoch": 3.2170160295930947 }, { "current_steps": 2609, "loss": 1.9893, "learning_rate": 3.6662452591656126e-07, "epoch": 3.218249075215783 }, { "current_steps": 2610, "loss": 1.7846, "learning_rate": 3.6637168141592915e-07, "epoch": 3.219482120838471 }, { "current_steps": 2611, "loss": 1.9423, "learning_rate": 3.661188369152971e-07, "epoch": 3.220715166461159 }, { "current_steps": 2612, "loss": 2.2077, "learning_rate": 3.65865992414665e-07, "epoch": 3.221948212083847 }, { "current_steps": 2613, "loss": 1.8471, "learning_rate": 3.6561314791403287e-07, "epoch": 3.223181257706535 }, { "current_steps": 2614, "loss": 1.461, "learning_rate": 3.6536030341340075e-07, "epoch": 3.2244143033292234 }, { "current_steps": 2615, "loss": 1.6315, "learning_rate": 3.6510745891276864e-07, "epoch": 3.2256473489519113 }, { "current_steps": 2616, "loss": 2.3305, "learning_rate": 3.648546144121365e-07, "epoch": 3.226880394574599 }, { "current_steps": 2617, "loss": 2.5713, "learning_rate": 3.646017699115044e-07, "epoch": 3.228113440197287 }, { "current_steps": 2618, "loss": 1.6224, "learning_rate": 3.643489254108723e-07, "epoch": 3.2293464858199754 }, { "current_steps": 2619, "loss": 2.1505, "learning_rate": 3.640960809102402e-07, "epoch": 3.2305795314426633 }, { "current_steps": 2620, "loss": 2.5002, "learning_rate": 3.638432364096081e-07, "epoch": 3.2318125770653516 }, { "current_steps": 2621, "loss": 2.1236, "learning_rate": 3.6359039190897597e-07, "epoch": 3.2330456226880395 }, { "current_steps": 2622, "loss": 2.145, "learning_rate": 3.633375474083438e-07, "epoch": 3.2342786683107274 }, { "current_steps": 2623, "loss": 1.6588, "learning_rate": 3.6308470290771174e-07, "epoch": 3.2355117139334153 }, { "current_steps": 2624, "loss": 2.1584, "learning_rate": 3.6283185840707963e-07, "epoch": 3.2367447595561036 }, { "current_steps": 2625, "loss": 2.3558, "learning_rate": 3.625790139064475e-07, "epoch": 3.2379778051787915 }, { "current_steps": 2626, "loss": 2.1134, "learning_rate": 3.623261694058154e-07, "epoch": 3.23921085080148 }, { "current_steps": 2627, "loss": 2.2288, "learning_rate": 3.620733249051833e-07, "epoch": 3.2404438964241677 }, { "current_steps": 2628, "loss": 1.9057, "learning_rate": 3.6182048040455123e-07, "epoch": 3.2416769420468556 }, { "current_steps": 2629, "loss": 2.1898, "learning_rate": 3.6156763590391907e-07, "epoch": 3.242909987669544 }, { "current_steps": 2630, "loss": 2.4867, "learning_rate": 3.6131479140328696e-07, "epoch": 3.244143033292232 }, { "current_steps": 2631, "loss": 2.3743, "learning_rate": 3.6106194690265484e-07, "epoch": 3.2453760789149197 }, { "current_steps": 2632, "loss": 1.8414, "learning_rate": 3.6080910240202273e-07, "epoch": 3.246609124537608 }, { "current_steps": 2633, "loss": 1.421, "learning_rate": 3.605562579013906e-07, "epoch": 3.247842170160296 }, { "current_steps": 2634, "loss": 2.2463, "learning_rate": 3.6030341340075856e-07, "epoch": 3.249075215782984 }, { "current_steps": 2635, "loss": 2.2363, "learning_rate": 3.600505689001264e-07, "epoch": 3.250308261405672 }, { "current_steps": 2636, "loss": 1.9199, "learning_rate": 3.597977243994943e-07, "epoch": 3.25154130702836 }, { "current_steps": 2637, "loss": 1.9141, "learning_rate": 3.5954487989886217e-07, "epoch": 3.252774352651048 }, { "current_steps": 2638, "loss": 2.0848, "learning_rate": 3.5929203539823006e-07, "epoch": 3.2540073982737363 }, { "current_steps": 2639, "loss": 2.3302, "learning_rate": 3.59039190897598e-07, "epoch": 3.255240443896424 }, { "current_steps": 2640, "loss": 1.3538, "learning_rate": 3.587863463969659e-07, "epoch": 3.256473489519112 }, { "current_steps": 2641, "loss": 2.3091, "learning_rate": 3.585335018963337e-07, "epoch": 3.2577065351418004 }, { "current_steps": 2642, "loss": 2.1895, "learning_rate": 3.582806573957016e-07, "epoch": 3.2589395807644883 }, { "current_steps": 2643, "loss": 2.1062, "learning_rate": 3.580278128950695e-07, "epoch": 3.260172626387176 }, { "current_steps": 2644, "loss": 1.7462, "learning_rate": 3.577749683944374e-07, "epoch": 3.2614056720098645 }, { "current_steps": 2645, "loss": 2.0428, "learning_rate": 3.575221238938053e-07, "epoch": 3.2626387176325524 }, { "current_steps": 2646, "loss": 1.7077, "learning_rate": 3.572692793931732e-07, "epoch": 3.2638717632552403 }, { "current_steps": 2647, "loss": 2.288, "learning_rate": 3.5701643489254105e-07, "epoch": 3.2651048088779286 }, { "current_steps": 2648, "loss": 2.1799, "learning_rate": 3.5676359039190894e-07, "epoch": 3.2663378545006165 }, { "current_steps": 2649, "loss": 2.2574, "learning_rate": 3.565107458912768e-07, "epoch": 3.2675709001233044 }, { "current_steps": 2650, "loss": 2.5168, "learning_rate": 3.562579013906447e-07, "epoch": 3.2688039457459928 }, { "current_steps": 2651, "loss": 2.3659, "learning_rate": 3.5600505689001265e-07, "epoch": 3.2700369913686806 }, { "current_steps": 2652, "loss": 2.2991, "learning_rate": 3.5575221238938054e-07, "epoch": 3.2712700369913685 }, { "current_steps": 2653, "loss": 2.6846, "learning_rate": 3.5549936788874843e-07, "epoch": 3.272503082614057 }, { "current_steps": 2654, "loss": 2.2893, "learning_rate": 3.5524652338811626e-07, "epoch": 3.2737361282367448 }, { "current_steps": 2655, "loss": 1.9137, "learning_rate": 3.5499367888748415e-07, "epoch": 3.2749691738594326 }, { "current_steps": 2656, "loss": 1.6457, "learning_rate": 3.547408343868521e-07, "epoch": 3.276202219482121 }, { "current_steps": 2657, "loss": 2.1879, "learning_rate": 3.5448798988622e-07, "epoch": 3.277435265104809 }, { "current_steps": 2658, "loss": 1.9334, "learning_rate": 3.5423514538558787e-07, "epoch": 3.2786683107274968 }, { "current_steps": 2659, "loss": 1.2215, "learning_rate": 3.5398230088495575e-07, "epoch": 3.279901356350185 }, { "current_steps": 2660, "loss": 2.0453, "learning_rate": 3.537294563843236e-07, "epoch": 3.281134401972873 }, { "current_steps": 2661, "loss": 1.7752, "learning_rate": 3.534766118836915e-07, "epoch": 3.282367447595561 }, { "current_steps": 2662, "loss": 2.2975, "learning_rate": 3.532237673830594e-07, "epoch": 3.283600493218249 }, { "current_steps": 2663, "loss": 2.3205, "learning_rate": 3.529709228824273e-07, "epoch": 3.284833538840937 }, { "current_steps": 2664, "loss": 1.9415, "learning_rate": 3.527180783817952e-07, "epoch": 3.286066584463625 }, { "current_steps": 2665, "loss": 2.3649, "learning_rate": 3.524652338811631e-07, "epoch": 3.2872996300863133 }, { "current_steps": 2666, "loss": 1.7282, "learning_rate": 3.522123893805309e-07, "epoch": 3.288532675709001 }, { "current_steps": 2667, "loss": 1.6414, "learning_rate": 3.5195954487989885e-07, "epoch": 3.289765721331689 }, { "current_steps": 2668, "loss": 2.1695, "learning_rate": 3.5170670037926674e-07, "epoch": 3.2909987669543774 }, { "current_steps": 2669, "loss": 2.1854, "learning_rate": 3.5145385587863463e-07, "epoch": 3.2922318125770653 }, { "current_steps": 2670, "loss": 2.1894, "learning_rate": 3.512010113780025e-07, "epoch": 3.293464858199753 }, { "current_steps": 2671, "loss": 2.1314, "learning_rate": 3.509481668773704e-07, "epoch": 3.2946979038224415 }, { "current_steps": 2672, "loss": 1.9681, "learning_rate": 3.506953223767383e-07, "epoch": 3.2959309494451294 }, { "current_steps": 2673, "loss": 1.6243, "learning_rate": 3.504424778761062e-07, "epoch": 3.2971639950678173 }, { "current_steps": 2674, "loss": 1.712, "learning_rate": 3.5018963337547407e-07, "epoch": 3.2983970406905057 }, { "current_steps": 2675, "loss": 1.9738, "learning_rate": 3.4993678887484196e-07, "epoch": 3.2996300863131935 }, { "current_steps": 2676, "loss": 1.6287, "learning_rate": 3.4968394437420984e-07, "epoch": 3.3008631319358814 }, { "current_steps": 2677, "loss": 1.3333, "learning_rate": 3.4943109987357773e-07, "epoch": 3.3020961775585698 }, { "current_steps": 2678, "loss": 2.0409, "learning_rate": 3.491782553729456e-07, "epoch": 3.3033292231812577 }, { "current_steps": 2679, "loss": 1.795, "learning_rate": 3.489254108723135e-07, "epoch": 3.304562268803946 }, { "current_steps": 2680, "loss": 1.8505, "learning_rate": 3.486725663716814e-07, "epoch": 3.305795314426634 }, { "current_steps": 2681, "loss": 2.1516, "learning_rate": 3.484197218710493e-07, "epoch": 3.3070283600493218 }, { "current_steps": 2682, "loss": 1.5914, "learning_rate": 3.4816687737041717e-07, "epoch": 3.3082614056720097 }, { "current_steps": 2683, "loss": 2.0836, "learning_rate": 3.4791403286978506e-07, "epoch": 3.309494451294698 }, { "current_steps": 2684, "loss": 2.0356, "learning_rate": 3.47661188369153e-07, "epoch": 3.310727496917386 }, { "current_steps": 2685, "loss": 1.9687, "learning_rate": 3.4740834386852083e-07, "epoch": 3.311960542540074 }, { "current_steps": 2686, "loss": 2.2286, "learning_rate": 3.471554993678887e-07, "epoch": 3.313193588162762 }, { "current_steps": 2687, "loss": 1.668, "learning_rate": 3.469026548672566e-07, "epoch": 3.31442663378545 }, { "current_steps": 2688, "loss": 1.574, "learning_rate": 3.466498103666245e-07, "epoch": 3.315659679408138 }, { "current_steps": 2689, "loss": 2.0918, "learning_rate": 3.463969658659924e-07, "epoch": 3.316892725030826 }, { "current_steps": 2690, "loss": 2.0789, "learning_rate": 3.461441213653603e-07, "epoch": 3.318125770653514 }, { "current_steps": 2691, "loss": 2.2285, "learning_rate": 3.458912768647282e-07, "epoch": 3.3193588162762024 }, { "current_steps": 2692, "loss": 1.8998, "learning_rate": 3.4563843236409605e-07, "epoch": 3.3205918618988903 }, { "current_steps": 2693, "loss": 1.83, "learning_rate": 3.4538558786346393e-07, "epoch": 3.321824907521578 }, { "current_steps": 2694, "loss": 2.0261, "learning_rate": 3.451327433628318e-07, "epoch": 3.323057953144266 }, { "current_steps": 2695, "loss": 2.3577, "learning_rate": 3.4487989886219976e-07, "epoch": 3.3242909987669544 }, { "current_steps": 2696, "loss": 2.1126, "learning_rate": 3.4462705436156765e-07, "epoch": 3.3255240443896423 }, { "current_steps": 2697, "loss": 1.6035, "learning_rate": 3.4437420986093554e-07, "epoch": 3.3267570900123307 }, { "current_steps": 2698, "loss": 2.4479, "learning_rate": 3.4412136536030337e-07, "epoch": 3.3279901356350186 }, { "current_steps": 2699, "loss": 1.887, "learning_rate": 3.4386852085967126e-07, "epoch": 3.3292231812577064 }, { "current_steps": 2700, "loss": 2.1545, "learning_rate": 3.4361567635903915e-07, "epoch": 3.3304562268803943 }, { "current_steps": 2701, "loss": 2.1286, "learning_rate": 3.433628318584071e-07, "epoch": 3.3316892725030827 }, { "current_steps": 2702, "loss": 2.2001, "learning_rate": 3.43109987357775e-07, "epoch": 3.3329223181257706 }, { "current_steps": 2703, "loss": 2.0359, "learning_rate": 3.4285714285714286e-07, "epoch": 3.334155363748459 }, { "current_steps": 2704, "loss": 1.6675, "learning_rate": 3.426042983565107e-07, "epoch": 3.335388409371147 }, { "current_steps": 2705, "loss": 2.1908, "learning_rate": 3.423514538558786e-07, "epoch": 3.3366214549938347 }, { "current_steps": 2706, "loss": 1.9868, "learning_rate": 3.420986093552465e-07, "epoch": 3.337854500616523 }, { "current_steps": 2707, "loss": 2.2358, "learning_rate": 3.418457648546144e-07, "epoch": 3.339087546239211 }, { "current_steps": 2708, "loss": 1.9857, "learning_rate": 3.415929203539823e-07, "epoch": 3.340320591861899 }, { "current_steps": 2709, "loss": 2.2537, "learning_rate": 3.413400758533502e-07, "epoch": 3.341553637484587 }, { "current_steps": 2710, "loss": 2.241, "learning_rate": 3.410872313527181e-07, "epoch": 3.342786683107275 }, { "current_steps": 2711, "loss": 2.1797, "learning_rate": 3.408343868520859e-07, "epoch": 3.344019728729963 }, { "current_steps": 2712, "loss": 2.0836, "learning_rate": 3.4058154235145385e-07, "epoch": 3.3452527743526512 }, { "current_steps": 2713, "loss": 2.0295, "learning_rate": 3.4032869785082174e-07, "epoch": 3.346485819975339 }, { "current_steps": 2714, "loss": 1.4624, "learning_rate": 3.4007585335018963e-07, "epoch": 3.347718865598027 }, { "current_steps": 2715, "loss": 1.9678, "learning_rate": 3.398230088495575e-07, "epoch": 3.3489519112207153 }, { "current_steps": 2716, "loss": 2.2732, "learning_rate": 3.395701643489254e-07, "epoch": 3.3501849568434032 }, { "current_steps": 2717, "loss": 1.8989, "learning_rate": 3.3931731984829324e-07, "epoch": 3.351418002466091 }, { "current_steps": 2718, "loss": 2.4075, "learning_rate": 3.390644753476612e-07, "epoch": 3.3526510480887795 }, { "current_steps": 2719, "loss": 2.2951, "learning_rate": 3.3881163084702907e-07, "epoch": 3.3538840937114673 }, { "current_steps": 2720, "loss": 1.764, "learning_rate": 3.3855878634639696e-07, "epoch": 3.3551171393341552 }, { "current_steps": 2721, "loss": 2.2214, "learning_rate": 3.3830594184576484e-07, "epoch": 3.3563501849568436 }, { "current_steps": 2722, "loss": 1.9801, "learning_rate": 3.3805309734513273e-07, "epoch": 3.3575832305795315 }, { "current_steps": 2723, "loss": 2.2058, "learning_rate": 3.378002528445006e-07, "epoch": 3.3588162762022193 }, { "current_steps": 2724, "loss": 2.5191, "learning_rate": 3.375474083438685e-07, "epoch": 3.3600493218249077 }, { "current_steps": 2725, "loss": 2.0069, "learning_rate": 3.372945638432364e-07, "epoch": 3.3612823674475956 }, { "current_steps": 2726, "loss": 2.1335, "learning_rate": 3.370417193426043e-07, "epoch": 3.3625154130702835 }, { "current_steps": 2727, "loss": 2.1819, "learning_rate": 3.3678887484197217e-07, "epoch": 3.363748458692972 }, { "current_steps": 2728, "loss": 1.9791, "learning_rate": 3.3653603034134006e-07, "epoch": 3.3649815043156597 }, { "current_steps": 2729, "loss": 2.1774, "learning_rate": 3.36283185840708e-07, "epoch": 3.3662145499383476 }, { "current_steps": 2730, "loss": 1.6717, "learning_rate": 3.3603034134007583e-07, "epoch": 3.367447595561036 }, { "current_steps": 2731, "loss": 2.2098, "learning_rate": 3.357774968394437e-07, "epoch": 3.368680641183724 }, { "current_steps": 2732, "loss": 2.5176, "learning_rate": 3.355246523388116e-07, "epoch": 3.3699136868064117 }, { "current_steps": 2733, "loss": 1.9188, "learning_rate": 3.352718078381795e-07, "epoch": 3.3711467324291 }, { "current_steps": 2734, "loss": 1.7916, "learning_rate": 3.350189633375474e-07, "epoch": 3.372379778051788 }, { "current_steps": 2735, "loss": 2.0697, "learning_rate": 3.347661188369153e-07, "epoch": 3.373612823674476 }, { "current_steps": 2736, "loss": 1.6762, "learning_rate": 3.3451327433628316e-07, "epoch": 3.374845869297164 }, { "current_steps": 2737, "loss": 1.8314, "learning_rate": 3.3426042983565105e-07, "epoch": 3.376078914919852 }, { "current_steps": 2738, "loss": 2.2585, "learning_rate": 3.3400758533501893e-07, "epoch": 3.37731196054254 }, { "current_steps": 2739, "loss": 2.4897, "learning_rate": 3.337547408343868e-07, "epoch": 3.3785450061652282 }, { "current_steps": 2740, "loss": 1.821, "learning_rate": 3.3350189633375476e-07, "epoch": 3.379778051787916 }, { "current_steps": 2741, "loss": 2.1705, "learning_rate": 3.3324905183312265e-07, "epoch": 3.381011097410604 }, { "current_steps": 2742, "loss": 2.3101, "learning_rate": 3.329962073324905e-07, "epoch": 3.3822441430332923 }, { "current_steps": 2743, "loss": 2.5563, "learning_rate": 3.3274336283185837e-07, "epoch": 3.3834771886559802 }, { "current_steps": 2744, "loss": 2.1102, "learning_rate": 3.3249051833122626e-07, "epoch": 3.384710234278668 }, { "current_steps": 2745, "loss": 2.5625, "learning_rate": 3.3223767383059415e-07, "epoch": 3.3859432799013565 }, { "current_steps": 2746, "loss": 2.004, "learning_rate": 3.319848293299621e-07, "epoch": 3.3871763255240444 }, { "current_steps": 2747, "loss": 2.2163, "learning_rate": 3.3173198482933e-07, "epoch": 3.3884093711467322 }, { "current_steps": 2748, "loss": 1.9165, "learning_rate": 3.3147914032869786e-07, "epoch": 3.3896424167694206 }, { "current_steps": 2749, "loss": 2.0127, "learning_rate": 3.312262958280657e-07, "epoch": 3.3908754623921085 }, { "current_steps": 2750, "loss": 2.2878, "learning_rate": 3.309734513274336e-07, "epoch": 3.392108508014797 }, { "current_steps": 2751, "loss": 2.4522, "learning_rate": 3.307206068268015e-07, "epoch": 3.3933415536374847 }, { "current_steps": 2752, "loss": 2.6235, "learning_rate": 3.304677623261694e-07, "epoch": 3.3945745992601726 }, { "current_steps": 2753, "loss": 1.3978, "learning_rate": 3.302149178255373e-07, "epoch": 3.3958076448828605 }, { "current_steps": 2754, "loss": 1.9667, "learning_rate": 3.299620733249052e-07, "epoch": 3.397040690505549 }, { "current_steps": 2755, "loss": 2.0266, "learning_rate": 3.29709228824273e-07, "epoch": 3.3982737361282367 }, { "current_steps": 2756, "loss": 2.62, "learning_rate": 3.294563843236409e-07, "epoch": 3.399506781750925 }, { "current_steps": 2757, "loss": 1.7234, "learning_rate": 3.2920353982300885e-07, "epoch": 3.400739827373613 }, { "current_steps": 2758, "loss": 1.5541, "learning_rate": 3.2895069532237674e-07, "epoch": 3.401972872996301 }, { "current_steps": 2759, "loss": 2.1718, "learning_rate": 3.2869785082174463e-07, "epoch": 3.4032059186189887 }, { "current_steps": 2760, "loss": 1.9255, "learning_rate": 3.284450063211125e-07, "epoch": 3.404438964241677 }, { "current_steps": 2761, "loss": 1.8705, "learning_rate": 3.2819216182048035e-07, "epoch": 3.405672009864365 }, { "current_steps": 2762, "loss": 2.4638, "learning_rate": 3.2793931731984824e-07, "epoch": 3.4069050554870532 }, { "current_steps": 2763, "loss": 1.8386, "learning_rate": 3.276864728192162e-07, "epoch": 3.408138101109741 }, { "current_steps": 2764, "loss": 1.6715, "learning_rate": 3.2743362831858407e-07, "epoch": 3.409371146732429 }, { "current_steps": 2765, "loss": 1.9983, "learning_rate": 3.2718078381795195e-07, "epoch": 3.410604192355117 }, { "current_steps": 2766, "loss": 2.1073, "learning_rate": 3.2692793931731984e-07, "epoch": 3.4118372379778052 }, { "current_steps": 2767, "loss": 2.3095, "learning_rate": 3.266750948166877e-07, "epoch": 3.413070283600493 }, { "current_steps": 2768, "loss": 1.963, "learning_rate": 3.264222503160556e-07, "epoch": 3.4143033292231815 }, { "current_steps": 2769, "loss": 1.9258, "learning_rate": 3.261694058154235e-07, "epoch": 3.4155363748458694 }, { "current_steps": 2770, "loss": 2.0242, "learning_rate": 3.259165613147914e-07, "epoch": 3.4167694204685573 }, { "current_steps": 2771, "loss": 2.3036, "learning_rate": 3.256637168141593e-07, "epoch": 3.418002466091245 }, { "current_steps": 2772, "loss": 2.0524, "learning_rate": 3.2541087231352717e-07, "epoch": 3.4192355117139335 }, { "current_steps": 2773, "loss": 2.0824, "learning_rate": 3.2515802781289506e-07, "epoch": 3.4204685573366214 }, { "current_steps": 2774, "loss": 1.779, "learning_rate": 3.2490518331226294e-07, "epoch": 3.4217016029593097 }, { "current_steps": 2775, "loss": 2.3587, "learning_rate": 3.2465233881163083e-07, "epoch": 3.4229346485819976 }, { "current_steps": 2776, "loss": 2.0868, "learning_rate": 3.243994943109987e-07, "epoch": 3.4241676942046855 }, { "current_steps": 2777, "loss": 2.4764, "learning_rate": 3.241466498103666e-07, "epoch": 3.4254007398273734 }, { "current_steps": 2778, "loss": 2.2482, "learning_rate": 3.238938053097345e-07, "epoch": 3.4266337854500617 }, { "current_steps": 2779, "loss": 2.1096, "learning_rate": 3.2364096080910243e-07, "epoch": 3.4278668310727496 }, { "current_steps": 2780, "loss": 2.3654, "learning_rate": 3.2338811630847027e-07, "epoch": 3.429099876695438 }, { "current_steps": 2781, "loss": 1.6018, "learning_rate": 3.2313527180783816e-07, "epoch": 3.430332922318126 }, { "current_steps": 2782, "loss": 2.2727, "learning_rate": 3.2288242730720604e-07, "epoch": 3.4315659679408137 }, { "current_steps": 2783, "loss": 1.6546, "learning_rate": 3.2262958280657393e-07, "epoch": 3.432799013563502 }, { "current_steps": 2784, "loss": 1.84, "learning_rate": 3.223767383059418e-07, "epoch": 3.43403205918619 }, { "current_steps": 2785, "loss": 2.2275, "learning_rate": 3.2212389380530976e-07, "epoch": 3.435265104808878 }, { "current_steps": 2786, "loss": 1.9588, "learning_rate": 3.2187104930467765e-07, "epoch": 3.436498150431566 }, { "current_steps": 2787, "loss": 2.3401, "learning_rate": 3.216182048040455e-07, "epoch": 3.437731196054254 }, { "current_steps": 2788, "loss": 2.2324, "learning_rate": 3.2136536030341337e-07, "epoch": 3.438964241676942 }, { "current_steps": 2789, "loss": 2.378, "learning_rate": 3.2111251580278126e-07, "epoch": 3.4401972872996303 }, { "current_steps": 2790, "loss": 1.816, "learning_rate": 3.2085967130214915e-07, "epoch": 3.441430332922318 }, { "current_steps": 2791, "loss": 1.8291, "learning_rate": 3.206068268015171e-07, "epoch": 3.442663378545006 }, { "current_steps": 2792, "loss": 2.2998, "learning_rate": 3.20353982300885e-07, "epoch": 3.4438964241676944 }, { "current_steps": 2793, "loss": 1.9555, "learning_rate": 3.201011378002528e-07, "epoch": 3.4451294697903823 }, { "current_steps": 2794, "loss": 1.9036, "learning_rate": 3.198482932996207e-07, "epoch": 3.44636251541307 }, { "current_steps": 2795, "loss": 2.25, "learning_rate": 3.195954487989886e-07, "epoch": 3.4475955610357585 }, { "current_steps": 2796, "loss": 2.3296, "learning_rate": 3.193426042983565e-07, "epoch": 3.4488286066584464 }, { "current_steps": 2797, "loss": 1.9413, "learning_rate": 3.190897597977244e-07, "epoch": 3.4500616522811343 }, { "current_steps": 2798, "loss": 2.2839, "learning_rate": 3.188369152970923e-07, "epoch": 3.4512946979038226 }, { "current_steps": 2799, "loss": 1.6424, "learning_rate": 3.1858407079646014e-07, "epoch": 3.4525277435265105 }, { "current_steps": 2800, "loss": 2.3737, "learning_rate": 3.18331226295828e-07, "epoch": 3.4537607891491984 }, { "current_steps": 2801, "loss": 2.4939, "learning_rate": 3.180783817951959e-07, "epoch": 3.4549938347718867 }, { "current_steps": 2802, "loss": 1.6654, "learning_rate": 3.1782553729456385e-07, "epoch": 3.4562268803945746 }, { "current_steps": 2803, "loss": 1.9172, "learning_rate": 3.1757269279393174e-07, "epoch": 3.4574599260172625 }, { "current_steps": 2804, "loss": 2.1799, "learning_rate": 3.1731984829329963e-07, "epoch": 3.458692971639951 }, { "current_steps": 2805, "loss": 1.7211, "learning_rate": 3.1706700379266746e-07, "epoch": 3.4599260172626387 }, { "current_steps": 2806, "loss": 1.8123, "learning_rate": 3.1681415929203535e-07, "epoch": 3.4611590628853266 }, { "current_steps": 2807, "loss": 2.1324, "learning_rate": 3.165613147914033e-07, "epoch": 3.462392108508015 }, { "current_steps": 2808, "loss": 1.9955, "learning_rate": 3.163084702907712e-07, "epoch": 3.463625154130703 }, { "current_steps": 2809, "loss": 2.0445, "learning_rate": 3.1605562579013907e-07, "epoch": 3.4648581997533907 }, { "current_steps": 2810, "loss": 2.1327, "learning_rate": 3.1580278128950695e-07, "epoch": 3.466091245376079 }, { "current_steps": 2811, "loss": 2.102, "learning_rate": 3.1554993678887484e-07, "epoch": 3.467324290998767 }, { "current_steps": 2812, "loss": 2.3914, "learning_rate": 3.152970922882427e-07, "epoch": 3.468557336621455 }, { "current_steps": 2813, "loss": 1.7467, "learning_rate": 3.150442477876106e-07, "epoch": 3.469790382244143 }, { "current_steps": 2814, "loss": 2.0337, "learning_rate": 3.147914032869785e-07, "epoch": 3.471023427866831 }, { "current_steps": 2815, "loss": 2.0961, "learning_rate": 3.145385587863464e-07, "epoch": 3.472256473489519 }, { "current_steps": 2816, "loss": 2.2227, "learning_rate": 3.142857142857143e-07, "epoch": 3.4734895191122073 }, { "current_steps": 2817, "loss": 1.9376, "learning_rate": 3.1403286978508217e-07, "epoch": 3.474722564734895 }, { "current_steps": 2818, "loss": 1.3402, "learning_rate": 3.1378002528445e-07, "epoch": 3.475955610357583 }, { "current_steps": 2819, "loss": 1.5818, "learning_rate": 3.1352718078381794e-07, "epoch": 3.4771886559802714 }, { "current_steps": 2820, "loss": 2.3885, "learning_rate": 3.1327433628318583e-07, "epoch": 3.4784217016029593 }, { "current_steps": 2821, "loss": 1.7528, "learning_rate": 3.130214917825537e-07, "epoch": 3.479654747225647 }, { "current_steps": 2822, "loss": 2.0183, "learning_rate": 3.127686472819216e-07, "epoch": 3.4808877928483355 }, { "current_steps": 2823, "loss": 1.7321, "learning_rate": 3.125158027812895e-07, "epoch": 3.4821208384710234 }, { "current_steps": 2824, "loss": 2.1119, "learning_rate": 3.122629582806574e-07, "epoch": 3.4833538840937113 }, { "current_steps": 2825, "loss": 1.9535, "learning_rate": 3.1201011378002527e-07, "epoch": 3.4845869297163996 }, { "current_steps": 2826, "loss": 1.8682, "learning_rate": 3.1175726927939316e-07, "epoch": 3.4858199753390875 }, { "current_steps": 2827, "loss": 2.0781, "learning_rate": 3.1150442477876104e-07, "epoch": 3.487053020961776 }, { "current_steps": 2828, "loss": 1.6808, "learning_rate": 3.1125158027812893e-07, "epoch": 3.4882860665844637 }, { "current_steps": 2829, "loss": 2.1282, "learning_rate": 3.109987357774968e-07, "epoch": 3.4895191122071516 }, { "current_steps": 2830, "loss": 2.2558, "learning_rate": 3.1074589127686476e-07, "epoch": 3.4907521578298395 }, { "current_steps": 2831, "loss": 2.136, "learning_rate": 3.104930467762326e-07, "epoch": 3.491985203452528 }, { "current_steps": 2832, "loss": 2.5517, "learning_rate": 3.102402022756005e-07, "epoch": 3.4932182490752157 }, { "current_steps": 2833, "loss": 2.0401, "learning_rate": 3.0998735777496837e-07, "epoch": 3.494451294697904 }, { "current_steps": 2834, "loss": 1.6108, "learning_rate": 3.0973451327433626e-07, "epoch": 3.495684340320592 }, { "current_steps": 2835, "loss": 2.339, "learning_rate": 3.094816687737042e-07, "epoch": 3.49691738594328 }, { "current_steps": 2836, "loss": 2.3644, "learning_rate": 3.092288242730721e-07, "epoch": 3.4981504315659677 }, { "current_steps": 2837, "loss": 1.907, "learning_rate": 3.089759797724399e-07, "epoch": 3.499383477188656 }, { "current_steps": 2838, "loss": 2.4028, "learning_rate": 3.087231352718078e-07, "epoch": 3.500616522811344 }, { "current_steps": 2839, "loss": 1.7228, "learning_rate": 3.084702907711757e-07, "epoch": 3.5018495684340323 }, { "current_steps": 2840, "loss": 1.9917, "learning_rate": 3.082174462705436e-07, "epoch": 3.50308261405672 }, { "current_steps": 2841, "loss": 2.0178, "learning_rate": 3.079646017699115e-07, "epoch": 3.504315659679408 }, { "current_steps": 2842, "loss": 1.3355, "learning_rate": 3.077117572692794e-07, "epoch": 3.505548705302096 }, { "current_steps": 2843, "loss": 2.0388, "learning_rate": 3.0745891276864725e-07, "epoch": 3.5067817509247843 }, { "current_steps": 2844, "loss": 2.4542, "learning_rate": 3.0720606826801513e-07, "epoch": 3.508014796547472 }, { "current_steps": 2845, "loss": 1.8923, "learning_rate": 3.06953223767383e-07, "epoch": 3.5092478421701605 }, { "current_steps": 2846, "loss": 2.0745, "learning_rate": 3.067003792667509e-07, "epoch": 3.5104808877928484 }, { "current_steps": 2847, "loss": 2.0513, "learning_rate": 3.0644753476611885e-07, "epoch": 3.5117139334155363 }, { "current_steps": 2848, "loss": 1.7924, "learning_rate": 3.0619469026548674e-07, "epoch": 3.512946979038224 }, { "current_steps": 2849, "loss": 2.3739, "learning_rate": 3.059418457648546e-07, "epoch": 3.5141800246609125 }, { "current_steps": 2850, "loss": 2.1002, "learning_rate": 3.0568900126422246e-07, "epoch": 3.5154130702836004 }, { "current_steps": 2851, "loss": 2.1291, "learning_rate": 3.0543615676359035e-07, "epoch": 3.5166461159062887 }, { "current_steps": 2852, "loss": 2.065, "learning_rate": 3.051833122629583e-07, "epoch": 3.5178791615289766 }, { "current_steps": 2853, "loss": 1.5969, "learning_rate": 3.049304677623262e-07, "epoch": 3.5191122071516645 }, { "current_steps": 2854, "loss": 1.763, "learning_rate": 3.0467762326169406e-07, "epoch": 3.5203452527743524 }, { "current_steps": 2855, "loss": 1.8471, "learning_rate": 3.0442477876106195e-07, "epoch": 3.5215782983970407 }, { "current_steps": 2856, "loss": 1.9627, "learning_rate": 3.041719342604298e-07, "epoch": 3.5228113440197286 }, { "current_steps": 2857, "loss": 2.6215, "learning_rate": 3.039190897597977e-07, "epoch": 3.524044389642417 }, { "current_steps": 2858, "loss": 2.1759, "learning_rate": 3.036662452591656e-07, "epoch": 3.525277435265105 }, { "current_steps": 2859, "loss": 2.1301, "learning_rate": 3.034134007585335e-07, "epoch": 3.5265104808877927 }, { "current_steps": 2860, "loss": 2.1309, "learning_rate": 3.031605562579014e-07, "epoch": 3.5277435265104806 }, { "current_steps": 2861, "loss": 2.4813, "learning_rate": 3.029077117572693e-07, "epoch": 3.528976572133169 }, { "current_steps": 2862, "loss": 1.6999, "learning_rate": 3.026548672566371e-07, "epoch": 3.530209617755857 }, { "current_steps": 2863, "loss": 2.3079, "learning_rate": 3.0240202275600505e-07, "epoch": 3.531442663378545 }, { "current_steps": 2864, "loss": 2.136, "learning_rate": 3.0214917825537294e-07, "epoch": 3.532675709001233 }, { "current_steps": 2865, "loss": 1.9686, "learning_rate": 3.0189633375474083e-07, "epoch": 3.533908754623921 }, { "current_steps": 2866, "loss": 2.1138, "learning_rate": 3.016434892541087e-07, "epoch": 3.5351418002466093 }, { "current_steps": 2867, "loss": 1.9156, "learning_rate": 3.013906447534766e-07, "epoch": 3.536374845869297 }, { "current_steps": 2868, "loss": 2.1965, "learning_rate": 3.0113780025284444e-07, "epoch": 3.537607891491985 }, { "current_steps": 2869, "loss": 2.2788, "learning_rate": 3.008849557522124e-07, "epoch": 3.5388409371146734 }, { "current_steps": 2870, "loss": 1.7447, "learning_rate": 3.0063211125158027e-07, "epoch": 3.5400739827373613 }, { "current_steps": 2871, "loss": 2.0175, "learning_rate": 3.0037926675094815e-07, "epoch": 3.541307028360049 }, { "current_steps": 2872, "loss": 2.0725, "learning_rate": 3.0012642225031604e-07, "epoch": 3.5425400739827375 }, { "current_steps": 2873, "loss": 2.2208, "learning_rate": 2.9987357774968393e-07, "epoch": 3.5437731196054254 }, { "current_steps": 2874, "loss": 2.3002, "learning_rate": 2.996207332490518e-07, "epoch": 3.5450061652281133 }, { "current_steps": 2875, "loss": 2.5163, "learning_rate": 2.993678887484197e-07, "epoch": 3.5462392108508016 }, { "current_steps": 2876, "loss": 1.9746, "learning_rate": 2.991150442477876e-07, "epoch": 3.5474722564734895 }, { "current_steps": 2877, "loss": 2.4361, "learning_rate": 2.988621997471555e-07, "epoch": 3.5487053020961774 }, { "current_steps": 2878, "loss": 2.1556, "learning_rate": 2.9860935524652337e-07, "epoch": 3.5499383477188657 }, { "current_steps": 2879, "loss": 1.9449, "learning_rate": 2.9835651074589126e-07, "epoch": 3.5511713933415536 }, { "current_steps": 2880, "loss": 2.0805, "learning_rate": 2.981036662452592e-07, "epoch": 3.5524044389642415 }, { "current_steps": 2881, "loss": 1.8109, "learning_rate": 2.9785082174462703e-07, "epoch": 3.55363748458693 }, { "current_steps": 2882, "loss": 1.9745, "learning_rate": 2.975979772439949e-07, "epoch": 3.5548705302096177 }, { "current_steps": 2883, "loss": 2.2719, "learning_rate": 2.973451327433628e-07, "epoch": 3.5561035758323056 }, { "current_steps": 2884, "loss": 2.1544, "learning_rate": 2.970922882427307e-07, "epoch": 3.557336621454994 }, { "current_steps": 2885, "loss": 1.9473, "learning_rate": 2.968394437420986e-07, "epoch": 3.558569667077682 }, { "current_steps": 2886, "loss": 2.1576, "learning_rate": 2.965865992414665e-07, "epoch": 3.55980271270037 }, { "current_steps": 2887, "loss": 2.4714, "learning_rate": 2.963337547408344e-07, "epoch": 3.561035758323058 }, { "current_steps": 2888, "loss": 1.893, "learning_rate": 2.9608091024020225e-07, "epoch": 3.562268803945746 }, { "current_steps": 2889, "loss": 1.8543, "learning_rate": 2.9582806573957013e-07, "epoch": 3.563501849568434 }, { "current_steps": 2890, "loss": 1.8858, "learning_rate": 2.95575221238938e-07, "epoch": 3.564734895191122 }, { "current_steps": 2891, "loss": 2.395, "learning_rate": 2.9532237673830596e-07, "epoch": 3.56596794081381 }, { "current_steps": 2892, "loss": 1.9789, "learning_rate": 2.9506953223767385e-07, "epoch": 3.5672009864364984 }, { "current_steps": 2893, "loss": 1.6458, "learning_rate": 2.9481668773704174e-07, "epoch": 3.5684340320591863 }, { "current_steps": 2894, "loss": 2.1751, "learning_rate": 2.9456384323640957e-07, "epoch": 3.569667077681874 }, { "current_steps": 2895, "loss": 1.8552, "learning_rate": 2.9431099873577746e-07, "epoch": 3.570900123304562 }, { "current_steps": 2896, "loss": 2.1905, "learning_rate": 2.9405815423514535e-07, "epoch": 3.5721331689272504 }, { "current_steps": 2897, "loss": 2.1743, "learning_rate": 2.938053097345133e-07, "epoch": 3.5733662145499383 }, { "current_steps": 2898, "loss": 2.0216, "learning_rate": 2.935524652338812e-07, "epoch": 3.5745992601726266 }, { "current_steps": 2899, "loss": 2.2517, "learning_rate": 2.9329962073324906e-07, "epoch": 3.5758323057953145 }, { "current_steps": 2900, "loss": 2.2873, "learning_rate": 2.930467762326169e-07, "epoch": 3.5770653514180024 }, { "current_steps": 2901, "loss": 2.4752, "learning_rate": 2.927939317319848e-07, "epoch": 3.5782983970406903 }, { "current_steps": 2902, "loss": 2.1375, "learning_rate": 2.9254108723135267e-07, "epoch": 3.5795314426633786 }, { "current_steps": 2903, "loss": 1.6355, "learning_rate": 2.922882427307206e-07, "epoch": 3.5807644882860665 }, { "current_steps": 2904, "loss": 2.041, "learning_rate": 2.920353982300885e-07, "epoch": 3.581997533908755 }, { "current_steps": 2905, "loss": 2.2436, "learning_rate": 2.917825537294564e-07, "epoch": 3.5832305795314427 }, { "current_steps": 2906, "loss": 2.1141, "learning_rate": 2.915297092288242e-07, "epoch": 3.5844636251541306 }, { "current_steps": 2907, "loss": 1.9126, "learning_rate": 2.912768647281921e-07, "epoch": 3.5856966707768185 }, { "current_steps": 2908, "loss": 2.3412, "learning_rate": 2.9102402022756005e-07, "epoch": 3.586929716399507 }, { "current_steps": 2909, "loss": 1.9878, "learning_rate": 2.9077117572692794e-07, "epoch": 3.5881627620221948 }, { "current_steps": 2910, "loss": 2.0196, "learning_rate": 2.9051833122629583e-07, "epoch": 3.589395807644883 }, { "current_steps": 2911, "loss": 2.1688, "learning_rate": 2.902654867256637e-07, "epoch": 3.590628853267571 }, { "current_steps": 2912, "loss": 2.4165, "learning_rate": 2.900126422250316e-07, "epoch": 3.591861898890259 }, { "current_steps": 2913, "loss": 2.1762, "learning_rate": 2.8975979772439944e-07, "epoch": 3.5930949445129468 }, { "current_steps": 2914, "loss": 2.2907, "learning_rate": 2.895069532237674e-07, "epoch": 3.594327990135635 }, { "current_steps": 2915, "loss": 2.2426, "learning_rate": 2.8925410872313527e-07, "epoch": 3.595561035758323 }, { "current_steps": 2916, "loss": 2.2506, "learning_rate": 2.8900126422250315e-07, "epoch": 3.5967940813810113 }, { "current_steps": 2917, "loss": 2.1336, "learning_rate": 2.8874841972187104e-07, "epoch": 3.598027127003699 }, { "current_steps": 2918, "loss": 2.1222, "learning_rate": 2.8849557522123893e-07, "epoch": 3.599260172626387 }, { "current_steps": 2919, "loss": 2.4679, "learning_rate": 2.882427307206068e-07, "epoch": 3.600493218249075 }, { "current_steps": 2920, "loss": 1.6806, "learning_rate": 2.879898862199747e-07, "epoch": 3.6017262638717633 }, { "current_steps": 2921, "loss": 2.1096, "learning_rate": 2.877370417193426e-07, "epoch": 3.602959309494451 }, { "current_steps": 2922, "loss": 2.5147, "learning_rate": 2.874841972187105e-07, "epoch": 3.6041923551171395 }, { "current_steps": 2923, "loss": 1.6507, "learning_rate": 2.8723135271807837e-07, "epoch": 3.6054254007398274 }, { "current_steps": 2924, "loss": 1.3179, "learning_rate": 2.8697850821744626e-07, "epoch": 3.6066584463625153 }, { "current_steps": 2925, "loss": 2.3799, "learning_rate": 2.867256637168142e-07, "epoch": 3.607891491985203 }, { "current_steps": 2926, "loss": 2.0551, "learning_rate": 2.8647281921618203e-07, "epoch": 3.6091245376078915 }, { "current_steps": 2927, "loss": 2.2459, "learning_rate": 2.862199747155499e-07, "epoch": 3.6103575832305794 }, { "current_steps": 2928, "loss": 2.2527, "learning_rate": 2.859671302149178e-07, "epoch": 3.6115906288532678 }, { "current_steps": 2929, "loss": 2.3853, "learning_rate": 2.857142857142857e-07, "epoch": 3.6128236744759556 }, { "current_steps": 2930, "loss": 2.4304, "learning_rate": 2.854614412136536e-07, "epoch": 3.6140567200986435 }, { "current_steps": 2931, "loss": 2.086, "learning_rate": 2.852085967130215e-07, "epoch": 3.6152897657213314 }, { "current_steps": 2932, "loss": 1.6197, "learning_rate": 2.8495575221238936e-07, "epoch": 3.6165228113440198 }, { "current_steps": 2933, "loss": 1.5001, "learning_rate": 2.8470290771175724e-07, "epoch": 3.6177558569667077 }, { "current_steps": 2934, "loss": 1.7615, "learning_rate": 2.8445006321112513e-07, "epoch": 3.618988902589396 }, { "current_steps": 2935, "loss": 2.1473, "learning_rate": 2.84197218710493e-07, "epoch": 3.620221948212084 }, { "current_steps": 2936, "loss": 1.9483, "learning_rate": 2.8394437420986096e-07, "epoch": 3.6214549938347718 }, { "current_steps": 2937, "loss": 2.1022, "learning_rate": 2.8369152970922885e-07, "epoch": 3.6226880394574597 }, { "current_steps": 2938, "loss": 2.2599, "learning_rate": 2.834386852085967e-07, "epoch": 3.623921085080148 }, { "current_steps": 2939, "loss": 1.6634, "learning_rate": 2.8318584070796457e-07, "epoch": 3.625154130702836 }, { "current_steps": 2940, "loss": 1.8362, "learning_rate": 2.8293299620733246e-07, "epoch": 3.626387176325524 }, { "current_steps": 2941, "loss": 1.4143, "learning_rate": 2.8268015170670035e-07, "epoch": 3.627620221948212 }, { "current_steps": 2942, "loss": 1.9613, "learning_rate": 2.824273072060683e-07, "epoch": 3.6288532675709 }, { "current_steps": 2943, "loss": 1.7103, "learning_rate": 2.821744627054362e-07, "epoch": 3.6300863131935883 }, { "current_steps": 2944, "loss": 2.0413, "learning_rate": 2.81921618204804e-07, "epoch": 3.631319358816276 }, { "current_steps": 2945, "loss": 2.2592, "learning_rate": 2.816687737041719e-07, "epoch": 3.632552404438964 }, { "current_steps": 2946, "loss": 2.176, "learning_rate": 2.814159292035398e-07, "epoch": 3.6337854500616524 }, { "current_steps": 2947, "loss": 2.2213, "learning_rate": 2.8116308470290767e-07, "epoch": 3.6350184956843403 }, { "current_steps": 2948, "loss": 1.9509, "learning_rate": 2.809102402022756e-07, "epoch": 3.636251541307028 }, { "current_steps": 2949, "loss": 1.7927, "learning_rate": 2.806573957016435e-07, "epoch": 3.6374845869297165 }, { "current_steps": 2950, "loss": 1.9293, "learning_rate": 2.804045512010114e-07, "epoch": 3.6387176325524044 }, { "current_steps": 2951, "loss": 1.9874, "learning_rate": 2.801517067003792e-07, "epoch": 3.6399506781750923 }, { "current_steps": 2952, "loss": 2.0487, "learning_rate": 2.798988621997471e-07, "epoch": 3.6411837237977807 }, { "current_steps": 2953, "loss": 2.1112, "learning_rate": 2.7964601769911505e-07, "epoch": 3.6424167694204685 }, { "current_steps": 2954, "loss": 2.1423, "learning_rate": 2.7939317319848294e-07, "epoch": 3.6436498150431564 }, { "current_steps": 2955, "loss": 2.1762, "learning_rate": 2.7914032869785083e-07, "epoch": 3.6448828606658448 }, { "current_steps": 2956, "loss": 2.1852, "learning_rate": 2.788874841972187e-07, "epoch": 3.6461159062885327 }, { "current_steps": 2957, "loss": 2.1233, "learning_rate": 2.7863463969658655e-07, "epoch": 3.6473489519112205 }, { "current_steps": 2958, "loss": 2.1697, "learning_rate": 2.7838179519595444e-07, "epoch": 3.648581997533909 }, { "current_steps": 2959, "loss": 1.9092, "learning_rate": 2.781289506953224e-07, "epoch": 3.6498150431565968 }, { "current_steps": 2960, "loss": 2.1795, "learning_rate": 2.7787610619469027e-07, "epoch": 3.6510480887792847 }, { "current_steps": 2961, "loss": 1.6668, "learning_rate": 2.7762326169405815e-07, "epoch": 3.652281134401973 }, { "current_steps": 2962, "loss": 2.0548, "learning_rate": 2.7737041719342604e-07, "epoch": 3.653514180024661 }, { "current_steps": 2963, "loss": 1.9653, "learning_rate": 2.771175726927939e-07, "epoch": 3.654747225647349 }, { "current_steps": 2964, "loss": 2.2814, "learning_rate": 2.768647281921618e-07, "epoch": 3.655980271270037 }, { "current_steps": 2965, "loss": 2.1218, "learning_rate": 2.766118836915297e-07, "epoch": 3.657213316892725 }, { "current_steps": 2966, "loss": 2.123, "learning_rate": 2.763590391908976e-07, "epoch": 3.658446362515413 }, { "current_steps": 2967, "loss": 1.9653, "learning_rate": 2.761061946902655e-07, "epoch": 3.659679408138101 }, { "current_steps": 2968, "loss": 2.1372, "learning_rate": 2.7585335018963337e-07, "epoch": 3.660912453760789 }, { "current_steps": 2969, "loss": 2.0645, "learning_rate": 2.7560050568900125e-07, "epoch": 3.6621454993834774 }, { "current_steps": 2970, "loss": 2.1458, "learning_rate": 2.7534766118836914e-07, "epoch": 3.6633785450061653 }, { "current_steps": 2971, "loss": 2.4278, "learning_rate": 2.7509481668773703e-07, "epoch": 3.664611590628853 }, { "current_steps": 2972, "loss": 2.3822, "learning_rate": 2.748419721871049e-07, "epoch": 3.665844636251541 }, { "current_steps": 2973, "loss": 1.9794, "learning_rate": 2.745891276864728e-07, "epoch": 3.6670776818742294 }, { "current_steps": 2974, "loss": 2.1175, "learning_rate": 2.743362831858407e-07, "epoch": 3.6683107274969173 }, { "current_steps": 2975, "loss": 2.3183, "learning_rate": 2.740834386852086e-07, "epoch": 3.6695437731196057 }, { "current_steps": 2976, "loss": 1.7491, "learning_rate": 2.7383059418457647e-07, "epoch": 3.6707768187422936 }, { "current_steps": 2977, "loss": 2.4829, "learning_rate": 2.7357774968394436e-07, "epoch": 3.6720098643649814 }, { "current_steps": 2978, "loss": 2.3081, "learning_rate": 2.7332490518331224e-07, "epoch": 3.6732429099876693 }, { "current_steps": 2979, "loss": 1.8652, "learning_rate": 2.7307206068268013e-07, "epoch": 3.6744759556103577 }, { "current_steps": 2980, "loss": 2.0377, "learning_rate": 2.72819216182048e-07, "epoch": 3.6757090012330456 }, { "current_steps": 2981, "loss": 2.4414, "learning_rate": 2.7256637168141596e-07, "epoch": 3.676942046855734 }, { "current_steps": 2982, "loss": 2.0572, "learning_rate": 2.723135271807838e-07, "epoch": 3.678175092478422 }, { "current_steps": 2983, "loss": 2.013, "learning_rate": 2.720606826801517e-07, "epoch": 3.6794081381011097 }, { "current_steps": 2984, "loss": 2.2069, "learning_rate": 2.7180783817951957e-07, "epoch": 3.6806411837237976 }, { "current_steps": 2985, "loss": 1.4974, "learning_rate": 2.7155499367888746e-07, "epoch": 3.681874229346486 }, { "current_steps": 2986, "loss": 2.3145, "learning_rate": 2.7130214917825535e-07, "epoch": 3.683107274969174 }, { "current_steps": 2987, "loss": 2.0404, "learning_rate": 2.710493046776233e-07, "epoch": 3.684340320591862 }, { "current_steps": 2988, "loss": 2.3231, "learning_rate": 2.7079646017699117e-07, "epoch": 3.68557336621455 }, { "current_steps": 2989, "loss": 2.2724, "learning_rate": 2.70543615676359e-07, "epoch": 3.686806411837238 }, { "current_steps": 2990, "loss": 2.2654, "learning_rate": 2.702907711757269e-07, "epoch": 3.688039457459926 }, { "current_steps": 2991, "loss": 2.0398, "learning_rate": 2.700379266750948e-07, "epoch": 3.689272503082614 }, { "current_steps": 2992, "loss": 1.6699, "learning_rate": 2.697850821744627e-07, "epoch": 3.690505548705302 }, { "current_steps": 2993, "loss": 2.0294, "learning_rate": 2.695322376738306e-07, "epoch": 3.6917385943279903 }, { "current_steps": 2994, "loss": 1.8424, "learning_rate": 2.692793931731985e-07, "epoch": 3.6929716399506782 }, { "current_steps": 2995, "loss": 2.1885, "learning_rate": 2.6902654867256633e-07, "epoch": 3.694204685573366 }, { "current_steps": 2996, "loss": 1.8841, "learning_rate": 2.687737041719342e-07, "epoch": 3.695437731196054 }, { "current_steps": 2997, "loss": 1.7444, "learning_rate": 2.685208596713021e-07, "epoch": 3.6966707768187423 }, { "current_steps": 2998, "loss": 1.745, "learning_rate": 2.6826801517067005e-07, "epoch": 3.6979038224414302 }, { "current_steps": 2999, "loss": 2.7378, "learning_rate": 2.6801517067003794e-07, "epoch": 3.6991368680641186 }, { "current_steps": 3000, "loss": 2.1828, "learning_rate": 2.677623261694058e-07, "epoch": 3.7003699136868065 }, { "current_steps": 3001, "loss": 1.9308, "learning_rate": 2.6750948166877366e-07, "epoch": 3.7016029593094943 }, { "current_steps": 3002, "loss": 2.2212, "learning_rate": 2.6725663716814155e-07, "epoch": 3.7028360049321822 }, { "current_steps": 3003, "loss": 2.2068, "learning_rate": 2.6700379266750944e-07, "epoch": 3.7040690505548706 }, { "current_steps": 3004, "loss": 1.3564, "learning_rate": 2.667509481668774e-07, "epoch": 3.7053020961775585 }, { "current_steps": 3005, "loss": 2.1177, "learning_rate": 2.6649810366624526e-07, "epoch": 3.706535141800247 }, { "current_steps": 3006, "loss": 1.9802, "learning_rate": 2.6624525916561315e-07, "epoch": 3.7077681874229347 }, { "current_steps": 3007, "loss": 2.0305, "learning_rate": 2.6599241466498104e-07, "epoch": 3.7090012330456226 }, { "current_steps": 3008, "loss": 1.5492, "learning_rate": 2.657395701643489e-07, "epoch": 3.7102342786683105 }, { "current_steps": 3009, "loss": 2.0733, "learning_rate": 2.654867256637168e-07, "epoch": 3.711467324290999 }, { "current_steps": 3010, "loss": 2.0335, "learning_rate": 2.652338811630847e-07, "epoch": 3.7127003699136867 }, { "current_steps": 3011, "loss": 1.9329, "learning_rate": 2.649810366624526e-07, "epoch": 3.713933415536375 }, { "current_steps": 3012, "loss": 2.1419, "learning_rate": 2.647281921618205e-07, "epoch": 3.715166461159063 }, { "current_steps": 3013, "loss": 2.1097, "learning_rate": 2.6447534766118837e-07, "epoch": 3.716399506781751 }, { "current_steps": 3014, "loss": 2.1631, "learning_rate": 2.642225031605562e-07, "epoch": 3.717632552404439 }, { "current_steps": 3015, "loss": 2.5009, "learning_rate": 2.6396965865992414e-07, "epoch": 3.718865598027127 }, { "current_steps": 3016, "loss": 2.5943, "learning_rate": 2.6371681415929203e-07, "epoch": 3.720098643649815 }, { "current_steps": 3017, "loss": 2.3364, "learning_rate": 2.634639696586599e-07, "epoch": 3.7213316892725032 }, { "current_steps": 3018, "loss": 2.1812, "learning_rate": 2.632111251580278e-07, "epoch": 3.722564734895191 }, { "current_steps": 3019, "loss": 2.1207, "learning_rate": 2.629582806573957e-07, "epoch": 3.723797780517879 }, { "current_steps": 3020, "loss": 1.7022, "learning_rate": 2.627054361567636e-07, "epoch": 3.7250308261405674 }, { "current_steps": 3021, "loss": 2.3453, "learning_rate": 2.6245259165613147e-07, "epoch": 3.7262638717632552 }, { "current_steps": 3022, "loss": 1.9237, "learning_rate": 2.6219974715549935e-07, "epoch": 3.727496917385943 }, { "current_steps": 3023, "loss": 2.1114, "learning_rate": 2.6194690265486724e-07, "epoch": 3.7287299630086315 }, { "current_steps": 3024, "loss": 2.2894, "learning_rate": 2.6169405815423513e-07, "epoch": 3.7299630086313194 }, { "current_steps": 3025, "loss": 2.3517, "learning_rate": 2.61441213653603e-07, "epoch": 3.7311960542540072 }, { "current_steps": 3026, "loss": 2.122, "learning_rate": 2.6118836915297096e-07, "epoch": 3.7324290998766956 }, { "current_steps": 3027, "loss": 2.327, "learning_rate": 2.609355246523388e-07, "epoch": 3.7336621454993835 }, { "current_steps": 3028, "loss": 2.1007, "learning_rate": 2.606826801517067e-07, "epoch": 3.7348951911220714 }, { "current_steps": 3029, "loss": 2.1471, "learning_rate": 2.6042983565107457e-07, "epoch": 3.7361282367447597 }, { "current_steps": 3030, "loss": 2.1521, "learning_rate": 2.6017699115044246e-07, "epoch": 3.7373612823674476 }, { "current_steps": 3031, "loss": 2.4319, "learning_rate": 2.5992414664981034e-07, "epoch": 3.7385943279901355 }, { "current_steps": 3032, "loss": 2.0132, "learning_rate": 2.596713021491783e-07, "epoch": 3.739827373612824 }, { "current_steps": 3033, "loss": 1.8702, "learning_rate": 2.594184576485461e-07, "epoch": 3.7410604192355117 }, { "current_steps": 3034, "loss": 1.6079, "learning_rate": 2.59165613147914e-07, "epoch": 3.7422934648582 }, { "current_steps": 3035, "loss": 1.7962, "learning_rate": 2.589127686472819e-07, "epoch": 3.743526510480888 }, { "current_steps": 3036, "loss": 2.0397, "learning_rate": 2.586599241466498e-07, "epoch": 3.744759556103576 }, { "current_steps": 3037, "loss": 1.7562, "learning_rate": 2.584070796460177e-07, "epoch": 3.7459926017262637 }, { "current_steps": 3038, "loss": 2.1959, "learning_rate": 2.581542351453856e-07, "epoch": 3.747225647348952 }, { "current_steps": 3039, "loss": 1.5294, "learning_rate": 2.5790139064475345e-07, "epoch": 3.74845869297164 }, { "current_steps": 3040, "loss": 1.9193, "learning_rate": 2.5764854614412133e-07, "epoch": 3.7496917385943282 }, { "current_steps": 3041, "loss": 1.7575, "learning_rate": 2.573957016434892e-07, "epoch": 3.750924784217016 }, { "current_steps": 3042, "loss": 1.9303, "learning_rate": 2.571428571428571e-07, "epoch": 3.752157829839704 }, { "current_steps": 3043, "loss": 2.2132, "learning_rate": 2.5689001264222505e-07, "epoch": 3.753390875462392 }, { "current_steps": 3044, "loss": 1.9306, "learning_rate": 2.5663716814159294e-07, "epoch": 3.7546239210850803 }, { "current_steps": 3045, "loss": 1.7179, "learning_rate": 2.5638432364096077e-07, "epoch": 3.755856966707768 }, { "current_steps": 3046, "loss": 2.0858, "learning_rate": 2.5613147914032866e-07, "epoch": 3.7570900123304565 }, { "current_steps": 3047, "loss": 1.8793, "learning_rate": 2.5587863463969655e-07, "epoch": 3.7583230579531444 }, { "current_steps": 3048, "loss": 1.7205, "learning_rate": 2.556257901390645e-07, "epoch": 3.7595561035758323 }, { "current_steps": 3049, "loss": 1.7329, "learning_rate": 2.553729456384324e-07, "epoch": 3.76078914919852 }, { "current_steps": 3050, "loss": 1.7686, "learning_rate": 2.5512010113780026e-07, "epoch": 3.7620221948212085 }, { "current_steps": 3051, "loss": 2.2707, "learning_rate": 2.5486725663716815e-07, "epoch": 3.7632552404438964 }, { "current_steps": 3052, "loss": 1.8413, "learning_rate": 2.54614412136536e-07, "epoch": 3.7644882860665847 }, { "current_steps": 3053, "loss": 1.8738, "learning_rate": 2.5436156763590387e-07, "epoch": 3.7657213316892726 }, { "current_steps": 3054, "loss": 1.9361, "learning_rate": 2.541087231352718e-07, "epoch": 3.7669543773119605 }, { "current_steps": 3055, "loss": 1.9114, "learning_rate": 2.538558786346397e-07, "epoch": 3.7681874229346484 }, { "current_steps": 3056, "loss": 1.9954, "learning_rate": 2.536030341340076e-07, "epoch": 3.7694204685573367 }, { "current_steps": 3057, "loss": 2.0386, "learning_rate": 2.533501896333755e-07, "epoch": 3.7706535141800246 }, { "current_steps": 3058, "loss": 1.9314, "learning_rate": 2.530973451327433e-07, "epoch": 3.771886559802713 }, { "current_steps": 3059, "loss": 2.3097, "learning_rate": 2.528445006321112e-07, "epoch": 3.773119605425401 }, { "current_steps": 3060, "loss": 2.1255, "learning_rate": 2.5259165613147914e-07, "epoch": 3.7743526510480887 }, { "current_steps": 3061, "loss": 1.8535, "learning_rate": 2.5233881163084703e-07, "epoch": 3.7755856966707766 }, { "current_steps": 3062, "loss": 1.4084, "learning_rate": 2.520859671302149e-07, "epoch": 3.776818742293465 }, { "current_steps": 3063, "loss": 1.9422, "learning_rate": 2.518331226295828e-07, "epoch": 3.778051787916153 }, { "current_steps": 3064, "loss": 2.1076, "learning_rate": 2.5158027812895064e-07, "epoch": 3.779284833538841 }, { "current_steps": 3065, "loss": 2.3385, "learning_rate": 2.513274336283186e-07, "epoch": 3.780517879161529 }, { "current_steps": 3066, "loss": 1.9001, "learning_rate": 2.5107458912768647e-07, "epoch": 3.781750924784217 }, { "current_steps": 3067, "loss": 2.0935, "learning_rate": 2.5082174462705435e-07, "epoch": 3.782983970406905 }, { "current_steps": 3068, "loss": 2.3444, "learning_rate": 2.5056890012642224e-07, "epoch": 3.784217016029593 }, { "current_steps": 3069, "loss": 1.944, "learning_rate": 2.5031605562579013e-07, "epoch": 3.785450061652281 }, { "current_steps": 3070, "loss": 2.0514, "learning_rate": 2.50063211125158e-07, "epoch": 3.7866831072749694 }, { "current_steps": 3071, "loss": 2.1602, "learning_rate": 2.498103666245259e-07, "epoch": 3.7879161528976573 }, { "current_steps": 3072, "loss": 1.9252, "learning_rate": 2.495575221238938e-07, "epoch": 3.789149198520345 }, { "current_steps": 3073, "loss": 2.3119, "learning_rate": 2.493046776232617e-07, "epoch": 3.790382244143033 }, { "current_steps": 3074, "loss": 1.7666, "learning_rate": 2.4905183312262957e-07, "epoch": 3.7916152897657214 }, { "current_steps": 3075, "loss": 2.2425, "learning_rate": 2.4879898862199746e-07, "epoch": 3.7928483353884093 }, { "current_steps": 3076, "loss": 2.2453, "learning_rate": 2.4854614412136534e-07, "epoch": 3.7940813810110976 }, { "current_steps": 3077, "loss": 2.0167, "learning_rate": 2.4829329962073323e-07, "epoch": 3.7953144266337855 }, { "current_steps": 3078, "loss": 0.9148, "learning_rate": 2.480404551201011e-07, "epoch": 3.7965474722564734 }, { "current_steps": 3079, "loss": 2.0747, "learning_rate": 2.47787610619469e-07, "epoch": 3.7977805178791613 }, { "current_steps": 3080, "loss": 2.0996, "learning_rate": 2.475347661188369e-07, "epoch": 3.7990135635018496 }, { "current_steps": 3081, "loss": 1.5655, "learning_rate": 2.472819216182048e-07, "epoch": 3.8002466091245375 }, { "current_steps": 3082, "loss": 2.116, "learning_rate": 2.4702907711757267e-07, "epoch": 3.801479654747226 }, { "current_steps": 3083, "loss": 1.5863, "learning_rate": 2.4677623261694056e-07, "epoch": 3.8027127003699137 }, { "current_steps": 3084, "loss": 1.9044, "learning_rate": 2.4652338811630844e-07, "epoch": 3.8039457459926016 }, { "current_steps": 3085, "loss": 2.2717, "learning_rate": 2.4627054361567633e-07, "epoch": 3.8051787916152895 }, { "current_steps": 3086, "loss": 1.1826, "learning_rate": 2.4601769911504427e-07, "epoch": 3.806411837237978 }, { "current_steps": 3087, "loss": 2.4459, "learning_rate": 2.457648546144121e-07, "epoch": 3.8076448828606657 }, { "current_steps": 3088, "loss": 1.9354, "learning_rate": 2.4551201011378e-07, "epoch": 3.808877928483354 }, { "current_steps": 3089, "loss": 2.5303, "learning_rate": 2.4525916561314794e-07, "epoch": 3.810110974106042 }, { "current_steps": 3090, "loss": 1.9426, "learning_rate": 2.4500632111251577e-07, "epoch": 3.81134401972873 }, { "current_steps": 3091, "loss": 1.9454, "learning_rate": 2.4475347661188366e-07, "epoch": 3.812577065351418 }, { "current_steps": 3092, "loss": 2.423, "learning_rate": 2.445006321112516e-07, "epoch": 3.813810110974106 }, { "current_steps": 3093, "loss": 2.044, "learning_rate": 2.4424778761061943e-07, "epoch": 3.815043156596794 }, { "current_steps": 3094, "loss": 2.0983, "learning_rate": 2.439949431099873e-07, "epoch": 3.8162762022194823 }, { "current_steps": 3095, "loss": 1.9691, "learning_rate": 2.4374209860935526e-07, "epoch": 3.81750924784217 }, { "current_steps": 3096, "loss": 1.9252, "learning_rate": 2.434892541087231e-07, "epoch": 3.818742293464858 }, { "current_steps": 3097, "loss": 2.1214, "learning_rate": 2.4323640960809104e-07, "epoch": 3.8199753390875464 }, { "current_steps": 3098, "loss": 1.4424, "learning_rate": 2.429835651074589e-07, "epoch": 3.8212083847102343 }, { "current_steps": 3099, "loss": 2.1097, "learning_rate": 2.4273072060682676e-07, "epoch": 3.822441430332922 }, { "current_steps": 3100, "loss": 1.533, "learning_rate": 2.424778761061947e-07, "epoch": 3.8236744759556105 }, { "current_steps": 3101, "loss": 2.6795, "learning_rate": 2.422250316055626e-07, "epoch": 3.8249075215782984 }, { "current_steps": 3102, "loss": 2.0742, "learning_rate": 2.419721871049304e-07, "epoch": 3.8261405672009863 }, { "current_steps": 3103, "loss": 2.3003, "learning_rate": 2.4171934260429836e-07, "epoch": 3.8273736128236746 }, { "current_steps": 3104, "loss": 2.1969, "learning_rate": 2.4146649810366625e-07, "epoch": 3.8286066584463625 }, { "current_steps": 3105, "loss": 1.7856, "learning_rate": 2.412136536030341e-07, "epoch": 3.829839704069051 }, { "current_steps": 3106, "loss": 2.0, "learning_rate": 2.4096080910240203e-07, "epoch": 3.8310727496917387 }, { "current_steps": 3107, "loss": 2.6575, "learning_rate": 2.407079646017699e-07, "epoch": 3.8323057953144266 }, { "current_steps": 3108, "loss": 1.263, "learning_rate": 2.404551201011378e-07, "epoch": 3.8335388409371145 }, { "current_steps": 3109, "loss": 2.1737, "learning_rate": 2.402022756005057e-07, "epoch": 3.834771886559803 }, { "current_steps": 3110, "loss": 2.074, "learning_rate": 2.399494310998736e-07, "epoch": 3.8360049321824907 }, { "current_steps": 3111, "loss": 2.0963, "learning_rate": 2.3969658659924146e-07, "epoch": 3.837237977805179 }, { "current_steps": 3112, "loss": 2.1848, "learning_rate": 2.3944374209860935e-07, "epoch": 3.838471023427867 }, { "current_steps": 3113, "loss": 1.9072, "learning_rate": 2.3919089759797724e-07, "epoch": 3.839704069050555 }, { "current_steps": 3114, "loss": 1.822, "learning_rate": 2.3893805309734513e-07, "epoch": 3.8409371146732427 }, { "current_steps": 3115, "loss": 2.2107, "learning_rate": 2.38685208596713e-07, "epoch": 3.842170160295931 }, { "current_steps": 3116, "loss": 1.7286, "learning_rate": 2.384323640960809e-07, "epoch": 3.843403205918619 }, { "current_steps": 3117, "loss": 1.579, "learning_rate": 2.381795195954488e-07, "epoch": 3.8446362515413073 }, { "current_steps": 3118, "loss": 1.7891, "learning_rate": 2.3792667509481668e-07, "epoch": 3.845869297163995 }, { "current_steps": 3119, "loss": 2.3048, "learning_rate": 2.3767383059418457e-07, "epoch": 3.847102342786683 }, { "current_steps": 3120, "loss": 1.6935, "learning_rate": 2.3742098609355245e-07, "epoch": 3.848335388409371 }, { "current_steps": 3121, "loss": 2.5638, "learning_rate": 2.3716814159292034e-07, "epoch": 3.8495684340320593 }, { "current_steps": 3122, "loss": 2.3498, "learning_rate": 2.3691529709228823e-07, "epoch": 3.850801479654747 }, { "current_steps": 3123, "loss": 2.41, "learning_rate": 2.3666245259165612e-07, "epoch": 3.8520345252774355 }, { "current_steps": 3124, "loss": 2.2632, "learning_rate": 2.36409608091024e-07, "epoch": 3.8532675709001234 }, { "current_steps": 3125, "loss": 2.3233, "learning_rate": 2.3615676359039192e-07, "epoch": 3.8545006165228113 }, { "current_steps": 3126, "loss": 1.5579, "learning_rate": 2.3590391908975978e-07, "epoch": 3.855733662145499 }, { "current_steps": 3127, "loss": 1.9057, "learning_rate": 2.3565107458912767e-07, "epoch": 3.8569667077681875 }, { "current_steps": 3128, "loss": 1.509, "learning_rate": 2.3539823008849558e-07, "epoch": 3.8581997533908754 }, { "current_steps": 3129, "loss": 1.8753, "learning_rate": 2.3514538558786344e-07, "epoch": 3.8594327990135637 }, { "current_steps": 3130, "loss": 2.1708, "learning_rate": 2.3489254108723133e-07, "epoch": 3.8606658446362516 }, { "current_steps": 3131, "loss": 1.8864, "learning_rate": 2.3463969658659925e-07, "epoch": 3.8618988902589395 }, { "current_steps": 3132, "loss": 2.0523, "learning_rate": 2.343868520859671e-07, "epoch": 3.8631319358816274 }, { "current_steps": 3133, "loss": 2.3285, "learning_rate": 2.34134007585335e-07, "epoch": 3.8643649815043157 }, { "current_steps": 3134, "loss": 2.2425, "learning_rate": 2.338811630847029e-07, "epoch": 3.8655980271270036 }, { "current_steps": 3135, "loss": 2.293, "learning_rate": 2.3362831858407077e-07, "epoch": 3.866831072749692 }, { "current_steps": 3136, "loss": 2.3174, "learning_rate": 2.3337547408343866e-07, "epoch": 3.86806411837238 }, { "current_steps": 3137, "loss": 2.2482, "learning_rate": 2.3312262958280657e-07, "epoch": 3.8692971639950677 }, { "current_steps": 3138, "loss": 2.4585, "learning_rate": 2.3286978508217446e-07, "epoch": 3.8705302096177556 }, { "current_steps": 3139, "loss": 2.0288, "learning_rate": 2.3261694058154235e-07, "epoch": 3.871763255240444 }, { "current_steps": 3140, "loss": 2.3773, "learning_rate": 2.3236409608091023e-07, "epoch": 3.872996300863132 }, { "current_steps": 3141, "loss": 1.9188, "learning_rate": 2.3211125158027812e-07, "epoch": 3.87422934648582 }, { "current_steps": 3142, "loss": 1.8558, "learning_rate": 2.31858407079646e-07, "epoch": 3.875462392108508 }, { "current_steps": 3143, "loss": 2.0882, "learning_rate": 2.316055625790139e-07, "epoch": 3.876695437731196 }, { "current_steps": 3144, "loss": 2.3017, "learning_rate": 2.3135271807838179e-07, "epoch": 3.877928483353884 }, { "current_steps": 3145, "loss": 1.7245, "learning_rate": 2.3109987357774967e-07, "epoch": 3.879161528976572 }, { "current_steps": 3146, "loss": 2.6247, "learning_rate": 2.3084702907711756e-07, "epoch": 3.88039457459926 }, { "current_steps": 3147, "loss": 2.185, "learning_rate": 2.3059418457648545e-07, "epoch": 3.8816276202219484 }, { "current_steps": 3148, "loss": 2.4566, "learning_rate": 2.3034134007585334e-07, "epoch": 3.8828606658446363 }, { "current_steps": 3149, "loss": 2.1887, "learning_rate": 2.3008849557522122e-07, "epoch": 3.884093711467324 }, { "current_steps": 3150, "loss": 2.1695, "learning_rate": 2.298356510745891e-07, "epoch": 3.885326757090012 }, { "current_steps": 3151, "loss": 2.4135, "learning_rate": 2.29582806573957e-07, "epoch": 3.8865598027127004 }, { "current_steps": 3152, "loss": 1.6525, "learning_rate": 2.2932996207332489e-07, "epoch": 3.8877928483353883 }, { "current_steps": 3153, "loss": 2.5431, "learning_rate": 2.290771175726928e-07, "epoch": 3.8890258939580766 }, { "current_steps": 3154, "loss": 2.1501, "learning_rate": 2.2882427307206066e-07, "epoch": 3.8902589395807645 }, { "current_steps": 3155, "loss": 1.8138, "learning_rate": 2.2857142857142855e-07, "epoch": 3.8914919852034524 }, { "current_steps": 3156, "loss": 2.2758, "learning_rate": 2.2831858407079646e-07, "epoch": 3.8927250308261403 }, { "current_steps": 3157, "loss": 2.3792, "learning_rate": 2.2806573957016435e-07, "epoch": 3.8939580764488286 }, { "current_steps": 3158, "loss": 2.304, "learning_rate": 2.278128950695322e-07, "epoch": 3.8951911220715165 }, { "current_steps": 3159, "loss": 2.3352, "learning_rate": 2.2756005056890013e-07, "epoch": 3.896424167694205 }, { "current_steps": 3160, "loss": 2.1256, "learning_rate": 2.2730720606826801e-07, "epoch": 3.8976572133168927 }, { "current_steps": 3161, "loss": 1.5292, "learning_rate": 2.2705436156763588e-07, "epoch": 3.8988902589395806 }, { "current_steps": 3162, "loss": 1.6856, "learning_rate": 2.268015170670038e-07, "epoch": 3.900123304562269 }, { "current_steps": 3163, "loss": 2.1623, "learning_rate": 2.2654867256637168e-07, "epoch": 3.901356350184957 }, { "current_steps": 3164, "loss": 2.3511, "learning_rate": 2.2629582806573954e-07, "epoch": 3.9025893958076447 }, { "current_steps": 3165, "loss": 1.9253, "learning_rate": 2.2604298356510745e-07, "epoch": 3.903822441430333 }, { "current_steps": 3166, "loss": 2.2177, "learning_rate": 2.2579013906447534e-07, "epoch": 3.905055487053021 }, { "current_steps": 3167, "loss": 1.3979, "learning_rate": 2.2553729456384323e-07, "epoch": 3.906288532675709 }, { "current_steps": 3168, "loss": 2.2405, "learning_rate": 2.2528445006321112e-07, "epoch": 3.907521578298397 }, { "current_steps": 3169, "loss": 2.2524, "learning_rate": 2.25031605562579e-07, "epoch": 3.908754623921085 }, { "current_steps": 3170, "loss": 1.8122, "learning_rate": 2.247787610619469e-07, "epoch": 3.909987669543773 }, { "current_steps": 3171, "loss": 2.0754, "learning_rate": 2.2452591656131478e-07, "epoch": 3.9112207151664613 }, { "current_steps": 3172, "loss": 2.0036, "learning_rate": 2.2427307206068267e-07, "epoch": 3.912453760789149 }, { "current_steps": 3173, "loss": 2.1245, "learning_rate": 2.2402022756005055e-07, "epoch": 3.913686806411837 }, { "current_steps": 3174, "loss": 2.1026, "learning_rate": 2.2376738305941844e-07, "epoch": 3.9149198520345254 }, { "current_steps": 3175, "loss": 2.2217, "learning_rate": 2.2351453855878633e-07, "epoch": 3.9161528976572133 }, { "current_steps": 3176, "loss": 2.1182, "learning_rate": 2.2326169405815424e-07, "epoch": 3.917385943279901 }, { "current_steps": 3177, "loss": 2.0469, "learning_rate": 2.230088495575221e-07, "epoch": 3.9186189889025895 }, { "current_steps": 3178, "loss": 1.9414, "learning_rate": 2.2275600505689e-07, "epoch": 3.9198520345252774 }, { "current_steps": 3179, "loss": 2.1452, "learning_rate": 2.225031605562579e-07, "epoch": 3.9210850801479653 }, { "current_steps": 3180, "loss": 2.1878, "learning_rate": 2.2225031605562577e-07, "epoch": 3.9223181257706536 }, { "current_steps": 3181, "loss": 2.1082, "learning_rate": 2.2199747155499368e-07, "epoch": 3.9235511713933415 }, { "current_steps": 3182, "loss": 2.3135, "learning_rate": 2.2174462705436157e-07, "epoch": 3.92478421701603 }, { "current_steps": 3183, "loss": 2.0261, "learning_rate": 2.2149178255372943e-07, "epoch": 3.9260172626387178 }, { "current_steps": 3184, "loss": 1.8155, "learning_rate": 2.2123893805309735e-07, "epoch": 3.9272503082614056 }, { "current_steps": 3185, "loss": 1.8243, "learning_rate": 2.2098609355246523e-07, "epoch": 3.9284833538840935 }, { "current_steps": 3186, "loss": 1.9905, "learning_rate": 2.207332490518331e-07, "epoch": 3.929716399506782 }, { "current_steps": 3187, "loss": 1.9208, "learning_rate": 2.20480404551201e-07, "epoch": 3.9309494451294698 }, { "current_steps": 3188, "loss": 1.6438, "learning_rate": 2.202275600505689e-07, "epoch": 3.932182490752158 }, { "current_steps": 3189, "loss": 1.7797, "learning_rate": 2.1997471554993676e-07, "epoch": 3.933415536374846 }, { "current_steps": 3190, "loss": 1.8852, "learning_rate": 2.1972187104930467e-07, "epoch": 3.934648581997534 }, { "current_steps": 3191, "loss": 1.9491, "learning_rate": 2.1946902654867256e-07, "epoch": 3.9358816276202218 }, { "current_steps": 3192, "loss": 2.0043, "learning_rate": 2.1921618204804042e-07, "epoch": 3.93711467324291 }, { "current_steps": 3193, "loss": 1.4225, "learning_rate": 2.1896333754740833e-07, "epoch": 3.938347718865598 }, { "current_steps": 3194, "loss": 2.5717, "learning_rate": 2.1871049304677622e-07, "epoch": 3.9395807644882863 }, { "current_steps": 3195, "loss": 2.1663, "learning_rate": 2.1845764854614414e-07, "epoch": 3.940813810110974 }, { "current_steps": 3196, "loss": 1.6714, "learning_rate": 2.18204804045512e-07, "epoch": 3.942046855733662 }, { "current_steps": 3197, "loss": 2.0882, "learning_rate": 2.1795195954487989e-07, "epoch": 3.94327990135635 }, { "current_steps": 3198, "loss": 2.2412, "learning_rate": 2.176991150442478e-07, "epoch": 3.9445129469790383 }, { "current_steps": 3199, "loss": 2.3644, "learning_rate": 2.1744627054361566e-07, "epoch": 3.945745992601726 }, { "current_steps": 3200, "loss": 1.9366, "learning_rate": 2.1719342604298355e-07, "epoch": 3.9469790382244145 }, { "current_steps": 3201, "loss": 1.7032, "learning_rate": 2.1694058154235146e-07, "epoch": 3.9482120838471024 }, { "current_steps": 3202, "loss": 1.959, "learning_rate": 2.1668773704171932e-07, "epoch": 3.9494451294697903 }, { "current_steps": 3203, "loss": 2.0876, "learning_rate": 2.164348925410872e-07, "epoch": 3.950678175092478 }, { "current_steps": 3204, "loss": 2.5878, "learning_rate": 2.1618204804045513e-07, "epoch": 3.9519112207151665 }, { "current_steps": 3205, "loss": 2.2309, "learning_rate": 2.15929203539823e-07, "epoch": 3.9531442663378544 }, { "current_steps": 3206, "loss": 1.9482, "learning_rate": 2.1567635903919087e-07, "epoch": 3.9543773119605428 }, { "current_steps": 3207, "loss": 1.7846, "learning_rate": 2.154235145385588e-07, "epoch": 3.9556103575832307 }, { "current_steps": 3208, "loss": 2.499, "learning_rate": 2.1517067003792665e-07, "epoch": 3.9568434032059185 }, { "current_steps": 3209, "loss": 2.0684, "learning_rate": 2.1491782553729456e-07, "epoch": 3.9580764488286064 }, { "current_steps": 3210, "loss": 2.4828, "learning_rate": 2.1466498103666245e-07, "epoch": 3.9593094944512948 }, { "current_steps": 3211, "loss": 1.9821, "learning_rate": 2.1441213653603031e-07, "epoch": 3.9605425400739827 }, { "current_steps": 3212, "loss": 1.9505, "learning_rate": 2.1415929203539823e-07, "epoch": 3.961775585696671 }, { "current_steps": 3213, "loss": 2.1242, "learning_rate": 2.1390644753476612e-07, "epoch": 3.963008631319359 }, { "current_steps": 3214, "loss": 1.5481, "learning_rate": 2.1365360303413398e-07, "epoch": 3.9642416769420468 }, { "current_steps": 3215, "loss": 2.0993, "learning_rate": 2.134007585335019e-07, "epoch": 3.9654747225647347 }, { "current_steps": 3216, "loss": 2.3977, "learning_rate": 2.1314791403286978e-07, "epoch": 3.966707768187423 }, { "current_steps": 3217, "loss": 2.3193, "learning_rate": 2.1289506953223767e-07, "epoch": 3.967940813810111 }, { "current_steps": 3218, "loss": 2.3612, "learning_rate": 2.1264222503160555e-07, "epoch": 3.969173859432799 }, { "current_steps": 3219, "loss": 1.9405, "learning_rate": 2.1238938053097344e-07, "epoch": 3.970406905055487 }, { "current_steps": 3220, "loss": 2.1897, "learning_rate": 2.1213653603034133e-07, "epoch": 3.971639950678175 }, { "current_steps": 3221, "loss": 2.1109, "learning_rate": 2.1188369152970922e-07, "epoch": 3.972872996300863 }, { "current_steps": 3222, "loss": 1.8543, "learning_rate": 2.116308470290771e-07, "epoch": 3.974106041923551 }, { "current_steps": 3223, "loss": 2.0379, "learning_rate": 2.1137800252844502e-07, "epoch": 3.975339087546239 }, { "current_steps": 3224, "loss": 1.9902, "learning_rate": 2.1112515802781288e-07, "epoch": 3.9765721331689274 }, { "current_steps": 3225, "loss": 2.3471, "learning_rate": 2.1087231352718077e-07, "epoch": 3.9778051787916153 }, { "current_steps": 3226, "loss": 1.8177, "learning_rate": 2.1061946902654868e-07, "epoch": 3.979038224414303 }, { "current_steps": 3227, "loss": 2.3397, "learning_rate": 2.1036662452591654e-07, "epoch": 3.980271270036991 }, { "current_steps": 3228, "loss": 2.2035, "learning_rate": 2.1011378002528443e-07, "epoch": 3.9815043156596794 }, { "current_steps": 3229, "loss": 2.283, "learning_rate": 2.0986093552465234e-07, "epoch": 3.9827373612823673 }, { "current_steps": 3230, "loss": 2.0067, "learning_rate": 2.096080910240202e-07, "epoch": 3.9839704069050557 }, { "current_steps": 3231, "loss": 2.0915, "learning_rate": 2.093552465233881e-07, "epoch": 3.9852034525277436 }, { "current_steps": 3232, "loss": 2.3285, "learning_rate": 2.09102402022756e-07, "epoch": 3.9864364981504314 }, { "current_steps": 3233, "loss": 2.0973, "learning_rate": 2.0884955752212387e-07, "epoch": 3.9876695437731193 }, { "current_steps": 3234, "loss": 2.0882, "learning_rate": 2.0859671302149176e-07, "epoch": 3.9889025893958077 }, { "current_steps": 3235, "loss": 1.055, "learning_rate": 2.0834386852085967e-07, "epoch": 3.9901356350184956 }, { "current_steps": 3236, "loss": 1.5889, "learning_rate": 2.0809102402022756e-07, "epoch": 3.991368680641184 }, { "current_steps": 3237, "loss": 1.7908, "learning_rate": 2.0783817951959545e-07, "epoch": 3.9926017262638718 }, { "current_steps": 3238, "loss": 2.2963, "learning_rate": 2.0758533501896333e-07, "epoch": 3.9938347718865597 }, { "current_steps": 3239, "loss": 2.6245, "learning_rate": 2.0733249051833122e-07, "epoch": 3.995067817509248 }, { "current_steps": 3240, "loss": 1.549, "learning_rate": 2.070796460176991e-07, "epoch": 3.996300863131936 }, { "current_steps": 3241, "loss": 2.0879, "learning_rate": 2.06826801517067e-07, "epoch": 3.9975339087546238 }, { "current_steps": 3242, "loss": 1.7907, "learning_rate": 2.0657395701643488e-07, "epoch": 3.998766954377312 }, { "current_steps": 3243, "loss": 1.448, "learning_rate": 2.0632111251580277e-07, "epoch": 4.0 }, { "current_steps": 3244, "loss": 2.2061, "learning_rate": 2.0606826801517066e-07, "epoch": 4.001233045622688 }, { "current_steps": 3245, "loss": 2.5205, "learning_rate": 2.0581542351453855e-07, "epoch": 4.002466091245376 }, { "current_steps": 3246, "loss": 2.074, "learning_rate": 2.0556257901390644e-07, "epoch": 4.003699136868065 }, { "current_steps": 3247, "loss": 1.8953, "learning_rate": 2.0530973451327432e-07, "epoch": 4.0049321824907524 }, { "current_steps": 3248, "loss": 2.3133, "learning_rate": 2.050568900126422e-07, "epoch": 4.00616522811344 }, { "current_steps": 3249, "loss": 1.9247, "learning_rate": 2.048040455120101e-07, "epoch": 4.007398273736128 }, { "current_steps": 3250, "loss": 2.2624, "learning_rate": 2.0455120101137799e-07, "epoch": 4.008631319358816 }, { "current_steps": 3251, "loss": 2.1333, "learning_rate": 2.042983565107459e-07, "epoch": 4.009864364981504 }, { "current_steps": 3252, "loss": 2.5553, "learning_rate": 2.0404551201011376e-07, "epoch": 4.011097410604193 }, { "current_steps": 3253, "loss": 1.9109, "learning_rate": 2.0379266750948165e-07, "epoch": 4.012330456226881 }, { "current_steps": 3254, "loss": 2.3846, "learning_rate": 2.0353982300884956e-07, "epoch": 4.013563501849569 }, { "current_steps": 3255, "loss": 2.0634, "learning_rate": 2.0328697850821745e-07, "epoch": 4.0147965474722564 }, { "current_steps": 3256, "loss": 2.2491, "learning_rate": 2.030341340075853e-07, "epoch": 4.016029593094944 }, { "current_steps": 3257, "loss": 2.3906, "learning_rate": 2.0278128950695323e-07, "epoch": 4.017262638717632 }, { "current_steps": 3258, "loss": 2.2435, "learning_rate": 2.0252844500632111e-07, "epoch": 4.018495684340321 }, { "current_steps": 3259, "loss": 2.3092, "learning_rate": 2.0227560050568898e-07, "epoch": 4.019728729963009 }, { "current_steps": 3260, "loss": 1.9533, "learning_rate": 2.020227560050569e-07, "epoch": 4.020961775585697 }, { "current_steps": 3261, "loss": 1.8924, "learning_rate": 2.0176991150442478e-07, "epoch": 4.022194821208385 }, { "current_steps": 3262, "loss": 1.9712, "learning_rate": 2.0151706700379264e-07, "epoch": 4.023427866831073 }, { "current_steps": 3263, "loss": 2.0631, "learning_rate": 2.0126422250316055e-07, "epoch": 4.0246609124537605 }, { "current_steps": 3264, "loss": 2.0438, "learning_rate": 2.0101137800252844e-07, "epoch": 4.025893958076449 }, { "current_steps": 3265, "loss": 1.8177, "learning_rate": 2.0075853350189633e-07, "epoch": 4.027127003699137 }, { "current_steps": 3266, "loss": 2.3215, "learning_rate": 2.0050568900126422e-07, "epoch": 4.028360049321825 }, { "current_steps": 3267, "loss": 2.3141, "learning_rate": 2.002528445006321e-07, "epoch": 4.029593094944513 }, { "current_steps": 3268, "loss": 2.0866, "learning_rate": 2e-07, "epoch": 4.030826140567201 }, { "current_steps": 3269, "loss": 2.1543, "learning_rate": 1.9974715549936788e-07, "epoch": 4.032059186189889 }, { "current_steps": 3270, "loss": 2.057, "learning_rate": 1.9949431099873577e-07, "epoch": 4.0332922318125775 }, { "current_steps": 3271, "loss": 2.2401, "learning_rate": 1.9924146649810365e-07, "epoch": 4.034525277435265 }, { "current_steps": 3272, "loss": 2.0161, "learning_rate": 1.9898862199747154e-07, "epoch": 4.035758323057953 }, { "current_steps": 3273, "loss": 2.0601, "learning_rate": 1.9873577749683943e-07, "epoch": 4.036991368680641 }, { "current_steps": 3274, "loss": 2.2206, "learning_rate": 1.9848293299620734e-07, "epoch": 4.038224414303329 }, { "current_steps": 3275, "loss": 2.0045, "learning_rate": 1.982300884955752e-07, "epoch": 4.039457459926017 }, { "current_steps": 3276, "loss": 1.3848, "learning_rate": 1.979772439949431e-07, "epoch": 4.040690505548706 }, { "current_steps": 3277, "loss": 2.0974, "learning_rate": 1.97724399494311e-07, "epoch": 4.041923551171394 }, { "current_steps": 3278, "loss": 2.1263, "learning_rate": 1.9747155499367887e-07, "epoch": 4.0431565967940815 }, { "current_steps": 3279, "loss": 2.1777, "learning_rate": 1.9721871049304678e-07, "epoch": 4.044389642416769 }, { "current_steps": 3280, "loss": 2.2958, "learning_rate": 1.9696586599241467e-07, "epoch": 4.045622688039457 }, { "current_steps": 3281, "loss": 1.8407, "learning_rate": 1.9671302149178253e-07, "epoch": 4.046855733662145 }, { "current_steps": 3282, "loss": 1.6308, "learning_rate": 1.9646017699115045e-07, "epoch": 4.048088779284834 }, { "current_steps": 3283, "loss": 2.2494, "learning_rate": 1.9620733249051833e-07, "epoch": 4.049321824907522 }, { "current_steps": 3284, "loss": 1.8955, "learning_rate": 1.959544879898862e-07, "epoch": 4.05055487053021 }, { "current_steps": 3285, "loss": 2.4174, "learning_rate": 1.957016434892541e-07, "epoch": 4.051787916152898 }, { "current_steps": 3286, "loss": 2.0034, "learning_rate": 1.95448798988622e-07, "epoch": 4.0530209617755855 }, { "current_steps": 3287, "loss": 2.0505, "learning_rate": 1.9519595448798986e-07, "epoch": 4.054254007398273 }, { "current_steps": 3288, "loss": 2.0292, "learning_rate": 1.9494310998735777e-07, "epoch": 4.055487053020962 }, { "current_steps": 3289, "loss": 2.0695, "learning_rate": 1.9469026548672566e-07, "epoch": 4.05672009864365 }, { "current_steps": 3290, "loss": 1.9686, "learning_rate": 1.9443742098609352e-07, "epoch": 4.057953144266338 }, { "current_steps": 3291, "loss": 2.1853, "learning_rate": 1.9418457648546143e-07, "epoch": 4.059186189889026 }, { "current_steps": 3292, "loss": 1.8407, "learning_rate": 1.9393173198482932e-07, "epoch": 4.060419235511714 }, { "current_steps": 3293, "loss": 2.3223, "learning_rate": 1.9367888748419724e-07, "epoch": 4.061652281134402 }, { "current_steps": 3294, "loss": 2.3798, "learning_rate": 1.934260429835651e-07, "epoch": 4.06288532675709 }, { "current_steps": 3295, "loss": 2.3244, "learning_rate": 1.9317319848293299e-07, "epoch": 4.064118372379778 }, { "current_steps": 3296, "loss": 2.0546, "learning_rate": 1.929203539823009e-07, "epoch": 4.065351418002466 }, { "current_steps": 3297, "loss": 2.2194, "learning_rate": 1.9266750948166876e-07, "epoch": 4.066584463625154 }, { "current_steps": 3298, "loss": 2.5809, "learning_rate": 1.9241466498103665e-07, "epoch": 4.067817509247842 }, { "current_steps": 3299, "loss": 2.2397, "learning_rate": 1.9216182048040456e-07, "epoch": 4.06905055487053 }, { "current_steps": 3300, "loss": 2.0184, "learning_rate": 1.9190897597977242e-07, "epoch": 4.070283600493219 }, { "current_steps": 3301, "loss": 2.1658, "learning_rate": 1.916561314791403e-07, "epoch": 4.0715166461159065 }, { "current_steps": 3302, "loss": 2.204, "learning_rate": 1.9140328697850823e-07, "epoch": 4.072749691738594 }, { "current_steps": 3303, "loss": 1.7097, "learning_rate": 1.9115044247787609e-07, "epoch": 4.073982737361282 }, { "current_steps": 3304, "loss": 2.2087, "learning_rate": 1.9089759797724397e-07, "epoch": 4.07521578298397 }, { "current_steps": 3305, "loss": 1.5249, "learning_rate": 1.906447534766119e-07, "epoch": 4.076448828606658 }, { "current_steps": 3306, "loss": 1.8921, "learning_rate": 1.9039190897597975e-07, "epoch": 4.077681874229347 }, { "current_steps": 3307, "loss": 2.4002, "learning_rate": 1.9013906447534766e-07, "epoch": 4.078914919852035 }, { "current_steps": 3308, "loss": 1.8627, "learning_rate": 1.8988621997471555e-07, "epoch": 4.080147965474723 }, { "current_steps": 3309, "loss": 1.9099, "learning_rate": 1.896333754740834e-07, "epoch": 4.0813810110974105 }, { "current_steps": 3310, "loss": 1.7044, "learning_rate": 1.8938053097345133e-07, "epoch": 4.082614056720098 }, { "current_steps": 3311, "loss": 1.9024, "learning_rate": 1.8912768647281921e-07, "epoch": 4.083847102342786 }, { "current_steps": 3312, "loss": 2.3876, "learning_rate": 1.8887484197218708e-07, "epoch": 4.085080147965475 }, { "current_steps": 3313, "loss": 2.2965, "learning_rate": 1.88621997471555e-07, "epoch": 4.086313193588163 }, { "current_steps": 3314, "loss": 2.3701, "learning_rate": 1.8836915297092288e-07, "epoch": 4.087546239210851 }, { "current_steps": 3315, "loss": 1.8157, "learning_rate": 1.8811630847029077e-07, "epoch": 4.088779284833539 }, { "current_steps": 3316, "loss": 1.3549, "learning_rate": 1.8786346396965865e-07, "epoch": 4.090012330456227 }, { "current_steps": 3317, "loss": 1.9633, "learning_rate": 1.8761061946902654e-07, "epoch": 4.0912453760789145 }, { "current_steps": 3318, "loss": 2.0616, "learning_rate": 1.8735777496839443e-07, "epoch": 4.092478421701603 }, { "current_steps": 3319, "loss": 2.3666, "learning_rate": 1.8710493046776232e-07, "epoch": 4.093711467324291 }, { "current_steps": 3320, "loss": 2.0625, "learning_rate": 1.868520859671302e-07, "epoch": 4.094944512946979 }, { "current_steps": 3321, "loss": 2.0456, "learning_rate": 1.8659924146649812e-07, "epoch": 4.096177558569667 }, { "current_steps": 3322, "loss": 1.7828, "learning_rate": 1.8634639696586598e-07, "epoch": 4.097410604192355 }, { "current_steps": 3323, "loss": 1.9216, "learning_rate": 1.8609355246523387e-07, "epoch": 4.098643649815044 }, { "current_steps": 3324, "loss": 2.1164, "learning_rate": 1.8584070796460178e-07, "epoch": 4.0998766954377315 }, { "current_steps": 3325, "loss": 1.9039, "learning_rate": 1.8558786346396964e-07, "epoch": 4.101109741060419 }, { "current_steps": 3326, "loss": 2.1535, "learning_rate": 1.8533501896333753e-07, "epoch": 4.102342786683107 }, { "current_steps": 3327, "loss": 1.626, "learning_rate": 1.8508217446270544e-07, "epoch": 4.103575832305795 }, { "current_steps": 3328, "loss": 1.8229, "learning_rate": 1.848293299620733e-07, "epoch": 4.104808877928483 }, { "current_steps": 3329, "loss": 2.1861, "learning_rate": 1.845764854614412e-07, "epoch": 4.106041923551172 }, { "current_steps": 3330, "loss": 2.136, "learning_rate": 1.843236409608091e-07, "epoch": 4.10727496917386 }, { "current_steps": 3331, "loss": 2.2443, "learning_rate": 1.8407079646017697e-07, "epoch": 4.108508014796548 }, { "current_steps": 3332, "loss": 1.6949, "learning_rate": 1.8381795195954486e-07, "epoch": 4.1097410604192355 }, { "current_steps": 3333, "loss": 2.1955, "learning_rate": 1.8356510745891277e-07, "epoch": 4.110974106041923 }, { "current_steps": 3334, "loss": 2.2717, "learning_rate": 1.8331226295828063e-07, "epoch": 4.112207151664611 }, { "current_steps": 3335, "loss": 2.1329, "learning_rate": 1.8305941845764855e-07, "epoch": 4.1134401972873 }, { "current_steps": 3336, "loss": 2.0469, "learning_rate": 1.8280657395701643e-07, "epoch": 4.114673242909988 }, { "current_steps": 3337, "loss": 1.8504, "learning_rate": 1.8255372945638432e-07, "epoch": 4.115906288532676 }, { "current_steps": 3338, "loss": 1.9896, "learning_rate": 1.823008849557522e-07, "epoch": 4.117139334155364 }, { "current_steps": 3339, "loss": 2.1812, "learning_rate": 1.820480404551201e-07, "epoch": 4.118372379778052 }, { "current_steps": 3340, "loss": 2.1421, "learning_rate": 1.8179519595448798e-07, "epoch": 4.1196054254007395 }, { "current_steps": 3341, "loss": 2.1325, "learning_rate": 1.8154235145385587e-07, "epoch": 4.120838471023428 }, { "current_steps": 3342, "loss": 2.4279, "learning_rate": 1.8128950695322376e-07, "epoch": 4.122071516646116 }, { "current_steps": 3343, "loss": 2.2692, "learning_rate": 1.8103666245259165e-07, "epoch": 4.123304562268804 }, { "current_steps": 3344, "loss": 1.874, "learning_rate": 1.8078381795195953e-07, "epoch": 4.124537607891492 }, { "current_steps": 3345, "loss": 1.8635, "learning_rate": 1.8053097345132742e-07, "epoch": 4.12577065351418 }, { "current_steps": 3346, "loss": 1.6178, "learning_rate": 1.802781289506953e-07, "epoch": 4.127003699136868 }, { "current_steps": 3347, "loss": 2.5118, "learning_rate": 1.800252844500632e-07, "epoch": 4.1282367447595565 }, { "current_steps": 3348, "loss": 1.7178, "learning_rate": 1.7977243994943109e-07, "epoch": 4.129469790382244 }, { "current_steps": 3349, "loss": 1.9431, "learning_rate": 1.79519595448799e-07, "epoch": 4.130702836004932 }, { "current_steps": 3350, "loss": 1.3608, "learning_rate": 1.7926675094816686e-07, "epoch": 4.13193588162762 }, { "current_steps": 3351, "loss": 1.5689, "learning_rate": 1.7901390644753475e-07, "epoch": 4.133168927250308 }, { "current_steps": 3352, "loss": 2.1493, "learning_rate": 1.7876106194690266e-07, "epoch": 4.134401972872996 }, { "current_steps": 3353, "loss": 2.0759, "learning_rate": 1.7850821744627052e-07, "epoch": 4.135635018495685 }, { "current_steps": 3354, "loss": 1.9349, "learning_rate": 1.782553729456384e-07, "epoch": 4.136868064118373 }, { "current_steps": 3355, "loss": 2.0979, "learning_rate": 1.7800252844500633e-07, "epoch": 4.1381011097410605 }, { "current_steps": 3356, "loss": 2.0873, "learning_rate": 1.7774968394437421e-07, "epoch": 4.139334155363748 }, { "current_steps": 3357, "loss": 1.9369, "learning_rate": 1.7749683944374207e-07, "epoch": 4.140567200986436 }, { "current_steps": 3358, "loss": 1.8813, "learning_rate": 1.7724399494311e-07, "epoch": 4.141800246609124 }, { "current_steps": 3359, "loss": 1.5802, "learning_rate": 1.7699115044247788e-07, "epoch": 4.143033292231813 }, { "current_steps": 3360, "loss": 2.2198, "learning_rate": 1.7673830594184574e-07, "epoch": 4.144266337854501 }, { "current_steps": 3361, "loss": 2.3963, "learning_rate": 1.7648546144121365e-07, "epoch": 4.145499383477189 }, { "current_steps": 3362, "loss": 1.6583, "learning_rate": 1.7623261694058154e-07, "epoch": 4.146732429099877 }, { "current_steps": 3363, "loss": 1.929, "learning_rate": 1.7597977243994943e-07, "epoch": 4.1479654747225645 }, { "current_steps": 3364, "loss": 2.2576, "learning_rate": 1.7572692793931731e-07, "epoch": 4.149198520345252 }, { "current_steps": 3365, "loss": 1.7606, "learning_rate": 1.754740834386852e-07, "epoch": 4.150431565967941 }, { "current_steps": 3366, "loss": 1.687, "learning_rate": 1.752212389380531e-07, "epoch": 4.151664611590629 }, { "current_steps": 3367, "loss": 1.9745, "learning_rate": 1.7496839443742098e-07, "epoch": 4.152897657213317 }, { "current_steps": 3368, "loss": 1.7951, "learning_rate": 1.7471554993678887e-07, "epoch": 4.154130702836005 }, { "current_steps": 3369, "loss": 2.1376, "learning_rate": 1.7446270543615675e-07, "epoch": 4.155363748458693 }, { "current_steps": 3370, "loss": 2.1982, "learning_rate": 1.7420986093552464e-07, "epoch": 4.156596794081381 }, { "current_steps": 3371, "loss": 2.2098, "learning_rate": 1.7395701643489253e-07, "epoch": 4.157829839704069 }, { "current_steps": 3372, "loss": 1.9277, "learning_rate": 1.7370417193426042e-07, "epoch": 4.159062885326757 }, { "current_steps": 3373, "loss": 2.0448, "learning_rate": 1.734513274336283e-07, "epoch": 4.160295930949445 }, { "current_steps": 3374, "loss": 2.6372, "learning_rate": 1.731984829329962e-07, "epoch": 4.161528976572133 }, { "current_steps": 3375, "loss": 1.7347, "learning_rate": 1.729456384323641e-07, "epoch": 4.162762022194821 }, { "current_steps": 3376, "loss": 2.3025, "learning_rate": 1.7269279393173197e-07, "epoch": 4.163995067817509 }, { "current_steps": 3377, "loss": 1.7546, "learning_rate": 1.7243994943109988e-07, "epoch": 4.165228113440198 }, { "current_steps": 3378, "loss": 2.0639, "learning_rate": 1.7218710493046777e-07, "epoch": 4.1664611590628855 }, { "current_steps": 3379, "loss": 2.5274, "learning_rate": 1.7193426042983563e-07, "epoch": 4.167694204685573 }, { "current_steps": 3380, "loss": 2.099, "learning_rate": 1.7168141592920354e-07, "epoch": 4.168927250308261 }, { "current_steps": 3381, "loss": 2.2031, "learning_rate": 1.7142857142857143e-07, "epoch": 4.170160295930949 }, { "current_steps": 3382, "loss": 2.1192, "learning_rate": 1.711757269279393e-07, "epoch": 4.171393341553637 }, { "current_steps": 3383, "loss": 1.8766, "learning_rate": 1.709228824273072e-07, "epoch": 4.172626387176326 }, { "current_steps": 3384, "loss": 2.1862, "learning_rate": 1.706700379266751e-07, "epoch": 4.173859432799014 }, { "current_steps": 3385, "loss": 1.8533, "learning_rate": 1.7041719342604296e-07, "epoch": 4.175092478421702 }, { "current_steps": 3386, "loss": 2.4115, "learning_rate": 1.7016434892541087e-07, "epoch": 4.1763255240443895 }, { "current_steps": 3387, "loss": 1.6681, "learning_rate": 1.6991150442477876e-07, "epoch": 4.177558569667077 }, { "current_steps": 3388, "loss": 2.0592, "learning_rate": 1.6965865992414662e-07, "epoch": 4.178791615289766 }, { "current_steps": 3389, "loss": 1.774, "learning_rate": 1.6940581542351453e-07, "epoch": 4.180024660912454 }, { "current_steps": 3390, "loss": 2.2347, "learning_rate": 1.6915297092288242e-07, "epoch": 4.181257706535142 }, { "current_steps": 3391, "loss": 1.8617, "learning_rate": 1.689001264222503e-07, "epoch": 4.18249075215783 }, { "current_steps": 3392, "loss": 2.3214, "learning_rate": 1.686472819216182e-07, "epoch": 4.183723797780518 }, { "current_steps": 3393, "loss": 1.5788, "learning_rate": 1.6839443742098608e-07, "epoch": 4.184956843403206 }, { "current_steps": 3394, "loss": 2.1122, "learning_rate": 1.68141592920354e-07, "epoch": 4.1861898890258935 }, { "current_steps": 3395, "loss": 2.4202, "learning_rate": 1.6788874841972186e-07, "epoch": 4.187422934648582 }, { "current_steps": 3396, "loss": 1.9667, "learning_rate": 1.6763590391908975e-07, "epoch": 4.18865598027127 }, { "current_steps": 3397, "loss": 1.8933, "learning_rate": 1.6738305941845766e-07, "epoch": 4.189889025893958 }, { "current_steps": 3398, "loss": 1.8612, "learning_rate": 1.6713021491782552e-07, "epoch": 4.191122071516646 }, { "current_steps": 3399, "loss": 2.2643, "learning_rate": 1.668773704171934e-07, "epoch": 4.192355117139334 }, { "current_steps": 3400, "loss": 2.2092, "learning_rate": 1.6662452591656132e-07, "epoch": 4.193588162762023 }, { "current_steps": 3401, "loss": 2.4275, "learning_rate": 1.6637168141592919e-07, "epoch": 4.1948212083847105 }, { "current_steps": 3402, "loss": 2.3613, "learning_rate": 1.6611883691529707e-07, "epoch": 4.196054254007398 }, { "current_steps": 3403, "loss": 2.4613, "learning_rate": 1.65865992414665e-07, "epoch": 4.197287299630086 }, { "current_steps": 3404, "loss": 2.5314, "learning_rate": 1.6561314791403285e-07, "epoch": 4.198520345252774 }, { "current_steps": 3405, "loss": 1.9192, "learning_rate": 1.6536030341340076e-07, "epoch": 4.199753390875462 }, { "current_steps": 3406, "loss": 1.6006, "learning_rate": 1.6510745891276865e-07, "epoch": 4.200986436498151 }, { "current_steps": 3407, "loss": 2.3218, "learning_rate": 1.648546144121365e-07, "epoch": 4.202219482120839 }, { "current_steps": 3408, "loss": 2.1093, "learning_rate": 1.6460176991150443e-07, "epoch": 4.203452527743527 }, { "current_steps": 3409, "loss": 1.9749, "learning_rate": 1.6434892541087231e-07, "epoch": 4.2046855733662145 }, { "current_steps": 3410, "loss": 1.9712, "learning_rate": 1.6409608091024018e-07, "epoch": 4.205918618988902 }, { "current_steps": 3411, "loss": 2.1781, "learning_rate": 1.638432364096081e-07, "epoch": 4.20715166461159 }, { "current_steps": 3412, "loss": 1.8543, "learning_rate": 1.6359039190897598e-07, "epoch": 4.208384710234279 }, { "current_steps": 3413, "loss": 1.7558, "learning_rate": 1.6333754740834384e-07, "epoch": 4.209617755856967 }, { "current_steps": 3414, "loss": 2.0034, "learning_rate": 1.6308470290771175e-07, "epoch": 4.210850801479655 }, { "current_steps": 3415, "loss": 1.9622, "learning_rate": 1.6283185840707964e-07, "epoch": 4.212083847102343 }, { "current_steps": 3416, "loss": 2.0299, "learning_rate": 1.6257901390644753e-07, "epoch": 4.213316892725031 }, { "current_steps": 3417, "loss": 2.0484, "learning_rate": 1.6232616940581542e-07, "epoch": 4.2145499383477185 }, { "current_steps": 3418, "loss": 2.0148, "learning_rate": 1.620733249051833e-07, "epoch": 4.215782983970407 }, { "current_steps": 3419, "loss": 2.2633, "learning_rate": 1.6182048040455122e-07, "epoch": 4.217016029593095 }, { "current_steps": 3420, "loss": 1.572, "learning_rate": 1.6156763590391908e-07, "epoch": 4.218249075215783 }, { "current_steps": 3421, "loss": 2.1987, "learning_rate": 1.6131479140328697e-07, "epoch": 4.219482120838471 }, { "current_steps": 3422, "loss": 1.9874, "learning_rate": 1.6106194690265488e-07, "epoch": 4.220715166461159 }, { "current_steps": 3423, "loss": 2.2804, "learning_rate": 1.6080910240202274e-07, "epoch": 4.221948212083847 }, { "current_steps": 3424, "loss": 1.9043, "learning_rate": 1.6055625790139063e-07, "epoch": 4.2231812577065355 }, { "current_steps": 3425, "loss": 2.3374, "learning_rate": 1.6030341340075854e-07, "epoch": 4.224414303329223 }, { "current_steps": 3426, "loss": 2.0677, "learning_rate": 1.600505689001264e-07, "epoch": 4.225647348951911 }, { "current_steps": 3427, "loss": 1.8952, "learning_rate": 1.597977243994943e-07, "epoch": 4.226880394574599 }, { "current_steps": 3428, "loss": 2.2047, "learning_rate": 1.595448798988622e-07, "epoch": 4.228113440197287 }, { "current_steps": 3429, "loss": 2.1382, "learning_rate": 1.5929203539823007e-07, "epoch": 4.229346485819975 }, { "current_steps": 3430, "loss": 1.6612, "learning_rate": 1.5903919089759796e-07, "epoch": 4.230579531442664 }, { "current_steps": 3431, "loss": 1.9522, "learning_rate": 1.5878634639696587e-07, "epoch": 4.231812577065352 }, { "current_steps": 3432, "loss": 2.1836, "learning_rate": 1.5853350189633373e-07, "epoch": 4.2330456226880395 }, { "current_steps": 3433, "loss": 1.7966, "learning_rate": 1.5828065739570164e-07, "epoch": 4.234278668310727 }, { "current_steps": 3434, "loss": 2.2029, "learning_rate": 1.5802781289506953e-07, "epoch": 4.235511713933415 }, { "current_steps": 3435, "loss": 2.0571, "learning_rate": 1.5777496839443742e-07, "epoch": 4.236744759556103 }, { "current_steps": 3436, "loss": 1.8339, "learning_rate": 1.575221238938053e-07, "epoch": 4.237977805178792 }, { "current_steps": 3437, "loss": 2.2963, "learning_rate": 1.572692793931732e-07, "epoch": 4.23921085080148 }, { "current_steps": 3438, "loss": 1.841, "learning_rate": 1.5701643489254108e-07, "epoch": 4.240443896424168 }, { "current_steps": 3439, "loss": 2.4232, "learning_rate": 1.5676359039190897e-07, "epoch": 4.241676942046856 }, { "current_steps": 3440, "loss": 2.1725, "learning_rate": 1.5651074589127686e-07, "epoch": 4.2429099876695435 }, { "current_steps": 3441, "loss": 2.0846, "learning_rate": 1.5625790139064475e-07, "epoch": 4.244143033292231 }, { "current_steps": 3442, "loss": 1.3173, "learning_rate": 1.5600505689001263e-07, "epoch": 4.24537607891492 }, { "current_steps": 3443, "loss": 2.2302, "learning_rate": 1.5575221238938052e-07, "epoch": 4.246609124537608 }, { "current_steps": 3444, "loss": 1.9978, "learning_rate": 1.554993678887484e-07, "epoch": 4.247842170160296 }, { "current_steps": 3445, "loss": 2.0877, "learning_rate": 1.552465233881163e-07, "epoch": 4.249075215782984 }, { "current_steps": 3446, "loss": 1.3996, "learning_rate": 1.5499367888748418e-07, "epoch": 4.250308261405672 }, { "current_steps": 3447, "loss": 2.0328, "learning_rate": 1.547408343868521e-07, "epoch": 4.25154130702836 }, { "current_steps": 3448, "loss": 1.9857, "learning_rate": 1.5448798988621996e-07, "epoch": 4.252774352651048 }, { "current_steps": 3449, "loss": 2.3337, "learning_rate": 1.5423514538558785e-07, "epoch": 4.254007398273736 }, { "current_steps": 3450, "loss": 2.0374, "learning_rate": 1.5398230088495576e-07, "epoch": 4.255240443896424 }, { "current_steps": 3451, "loss": 2.1608, "learning_rate": 1.5372945638432362e-07, "epoch": 4.256473489519112 }, { "current_steps": 3452, "loss": 2.2582, "learning_rate": 1.534766118836915e-07, "epoch": 4.2577065351418 }, { "current_steps": 3453, "loss": 2.1522, "learning_rate": 1.5322376738305943e-07, "epoch": 4.258939580764489 }, { "current_steps": 3454, "loss": 1.9665, "learning_rate": 1.529709228824273e-07, "epoch": 4.260172626387177 }, { "current_steps": 3455, "loss": 1.9147, "learning_rate": 1.5271807838179517e-07, "epoch": 4.2614056720098645 }, { "current_steps": 3456, "loss": 2.2452, "learning_rate": 1.524652338811631e-07, "epoch": 4.262638717632552 }, { "current_steps": 3457, "loss": 2.3099, "learning_rate": 1.5221238938053098e-07, "epoch": 4.26387176325524 }, { "current_steps": 3458, "loss": 2.2773, "learning_rate": 1.5195954487989884e-07, "epoch": 4.265104808877928 }, { "current_steps": 3459, "loss": 2.0703, "learning_rate": 1.5170670037926675e-07, "epoch": 4.266337854500616 }, { "current_steps": 3460, "loss": 2.1778, "learning_rate": 1.5145385587863464e-07, "epoch": 4.267570900123305 }, { "current_steps": 3461, "loss": 2.2373, "learning_rate": 1.5120101137800253e-07, "epoch": 4.268803945745993 }, { "current_steps": 3462, "loss": 2.0902, "learning_rate": 1.5094816687737041e-07, "epoch": 4.270036991368681 }, { "current_steps": 3463, "loss": 2.1778, "learning_rate": 1.506953223767383e-07, "epoch": 4.2712700369913685 }, { "current_steps": 3464, "loss": 2.5243, "learning_rate": 1.504424778761062e-07, "epoch": 4.272503082614056 }, { "current_steps": 3465, "loss": 1.8891, "learning_rate": 1.5018963337547408e-07, "epoch": 4.273736128236745 }, { "current_steps": 3466, "loss": 1.4558, "learning_rate": 1.4993678887484197e-07, "epoch": 4.274969173859433 }, { "current_steps": 3467, "loss": 2.146, "learning_rate": 1.4968394437420985e-07, "epoch": 4.276202219482121 }, { "current_steps": 3468, "loss": 2.4157, "learning_rate": 1.4943109987357774e-07, "epoch": 4.277435265104809 }, { "current_steps": 3469, "loss": 2.3997, "learning_rate": 1.4917825537294563e-07, "epoch": 4.278668310727497 }, { "current_steps": 3470, "loss": 2.1995, "learning_rate": 1.4892541087231352e-07, "epoch": 4.279901356350185 }, { "current_steps": 3471, "loss": 2.2326, "learning_rate": 1.486725663716814e-07, "epoch": 4.2811344019728725 }, { "current_steps": 3472, "loss": 2.2214, "learning_rate": 1.484197218710493e-07, "epoch": 4.282367447595561 }, { "current_steps": 3473, "loss": 1.8677, "learning_rate": 1.481668773704172e-07, "epoch": 4.283600493218249 }, { "current_steps": 3474, "loss": 2.2103, "learning_rate": 1.4791403286978507e-07, "epoch": 4.284833538840937 }, { "current_steps": 3475, "loss": 1.7422, "learning_rate": 1.4766118836915298e-07, "epoch": 4.286066584463625 }, { "current_steps": 3476, "loss": 2.1493, "learning_rate": 1.4740834386852087e-07, "epoch": 4.287299630086313 }, { "current_steps": 3477, "loss": 1.943, "learning_rate": 1.4715549936788873e-07, "epoch": 4.288532675709002 }, { "current_steps": 3478, "loss": 2.1675, "learning_rate": 1.4690265486725664e-07, "epoch": 4.2897657213316895 }, { "current_steps": 3479, "loss": 1.6785, "learning_rate": 1.4664981036662453e-07, "epoch": 4.290998766954377 }, { "current_steps": 3480, "loss": 2.3598, "learning_rate": 1.463969658659924e-07, "epoch": 4.292231812577065 }, { "current_steps": 3481, "loss": 2.2165, "learning_rate": 1.461441213653603e-07, "epoch": 4.293464858199753 }, { "current_steps": 3482, "loss": 2.2546, "learning_rate": 1.458912768647282e-07, "epoch": 4.294697903822441 }, { "current_steps": 3483, "loss": 1.7758, "learning_rate": 1.4563843236409606e-07, "epoch": 4.29593094944513 }, { "current_steps": 3484, "loss": 1.9678, "learning_rate": 1.4538558786346397e-07, "epoch": 4.297163995067818 }, { "current_steps": 3485, "loss": 2.1409, "learning_rate": 1.4513274336283186e-07, "epoch": 4.298397040690506 }, { "current_steps": 3486, "loss": 2.117, "learning_rate": 1.4487989886219972e-07, "epoch": 4.2996300863131935 }, { "current_steps": 3487, "loss": 1.8911, "learning_rate": 1.4462705436156763e-07, "epoch": 4.300863131935881 }, { "current_steps": 3488, "loss": 1.8901, "learning_rate": 1.4437420986093552e-07, "epoch": 4.302096177558569 }, { "current_steps": 3489, "loss": 1.8232, "learning_rate": 1.441213653603034e-07, "epoch": 4.303329223181258 }, { "current_steps": 3490, "loss": 1.7785, "learning_rate": 1.438685208596713e-07, "epoch": 4.304562268803946 }, { "current_steps": 3491, "loss": 2.1592, "learning_rate": 1.4361567635903918e-07, "epoch": 4.305795314426634 }, { "current_steps": 3492, "loss": 2.0097, "learning_rate": 1.433628318584071e-07, "epoch": 4.307028360049322 }, { "current_steps": 3493, "loss": 1.9291, "learning_rate": 1.4310998735777496e-07, "epoch": 4.30826140567201 }, { "current_steps": 3494, "loss": 2.2285, "learning_rate": 1.4285714285714285e-07, "epoch": 4.3094944512946975 }, { "current_steps": 3495, "loss": 2.1016, "learning_rate": 1.4260429835651076e-07, "epoch": 4.310727496917386 }, { "current_steps": 3496, "loss": 2.4943, "learning_rate": 1.4235145385587862e-07, "epoch": 4.311960542540074 }, { "current_steps": 3497, "loss": 2.2147, "learning_rate": 1.420986093552465e-07, "epoch": 4.313193588162762 }, { "current_steps": 3498, "loss": 2.2724, "learning_rate": 1.4184576485461442e-07, "epoch": 4.31442663378545 }, { "current_steps": 3499, "loss": 2.197, "learning_rate": 1.4159292035398229e-07, "epoch": 4.315659679408138 }, { "current_steps": 3500, "loss": 1.6639, "learning_rate": 1.4134007585335017e-07, "epoch": 4.316892725030826 }, { "current_steps": 3501, "loss": 2.3508, "learning_rate": 1.410872313527181e-07, "epoch": 4.3181257706535146 }, { "current_steps": 3502, "loss": 2.1331, "learning_rate": 1.4083438685208595e-07, "epoch": 4.319358816276202 }, { "current_steps": 3503, "loss": 2.0904, "learning_rate": 1.4058154235145384e-07, "epoch": 4.32059186189889 }, { "current_steps": 3504, "loss": 2.4246, "learning_rate": 1.4032869785082175e-07, "epoch": 4.321824907521578 }, { "current_steps": 3505, "loss": 2.1964, "learning_rate": 1.400758533501896e-07, "epoch": 4.323057953144266 }, { "current_steps": 3506, "loss": 2.2055, "learning_rate": 1.3982300884955753e-07, "epoch": 4.324290998766954 }, { "current_steps": 3507, "loss": 1.807, "learning_rate": 1.3957016434892541e-07, "epoch": 4.325524044389643 }, { "current_steps": 3508, "loss": 2.1246, "learning_rate": 1.3931731984829327e-07, "epoch": 4.326757090012331 }, { "current_steps": 3509, "loss": 1.7091, "learning_rate": 1.390644753476612e-07, "epoch": 4.3279901356350186 }, { "current_steps": 3510, "loss": 1.8081, "learning_rate": 1.3881163084702908e-07, "epoch": 4.329223181257706 }, { "current_steps": 3511, "loss": 1.6546, "learning_rate": 1.3855878634639694e-07, "epoch": 4.330456226880394 }, { "current_steps": 3512, "loss": 2.0903, "learning_rate": 1.3830594184576485e-07, "epoch": 4.331689272503082 }, { "current_steps": 3513, "loss": 1.795, "learning_rate": 1.3805309734513274e-07, "epoch": 4.332922318125771 }, { "current_steps": 3514, "loss": 2.1103, "learning_rate": 1.3780025284450063e-07, "epoch": 4.334155363748459 }, { "current_steps": 3515, "loss": 2.1073, "learning_rate": 1.3754740834386851e-07, "epoch": 4.335388409371147 }, { "current_steps": 3516, "loss": 1.9917, "learning_rate": 1.372945638432364e-07, "epoch": 4.336621454993835 }, { "current_steps": 3517, "loss": 1.8741, "learning_rate": 1.370417193426043e-07, "epoch": 4.337854500616523 }, { "current_steps": 3518, "loss": 2.129, "learning_rate": 1.3678887484197218e-07, "epoch": 4.3390875462392104 }, { "current_steps": 3519, "loss": 1.7563, "learning_rate": 1.3653603034134007e-07, "epoch": 4.340320591861899 }, { "current_steps": 3520, "loss": 2.382, "learning_rate": 1.3628318584070798e-07, "epoch": 4.341553637484587 }, { "current_steps": 3521, "loss": 2.1672, "learning_rate": 1.3603034134007584e-07, "epoch": 4.342786683107275 }, { "current_steps": 3522, "loss": 2.4029, "learning_rate": 1.3577749683944373e-07, "epoch": 4.344019728729963 }, { "current_steps": 3523, "loss": 1.8895, "learning_rate": 1.3552465233881164e-07, "epoch": 4.345252774352651 }, { "current_steps": 3524, "loss": 2.2424, "learning_rate": 1.352718078381795e-07, "epoch": 4.346485819975339 }, { "current_steps": 3525, "loss": 2.0835, "learning_rate": 1.350189633375474e-07, "epoch": 4.3477188655980274 }, { "current_steps": 3526, "loss": 2.049, "learning_rate": 1.347661188369153e-07, "epoch": 4.348951911220715 }, { "current_steps": 3527, "loss": 2.2856, "learning_rate": 1.3451327433628317e-07, "epoch": 4.350184956843403 }, { "current_steps": 3528, "loss": 1.7135, "learning_rate": 1.3426042983565105e-07, "epoch": 4.351418002466091 }, { "current_steps": 3529, "loss": 2.0588, "learning_rate": 1.3400758533501897e-07, "epoch": 4.352651048088779 }, { "current_steps": 3530, "loss": 1.3268, "learning_rate": 1.3375474083438683e-07, "epoch": 4.353884093711468 }, { "current_steps": 3531, "loss": 2.2692, "learning_rate": 1.3350189633375472e-07, "epoch": 4.355117139334156 }, { "current_steps": 3532, "loss": 2.1179, "learning_rate": 1.3324905183312263e-07, "epoch": 4.356350184956844 }, { "current_steps": 3533, "loss": 1.7335, "learning_rate": 1.3299620733249052e-07, "epoch": 4.3575832305795315 }, { "current_steps": 3534, "loss": 1.7898, "learning_rate": 1.327433628318584e-07, "epoch": 4.358816276202219 }, { "current_steps": 3535, "loss": 2.2015, "learning_rate": 1.324905183312263e-07, "epoch": 4.360049321824907 }, { "current_steps": 3536, "loss": 1.9902, "learning_rate": 1.3223767383059418e-07, "epoch": 4.361282367447595 }, { "current_steps": 3537, "loss": 2.0692, "learning_rate": 1.3198482932996207e-07, "epoch": 4.362515413070284 }, { "current_steps": 3538, "loss": 2.188, "learning_rate": 1.3173198482932996e-07, "epoch": 4.363748458692972 }, { "current_steps": 3539, "loss": 2.0394, "learning_rate": 1.3147914032869785e-07, "epoch": 4.36498150431566 }, { "current_steps": 3540, "loss": 1.7687, "learning_rate": 1.3122629582806573e-07, "epoch": 4.366214549938348 }, { "current_steps": 3541, "loss": 2.0978, "learning_rate": 1.3097345132743362e-07, "epoch": 4.3674475955610355 }, { "current_steps": 3542, "loss": 1.2792, "learning_rate": 1.307206068268015e-07, "epoch": 4.368680641183724 }, { "current_steps": 3543, "loss": 2.3255, "learning_rate": 1.304677623261694e-07, "epoch": 4.369913686806412 }, { "current_steps": 3544, "loss": 2.1328, "learning_rate": 1.3021491782553728e-07, "epoch": 4.3711467324291 }, { "current_steps": 3545, "loss": 1.7316, "learning_rate": 1.2996207332490517e-07, "epoch": 4.372379778051788 }, { "current_steps": 3546, "loss": 1.8836, "learning_rate": 1.2970922882427306e-07, "epoch": 4.373612823674476 }, { "current_steps": 3547, "loss": 2.0939, "learning_rate": 1.2945638432364095e-07, "epoch": 4.374845869297164 }, { "current_steps": 3548, "loss": 2.3442, "learning_rate": 1.2920353982300886e-07, "epoch": 4.376078914919852 }, { "current_steps": 3549, "loss": 1.9313, "learning_rate": 1.2895069532237672e-07, "epoch": 4.37731196054254 }, { "current_steps": 3550, "loss": 1.8327, "learning_rate": 1.286978508217446e-07, "epoch": 4.378545006165228 }, { "current_steps": 3551, "loss": 2.018, "learning_rate": 1.2844500632111252e-07, "epoch": 4.379778051787916 }, { "current_steps": 3552, "loss": 2.4267, "learning_rate": 1.2819216182048039e-07, "epoch": 4.381011097410604 }, { "current_steps": 3553, "loss": 2.4111, "learning_rate": 1.2793931731984827e-07, "epoch": 4.382244143033292 }, { "current_steps": 3554, "loss": 2.1182, "learning_rate": 1.276864728192162e-07, "epoch": 4.383477188655981 }, { "current_steps": 3555, "loss": 1.8545, "learning_rate": 1.2743362831858408e-07, "epoch": 4.384710234278669 }, { "current_steps": 3556, "loss": 2.1069, "learning_rate": 1.2718078381795194e-07, "epoch": 4.3859432799013565 }, { "current_steps": 3557, "loss": 2.0785, "learning_rate": 1.2692793931731985e-07, "epoch": 4.387176325524044 }, { "current_steps": 3558, "loss": 2.0467, "learning_rate": 1.2667509481668774e-07, "epoch": 4.388409371146732 }, { "current_steps": 3559, "loss": 2.0993, "learning_rate": 1.264222503160556e-07, "epoch": 4.38964241676942 }, { "current_steps": 3560, "loss": 2.5361, "learning_rate": 1.2616940581542351e-07, "epoch": 4.390875462392109 }, { "current_steps": 3561, "loss": 1.8991, "learning_rate": 1.259165613147914e-07, "epoch": 4.392108508014797 }, { "current_steps": 3562, "loss": 2.1551, "learning_rate": 1.256637168141593e-07, "epoch": 4.393341553637485 }, { "current_steps": 3563, "loss": 2.0094, "learning_rate": 1.2541087231352718e-07, "epoch": 4.394574599260173 }, { "current_steps": 3564, "loss": 1.879, "learning_rate": 1.2515802781289506e-07, "epoch": 4.3958076448828605 }, { "current_steps": 3565, "loss": 1.9674, "learning_rate": 1.2490518331226295e-07, "epoch": 4.397040690505548 }, { "current_steps": 3566, "loss": 1.9013, "learning_rate": 1.2465233881163084e-07, "epoch": 4.398273736128237 }, { "current_steps": 3567, "loss": 2.0373, "learning_rate": 1.2439949431099873e-07, "epoch": 4.399506781750925 }, { "current_steps": 3568, "loss": 1.9635, "learning_rate": 1.2414664981036662e-07, "epoch": 4.400739827373613 }, { "current_steps": 3569, "loss": 2.4113, "learning_rate": 1.238938053097345e-07, "epoch": 4.401972872996301 }, { "current_steps": 3570, "loss": 2.4712, "learning_rate": 1.236409608091024e-07, "epoch": 4.403205918618989 }, { "current_steps": 3571, "loss": 1.8895, "learning_rate": 1.2338811630847028e-07, "epoch": 4.404438964241677 }, { "current_steps": 3572, "loss": 1.7787, "learning_rate": 1.2313527180783817e-07, "epoch": 4.405672009864365 }, { "current_steps": 3573, "loss": 1.7352, "learning_rate": 1.2288242730720605e-07, "epoch": 4.406905055487053 }, { "current_steps": 3574, "loss": 2.2899, "learning_rate": 1.2262958280657397e-07, "epoch": 4.408138101109741 }, { "current_steps": 3575, "loss": 1.8552, "learning_rate": 1.2237673830594183e-07, "epoch": 4.409371146732429 }, { "current_steps": 3576, "loss": 2.2107, "learning_rate": 1.2212389380530972e-07, "epoch": 4.410604192355117 }, { "current_steps": 3577, "loss": 2.0058, "learning_rate": 1.2187104930467763e-07, "epoch": 4.411837237977805 }, { "current_steps": 3578, "loss": 2.4685, "learning_rate": 1.2161820480404552e-07, "epoch": 4.413070283600494 }, { "current_steps": 3579, "loss": 1.9591, "learning_rate": 1.2136536030341338e-07, "epoch": 4.4143033292231815 }, { "current_steps": 3580, "loss": 1.9253, "learning_rate": 1.211125158027813e-07, "epoch": 4.415536374845869 }, { "current_steps": 3581, "loss": 1.8457, "learning_rate": 1.2085967130214918e-07, "epoch": 4.416769420468557 }, { "current_steps": 3582, "loss": 1.9806, "learning_rate": 1.2060682680151704e-07, "epoch": 4.418002466091245 }, { "current_steps": 3583, "loss": 2.1022, "learning_rate": 1.2035398230088496e-07, "epoch": 4.419235511713933 }, { "current_steps": 3584, "loss": 2.191, "learning_rate": 1.2010113780025284e-07, "epoch": 4.420468557336622 }, { "current_steps": 3585, "loss": 2.3172, "learning_rate": 1.1984829329962073e-07, "epoch": 4.42170160295931 }, { "current_steps": 3586, "loss": 2.1126, "learning_rate": 1.1959544879898862e-07, "epoch": 4.422934648581998 }, { "current_steps": 3587, "loss": 2.0793, "learning_rate": 1.193426042983565e-07, "epoch": 4.4241676942046855 }, { "current_steps": 3588, "loss": 1.6674, "learning_rate": 1.190897597977244e-07, "epoch": 4.425400739827373 }, { "current_steps": 3589, "loss": 2.146, "learning_rate": 1.1883691529709228e-07, "epoch": 4.426633785450061 }, { "current_steps": 3590, "loss": 2.3294, "learning_rate": 1.1858407079646017e-07, "epoch": 4.42786683107275 }, { "current_steps": 3591, "loss": 2.6531, "learning_rate": 1.1833122629582806e-07, "epoch": 4.429099876695438 }, { "current_steps": 3592, "loss": 1.7885, "learning_rate": 1.1807838179519596e-07, "epoch": 4.430332922318126 }, { "current_steps": 3593, "loss": 1.7155, "learning_rate": 1.1782553729456383e-07, "epoch": 4.431565967940814 }, { "current_steps": 3594, "loss": 2.2231, "learning_rate": 1.1757269279393172e-07, "epoch": 4.432799013563502 }, { "current_steps": 3595, "loss": 2.0078, "learning_rate": 1.1731984829329962e-07, "epoch": 4.4340320591861895 }, { "current_steps": 3596, "loss": 2.0544, "learning_rate": 1.170670037926675e-07, "epoch": 4.435265104808878 }, { "current_steps": 3597, "loss": 2.3325, "learning_rate": 1.1681415929203538e-07, "epoch": 4.436498150431566 }, { "current_steps": 3598, "loss": 2.1302, "learning_rate": 1.1656131479140329e-07, "epoch": 4.437731196054254 }, { "current_steps": 3599, "loss": 2.1582, "learning_rate": 1.1630847029077117e-07, "epoch": 4.438964241676942 }, { "current_steps": 3600, "loss": 2.1831, "learning_rate": 1.1605562579013906e-07, "epoch": 4.44019728729963 }, { "current_steps": 3601, "loss": 1.6993, "learning_rate": 1.1580278128950695e-07, "epoch": 4.441430332922318 }, { "current_steps": 3602, "loss": 1.9903, "learning_rate": 1.1554993678887484e-07, "epoch": 4.4426633785450065 }, { "current_steps": 3603, "loss": 1.546, "learning_rate": 1.1529709228824272e-07, "epoch": 4.443896424167694 }, { "current_steps": 3604, "loss": 1.724, "learning_rate": 1.1504424778761061e-07, "epoch": 4.445129469790382 }, { "current_steps": 3605, "loss": 2.1706, "learning_rate": 1.147914032869785e-07, "epoch": 4.44636251541307 }, { "current_steps": 3606, "loss": 1.9602, "learning_rate": 1.145385587863464e-07, "epoch": 4.447595561035758 }, { "current_steps": 3607, "loss": 2.4748, "learning_rate": 1.1428571428571427e-07, "epoch": 4.448828606658447 }, { "current_steps": 3608, "loss": 2.5092, "learning_rate": 1.1403286978508218e-07, "epoch": 4.450061652281135 }, { "current_steps": 3609, "loss": 2.408, "learning_rate": 1.1378002528445006e-07, "epoch": 4.451294697903823 }, { "current_steps": 3610, "loss": 1.861, "learning_rate": 1.1352718078381794e-07, "epoch": 4.4525277435265105 }, { "current_steps": 3611, "loss": 2.0053, "learning_rate": 1.1327433628318584e-07, "epoch": 4.453760789149198 }, { "current_steps": 3612, "loss": 1.7512, "learning_rate": 1.1302149178255373e-07, "epoch": 4.454993834771886 }, { "current_steps": 3613, "loss": 1.7818, "learning_rate": 1.1276864728192161e-07, "epoch": 4.456226880394574 }, { "current_steps": 3614, "loss": 2.2337, "learning_rate": 1.125158027812895e-07, "epoch": 4.457459926017263 }, { "current_steps": 3615, "loss": 1.8402, "learning_rate": 1.1226295828065739e-07, "epoch": 4.458692971639951 }, { "current_steps": 3616, "loss": 2.255, "learning_rate": 1.1201011378002528e-07, "epoch": 4.459926017262639 }, { "current_steps": 3617, "loss": 2.054, "learning_rate": 1.1175726927939316e-07, "epoch": 4.461159062885327 }, { "current_steps": 3618, "loss": 2.3286, "learning_rate": 1.1150442477876105e-07, "epoch": 4.4623921085080145 }, { "current_steps": 3619, "loss": 1.8821, "learning_rate": 1.1125158027812895e-07, "epoch": 4.463625154130703 }, { "current_steps": 3620, "loss": 1.9074, "learning_rate": 1.1099873577749684e-07, "epoch": 4.464858199753391 }, { "current_steps": 3621, "loss": 2.0502, "learning_rate": 1.1074589127686472e-07, "epoch": 4.466091245376079 }, { "current_steps": 3622, "loss": 2.1569, "learning_rate": 1.1049304677623262e-07, "epoch": 4.467324290998767 }, { "current_steps": 3623, "loss": 2.4713, "learning_rate": 1.102402022756005e-07, "epoch": 4.468557336621455 }, { "current_steps": 3624, "loss": 2.5546, "learning_rate": 1.0998735777496838e-07, "epoch": 4.469790382244143 }, { "current_steps": 3625, "loss": 1.9438, "learning_rate": 1.0973451327433628e-07, "epoch": 4.471023427866831 }, { "current_steps": 3626, "loss": 2.1109, "learning_rate": 1.0948166877370417e-07, "epoch": 4.472256473489519 }, { "current_steps": 3627, "loss": 1.9443, "learning_rate": 1.0922882427307207e-07, "epoch": 4.473489519112207 }, { "current_steps": 3628, "loss": 2.3046, "learning_rate": 1.0897597977243994e-07, "epoch": 4.474722564734895 }, { "current_steps": 3629, "loss": 1.8958, "learning_rate": 1.0872313527180783e-07, "epoch": 4.475955610357583 }, { "current_steps": 3630, "loss": 2.1583, "learning_rate": 1.0847029077117573e-07, "epoch": 4.477188655980271 }, { "current_steps": 3631, "loss": 2.1781, "learning_rate": 1.082174462705436e-07, "epoch": 4.47842170160296 }, { "current_steps": 3632, "loss": 1.7052, "learning_rate": 1.079646017699115e-07, "epoch": 4.479654747225648 }, { "current_steps": 3633, "loss": 2.3695, "learning_rate": 1.077117572692794e-07, "epoch": 4.4808877928483355 }, { "current_steps": 3634, "loss": 1.8473, "learning_rate": 1.0745891276864728e-07, "epoch": 4.482120838471023 }, { "current_steps": 3635, "loss": 2.2304, "learning_rate": 1.0720606826801516e-07, "epoch": 4.483353884093711 }, { "current_steps": 3636, "loss": 2.0204, "learning_rate": 1.0695322376738306e-07, "epoch": 4.484586929716399 }, { "current_steps": 3637, "loss": 2.1199, "learning_rate": 1.0670037926675095e-07, "epoch": 4.485819975339088 }, { "current_steps": 3638, "loss": 1.6278, "learning_rate": 1.0644753476611883e-07, "epoch": 4.487053020961776 }, { "current_steps": 3639, "loss": 1.6199, "learning_rate": 1.0619469026548672e-07, "epoch": 4.488286066584464 }, { "current_steps": 3640, "loss": 2.2142, "learning_rate": 1.0594184576485461e-07, "epoch": 4.489519112207152 }, { "current_steps": 3641, "loss": 1.4724, "learning_rate": 1.0568900126422251e-07, "epoch": 4.4907521578298395 }, { "current_steps": 3642, "loss": 1.6365, "learning_rate": 1.0543615676359038e-07, "epoch": 4.491985203452527 }, { "current_steps": 3643, "loss": 1.6341, "learning_rate": 1.0518331226295827e-07, "epoch": 4.493218249075216 }, { "current_steps": 3644, "loss": 2.4218, "learning_rate": 1.0493046776232617e-07, "epoch": 4.494451294697904 }, { "current_steps": 3645, "loss": 1.6439, "learning_rate": 1.0467762326169405e-07, "epoch": 4.495684340320592 }, { "current_steps": 3646, "loss": 2.3392, "learning_rate": 1.0442477876106193e-07, "epoch": 4.49691738594328 }, { "current_steps": 3647, "loss": 1.7788, "learning_rate": 1.0417193426042984e-07, "epoch": 4.498150431565968 }, { "current_steps": 3648, "loss": 2.1862, "learning_rate": 1.0391908975979772e-07, "epoch": 4.499383477188656 }, { "current_steps": 3649, "loss": 1.867, "learning_rate": 1.0366624525916561e-07, "epoch": 4.500616522811344 }, { "current_steps": 3650, "loss": 2.2438, "learning_rate": 1.034134007585335e-07, "epoch": 4.501849568434032 }, { "current_steps": 3651, "loss": 2.1571, "learning_rate": 1.0316055625790139e-07, "epoch": 4.50308261405672 }, { "current_steps": 3652, "loss": 2.115, "learning_rate": 1.0290771175726927e-07, "epoch": 4.504315659679408 }, { "current_steps": 3653, "loss": 2.1677, "learning_rate": 1.0265486725663716e-07, "epoch": 4.505548705302096 }, { "current_steps": 3654, "loss": 1.553, "learning_rate": 1.0240202275600505e-07, "epoch": 4.506781750924784 }, { "current_steps": 3655, "loss": 1.8441, "learning_rate": 1.0214917825537295e-07, "epoch": 4.508014796547473 }, { "current_steps": 3656, "loss": 1.8749, "learning_rate": 1.0189633375474082e-07, "epoch": 4.5092478421701605 }, { "current_steps": 3657, "loss": 1.5834, "learning_rate": 1.0164348925410873e-07, "epoch": 4.510480887792848 }, { "current_steps": 3658, "loss": 1.9641, "learning_rate": 1.0139064475347661e-07, "epoch": 4.511713933415536 }, { "current_steps": 3659, "loss": 1.6632, "learning_rate": 1.0113780025284449e-07, "epoch": 4.512946979038224 }, { "current_steps": 3660, "loss": 2.169, "learning_rate": 1.0088495575221239e-07, "epoch": 4.514180024660913 }, { "current_steps": 3661, "loss": 2.2813, "learning_rate": 1.0063211125158028e-07, "epoch": 4.515413070283601 }, { "current_steps": 3662, "loss": 1.6002, "learning_rate": 1.0037926675094816e-07, "epoch": 4.516646115906289 }, { "current_steps": 3663, "loss": 1.649, "learning_rate": 1.0012642225031605e-07, "epoch": 4.517879161528977 }, { "current_steps": 3664, "loss": 2.3388, "learning_rate": 9.987357774968394e-08, "epoch": 4.5191122071516645 }, { "current_steps": 3665, "loss": 2.0161, "learning_rate": 9.962073324905183e-08, "epoch": 4.520345252774352 }, { "current_steps": 3666, "loss": 2.1551, "learning_rate": 9.936788874841971e-08, "epoch": 4.52157829839704 }, { "current_steps": 3667, "loss": 1.6812, "learning_rate": 9.91150442477876e-08, "epoch": 4.522811344019729 }, { "current_steps": 3668, "loss": 1.8625, "learning_rate": 9.88621997471555e-08, "epoch": 4.524044389642417 }, { "current_steps": 3669, "loss": 1.9863, "learning_rate": 9.860935524652339e-08, "epoch": 4.525277435265105 }, { "current_steps": 3670, "loss": 1.6006, "learning_rate": 9.835651074589127e-08, "epoch": 4.526510480887793 }, { "current_steps": 3671, "loss": 2.1186, "learning_rate": 9.810366624525917e-08, "epoch": 4.527743526510481 }, { "current_steps": 3672, "loss": 1.6627, "learning_rate": 9.785082174462705e-08, "epoch": 4.528976572133169 }, { "current_steps": 3673, "loss": 1.737, "learning_rate": 9.759797724399493e-08, "epoch": 4.530209617755857 }, { "current_steps": 3674, "loss": 2.5104, "learning_rate": 9.734513274336283e-08, "epoch": 4.531442663378545 }, { "current_steps": 3675, "loss": 1.9377, "learning_rate": 9.709228824273072e-08, "epoch": 4.532675709001233 }, { "current_steps": 3676, "loss": 2.2271, "learning_rate": 9.683944374209862e-08, "epoch": 4.533908754623921 }, { "current_steps": 3677, "loss": 2.0366, "learning_rate": 9.658659924146649e-08, "epoch": 4.535141800246609 }, { "current_steps": 3678, "loss": 2.4141, "learning_rate": 9.633375474083438e-08, "epoch": 4.536374845869297 }, { "current_steps": 3679, "loss": 2.0062, "learning_rate": 9.608091024020228e-08, "epoch": 4.5376078914919855 }, { "current_steps": 3680, "loss": 2.0903, "learning_rate": 9.582806573957016e-08, "epoch": 4.538840937114673 }, { "current_steps": 3681, "loss": 2.3788, "learning_rate": 9.557522123893804e-08, "epoch": 4.540073982737361 }, { "current_steps": 3682, "loss": 1.0229, "learning_rate": 9.532237673830594e-08, "epoch": 4.541307028360049 }, { "current_steps": 3683, "loss": 2.3583, "learning_rate": 9.506953223767383e-08, "epoch": 4.542540073982737 }, { "current_steps": 3684, "loss": 1.4968, "learning_rate": 9.48166877370417e-08, "epoch": 4.543773119605426 }, { "current_steps": 3685, "loss": 2.2423, "learning_rate": 9.456384323640961e-08, "epoch": 4.545006165228114 }, { "current_steps": 3686, "loss": 2.5446, "learning_rate": 9.43109987357775e-08, "epoch": 4.546239210850802 }, { "current_steps": 3687, "loss": 1.9408, "learning_rate": 9.405815423514538e-08, "epoch": 4.5474722564734895 }, { "current_steps": 3688, "loss": 2.1109, "learning_rate": 9.380530973451327e-08, "epoch": 4.548705302096177 }, { "current_steps": 3689, "loss": 2.113, "learning_rate": 9.355246523388116e-08, "epoch": 4.549938347718865 }, { "current_steps": 3690, "loss": 2.2294, "learning_rate": 9.329962073324906e-08, "epoch": 4.551171393341553 }, { "current_steps": 3691, "loss": 2.1132, "learning_rate": 9.304677623261693e-08, "epoch": 4.552404438964242 }, { "current_steps": 3692, "loss": 2.4263, "learning_rate": 9.279393173198482e-08, "epoch": 4.55363748458693 }, { "current_steps": 3693, "loss": 2.1077, "learning_rate": 9.254108723135272e-08, "epoch": 4.554870530209618 }, { "current_steps": 3694, "loss": 2.2366, "learning_rate": 9.22882427307206e-08, "epoch": 4.556103575832306 }, { "current_steps": 3695, "loss": 2.35, "learning_rate": 9.203539823008848e-08, "epoch": 4.5573366214549935 }, { "current_steps": 3696, "loss": 1.7984, "learning_rate": 9.178255372945639e-08, "epoch": 4.558569667077682 }, { "current_steps": 3697, "loss": 1.9204, "learning_rate": 9.152970922882427e-08, "epoch": 4.55980271270037 }, { "current_steps": 3698, "loss": 2.1061, "learning_rate": 9.127686472819216e-08, "epoch": 4.561035758323058 }, { "current_steps": 3699, "loss": 1.8094, "learning_rate": 9.102402022756005e-08, "epoch": 4.562268803945746 }, { "current_steps": 3700, "loss": 2.1924, "learning_rate": 9.077117572692794e-08, "epoch": 4.563501849568434 }, { "current_steps": 3701, "loss": 2.2209, "learning_rate": 9.051833122629582e-08, "epoch": 4.564734895191122 }, { "current_steps": 3702, "loss": 2.4671, "learning_rate": 9.026548672566371e-08, "epoch": 4.56596794081381 }, { "current_steps": 3703, "loss": 1.6692, "learning_rate": 9.00126422250316e-08, "epoch": 4.567200986436498 }, { "current_steps": 3704, "loss": 1.5495, "learning_rate": 8.97597977243995e-08, "epoch": 4.568434032059186 }, { "current_steps": 3705, "loss": 2.0466, "learning_rate": 8.950695322376737e-08, "epoch": 4.569667077681874 }, { "current_steps": 3706, "loss": 1.774, "learning_rate": 8.925410872313526e-08, "epoch": 4.570900123304562 }, { "current_steps": 3707, "loss": 2.5956, "learning_rate": 8.900126422250316e-08, "epoch": 4.57213316892725 }, { "current_steps": 3708, "loss": 1.9526, "learning_rate": 8.874841972187104e-08, "epoch": 4.573366214549939 }, { "current_steps": 3709, "loss": 2.0912, "learning_rate": 8.849557522123894e-08, "epoch": 4.574599260172627 }, { "current_steps": 3710, "loss": 2.2628, "learning_rate": 8.824273072060683e-08, "epoch": 4.5758323057953145 }, { "current_steps": 3711, "loss": 2.1903, "learning_rate": 8.798988621997471e-08, "epoch": 4.577065351418002 }, { "current_steps": 3712, "loss": 2.4061, "learning_rate": 8.77370417193426e-08, "epoch": 4.57829839704069 }, { "current_steps": 3713, "loss": 2.3376, "learning_rate": 8.748419721871049e-08, "epoch": 4.579531442663378 }, { "current_steps": 3714, "loss": 1.9533, "learning_rate": 8.723135271807838e-08, "epoch": 4.580764488286067 }, { "current_steps": 3715, "loss": 2.0417, "learning_rate": 8.697850821744626e-08, "epoch": 4.581997533908755 }, { "current_steps": 3716, "loss": 2.478, "learning_rate": 8.672566371681415e-08, "epoch": 4.583230579531443 }, { "current_steps": 3717, "loss": 1.9029, "learning_rate": 8.647281921618205e-08, "epoch": 4.584463625154131 }, { "current_steps": 3718, "loss": 1.9143, "learning_rate": 8.621997471554994e-08, "epoch": 4.5856966707768185 }, { "current_steps": 3719, "loss": 1.9951, "learning_rate": 8.596713021491782e-08, "epoch": 4.586929716399506 }, { "current_steps": 3720, "loss": 1.9894, "learning_rate": 8.571428571428572e-08, "epoch": 4.588162762022195 }, { "current_steps": 3721, "loss": 2.1774, "learning_rate": 8.54614412136536e-08, "epoch": 4.589395807644883 }, { "current_steps": 3722, "loss": 2.013, "learning_rate": 8.520859671302148e-08, "epoch": 4.590628853267571 }, { "current_steps": 3723, "loss": 1.8289, "learning_rate": 8.495575221238938e-08, "epoch": 4.591861898890259 }, { "current_steps": 3724, "loss": 2.0779, "learning_rate": 8.470290771175727e-08, "epoch": 4.593094944512947 }, { "current_steps": 3725, "loss": 1.6486, "learning_rate": 8.445006321112515e-08, "epoch": 4.594327990135635 }, { "current_steps": 3726, "loss": 1.8346, "learning_rate": 8.419721871049304e-08, "epoch": 4.595561035758323 }, { "current_steps": 3727, "loss": 1.6993, "learning_rate": 8.394437420986093e-08, "epoch": 4.596794081381011 }, { "current_steps": 3728, "loss": 1.9749, "learning_rate": 8.369152970922883e-08, "epoch": 4.598027127003699 }, { "current_steps": 3729, "loss": 2.4128, "learning_rate": 8.34386852085967e-08, "epoch": 4.599260172626387 }, { "current_steps": 3730, "loss": 2.3341, "learning_rate": 8.318584070796459e-08, "epoch": 4.600493218249075 }, { "current_steps": 3731, "loss": 2.1091, "learning_rate": 8.29329962073325e-08, "epoch": 4.601726263871763 }, { "current_steps": 3732, "loss": 1.7129, "learning_rate": 8.268015170670038e-08, "epoch": 4.602959309494452 }, { "current_steps": 3733, "loss": 1.3931, "learning_rate": 8.242730720606826e-08, "epoch": 4.6041923551171395 }, { "current_steps": 3734, "loss": 2.1622, "learning_rate": 8.217446270543616e-08, "epoch": 4.605425400739827 }, { "current_steps": 3735, "loss": 1.7637, "learning_rate": 8.192161820480404e-08, "epoch": 4.606658446362515 }, { "current_steps": 3736, "loss": 2.2406, "learning_rate": 8.166877370417192e-08, "epoch": 4.607891491985203 }, { "current_steps": 3737, "loss": 2.0214, "learning_rate": 8.141592920353982e-08, "epoch": 4.609124537607892 }, { "current_steps": 3738, "loss": 1.7179, "learning_rate": 8.116308470290771e-08, "epoch": 4.61035758323058 }, { "current_steps": 3739, "loss": 2.0092, "learning_rate": 8.091024020227561e-08, "epoch": 4.611590628853268 }, { "current_steps": 3740, "loss": 1.9825, "learning_rate": 8.065739570164348e-08, "epoch": 4.612823674475956 }, { "current_steps": 3741, "loss": 2.1784, "learning_rate": 8.040455120101137e-08, "epoch": 4.6140567200986435 }, { "current_steps": 3742, "loss": 2.0948, "learning_rate": 8.015170670037927e-08, "epoch": 4.615289765721331 }, { "current_steps": 3743, "loss": 1.9443, "learning_rate": 7.989886219974715e-08, "epoch": 4.616522811344019 }, { "current_steps": 3744, "loss": 2.1371, "learning_rate": 7.964601769911503e-08, "epoch": 4.617755856966708 }, { "current_steps": 3745, "loss": 1.8688, "learning_rate": 7.939317319848293e-08, "epoch": 4.618988902589396 }, { "current_steps": 3746, "loss": 2.3693, "learning_rate": 7.914032869785082e-08, "epoch": 4.620221948212084 }, { "current_steps": 3747, "loss": 2.1248, "learning_rate": 7.888748419721871e-08, "epoch": 4.621454993834772 }, { "current_steps": 3748, "loss": 1.9044, "learning_rate": 7.86346396965866e-08, "epoch": 4.62268803945746 }, { "current_steps": 3749, "loss": 2.068, "learning_rate": 7.838179519595449e-08, "epoch": 4.623921085080148 }, { "current_steps": 3750, "loss": 1.7095, "learning_rate": 7.812895069532237e-08, "epoch": 4.625154130702836 }, { "current_steps": 3751, "loss": 2.0958, "learning_rate": 7.787610619469026e-08, "epoch": 4.626387176325524 }, { "current_steps": 3752, "loss": 1.7873, "learning_rate": 7.762326169405815e-08, "epoch": 4.627620221948212 }, { "current_steps": 3753, "loss": 1.5438, "learning_rate": 7.737041719342605e-08, "epoch": 4.6288532675709 }, { "current_steps": 3754, "loss": 1.9355, "learning_rate": 7.711757269279392e-08, "epoch": 4.630086313193588 }, { "current_steps": 3755, "loss": 1.8469, "learning_rate": 7.686472819216181e-08, "epoch": 4.631319358816276 }, { "current_steps": 3756, "loss": 2.4824, "learning_rate": 7.661188369152971e-08, "epoch": 4.6325524044389645 }, { "current_steps": 3757, "loss": 1.717, "learning_rate": 7.635903919089759e-08, "epoch": 4.633785450061652 }, { "current_steps": 3758, "loss": 1.4438, "learning_rate": 7.610619469026549e-08, "epoch": 4.63501849568434 }, { "current_steps": 3759, "loss": 1.6446, "learning_rate": 7.585335018963338e-08, "epoch": 4.636251541307028 }, { "current_steps": 3760, "loss": 1.7509, "learning_rate": 7.560050568900126e-08, "epoch": 4.637484586929716 }, { "current_steps": 3761, "loss": 2.0966, "learning_rate": 7.534766118836915e-08, "epoch": 4.638717632552405 }, { "current_steps": 3762, "loss": 1.8453, "learning_rate": 7.509481668773704e-08, "epoch": 4.639950678175093 }, { "current_steps": 3763, "loss": 2.071, "learning_rate": 7.484197218710493e-08, "epoch": 4.641183723797781 }, { "current_steps": 3764, "loss": 1.4654, "learning_rate": 7.458912768647281e-08, "epoch": 4.6424167694204685 }, { "current_steps": 3765, "loss": 2.0502, "learning_rate": 7.43362831858407e-08, "epoch": 4.643649815043156 }, { "current_steps": 3766, "loss": 1.8736, "learning_rate": 7.40834386852086e-08, "epoch": 4.644882860665844 }, { "current_steps": 3767, "loss": 2.3288, "learning_rate": 7.383059418457649e-08, "epoch": 4.646115906288532 }, { "current_steps": 3768, "loss": 2.0076, "learning_rate": 7.357774968394436e-08, "epoch": 4.647348951911221 }, { "current_steps": 3769, "loss": 1.8641, "learning_rate": 7.332490518331227e-08, "epoch": 4.648581997533909 }, { "current_steps": 3770, "loss": 2.5655, "learning_rate": 7.307206068268015e-08, "epoch": 4.649815043156597 }, { "current_steps": 3771, "loss": 2.1376, "learning_rate": 7.281921618204803e-08, "epoch": 4.651048088779285 }, { "current_steps": 3772, "loss": 1.5935, "learning_rate": 7.256637168141593e-08, "epoch": 4.6522811344019726 }, { "current_steps": 3773, "loss": 2.448, "learning_rate": 7.231352718078382e-08, "epoch": 4.653514180024661 }, { "current_steps": 3774, "loss": 2.1772, "learning_rate": 7.20606826801517e-08, "epoch": 4.654747225647349 }, { "current_steps": 3775, "loss": 1.9992, "learning_rate": 7.180783817951959e-08, "epoch": 4.655980271270037 }, { "current_steps": 3776, "loss": 2.3697, "learning_rate": 7.155499367888748e-08, "epoch": 4.657213316892725 }, { "current_steps": 3777, "loss": 1.0498, "learning_rate": 7.130214917825538e-08, "epoch": 4.658446362515413 }, { "current_steps": 3778, "loss": 1.8359, "learning_rate": 7.104930467762325e-08, "epoch": 4.659679408138101 }, { "current_steps": 3779, "loss": 2.3889, "learning_rate": 7.079646017699114e-08, "epoch": 4.660912453760789 }, { "current_steps": 3780, "loss": 2.1294, "learning_rate": 7.054361567635904e-08, "epoch": 4.662145499383477 }, { "current_steps": 3781, "loss": 1.945, "learning_rate": 7.029077117572692e-08, "epoch": 4.663378545006165 }, { "current_steps": 3782, "loss": 1.9602, "learning_rate": 7.00379266750948e-08, "epoch": 4.664611590628853 }, { "current_steps": 3783, "loss": 1.9861, "learning_rate": 6.978508217446271e-08, "epoch": 4.665844636251541 }, { "current_steps": 3784, "loss": 2.2988, "learning_rate": 6.95322376738306e-08, "epoch": 4.667077681874229 }, { "current_steps": 3785, "loss": 2.2185, "learning_rate": 6.927939317319847e-08, "epoch": 4.668310727496918 }, { "current_steps": 3786, "loss": 2.0053, "learning_rate": 6.902654867256637e-08, "epoch": 4.669543773119606 }, { "current_steps": 3787, "loss": 2.0639, "learning_rate": 6.877370417193426e-08, "epoch": 4.670776818742294 }, { "current_steps": 3788, "loss": 1.8695, "learning_rate": 6.852085967130215e-08, "epoch": 4.6720098643649814 }, { "current_steps": 3789, "loss": 1.7876, "learning_rate": 6.826801517067003e-08, "epoch": 4.673242909987669 }, { "current_steps": 3790, "loss": 2.2785, "learning_rate": 6.801517067003792e-08, "epoch": 4.674475955610357 }, { "current_steps": 3791, "loss": 1.4736, "learning_rate": 6.776232616940582e-08, "epoch": 4.675709001233046 }, { "current_steps": 3792, "loss": 2.0089, "learning_rate": 6.75094816687737e-08, "epoch": 4.676942046855734 }, { "current_steps": 3793, "loss": 1.9086, "learning_rate": 6.725663716814158e-08, "epoch": 4.678175092478422 }, { "current_steps": 3794, "loss": 1.6489, "learning_rate": 6.700379266750948e-08, "epoch": 4.67940813810111 }, { "current_steps": 3795, "loss": 2.2376, "learning_rate": 6.675094816687736e-08, "epoch": 4.680641183723798 }, { "current_steps": 3796, "loss": 1.3303, "learning_rate": 6.649810366624526e-08, "epoch": 4.6818742293464854 }, { "current_steps": 3797, "loss": 2.0696, "learning_rate": 6.624525916561315e-08, "epoch": 4.683107274969174 }, { "current_steps": 3798, "loss": 1.3245, "learning_rate": 6.599241466498104e-08, "epoch": 4.684340320591862 }, { "current_steps": 3799, "loss": 2.1731, "learning_rate": 6.573957016434892e-08, "epoch": 4.68557336621455 }, { "current_steps": 3800, "loss": 2.1931, "learning_rate": 6.548672566371681e-08, "epoch": 4.686806411837238 }, { "current_steps": 3801, "loss": 2.1726, "learning_rate": 6.52338811630847e-08, "epoch": 4.688039457459926 }, { "current_steps": 3802, "loss": 1.7514, "learning_rate": 6.498103666245259e-08, "epoch": 4.689272503082614 }, { "current_steps": 3803, "loss": 1.8257, "learning_rate": 6.472819216182047e-08, "epoch": 4.6905055487053025 }, { "current_steps": 3804, "loss": 1.8321, "learning_rate": 6.447534766118836e-08, "epoch": 4.69173859432799 }, { "current_steps": 3805, "loss": 1.9215, "learning_rate": 6.422250316055626e-08, "epoch": 4.692971639950678 }, { "current_steps": 3806, "loss": 2.2413, "learning_rate": 6.396965865992414e-08, "epoch": 4.694204685573366 }, { "current_steps": 3807, "loss": 1.925, "learning_rate": 6.371681415929204e-08, "epoch": 4.695437731196054 }, { "current_steps": 3808, "loss": 2.1302, "learning_rate": 6.346396965865993e-08, "epoch": 4.696670776818742 }, { "current_steps": 3809, "loss": 1.983, "learning_rate": 6.32111251580278e-08, "epoch": 4.697903822441431 }, { "current_steps": 3810, "loss": 2.1277, "learning_rate": 6.29582806573957e-08, "epoch": 4.699136868064119 }, { "current_steps": 3811, "loss": 1.8056, "learning_rate": 6.270543615676359e-08, "epoch": 4.7003699136868065 }, { "current_steps": 3812, "loss": 1.2685, "learning_rate": 6.245259165613148e-08, "epoch": 4.701602959309494 }, { "current_steps": 3813, "loss": 1.9545, "learning_rate": 6.219974715549936e-08, "epoch": 4.702836004932182 }, { "current_steps": 3814, "loss": 1.6411, "learning_rate": 6.194690265486725e-08, "epoch": 4.704069050554871 }, { "current_steps": 3815, "loss": 1.903, "learning_rate": 6.169405815423514e-08, "epoch": 4.705302096177559 }, { "current_steps": 3816, "loss": 2.3648, "learning_rate": 6.144121365360303e-08, "epoch": 4.706535141800247 }, { "current_steps": 3817, "loss": 1.9531, "learning_rate": 6.118836915297091e-08, "epoch": 4.707768187422935 }, { "current_steps": 3818, "loss": 2.0897, "learning_rate": 6.093552465233882e-08, "epoch": 4.709001233045623 }, { "current_steps": 3819, "loss": 1.9278, "learning_rate": 6.068268015170669e-08, "epoch": 4.7102342786683105 }, { "current_steps": 3820, "loss": 2.7011, "learning_rate": 6.042983565107459e-08, "epoch": 4.711467324290998 }, { "current_steps": 3821, "loss": 1.8557, "learning_rate": 6.017699115044248e-08, "epoch": 4.712700369913687 }, { "current_steps": 3822, "loss": 2.1838, "learning_rate": 5.992414664981037e-08, "epoch": 4.713933415536375 }, { "current_steps": 3823, "loss": 1.9907, "learning_rate": 5.967130214917825e-08, "epoch": 4.715166461159063 }, { "current_steps": 3824, "loss": 2.3216, "learning_rate": 5.941845764854614e-08, "epoch": 4.716399506781751 }, { "current_steps": 3825, "loss": 1.7856, "learning_rate": 5.916561314791403e-08, "epoch": 4.717632552404439 }, { "current_steps": 3826, "loss": 2.0898, "learning_rate": 5.891276864728192e-08, "epoch": 4.7188655980271275 }, { "current_steps": 3827, "loss": 1.397, "learning_rate": 5.865992414664981e-08, "epoch": 4.720098643649815 }, { "current_steps": 3828, "loss": 2.1169, "learning_rate": 5.840707964601769e-08, "epoch": 4.721331689272503 }, { "current_steps": 3829, "loss": 2.225, "learning_rate": 5.8154235145385587e-08, "epoch": 4.722564734895191 }, { "current_steps": 3830, "loss": 1.7342, "learning_rate": 5.7901390644753474e-08, "epoch": 4.723797780517879 }, { "current_steps": 3831, "loss": 2.143, "learning_rate": 5.764854614412136e-08, "epoch": 4.725030826140567 }, { "current_steps": 3832, "loss": 2.2961, "learning_rate": 5.739570164348925e-08, "epoch": 4.726263871763255 }, { "current_steps": 3833, "loss": 2.259, "learning_rate": 5.714285714285714e-08, "epoch": 4.727496917385944 }, { "current_steps": 3834, "loss": 1.9546, "learning_rate": 5.689001264222503e-08, "epoch": 4.7287299630086315 }, { "current_steps": 3835, "loss": 2.4164, "learning_rate": 5.663716814159292e-08, "epoch": 4.729963008631319 }, { "current_steps": 3836, "loss": 1.6658, "learning_rate": 5.638432364096081e-08, "epoch": 4.731196054254007 }, { "current_steps": 3837, "loss": 1.7973, "learning_rate": 5.6131479140328695e-08, "epoch": 4.732429099876695 }, { "current_steps": 3838, "loss": 2.1512, "learning_rate": 5.587863463969658e-08, "epoch": 4.733662145499384 }, { "current_steps": 3839, "loss": 1.4261, "learning_rate": 5.562579013906448e-08, "epoch": 4.734895191122072 }, { "current_steps": 3840, "loss": 2.1958, "learning_rate": 5.537294563843236e-08, "epoch": 4.73612823674476 }, { "current_steps": 3841, "loss": 2.2528, "learning_rate": 5.512010113780025e-08, "epoch": 4.737361282367448 }, { "current_steps": 3842, "loss": 2.1612, "learning_rate": 5.486725663716814e-08, "epoch": 4.7385943279901355 }, { "current_steps": 3843, "loss": 2.0439, "learning_rate": 5.4614412136536034e-08, "epoch": 4.739827373612823 }, { "current_steps": 3844, "loss": 1.8666, "learning_rate": 5.4361567635903915e-08, "epoch": 4.741060419235511 }, { "current_steps": 3845, "loss": 1.703, "learning_rate": 5.41087231352718e-08, "epoch": 4.7422934648582 }, { "current_steps": 3846, "loss": 1.8486, "learning_rate": 5.38558786346397e-08, "epoch": 4.743526510480888 }, { "current_steps": 3847, "loss": 1.7854, "learning_rate": 5.360303413400758e-08, "epoch": 4.744759556103576 }, { "current_steps": 3848, "loss": 2.2388, "learning_rate": 5.335018963337547e-08, "epoch": 4.745992601726264 }, { "current_steps": 3849, "loss": 2.3932, "learning_rate": 5.309734513274336e-08, "epoch": 4.747225647348952 }, { "current_steps": 3850, "loss": 1.893, "learning_rate": 5.2844500632111255e-08, "epoch": 4.74845869297164 }, { "current_steps": 3851, "loss": 2.1585, "learning_rate": 5.2591656131479136e-08, "epoch": 4.749691738594328 }, { "current_steps": 3852, "loss": 1.933, "learning_rate": 5.2338811630847023e-08, "epoch": 4.750924784217016 }, { "current_steps": 3853, "loss": 1.8513, "learning_rate": 5.208596713021492e-08, "epoch": 4.752157829839704 }, { "current_steps": 3854, "loss": 1.9212, "learning_rate": 5.1833122629582805e-08, "epoch": 4.753390875462392 }, { "current_steps": 3855, "loss": 1.2991, "learning_rate": 5.158027812895069e-08, "epoch": 4.75462392108508 }, { "current_steps": 3856, "loss": 1.8143, "learning_rate": 5.132743362831858e-08, "epoch": 4.755856966707768 }, { "current_steps": 3857, "loss": 1.8639, "learning_rate": 5.1074589127686475e-08, "epoch": 4.7570900123304565 }, { "current_steps": 3858, "loss": 1.6307, "learning_rate": 5.082174462705436e-08, "epoch": 4.758323057953144 }, { "current_steps": 3859, "loss": 2.1871, "learning_rate": 5.0568900126422244e-08, "epoch": 4.759556103575832 }, { "current_steps": 3860, "loss": 1.8601, "learning_rate": 5.031605562579014e-08, "epoch": 4.76078914919852 }, { "current_steps": 3861, "loss": 1.7421, "learning_rate": 5.0063211125158026e-08, "epoch": 4.762022194821208 }, { "current_steps": 3862, "loss": 2.0008, "learning_rate": 4.9810366624525914e-08, "epoch": 4.763255240443897 }, { "current_steps": 3863, "loss": 2.244, "learning_rate": 4.95575221238938e-08, "epoch": 4.764488286066585 }, { "current_steps": 3864, "loss": 2.2136, "learning_rate": 4.9304677623261695e-08, "epoch": 4.765721331689273 }, { "current_steps": 3865, "loss": 1.6038, "learning_rate": 4.905183312262958e-08, "epoch": 4.7669543773119605 }, { "current_steps": 3866, "loss": 2.5371, "learning_rate": 4.8798988621997464e-08, "epoch": 4.768187422934648 }, { "current_steps": 3867, "loss": 2.1654, "learning_rate": 4.854614412136536e-08, "epoch": 4.769420468557336 }, { "current_steps": 3868, "loss": 1.5183, "learning_rate": 4.8293299620733246e-08, "epoch": 4.770653514180025 }, { "current_steps": 3869, "loss": 2.1552, "learning_rate": 4.804045512010114e-08, "epoch": 4.771886559802713 }, { "current_steps": 3870, "loss": 2.6764, "learning_rate": 4.778761061946902e-08, "epoch": 4.773119605425401 }, { "current_steps": 3871, "loss": 2.1965, "learning_rate": 4.7534766118836916e-08, "epoch": 4.774352651048089 }, { "current_steps": 3872, "loss": 1.8547, "learning_rate": 4.7281921618204804e-08, "epoch": 4.775585696670777 }, { "current_steps": 3873, "loss": 1.9028, "learning_rate": 4.702907711757269e-08, "epoch": 4.7768187422934645 }, { "current_steps": 3874, "loss": 2.0995, "learning_rate": 4.677623261694058e-08, "epoch": 4.778051787916153 }, { "current_steps": 3875, "loss": 2.1469, "learning_rate": 4.652338811630847e-08, "epoch": 4.779284833538841 }, { "current_steps": 3876, "loss": 2.4324, "learning_rate": 4.627054361567636e-08, "epoch": 4.780517879161529 }, { "current_steps": 3877, "loss": 2.0408, "learning_rate": 4.601769911504424e-08, "epoch": 4.781750924784217 }, { "current_steps": 3878, "loss": 1.7386, "learning_rate": 4.5764854614412136e-08, "epoch": 4.782983970406905 }, { "current_steps": 3879, "loss": 2.4054, "learning_rate": 4.5512010113780024e-08, "epoch": 4.784217016029594 }, { "current_steps": 3880, "loss": 2.1706, "learning_rate": 4.525916561314791e-08, "epoch": 4.7854500616522815 }, { "current_steps": 3881, "loss": 2.1183, "learning_rate": 4.50063211125158e-08, "epoch": 4.786683107274969 }, { "current_steps": 3882, "loss": 1.3995, "learning_rate": 4.475347661188369e-08, "epoch": 4.787916152897657 }, { "current_steps": 3883, "loss": 2.0226, "learning_rate": 4.450063211125158e-08, "epoch": 4.789149198520345 }, { "current_steps": 3884, "loss": 2.1921, "learning_rate": 4.424778761061947e-08, "epoch": 4.790382244143033 }, { "current_steps": 3885, "loss": 2.1215, "learning_rate": 4.399494310998736e-08, "epoch": 4.791615289765721 }, { "current_steps": 3886, "loss": 1.8543, "learning_rate": 4.3742098609355245e-08, "epoch": 4.79284833538841 }, { "current_steps": 3887, "loss": 1.7512, "learning_rate": 4.348925410872313e-08, "epoch": 4.794081381011098 }, { "current_steps": 3888, "loss": 1.9142, "learning_rate": 4.3236409608091027e-08, "epoch": 4.7953144266337855 }, { "current_steps": 3889, "loss": 2.3203, "learning_rate": 4.298356510745891e-08, "epoch": 4.796547472256473 }, { "current_steps": 3890, "loss": 2.2747, "learning_rate": 4.27307206068268e-08, "epoch": 4.797780517879161 }, { "current_steps": 3891, "loss": 1.973, "learning_rate": 4.247787610619469e-08, "epoch": 4.79901356350185 }, { "current_steps": 3892, "loss": 1.5244, "learning_rate": 4.222503160556258e-08, "epoch": 4.800246609124538 }, { "current_steps": 3893, "loss": 2.1168, "learning_rate": 4.1972187104930465e-08, "epoch": 4.801479654747226 }, { "current_steps": 3894, "loss": 2.3141, "learning_rate": 4.171934260429835e-08, "epoch": 4.802712700369914 }, { "current_steps": 3895, "loss": 2.1737, "learning_rate": 4.146649810366625e-08, "epoch": 4.803945745992602 }, { "current_steps": 3896, "loss": 1.7507, "learning_rate": 4.121365360303413e-08, "epoch": 4.8051787916152895 }, { "current_steps": 3897, "loss": 1.4999, "learning_rate": 4.096080910240202e-08, "epoch": 4.806411837237977 }, { "current_steps": 3898, "loss": 2.1363, "learning_rate": 4.070796460176991e-08, "epoch": 4.807644882860666 }, { "current_steps": 3899, "loss": 2.2388, "learning_rate": 4.0455120101137804e-08, "epoch": 4.808877928483354 }, { "current_steps": 3900, "loss": 1.3417, "learning_rate": 4.0202275600505685e-08, "epoch": 4.810110974106042 }, { "current_steps": 3901, "loss": 1.9979, "learning_rate": 3.994943109987357e-08, "epoch": 4.81134401972873 }, { "current_steps": 3902, "loss": 2.1311, "learning_rate": 3.969658659924147e-08, "epoch": 4.812577065351418 }, { "current_steps": 3903, "loss": 2.2825, "learning_rate": 3.9443742098609355e-08, "epoch": 4.8138101109741065 }, { "current_steps": 3904, "loss": 2.3151, "learning_rate": 3.919089759797724e-08, "epoch": 4.815043156596794 }, { "current_steps": 3905, "loss": 1.7886, "learning_rate": 3.893805309734513e-08, "epoch": 4.816276202219482 }, { "current_steps": 3906, "loss": 2.25, "learning_rate": 3.8685208596713025e-08, "epoch": 4.81750924784217 }, { "current_steps": 3907, "loss": 2.0516, "learning_rate": 3.8432364096080906e-08, "epoch": 4.818742293464858 }, { "current_steps": 3908, "loss": 2.7424, "learning_rate": 3.8179519595448794e-08, "epoch": 4.819975339087546 }, { "current_steps": 3909, "loss": 2.1784, "learning_rate": 3.792667509481669e-08, "epoch": 4.821208384710234 }, { "current_steps": 3910, "loss": 1.9922, "learning_rate": 3.7673830594184576e-08, "epoch": 4.822441430332923 }, { "current_steps": 3911, "loss": 1.6656, "learning_rate": 3.742098609355246e-08, "epoch": 4.8236744759556105 }, { "current_steps": 3912, "loss": 2.2097, "learning_rate": 3.716814159292035e-08, "epoch": 4.824907521578298 }, { "current_steps": 3913, "loss": 2.1207, "learning_rate": 3.6915297092288245e-08, "epoch": 4.826140567200986 }, { "current_steps": 3914, "loss": 1.8781, "learning_rate": 3.666245259165613e-08, "epoch": 4.827373612823674 }, { "current_steps": 3915, "loss": 2.008, "learning_rate": 3.6409608091024014e-08, "epoch": 4.828606658446363 }, { "current_steps": 3916, "loss": 1.9881, "learning_rate": 3.615676359039191e-08, "epoch": 4.829839704069051 }, { "current_steps": 3917, "loss": 1.9231, "learning_rate": 3.5903919089759796e-08, "epoch": 4.831072749691739 }, { "current_steps": 3918, "loss": 2.4526, "learning_rate": 3.565107458912769e-08, "epoch": 4.832305795314427 }, { "current_steps": 3919, "loss": 1.9499, "learning_rate": 3.539823008849557e-08, "epoch": 4.8335388409371145 }, { "current_steps": 3920, "loss": 2.0828, "learning_rate": 3.514538558786346e-08, "epoch": 4.834771886559802 }, { "current_steps": 3921, "loss": 1.4812, "learning_rate": 3.4892541087231353e-08, "epoch": 4.83600493218249 }, { "current_steps": 3922, "loss": 2.0198, "learning_rate": 3.4639696586599234e-08, "epoch": 4.837237977805179 }, { "current_steps": 3923, "loss": 2.1759, "learning_rate": 3.438685208596713e-08, "epoch": 4.838471023427867 }, { "current_steps": 3924, "loss": 2.2164, "learning_rate": 3.4134007585335016e-08, "epoch": 4.839704069050555 }, { "current_steps": 3925, "loss": 2.3424, "learning_rate": 3.388116308470291e-08, "epoch": 4.840937114673243 }, { "current_steps": 3926, "loss": 1.5962, "learning_rate": 3.362831858407079e-08, "epoch": 4.842170160295931 }, { "current_steps": 3927, "loss": 1.9701, "learning_rate": 3.337547408343868e-08, "epoch": 4.843403205918619 }, { "current_steps": 3928, "loss": 1.9393, "learning_rate": 3.3122629582806574e-08, "epoch": 4.844636251541307 }, { "current_steps": 3929, "loss": 2.0671, "learning_rate": 3.286978508217446e-08, "epoch": 4.845869297163995 }, { "current_steps": 3930, "loss": 1.9137, "learning_rate": 3.261694058154235e-08, "epoch": 4.847102342786683 }, { "current_steps": 3931, "loss": 2.2991, "learning_rate": 3.236409608091024e-08, "epoch": 4.848335388409371 }, { "current_steps": 3932, "loss": 1.9556, "learning_rate": 3.211125158027813e-08, "epoch": 4.849568434032059 }, { "current_steps": 3933, "loss": 2.1715, "learning_rate": 3.185840707964602e-08, "epoch": 4.850801479654747 }, { "current_steps": 3934, "loss": 1.7735, "learning_rate": 3.16055625790139e-08, "epoch": 4.8520345252774355 }, { "current_steps": 3935, "loss": 1.4865, "learning_rate": 3.1352718078381794e-08, "epoch": 4.853267570900123 }, { "current_steps": 3936, "loss": 1.299, "learning_rate": 3.109987357774968e-08, "epoch": 4.854500616522811 }, { "current_steps": 3937, "loss": 2.2757, "learning_rate": 3.084702907711757e-08, "epoch": 4.855733662145499 }, { "current_steps": 3938, "loss": 2.0915, "learning_rate": 3.059418457648546e-08, "epoch": 4.856966707768187 }, { "current_steps": 3939, "loss": 2.5021, "learning_rate": 3.0341340075853345e-08, "epoch": 4.858199753390876 }, { "current_steps": 3940, "loss": 2.0298, "learning_rate": 3.008849557522124e-08, "epoch": 4.859432799013564 }, { "current_steps": 3941, "loss": 1.4453, "learning_rate": 2.983565107458913e-08, "epoch": 4.860665844636252 }, { "current_steps": 3942, "loss": 2.0553, "learning_rate": 2.9582806573957015e-08, "epoch": 4.8618988902589395 }, { "current_steps": 3943, "loss": 1.642, "learning_rate": 2.9329962073324906e-08, "epoch": 4.863131935881627 }, { "current_steps": 3944, "loss": 2.1005, "learning_rate": 2.9077117572692793e-08, "epoch": 4.864364981504315 }, { "current_steps": 3945, "loss": 1.7645, "learning_rate": 2.882427307206068e-08, "epoch": 4.865598027127004 }, { "current_steps": 3946, "loss": 2.1324, "learning_rate": 2.857142857142857e-08, "epoch": 4.866831072749692 }, { "current_steps": 3947, "loss": 2.1804, "learning_rate": 2.831858407079646e-08, "epoch": 4.86806411837238 }, { "current_steps": 3948, "loss": 1.422, "learning_rate": 2.8065739570164347e-08, "epoch": 4.869297163995068 }, { "current_steps": 3949, "loss": 1.9376, "learning_rate": 2.781289506953224e-08, "epoch": 4.870530209617756 }, { "current_steps": 3950, "loss": 2.3571, "learning_rate": 2.7560050568900126e-08, "epoch": 4.8717632552404435 }, { "current_steps": 3951, "loss": 2.6023, "learning_rate": 2.7307206068268017e-08, "epoch": 4.872996300863132 }, { "current_steps": 3952, "loss": 2.6498, "learning_rate": 2.70543615676359e-08, "epoch": 4.87422934648582 }, { "current_steps": 3953, "loss": 1.7078, "learning_rate": 2.680151706700379e-08, "epoch": 4.875462392108508 }, { "current_steps": 3954, "loss": 2.2917, "learning_rate": 2.654867256637168e-08, "epoch": 4.876695437731196 }, { "current_steps": 3955, "loss": 1.4253, "learning_rate": 2.6295828065739568e-08, "epoch": 4.877928483353884 }, { "current_steps": 3956, "loss": 1.7945, "learning_rate": 2.604298356510746e-08, "epoch": 4.879161528976573 }, { "current_steps": 3957, "loss": 2.3546, "learning_rate": 2.5790139064475347e-08, "epoch": 4.8803945745992605 }, { "current_steps": 3958, "loss": 1.6789, "learning_rate": 2.5537294563843238e-08, "epoch": 4.881627620221948 }, { "current_steps": 3959, "loss": 1.6929, "learning_rate": 2.5284450063211122e-08, "epoch": 4.882860665844636 }, { "current_steps": 3960, "loss": 1.8293, "learning_rate": 2.5031605562579013e-08, "epoch": 4.884093711467324 }, { "current_steps": 3961, "loss": 2.0187, "learning_rate": 2.47787610619469e-08, "epoch": 4.885326757090012 }, { "current_steps": 3962, "loss": 2.22, "learning_rate": 2.452591656131479e-08, "epoch": 4.8865598027127 }, { "current_steps": 3963, "loss": 2.3194, "learning_rate": 2.427307206068268e-08, "epoch": 4.887792848335389 }, { "current_steps": 3964, "loss": 2.2265, "learning_rate": 2.402022756005057e-08, "epoch": 4.889025893958077 }, { "current_steps": 3965, "loss": 2.068, "learning_rate": 2.3767383059418458e-08, "epoch": 4.8902589395807645 }, { "current_steps": 3966, "loss": 2.4347, "learning_rate": 2.3514538558786346e-08, "epoch": 4.891491985203452 }, { "current_steps": 3967, "loss": 2.0345, "learning_rate": 2.3261694058154233e-08, "epoch": 4.89272503082614 }, { "current_steps": 3968, "loss": 2.0578, "learning_rate": 2.300884955752212e-08, "epoch": 4.893958076448829 }, { "current_steps": 3969, "loss": 2.557, "learning_rate": 2.2756005056890012e-08, "epoch": 4.895191122071517 }, { "current_steps": 3970, "loss": 2.036, "learning_rate": 2.25031605562579e-08, "epoch": 4.896424167694205 }, { "current_steps": 3971, "loss": 1.82, "learning_rate": 2.225031605562579e-08, "epoch": 4.897657213316893 }, { "current_steps": 3972, "loss": 2.2203, "learning_rate": 2.199747155499368e-08, "epoch": 4.898890258939581 }, { "current_steps": 3973, "loss": 1.69, "learning_rate": 2.1744627054361566e-08, "epoch": 4.9001233045622685 }, { "current_steps": 3974, "loss": 2.3301, "learning_rate": 2.1491782553729454e-08, "epoch": 4.901356350184956 }, { "current_steps": 3975, "loss": 2.3685, "learning_rate": 2.1238938053097345e-08, "epoch": 4.902589395807645 }, { "current_steps": 3976, "loss": 1.932, "learning_rate": 2.0986093552465232e-08, "epoch": 4.903822441430333 }, { "current_steps": 3977, "loss": 2.1986, "learning_rate": 2.0733249051833123e-08, "epoch": 4.905055487053021 }, { "current_steps": 3978, "loss": 1.4907, "learning_rate": 2.048040455120101e-08, "epoch": 4.906288532675709 }, { "current_steps": 3979, "loss": 2.3563, "learning_rate": 2.0227560050568902e-08, "epoch": 4.907521578298397 }, { "current_steps": 3980, "loss": 2.1308, "learning_rate": 1.9974715549936787e-08, "epoch": 4.9087546239210855 }, { "current_steps": 3981, "loss": 1.963, "learning_rate": 1.9721871049304678e-08, "epoch": 4.909987669543773 }, { "current_steps": 3982, "loss": 2.2543, "learning_rate": 1.9469026548672565e-08, "epoch": 4.911220715166461 }, { "current_steps": 3983, "loss": 1.9155, "learning_rate": 1.9216182048040453e-08, "epoch": 4.912453760789149 }, { "current_steps": 3984, "loss": 1.9624, "learning_rate": 1.8963337547408344e-08, "epoch": 4.913686806411837 }, { "current_steps": 3985, "loss": 2.2201, "learning_rate": 1.871049304677623e-08, "epoch": 4.914919852034525 }, { "current_steps": 3986, "loss": 2.0608, "learning_rate": 1.8457648546144123e-08, "epoch": 4.916152897657213 }, { "current_steps": 3987, "loss": 1.3438, "learning_rate": 1.8204804045512007e-08, "epoch": 4.917385943279902 }, { "current_steps": 3988, "loss": 2.2527, "learning_rate": 1.7951959544879898e-08, "epoch": 4.9186189889025895 }, { "current_steps": 3989, "loss": 2.1652, "learning_rate": 1.7699115044247786e-08, "epoch": 4.919852034525277 }, { "current_steps": 3990, "loss": 2.0309, "learning_rate": 1.7446270543615677e-08, "epoch": 4.921085080147965 }, { "current_steps": 3991, "loss": 2.2538, "learning_rate": 1.7193426042983564e-08, "epoch": 4.922318125770653 }, { "current_steps": 3992, "loss": 1.9551, "learning_rate": 1.6940581542351455e-08, "epoch": 4.923551171393342 }, { "current_steps": 3993, "loss": 1.768, "learning_rate": 1.668773704171934e-08, "epoch": 4.92478421701603 }, { "current_steps": 3994, "loss": 2.2058, "learning_rate": 1.643489254108723e-08, "epoch": 4.926017262638718 }, { "current_steps": 3995, "loss": 1.9235, "learning_rate": 1.618204804045512e-08, "epoch": 4.927250308261406 }, { "current_steps": 3996, "loss": 1.9354, "learning_rate": 1.592920353982301e-08, "epoch": 4.9284833538840935 }, { "current_steps": 3997, "loss": 1.9066, "learning_rate": 1.5676359039190897e-08, "epoch": 4.929716399506781 }, { "current_steps": 3998, "loss": 1.6991, "learning_rate": 1.5423514538558785e-08, "epoch": 4.930949445129469 }, { "current_steps": 3999, "loss": 2.1897, "learning_rate": 1.5170670037926672e-08, "epoch": 4.932182490752158 }, { "current_steps": 4000, "loss": 1.2265, "learning_rate": 1.4917825537294563e-08, "epoch": 4.933415536374846 }, { "current_steps": 4001, "loss": 2.4528, "learning_rate": 1.4664981036662453e-08, "epoch": 4.934648581997534 }, { "current_steps": 4002, "loss": 2.3238, "learning_rate": 1.441213653603034e-08, "epoch": 4.935881627620222 }, { "current_steps": 4003, "loss": 1.9265, "learning_rate": 1.415929203539823e-08, "epoch": 4.93711467324291 }, { "current_steps": 4004, "loss": 1.8917, "learning_rate": 1.390644753476612e-08, "epoch": 4.938347718865598 }, { "current_steps": 4005, "loss": 2.0928, "learning_rate": 1.3653603034134009e-08, "epoch": 4.939580764488286 }, { "current_steps": 4006, "loss": 2.0113, "learning_rate": 1.3400758533501895e-08, "epoch": 4.940813810110974 }, { "current_steps": 4007, "loss": 2.0823, "learning_rate": 1.3147914032869784e-08, "epoch": 4.942046855733662 }, { "current_steps": 4008, "loss": 1.5061, "learning_rate": 1.2895069532237673e-08, "epoch": 4.94327990135635 }, { "current_steps": 4009, "loss": 1.9407, "learning_rate": 1.2642225031605561e-08, "epoch": 4.944512946979038 }, { "current_steps": 4010, "loss": 2.1621, "learning_rate": 1.238938053097345e-08, "epoch": 4.945745992601727 }, { "current_steps": 4011, "loss": 2.1126, "learning_rate": 1.213653603034134e-08, "epoch": 4.9469790382244145 }, { "current_steps": 4012, "loss": 1.9147, "learning_rate": 1.1883691529709229e-08, "epoch": 4.948212083847102 }, { "current_steps": 4013, "loss": 2.1101, "learning_rate": 1.1630847029077117e-08, "epoch": 4.94944512946979 }, { "current_steps": 4014, "loss": 2.1971, "learning_rate": 1.1378002528445006e-08, "epoch": 4.950678175092478 }, { "current_steps": 4015, "loss": 1.9668, "learning_rate": 1.1125158027812895e-08, "epoch": 4.951911220715166 }, { "current_steps": 4016, "loss": 2.3503, "learning_rate": 1.0872313527180783e-08, "epoch": 4.953144266337855 }, { "current_steps": 4017, "loss": 1.9054, "learning_rate": 1.0619469026548672e-08, "epoch": 4.954377311960543 }, { "current_steps": 4018, "loss": 2.1603, "learning_rate": 1.0366624525916562e-08, "epoch": 4.955610357583231 }, { "current_steps": 4019, "loss": 2.1062, "learning_rate": 1.0113780025284451e-08, "epoch": 4.9568434032059185 }, { "current_steps": 4020, "loss": 1.7886, "learning_rate": 9.860935524652339e-09, "epoch": 4.958076448828606 }, { "current_steps": 4021, "loss": 1.9887, "learning_rate": 9.608091024020226e-09, "epoch": 4.959309494451294 }, { "current_steps": 4022, "loss": 2.4219, "learning_rate": 9.355246523388116e-09, "epoch": 4.960542540073983 }, { "current_steps": 4023, "loss": 2.1142, "learning_rate": 9.102402022756003e-09, "epoch": 4.961775585696671 }, { "current_steps": 4024, "loss": 1.8474, "learning_rate": 8.849557522123893e-09, "epoch": 4.963008631319359 }, { "current_steps": 4025, "loss": 1.8649, "learning_rate": 8.596713021491782e-09, "epoch": 4.964241676942047 }, { "current_steps": 4026, "loss": 2.2141, "learning_rate": 8.34386852085967e-09, "epoch": 4.965474722564735 }, { "current_steps": 4027, "loss": 2.1674, "learning_rate": 8.09102402022756e-09, "epoch": 4.9667077681874225 }, { "current_steps": 4028, "loss": 2.3148, "learning_rate": 7.838179519595449e-09, "epoch": 4.967940813810111 }, { "current_steps": 4029, "loss": 2.0159, "learning_rate": 7.585335018963336e-09, "epoch": 4.969173859432799 }, { "current_steps": 4030, "loss": 2.0708, "learning_rate": 7.332490518331226e-09, "epoch": 4.970406905055487 }, { "current_steps": 4031, "loss": 2.1078, "learning_rate": 7.079646017699115e-09, "epoch": 4.971639950678175 }, { "current_steps": 4032, "loss": 2.4158, "learning_rate": 6.826801517067004e-09, "epoch": 4.972872996300863 }, { "current_steps": 4033, "loss": 2.1796, "learning_rate": 6.573957016434892e-09, "epoch": 4.974106041923552 }, { "current_steps": 4034, "loss": 1.9184, "learning_rate": 6.3211125158027805e-09, "epoch": 4.9753390875462395 }, { "current_steps": 4035, "loss": 1.9674, "learning_rate": 6.06826801517067e-09, "epoch": 4.976572133168927 }, { "current_steps": 4036, "loss": 2.0667, "learning_rate": 5.815423514538558e-09, "epoch": 4.977805178791615 }, { "current_steps": 4037, "loss": 1.8317, "learning_rate": 5.562579013906448e-09, "epoch": 4.979038224414303 }, { "current_steps": 4038, "loss": 1.5056, "learning_rate": 5.309734513274336e-09, "epoch": 4.980271270036991 }, { "current_steps": 4039, "loss": 1.7233, "learning_rate": 5.0568900126422255e-09, "epoch": 4.981504315659679 }, { "current_steps": 4040, "loss": 1.6468, "learning_rate": 4.804045512010113e-09, "epoch": 4.982737361282368 }, { "current_steps": 4041, "loss": 1.7497, "learning_rate": 4.551201011378002e-09, "epoch": 4.983970406905056 }, { "current_steps": 4042, "loss": 2.2929, "learning_rate": 4.298356510745891e-09, "epoch": 4.9852034525277436 }, { "current_steps": 4043, "loss": 1.714, "learning_rate": 4.04551201011378e-09, "epoch": 4.986436498150431 }, { "current_steps": 4044, "loss": 2.2331, "learning_rate": 3.792667509481668e-09, "epoch": 4.987669543773119 }, { "current_steps": 4045, "loss": 1.8932, "learning_rate": 3.5398230088495575e-09, "epoch": 4.988902589395808 }, { "current_steps": 4046, "loss": 2.3041, "learning_rate": 3.286978508217446e-09, "epoch": 4.990135635018496 }, { "current_steps": 4047, "loss": 2.1986, "learning_rate": 3.034134007585335e-09, "epoch": 4.991368680641184 }, { "current_steps": 4048, "loss": 1.8542, "learning_rate": 2.781289506953224e-09, "epoch": 4.992601726263872 }, { "current_steps": 4049, "loss": 1.6498, "learning_rate": 2.5284450063211128e-09, "epoch": 4.99383477188656 }, { "current_steps": 4050, "loss": 1.9363, "learning_rate": 2.275600505689001e-09, "epoch": 4.9950678175092476 }, { "current_steps": 4051, "loss": 2.3164, "learning_rate": 2.02275600505689e-09, "epoch": 4.996300863131935 }, { "current_steps": 4052, "loss": 2.0909, "learning_rate": 1.7699115044247787e-09, "epoch": 4.997533908754624 }, { "current_steps": 4053, "loss": 1.5795, "learning_rate": 1.5170670037926675e-09, "epoch": 4.998766954377312 }, { "current_steps": 4054, "loss": 2.085, "learning_rate": 1.2642225031605564e-09, "epoch": 5.0 }, { "current_steps": 4054, "loss": 2.085, "learning_rate": 1.2642225031605564e-09, "epoch": 5.0 } ]