TinyLlama-Turkish-QLora-0.25 / training_graph.json
Ba2han's picture
Upload 2 files
1c354e6
[
{
"current_steps": 1,
"loss": 2.243,
"learning_rate": 0.0,
"epoch": 0.0
},
{
"current_steps": 3,
"loss": 2.4906,
"learning_rate": 0.0,
"epoch": 0.0
},
{
"current_steps": 5,
"loss": 2.4031,
"learning_rate": 0.0,
"epoch": 0.0
},
{
"current_steps": 7,
"loss": 2.0465,
"learning_rate": 4.9999999999999996e-06,
"epoch": 0.0
},
{
"current_steps": 9,
"loss": 2.1139,
"learning_rate": 4.9999999999999996e-06,
"epoch": 0.0
},
{
"current_steps": 11,
"loss": 2.3079,
"learning_rate": 9.999999999999999e-06,
"epoch": 0.0
},
{
"current_steps": 13,
"loss": 2.4454,
"learning_rate": 9.999999999999999e-06,
"epoch": 0.0
},
{
"current_steps": 15,
"loss": 2.1482,
"learning_rate": 1.4999999999999999e-05,
"epoch": 0.0
},
{
"current_steps": 17,
"loss": 2.2773,
"learning_rate": 1.9999999999999998e-05,
"epoch": 0.0
},
{
"current_steps": 19,
"loss": 2.1211,
"learning_rate": 1.9999999999999998e-05,
"epoch": 0.0
},
{
"current_steps": 21,
"loss": 2.4949,
"learning_rate": 2.4999999999999998e-05,
"epoch": 0.0
},
{
"current_steps": 23,
"loss": 1.9488,
"learning_rate": 2.9999999999999997e-05,
"epoch": 0.0
},
{
"current_steps": 25,
"loss": 2.0129,
"learning_rate": 3.5e-05,
"epoch": 0.0
},
{
"current_steps": 27,
"loss": 2.3205,
"learning_rate": 3.9999999999999996e-05,
"epoch": 0.0
},
{
"current_steps": 29,
"loss": 1.7864,
"learning_rate": 4.4999999999999996e-05,
"epoch": 0.0
},
{
"current_steps": 31,
"loss": 2.161,
"learning_rate": 4.9999999999999996e-05,
"epoch": 0.0
},
{
"current_steps": 33,
"loss": 2.0796,
"learning_rate": 5.499999999999999e-05,
"epoch": 0.0
},
{
"current_steps": 35,
"loss": 2.0092,
"learning_rate": 5.9999999999999995e-05,
"epoch": 0.0
},
{
"current_steps": 37,
"loss": 1.9249,
"learning_rate": 6.5e-05,
"epoch": 0.0
},
{
"current_steps": 39,
"loss": 2.0162,
"learning_rate": 7e-05,
"epoch": 0.0
},
{
"current_steps": 41,
"loss": 1.9155,
"learning_rate": 7.5e-05,
"epoch": 0.0
},
{
"current_steps": 43,
"loss": 1.9234,
"learning_rate": 7.999999999999999e-05,
"epoch": 0.01
},
{
"current_steps": 45,
"loss": 1.8821,
"learning_rate": 8.499999999999999e-05,
"epoch": 0.01
},
{
"current_steps": 47,
"loss": 1.8072,
"learning_rate": 8.999999999999999e-05,
"epoch": 0.01
},
{
"current_steps": 49,
"loss": 1.9472,
"learning_rate": 9.499999999999999e-05,
"epoch": 0.01
},
{
"current_steps": 51,
"loss": 1.8368,
"learning_rate": 9.999999999999999e-05,
"epoch": 0.01
},
{
"current_steps": 53,
"loss": 1.6759,
"learning_rate": 0.00010499999999999999,
"epoch": 0.01
},
{
"current_steps": 55,
"loss": 1.9959,
"learning_rate": 0.00010999999999999998,
"epoch": 0.01
},
{
"current_steps": 57,
"loss": 1.7571,
"learning_rate": 0.000115,
"epoch": 0.01
},
{
"current_steps": 59,
"loss": 1.6964,
"learning_rate": 0.00011999999999999999,
"epoch": 0.01
},
{
"current_steps": 61,
"loss": 1.7602,
"learning_rate": 0.000125,
"epoch": 0.01
},
{
"current_steps": 63,
"loss": 1.6103,
"learning_rate": 0.00013,
"epoch": 0.01
},
{
"current_steps": 65,
"loss": 1.73,
"learning_rate": 0.000135,
"epoch": 0.01
},
{
"current_steps": 67,
"loss": 1.6363,
"learning_rate": 0.00014,
"epoch": 0.01
},
{
"current_steps": 69,
"loss": 1.6342,
"learning_rate": 0.000145,
"epoch": 0.01
},
{
"current_steps": 71,
"loss": 1.6227,
"learning_rate": 0.00015,
"epoch": 0.01
},
{
"current_steps": 73,
"loss": 1.7408,
"learning_rate": 0.000155,
"epoch": 0.01
},
{
"current_steps": 75,
"loss": 1.6787,
"learning_rate": 0.00015999999999999999,
"epoch": 0.01
},
{
"current_steps": 77,
"loss": 1.998,
"learning_rate": 0.000165,
"epoch": 0.01
},
{
"current_steps": 79,
"loss": 1.5218,
"learning_rate": 0.00016999999999999999,
"epoch": 0.01
},
{
"current_steps": 81,
"loss": 1.648,
"learning_rate": 0.000175,
"epoch": 0.01
},
{
"current_steps": 83,
"loss": 1.8393,
"learning_rate": 0.00017999999999999998,
"epoch": 0.01
},
{
"current_steps": 85,
"loss": 1.5411,
"learning_rate": 0.000185,
"epoch": 0.01
},
{
"current_steps": 87,
"loss": 1.5455,
"learning_rate": 0.00018999999999999998,
"epoch": 0.01
},
{
"current_steps": 89,
"loss": 1.8884,
"learning_rate": 0.000195,
"epoch": 0.01
},
{
"current_steps": 91,
"loss": 1.7582,
"learning_rate": 0.00019999999999999998,
"epoch": 0.01
},
{
"current_steps": 93,
"loss": 1.6639,
"learning_rate": 0.000205,
"epoch": 0.01
},
{
"current_steps": 95,
"loss": 1.5912,
"learning_rate": 0.00020999999999999998,
"epoch": 0.01
},
{
"current_steps": 97,
"loss": 1.8708,
"learning_rate": 0.000215,
"epoch": 0.01
},
{
"current_steps": 99,
"loss": 1.776,
"learning_rate": 0.00021999999999999995,
"epoch": 0.01
},
{
"current_steps": 101,
"loss": 1.726,
"learning_rate": 0.000225,
"epoch": 0.01
},
{
"current_steps": 103,
"loss": 1.7652,
"learning_rate": 0.00023,
"epoch": 0.01
},
{
"current_steps": 105,
"loss": 1.655,
"learning_rate": 0.00023499999999999997,
"epoch": 0.01
},
{
"current_steps": 107,
"loss": 1.552,
"learning_rate": 0.00023999999999999998,
"epoch": 0.01
},
{
"current_steps": 109,
"loss": 1.8468,
"learning_rate": 0.000245,
"epoch": 0.01
},
{
"current_steps": 111,
"loss": 1.7365,
"learning_rate": 0.00025,
"epoch": 0.01
},
{
"current_steps": 113,
"loss": 1.8068,
"learning_rate": 0.00025499999999999996,
"epoch": 0.01
},
{
"current_steps": 115,
"loss": 1.6334,
"learning_rate": 0.00026,
"epoch": 0.01
},
{
"current_steps": 117,
"loss": 1.7037,
"learning_rate": 0.000265,
"epoch": 0.01
},
{
"current_steps": 119,
"loss": 1.7575,
"learning_rate": 0.00027,
"epoch": 0.01
},
{
"current_steps": 121,
"loss": 1.5549,
"learning_rate": 0.00027499999999999996,
"epoch": 0.01
},
{
"current_steps": 123,
"loss": 1.7386,
"learning_rate": 0.00028,
"epoch": 0.01
},
{
"current_steps": 125,
"loss": 1.4847,
"learning_rate": 0.000285,
"epoch": 0.01
},
{
"current_steps": 127,
"loss": 1.912,
"learning_rate": 0.00029,
"epoch": 0.01
},
{
"current_steps": 129,
"loss": 1.4259,
"learning_rate": 0.00029499999999999996,
"epoch": 0.02
},
{
"current_steps": 131,
"loss": 1.7833,
"learning_rate": 0.0003,
"epoch": 0.02
},
{
"current_steps": 133,
"loss": 1.6208,
"learning_rate": 0.00029970443349753694,
"epoch": 0.02
},
{
"current_steps": 135,
"loss": 1.7041,
"learning_rate": 0.00029940886699507385,
"epoch": 0.02
},
{
"current_steps": 137,
"loss": 1.6153,
"learning_rate": 0.0002991133004926108,
"epoch": 0.02
},
{
"current_steps": 139,
"loss": 1.716,
"learning_rate": 0.00029881773399014774,
"epoch": 0.02
},
{
"current_steps": 141,
"loss": 1.5853,
"learning_rate": 0.0002985221674876847,
"epoch": 0.02
},
{
"current_steps": 143,
"loss": 1.728,
"learning_rate": 0.00029822660098522167,
"epoch": 0.02
},
{
"current_steps": 145,
"loss": 1.9886,
"learning_rate": 0.0002979310344827586,
"epoch": 0.02
},
{
"current_steps": 147,
"loss": 1.7207,
"learning_rate": 0.00029763546798029555,
"epoch": 0.02
},
{
"current_steps": 149,
"loss": 1.6006,
"learning_rate": 0.00029733990147783247,
"epoch": 0.02
},
{
"current_steps": 151,
"loss": 1.3388,
"learning_rate": 0.00029704433497536943,
"epoch": 0.02
},
{
"current_steps": 153,
"loss": 1.5003,
"learning_rate": 0.00029674876847290635,
"epoch": 0.02
},
{
"current_steps": 155,
"loss": 1.479,
"learning_rate": 0.0002964532019704433,
"epoch": 0.02
},
{
"current_steps": 157,
"loss": 1.58,
"learning_rate": 0.0002961576354679803,
"epoch": 0.02
},
{
"current_steps": 159,
"loss": 1.8409,
"learning_rate": 0.0002958620689655172,
"epoch": 0.02
},
{
"current_steps": 161,
"loss": 1.6689,
"learning_rate": 0.00029556650246305416,
"epoch": 0.02
},
{
"current_steps": 163,
"loss": 1.5511,
"learning_rate": 0.0002952709359605911,
"epoch": 0.02
},
{
"current_steps": 165,
"loss": 1.6765,
"learning_rate": 0.00029497536945812805,
"epoch": 0.02
},
{
"current_steps": 167,
"loss": 1.8226,
"learning_rate": 0.000294679802955665,
"epoch": 0.02
},
{
"current_steps": 169,
"loss": 1.7741,
"learning_rate": 0.0002943842364532019,
"epoch": 0.02
},
{
"current_steps": 171,
"loss": 1.59,
"learning_rate": 0.0002940886699507389,
"epoch": 0.02
},
{
"current_steps": 173,
"loss": 1.6031,
"learning_rate": 0.0002937931034482758,
"epoch": 0.02
},
{
"current_steps": 175,
"loss": 1.568,
"learning_rate": 0.0002934975369458128,
"epoch": 0.02
},
{
"current_steps": 177,
"loss": 1.2813,
"learning_rate": 0.00029320197044334974,
"epoch": 0.02
},
{
"current_steps": 179,
"loss": 1.6867,
"learning_rate": 0.00029290640394088666,
"epoch": 0.02
},
{
"current_steps": 181,
"loss": 1.7537,
"learning_rate": 0.0002926108374384236,
"epoch": 0.02
},
{
"current_steps": 183,
"loss": 1.6056,
"learning_rate": 0.00029231527093596054,
"epoch": 0.02
},
{
"current_steps": 185,
"loss": 1.6866,
"learning_rate": 0.0002920197044334975,
"epoch": 0.02
},
{
"current_steps": 187,
"loss": 1.7206,
"learning_rate": 0.0002917241379310344,
"epoch": 0.02
},
{
"current_steps": 189,
"loss": 1.6578,
"learning_rate": 0.0002914285714285714,
"epoch": 0.02
},
{
"current_steps": 191,
"loss": 1.5329,
"learning_rate": 0.00029113300492610836,
"epoch": 0.02
},
{
"current_steps": 193,
"loss": 1.6231,
"learning_rate": 0.00029083743842364527,
"epoch": 0.02
},
{
"current_steps": 195,
"loss": 1.4756,
"learning_rate": 0.00029054187192118224,
"epoch": 0.02
},
{
"current_steps": 197,
"loss": 1.4725,
"learning_rate": 0.00029024630541871915,
"epoch": 0.02
},
{
"current_steps": 199,
"loss": 1.7892,
"learning_rate": 0.0002899507389162561,
"epoch": 0.02
},
{
"current_steps": 201,
"loss": 1.7455,
"learning_rate": 0.0002896551724137931,
"epoch": 0.02
},
{
"current_steps": 203,
"loss": 1.6733,
"learning_rate": 0.00028935960591133,
"epoch": 0.02
},
{
"current_steps": 205,
"loss": 1.4954,
"learning_rate": 0.00028906403940886697,
"epoch": 0.02
},
{
"current_steps": 207,
"loss": 1.6448,
"learning_rate": 0.0002887684729064039,
"epoch": 0.02
},
{
"current_steps": 209,
"loss": 1.7882,
"learning_rate": 0.00028847290640394085,
"epoch": 0.02
},
{
"current_steps": 211,
"loss": 1.7796,
"learning_rate": 0.0002881773399014778,
"epoch": 0.02
},
{
"current_steps": 213,
"loss": 1.6399,
"learning_rate": 0.00028788177339901473,
"epoch": 0.02
},
{
"current_steps": 215,
"loss": 1.6645,
"learning_rate": 0.0002875862068965517,
"epoch": 0.03
},
{
"current_steps": 217,
"loss": 1.7824,
"learning_rate": 0.0002872906403940886,
"epoch": 0.03
},
{
"current_steps": 219,
"loss": 1.5053,
"learning_rate": 0.0002869950738916256,
"epoch": 0.03
},
{
"current_steps": 221,
"loss": 1.5483,
"learning_rate": 0.00028669950738916255,
"epoch": 0.03
},
{
"current_steps": 223,
"loss": 1.5874,
"learning_rate": 0.00028640394088669946,
"epoch": 0.03
},
{
"current_steps": 225,
"loss": 1.6807,
"learning_rate": 0.00028610837438423643,
"epoch": 0.03
},
{
"current_steps": 227,
"loss": 1.502,
"learning_rate": 0.00028581280788177334,
"epoch": 0.03
},
{
"current_steps": 229,
"loss": 1.6234,
"learning_rate": 0.0002855172413793103,
"epoch": 0.03
},
{
"current_steps": 231,
"loss": 1.8706,
"learning_rate": 0.0002852216748768473,
"epoch": 0.03
},
{
"current_steps": 233,
"loss": 1.456,
"learning_rate": 0.0002849261083743842,
"epoch": 0.03
},
{
"current_steps": 235,
"loss": 1.4441,
"learning_rate": 0.00028463054187192116,
"epoch": 0.03
},
{
"current_steps": 237,
"loss": 1.7222,
"learning_rate": 0.00028433497536945807,
"epoch": 0.03
},
{
"current_steps": 239,
"loss": 1.6667,
"learning_rate": 0.00028403940886699504,
"epoch": 0.03
},
{
"current_steps": 241,
"loss": 1.6506,
"learning_rate": 0.000283743842364532,
"epoch": 0.03
},
{
"current_steps": 243,
"loss": 1.4267,
"learning_rate": 0.0002834482758620689,
"epoch": 0.03
},
{
"current_steps": 245,
"loss": 1.6695,
"learning_rate": 0.0002831527093596059,
"epoch": 0.03
},
{
"current_steps": 247,
"loss": 1.6691,
"learning_rate": 0.0002828571428571428,
"epoch": 0.03
},
{
"current_steps": 249,
"loss": 1.5338,
"learning_rate": 0.00028256157635467977,
"epoch": 0.03
},
{
"current_steps": 251,
"loss": 1.6988,
"learning_rate": 0.00028226600985221674,
"epoch": 0.03
},
{
"current_steps": 253,
"loss": 1.4844,
"learning_rate": 0.00028197044334975365,
"epoch": 0.03
},
{
"current_steps": 255,
"loss": 1.7708,
"learning_rate": 0.0002816748768472906,
"epoch": 0.03
},
{
"current_steps": 257,
"loss": 1.6342,
"learning_rate": 0.00028137931034482753,
"epoch": 0.03
},
{
"current_steps": 259,
"loss": 1.4992,
"learning_rate": 0.0002810837438423645,
"epoch": 0.03
},
{
"current_steps": 261,
"loss": 1.5933,
"learning_rate": 0.00028078817733990147,
"epoch": 0.03
},
{
"current_steps": 263,
"loss": 1.5269,
"learning_rate": 0.0002804926108374384,
"epoch": 0.03
},
{
"current_steps": 265,
"loss": 1.7497,
"learning_rate": 0.00028019704433497535,
"epoch": 0.03
},
{
"current_steps": 267,
"loss": 1.3407,
"learning_rate": 0.00027990147783251226,
"epoch": 0.03
},
{
"current_steps": 269,
"loss": 1.611,
"learning_rate": 0.00027960591133004923,
"epoch": 0.03
},
{
"current_steps": 271,
"loss": 1.5756,
"learning_rate": 0.0002793103448275862,
"epoch": 0.03
},
{
"current_steps": 273,
"loss": 1.566,
"learning_rate": 0.0002790147783251231,
"epoch": 0.03
},
{
"current_steps": 275,
"loss": 1.5732,
"learning_rate": 0.0002787192118226601,
"epoch": 0.03
},
{
"current_steps": 277,
"loss": 1.5015,
"learning_rate": 0.00027842364532019705,
"epoch": 0.03
},
{
"current_steps": 279,
"loss": 1.5882,
"learning_rate": 0.00027812807881773396,
"epoch": 0.03
},
{
"current_steps": 281,
"loss": 1.7877,
"learning_rate": 0.00027783251231527093,
"epoch": 0.03
},
{
"current_steps": 283,
"loss": 1.4578,
"learning_rate": 0.00027753694581280784,
"epoch": 0.03
},
{
"current_steps": 285,
"loss": 1.4568,
"learning_rate": 0.0002772413793103448,
"epoch": 0.03
},
{
"current_steps": 287,
"loss": 1.4897,
"learning_rate": 0.0002769458128078818,
"epoch": 0.03
},
{
"current_steps": 289,
"loss": 1.642,
"learning_rate": 0.0002766502463054187,
"epoch": 0.03
},
{
"current_steps": 291,
"loss": 1.5697,
"learning_rate": 0.00027635467980295566,
"epoch": 0.03
},
{
"current_steps": 293,
"loss": 1.4063,
"learning_rate": 0.00027605911330049257,
"epoch": 0.03
},
{
"current_steps": 295,
"loss": 1.5415,
"learning_rate": 0.00027576354679802954,
"epoch": 0.03
},
{
"current_steps": 297,
"loss": 1.497,
"learning_rate": 0.0002754679802955665,
"epoch": 0.03
},
{
"current_steps": 299,
"loss": 1.561,
"learning_rate": 0.0002751724137931034,
"epoch": 0.03
},
{
"current_steps": 301,
"loss": 1.5017,
"learning_rate": 0.0002748768472906404,
"epoch": 0.04
},
{
"current_steps": 303,
"loss": 1.478,
"learning_rate": 0.0002745812807881773,
"epoch": 0.04
},
{
"current_steps": 305,
"loss": 1.4427,
"learning_rate": 0.00027428571428571427,
"epoch": 0.04
},
{
"current_steps": 307,
"loss": 1.6162,
"learning_rate": 0.00027399014778325124,
"epoch": 0.04
},
{
"current_steps": 309,
"loss": 1.5174,
"learning_rate": 0.00027369458128078815,
"epoch": 0.04
},
{
"current_steps": 311,
"loss": 1.6452,
"learning_rate": 0.0002733990147783251,
"epoch": 0.04
},
{
"current_steps": 313,
"loss": 1.6152,
"learning_rate": 0.00027310344827586203,
"epoch": 0.04
},
{
"current_steps": 315,
"loss": 1.5749,
"learning_rate": 0.000272807881773399,
"epoch": 0.04
},
{
"current_steps": 317,
"loss": 1.6,
"learning_rate": 0.00027251231527093597,
"epoch": 0.04
},
{
"current_steps": 319,
"loss": 1.4854,
"learning_rate": 0.0002722167487684729,
"epoch": 0.04
},
{
"current_steps": 321,
"loss": 1.5066,
"learning_rate": 0.00027192118226600985,
"epoch": 0.04
},
{
"current_steps": 323,
"loss": 1.6054,
"learning_rate": 0.00027162561576354676,
"epoch": 0.04
},
{
"current_steps": 325,
"loss": 1.5749,
"learning_rate": 0.00027133004926108373,
"epoch": 0.04
},
{
"current_steps": 327,
"loss": 1.6021,
"learning_rate": 0.0002710344827586207,
"epoch": 0.04
},
{
"current_steps": 329,
"loss": 1.4024,
"learning_rate": 0.0002707389162561576,
"epoch": 0.04
},
{
"current_steps": 331,
"loss": 1.6317,
"learning_rate": 0.0002704433497536946,
"epoch": 0.04
},
{
"current_steps": 333,
"loss": 1.4919,
"learning_rate": 0.00027014778325123155,
"epoch": 0.04
},
{
"current_steps": 335,
"loss": 1.4037,
"learning_rate": 0.00026985221674876846,
"epoch": 0.04
},
{
"current_steps": 337,
"loss": 1.5371,
"learning_rate": 0.00026955665024630543,
"epoch": 0.04
},
{
"current_steps": 339,
"loss": 1.5984,
"learning_rate": 0.00026926108374384234,
"epoch": 0.04
},
{
"current_steps": 341,
"loss": 1.5313,
"learning_rate": 0.0002689655172413793,
"epoch": 0.04
},
{
"current_steps": 343,
"loss": 1.586,
"learning_rate": 0.0002686699507389162,
"epoch": 0.04
},
{
"current_steps": 345,
"loss": 1.5473,
"learning_rate": 0.0002683743842364532,
"epoch": 0.04
},
{
"current_steps": 347,
"loss": 1.6482,
"learning_rate": 0.00026807881773399016,
"epoch": 0.04
},
{
"current_steps": 349,
"loss": 1.5501,
"learning_rate": 0.00026778325123152707,
"epoch": 0.04
},
{
"current_steps": 351,
"loss": 1.8194,
"learning_rate": 0.00026748768472906404,
"epoch": 0.04
},
{
"current_steps": 353,
"loss": 1.4082,
"learning_rate": 0.00026719211822660095,
"epoch": 0.04
},
{
"current_steps": 355,
"loss": 1.7386,
"learning_rate": 0.0002668965517241379,
"epoch": 0.04
},
{
"current_steps": 357,
"loss": 1.591,
"learning_rate": 0.0002666009852216749,
"epoch": 0.04
},
{
"current_steps": 359,
"loss": 1.6685,
"learning_rate": 0.0002663054187192118,
"epoch": 0.04
},
{
"current_steps": 361,
"loss": 1.7065,
"learning_rate": 0.00026600985221674877,
"epoch": 0.04
},
{
"current_steps": 363,
"loss": 1.6918,
"learning_rate": 0.0002657142857142857,
"epoch": 0.04
},
{
"current_steps": 365,
"loss": 1.5165,
"learning_rate": 0.00026541871921182265,
"epoch": 0.04
},
{
"current_steps": 367,
"loss": 1.535,
"learning_rate": 0.0002651231527093596,
"epoch": 0.04
},
{
"current_steps": 369,
"loss": 1.6364,
"learning_rate": 0.00026482758620689653,
"epoch": 0.04
},
{
"current_steps": 371,
"loss": 1.4763,
"learning_rate": 0.0002645320197044335,
"epoch": 0.04
},
{
"current_steps": 373,
"loss": 1.4425,
"learning_rate": 0.0002642364532019704,
"epoch": 0.04
},
{
"current_steps": 375,
"loss": 1.3717,
"learning_rate": 0.0002639408866995074,
"epoch": 0.04
},
{
"current_steps": 377,
"loss": 1.4793,
"learning_rate": 0.0002636453201970443,
"epoch": 0.04
},
{
"current_steps": 379,
"loss": 1.7218,
"learning_rate": 0.00026334975369458126,
"epoch": 0.04
},
{
"current_steps": 381,
"loss": 1.4009,
"learning_rate": 0.00026305418719211823,
"epoch": 0.04
},
{
"current_steps": 383,
"loss": 1.7674,
"learning_rate": 0.00026275862068965514,
"epoch": 0.04
},
{
"current_steps": 385,
"loss": 1.5999,
"learning_rate": 0.0002624630541871921,
"epoch": 0.04
},
{
"current_steps": 387,
"loss": 1.5977,
"learning_rate": 0.000262167487684729,
"epoch": 0.05
},
{
"current_steps": 389,
"loss": 1.5001,
"learning_rate": 0.000261871921182266,
"epoch": 0.05
},
{
"current_steps": 391,
"loss": 1.6639,
"learning_rate": 0.00026157635467980296,
"epoch": 0.05
},
{
"current_steps": 393,
"loss": 1.578,
"learning_rate": 0.0002612807881773399,
"epoch": 0.05
},
{
"current_steps": 395,
"loss": 1.7178,
"learning_rate": 0.00026098522167487684,
"epoch": 0.05
},
{
"current_steps": 397,
"loss": 1.6051,
"learning_rate": 0.00026068965517241376,
"epoch": 0.05
},
{
"current_steps": 399,
"loss": 1.7263,
"learning_rate": 0.0002603940886699507,
"epoch": 0.05
},
{
"current_steps": 401,
"loss": 1.4514,
"learning_rate": 0.0002600985221674877,
"epoch": 0.05
},
{
"current_steps": 403,
"loss": 1.5681,
"learning_rate": 0.0002598029556650246,
"epoch": 0.05
},
{
"current_steps": 405,
"loss": 1.3968,
"learning_rate": 0.00025950738916256157,
"epoch": 0.05
},
{
"current_steps": 407,
"loss": 1.5056,
"learning_rate": 0.0002592118226600985,
"epoch": 0.05
},
{
"current_steps": 409,
"loss": 1.4911,
"learning_rate": 0.00025891625615763545,
"epoch": 0.05
},
{
"current_steps": 411,
"loss": 1.1793,
"learning_rate": 0.00025862068965517237,
"epoch": 0.05
},
{
"current_steps": 413,
"loss": 1.6403,
"learning_rate": 0.00025832512315270933,
"epoch": 0.05
},
{
"current_steps": 415,
"loss": 1.4649,
"learning_rate": 0.0002580295566502463,
"epoch": 0.05
},
{
"current_steps": 417,
"loss": 1.4229,
"learning_rate": 0.0002577339901477832,
"epoch": 0.05
},
{
"current_steps": 419,
"loss": 1.5858,
"learning_rate": 0.0002574384236453202,
"epoch": 0.05
},
{
"current_steps": 421,
"loss": 1.5472,
"learning_rate": 0.0002571428571428571,
"epoch": 0.05
},
{
"current_steps": 423,
"loss": 1.521,
"learning_rate": 0.00025684729064039406,
"epoch": 0.05
},
{
"current_steps": 425,
"loss": 1.3612,
"learning_rate": 0.00025655172413793103,
"epoch": 0.05
},
{
"current_steps": 427,
"loss": 1.6238,
"learning_rate": 0.00025625615763546795,
"epoch": 0.05
},
{
"current_steps": 429,
"loss": 1.8308,
"learning_rate": 0.0002559605911330049,
"epoch": 0.05
},
{
"current_steps": 431,
"loss": 1.3374,
"learning_rate": 0.00025566502463054183,
"epoch": 0.05
},
{
"current_steps": 433,
"loss": 1.4902,
"learning_rate": 0.0002553694581280788,
"epoch": 0.05
},
{
"current_steps": 435,
"loss": 1.5373,
"learning_rate": 0.00025507389162561576,
"epoch": 0.05
},
{
"current_steps": 437,
"loss": 1.5823,
"learning_rate": 0.0002547783251231527,
"epoch": 0.05
},
{
"current_steps": 439,
"loss": 1.4721,
"learning_rate": 0.00025448275862068964,
"epoch": 0.05
},
{
"current_steps": 441,
"loss": 1.5371,
"learning_rate": 0.00025418719211822656,
"epoch": 0.05
},
{
"current_steps": 443,
"loss": 1.6405,
"learning_rate": 0.0002538916256157635,
"epoch": 0.05
},
{
"current_steps": 445,
"loss": 1.4804,
"learning_rate": 0.00025359605911330044,
"epoch": 0.05
},
{
"current_steps": 447,
"loss": 1.3803,
"learning_rate": 0.0002533004926108374,
"epoch": 0.05
},
{
"current_steps": 449,
"loss": 1.5504,
"learning_rate": 0.0002530049261083744,
"epoch": 0.05
},
{
"current_steps": 451,
"loss": 1.5377,
"learning_rate": 0.0002527093596059113,
"epoch": 0.05
},
{
"current_steps": 453,
"loss": 1.5253,
"learning_rate": 0.00025241379310344826,
"epoch": 0.05
},
{
"current_steps": 455,
"loss": 1.4337,
"learning_rate": 0.00025211822660098517,
"epoch": 0.05
},
{
"current_steps": 457,
"loss": 1.4687,
"learning_rate": 0.00025182266009852214,
"epoch": 0.05
},
{
"current_steps": 459,
"loss": 1.6957,
"learning_rate": 0.0002515270935960591,
"epoch": 0.05
},
{
"current_steps": 461,
"loss": 1.4894,
"learning_rate": 0.000251231527093596,
"epoch": 0.05
},
{
"current_steps": 463,
"loss": 1.5078,
"learning_rate": 0.000250935960591133,
"epoch": 0.05
},
{
"current_steps": 465,
"loss": 1.5807,
"learning_rate": 0.0002506403940886699,
"epoch": 0.05
},
{
"current_steps": 467,
"loss": 1.5709,
"learning_rate": 0.00025034482758620687,
"epoch": 0.05
},
{
"current_steps": 469,
"loss": 1.6079,
"learning_rate": 0.00025004926108374383,
"epoch": 0.05
},
{
"current_steps": 471,
"loss": 1.6427,
"learning_rate": 0.00024975369458128075,
"epoch": 0.05
},
{
"current_steps": 473,
"loss": 1.3753,
"learning_rate": 0.0002494581280788177,
"epoch": 0.06
},
{
"current_steps": 475,
"loss": 1.6384,
"learning_rate": 0.00024916256157635463,
"epoch": 0.06
},
{
"current_steps": 477,
"loss": 1.5394,
"learning_rate": 0.0002488669950738916,
"epoch": 0.06
},
{
"current_steps": 479,
"loss": 1.4127,
"learning_rate": 0.00024857142857142857,
"epoch": 0.06
},
{
"current_steps": 481,
"loss": 1.5786,
"learning_rate": 0.0002482758620689655,
"epoch": 0.06
},
{
"current_steps": 483,
"loss": 1.4848,
"learning_rate": 0.00024798029556650245,
"epoch": 0.06
},
{
"current_steps": 485,
"loss": 1.4502,
"learning_rate": 0.00024768472906403936,
"epoch": 0.06
},
{
"current_steps": 487,
"loss": 1.6508,
"learning_rate": 0.00024738916256157633,
"epoch": 0.06
},
{
"current_steps": 489,
"loss": 1.5376,
"learning_rate": 0.00024709359605911324,
"epoch": 0.06
},
{
"current_steps": 491,
"loss": 1.5825,
"learning_rate": 0.0002467980295566502,
"epoch": 0.06
},
{
"current_steps": 493,
"loss": 1.6356,
"learning_rate": 0.0002465024630541872,
"epoch": 0.06
},
{
"current_steps": 495,
"loss": 1.5132,
"learning_rate": 0.0002462068965517241,
"epoch": 0.06
},
{
"current_steps": 497,
"loss": 1.6001,
"learning_rate": 0.00024591133004926106,
"epoch": 0.06
},
{
"current_steps": 499,
"loss": 1.4645,
"learning_rate": 0.00024561576354679797,
"epoch": 0.06
},
{
"current_steps": 501,
"loss": 1.592,
"learning_rate": 0.00024532019704433494,
"epoch": 0.06
},
{
"current_steps": 503,
"loss": 1.4911,
"learning_rate": 0.0002450246305418719,
"epoch": 0.06
},
{
"current_steps": 505,
"loss": 1.6054,
"learning_rate": 0.0002447290640394088,
"epoch": 0.06
},
{
"current_steps": 507,
"loss": 1.6558,
"learning_rate": 0.0002444334975369458,
"epoch": 0.06
},
{
"current_steps": 509,
"loss": 1.4264,
"learning_rate": 0.00024413793103448273,
"epoch": 0.06
},
{
"current_steps": 511,
"loss": 1.7094,
"learning_rate": 0.0002438423645320197,
"epoch": 0.06
},
{
"current_steps": 513,
"loss": 1.4223,
"learning_rate": 0.00024354679802955664,
"epoch": 0.06
},
{
"current_steps": 515,
"loss": 1.3422,
"learning_rate": 0.00024325123152709358,
"epoch": 0.06
},
{
"current_steps": 517,
"loss": 1.6417,
"learning_rate": 0.00024295566502463052,
"epoch": 0.06
},
{
"current_steps": 519,
"loss": 1.4039,
"learning_rate": 0.00024266009852216746,
"epoch": 0.06
},
{
"current_steps": 521,
"loss": 1.4574,
"learning_rate": 0.00024236453201970443,
"epoch": 0.06
},
{
"current_steps": 523,
"loss": 1.4048,
"learning_rate": 0.00024206896551724134,
"epoch": 0.06
},
{
"current_steps": 525,
"loss": 1.5909,
"learning_rate": 0.0002417733990147783,
"epoch": 0.06
},
{
"current_steps": 527,
"loss": 1.3579,
"learning_rate": 0.00024147783251231525,
"epoch": 0.06
},
{
"current_steps": 529,
"loss": 1.4976,
"learning_rate": 0.0002411822660098522,
"epoch": 0.06
},
{
"current_steps": 531,
"loss": 1.5886,
"learning_rate": 0.00024088669950738916,
"epoch": 0.06
},
{
"current_steps": 533,
"loss": 1.6522,
"learning_rate": 0.00024059113300492607,
"epoch": 0.06
},
{
"current_steps": 535,
"loss": 1.4365,
"learning_rate": 0.00024029556650246304,
"epoch": 0.06
},
{
"current_steps": 537,
"loss": 1.5958,
"learning_rate": 0.00023999999999999998,
"epoch": 0.06
},
{
"current_steps": 539,
"loss": 1.5983,
"learning_rate": 0.00023970443349753692,
"epoch": 0.06
},
{
"current_steps": 541,
"loss": 1.5317,
"learning_rate": 0.0002394088669950739,
"epoch": 0.06
},
{
"current_steps": 543,
"loss": 1.6437,
"learning_rate": 0.0002391133004926108,
"epoch": 0.06
},
{
"current_steps": 545,
"loss": 1.3556,
"learning_rate": 0.00023881773399014777,
"epoch": 0.06
},
{
"current_steps": 547,
"loss": 1.3637,
"learning_rate": 0.00023852216748768474,
"epoch": 0.06
},
{
"current_steps": 549,
"loss": 1.3729,
"learning_rate": 0.00023822660098522165,
"epoch": 0.06
},
{
"current_steps": 551,
"loss": 1.656,
"learning_rate": 0.00023793103448275862,
"epoch": 0.06
},
{
"current_steps": 553,
"loss": 1.5467,
"learning_rate": 0.00023763546798029553,
"epoch": 0.06
},
{
"current_steps": 555,
"loss": 1.4409,
"learning_rate": 0.0002373399014778325,
"epoch": 0.06
},
{
"current_steps": 557,
"loss": 1.4597,
"learning_rate": 0.0002370443349753694,
"epoch": 0.06
},
{
"current_steps": 559,
"loss": 1.4803,
"learning_rate": 0.00023674876847290638,
"epoch": 0.07
},
{
"current_steps": 561,
"loss": 1.4419,
"learning_rate": 0.00023645320197044335,
"epoch": 0.07
},
{
"current_steps": 563,
"loss": 1.5701,
"learning_rate": 0.00023615763546798026,
"epoch": 0.07
},
{
"current_steps": 565,
"loss": 1.543,
"learning_rate": 0.00023586206896551723,
"epoch": 0.07
},
{
"current_steps": 567,
"loss": 1.625,
"learning_rate": 0.00023556650246305414,
"epoch": 0.07
},
{
"current_steps": 569,
"loss": 1.7143,
"learning_rate": 0.0002352709359605911,
"epoch": 0.07
},
{
"current_steps": 571,
"loss": 1.4128,
"learning_rate": 0.00023497536945812808,
"epoch": 0.07
},
{
"current_steps": 573,
"loss": 1.6889,
"learning_rate": 0.000234679802955665,
"epoch": 0.07
},
{
"current_steps": 575,
"loss": 1.5572,
"learning_rate": 0.00023438423645320196,
"epoch": 0.07
},
{
"current_steps": 577,
"loss": 1.3485,
"learning_rate": 0.00023408866995073887,
"epoch": 0.07
},
{
"current_steps": 579,
"loss": 1.6611,
"learning_rate": 0.00023379310344827584,
"epoch": 0.07
},
{
"current_steps": 581,
"loss": 1.503,
"learning_rate": 0.0002334975369458128,
"epoch": 0.07
},
{
"current_steps": 583,
"loss": 1.4661,
"learning_rate": 0.00023320197044334972,
"epoch": 0.07
},
{
"current_steps": 585,
"loss": 1.4585,
"learning_rate": 0.0002329064039408867,
"epoch": 0.07
},
{
"current_steps": 587,
"loss": 1.5381,
"learning_rate": 0.0002326108374384236,
"epoch": 0.07
},
{
"current_steps": 589,
"loss": 1.5061,
"learning_rate": 0.00023231527093596057,
"epoch": 0.07
},
{
"current_steps": 591,
"loss": 1.5201,
"learning_rate": 0.00023201970443349754,
"epoch": 0.07
},
{
"current_steps": 593,
"loss": 1.5733,
"learning_rate": 0.00023172413793103445,
"epoch": 0.07
},
{
"current_steps": 595,
"loss": 1.4034,
"learning_rate": 0.00023142857142857142,
"epoch": 0.07
},
{
"current_steps": 597,
"loss": 1.6211,
"learning_rate": 0.00023113300492610833,
"epoch": 0.07
},
{
"current_steps": 599,
"loss": 1.6118,
"learning_rate": 0.0002308374384236453,
"epoch": 0.07
},
{
"current_steps": 601,
"loss": 1.5431,
"learning_rate": 0.00023054187192118224,
"epoch": 0.07
},
{
"current_steps": 603,
"loss": 1.6332,
"learning_rate": 0.00023024630541871918,
"epoch": 0.07
},
{
"current_steps": 605,
"loss": 1.4212,
"learning_rate": 0.00022995073891625615,
"epoch": 0.07
},
{
"current_steps": 607,
"loss": 1.5234,
"learning_rate": 0.00022965517241379306,
"epoch": 0.07
},
{
"current_steps": 609,
"loss": 1.7482,
"learning_rate": 0.00022935960591133003,
"epoch": 0.07
},
{
"current_steps": 611,
"loss": 1.483,
"learning_rate": 0.00022906403940886697,
"epoch": 0.07
},
{
"current_steps": 613,
"loss": 1.3697,
"learning_rate": 0.0002287684729064039,
"epoch": 0.07
},
{
"current_steps": 615,
"loss": 1.484,
"learning_rate": 0.00022847290640394088,
"epoch": 0.07
},
{
"current_steps": 617,
"loss": 1.7224,
"learning_rate": 0.0002281773399014778,
"epoch": 0.07
},
{
"current_steps": 619,
"loss": 1.3295,
"learning_rate": 0.00022788177339901476,
"epoch": 0.07
},
{
"current_steps": 621,
"loss": 1.7212,
"learning_rate": 0.0002275862068965517,
"epoch": 0.07
},
{
"current_steps": 623,
"loss": 1.4627,
"learning_rate": 0.00022729064039408864,
"epoch": 0.07
},
{
"current_steps": 625,
"loss": 1.6296,
"learning_rate": 0.0002269950738916256,
"epoch": 0.07
},
{
"current_steps": 627,
"loss": 1.6696,
"learning_rate": 0.00022669950738916255,
"epoch": 0.07
},
{
"current_steps": 629,
"loss": 1.5944,
"learning_rate": 0.0002264039408866995,
"epoch": 0.07
},
{
"current_steps": 631,
"loss": 1.6079,
"learning_rate": 0.00022610837438423643,
"epoch": 0.07
},
{
"current_steps": 633,
"loss": 1.4051,
"learning_rate": 0.00022581280788177337,
"epoch": 0.07
},
{
"current_steps": 635,
"loss": 1.4022,
"learning_rate": 0.00022551724137931031,
"epoch": 0.07
},
{
"current_steps": 637,
"loss": 1.6161,
"learning_rate": 0.00022522167487684728,
"epoch": 0.07
},
{
"current_steps": 639,
"loss": 1.5758,
"learning_rate": 0.00022492610837438422,
"epoch": 0.07
},
{
"current_steps": 641,
"loss": 1.4163,
"learning_rate": 0.00022463054187192116,
"epoch": 0.07
},
{
"current_steps": 643,
"loss": 1.4388,
"learning_rate": 0.0002243349753694581,
"epoch": 0.07
},
{
"current_steps": 645,
"loss": 1.5208,
"learning_rate": 0.00022403940886699504,
"epoch": 0.08
},
{
"current_steps": 647,
"loss": 1.6124,
"learning_rate": 0.000223743842364532,
"epoch": 0.08
},
{
"current_steps": 649,
"loss": 1.6304,
"learning_rate": 0.00022344827586206895,
"epoch": 0.08
},
{
"current_steps": 651,
"loss": 1.2748,
"learning_rate": 0.0002231527093596059,
"epoch": 0.08
},
{
"current_steps": 653,
"loss": 1.4693,
"learning_rate": 0.00022285714285714283,
"epoch": 0.08
},
{
"current_steps": 655,
"loss": 1.4548,
"learning_rate": 0.00022256157635467977,
"epoch": 0.08
},
{
"current_steps": 657,
"loss": 1.5686,
"learning_rate": 0.00022226600985221674,
"epoch": 0.08
},
{
"current_steps": 659,
"loss": 1.467,
"learning_rate": 0.00022197044334975368,
"epoch": 0.08
},
{
"current_steps": 661,
"loss": 1.4681,
"learning_rate": 0.00022167487684729062,
"epoch": 0.08
},
{
"current_steps": 663,
"loss": 1.6515,
"learning_rate": 0.00022137931034482756,
"epoch": 0.08
},
{
"current_steps": 665,
"loss": 1.4428,
"learning_rate": 0.0002210837438423645,
"epoch": 0.08
},
{
"current_steps": 667,
"loss": 1.4966,
"learning_rate": 0.00022078817733990147,
"epoch": 0.08
},
{
"current_steps": 669,
"loss": 1.526,
"learning_rate": 0.00022049261083743839,
"epoch": 0.08
},
{
"current_steps": 671,
"loss": 1.5171,
"learning_rate": 0.00022019704433497535,
"epoch": 0.08
},
{
"current_steps": 673,
"loss": 1.478,
"learning_rate": 0.0002199014778325123,
"epoch": 0.08
},
{
"current_steps": 675,
"loss": 1.3102,
"learning_rate": 0.00021960591133004923,
"epoch": 0.08
},
{
"current_steps": 677,
"loss": 1.4725,
"learning_rate": 0.0002193103448275862,
"epoch": 0.08
},
{
"current_steps": 679,
"loss": 1.6475,
"learning_rate": 0.00021901477832512312,
"epoch": 0.08
},
{
"current_steps": 681,
"loss": 1.4398,
"learning_rate": 0.00021871921182266008,
"epoch": 0.08
},
{
"current_steps": 683,
"loss": 1.4224,
"learning_rate": 0.00021842364532019705,
"epoch": 0.08
},
{
"current_steps": 685,
"loss": 1.4644,
"learning_rate": 0.00021812807881773397,
"epoch": 0.08
},
{
"current_steps": 687,
"loss": 1.4869,
"learning_rate": 0.00021783251231527093,
"epoch": 0.08
},
{
"current_steps": 689,
"loss": 1.5006,
"learning_rate": 0.00021753694581280785,
"epoch": 0.08
},
{
"current_steps": 691,
"loss": 1.385,
"learning_rate": 0.00021724137931034481,
"epoch": 0.08
},
{
"current_steps": 693,
"loss": 1.4807,
"learning_rate": 0.00021694581280788178,
"epoch": 0.08
},
{
"current_steps": 695,
"loss": 1.4785,
"learning_rate": 0.0002166502463054187,
"epoch": 0.08
},
{
"current_steps": 697,
"loss": 1.3646,
"learning_rate": 0.00021635467980295566,
"epoch": 0.08
},
{
"current_steps": 699,
"loss": 1.5786,
"learning_rate": 0.00021605911330049258,
"epoch": 0.08
},
{
"current_steps": 701,
"loss": 1.5026,
"learning_rate": 0.00021576354679802954,
"epoch": 0.08
},
{
"current_steps": 703,
"loss": 1.5796,
"learning_rate": 0.0002154679802955665,
"epoch": 0.08
},
{
"current_steps": 705,
"loss": 1.3805,
"learning_rate": 0.00021517241379310343,
"epoch": 0.08
},
{
"current_steps": 707,
"loss": 1.3003,
"learning_rate": 0.0002148768472906404,
"epoch": 0.08
},
{
"current_steps": 709,
"loss": 1.6073,
"learning_rate": 0.0002145812807881773,
"epoch": 0.08
},
{
"current_steps": 711,
"loss": 1.6322,
"learning_rate": 0.00021428571428571427,
"epoch": 0.08
},
{
"current_steps": 713,
"loss": 1.4224,
"learning_rate": 0.0002139901477832512,
"epoch": 0.08
},
{
"current_steps": 715,
"loss": 1.5127,
"learning_rate": 0.00021369458128078816,
"epoch": 0.08
},
{
"current_steps": 717,
"loss": 1.5959,
"learning_rate": 0.00021339901477832512,
"epoch": 0.08
},
{
"current_steps": 719,
"loss": 1.5482,
"learning_rate": 0.00021310344827586204,
"epoch": 0.08
},
{
"current_steps": 721,
"loss": 1.4671,
"learning_rate": 0.000212807881773399,
"epoch": 0.08
},
{
"current_steps": 723,
"loss": 1.3553,
"learning_rate": 0.00021251231527093592,
"epoch": 0.08
},
{
"current_steps": 725,
"loss": 1.3771,
"learning_rate": 0.00021221674876847289,
"epoch": 0.08
},
{
"current_steps": 727,
"loss": 1.3838,
"learning_rate": 0.00021192118226600985,
"epoch": 0.08
},
{
"current_steps": 729,
"loss": 1.5396,
"learning_rate": 0.00021162561576354677,
"epoch": 0.08
},
{
"current_steps": 731,
"loss": 1.4486,
"learning_rate": 0.00021133004926108374,
"epoch": 0.09
},
{
"current_steps": 733,
"loss": 1.7393,
"learning_rate": 0.00021103448275862065,
"epoch": 0.09
},
{
"current_steps": 735,
"loss": 1.6453,
"learning_rate": 0.00021073891625615762,
"epoch": 0.09
},
{
"current_steps": 737,
"loss": 1.531,
"learning_rate": 0.00021044334975369458,
"epoch": 0.09
},
{
"current_steps": 739,
"loss": 1.4001,
"learning_rate": 0.0002101477832512315,
"epoch": 0.09
},
{
"current_steps": 741,
"loss": 1.715,
"learning_rate": 0.00020985221674876847,
"epoch": 0.09
},
{
"current_steps": 743,
"loss": 1.5752,
"learning_rate": 0.00020955665024630538,
"epoch": 0.09
},
{
"current_steps": 745,
"loss": 1.6571,
"learning_rate": 0.00020926108374384235,
"epoch": 0.09
},
{
"current_steps": 747,
"loss": 1.3944,
"learning_rate": 0.0002089655172413793,
"epoch": 0.09
},
{
"current_steps": 749,
"loss": 1.4435,
"learning_rate": 0.00020866995073891623,
"epoch": 0.09
},
{
"current_steps": 751,
"loss": 1.5315,
"learning_rate": 0.0002083743842364532,
"epoch": 0.09
},
{
"current_steps": 753,
"loss": 1.5234,
"learning_rate": 0.0002080788177339901,
"epoch": 0.09
},
{
"current_steps": 755,
"loss": 1.3887,
"learning_rate": 0.00020778325123152708,
"epoch": 0.09
},
{
"current_steps": 757,
"loss": 1.7222,
"learning_rate": 0.00020748768472906402,
"epoch": 0.09
},
{
"current_steps": 759,
"loss": 1.3696,
"learning_rate": 0.00020719211822660096,
"epoch": 0.09
},
{
"current_steps": 761,
"loss": 1.7286,
"learning_rate": 0.00020689655172413793,
"epoch": 0.09
},
{
"current_steps": 763,
"loss": 1.4161,
"learning_rate": 0.00020660098522167484,
"epoch": 0.09
},
{
"current_steps": 765,
"loss": 1.3924,
"learning_rate": 0.0002063054187192118,
"epoch": 0.09
},
{
"current_steps": 767,
"loss": 1.5465,
"learning_rate": 0.00020600985221674875,
"epoch": 0.09
},
{
"current_steps": 769,
"loss": 1.4817,
"learning_rate": 0.0002057142857142857,
"epoch": 0.09
},
{
"current_steps": 771,
"loss": 1.3681,
"learning_rate": 0.00020541871921182266,
"epoch": 0.09
},
{
"current_steps": 773,
"loss": 1.4576,
"learning_rate": 0.0002051231527093596,
"epoch": 0.09
},
{
"current_steps": 775,
"loss": 1.6408,
"learning_rate": 0.00020482758620689654,
"epoch": 0.09
},
{
"current_steps": 777,
"loss": 1.3061,
"learning_rate": 0.00020453201970443348,
"epoch": 0.09
},
{
"current_steps": 779,
"loss": 1.3361,
"learning_rate": 0.00020423645320197042,
"epoch": 0.09
},
{
"current_steps": 781,
"loss": 1.6419,
"learning_rate": 0.00020394088669950736,
"epoch": 0.09
},
{
"current_steps": 783,
"loss": 1.5729,
"learning_rate": 0.00020364532019704433,
"epoch": 0.09
},
{
"current_steps": 785,
"loss": 1.4599,
"learning_rate": 0.00020334975369458127,
"epoch": 0.09
},
{
"current_steps": 787,
"loss": 1.6963,
"learning_rate": 0.0002030541871921182,
"epoch": 0.09
},
{
"current_steps": 789,
"loss": 1.34,
"learning_rate": 0.00020275862068965515,
"epoch": 0.09
},
{
"current_steps": 791,
"loss": 1.5912,
"learning_rate": 0.0002024630541871921,
"epoch": 0.09
},
{
"current_steps": 793,
"loss": 1.2494,
"learning_rate": 0.00020216748768472906,
"epoch": 0.09
},
{
"current_steps": 795,
"loss": 1.6001,
"learning_rate": 0.000201871921182266,
"epoch": 0.09
},
{
"current_steps": 797,
"loss": 1.476,
"learning_rate": 0.00020157635467980294,
"epoch": 0.09
},
{
"current_steps": 799,
"loss": 1.4616,
"learning_rate": 0.00020128078817733988,
"epoch": 0.09
},
{
"current_steps": 801,
"loss": 1.4727,
"learning_rate": 0.00020098522167487682,
"epoch": 0.09
},
{
"current_steps": 803,
"loss": 1.6162,
"learning_rate": 0.0002006896551724138,
"epoch": 0.09
},
{
"current_steps": 805,
"loss": 1.4267,
"learning_rate": 0.00020039408866995073,
"epoch": 0.09
},
{
"current_steps": 807,
"loss": 1.3787,
"learning_rate": 0.00020009852216748767,
"epoch": 0.09
},
{
"current_steps": 809,
"loss": 1.6201,
"learning_rate": 0.0001998029556650246,
"epoch": 0.09
},
{
"current_steps": 811,
"loss": 1.5628,
"learning_rate": 0.00019950738916256155,
"epoch": 0.09
},
{
"current_steps": 813,
"loss": 1.5113,
"learning_rate": 0.00019921182266009852,
"epoch": 0.09
},
{
"current_steps": 815,
"loss": 1.5544,
"learning_rate": 0.00019891625615763543,
"epoch": 0.09
},
{
"current_steps": 817,
"loss": 1.4982,
"learning_rate": 0.0001986206896551724,
"epoch": 0.1
},
{
"current_steps": 819,
"loss": 1.4648,
"learning_rate": 0.00019832512315270934,
"epoch": 0.1
},
{
"current_steps": 821,
"loss": 1.3321,
"learning_rate": 0.00019802955665024628,
"epoch": 0.1
},
{
"current_steps": 823,
"loss": 1.3048,
"learning_rate": 0.00019773399014778325,
"epoch": 0.1
},
{
"current_steps": 825,
"loss": 1.4533,
"learning_rate": 0.00019743842364532016,
"epoch": 0.1
},
{
"current_steps": 827,
"loss": 1.6091,
"learning_rate": 0.00019714285714285713,
"epoch": 0.1
},
{
"current_steps": 829,
"loss": 1.5188,
"learning_rate": 0.0001968472906403941,
"epoch": 0.1
},
{
"current_steps": 831,
"loss": 1.612,
"learning_rate": 0.000196551724137931,
"epoch": 0.1
},
{
"current_steps": 833,
"loss": 1.3739,
"learning_rate": 0.00019625615763546798,
"epoch": 0.1
},
{
"current_steps": 835,
"loss": 1.5841,
"learning_rate": 0.0001959605911330049,
"epoch": 0.1
},
{
"current_steps": 837,
"loss": 1.382,
"learning_rate": 0.00019566502463054186,
"epoch": 0.1
},
{
"current_steps": 839,
"loss": 1.4158,
"learning_rate": 0.00019536945812807883,
"epoch": 0.1
},
{
"current_steps": 841,
"loss": 1.4126,
"learning_rate": 0.00019507389162561574,
"epoch": 0.1
},
{
"current_steps": 843,
"loss": 1.5635,
"learning_rate": 0.0001947783251231527,
"epoch": 0.1
},
{
"current_steps": 845,
"loss": 1.375,
"learning_rate": 0.00019448275862068962,
"epoch": 0.1
},
{
"current_steps": 847,
"loss": 1.6939,
"learning_rate": 0.0001941871921182266,
"epoch": 0.1
},
{
"current_steps": 849,
"loss": 1.6394,
"learning_rate": 0.00019389162561576356,
"epoch": 0.1
},
{
"current_steps": 851,
"loss": 1.6799,
"learning_rate": 0.00019359605911330047,
"epoch": 0.1
},
{
"current_steps": 853,
"loss": 1.393,
"learning_rate": 0.00019330049261083744,
"epoch": 0.1
},
{
"current_steps": 855,
"loss": 1.3931,
"learning_rate": 0.00019300492610837435,
"epoch": 0.1
},
{
"current_steps": 857,
"loss": 1.5691,
"learning_rate": 0.00019270935960591132,
"epoch": 0.1
},
{
"current_steps": 859,
"loss": 1.3421,
"learning_rate": 0.00019241379310344823,
"epoch": 0.1
},
{
"current_steps": 861,
"loss": 1.3255,
"learning_rate": 0.0001921182266009852,
"epoch": 0.1
},
{
"current_steps": 863,
"loss": 1.468,
"learning_rate": 0.00019182266009852217,
"epoch": 0.1
},
{
"current_steps": 865,
"loss": 1.6568,
"learning_rate": 0.00019152709359605908,
"epoch": 0.1
},
{
"current_steps": 867,
"loss": 1.6134,
"learning_rate": 0.00019123152709359605,
"epoch": 0.1
},
{
"current_steps": 869,
"loss": 1.5257,
"learning_rate": 0.00019093596059113296,
"epoch": 0.1
},
{
"current_steps": 871,
"loss": 1.3118,
"learning_rate": 0.00019064039408866993,
"epoch": 0.1
},
{
"current_steps": 873,
"loss": 1.449,
"learning_rate": 0.0001903448275862069,
"epoch": 0.1
},
{
"current_steps": 875,
"loss": 1.627,
"learning_rate": 0.0001900492610837438,
"epoch": 0.1
},
{
"current_steps": 877,
"loss": 1.4511,
"learning_rate": 0.00018975369458128078,
"epoch": 0.1
},
{
"current_steps": 879,
"loss": 1.6959,
"learning_rate": 0.0001894581280788177,
"epoch": 0.1
},
{
"current_steps": 881,
"loss": 1.5174,
"learning_rate": 0.00018916256157635466,
"epoch": 0.1
},
{
"current_steps": 883,
"loss": 1.5205,
"learning_rate": 0.00018886699507389163,
"epoch": 0.1
},
{
"current_steps": 885,
"loss": 1.4595,
"learning_rate": 0.00018857142857142854,
"epoch": 0.1
},
{
"current_steps": 887,
"loss": 1.5407,
"learning_rate": 0.0001882758620689655,
"epoch": 0.1
},
{
"current_steps": 889,
"loss": 1.8308,
"learning_rate": 0.00018798029556650242,
"epoch": 0.1
},
{
"current_steps": 891,
"loss": 1.3327,
"learning_rate": 0.0001876847290640394,
"epoch": 0.1
},
{
"current_steps": 893,
"loss": 1.5598,
"learning_rate": 0.00018738916256157633,
"epoch": 0.1
},
{
"current_steps": 895,
"loss": 1.5892,
"learning_rate": 0.00018709359605911327,
"epoch": 0.1
},
{
"current_steps": 897,
"loss": 1.5615,
"learning_rate": 0.00018679802955665024,
"epoch": 0.1
},
{
"current_steps": 899,
"loss": 1.3798,
"learning_rate": 0.00018650246305418715,
"epoch": 0.1
},
{
"current_steps": 901,
"loss": 1.7268,
"learning_rate": 0.00018620689655172412,
"epoch": 0.1
},
{
"current_steps": 903,
"loss": 1.3698,
"learning_rate": 0.00018591133004926106,
"epoch": 0.11
},
{
"current_steps": 905,
"loss": 1.6856,
"learning_rate": 0.000185615763546798,
"epoch": 0.11
},
{
"current_steps": 907,
"loss": 1.4168,
"learning_rate": 0.00018532019704433497,
"epoch": 0.11
},
{
"current_steps": 909,
"loss": 1.3779,
"learning_rate": 0.0001850246305418719,
"epoch": 0.11
},
{
"current_steps": 911,
"loss": 1.4477,
"learning_rate": 0.00018472906403940885,
"epoch": 0.11
},
{
"current_steps": 913,
"loss": 1.5091,
"learning_rate": 0.0001844334975369458,
"epoch": 0.11
},
{
"current_steps": 915,
"loss": 1.4676,
"learning_rate": 0.00018413793103448273,
"epoch": 0.11
},
{
"current_steps": 917,
"loss": 1.417,
"learning_rate": 0.0001838423645320197,
"epoch": 0.11
},
{
"current_steps": 919,
"loss": 1.537,
"learning_rate": 0.00018354679802955664,
"epoch": 0.11
},
{
"current_steps": 921,
"loss": 1.4578,
"learning_rate": 0.00018325123152709358,
"epoch": 0.11
},
{
"current_steps": 923,
"loss": 1.5152,
"learning_rate": 0.00018295566502463052,
"epoch": 0.11
},
{
"current_steps": 925,
"loss": 1.5826,
"learning_rate": 0.00018266009852216746,
"epoch": 0.11
},
{
"current_steps": 927,
"loss": 1.4088,
"learning_rate": 0.0001823645320197044,
"epoch": 0.11
},
{
"current_steps": 929,
"loss": 1.4405,
"learning_rate": 0.00018206896551724137,
"epoch": 0.11
},
{
"current_steps": 931,
"loss": 1.4493,
"learning_rate": 0.0001817733990147783,
"epoch": 0.11
},
{
"current_steps": 933,
"loss": 1.4781,
"learning_rate": 0.00018147783251231525,
"epoch": 0.11
},
{
"current_steps": 935,
"loss": 1.5156,
"learning_rate": 0.0001811822660098522,
"epoch": 0.11
},
{
"current_steps": 937,
"loss": 1.3937,
"learning_rate": 0.00018088669950738914,
"epoch": 0.11
},
{
"current_steps": 939,
"loss": 1.4797,
"learning_rate": 0.0001805911330049261,
"epoch": 0.11
},
{
"current_steps": 941,
"loss": 1.504,
"learning_rate": 0.00018029556650246304,
"epoch": 0.11
},
{
"current_steps": 943,
"loss": 1.5977,
"learning_rate": 0.00017999999999999998,
"epoch": 0.11
},
{
"current_steps": 945,
"loss": 1.6771,
"learning_rate": 0.00017970443349753692,
"epoch": 0.11
},
{
"current_steps": 947,
"loss": 1.5078,
"learning_rate": 0.00017940886699507387,
"epoch": 0.11
},
{
"current_steps": 949,
"loss": 1.3564,
"learning_rate": 0.00017911330049261083,
"epoch": 0.11
},
{
"current_steps": 951,
"loss": 1.4263,
"learning_rate": 0.00017881773399014777,
"epoch": 0.11
},
{
"current_steps": 953,
"loss": 1.4985,
"learning_rate": 0.00017852216748768471,
"epoch": 0.11
},
{
"current_steps": 955,
"loss": 1.3852,
"learning_rate": 0.00017822660098522166,
"epoch": 0.11
},
{
"current_steps": 957,
"loss": 1.2973,
"learning_rate": 0.0001779310344827586,
"epoch": 0.11
},
{
"current_steps": 959,
"loss": 1.3745,
"learning_rate": 0.00017763546798029556,
"epoch": 0.11
},
{
"current_steps": 961,
"loss": 1.5937,
"learning_rate": 0.0001773399014778325,
"epoch": 0.11
},
{
"current_steps": 963,
"loss": 1.5452,
"learning_rate": 0.00017704433497536944,
"epoch": 0.11
},
{
"current_steps": 965,
"loss": 1.5924,
"learning_rate": 0.0001767487684729064,
"epoch": 0.11
},
{
"current_steps": 967,
"loss": 1.5265,
"learning_rate": 0.00017645320197044333,
"epoch": 0.11
},
{
"current_steps": 969,
"loss": 1.4742,
"learning_rate": 0.0001761576354679803,
"epoch": 0.11
},
{
"current_steps": 971,
"loss": 1.45,
"learning_rate": 0.0001758620689655172,
"epoch": 0.11
},
{
"current_steps": 973,
"loss": 1.5054,
"learning_rate": 0.00017556650246305418,
"epoch": 0.11
},
{
"current_steps": 975,
"loss": 1.2477,
"learning_rate": 0.00017527093596059114,
"epoch": 0.11
},
{
"current_steps": 977,
"loss": 1.4494,
"learning_rate": 0.00017497536945812806,
"epoch": 0.11
},
{
"current_steps": 979,
"loss": 1.4671,
"learning_rate": 0.00017467980295566502,
"epoch": 0.11
},
{
"current_steps": 981,
"loss": 1.4439,
"learning_rate": 0.00017438423645320194,
"epoch": 0.11
},
{
"current_steps": 983,
"loss": 1.457,
"learning_rate": 0.0001740886699507389,
"epoch": 0.11
},
{
"current_steps": 985,
"loss": 1.3146,
"learning_rate": 0.00017379310344827587,
"epoch": 0.11
},
{
"current_steps": 987,
"loss": 1.6909,
"learning_rate": 0.0001734975369458128,
"epoch": 0.11
},
{
"current_steps": 989,
"loss": 1.4458,
"learning_rate": 0.00017320197044334975,
"epoch": 0.12
},
{
"current_steps": 991,
"loss": 1.4335,
"learning_rate": 0.00017290640394088667,
"epoch": 0.12
},
{
"current_steps": 993,
"loss": 1.4131,
"learning_rate": 0.00017261083743842364,
"epoch": 0.12
},
{
"current_steps": 995,
"loss": 1.6884,
"learning_rate": 0.0001723152709359606,
"epoch": 0.12
},
{
"current_steps": 997,
"loss": 1.4066,
"learning_rate": 0.00017201970443349752,
"epoch": 0.12
},
{
"current_steps": 999,
"loss": 1.4257,
"learning_rate": 0.00017172413793103448,
"epoch": 0.12
},
{
"current_steps": 1001,
"loss": 1.452,
"learning_rate": 0.0001714285714285714,
"epoch": 0.12
},
{
"current_steps": 1003,
"loss": 1.4184,
"learning_rate": 0.00017113300492610837,
"epoch": 0.12
},
{
"current_steps": 1005,
"loss": 1.3286,
"learning_rate": 0.00017083743842364528,
"epoch": 0.12
},
{
"current_steps": 1007,
"loss": 1.3791,
"learning_rate": 0.00017054187192118225,
"epoch": 0.12
},
{
"current_steps": 1009,
"loss": 1.6685,
"learning_rate": 0.00017024630541871921,
"epoch": 0.12
},
{
"current_steps": 1011,
"loss": 1.6164,
"learning_rate": 0.00016995073891625613,
"epoch": 0.12
},
{
"current_steps": 1013,
"loss": 1.4075,
"learning_rate": 0.0001696551724137931,
"epoch": 0.12
},
{
"current_steps": 1015,
"loss": 1.5294,
"learning_rate": 0.00016935960591133,
"epoch": 0.12
},
{
"current_steps": 1017,
"loss": 1.3351,
"learning_rate": 0.00016906403940886698,
"epoch": 0.12
},
{
"current_steps": 1019,
"loss": 1.6398,
"learning_rate": 0.00016876847290640395,
"epoch": 0.12
},
{
"current_steps": 1021,
"loss": 1.2795,
"learning_rate": 0.00016847290640394086,
"epoch": 0.12
},
{
"current_steps": 1023,
"loss": 1.3947,
"learning_rate": 0.00016817733990147783,
"epoch": 0.12
},
{
"current_steps": 1025,
"loss": 1.4945,
"learning_rate": 0.00016788177339901474,
"epoch": 0.12
},
{
"current_steps": 1027,
"loss": 1.4894,
"learning_rate": 0.0001675862068965517,
"epoch": 0.12
},
{
"current_steps": 1029,
"loss": 1.3364,
"learning_rate": 0.00016729064039408868,
"epoch": 0.12
},
{
"current_steps": 1031,
"loss": 1.4511,
"learning_rate": 0.0001669950738916256,
"epoch": 0.12
},
{
"current_steps": 1033,
"loss": 1.4264,
"learning_rate": 0.00016669950738916256,
"epoch": 0.12
},
{
"current_steps": 1035,
"loss": 1.455,
"learning_rate": 0.00016640394088669947,
"epoch": 0.12
},
{
"current_steps": 1037,
"loss": 1.4287,
"learning_rate": 0.00016610837438423644,
"epoch": 0.12
},
{
"current_steps": 1039,
"loss": 1.6167,
"learning_rate": 0.00016581280788177338,
"epoch": 0.12
},
{
"current_steps": 1041,
"loss": 1.6667,
"learning_rate": 0.00016551724137931032,
"epoch": 0.12
},
{
"current_steps": 1043,
"loss": 1.5287,
"learning_rate": 0.0001652216748768473,
"epoch": 0.12
},
{
"current_steps": 1045,
"loss": 1.5189,
"learning_rate": 0.0001649261083743842,
"epoch": 0.12
},
{
"current_steps": 1047,
"loss": 1.6377,
"learning_rate": 0.00016463054187192117,
"epoch": 0.12
},
{
"current_steps": 1049,
"loss": 1.6074,
"learning_rate": 0.0001643349753694581,
"epoch": 0.12
},
{
"current_steps": 1051,
"loss": 1.4497,
"learning_rate": 0.00016403940886699505,
"epoch": 0.12
},
{
"current_steps": 1053,
"loss": 1.5705,
"learning_rate": 0.00016374384236453202,
"epoch": 0.12
},
{
"current_steps": 1055,
"loss": 1.7055,
"learning_rate": 0.00016344827586206896,
"epoch": 0.12
},
{
"current_steps": 1057,
"loss": 1.4076,
"learning_rate": 0.0001631527093596059,
"epoch": 0.12
},
{
"current_steps": 1059,
"loss": 1.4607,
"learning_rate": 0.00016285714285714284,
"epoch": 0.12
},
{
"current_steps": 1061,
"loss": 1.4689,
"learning_rate": 0.00016256157635467978,
"epoch": 0.12
},
{
"current_steps": 1063,
"loss": 1.3713,
"learning_rate": 0.00016226600985221675,
"epoch": 0.12
},
{
"current_steps": 1065,
"loss": 1.407,
"learning_rate": 0.0001619704433497537,
"epoch": 0.12
},
{
"current_steps": 1067,
"loss": 1.5291,
"learning_rate": 0.00016167487684729063,
"epoch": 0.12
},
{
"current_steps": 1069,
"loss": 1.382,
"learning_rate": 0.00016137931034482757,
"epoch": 0.12
},
{
"current_steps": 1071,
"loss": 1.553,
"learning_rate": 0.0001610837438423645,
"epoch": 0.12
},
{
"current_steps": 1073,
"loss": 1.6119,
"learning_rate": 0.00016078817733990145,
"epoch": 0.12
},
{
"current_steps": 1075,
"loss": 1.5013,
"learning_rate": 0.00016049261083743842,
"epoch": 0.13
},
{
"current_steps": 1077,
"loss": 1.6357,
"learning_rate": 0.00016019704433497536,
"epoch": 0.13
},
{
"current_steps": 1079,
"loss": 1.4906,
"learning_rate": 0.0001599014778325123,
"epoch": 0.13
},
{
"current_steps": 1081,
"loss": 1.3798,
"learning_rate": 0.00015960591133004924,
"epoch": 0.13
},
{
"current_steps": 1083,
"loss": 1.4316,
"learning_rate": 0.00015931034482758618,
"epoch": 0.13
},
{
"current_steps": 1085,
"loss": 1.4263,
"learning_rate": 0.00015901477832512315,
"epoch": 0.13
},
{
"current_steps": 1087,
"loss": 1.2947,
"learning_rate": 0.0001587192118226601,
"epoch": 0.13
},
{
"current_steps": 1089,
"loss": 1.5948,
"learning_rate": 0.00015842364532019703,
"epoch": 0.13
},
{
"current_steps": 1091,
"loss": 1.4573,
"learning_rate": 0.00015812807881773397,
"epoch": 0.13
},
{
"current_steps": 1093,
"loss": 1.267,
"learning_rate": 0.0001578325123152709,
"epoch": 0.13
},
{
"current_steps": 1095,
"loss": 1.4615,
"learning_rate": 0.00015753694581280788,
"epoch": 0.13
},
{
"current_steps": 1097,
"loss": 1.4817,
"learning_rate": 0.00015724137931034482,
"epoch": 0.13
},
{
"current_steps": 1099,
"loss": 1.5427,
"learning_rate": 0.00015694581280788176,
"epoch": 0.13
},
{
"current_steps": 1101,
"loss": 1.5629,
"learning_rate": 0.0001566502463054187,
"epoch": 0.13
},
{
"current_steps": 1103,
"loss": 1.5958,
"learning_rate": 0.00015635467980295564,
"epoch": 0.13
},
{
"current_steps": 1105,
"loss": 1.5782,
"learning_rate": 0.0001560591133004926,
"epoch": 0.13
},
{
"current_steps": 1107,
"loss": 1.533,
"learning_rate": 0.00015576354679802955,
"epoch": 0.13
},
{
"current_steps": 1109,
"loss": 1.486,
"learning_rate": 0.0001554679802955665,
"epoch": 0.13
},
{
"current_steps": 1111,
"loss": 1.6014,
"learning_rate": 0.00015517241379310346,
"epoch": 0.13
},
{
"current_steps": 1113,
"loss": 1.5115,
"learning_rate": 0.00015487684729064037,
"epoch": 0.13
},
{
"current_steps": 1115,
"loss": 1.4503,
"learning_rate": 0.00015458128078817734,
"epoch": 0.13
},
{
"current_steps": 1117,
"loss": 1.4544,
"learning_rate": 0.00015428571428571425,
"epoch": 0.13
},
{
"current_steps": 1119,
"loss": 1.4324,
"learning_rate": 0.00015399014778325122,
"epoch": 0.13
},
{
"current_steps": 1121,
"loss": 1.6173,
"learning_rate": 0.0001536945812807882,
"epoch": 0.13
},
{
"current_steps": 1123,
"loss": 1.2486,
"learning_rate": 0.0001533990147783251,
"epoch": 0.13
},
{
"current_steps": 1125,
"loss": 1.3345,
"learning_rate": 0.00015310344827586207,
"epoch": 0.13
},
{
"current_steps": 1127,
"loss": 1.3289,
"learning_rate": 0.00015280788177339898,
"epoch": 0.13
},
{
"current_steps": 1129,
"loss": 1.4645,
"learning_rate": 0.00015251231527093595,
"epoch": 0.13
},
{
"current_steps": 1131,
"loss": 1.2622,
"learning_rate": 0.00015221674876847292,
"epoch": 0.13
},
{
"current_steps": 1133,
"loss": 1.6447,
"learning_rate": 0.00015192118226600983,
"epoch": 0.13
},
{
"current_steps": 1135,
"loss": 1.4101,
"learning_rate": 0.0001516256157635468,
"epoch": 0.13
},
{
"current_steps": 1137,
"loss": 1.2335,
"learning_rate": 0.0001513300492610837,
"epoch": 0.13
},
{
"current_steps": 1139,
"loss": 1.5197,
"learning_rate": 0.00015103448275862068,
"epoch": 0.13
},
{
"current_steps": 1141,
"loss": 1.4544,
"learning_rate": 0.00015073891625615765,
"epoch": 0.13
},
{
"current_steps": 1143,
"loss": 1.653,
"learning_rate": 0.00015044334975369456,
"epoch": 0.13
},
{
"current_steps": 1145,
"loss": 1.4697,
"learning_rate": 0.00015014778325123153,
"epoch": 0.13
},
{
"current_steps": 1147,
"loss": 1.5149,
"learning_rate": 0.00014985221674876847,
"epoch": 0.13
},
{
"current_steps": 1149,
"loss": 1.5345,
"learning_rate": 0.0001495566502463054,
"epoch": 0.13
},
{
"current_steps": 1151,
"loss": 1.4074,
"learning_rate": 0.00014926108374384235,
"epoch": 0.13
},
{
"current_steps": 1153,
"loss": 1.6013,
"learning_rate": 0.0001489655172413793,
"epoch": 0.13
},
{
"current_steps": 1155,
"loss": 1.4285,
"learning_rate": 0.00014866995073891623,
"epoch": 0.13
},
{
"current_steps": 1157,
"loss": 1.4755,
"learning_rate": 0.00014837438423645317,
"epoch": 0.13
},
{
"current_steps": 1159,
"loss": 1.3552,
"learning_rate": 0.00014807881773399014,
"epoch": 0.13
},
{
"current_steps": 1161,
"loss": 1.3876,
"learning_rate": 0.00014778325123152708,
"epoch": 0.14
},
{
"current_steps": 1163,
"loss": 1.3403,
"learning_rate": 0.00014748768472906402,
"epoch": 0.14
},
{
"current_steps": 1165,
"loss": 1.305,
"learning_rate": 0.00014719211822660096,
"epoch": 0.14
},
{
"current_steps": 1167,
"loss": 1.523,
"learning_rate": 0.0001468965517241379,
"epoch": 0.14
},
{
"current_steps": 1169,
"loss": 1.575,
"learning_rate": 0.00014660098522167487,
"epoch": 0.14
},
{
"current_steps": 1171,
"loss": 1.6639,
"learning_rate": 0.0001463054187192118,
"epoch": 0.14
},
{
"current_steps": 1173,
"loss": 1.4092,
"learning_rate": 0.00014600985221674875,
"epoch": 0.14
},
{
"current_steps": 1175,
"loss": 1.6079,
"learning_rate": 0.0001457142857142857,
"epoch": 0.14
},
{
"current_steps": 1177,
"loss": 1.6555,
"learning_rate": 0.00014541871921182263,
"epoch": 0.14
},
{
"current_steps": 1179,
"loss": 1.5001,
"learning_rate": 0.00014512315270935958,
"epoch": 0.14
},
{
"current_steps": 1181,
"loss": 1.4116,
"learning_rate": 0.00014482758620689654,
"epoch": 0.14
},
{
"current_steps": 1183,
"loss": 1.4401,
"learning_rate": 0.00014453201970443348,
"epoch": 0.14
},
{
"current_steps": 1185,
"loss": 1.6073,
"learning_rate": 0.00014423645320197042,
"epoch": 0.14
},
{
"current_steps": 1187,
"loss": 1.5916,
"learning_rate": 0.00014394088669950736,
"epoch": 0.14
},
{
"current_steps": 1189,
"loss": 1.4508,
"learning_rate": 0.0001436453201970443,
"epoch": 0.14
},
{
"current_steps": 1191,
"loss": 1.2755,
"learning_rate": 0.00014334975369458127,
"epoch": 0.14
},
{
"current_steps": 1193,
"loss": 1.4433,
"learning_rate": 0.00014305418719211821,
"epoch": 0.14
},
{
"current_steps": 1195,
"loss": 1.6638,
"learning_rate": 0.00014275862068965515,
"epoch": 0.14
},
{
"current_steps": 1197,
"loss": 1.5391,
"learning_rate": 0.0001424630541871921,
"epoch": 0.14
},
{
"current_steps": 1199,
"loss": 1.2966,
"learning_rate": 0.00014216748768472904,
"epoch": 0.14
},
{
"current_steps": 1201,
"loss": 1.6218,
"learning_rate": 0.000141871921182266,
"epoch": 0.14
},
{
"current_steps": 1203,
"loss": 1.3367,
"learning_rate": 0.00014157635467980294,
"epoch": 0.14
},
{
"current_steps": 1205,
"loss": 1.5205,
"learning_rate": 0.00014128078817733988,
"epoch": 0.14
},
{
"current_steps": 1207,
"loss": 1.2991,
"learning_rate": 0.00014098522167487683,
"epoch": 0.14
},
{
"current_steps": 1209,
"loss": 1.5471,
"learning_rate": 0.00014068965517241377,
"epoch": 0.14
},
{
"current_steps": 1211,
"loss": 1.3356,
"learning_rate": 0.00014039408866995073,
"epoch": 0.14
},
{
"current_steps": 1213,
"loss": 1.4511,
"learning_rate": 0.00014009852216748767,
"epoch": 0.14
},
{
"current_steps": 1215,
"loss": 1.3815,
"learning_rate": 0.00013980295566502461,
"epoch": 0.14
},
{
"current_steps": 1217,
"loss": 1.5201,
"learning_rate": 0.00013950738916256156,
"epoch": 0.14
},
{
"current_steps": 1219,
"loss": 1.5778,
"learning_rate": 0.00013921182266009852,
"epoch": 0.14
},
{
"current_steps": 1221,
"loss": 1.3516,
"learning_rate": 0.00013891625615763546,
"epoch": 0.14
},
{
"current_steps": 1223,
"loss": 1.4532,
"learning_rate": 0.0001386206896551724,
"epoch": 0.14
},
{
"current_steps": 1225,
"loss": 1.468,
"learning_rate": 0.00013832512315270935,
"epoch": 0.14
},
{
"current_steps": 1227,
"loss": 1.4443,
"learning_rate": 0.00013802955665024629,
"epoch": 0.14
},
{
"current_steps": 1229,
"loss": 1.2362,
"learning_rate": 0.00013773399014778325,
"epoch": 0.14
},
{
"current_steps": 1231,
"loss": 1.6585,
"learning_rate": 0.0001374384236453202,
"epoch": 0.14
},
{
"current_steps": 1233,
"loss": 1.7561,
"learning_rate": 0.00013714285714285713,
"epoch": 0.14
},
{
"current_steps": 1235,
"loss": 1.6997,
"learning_rate": 0.00013684729064039408,
"epoch": 0.14
},
{
"current_steps": 1237,
"loss": 1.4349,
"learning_rate": 0.00013655172413793102,
"epoch": 0.14
},
{
"current_steps": 1239,
"loss": 1.3657,
"learning_rate": 0.00013625615763546798,
"epoch": 0.14
},
{
"current_steps": 1241,
"loss": 1.6985,
"learning_rate": 0.00013596059113300492,
"epoch": 0.14
},
{
"current_steps": 1243,
"loss": 1.5181,
"learning_rate": 0.00013566502463054186,
"epoch": 0.14
},
{
"current_steps": 1245,
"loss": 1.4471,
"learning_rate": 0.0001353694581280788,
"epoch": 0.14
},
{
"current_steps": 1247,
"loss": 1.65,
"learning_rate": 0.00013507389162561577,
"epoch": 0.15
},
{
"current_steps": 1249,
"loss": 1.1517,
"learning_rate": 0.00013477832512315271,
"epoch": 0.15
},
{
"current_steps": 1251,
"loss": 1.4166,
"learning_rate": 0.00013448275862068965,
"epoch": 0.15
},
{
"current_steps": 1253,
"loss": 1.4719,
"learning_rate": 0.0001341871921182266,
"epoch": 0.15
},
{
"current_steps": 1255,
"loss": 1.4172,
"learning_rate": 0.00013389162561576354,
"epoch": 0.15
},
{
"current_steps": 1257,
"loss": 1.3626,
"learning_rate": 0.00013359605911330048,
"epoch": 0.15
},
{
"current_steps": 1259,
"loss": 1.4102,
"learning_rate": 0.00013330049261083744,
"epoch": 0.15
},
{
"current_steps": 1261,
"loss": 1.3181,
"learning_rate": 0.00013300492610837438,
"epoch": 0.15
},
{
"current_steps": 1263,
"loss": 1.4147,
"learning_rate": 0.00013270935960591133,
"epoch": 0.15
},
{
"current_steps": 1265,
"loss": 1.5692,
"learning_rate": 0.00013241379310344827,
"epoch": 0.15
},
{
"current_steps": 1267,
"loss": 1.3556,
"learning_rate": 0.0001321182266009852,
"epoch": 0.15
},
{
"current_steps": 1269,
"loss": 1.5808,
"learning_rate": 0.00013182266009852215,
"epoch": 0.15
},
{
"current_steps": 1271,
"loss": 1.2675,
"learning_rate": 0.00013152709359605912,
"epoch": 0.15
},
{
"current_steps": 1273,
"loss": 1.4857,
"learning_rate": 0.00013123152709359606,
"epoch": 0.15
},
{
"current_steps": 1275,
"loss": 1.445,
"learning_rate": 0.000130935960591133,
"epoch": 0.15
},
{
"current_steps": 1277,
"loss": 1.6075,
"learning_rate": 0.00013064039408866994,
"epoch": 0.15
},
{
"current_steps": 1279,
"loss": 1.3829,
"learning_rate": 0.00013034482758620688,
"epoch": 0.15
},
{
"current_steps": 1281,
"loss": 1.3487,
"learning_rate": 0.00013004926108374385,
"epoch": 0.15
},
{
"current_steps": 1283,
"loss": 1.3105,
"learning_rate": 0.00012975369458128079,
"epoch": 0.15
},
{
"current_steps": 1285,
"loss": 1.5299,
"learning_rate": 0.00012945812807881773,
"epoch": 0.15
},
{
"current_steps": 1287,
"loss": 1.5279,
"learning_rate": 0.00012916256157635467,
"epoch": 0.15
},
{
"current_steps": 1289,
"loss": 1.5722,
"learning_rate": 0.0001288669950738916,
"epoch": 0.15
},
{
"current_steps": 1291,
"loss": 1.4816,
"learning_rate": 0.00012857142857142855,
"epoch": 0.15
},
{
"current_steps": 1293,
"loss": 1.3207,
"learning_rate": 0.00012827586206896552,
"epoch": 0.15
},
{
"current_steps": 1295,
"loss": 1.566,
"learning_rate": 0.00012798029556650246,
"epoch": 0.15
},
{
"current_steps": 1297,
"loss": 1.5285,
"learning_rate": 0.0001276847290640394,
"epoch": 0.15
},
{
"current_steps": 1299,
"loss": 1.6871,
"learning_rate": 0.00012738916256157634,
"epoch": 0.15
},
{
"current_steps": 1301,
"loss": 1.3097,
"learning_rate": 0.00012709359605911328,
"epoch": 0.15
},
{
"current_steps": 1303,
"loss": 1.408,
"learning_rate": 0.00012679802955665022,
"epoch": 0.15
},
{
"current_steps": 1305,
"loss": 1.3767,
"learning_rate": 0.0001265024630541872,
"epoch": 0.15
},
{
"current_steps": 1307,
"loss": 1.3903,
"learning_rate": 0.00012620689655172413,
"epoch": 0.15
},
{
"current_steps": 1309,
"loss": 1.433,
"learning_rate": 0.00012591133004926107,
"epoch": 0.15
},
{
"current_steps": 1311,
"loss": 1.3245,
"learning_rate": 0.000125615763546798,
"epoch": 0.15
},
{
"current_steps": 1313,
"loss": 1.3325,
"learning_rate": 0.00012532019704433495,
"epoch": 0.15
},
{
"current_steps": 1315,
"loss": 1.4503,
"learning_rate": 0.00012502463054187192,
"epoch": 0.15
},
{
"current_steps": 1317,
"loss": 1.49,
"learning_rate": 0.00012472906403940886,
"epoch": 0.15
},
{
"current_steps": 1319,
"loss": 1.485,
"learning_rate": 0.0001244334975369458,
"epoch": 0.15
},
{
"current_steps": 1321,
"loss": 1.4404,
"learning_rate": 0.00012413793103448274,
"epoch": 0.15
},
{
"current_steps": 1323,
"loss": 1.3019,
"learning_rate": 0.00012384236453201968,
"epoch": 0.15
},
{
"current_steps": 1325,
"loss": 1.5872,
"learning_rate": 0.00012354679802955662,
"epoch": 0.15
},
{
"current_steps": 1327,
"loss": 1.5612,
"learning_rate": 0.0001232512315270936,
"epoch": 0.15
},
{
"current_steps": 1329,
"loss": 1.4557,
"learning_rate": 0.00012295566502463053,
"epoch": 0.15
},
{
"current_steps": 1331,
"loss": 1.5886,
"learning_rate": 0.00012266009852216747,
"epoch": 0.15
},
{
"current_steps": 1333,
"loss": 1.3797,
"learning_rate": 0.0001223645320197044,
"epoch": 0.16
},
{
"current_steps": 1335,
"loss": 1.3713,
"learning_rate": 0.00012206896551724136,
"epoch": 0.16
},
{
"current_steps": 1337,
"loss": 1.4158,
"learning_rate": 0.00012177339901477832,
"epoch": 0.16
},
{
"current_steps": 1339,
"loss": 1.4021,
"learning_rate": 0.00012147783251231526,
"epoch": 0.16
},
{
"current_steps": 1341,
"loss": 1.4421,
"learning_rate": 0.00012118226600985221,
"epoch": 0.16
},
{
"current_steps": 1343,
"loss": 1.2342,
"learning_rate": 0.00012088669950738915,
"epoch": 0.16
},
{
"current_steps": 1345,
"loss": 1.3491,
"learning_rate": 0.0001205911330049261,
"epoch": 0.16
},
{
"current_steps": 1347,
"loss": 1.3119,
"learning_rate": 0.00012029556650246304,
"epoch": 0.16
},
{
"current_steps": 1349,
"loss": 1.328,
"learning_rate": 0.00011999999999999999,
"epoch": 0.16
},
{
"current_steps": 1351,
"loss": 1.5423,
"learning_rate": 0.00011970443349753694,
"epoch": 0.16
},
{
"current_steps": 1353,
"loss": 1.4857,
"learning_rate": 0.00011940886699507388,
"epoch": 0.16
},
{
"current_steps": 1355,
"loss": 1.6532,
"learning_rate": 0.00011911330049261082,
"epoch": 0.16
},
{
"current_steps": 1357,
"loss": 1.3726,
"learning_rate": 0.00011881773399014777,
"epoch": 0.16
},
{
"current_steps": 1359,
"loss": 1.1621,
"learning_rate": 0.0001185221674876847,
"epoch": 0.16
},
{
"current_steps": 1361,
"loss": 1.4227,
"learning_rate": 0.00011822660098522167,
"epoch": 0.16
},
{
"current_steps": 1363,
"loss": 1.3642,
"learning_rate": 0.00011793103448275861,
"epoch": 0.16
},
{
"current_steps": 1365,
"loss": 1.5377,
"learning_rate": 0.00011763546798029556,
"epoch": 0.16
},
{
"current_steps": 1367,
"loss": 1.4044,
"learning_rate": 0.0001173399014778325,
"epoch": 0.16
},
{
"current_steps": 1369,
"loss": 1.4965,
"learning_rate": 0.00011704433497536944,
"epoch": 0.16
},
{
"current_steps": 1371,
"loss": 1.2706,
"learning_rate": 0.0001167487684729064,
"epoch": 0.16
},
{
"current_steps": 1373,
"loss": 1.5986,
"learning_rate": 0.00011645320197044334,
"epoch": 0.16
},
{
"current_steps": 1375,
"loss": 1.3399,
"learning_rate": 0.00011615763546798029,
"epoch": 0.16
},
{
"current_steps": 1377,
"loss": 1.5509,
"learning_rate": 0.00011586206896551723,
"epoch": 0.16
},
{
"current_steps": 1379,
"loss": 1.473,
"learning_rate": 0.00011556650246305417,
"epoch": 0.16
},
{
"current_steps": 1381,
"loss": 1.487,
"learning_rate": 0.00011527093596059112,
"epoch": 0.16
},
{
"current_steps": 1383,
"loss": 1.5032,
"learning_rate": 0.00011497536945812808,
"epoch": 0.16
},
{
"current_steps": 1385,
"loss": 1.4614,
"learning_rate": 0.00011467980295566502,
"epoch": 0.16
},
{
"current_steps": 1387,
"loss": 1.756,
"learning_rate": 0.00011438423645320196,
"epoch": 0.16
},
{
"current_steps": 1389,
"loss": 1.4472,
"learning_rate": 0.0001140886699507389,
"epoch": 0.16
},
{
"current_steps": 1391,
"loss": 1.3235,
"learning_rate": 0.00011379310344827585,
"epoch": 0.16
},
{
"current_steps": 1393,
"loss": 1.4611,
"learning_rate": 0.0001134975369458128,
"epoch": 0.16
},
{
"current_steps": 1395,
"loss": 1.5012,
"learning_rate": 0.00011320197044334975,
"epoch": 0.16
},
{
"current_steps": 1397,
"loss": 1.4022,
"learning_rate": 0.00011290640394088669,
"epoch": 0.16
},
{
"current_steps": 1399,
"loss": 1.6699,
"learning_rate": 0.00011261083743842364,
"epoch": 0.16
},
{
"current_steps": 1401,
"loss": 1.3621,
"learning_rate": 0.00011231527093596058,
"epoch": 0.16
},
{
"current_steps": 1403,
"loss": 1.414,
"learning_rate": 0.00011201970443349752,
"epoch": 0.16
},
{
"current_steps": 1405,
"loss": 1.2507,
"learning_rate": 0.00011172413793103448,
"epoch": 0.16
},
{
"current_steps": 1407,
"loss": 1.3533,
"learning_rate": 0.00011142857142857142,
"epoch": 0.16
},
{
"current_steps": 1409,
"loss": 1.3269,
"learning_rate": 0.00011113300492610837,
"epoch": 0.16
},
{
"current_steps": 1411,
"loss": 1.3367,
"learning_rate": 0.00011083743842364531,
"epoch": 0.16
},
{
"current_steps": 1413,
"loss": 1.354,
"learning_rate": 0.00011054187192118225,
"epoch": 0.16
},
{
"current_steps": 1415,
"loss": 1.4322,
"learning_rate": 0.00011024630541871919,
"epoch": 0.16
},
{
"current_steps": 1417,
"loss": 1.3051,
"learning_rate": 0.00010995073891625615,
"epoch": 0.16
},
{
"current_steps": 1419,
"loss": 1.4336,
"learning_rate": 0.0001096551724137931,
"epoch": 0.17
},
{
"current_steps": 1421,
"loss": 1.4927,
"learning_rate": 0.00010935960591133004,
"epoch": 0.17
},
{
"current_steps": 1423,
"loss": 1.2925,
"learning_rate": 0.00010906403940886698,
"epoch": 0.17
},
{
"current_steps": 1425,
"loss": 1.5501,
"learning_rate": 0.00010876847290640392,
"epoch": 0.17
},
{
"current_steps": 1427,
"loss": 1.4563,
"learning_rate": 0.00010847290640394089,
"epoch": 0.17
},
{
"current_steps": 1429,
"loss": 1.5805,
"learning_rate": 0.00010817733990147783,
"epoch": 0.17
},
{
"current_steps": 1431,
"loss": 1.603,
"learning_rate": 0.00010788177339901477,
"epoch": 0.17
},
{
"current_steps": 1433,
"loss": 1.5101,
"learning_rate": 0.00010758620689655171,
"epoch": 0.17
},
{
"current_steps": 1435,
"loss": 1.3574,
"learning_rate": 0.00010729064039408865,
"epoch": 0.17
},
{
"current_steps": 1437,
"loss": 1.527,
"learning_rate": 0.0001069950738916256,
"epoch": 0.17
},
{
"current_steps": 1439,
"loss": 1.6483,
"learning_rate": 0.00010669950738916256,
"epoch": 0.17
},
{
"current_steps": 1441,
"loss": 1.5076,
"learning_rate": 0.0001064039408866995,
"epoch": 0.17
},
{
"current_steps": 1443,
"loss": 1.1374,
"learning_rate": 0.00010610837438423644,
"epoch": 0.17
},
{
"current_steps": 1445,
"loss": 1.4851,
"learning_rate": 0.00010581280788177338,
"epoch": 0.17
},
{
"current_steps": 1447,
"loss": 1.4381,
"learning_rate": 0.00010551724137931032,
"epoch": 0.17
},
{
"current_steps": 1449,
"loss": 1.4604,
"learning_rate": 0.00010522167487684729,
"epoch": 0.17
},
{
"current_steps": 1451,
"loss": 1.3896,
"learning_rate": 0.00010492610837438423,
"epoch": 0.17
},
{
"current_steps": 1453,
"loss": 1.6003,
"learning_rate": 0.00010463054187192117,
"epoch": 0.17
},
{
"current_steps": 1455,
"loss": 1.4545,
"learning_rate": 0.00010433497536945811,
"epoch": 0.17
},
{
"current_steps": 1457,
"loss": 1.5914,
"learning_rate": 0.00010403940886699505,
"epoch": 0.17
},
{
"current_steps": 1459,
"loss": 1.4277,
"learning_rate": 0.00010374384236453201,
"epoch": 0.17
},
{
"current_steps": 1461,
"loss": 1.4904,
"learning_rate": 0.00010344827586206896,
"epoch": 0.17
},
{
"current_steps": 1463,
"loss": 1.2917,
"learning_rate": 0.0001031527093596059,
"epoch": 0.17
},
{
"current_steps": 1465,
"loss": 1.5209,
"learning_rate": 0.00010285714285714284,
"epoch": 0.17
},
{
"current_steps": 1467,
"loss": 1.6083,
"learning_rate": 0.0001025615763546798,
"epoch": 0.17
},
{
"current_steps": 1469,
"loss": 1.322,
"learning_rate": 0.00010226600985221674,
"epoch": 0.17
},
{
"current_steps": 1471,
"loss": 1.3146,
"learning_rate": 0.00010197044334975368,
"epoch": 0.17
},
{
"current_steps": 1473,
"loss": 1.4603,
"learning_rate": 0.00010167487684729063,
"epoch": 0.17
},
{
"current_steps": 1475,
"loss": 1.2714,
"learning_rate": 0.00010137931034482757,
"epoch": 0.17
},
{
"current_steps": 1477,
"loss": 1.3224,
"learning_rate": 0.00010108374384236453,
"epoch": 0.17
},
{
"current_steps": 1479,
"loss": 1.3709,
"learning_rate": 0.00010078817733990147,
"epoch": 0.17
},
{
"current_steps": 1481,
"loss": 1.6218,
"learning_rate": 0.00010049261083743841,
"epoch": 0.17
},
{
"current_steps": 1483,
"loss": 1.5694,
"learning_rate": 0.00010019704433497536,
"epoch": 0.17
},
{
"current_steps": 1485,
"loss": 1.4105,
"learning_rate": 9.99014778325123e-05,
"epoch": 0.17
},
{
"current_steps": 1487,
"loss": 1.5503,
"learning_rate": 9.960591133004926e-05,
"epoch": 0.17
},
{
"current_steps": 1489,
"loss": 1.4169,
"learning_rate": 9.93103448275862e-05,
"epoch": 0.17
},
{
"current_steps": 1491,
"loss": 1.5529,
"learning_rate": 9.901477832512314e-05,
"epoch": 0.17
},
{
"current_steps": 1493,
"loss": 1.339,
"learning_rate": 9.871921182266008e-05,
"epoch": 0.17
},
{
"current_steps": 1495,
"loss": 1.4177,
"learning_rate": 9.842364532019705e-05,
"epoch": 0.17
},
{
"current_steps": 1497,
"loss": 1.3418,
"learning_rate": 9.812807881773399e-05,
"epoch": 0.17
},
{
"current_steps": 1499,
"loss": 1.405,
"learning_rate": 9.783251231527093e-05,
"epoch": 0.17
},
{
"current_steps": 1501,
"loss": 1.3082,
"learning_rate": 9.753694581280787e-05,
"epoch": 0.17
},
{
"current_steps": 1503,
"loss": 1.5519,
"learning_rate": 9.724137931034481e-05,
"epoch": 0.18
},
{
"current_steps": 1505,
"loss": 1.4063,
"learning_rate": 9.694581280788178e-05,
"epoch": 0.18
},
{
"current_steps": 1507,
"loss": 1.4751,
"learning_rate": 9.694581280788178e-05,
"epoch": 0.18
},
{
"current_steps": 1509,
"loss": 1.6339,
"learning_rate": 9.665024630541872e-05,
"epoch": 0.18
},
{
"current_steps": 1511,
"loss": 1.4362,
"learning_rate": 9.635467980295566e-05,
"epoch": 0.18
},
{
"current_steps": 1513,
"loss": 1.5477,
"learning_rate": 9.60591133004926e-05,
"epoch": 0.18
},
{
"current_steps": 1515,
"loss": 1.3371,
"learning_rate": 9.576354679802954e-05,
"epoch": 0.18
},
{
"current_steps": 1517,
"loss": 1.2858,
"learning_rate": 9.546798029556648e-05,
"epoch": 0.18
},
{
"current_steps": 1519,
"loss": 1.2767,
"learning_rate": 9.517241379310345e-05,
"epoch": 0.18
},
{
"current_steps": 1521,
"loss": 1.374,
"learning_rate": 9.487684729064039e-05,
"epoch": 0.18
},
{
"current_steps": 1523,
"loss": 1.264,
"learning_rate": 9.458128078817733e-05,
"epoch": 0.18
},
{
"current_steps": 1525,
"loss": 1.3859,
"learning_rate": 9.428571428571427e-05,
"epoch": 0.18
},
{
"current_steps": 1527,
"loss": 1.538,
"learning_rate": 9.399014778325121e-05,
"epoch": 0.18
},
{
"current_steps": 1529,
"loss": 1.3722,
"learning_rate": 9.369458128078817e-05,
"epoch": 0.18
},
{
"current_steps": 1531,
"loss": 1.5568,
"learning_rate": 9.339901477832512e-05,
"epoch": 0.18
},
{
"current_steps": 1533,
"loss": 1.5843,
"learning_rate": 9.310344827586206e-05,
"epoch": 0.18
},
{
"current_steps": 1535,
"loss": 1.3413,
"learning_rate": 9.2807881773399e-05,
"epoch": 0.18
},
{
"current_steps": 1537,
"loss": 1.4297,
"learning_rate": 9.251231527093596e-05,
"epoch": 0.18
},
{
"current_steps": 1539,
"loss": 1.359,
"learning_rate": 9.22167487684729e-05,
"epoch": 0.18
},
{
"current_steps": 1541,
"loss": 1.3543,
"learning_rate": 9.192118226600985e-05,
"epoch": 0.18
},
{
"current_steps": 1543,
"loss": 1.5025,
"learning_rate": 9.162561576354679e-05,
"epoch": 0.18
},
{
"current_steps": 1545,
"loss": 1.4669,
"learning_rate": 9.133004926108373e-05,
"epoch": 0.18
},
{
"current_steps": 1547,
"loss": 1.3861,
"learning_rate": 9.103448275862069e-05,
"epoch": 0.18
},
{
"current_steps": 1549,
"loss": 1.2934,
"learning_rate": 9.073891625615763e-05,
"epoch": 0.18
},
{
"current_steps": 1551,
"loss": 1.3889,
"learning_rate": 9.044334975369457e-05,
"epoch": 0.18
},
{
"current_steps": 1553,
"loss": 1.4878,
"learning_rate": 9.014778325123152e-05,
"epoch": 0.18
},
{
"current_steps": 1555,
"loss": 1.6128,
"learning_rate": 8.985221674876846e-05,
"epoch": 0.18
},
{
"current_steps": 1557,
"loss": 1.3818,
"learning_rate": 8.955665024630542e-05,
"epoch": 0.18
},
{
"current_steps": 1559,
"loss": 1.5759,
"learning_rate": 8.926108374384236e-05,
"epoch": 0.18
},
{
"current_steps": 1561,
"loss": 1.3481,
"learning_rate": 8.89655172413793e-05,
"epoch": 0.18
},
{
"current_steps": 1563,
"loss": 1.402,
"learning_rate": 8.866995073891625e-05,
"epoch": 0.18
},
{
"current_steps": 1565,
"loss": 1.4652,
"learning_rate": 8.83743842364532e-05,
"epoch": 0.18
},
{
"current_steps": 1567,
"loss": 1.4362,
"learning_rate": 8.807881773399015e-05,
"epoch": 0.18
},
{
"current_steps": 1569,
"loss": 1.1903,
"learning_rate": 8.778325123152709e-05,
"epoch": 0.18
},
{
"current_steps": 1571,
"loss": 1.6198,
"learning_rate": 8.748768472906403e-05,
"epoch": 0.18
},
{
"current_steps": 1573,
"loss": 1.3341,
"learning_rate": 8.719211822660097e-05,
"epoch": 0.18
},
{
"current_steps": 1575,
"loss": 1.4117,
"learning_rate": 8.689655172413794e-05,
"epoch": 0.18
},
{
"current_steps": 1577,
"loss": 1.4784,
"learning_rate": 8.660098522167488e-05,
"epoch": 0.18
},
{
"current_steps": 1579,
"loss": 1.3749,
"learning_rate": 8.630541871921182e-05,
"epoch": 0.18
},
{
"current_steps": 1581,
"loss": 1.5399,
"learning_rate": 8.600985221674876e-05,
"epoch": 0.18
},
{
"current_steps": 1583,
"loss": 1.48,
"learning_rate": 8.57142857142857e-05,
"epoch": 0.18
},
{
"current_steps": 1585,
"loss": 1.5269,
"learning_rate": 8.541871921182264e-05,
"epoch": 0.18
},
{
"current_steps": 1587,
"loss": 1.6458,
"learning_rate": 8.512315270935961e-05,
"epoch": 0.18
},
{
"current_steps": 1589,
"loss": 1.4381,
"learning_rate": 8.482758620689655e-05,
"epoch": 0.19
},
{
"current_steps": 1591,
"loss": 1.5807,
"learning_rate": 8.453201970443349e-05,
"epoch": 0.19
},
{
"current_steps": 1593,
"loss": 1.1841,
"learning_rate": 8.423645320197043e-05,
"epoch": 0.19
},
{
"current_steps": 1595,
"loss": 1.4327,
"learning_rate": 8.394088669950737e-05,
"epoch": 0.19
},
{
"current_steps": 1597,
"loss": 1.3586,
"learning_rate": 8.364532019704434e-05,
"epoch": 0.19
},
{
"current_steps": 1599,
"loss": 1.4018,
"learning_rate": 8.334975369458128e-05,
"epoch": 0.19
},
{
"current_steps": 1601,
"loss": 1.3762,
"learning_rate": 8.305418719211822e-05,
"epoch": 0.19
},
{
"current_steps": 1603,
"loss": 1.4126,
"learning_rate": 8.275862068965516e-05,
"epoch": 0.19
},
{
"current_steps": 1605,
"loss": 1.2462,
"learning_rate": 8.24630541871921e-05,
"epoch": 0.19
},
{
"current_steps": 1607,
"loss": 1.4754,
"learning_rate": 8.216748768472905e-05,
"epoch": 0.19
},
{
"current_steps": 1609,
"loss": 1.5572,
"learning_rate": 8.187192118226601e-05,
"epoch": 0.19
},
{
"current_steps": 1611,
"loss": 1.5467,
"learning_rate": 8.157635467980295e-05,
"epoch": 0.19
},
{
"current_steps": 1613,
"loss": 1.5242,
"learning_rate": 8.128078817733989e-05,
"epoch": 0.19
},
{
"current_steps": 1615,
"loss": 1.5588,
"learning_rate": 8.098522167487684e-05,
"epoch": 0.19
},
{
"current_steps": 1617,
"loss": 1.6107,
"learning_rate": 8.068965517241378e-05,
"epoch": 0.19
},
{
"current_steps": 1619,
"loss": 1.3845,
"learning_rate": 8.039408866995073e-05,
"epoch": 0.19
},
{
"current_steps": 1621,
"loss": 1.391,
"learning_rate": 8.009852216748768e-05,
"epoch": 0.19
},
{
"current_steps": 1623,
"loss": 1.3375,
"learning_rate": 7.980295566502462e-05,
"epoch": 0.19
},
{
"current_steps": 1625,
"loss": 1.5187,
"learning_rate": 7.950738916256157e-05,
"epoch": 0.19
},
{
"current_steps": 1627,
"loss": 1.4744,
"learning_rate": 7.921182266009851e-05,
"epoch": 0.19
},
{
"current_steps": 1629,
"loss": 1.4449,
"learning_rate": 7.891625615763546e-05,
"epoch": 0.19
},
{
"current_steps": 1631,
"loss": 1.4237,
"learning_rate": 7.862068965517241e-05,
"epoch": 0.19
},
{
"current_steps": 1633,
"loss": 1.3786,
"learning_rate": 7.832512315270935e-05,
"epoch": 0.19
},
{
"current_steps": 1635,
"loss": 1.6633,
"learning_rate": 7.80295566502463e-05,
"epoch": 0.19
},
{
"current_steps": 1637,
"loss": 1.3197,
"learning_rate": 7.773399014778325e-05,
"epoch": 0.19
},
{
"current_steps": 1639,
"loss": 1.2572,
"learning_rate": 7.743842364532019e-05,
"epoch": 0.19
},
{
"current_steps": 1641,
"loss": 1.6133,
"learning_rate": 7.714285714285713e-05,
"epoch": 0.19
},
{
"current_steps": 1643,
"loss": 1.3526,
"learning_rate": 7.68472906403941e-05,
"epoch": 0.19
},
{
"current_steps": 1645,
"loss": 1.2797,
"learning_rate": 7.655172413793103e-05,
"epoch": 0.19
},
{
"current_steps": 1647,
"loss": 1.3872,
"learning_rate": 7.625615763546798e-05,
"epoch": 0.19
},
{
"current_steps": 1649,
"loss": 1.2822,
"learning_rate": 7.596059113300492e-05,
"epoch": 0.19
},
{
"current_steps": 1651,
"loss": 1.3557,
"learning_rate": 7.566502463054186e-05,
"epoch": 0.19
},
{
"current_steps": 1653,
"loss": 1.1847,
"learning_rate": 7.536945812807882e-05,
"epoch": 0.19
},
{
"current_steps": 1655,
"loss": 1.4428,
"learning_rate": 7.507389162561577e-05,
"epoch": 0.19
},
{
"current_steps": 1657,
"loss": 1.443,
"learning_rate": 7.47783251231527e-05,
"epoch": 0.19
},
{
"current_steps": 1659,
"loss": 1.4633,
"learning_rate": 7.448275862068965e-05,
"epoch": 0.19
},
{
"current_steps": 1661,
"loss": 1.5413,
"learning_rate": 7.418719211822659e-05,
"epoch": 0.19
},
{
"current_steps": 1663,
"loss": 1.3113,
"learning_rate": 7.389162561576354e-05,
"epoch": 0.19
},
{
"current_steps": 1665,
"loss": 1.4703,
"learning_rate": 7.359605911330048e-05,
"epoch": 0.19
},
{
"current_steps": 1667,
"loss": 1.3975,
"learning_rate": 7.330049261083744e-05,
"epoch": 0.19
},
{
"current_steps": 1669,
"loss": 1.3937,
"learning_rate": 7.300492610837438e-05,
"epoch": 0.19
},
{
"current_steps": 1671,
"loss": 1.4717,
"learning_rate": 7.270935960591132e-05,
"epoch": 0.19
},
{
"current_steps": 1673,
"loss": 1.3512,
"learning_rate": 7.241379310344827e-05,
"epoch": 0.19
},
{
"current_steps": 1675,
"loss": 1.4515,
"learning_rate": 7.211822660098521e-05,
"epoch": 0.2
},
{
"current_steps": 1677,
"loss": 1.6275,
"learning_rate": 7.182266009852215e-05,
"epoch": 0.2
},
{
"current_steps": 1679,
"loss": 1.3561,
"learning_rate": 7.152709359605911e-05,
"epoch": 0.2
},
{
"current_steps": 1681,
"loss": 1.2888,
"learning_rate": 7.123152709359605e-05,
"epoch": 0.2
},
{
"current_steps": 1683,
"loss": 1.351,
"learning_rate": 7.0935960591133e-05,
"epoch": 0.2
},
{
"current_steps": 1685,
"loss": 1.4117,
"learning_rate": 7.064039408866994e-05,
"epoch": 0.2
},
{
"current_steps": 1687,
"loss": 1.2991,
"learning_rate": 7.034482758620688e-05,
"epoch": 0.2
},
{
"current_steps": 1689,
"loss": 1.413,
"learning_rate": 7.004926108374384e-05,
"epoch": 0.2
},
{
"current_steps": 1691,
"loss": 1.3917,
"learning_rate": 6.975369458128078e-05,
"epoch": 0.2
},
{
"current_steps": 1693,
"loss": 1.3873,
"learning_rate": 6.945812807881773e-05,
"epoch": 0.2
},
{
"current_steps": 1695,
"loss": 1.4783,
"learning_rate": 6.916256157635467e-05,
"epoch": 0.2
},
{
"current_steps": 1697,
"loss": 1.5639,
"learning_rate": 6.886699507389163e-05,
"epoch": 0.2
},
{
"current_steps": 1699,
"loss": 1.5239,
"learning_rate": 6.857142857142857e-05,
"epoch": 0.2
},
{
"current_steps": 1701,
"loss": 1.5281,
"learning_rate": 6.827586206896551e-05,
"epoch": 0.2
},
{
"current_steps": 1703,
"loss": 1.2794,
"learning_rate": 6.798029556650246e-05,
"epoch": 0.2
},
{
"current_steps": 1705,
"loss": 1.2762,
"learning_rate": 6.76847290640394e-05,
"epoch": 0.2
},
{
"current_steps": 1707,
"loss": 1.2824,
"learning_rate": 6.738916256157636e-05,
"epoch": 0.2
},
{
"current_steps": 1709,
"loss": 1.2033,
"learning_rate": 6.70935960591133e-05,
"epoch": 0.2
},
{
"current_steps": 1711,
"loss": 1.4473,
"learning_rate": 6.679802955665024e-05,
"epoch": 0.2
},
{
"current_steps": 1713,
"loss": 1.6409,
"learning_rate": 6.650246305418719e-05,
"epoch": 0.2
},
{
"current_steps": 1715,
"loss": 1.6363,
"learning_rate": 6.620689655172413e-05,
"epoch": 0.2
},
{
"current_steps": 1717,
"loss": 1.3944,
"learning_rate": 6.591133004926107e-05,
"epoch": 0.2
},
{
"current_steps": 1719,
"loss": 1.5966,
"learning_rate": 6.561576354679803e-05,
"epoch": 0.2
},
{
"current_steps": 1721,
"loss": 1.4439,
"learning_rate": 6.532019704433497e-05,
"epoch": 0.2
},
{
"current_steps": 1723,
"loss": 1.4548,
"learning_rate": 6.502463054187192e-05,
"epoch": 0.2
},
{
"current_steps": 1725,
"loss": 1.2948,
"learning_rate": 6.472906403940886e-05,
"epoch": 0.2
},
{
"current_steps": 1727,
"loss": 1.5007,
"learning_rate": 6.44334975369458e-05,
"epoch": 0.2
},
{
"current_steps": 1729,
"loss": 1.3214,
"learning_rate": 6.413793103448276e-05,
"epoch": 0.2
},
{
"current_steps": 1731,
"loss": 1.4571,
"learning_rate": 6.38423645320197e-05,
"epoch": 0.2
},
{
"current_steps": 1733,
"loss": 1.5113,
"learning_rate": 6.354679802955664e-05,
"epoch": 0.2
},
{
"current_steps": 1735,
"loss": 1.4404,
"learning_rate": 6.32512315270936e-05,
"epoch": 0.2
},
{
"current_steps": 1737,
"loss": 1.3533,
"learning_rate": 6.295566502463053e-05,
"epoch": 0.2
},
{
"current_steps": 1739,
"loss": 1.4039,
"learning_rate": 6.266009852216747e-05,
"epoch": 0.2
},
{
"current_steps": 1741,
"loss": 1.4558,
"learning_rate": 6.236453201970443e-05,
"epoch": 0.2
},
{
"current_steps": 1743,
"loss": 1.4169,
"learning_rate": 6.206896551724137e-05,
"epoch": 0.2
},
{
"current_steps": 1745,
"loss": 1.5056,
"learning_rate": 6.177339901477831e-05,
"epoch": 0.2
},
{
"current_steps": 1747,
"loss": 1.4671,
"learning_rate": 6.147783251231526e-05,
"epoch": 0.2
},
{
"current_steps": 1749,
"loss": 1.3122,
"learning_rate": 6.11822660098522e-05,
"epoch": 0.2
},
{
"current_steps": 1751,
"loss": 1.3384,
"learning_rate": 6.088669950738916e-05,
"epoch": 0.2
},
{
"current_steps": 1753,
"loss": 1.3222,
"learning_rate": 6.059113300492611e-05,
"epoch": 0.2
},
{
"current_steps": 1755,
"loss": 1.4684,
"learning_rate": 6.029556650246305e-05,
"epoch": 0.2
},
{
"current_steps": 1757,
"loss": 1.2302,
"learning_rate": 5.9999999999999995e-05,
"epoch": 0.2
},
{
"current_steps": 1759,
"loss": 1.3662,
"learning_rate": 5.970443349753694e-05,
"epoch": 0.2
},
{
"current_steps": 1761,
"loss": 1.4275,
"learning_rate": 5.940886699507388e-05,
"epoch": 0.21
},
{
"current_steps": 1763,
"loss": 1.2949,
"learning_rate": 5.911330049261084e-05,
"epoch": 0.21
},
{
"current_steps": 1765,
"loss": 1.4499,
"learning_rate": 5.881773399014778e-05,
"epoch": 0.21
},
{
"current_steps": 1767,
"loss": 1.3698,
"learning_rate": 5.852216748768472e-05,
"epoch": 0.21
},
{
"current_steps": 1769,
"loss": 1.217,
"learning_rate": 5.822660098522167e-05,
"epoch": 0.21
},
{
"current_steps": 1771,
"loss": 1.5283,
"learning_rate": 5.793103448275861e-05,
"epoch": 0.21
},
{
"current_steps": 1773,
"loss": 1.4191,
"learning_rate": 5.763546798029556e-05,
"epoch": 0.21
},
{
"current_steps": 1775,
"loss": 1.4649,
"learning_rate": 5.733990147783251e-05,
"epoch": 0.21
},
{
"current_steps": 1777,
"loss": 1.3919,
"learning_rate": 5.704433497536945e-05,
"epoch": 0.21
},
{
"current_steps": 1779,
"loss": 1.5556,
"learning_rate": 5.67487684729064e-05,
"epoch": 0.21
},
{
"current_steps": 1781,
"loss": 1.5268,
"learning_rate": 5.645320197044334e-05,
"epoch": 0.21
},
{
"current_steps": 1783,
"loss": 1.3228,
"learning_rate": 5.615763546798029e-05,
"epoch": 0.21
},
{
"current_steps": 1785,
"loss": 1.5109,
"learning_rate": 5.586206896551724e-05,
"epoch": 0.21
},
{
"current_steps": 1787,
"loss": 1.4711,
"learning_rate": 5.5566502463054186e-05,
"epoch": 0.21
},
{
"current_steps": 1789,
"loss": 1.5531,
"learning_rate": 5.5270935960591126e-05,
"epoch": 0.21
},
{
"current_steps": 1791,
"loss": 1.4627,
"learning_rate": 5.4975369458128074e-05,
"epoch": 0.21
},
{
"current_steps": 1793,
"loss": 1.375,
"learning_rate": 5.467980295566502e-05,
"epoch": 0.21
},
{
"current_steps": 1795,
"loss": 1.3208,
"learning_rate": 5.438423645320196e-05,
"epoch": 0.21
},
{
"current_steps": 1797,
"loss": 1.5061,
"learning_rate": 5.4088669950738916e-05,
"epoch": 0.21
},
{
"current_steps": 1799,
"loss": 1.4171,
"learning_rate": 5.3793103448275856e-05,
"epoch": 0.21
},
{
"current_steps": 1801,
"loss": 1.2523,
"learning_rate": 5.34975369458128e-05,
"epoch": 0.21
},
{
"current_steps": 1803,
"loss": 1.1953,
"learning_rate": 5.320197044334975e-05,
"epoch": 0.21
},
{
"current_steps": 1805,
"loss": 1.5136,
"learning_rate": 5.290640394088669e-05,
"epoch": 0.21
},
{
"current_steps": 1807,
"loss": 1.2321,
"learning_rate": 5.2610837438423646e-05,
"epoch": 0.21
},
{
"current_steps": 1809,
"loss": 1.2477,
"learning_rate": 5.231527093596059e-05,
"epoch": 0.21
},
{
"current_steps": 1811,
"loss": 1.4657,
"learning_rate": 5.201970443349753e-05,
"epoch": 0.21
},
{
"current_steps": 1813,
"loss": 1.6217,
"learning_rate": 5.172413793103448e-05,
"epoch": 0.21
},
{
"current_steps": 1815,
"loss": 1.4777,
"learning_rate": 5.142857142857142e-05,
"epoch": 0.21
},
{
"current_steps": 1817,
"loss": 1.2844,
"learning_rate": 5.113300492610837e-05,
"epoch": 0.21
},
{
"current_steps": 1819,
"loss": 1.5179,
"learning_rate": 5.083743842364532e-05,
"epoch": 0.21
},
{
"current_steps": 1821,
"loss": 1.5915,
"learning_rate": 5.0541871921182264e-05,
"epoch": 0.21
},
{
"current_steps": 1823,
"loss": 1.2204,
"learning_rate": 5.0246305418719205e-05,
"epoch": 0.21
},
{
"current_steps": 1825,
"loss": 1.3171,
"learning_rate": 4.995073891625615e-05,
"epoch": 0.21
},
{
"current_steps": 1827,
"loss": 1.4406,
"learning_rate": 4.96551724137931e-05,
"epoch": 0.21
},
{
"current_steps": 1829,
"loss": 1.4661,
"learning_rate": 4.935960591133004e-05,
"epoch": 0.21
},
{
"current_steps": 1831,
"loss": 1.4711,
"learning_rate": 4.9064039408866995e-05,
"epoch": 0.21
},
{
"current_steps": 1833,
"loss": 1.3958,
"learning_rate": 4.8768472906403935e-05,
"epoch": 0.21
},
{
"current_steps": 1835,
"loss": 1.3915,
"learning_rate": 4.847290640394089e-05,
"epoch": 0.21
},
{
"current_steps": 1837,
"loss": 1.412,
"learning_rate": 4.817733990147783e-05,
"epoch": 0.21
},
{
"current_steps": 1839,
"loss": 1.3662,
"learning_rate": 4.788177339901477e-05,
"epoch": 0.21
},
{
"current_steps": 1841,
"loss": 1.3259,
"learning_rate": 4.788177339901477e-05,
"epoch": 0.21
},
{
"current_steps": 1843,
"loss": 1.499,
"learning_rate": 4.7586206896551725e-05,
"epoch": 0.21
},
{
"current_steps": 1845,
"loss": 1.4059,
"learning_rate": 4.7290640394088666e-05,
"epoch": 0.21
},
{
"current_steps": 1847,
"loss": 1.4343,
"learning_rate": 4.6995073891625606e-05,
"epoch": 0.22
},
{
"current_steps": 1849,
"loss": 1.3018,
"learning_rate": 4.669950738916256e-05,
"epoch": 0.22
},
{
"current_steps": 1851,
"loss": 1.3917,
"learning_rate": 4.64039408866995e-05,
"epoch": 0.22
},
{
"current_steps": 1853,
"loss": 1.2198,
"learning_rate": 4.610837438423645e-05,
"epoch": 0.22
},
{
"current_steps": 1855,
"loss": 1.3396,
"learning_rate": 4.5812807881773396e-05,
"epoch": 0.22
},
{
"current_steps": 1857,
"loss": 1.515,
"learning_rate": 4.551724137931034e-05,
"epoch": 0.22
},
{
"current_steps": 1859,
"loss": 1.2641,
"learning_rate": 4.5221674876847284e-05,
"epoch": 0.22
},
{
"current_steps": 1861,
"loss": 1.3899,
"learning_rate": 4.492610837438423e-05,
"epoch": 0.22
},
{
"current_steps": 1863,
"loss": 1.5666,
"learning_rate": 4.463054187192118e-05,
"epoch": 0.22
},
{
"current_steps": 1865,
"loss": 1.2917,
"learning_rate": 4.4334975369458126e-05,
"epoch": 0.22
},
{
"current_steps": 1867,
"loss": 1.4541,
"learning_rate": 4.4039408866995073e-05,
"epoch": 0.22
},
{
"current_steps": 1869,
"loss": 1.2695,
"learning_rate": 4.3743842364532014e-05,
"epoch": 0.22
},
{
"current_steps": 1871,
"loss": 1.4244,
"learning_rate": 4.344827586206897e-05,
"epoch": 0.22
},
{
"current_steps": 1873,
"loss": 1.325,
"learning_rate": 4.315270935960591e-05,
"epoch": 0.22
},
{
"current_steps": 1875,
"loss": 1.3828,
"learning_rate": 4.285714285714285e-05,
"epoch": 0.22
},
{
"current_steps": 1877,
"loss": 1.4658,
"learning_rate": 4.2561576354679804e-05,
"epoch": 0.22
},
{
"current_steps": 1879,
"loss": 1.4219,
"learning_rate": 4.2266009852216744e-05,
"epoch": 0.22
},
{
"current_steps": 1881,
"loss": 1.3448,
"learning_rate": 4.1970443349753685e-05,
"epoch": 0.22
},
{
"current_steps": 1883,
"loss": 1.3195,
"learning_rate": 4.167487684729064e-05,
"epoch": 0.22
},
{
"current_steps": 1885,
"loss": 1.3979,
"learning_rate": 4.137931034482758e-05,
"epoch": 0.22
},
{
"current_steps": 1887,
"loss": 1.4834,
"learning_rate": 4.108374384236453e-05,
"epoch": 0.22
},
{
"current_steps": 1889,
"loss": 1.4958,
"learning_rate": 4.0788177339901475e-05,
"epoch": 0.22
},
{
"current_steps": 1891,
"loss": 1.5403,
"learning_rate": 4.049261083743842e-05,
"epoch": 0.22
},
{
"current_steps": 1893,
"loss": 1.2163,
"learning_rate": 4.019704433497536e-05,
"epoch": 0.22
},
{
"current_steps": 1895,
"loss": 1.2908,
"learning_rate": 3.990147783251231e-05,
"epoch": 0.22
},
{
"current_steps": 1897,
"loss": 1.4287,
"learning_rate": 3.960591133004926e-05,
"epoch": 0.22
},
{
"current_steps": 1899,
"loss": 1.409,
"learning_rate": 3.9310344827586205e-05,
"epoch": 0.22
},
{
"current_steps": 1901,
"loss": 1.3897,
"learning_rate": 3.901477832512315e-05,
"epoch": 0.22
},
{
"current_steps": 1903,
"loss": 1.3014,
"learning_rate": 3.871921182266009e-05,
"epoch": 0.22
},
{
"current_steps": 1905,
"loss": 1.3687,
"learning_rate": 3.842364532019705e-05,
"epoch": 0.22
},
{
"current_steps": 1907,
"loss": 1.4001,
"learning_rate": 3.812807881773399e-05,
"epoch": 0.22
},
{
"current_steps": 1909,
"loss": 1.3124,
"learning_rate": 3.783251231527093e-05,
"epoch": 0.22
},
{
"current_steps": 1911,
"loss": 1.3053,
"learning_rate": 3.753694581280788e-05,
"epoch": 0.22
},
{
"current_steps": 1913,
"loss": 1.435,
"learning_rate": 3.724137931034482e-05,
"epoch": 0.22
},
{
"current_steps": 1915,
"loss": 1.2448,
"learning_rate": 3.694581280788177e-05,
"epoch": 0.22
},
{
"current_steps": 1917,
"loss": 1.3858,
"learning_rate": 3.665024630541872e-05,
"epoch": 0.22
},
{
"current_steps": 1919,
"loss": 1.3794,
"learning_rate": 3.635467980295566e-05,
"epoch": 0.22
},
{
"current_steps": 1921,
"loss": 1.3404,
"learning_rate": 3.6059113300492606e-05,
"epoch": 0.22
},
{
"current_steps": 1923,
"loss": 1.5008,
"learning_rate": 3.5763546798029553e-05,
"epoch": 0.22
},
{
"current_steps": 1925,
"loss": 1.5155,
"learning_rate": 3.54679802955665e-05,
"epoch": 0.22
},
{
"current_steps": 1927,
"loss": 1.2666,
"learning_rate": 3.517241379310344e-05,
"epoch": 0.22
},
{
"current_steps": 1929,
"loss": 1.457,
"learning_rate": 3.487684729064039e-05,
"epoch": 0.22
},
{
"current_steps": 1931,
"loss": 1.5565,
"learning_rate": 3.4581280788177336e-05,
"epoch": 0.22
},
{
"current_steps": 1933,
"loss": 1.3359,
"learning_rate": 3.4285714285714284e-05,
"epoch": 0.23
},
{
"current_steps": 1935,
"loss": 1.4157,
"learning_rate": 3.399014778325123e-05,
"epoch": 0.23
},
{
"current_steps": 1937,
"loss": 1.322,
"learning_rate": 3.369458128078818e-05,
"epoch": 0.23
},
{
"current_steps": 1939,
"loss": 1.6225,
"learning_rate": 3.339901477832512e-05,
"epoch": 0.23
},
{
"current_steps": 1941,
"loss": 1.4268,
"learning_rate": 3.3103448275862067e-05,
"epoch": 0.23
},
{
"current_steps": 1943,
"loss": 1.3425,
"learning_rate": 3.2807881773399014e-05,
"epoch": 0.23
},
{
"current_steps": 1945,
"loss": 1.3537,
"learning_rate": 3.251231527093596e-05,
"epoch": 0.23
},
{
"current_steps": 1947,
"loss": 1.3513,
"learning_rate": 3.22167487684729e-05,
"epoch": 0.23
},
{
"current_steps": 1949,
"loss": 1.4673,
"learning_rate": 3.192118226600985e-05,
"epoch": 0.23
},
{
"current_steps": 1951,
"loss": 1.4394,
"learning_rate": 3.16256157635468e-05,
"epoch": 0.23
},
{
"current_steps": 1953,
"loss": 1.4502,
"learning_rate": 3.133004926108374e-05,
"epoch": 0.23
},
{
"current_steps": 1955,
"loss": 1.415,
"learning_rate": 3.1034482758620685e-05,
"epoch": 0.23
},
{
"current_steps": 1957,
"loss": 1.2713,
"learning_rate": 3.073891625615763e-05,
"epoch": 0.23
},
{
"current_steps": 1959,
"loss": 1.2928,
"learning_rate": 3.044334975369458e-05,
"epoch": 0.23
},
{
"current_steps": 1961,
"loss": 1.3281,
"learning_rate": 3.0147783251231524e-05,
"epoch": 0.23
},
{
"current_steps": 1963,
"loss": 1.3303,
"learning_rate": 2.985221674876847e-05,
"epoch": 0.23
},
{
"current_steps": 1965,
"loss": 1.3034,
"learning_rate": 2.955665024630542e-05,
"epoch": 0.23
},
{
"current_steps": 1967,
"loss": 1.5266,
"learning_rate": 2.926108374384236e-05,
"epoch": 0.23
},
{
"current_steps": 1969,
"loss": 1.6226,
"learning_rate": 2.8965517241379307e-05,
"epoch": 0.23
},
{
"current_steps": 1971,
"loss": 1.4254,
"learning_rate": 2.8669950738916254e-05,
"epoch": 0.23
},
{
"current_steps": 1973,
"loss": 1.4346,
"learning_rate": 2.83743842364532e-05,
"epoch": 0.23
},
{
"current_steps": 1975,
"loss": 1.4151,
"learning_rate": 2.8078817733990145e-05,
"epoch": 0.23
},
{
"current_steps": 1977,
"loss": 1.5342,
"learning_rate": 2.7783251231527093e-05,
"epoch": 0.23
},
{
"current_steps": 1979,
"loss": 1.4544,
"learning_rate": 2.7487684729064037e-05,
"epoch": 0.23
},
{
"current_steps": 1981,
"loss": 1.2191,
"learning_rate": 2.719211822660098e-05,
"epoch": 0.23
},
{
"current_steps": 1983,
"loss": 1.4886,
"learning_rate": 2.6896551724137928e-05,
"epoch": 0.23
},
{
"current_steps": 1985,
"loss": 1.4416,
"learning_rate": 2.6600985221674876e-05,
"epoch": 0.23
},
{
"current_steps": 1987,
"loss": 1.3084,
"learning_rate": 2.6305418719211823e-05,
"epoch": 0.23
},
{
"current_steps": 1989,
"loss": 1.4037,
"learning_rate": 2.6009852216748764e-05,
"epoch": 0.23
},
{
"current_steps": 1991,
"loss": 1.4999,
"learning_rate": 2.571428571428571e-05,
"epoch": 0.23
},
{
"current_steps": 1993,
"loss": 1.3526,
"learning_rate": 2.541871921182266e-05,
"epoch": 0.23
},
{
"current_steps": 1995,
"loss": 1.2452,
"learning_rate": 2.5123152709359602e-05,
"epoch": 0.23
},
{
"current_steps": 1997,
"loss": 1.4491,
"learning_rate": 2.482758620689655e-05,
"epoch": 0.23
},
{
"current_steps": 1999,
"loss": 1.5146,
"learning_rate": 2.4532019704433497e-05,
"epoch": 0.23
},
{
"current_steps": 2001,
"loss": 1.3113,
"learning_rate": 2.4236453201970445e-05,
"epoch": 0.23
},
{
"current_steps": 2003,
"loss": 1.4204,
"learning_rate": 2.3940886699507385e-05,
"epoch": 0.23
},
{
"current_steps": 2005,
"loss": 1.4282,
"learning_rate": 2.3645320197044333e-05,
"epoch": 0.23
},
{
"current_steps": 2007,
"loss": 1.2543,
"learning_rate": 2.334975369458128e-05,
"epoch": 0.23
},
{
"current_steps": 2009,
"loss": 1.4168,
"learning_rate": 2.3054187192118224e-05,
"epoch": 0.23
},
{
"current_steps": 2011,
"loss": 1.3244,
"learning_rate": 2.275862068965517e-05,
"epoch": 0.23
},
{
"current_steps": 2013,
"loss": 1.4299,
"learning_rate": 2.2463054187192116e-05,
"epoch": 0.23
},
{
"current_steps": 2015,
"loss": 1.6615,
"learning_rate": 2.2167487684729063e-05,
"epoch": 0.23
},
{
"current_steps": 2017,
"loss": 1.494,
"learning_rate": 2.1871921182266007e-05,
"epoch": 0.23
},
{
"current_steps": 2019,
"loss": 1.4014,
"learning_rate": 2.1576354679802954e-05,
"epoch": 0.24
},
{
"current_steps": 2021,
"loss": 1.6533,
"learning_rate": 2.1280788177339902e-05,
"epoch": 0.24
},
{
"current_steps": 2023,
"loss": 1.3424,
"learning_rate": 2.0985221674876842e-05,
"epoch": 0.24
},
{
"current_steps": 2025,
"loss": 1.2958,
"learning_rate": 2.068965517241379e-05,
"epoch": 0.24
},
{
"current_steps": 2027,
"loss": 1.4598,
"learning_rate": 2.0394088669950737e-05,
"epoch": 0.24
},
{
"current_steps": 2029,
"loss": 1.2353,
"learning_rate": 2.009852216748768e-05,
"epoch": 0.24
},
{
"current_steps": 2031,
"loss": 1.3546,
"learning_rate": 1.980295566502463e-05,
"epoch": 0.24
},
{
"current_steps": 2033,
"loss": 1.3564,
"learning_rate": 1.9507389162561576e-05,
"epoch": 0.24
},
{
"current_steps": 2035,
"loss": 1.374,
"learning_rate": 1.9211822660098524e-05,
"epoch": 0.24
},
{
"current_steps": 2037,
"loss": 1.3581,
"learning_rate": 1.8916256157635464e-05,
"epoch": 0.24
},
{
"current_steps": 2039,
"loss": 1.5073,
"learning_rate": 1.862068965517241e-05,
"epoch": 0.24
},
{
"current_steps": 2041,
"loss": 1.4059,
"learning_rate": 1.832512315270936e-05,
"epoch": 0.24
},
{
"current_steps": 2043,
"loss": 1.3828,
"learning_rate": 1.8029556650246303e-05,
"epoch": 0.24
},
{
"current_steps": 2045,
"loss": 1.4461,
"learning_rate": 1.773399014778325e-05,
"epoch": 0.24
},
{
"current_steps": 2047,
"loss": 1.3237,
"learning_rate": 1.7438423645320194e-05,
"epoch": 0.24
},
{
"current_steps": 2049,
"loss": 1.3452,
"learning_rate": 1.7142857142857142e-05,
"epoch": 0.24
},
{
"current_steps": 2051,
"loss": 1.3552,
"learning_rate": 1.684729064039409e-05,
"epoch": 0.24
},
{
"current_steps": 2053,
"loss": 1.3517,
"learning_rate": 1.6551724137931033e-05,
"epoch": 0.24
},
{
"current_steps": 2055,
"loss": 1.4637,
"learning_rate": 1.625615763546798e-05,
"epoch": 0.24
},
{
"current_steps": 2057,
"loss": 1.4506,
"learning_rate": 1.5960591133004925e-05,
"epoch": 0.24
},
{
"current_steps": 2059,
"loss": 1.2913,
"learning_rate": 1.566502463054187e-05,
"epoch": 0.24
},
{
"current_steps": 2061,
"loss": 1.4415,
"learning_rate": 1.5369458128078816e-05,
"epoch": 0.24
},
{
"current_steps": 2063,
"loss": 1.3982,
"learning_rate": 1.5073891625615762e-05,
"epoch": 0.24
},
{
"current_steps": 2065,
"loss": 1.476,
"learning_rate": 1.477832512315271e-05,
"epoch": 0.24
},
{
"current_steps": 2067,
"loss": 1.4763,
"learning_rate": 1.4482758620689653e-05,
"epoch": 0.24
},
{
"current_steps": 2069,
"loss": 1.3806,
"learning_rate": 1.41871921182266e-05,
"epoch": 0.24
},
{
"current_steps": 2071,
"loss": 1.6809,
"learning_rate": 1.3891625615763546e-05,
"epoch": 0.24
},
{
"current_steps": 2073,
"loss": 1.3225,
"learning_rate": 1.359605911330049e-05,
"epoch": 0.24
},
{
"current_steps": 2075,
"loss": 1.5042,
"learning_rate": 1.3300492610837438e-05,
"epoch": 0.24
},
{
"current_steps": 2077,
"loss": 1.3123,
"learning_rate": 1.3004926108374382e-05,
"epoch": 0.24
},
{
"current_steps": 2079,
"loss": 1.2716,
"learning_rate": 1.270935960591133e-05,
"epoch": 0.24
},
{
"current_steps": 2081,
"loss": 1.4565,
"learning_rate": 1.2413793103448275e-05,
"epoch": 0.24
},
{
"current_steps": 2083,
"loss": 1.268,
"learning_rate": 1.2118226600985222e-05,
"epoch": 0.24
},
{
"current_steps": 2085,
"loss": 1.3143,
"learning_rate": 1.1822660098522166e-05,
"epoch": 0.24
},
{
"current_steps": 2087,
"loss": 1.4247,
"learning_rate": 1.1527093596059112e-05,
"epoch": 0.24
},
{
"current_steps": 2089,
"loss": 1.3091,
"learning_rate": 1.1231527093596058e-05,
"epoch": 0.24
},
{
"current_steps": 2091,
"loss": 1.4655,
"learning_rate": 1.0935960591133004e-05,
"epoch": 0.24
},
{
"current_steps": 2093,
"loss": 1.4689,
"learning_rate": 1.0640394088669951e-05,
"epoch": 0.24
},
{
"current_steps": 2095,
"loss": 1.3892,
"learning_rate": 1.0344827586206895e-05,
"epoch": 0.24
},
{
"current_steps": 2097,
"loss": 1.2808,
"learning_rate": 1.004926108374384e-05,
"epoch": 0.24
},
{
"current_steps": 2099,
"loss": 1.5469,
"learning_rate": 9.753694581280788e-06,
"epoch": 0.24
},
{
"current_steps": 2101,
"loss": 1.3735,
"learning_rate": 9.458128078817732e-06,
"epoch": 0.24
},
{
"current_steps": 2103,
"loss": 1.4516,
"learning_rate": 9.16256157635468e-06,
"epoch": 0.24
},
{
"current_steps": 2105,
"loss": 1.2961,
"learning_rate": 8.866995073891625e-06,
"epoch": 0.25
},
{
"current_steps": 2107,
"loss": 1.4416,
"learning_rate": 8.571428571428571e-06,
"epoch": 0.25
},
{
"current_steps": 2109,
"loss": 1.309,
"learning_rate": 8.275862068965517e-06,
"epoch": 0.25
},
{
"current_steps": 2111,
"loss": 1.4846,
"learning_rate": 7.980295566502462e-06,
"epoch": 0.25
},
{
"current_steps": 2113,
"loss": 1.4353,
"learning_rate": 7.684729064039408e-06,
"epoch": 0.25
},
{
"current_steps": 2115,
"loss": 1.3507,
"learning_rate": 7.389162561576355e-06,
"epoch": 0.25
},
{
"current_steps": 2117,
"loss": 1.1822,
"learning_rate": 7.0935960591133e-06,
"epoch": 0.25
},
{
"current_steps": 2119,
"loss": 1.3353,
"learning_rate": 6.798029556650245e-06,
"epoch": 0.25
},
{
"current_steps": 2121,
"loss": 1.3504,
"learning_rate": 6.502463054187191e-06,
"epoch": 0.25
},
{
"current_steps": 2123,
"loss": 1.3049,
"learning_rate": 6.2068965517241375e-06,
"epoch": 0.25
},
{
"current_steps": 2125,
"loss": 1.4379,
"learning_rate": 5.911330049261083e-06,
"epoch": 0.25
},
{
"current_steps": 2127,
"loss": 1.4737,
"learning_rate": 5.615763546798029e-06,
"epoch": 0.25
},
{
"current_steps": 2129,
"loss": 1.4593,
"learning_rate": 5.3201970443349755e-06,
"epoch": 0.25
},
{
"current_steps": 2131,
"loss": 1.5299,
"learning_rate": 5.02463054187192e-06,
"epoch": 0.25
},
{
"current_steps": 2133,
"loss": 1.2848,
"learning_rate": 4.729064039408866e-06,
"epoch": 0.25
},
{
"current_steps": 2135,
"loss": 1.1764,
"learning_rate": 4.433497536945813e-06,
"epoch": 0.25
},
{
"current_steps": 2137,
"loss": 1.3565,
"learning_rate": 4.137931034482758e-06,
"epoch": 0.25
},
{
"current_steps": 2139,
"loss": 1.47,
"learning_rate": 3.842364532019704e-06,
"epoch": 0.25
},
{
"current_steps": 2141,
"loss": 1.4827,
"learning_rate": 3.54679802955665e-06,
"epoch": 0.25
},
{
"current_steps": 2143,
"loss": 1.4833,
"learning_rate": 3.2512315270935955e-06,
"epoch": 0.25
},
{
"current_steps": 2145,
"loss": 1.3252,
"learning_rate": 2.9556650246305416e-06,
"epoch": 0.25
},
{
"current_steps": 2147,
"loss": 1.4242,
"learning_rate": 2.6600985221674877e-06,
"epoch": 0.25
},
{
"current_steps": 2149,
"loss": 1.4473,
"learning_rate": 2.364532019704433e-06,
"epoch": 0.25
},
{
"current_steps": 2149,
"loss": 1.4473,
"learning_rate": 2.364532019704433e-06,
"epoch": 0.25
}
]