|
[
|
|
{
|
|
"current_steps": 0,
|
|
"loss": 0.6046,
|
|
"learning_rate": 1e-08,
|
|
"epoch": 0.008928571428571428
|
|
},
|
|
{
|
|
"current_steps": 1,
|
|
"loss": 0.6431,
|
|
"learning_rate": 2e-08,
|
|
"epoch": 0.017857142857142856
|
|
},
|
|
{
|
|
"current_steps": 2,
|
|
"loss": 0.6447,
|
|
"learning_rate": 3e-08,
|
|
"epoch": 0.026785714285714284
|
|
},
|
|
{
|
|
"current_steps": 3,
|
|
"loss": 0.7972,
|
|
"learning_rate": 4e-08,
|
|
"epoch": 0.03571428571428571
|
|
},
|
|
{
|
|
"current_steps": 4,
|
|
"loss": 0.6911,
|
|
"learning_rate": 5e-08,
|
|
"epoch": 0.044642857142857144
|
|
},
|
|
{
|
|
"current_steps": 5,
|
|
"loss": 0.8546,
|
|
"learning_rate": 6e-08,
|
|
"epoch": 0.05357142857142857
|
|
},
|
|
{
|
|
"current_steps": 6,
|
|
"loss": 0.7624,
|
|
"learning_rate": 6e-08,
|
|
"epoch": 0.0625
|
|
},
|
|
{
|
|
"current_steps": 7,
|
|
"loss": 0.6565,
|
|
"learning_rate": 7e-08,
|
|
"epoch": 0.07142857142857142
|
|
},
|
|
{
|
|
"current_steps": 8,
|
|
"loss": 0.6789,
|
|
"learning_rate": 8e-08,
|
|
"epoch": 0.08035714285714286
|
|
},
|
|
{
|
|
"current_steps": 9,
|
|
"loss": 0.8562,
|
|
"learning_rate": 8e-08,
|
|
"epoch": 0.08928571428571429
|
|
},
|
|
{
|
|
"current_steps": 10,
|
|
"loss": 0.8084,
|
|
"learning_rate": 9e-08,
|
|
"epoch": 0.09821428571428571
|
|
},
|
|
{
|
|
"current_steps": 11,
|
|
"loss": 0.7024,
|
|
"learning_rate": 1e-07,
|
|
"epoch": 0.10714285714285714
|
|
},
|
|
{
|
|
"current_steps": 12,
|
|
"loss": 0.7454,
|
|
"learning_rate": 1.0999999999999999e-07,
|
|
"epoch": 0.11607142857142858
|
|
},
|
|
{
|
|
"current_steps": 13,
|
|
"loss": 0.5896,
|
|
"learning_rate": 1.2e-07,
|
|
"epoch": 0.125
|
|
},
|
|
{
|
|
"current_steps": 14,
|
|
"loss": 1.1716,
|
|
"learning_rate": 1.3e-07,
|
|
"epoch": 0.13392857142857142
|
|
},
|
|
{
|
|
"current_steps": 15,
|
|
"loss": 0.8561,
|
|
"learning_rate": 1.4e-07,
|
|
"epoch": 0.14285714285714285
|
|
},
|
|
{
|
|
"current_steps": 16,
|
|
"loss": 0.9048,
|
|
"learning_rate": 1.5e-07,
|
|
"epoch": 0.15178571428571427
|
|
},
|
|
{
|
|
"current_steps": 17,
|
|
"loss": 0.6079,
|
|
"learning_rate": 1.6e-07,
|
|
"epoch": 0.16071428571428573
|
|
},
|
|
{
|
|
"current_steps": 18,
|
|
"loss": 0.9004,
|
|
"learning_rate": 1.7000000000000001e-07,
|
|
"epoch": 0.16964285714285715
|
|
},
|
|
{
|
|
"current_steps": 19,
|
|
"loss": 0.5512,
|
|
"learning_rate": 1.8e-07,
|
|
"epoch": 0.17857142857142858
|
|
},
|
|
{
|
|
"current_steps": 20,
|
|
"loss": 0.7782,
|
|
"learning_rate": 1.8999999999999998e-07,
|
|
"epoch": 0.1875
|
|
},
|
|
{
|
|
"current_steps": 21,
|
|
"loss": 0.7905,
|
|
"learning_rate": 2e-07,
|
|
"epoch": 0.19642857142857142
|
|
},
|
|
{
|
|
"current_steps": 22,
|
|
"loss": 0.6049,
|
|
"learning_rate": 2.0999999999999997e-07,
|
|
"epoch": 0.20535714285714285
|
|
},
|
|
{
|
|
"current_steps": 23,
|
|
"loss": 0.685,
|
|
"learning_rate": 2.1999999999999998e-07,
|
|
"epoch": 0.21428571428571427
|
|
},
|
|
{
|
|
"current_steps": 24,
|
|
"loss": 0.8171,
|
|
"learning_rate": 2.3e-07,
|
|
"epoch": 0.22321428571428573
|
|
},
|
|
{
|
|
"current_steps": 25,
|
|
"loss": 0.8018,
|
|
"learning_rate": 2.4e-07,
|
|
"epoch": 0.23214285714285715
|
|
},
|
|
{
|
|
"current_steps": 26,
|
|
"loss": 0.4959,
|
|
"learning_rate": 2.5e-07,
|
|
"epoch": 0.24107142857142858
|
|
},
|
|
{
|
|
"current_steps": 27,
|
|
"loss": 0.6348,
|
|
"learning_rate": 2.6e-07,
|
|
"epoch": 0.25
|
|
},
|
|
{
|
|
"current_steps": 28,
|
|
"loss": 0.8005,
|
|
"learning_rate": 2.7e-07,
|
|
"epoch": 0.25892857142857145
|
|
},
|
|
{
|
|
"current_steps": 29,
|
|
"loss": 0.6777,
|
|
"learning_rate": 2.8e-07,
|
|
"epoch": 0.26785714285714285
|
|
},
|
|
{
|
|
"current_steps": 30,
|
|
"loss": 0.9042,
|
|
"learning_rate": 2.9e-07,
|
|
"epoch": 0.2767857142857143
|
|
},
|
|
{
|
|
"current_steps": 31,
|
|
"loss": 0.6491,
|
|
"learning_rate": 3e-07,
|
|
"epoch": 0.2857142857142857
|
|
},
|
|
{
|
|
"current_steps": 32,
|
|
"loss": 1.0966,
|
|
"learning_rate": 3.1e-07,
|
|
"epoch": 0.29464285714285715
|
|
},
|
|
{
|
|
"current_steps": 33,
|
|
"loss": 0.7451,
|
|
"learning_rate": 3.2e-07,
|
|
"epoch": 0.30357142857142855
|
|
},
|
|
{
|
|
"current_steps": 34,
|
|
"loss": 1.1446,
|
|
"learning_rate": 3.2e-07,
|
|
"epoch": 0.3125
|
|
},
|
|
{
|
|
"current_steps": 35,
|
|
"loss": 0.7644,
|
|
"learning_rate": 3.3e-07,
|
|
"epoch": 0.32142857142857145
|
|
},
|
|
{
|
|
"current_steps": 36,
|
|
"loss": 0.7742,
|
|
"learning_rate": 3.4000000000000003e-07,
|
|
"epoch": 0.33035714285714285
|
|
},
|
|
{
|
|
"current_steps": 37,
|
|
"loss": 0.8247,
|
|
"learning_rate": 3.5e-07,
|
|
"epoch": 0.3392857142857143
|
|
},
|
|
{
|
|
"current_steps": 38,
|
|
"loss": 0.8667,
|
|
"learning_rate": 3.6e-07,
|
|
"epoch": 0.3482142857142857
|
|
},
|
|
{
|
|
"current_steps": 39,
|
|
"loss": 0.8309,
|
|
"learning_rate": 3.7e-07,
|
|
"epoch": 0.35714285714285715
|
|
},
|
|
{
|
|
"current_steps": 40,
|
|
"loss": 0.5913,
|
|
"learning_rate": 3.7999999999999996e-07,
|
|
"epoch": 0.36607142857142855
|
|
},
|
|
{
|
|
"current_steps": 41,
|
|
"loss": 0.5562,
|
|
"learning_rate": 3.8999999999999997e-07,
|
|
"epoch": 0.375
|
|
},
|
|
{
|
|
"current_steps": 42,
|
|
"loss": 1.6276,
|
|
"learning_rate": 4e-07,
|
|
"epoch": 0.38392857142857145
|
|
},
|
|
{
|
|
"current_steps": 43,
|
|
"loss": 0.682,
|
|
"learning_rate": 4.0999999999999994e-07,
|
|
"epoch": 0.39285714285714285
|
|
},
|
|
{
|
|
"current_steps": 44,
|
|
"loss": 0.8022,
|
|
"learning_rate": 4.1999999999999995e-07,
|
|
"epoch": 0.4017857142857143
|
|
},
|
|
{
|
|
"current_steps": 45,
|
|
"loss": 0.6702,
|
|
"learning_rate": 4.2999999999999996e-07,
|
|
"epoch": 0.4107142857142857
|
|
},
|
|
{
|
|
"current_steps": 46,
|
|
"loss": 0.6993,
|
|
"learning_rate": 4.3999999999999997e-07,
|
|
"epoch": 0.41964285714285715
|
|
},
|
|
{
|
|
"current_steps": 47,
|
|
"loss": 0.9685,
|
|
"learning_rate": 4.5e-07,
|
|
"epoch": 0.42857142857142855
|
|
},
|
|
{
|
|
"current_steps": 48,
|
|
"loss": 0.6637,
|
|
"learning_rate": 4.6e-07,
|
|
"epoch": 0.4375
|
|
},
|
|
{
|
|
"current_steps": 49,
|
|
"loss": 0.908,
|
|
"learning_rate": 4.6999999999999995e-07,
|
|
"epoch": 0.44642857142857145
|
|
},
|
|
{
|
|
"current_steps": 50,
|
|
"loss": 0.8683,
|
|
"learning_rate": 4.8e-07,
|
|
"epoch": 0.45535714285714285
|
|
},
|
|
{
|
|
"current_steps": 51,
|
|
"loss": 0.9243,
|
|
"learning_rate": 4.9e-07,
|
|
"epoch": 0.4642857142857143
|
|
},
|
|
{
|
|
"current_steps": 52,
|
|
"loss": 0.7933,
|
|
"learning_rate": 5e-07,
|
|
"epoch": 0.4732142857142857
|
|
},
|
|
{
|
|
"current_steps": 53,
|
|
"loss": 0.5856,
|
|
"learning_rate": 5.1e-07,
|
|
"epoch": 0.48214285714285715
|
|
},
|
|
{
|
|
"current_steps": 54,
|
|
"loss": 0.7097,
|
|
"learning_rate": 5.2e-07,
|
|
"epoch": 0.49107142857142855
|
|
},
|
|
{
|
|
"current_steps": 55,
|
|
"loss": 0.6476,
|
|
"learning_rate": 5.3e-07,
|
|
"epoch": 0.5
|
|
},
|
|
{
|
|
"current_steps": 56,
|
|
"loss": 0.8212,
|
|
"learning_rate": 5.4e-07,
|
|
"epoch": 0.5089285714285714
|
|
},
|
|
{
|
|
"current_steps": 57,
|
|
"loss": 0.7932,
|
|
"learning_rate": 5.5e-07,
|
|
"epoch": 0.5178571428571429
|
|
},
|
|
{
|
|
"current_steps": 58,
|
|
"loss": 0.8155,
|
|
"learning_rate": 5.6e-07,
|
|
"epoch": 0.5267857142857143
|
|
},
|
|
{
|
|
"current_steps": 59,
|
|
"loss": 0.5644,
|
|
"learning_rate": 5.699999999999999e-07,
|
|
"epoch": 0.5357142857142857
|
|
},
|
|
{
|
|
"current_steps": 60,
|
|
"loss": 0.8935,
|
|
"learning_rate": 5.8e-07,
|
|
"epoch": 0.5446428571428571
|
|
},
|
|
{
|
|
"current_steps": 61,
|
|
"loss": 0.6935,
|
|
"learning_rate": 5.9e-07,
|
|
"epoch": 0.5535714285714286
|
|
},
|
|
{
|
|
"current_steps": 62,
|
|
"loss": 0.6186,
|
|
"learning_rate": 6e-07,
|
|
"epoch": 0.5625
|
|
},
|
|
{
|
|
"current_steps": 63,
|
|
"loss": 0.7528,
|
|
"learning_rate": 6.1e-07,
|
|
"epoch": 0.5714285714285714
|
|
},
|
|
{
|
|
"current_steps": 64,
|
|
"loss": 0.7043,
|
|
"learning_rate": 6.2e-07,
|
|
"epoch": 0.5803571428571429
|
|
},
|
|
{
|
|
"current_steps": 65,
|
|
"loss": 0.5926,
|
|
"learning_rate": 6.3e-07,
|
|
"epoch": 0.5892857142857143
|
|
},
|
|
{
|
|
"current_steps": 66,
|
|
"loss": 0.7927,
|
|
"learning_rate": 6.4e-07,
|
|
"epoch": 0.5982142857142857
|
|
},
|
|
{
|
|
"current_steps": 67,
|
|
"loss": 0.5625,
|
|
"learning_rate": 6.5e-07,
|
|
"epoch": 0.6071428571428571
|
|
},
|
|
{
|
|
"current_steps": 68,
|
|
"loss": 0.707,
|
|
"learning_rate": 6.6e-07,
|
|
"epoch": 0.6160714285714286
|
|
},
|
|
{
|
|
"current_steps": 69,
|
|
"loss": 0.7023,
|
|
"learning_rate": 6.7e-07,
|
|
"epoch": 0.625
|
|
},
|
|
{
|
|
"current_steps": 70,
|
|
"loss": 0.586,
|
|
"learning_rate": 6.800000000000001e-07,
|
|
"epoch": 0.6339285714285714
|
|
},
|
|
{
|
|
"current_steps": 71,
|
|
"loss": 0.5741,
|
|
"learning_rate": 6.9e-07,
|
|
"epoch": 0.6428571428571429
|
|
},
|
|
{
|
|
"current_steps": 72,
|
|
"loss": 1.086,
|
|
"learning_rate": 7e-07,
|
|
"epoch": 0.6517857142857143
|
|
},
|
|
{
|
|
"current_steps": 73,
|
|
"loss": 0.6381,
|
|
"learning_rate": 7.1e-07,
|
|
"epoch": 0.6607142857142857
|
|
},
|
|
{
|
|
"current_steps": 74,
|
|
"loss": 0.7509,
|
|
"learning_rate": 7.2e-07,
|
|
"epoch": 0.6696428571428571
|
|
},
|
|
{
|
|
"current_steps": 75,
|
|
"loss": 0.8276,
|
|
"learning_rate": 7.3e-07,
|
|
"epoch": 0.6785714285714286
|
|
},
|
|
{
|
|
"current_steps": 76,
|
|
"loss": 0.7623,
|
|
"learning_rate": 7.4e-07,
|
|
"epoch": 0.6875
|
|
},
|
|
{
|
|
"current_steps": 77,
|
|
"loss": 0.9499,
|
|
"learning_rate": 7.5e-07,
|
|
"epoch": 0.6964285714285714
|
|
},
|
|
{
|
|
"current_steps": 78,
|
|
"loss": 0.8563,
|
|
"learning_rate": 7.599999999999999e-07,
|
|
"epoch": 0.7053571428571429
|
|
},
|
|
{
|
|
"current_steps": 79,
|
|
"loss": 0.6512,
|
|
"learning_rate": 7.699999999999999e-07,
|
|
"epoch": 0.7142857142857143
|
|
},
|
|
{
|
|
"current_steps": 80,
|
|
"loss": 0.843,
|
|
"learning_rate": 7.799999999999999e-07,
|
|
"epoch": 0.7232142857142857
|
|
},
|
|
{
|
|
"current_steps": 81,
|
|
"loss": 0.7272,
|
|
"learning_rate": 7.9e-07,
|
|
"epoch": 0.7321428571428571
|
|
},
|
|
{
|
|
"current_steps": 82,
|
|
"loss": 0.5161,
|
|
"learning_rate": 8e-07,
|
|
"epoch": 0.7410714285714286
|
|
},
|
|
{
|
|
"current_steps": 83,
|
|
"loss": 0.8293,
|
|
"learning_rate": 8.1e-07,
|
|
"epoch": 0.75
|
|
},
|
|
{
|
|
"current_steps": 84,
|
|
"loss": 0.8704,
|
|
"learning_rate": 8.199999999999999e-07,
|
|
"epoch": 0.7589285714285714
|
|
},
|
|
{
|
|
"current_steps": 85,
|
|
"loss": 0.7255,
|
|
"learning_rate": 8.299999999999999e-07,
|
|
"epoch": 0.7678571428571429
|
|
},
|
|
{
|
|
"current_steps": 86,
|
|
"loss": 0.6252,
|
|
"learning_rate": 8.399999999999999e-07,
|
|
"epoch": 0.7767857142857143
|
|
},
|
|
{
|
|
"current_steps": 87,
|
|
"loss": 0.8116,
|
|
"learning_rate": 8.499999999999999e-07,
|
|
"epoch": 0.7857142857142857
|
|
},
|
|
{
|
|
"current_steps": 88,
|
|
"loss": 0.7703,
|
|
"learning_rate": 8.599999999999999e-07,
|
|
"epoch": 0.7946428571428571
|
|
},
|
|
{
|
|
"current_steps": 89,
|
|
"loss": 0.6496,
|
|
"learning_rate": 8.699999999999999e-07,
|
|
"epoch": 0.8035714285714286
|
|
},
|
|
{
|
|
"current_steps": 90,
|
|
"loss": 0.8585,
|
|
"learning_rate": 8.799999999999999e-07,
|
|
"epoch": 0.8125
|
|
},
|
|
{
|
|
"current_steps": 91,
|
|
"loss": 0.905,
|
|
"learning_rate": 8.9e-07,
|
|
"epoch": 0.8214285714285714
|
|
},
|
|
{
|
|
"current_steps": 92,
|
|
"loss": 0.9139,
|
|
"learning_rate": 9e-07,
|
|
"epoch": 0.8303571428571429
|
|
},
|
|
{
|
|
"current_steps": 93,
|
|
"loss": 0.9925,
|
|
"learning_rate": 9.1e-07,
|
|
"epoch": 0.8392857142857143
|
|
},
|
|
{
|
|
"current_steps": 94,
|
|
"loss": 0.7344,
|
|
"learning_rate": 9.2e-07,
|
|
"epoch": 0.8482142857142857
|
|
},
|
|
{
|
|
"current_steps": 95,
|
|
"loss": 0.7477,
|
|
"learning_rate": 9.3e-07,
|
|
"epoch": 0.8571428571428571
|
|
},
|
|
{
|
|
"current_steps": 96,
|
|
"loss": 0.671,
|
|
"learning_rate": 9.399999999999999e-07,
|
|
"epoch": 0.8660714285714286
|
|
},
|
|
{
|
|
"current_steps": 97,
|
|
"loss": 0.9654,
|
|
"learning_rate": 9.499999999999999e-07,
|
|
"epoch": 0.875
|
|
},
|
|
{
|
|
"current_steps": 98,
|
|
"loss": 0.6788,
|
|
"learning_rate": 9.6e-07,
|
|
"epoch": 0.8839285714285714
|
|
},
|
|
{
|
|
"current_steps": 99,
|
|
"loss": 0.764,
|
|
"learning_rate": 9.7e-07,
|
|
"epoch": 0.8928571428571429
|
|
},
|
|
{
|
|
"current_steps": 100,
|
|
"loss": 0.7536,
|
|
"learning_rate": 9.8e-07,
|
|
"epoch": 0.9017857142857143
|
|
},
|
|
{
|
|
"current_steps": 101,
|
|
"loss": 0.6409,
|
|
"learning_rate": 9.9e-07,
|
|
"epoch": 0.9107142857142857
|
|
},
|
|
{
|
|
"current_steps": 102,
|
|
"loss": 0.904,
|
|
"learning_rate": 1e-06,
|
|
"epoch": 0.9196428571428571
|
|
},
|
|
{
|
|
"current_steps": 103,
|
|
"loss": 0.7079,
|
|
"learning_rate": 9.978260869565217e-07,
|
|
"epoch": 0.9285714285714286
|
|
},
|
|
{
|
|
"current_steps": 104,
|
|
"loss": 0.748,
|
|
"learning_rate": 9.956521739130434e-07,
|
|
"epoch": 0.9375
|
|
},
|
|
{
|
|
"current_steps": 105,
|
|
"loss": 0.7228,
|
|
"learning_rate": 9.934782608695653e-07,
|
|
"epoch": 0.9464285714285714
|
|
},
|
|
{
|
|
"current_steps": 106,
|
|
"loss": 0.722,
|
|
"learning_rate": 9.91304347826087e-07,
|
|
"epoch": 0.9553571428571429
|
|
},
|
|
{
|
|
"current_steps": 107,
|
|
"loss": 0.8011,
|
|
"learning_rate": 9.891304347826085e-07,
|
|
"epoch": 0.9642857142857143
|
|
},
|
|
{
|
|
"current_steps": 108,
|
|
"loss": 0.8125,
|
|
"learning_rate": 9.869565217391304e-07,
|
|
"epoch": 0.9732142857142857
|
|
},
|
|
{
|
|
"current_steps": 109,
|
|
"loss": 0.8091,
|
|
"learning_rate": 9.847826086956522e-07,
|
|
"epoch": 0.9821428571428571
|
|
},
|
|
{
|
|
"current_steps": 110,
|
|
"loss": 0.9399,
|
|
"learning_rate": 9.826086956521739e-07,
|
|
"epoch": 0.9910714285714286
|
|
},
|
|
{
|
|
"current_steps": 111,
|
|
"loss": 1.0917,
|
|
"learning_rate": 9.804347826086956e-07,
|
|
"epoch": 1.0
|
|
},
|
|
{
|
|
"current_steps": 112,
|
|
"loss": 0.9014,
|
|
"learning_rate": 9.782608695652173e-07,
|
|
"epoch": 1.0089285714285714
|
|
},
|
|
{
|
|
"current_steps": 113,
|
|
"loss": 0.873,
|
|
"learning_rate": 9.782608695652173e-07,
|
|
"epoch": 1.0178571428571428
|
|
},
|
|
{
|
|
"current_steps": 114,
|
|
"loss": 0.7153,
|
|
"learning_rate": 9.76086956521739e-07,
|
|
"epoch": 1.0267857142857142
|
|
},
|
|
{
|
|
"current_steps": 115,
|
|
"loss": 0.8828,
|
|
"learning_rate": 9.73913043478261e-07,
|
|
"epoch": 1.0357142857142858
|
|
},
|
|
{
|
|
"current_steps": 116,
|
|
"loss": 1.0329,
|
|
"learning_rate": 9.717391304347827e-07,
|
|
"epoch": 1.0446428571428572
|
|
},
|
|
{
|
|
"current_steps": 117,
|
|
"loss": 1.057,
|
|
"learning_rate": 9.695652173913042e-07,
|
|
"epoch": 1.0535714285714286
|
|
},
|
|
{
|
|
"current_steps": 118,
|
|
"loss": 0.8047,
|
|
"learning_rate": 9.67391304347826e-07,
|
|
"epoch": 1.0625
|
|
},
|
|
{
|
|
"current_steps": 119,
|
|
"loss": 0.7098,
|
|
"learning_rate": 9.652173913043478e-07,
|
|
"epoch": 1.0714285714285714
|
|
},
|
|
{
|
|
"current_steps": 120,
|
|
"loss": 1.094,
|
|
"learning_rate": 9.630434782608695e-07,
|
|
"epoch": 1.0803571428571428
|
|
},
|
|
{
|
|
"current_steps": 121,
|
|
"loss": 0.7521,
|
|
"learning_rate": 9.608695652173912e-07,
|
|
"epoch": 1.0892857142857142
|
|
},
|
|
{
|
|
"current_steps": 122,
|
|
"loss": 0.9738,
|
|
"learning_rate": 9.58695652173913e-07,
|
|
"epoch": 1.0982142857142858
|
|
},
|
|
{
|
|
"current_steps": 123,
|
|
"loss": 0.5577,
|
|
"learning_rate": 9.565217391304349e-07,
|
|
"epoch": 1.1071428571428572
|
|
},
|
|
{
|
|
"current_steps": 124,
|
|
"loss": 1.046,
|
|
"learning_rate": 9.543478260869566e-07,
|
|
"epoch": 1.1160714285714286
|
|
},
|
|
{
|
|
"current_steps": 125,
|
|
"loss": 0.597,
|
|
"learning_rate": 9.521739130434783e-07,
|
|
"epoch": 1.125
|
|
},
|
|
{
|
|
"current_steps": 126,
|
|
"loss": 0.7996,
|
|
"learning_rate": 9.499999999999999e-07,
|
|
"epoch": 1.1339285714285714
|
|
},
|
|
{
|
|
"current_steps": 127,
|
|
"loss": 0.9885,
|
|
"learning_rate": 9.478260869565216e-07,
|
|
"epoch": 1.1428571428571428
|
|
},
|
|
{
|
|
"current_steps": 128,
|
|
"loss": 0.6274,
|
|
"learning_rate": 9.456521739130434e-07,
|
|
"epoch": 1.1517857142857142
|
|
},
|
|
{
|
|
"current_steps": 129,
|
|
"loss": 0.8557,
|
|
"learning_rate": 9.434782608695652e-07,
|
|
"epoch": 1.1607142857142858
|
|
},
|
|
{
|
|
"current_steps": 130,
|
|
"loss": 0.702,
|
|
"learning_rate": 9.41304347826087e-07,
|
|
"epoch": 1.1696428571428572
|
|
},
|
|
{
|
|
"current_steps": 131,
|
|
"loss": 0.6905,
|
|
"learning_rate": 9.391304347826087e-07,
|
|
"epoch": 1.1785714285714286
|
|
},
|
|
{
|
|
"current_steps": 132,
|
|
"loss": 0.5707,
|
|
"learning_rate": 9.369565217391304e-07,
|
|
"epoch": 1.1875
|
|
},
|
|
{
|
|
"current_steps": 133,
|
|
"loss": 0.6121,
|
|
"learning_rate": 9.347826086956522e-07,
|
|
"epoch": 1.1964285714285714
|
|
},
|
|
{
|
|
"current_steps": 134,
|
|
"loss": 0.8348,
|
|
"learning_rate": 9.326086956521738e-07,
|
|
"epoch": 1.2053571428571428
|
|
},
|
|
{
|
|
"current_steps": 135,
|
|
"loss": 0.8768,
|
|
"learning_rate": 9.304347826086955e-07,
|
|
"epoch": 1.2142857142857142
|
|
},
|
|
{
|
|
"current_steps": 136,
|
|
"loss": 0.5648,
|
|
"learning_rate": 9.282608695652174e-07,
|
|
"epoch": 1.2232142857142858
|
|
},
|
|
{
|
|
"current_steps": 137,
|
|
"loss": 0.6316,
|
|
"learning_rate": 9.260869565217391e-07,
|
|
"epoch": 1.2321428571428572
|
|
},
|
|
{
|
|
"current_steps": 138,
|
|
"loss": 1.1728,
|
|
"learning_rate": 9.239130434782608e-07,
|
|
"epoch": 1.2410714285714286
|
|
},
|
|
{
|
|
"current_steps": 139,
|
|
"loss": 0.7299,
|
|
"learning_rate": 9.217391304347826e-07,
|
|
"epoch": 1.25
|
|
},
|
|
{
|
|
"current_steps": 140,
|
|
"loss": 0.6284,
|
|
"learning_rate": 9.195652173913043e-07,
|
|
"epoch": 1.2589285714285714
|
|
},
|
|
{
|
|
"current_steps": 141,
|
|
"loss": 0.6366,
|
|
"learning_rate": 9.17391304347826e-07,
|
|
"epoch": 1.2678571428571428
|
|
},
|
|
{
|
|
"current_steps": 142,
|
|
"loss": 0.7357,
|
|
"learning_rate": 9.152173913043479e-07,
|
|
"epoch": 1.2767857142857144
|
|
},
|
|
{
|
|
"current_steps": 143,
|
|
"loss": 0.8618,
|
|
"learning_rate": 9.130434782608695e-07,
|
|
"epoch": 1.2857142857142856
|
|
},
|
|
{
|
|
"current_steps": 144,
|
|
"loss": 0.6803,
|
|
"learning_rate": 9.108695652173912e-07,
|
|
"epoch": 1.2946428571428572
|
|
},
|
|
{
|
|
"current_steps": 145,
|
|
"loss": 0.8093,
|
|
"learning_rate": 9.08695652173913e-07,
|
|
"epoch": 1.3035714285714286
|
|
},
|
|
{
|
|
"current_steps": 146,
|
|
"loss": 0.6808,
|
|
"learning_rate": 9.065217391304347e-07,
|
|
"epoch": 1.3125
|
|
},
|
|
{
|
|
"current_steps": 147,
|
|
"loss": 0.7173,
|
|
"learning_rate": 9.043478260869564e-07,
|
|
"epoch": 1.3214285714285714
|
|
},
|
|
{
|
|
"current_steps": 148,
|
|
"loss": 0.6964,
|
|
"learning_rate": 9.021739130434782e-07,
|
|
"epoch": 1.3303571428571428
|
|
},
|
|
{
|
|
"current_steps": 149,
|
|
"loss": 0.5458,
|
|
"learning_rate": 9e-07,
|
|
"epoch": 1.3392857142857144
|
|
},
|
|
{
|
|
"current_steps": 150,
|
|
"loss": 0.5362,
|
|
"learning_rate": 8.978260869565218e-07,
|
|
"epoch": 1.3482142857142856
|
|
},
|
|
{
|
|
"current_steps": 151,
|
|
"loss": 0.7248,
|
|
"learning_rate": 8.956521739130435e-07,
|
|
"epoch": 1.3571428571428572
|
|
},
|
|
{
|
|
"current_steps": 152,
|
|
"loss": 0.9701,
|
|
"learning_rate": 8.934782608695651e-07,
|
|
"epoch": 1.3660714285714286
|
|
},
|
|
{
|
|
"current_steps": 153,
|
|
"loss": 0.6072,
|
|
"learning_rate": 8.913043478260869e-07,
|
|
"epoch": 1.375
|
|
},
|
|
{
|
|
"current_steps": 154,
|
|
"loss": 0.8135,
|
|
"learning_rate": 8.891304347826086e-07,
|
|
"epoch": 1.3839285714285714
|
|
},
|
|
{
|
|
"current_steps": 155,
|
|
"loss": 0.6519,
|
|
"learning_rate": 8.869565217391303e-07,
|
|
"epoch": 1.3928571428571428
|
|
},
|
|
{
|
|
"current_steps": 156,
|
|
"loss": 0.7911,
|
|
"learning_rate": 8.847826086956522e-07,
|
|
"epoch": 1.4017857142857144
|
|
},
|
|
{
|
|
"current_steps": 157,
|
|
"loss": 0.7084,
|
|
"learning_rate": 8.826086956521739e-07,
|
|
"epoch": 1.4107142857142856
|
|
},
|
|
{
|
|
"current_steps": 158,
|
|
"loss": 0.6062,
|
|
"learning_rate": 8.804347826086956e-07,
|
|
"epoch": 1.4196428571428572
|
|
},
|
|
{
|
|
"current_steps": 159,
|
|
"loss": 0.5372,
|
|
"learning_rate": 8.782608695652174e-07,
|
|
"epoch": 1.4285714285714286
|
|
},
|
|
{
|
|
"current_steps": 160,
|
|
"loss": 0.7001,
|
|
"learning_rate": 8.760869565217391e-07,
|
|
"epoch": 1.4375
|
|
},
|
|
{
|
|
"current_steps": 161,
|
|
"loss": 0.628,
|
|
"learning_rate": 8.739130434782607e-07,
|
|
"epoch": 1.4464285714285714
|
|
},
|
|
{
|
|
"current_steps": 162,
|
|
"loss": 0.6766,
|
|
"learning_rate": 8.717391304347826e-07,
|
|
"epoch": 1.4553571428571428
|
|
},
|
|
{
|
|
"current_steps": 163,
|
|
"loss": 0.7406,
|
|
"learning_rate": 8.695652173913043e-07,
|
|
"epoch": 1.4642857142857144
|
|
},
|
|
{
|
|
"current_steps": 164,
|
|
"loss": 0.7032,
|
|
"learning_rate": 8.67391304347826e-07,
|
|
"epoch": 1.4732142857142856
|
|
},
|
|
{
|
|
"current_steps": 165,
|
|
"loss": 0.8338,
|
|
"learning_rate": 8.652173913043478e-07,
|
|
"epoch": 1.4821428571428572
|
|
},
|
|
{
|
|
"current_steps": 166,
|
|
"loss": 0.6067,
|
|
"learning_rate": 8.630434782608695e-07,
|
|
"epoch": 1.4910714285714286
|
|
},
|
|
{
|
|
"current_steps": 167,
|
|
"loss": 0.6988,
|
|
"learning_rate": 8.608695652173913e-07,
|
|
"epoch": 1.5
|
|
},
|
|
{
|
|
"current_steps": 168,
|
|
"loss": 0.6294,
|
|
"learning_rate": 8.586956521739131e-07,
|
|
"epoch": 1.5089285714285714
|
|
},
|
|
{
|
|
"current_steps": 169,
|
|
"loss": 0.7358,
|
|
"learning_rate": 8.565217391304348e-07,
|
|
"epoch": 1.5178571428571428
|
|
},
|
|
{
|
|
"current_steps": 170,
|
|
"loss": 0.7709,
|
|
"learning_rate": 8.543478260869565e-07,
|
|
"epoch": 1.5267857142857144
|
|
},
|
|
{
|
|
"current_steps": 171,
|
|
"loss": 0.8913,
|
|
"learning_rate": 8.521739130434782e-07,
|
|
"epoch": 1.5357142857142856
|
|
},
|
|
{
|
|
"current_steps": 172,
|
|
"loss": 0.697,
|
|
"learning_rate": 8.499999999999999e-07,
|
|
"epoch": 1.5446428571428572
|
|
},
|
|
{
|
|
"current_steps": 173,
|
|
"loss": 0.7902,
|
|
"learning_rate": 8.478260869565217e-07,
|
|
"epoch": 1.5535714285714286
|
|
},
|
|
{
|
|
"current_steps": 174,
|
|
"loss": 0.7858,
|
|
"learning_rate": 8.456521739130434e-07,
|
|
"epoch": 1.5625
|
|
},
|
|
{
|
|
"current_steps": 175,
|
|
"loss": 0.8903,
|
|
"learning_rate": 8.434782608695652e-07,
|
|
"epoch": 1.5714285714285714
|
|
},
|
|
{
|
|
"current_steps": 176,
|
|
"loss": 0.8324,
|
|
"learning_rate": 8.41304347826087e-07,
|
|
"epoch": 1.5803571428571428
|
|
},
|
|
{
|
|
"current_steps": 177,
|
|
"loss": 0.7323,
|
|
"learning_rate": 8.391304347826087e-07,
|
|
"epoch": 1.5892857142857144
|
|
},
|
|
{
|
|
"current_steps": 178,
|
|
"loss": 0.7527,
|
|
"learning_rate": 8.369565217391304e-07,
|
|
"epoch": 1.5982142857142856
|
|
},
|
|
{
|
|
"current_steps": 179,
|
|
"loss": 0.8336,
|
|
"learning_rate": 8.347826086956521e-07,
|
|
"epoch": 1.6071428571428572
|
|
},
|
|
{
|
|
"current_steps": 180,
|
|
"loss": 0.7886,
|
|
"learning_rate": 8.326086956521738e-07,
|
|
"epoch": 1.6160714285714286
|
|
},
|
|
{
|
|
"current_steps": 181,
|
|
"loss": 0.7455,
|
|
"learning_rate": 8.304347826086955e-07,
|
|
"epoch": 1.625
|
|
},
|
|
{
|
|
"current_steps": 182,
|
|
"loss": 0.7702,
|
|
"learning_rate": 8.282608695652174e-07,
|
|
"epoch": 1.6339285714285714
|
|
},
|
|
{
|
|
"current_steps": 183,
|
|
"loss": 0.6935,
|
|
"learning_rate": 8.260869565217391e-07,
|
|
"epoch": 1.6428571428571428
|
|
},
|
|
{
|
|
"current_steps": 184,
|
|
"loss": 0.6778,
|
|
"learning_rate": 8.239130434782609e-07,
|
|
"epoch": 1.6517857142857144
|
|
},
|
|
{
|
|
"current_steps": 185,
|
|
"loss": 0.7623,
|
|
"learning_rate": 8.217391304347826e-07,
|
|
"epoch": 1.6607142857142856
|
|
},
|
|
{
|
|
"current_steps": 186,
|
|
"loss": 0.8068,
|
|
"learning_rate": 8.195652173913043e-07,
|
|
"epoch": 1.6696428571428572
|
|
},
|
|
{
|
|
"current_steps": 187,
|
|
"loss": 0.6384,
|
|
"learning_rate": 8.173913043478261e-07,
|
|
"epoch": 1.6785714285714286
|
|
},
|
|
{
|
|
"current_steps": 188,
|
|
"loss": 0.9876,
|
|
"learning_rate": 8.152173913043478e-07,
|
|
"epoch": 1.6875
|
|
},
|
|
{
|
|
"current_steps": 189,
|
|
"loss": 0.5316,
|
|
"learning_rate": 8.130434782608695e-07,
|
|
"epoch": 1.6964285714285714
|
|
},
|
|
{
|
|
"current_steps": 190,
|
|
"loss": 0.6117,
|
|
"learning_rate": 8.108695652173913e-07,
|
|
"epoch": 1.7053571428571428
|
|
},
|
|
{
|
|
"current_steps": 191,
|
|
"loss": 0.5897,
|
|
"learning_rate": 8.08695652173913e-07,
|
|
"epoch": 1.7142857142857144
|
|
},
|
|
{
|
|
"current_steps": 192,
|
|
"loss": 0.7045,
|
|
"learning_rate": 8.065217391304347e-07,
|
|
"epoch": 1.7232142857142856
|
|
},
|
|
{
|
|
"current_steps": 193,
|
|
"loss": 0.7491,
|
|
"learning_rate": 8.043478260869565e-07,
|
|
"epoch": 1.7321428571428572
|
|
},
|
|
{
|
|
"current_steps": 194,
|
|
"loss": 0.8067,
|
|
"learning_rate": 8.021739130434782e-07,
|
|
"epoch": 1.7410714285714286
|
|
},
|
|
{
|
|
"current_steps": 195,
|
|
"loss": 0.9085,
|
|
"learning_rate": 8e-07,
|
|
"epoch": 1.75
|
|
},
|
|
{
|
|
"current_steps": 196,
|
|
"loss": 0.7977,
|
|
"learning_rate": 7.978260869565217e-07,
|
|
"epoch": 1.7589285714285714
|
|
},
|
|
{
|
|
"current_steps": 197,
|
|
"loss": 0.7509,
|
|
"learning_rate": 7.956521739130434e-07,
|
|
"epoch": 1.7678571428571428
|
|
},
|
|
{
|
|
"current_steps": 198,
|
|
"loss": 0.7048,
|
|
"learning_rate": 7.934782608695651e-07,
|
|
"epoch": 1.7767857142857144
|
|
},
|
|
{
|
|
"current_steps": 199,
|
|
"loss": 0.6452,
|
|
"learning_rate": 7.913043478260869e-07,
|
|
"epoch": 1.7857142857142856
|
|
},
|
|
{
|
|
"current_steps": 200,
|
|
"loss": 0.7265,
|
|
"learning_rate": 7.891304347826086e-07,
|
|
"epoch": 1.7946428571428572
|
|
},
|
|
{
|
|
"current_steps": 201,
|
|
"loss": 0.7936,
|
|
"learning_rate": 7.869565217391305e-07,
|
|
"epoch": 1.8035714285714286
|
|
},
|
|
{
|
|
"current_steps": 202,
|
|
"loss": 0.7336,
|
|
"learning_rate": 7.847826086956522e-07,
|
|
"epoch": 1.8125
|
|
},
|
|
{
|
|
"current_steps": 203,
|
|
"loss": 0.6462,
|
|
"learning_rate": 7.826086956521739e-07,
|
|
"epoch": 1.8214285714285714
|
|
},
|
|
{
|
|
"current_steps": 204,
|
|
"loss": 0.579,
|
|
"learning_rate": 7.804347826086957e-07,
|
|
"epoch": 1.8303571428571428
|
|
},
|
|
{
|
|
"current_steps": 205,
|
|
"loss": 0.6014,
|
|
"learning_rate": 7.782608695652173e-07,
|
|
"epoch": 1.8392857142857144
|
|
},
|
|
{
|
|
"current_steps": 206,
|
|
"loss": 0.684,
|
|
"learning_rate": 7.76086956521739e-07,
|
|
"epoch": 1.8482142857142856
|
|
},
|
|
{
|
|
"current_steps": 207,
|
|
"loss": 0.5932,
|
|
"learning_rate": 7.739130434782608e-07,
|
|
"epoch": 1.8571428571428572
|
|
},
|
|
{
|
|
"current_steps": 208,
|
|
"loss": 0.7736,
|
|
"learning_rate": 7.717391304347826e-07,
|
|
"epoch": 1.8660714285714286
|
|
},
|
|
{
|
|
"current_steps": 209,
|
|
"loss": 0.7601,
|
|
"learning_rate": 7.695652173913043e-07,
|
|
"epoch": 1.875
|
|
},
|
|
{
|
|
"current_steps": 210,
|
|
"loss": 0.8428,
|
|
"learning_rate": 7.673913043478261e-07,
|
|
"epoch": 1.8839285714285714
|
|
},
|
|
{
|
|
"current_steps": 211,
|
|
"loss": 0.8017,
|
|
"learning_rate": 7.652173913043478e-07,
|
|
"epoch": 1.8928571428571428
|
|
},
|
|
{
|
|
"current_steps": 212,
|
|
"loss": 0.5998,
|
|
"learning_rate": 7.630434782608695e-07,
|
|
"epoch": 1.9017857142857144
|
|
},
|
|
{
|
|
"current_steps": 213,
|
|
"loss": 0.9071,
|
|
"learning_rate": 7.608695652173913e-07,
|
|
"epoch": 1.9107142857142856
|
|
},
|
|
{
|
|
"current_steps": 214,
|
|
"loss": 0.8255,
|
|
"learning_rate": 7.58695652173913e-07,
|
|
"epoch": 1.9196428571428572
|
|
},
|
|
{
|
|
"current_steps": 215,
|
|
"loss": 0.9256,
|
|
"learning_rate": 7.565217391304347e-07,
|
|
"epoch": 1.9285714285714286
|
|
},
|
|
{
|
|
"current_steps": 216,
|
|
"loss": 0.6745,
|
|
"learning_rate": 7.543478260869565e-07,
|
|
"epoch": 1.9375
|
|
},
|
|
{
|
|
"current_steps": 217,
|
|
"loss": 0.6372,
|
|
"learning_rate": 7.521739130434782e-07,
|
|
"epoch": 1.9464285714285714
|
|
},
|
|
{
|
|
"current_steps": 218,
|
|
"loss": 0.6495,
|
|
"learning_rate": 7.5e-07,
|
|
"epoch": 1.9553571428571428
|
|
},
|
|
{
|
|
"current_steps": 219,
|
|
"loss": 0.6054,
|
|
"learning_rate": 7.478260869565217e-07,
|
|
"epoch": 1.9642857142857144
|
|
},
|
|
{
|
|
"current_steps": 220,
|
|
"loss": 0.9751,
|
|
"learning_rate": 7.478260869565217e-07,
|
|
"epoch": 1.9732142857142856
|
|
},
|
|
{
|
|
"current_steps": 221,
|
|
"loss": 0.6258,
|
|
"learning_rate": 7.456521739130434e-07,
|
|
"epoch": 1.9821428571428572
|
|
},
|
|
{
|
|
"current_steps": 222,
|
|
"loss": 0.794,
|
|
"learning_rate": 7.434782608695653e-07,
|
|
"epoch": 1.9910714285714286
|
|
},
|
|
{
|
|
"current_steps": 223,
|
|
"loss": 0.9991,
|
|
"learning_rate": 7.41304347826087e-07,
|
|
"epoch": 2.0
|
|
},
|
|
{
|
|
"current_steps": 224,
|
|
"loss": 0.8048,
|
|
"learning_rate": 7.391304347826086e-07,
|
|
"epoch": 2.0089285714285716
|
|
},
|
|
{
|
|
"current_steps": 225,
|
|
"loss": 0.8439,
|
|
"learning_rate": 7.369565217391304e-07,
|
|
"epoch": 2.017857142857143
|
|
},
|
|
{
|
|
"current_steps": 226,
|
|
"loss": 0.7546,
|
|
"learning_rate": 7.347826086956521e-07,
|
|
"epoch": 2.0267857142857144
|
|
},
|
|
{
|
|
"current_steps": 227,
|
|
"loss": 0.8195,
|
|
"learning_rate": 7.326086956521738e-07,
|
|
"epoch": 2.0357142857142856
|
|
},
|
|
{
|
|
"current_steps": 228,
|
|
"loss": 0.6988,
|
|
"learning_rate": 7.304347826086957e-07,
|
|
"epoch": 2.044642857142857
|
|
},
|
|
{
|
|
"current_steps": 229,
|
|
"loss": 0.8419,
|
|
"learning_rate": 7.282608695652174e-07,
|
|
"epoch": 2.0535714285714284
|
|
},
|
|
{
|
|
"current_steps": 230,
|
|
"loss": 0.6133,
|
|
"learning_rate": 7.260869565217391e-07,
|
|
"epoch": 2.0625
|
|
},
|
|
{
|
|
"current_steps": 231,
|
|
"loss": 0.6307,
|
|
"learning_rate": 7.239130434782609e-07,
|
|
"epoch": 2.0714285714285716
|
|
},
|
|
{
|
|
"current_steps": 232,
|
|
"loss": 0.7852,
|
|
"learning_rate": 7.217391304347826e-07,
|
|
"epoch": 2.080357142857143
|
|
},
|
|
{
|
|
"current_steps": 233,
|
|
"loss": 0.4894,
|
|
"learning_rate": 7.195652173913042e-07,
|
|
"epoch": 2.0892857142857144
|
|
},
|
|
{
|
|
"current_steps": 234,
|
|
"loss": 0.6806,
|
|
"learning_rate": 7.17391304347826e-07,
|
|
"epoch": 2.0982142857142856
|
|
},
|
|
{
|
|
"current_steps": 235,
|
|
"loss": 0.7798,
|
|
"learning_rate": 7.152173913043478e-07,
|
|
"epoch": 2.107142857142857
|
|
},
|
|
{
|
|
"current_steps": 236,
|
|
"loss": 0.934,
|
|
"learning_rate": 7.130434782608695e-07,
|
|
"epoch": 2.1160714285714284
|
|
},
|
|
{
|
|
"current_steps": 237,
|
|
"loss": 0.8044,
|
|
"learning_rate": 7.108695652173913e-07,
|
|
"epoch": 2.125
|
|
},
|
|
{
|
|
"current_steps": 238,
|
|
"loss": 0.8984,
|
|
"learning_rate": 7.08695652173913e-07,
|
|
"epoch": 2.1339285714285716
|
|
},
|
|
{
|
|
"current_steps": 239,
|
|
"loss": 0.7468,
|
|
"learning_rate": 7.065217391304348e-07,
|
|
"epoch": 2.142857142857143
|
|
},
|
|
{
|
|
"current_steps": 240,
|
|
"loss": 0.744,
|
|
"learning_rate": 7.043478260869565e-07,
|
|
"epoch": 2.1517857142857144
|
|
},
|
|
{
|
|
"current_steps": 241,
|
|
"loss": 0.5531,
|
|
"learning_rate": 7.021739130434783e-07,
|
|
"epoch": 2.1607142857142856
|
|
},
|
|
{
|
|
"current_steps": 242,
|
|
"loss": 0.8155,
|
|
"learning_rate": 7e-07,
|
|
"epoch": 2.169642857142857
|
|
},
|
|
{
|
|
"current_steps": 243,
|
|
"loss": 0.7626,
|
|
"learning_rate": 6.978260869565217e-07,
|
|
"epoch": 2.1785714285714284
|
|
},
|
|
{
|
|
"current_steps": 244,
|
|
"loss": 0.5438,
|
|
"learning_rate": 6.956521739130434e-07,
|
|
"epoch": 2.1875
|
|
},
|
|
{
|
|
"current_steps": 245,
|
|
"loss": 0.7638,
|
|
"learning_rate": 6.934782608695652e-07,
|
|
"epoch": 2.1964285714285716
|
|
},
|
|
{
|
|
"current_steps": 246,
|
|
"loss": 0.5092,
|
|
"learning_rate": 6.913043478260869e-07,
|
|
"epoch": 2.205357142857143
|
|
},
|
|
{
|
|
"current_steps": 247,
|
|
"loss": 0.7026,
|
|
"learning_rate": 6.891304347826086e-07,
|
|
"epoch": 2.2142857142857144
|
|
},
|
|
{
|
|
"current_steps": 248,
|
|
"loss": 0.727,
|
|
"learning_rate": 6.869565217391305e-07,
|
|
"epoch": 2.2232142857142856
|
|
},
|
|
{
|
|
"current_steps": 249,
|
|
"loss": 0.6229,
|
|
"learning_rate": 6.847826086956522e-07,
|
|
"epoch": 2.232142857142857
|
|
},
|
|
{
|
|
"current_steps": 250,
|
|
"loss": 0.6695,
|
|
"learning_rate": 6.826086956521738e-07,
|
|
"epoch": 2.2410714285714284
|
|
},
|
|
{
|
|
"current_steps": 251,
|
|
"loss": 0.6603,
|
|
"learning_rate": 6.804347826086956e-07,
|
|
"epoch": 2.25
|
|
},
|
|
{
|
|
"current_steps": 252,
|
|
"loss": 0.7804,
|
|
"learning_rate": 6.782608695652173e-07,
|
|
"epoch": 2.2589285714285716
|
|
},
|
|
{
|
|
"current_steps": 253,
|
|
"loss": 0.9138,
|
|
"learning_rate": 6.76086956521739e-07,
|
|
"epoch": 2.267857142857143
|
|
},
|
|
{
|
|
"current_steps": 254,
|
|
"loss": 0.7793,
|
|
"learning_rate": 6.739130434782609e-07,
|
|
"epoch": 2.2767857142857144
|
|
},
|
|
{
|
|
"current_steps": 255,
|
|
"loss": 0.7045,
|
|
"learning_rate": 6.717391304347826e-07,
|
|
"epoch": 2.2857142857142856
|
|
},
|
|
{
|
|
"current_steps": 256,
|
|
"loss": 0.8594,
|
|
"learning_rate": 6.695652173913044e-07,
|
|
"epoch": 2.294642857142857
|
|
},
|
|
{
|
|
"current_steps": 257,
|
|
"loss": 0.9529,
|
|
"learning_rate": 6.673913043478261e-07,
|
|
"epoch": 2.3035714285714284
|
|
},
|
|
{
|
|
"current_steps": 258,
|
|
"loss": 0.7477,
|
|
"learning_rate": 6.652173913043478e-07,
|
|
"epoch": 2.3125
|
|
},
|
|
{
|
|
"current_steps": 259,
|
|
"loss": 0.7676,
|
|
"learning_rate": 6.630434782608695e-07,
|
|
"epoch": 2.3214285714285716
|
|
},
|
|
{
|
|
"current_steps": 260,
|
|
"loss": 0.6468,
|
|
"learning_rate": 6.608695652173912e-07,
|
|
"epoch": 2.330357142857143
|
|
},
|
|
{
|
|
"current_steps": 261,
|
|
"loss": 0.6665,
|
|
"learning_rate": 6.58695652173913e-07,
|
|
"epoch": 2.3392857142857144
|
|
},
|
|
{
|
|
"current_steps": 262,
|
|
"loss": 0.838,
|
|
"learning_rate": 6.565217391304348e-07,
|
|
"epoch": 2.3482142857142856
|
|
},
|
|
{
|
|
"current_steps": 263,
|
|
"loss": 0.7129,
|
|
"learning_rate": 6.543478260869565e-07,
|
|
"epoch": 2.357142857142857
|
|
},
|
|
{
|
|
"current_steps": 264,
|
|
"loss": 0.8685,
|
|
"learning_rate": 6.521739130434782e-07,
|
|
"epoch": 2.3660714285714284
|
|
},
|
|
{
|
|
"current_steps": 265,
|
|
"loss": 0.7224,
|
|
"learning_rate": 6.5e-07,
|
|
"epoch": 2.375
|
|
},
|
|
{
|
|
"current_steps": 266,
|
|
"loss": 0.7037,
|
|
"learning_rate": 6.478260869565217e-07,
|
|
"epoch": 2.3839285714285716
|
|
},
|
|
{
|
|
"current_steps": 267,
|
|
"loss": 0.5596,
|
|
"learning_rate": 6.456521739130435e-07,
|
|
"epoch": 2.392857142857143
|
|
},
|
|
{
|
|
"current_steps": 268,
|
|
"loss": 0.8887,
|
|
"learning_rate": 6.434782608695652e-07,
|
|
"epoch": 2.4017857142857144
|
|
},
|
|
{
|
|
"current_steps": 269,
|
|
"loss": 0.6721,
|
|
"learning_rate": 6.413043478260869e-07,
|
|
"epoch": 2.4107142857142856
|
|
},
|
|
{
|
|
"current_steps": 270,
|
|
"loss": 0.7387,
|
|
"learning_rate": 6.391304347826086e-07,
|
|
"epoch": 2.419642857142857
|
|
},
|
|
{
|
|
"current_steps": 271,
|
|
"loss": 0.6304,
|
|
"learning_rate": 6.369565217391304e-07,
|
|
"epoch": 2.4285714285714284
|
|
},
|
|
{
|
|
"current_steps": 272,
|
|
"loss": 0.7563,
|
|
"learning_rate": 6.347826086956521e-07,
|
|
"epoch": 2.4375
|
|
},
|
|
{
|
|
"current_steps": 273,
|
|
"loss": 0.6833,
|
|
"learning_rate": 6.326086956521739e-07,
|
|
"epoch": 2.4464285714285716
|
|
},
|
|
{
|
|
"current_steps": 274,
|
|
"loss": 0.722,
|
|
"learning_rate": 6.304347826086957e-07,
|
|
"epoch": 2.455357142857143
|
|
},
|
|
{
|
|
"current_steps": 275,
|
|
"loss": 0.8583,
|
|
"learning_rate": 6.282608695652174e-07,
|
|
"epoch": 2.4642857142857144
|
|
},
|
|
{
|
|
"current_steps": 276,
|
|
"loss": 0.8988,
|
|
"learning_rate": 6.260869565217392e-07,
|
|
"epoch": 2.4732142857142856
|
|
},
|
|
{
|
|
"current_steps": 277,
|
|
"loss": 0.6269,
|
|
"learning_rate": 6.239130434782608e-07,
|
|
"epoch": 2.482142857142857
|
|
},
|
|
{
|
|
"current_steps": 278,
|
|
"loss": 0.473,
|
|
"learning_rate": 6.217391304347825e-07,
|
|
"epoch": 2.4910714285714284
|
|
},
|
|
{
|
|
"current_steps": 279,
|
|
"loss": 0.7065,
|
|
"learning_rate": 6.195652173913043e-07,
|
|
"epoch": 2.5
|
|
},
|
|
{
|
|
"current_steps": 280,
|
|
"loss": 0.7912,
|
|
"learning_rate": 6.17391304347826e-07,
|
|
"epoch": 2.508928571428571
|
|
},
|
|
{
|
|
"current_steps": 281,
|
|
"loss": 0.6589,
|
|
"learning_rate": 6.152173913043478e-07,
|
|
"epoch": 2.517857142857143
|
|
},
|
|
{
|
|
"current_steps": 282,
|
|
"loss": 0.5908,
|
|
"learning_rate": 6.130434782608696e-07,
|
|
"epoch": 2.5267857142857144
|
|
},
|
|
{
|
|
"current_steps": 283,
|
|
"loss": 0.839,
|
|
"learning_rate": 6.108695652173913e-07,
|
|
"epoch": 2.5357142857142856
|
|
},
|
|
{
|
|
"current_steps": 284,
|
|
"loss": 0.9573,
|
|
"learning_rate": 6.08695652173913e-07,
|
|
"epoch": 2.544642857142857
|
|
},
|
|
{
|
|
"current_steps": 285,
|
|
"loss": 0.8881,
|
|
"learning_rate": 6.065217391304348e-07,
|
|
"epoch": 2.553571428571429
|
|
},
|
|
{
|
|
"current_steps": 286,
|
|
"loss": 0.5213,
|
|
"learning_rate": 6.043478260869564e-07,
|
|
"epoch": 2.5625
|
|
},
|
|
{
|
|
"current_steps": 287,
|
|
"loss": 0.5668,
|
|
"learning_rate": 6.021739130434782e-07,
|
|
"epoch": 2.571428571428571
|
|
},
|
|
{
|
|
"current_steps": 288,
|
|
"loss": 0.6856,
|
|
"learning_rate": 6e-07,
|
|
"epoch": 2.580357142857143
|
|
},
|
|
{
|
|
"current_steps": 289,
|
|
"loss": 0.6793,
|
|
"learning_rate": 5.978260869565217e-07,
|
|
"epoch": 2.5892857142857144
|
|
},
|
|
{
|
|
"current_steps": 290,
|
|
"loss": 0.6176,
|
|
"learning_rate": 5.956521739130435e-07,
|
|
"epoch": 2.5982142857142856
|
|
},
|
|
{
|
|
"current_steps": 291,
|
|
"loss": 0.5633,
|
|
"learning_rate": 5.934782608695652e-07,
|
|
"epoch": 2.607142857142857
|
|
},
|
|
{
|
|
"current_steps": 292,
|
|
"loss": 0.8512,
|
|
"learning_rate": 5.913043478260869e-07,
|
|
"epoch": 2.616071428571429
|
|
},
|
|
{
|
|
"current_steps": 293,
|
|
"loss": 0.9664,
|
|
"learning_rate": 5.891304347826088e-07,
|
|
"epoch": 2.625
|
|
},
|
|
{
|
|
"current_steps": 294,
|
|
"loss": 0.6124,
|
|
"learning_rate": 5.869565217391305e-07,
|
|
"epoch": 2.633928571428571
|
|
},
|
|
{
|
|
"current_steps": 295,
|
|
"loss": 0.6244,
|
|
"learning_rate": 5.847826086956521e-07,
|
|
"epoch": 2.642857142857143
|
|
},
|
|
{
|
|
"current_steps": 296,
|
|
"loss": 0.7879,
|
|
"learning_rate": 5.826086956521739e-07,
|
|
"epoch": 2.6517857142857144
|
|
},
|
|
{
|
|
"current_steps": 297,
|
|
"loss": 0.6862,
|
|
"learning_rate": 5.804347826086956e-07,
|
|
"epoch": 2.6607142857142856
|
|
},
|
|
{
|
|
"current_steps": 298,
|
|
"loss": 0.6368,
|
|
"learning_rate": 5.782608695652173e-07,
|
|
"epoch": 2.669642857142857
|
|
},
|
|
{
|
|
"current_steps": 299,
|
|
"loss": 0.8478,
|
|
"learning_rate": 5.760869565217391e-07,
|
|
"epoch": 2.678571428571429
|
|
},
|
|
{
|
|
"current_steps": 300,
|
|
"loss": 0.6466,
|
|
"learning_rate": 5.739130434782609e-07,
|
|
"epoch": 2.6875
|
|
},
|
|
{
|
|
"current_steps": 301,
|
|
"loss": 0.7323,
|
|
"learning_rate": 5.717391304347826e-07,
|
|
"epoch": 2.696428571428571
|
|
},
|
|
{
|
|
"current_steps": 302,
|
|
"loss": 0.7611,
|
|
"learning_rate": 5.695652173913044e-07,
|
|
"epoch": 2.705357142857143
|
|
},
|
|
{
|
|
"current_steps": 303,
|
|
"loss": 0.7075,
|
|
"learning_rate": 5.673913043478261e-07,
|
|
"epoch": 2.7142857142857144
|
|
},
|
|
{
|
|
"current_steps": 304,
|
|
"loss": 0.5448,
|
|
"learning_rate": 5.652173913043477e-07,
|
|
"epoch": 2.7232142857142856
|
|
},
|
|
{
|
|
"current_steps": 305,
|
|
"loss": 0.704,
|
|
"learning_rate": 5.630434782608695e-07,
|
|
"epoch": 2.732142857142857
|
|
},
|
|
{
|
|
"current_steps": 306,
|
|
"loss": 0.8591,
|
|
"learning_rate": 5.608695652173912e-07,
|
|
"epoch": 2.741071428571429
|
|
},
|
|
{
|
|
"current_steps": 307,
|
|
"loss": 0.6702,
|
|
"learning_rate": 5.58695652173913e-07,
|
|
"epoch": 2.75
|
|
},
|
|
{
|
|
"current_steps": 308,
|
|
"loss": 0.6652,
|
|
"learning_rate": 5.565217391304348e-07,
|
|
"epoch": 2.758928571428571
|
|
},
|
|
{
|
|
"current_steps": 309,
|
|
"loss": 0.7208,
|
|
"learning_rate": 5.543478260869565e-07,
|
|
"epoch": 2.767857142857143
|
|
},
|
|
{
|
|
"current_steps": 310,
|
|
"loss": 0.7334,
|
|
"learning_rate": 5.521739130434783e-07,
|
|
"epoch": 2.7767857142857144
|
|
},
|
|
{
|
|
"current_steps": 311,
|
|
"loss": 0.865,
|
|
"learning_rate": 5.5e-07,
|
|
"epoch": 2.7857142857142856
|
|
},
|
|
{
|
|
"current_steps": 312,
|
|
"loss": 0.5955,
|
|
"learning_rate": 5.478260869565216e-07,
|
|
"epoch": 2.794642857142857
|
|
},
|
|
{
|
|
"current_steps": 313,
|
|
"loss": 0.5059,
|
|
"learning_rate": 5.456521739130435e-07,
|
|
"epoch": 2.803571428571429
|
|
},
|
|
{
|
|
"current_steps": 314,
|
|
"loss": 1.0855,
|
|
"learning_rate": 5.434782608695652e-07,
|
|
"epoch": 2.8125
|
|
},
|
|
{
|
|
"current_steps": 315,
|
|
"loss": 0.7484,
|
|
"learning_rate": 5.413043478260869e-07,
|
|
"epoch": 2.821428571428571
|
|
},
|
|
{
|
|
"current_steps": 316,
|
|
"loss": 0.8017,
|
|
"learning_rate": 5.391304347826087e-07,
|
|
"epoch": 2.830357142857143
|
|
},
|
|
{
|
|
"current_steps": 317,
|
|
"loss": 0.7272,
|
|
"learning_rate": 5.369565217391304e-07,
|
|
"epoch": 2.8392857142857144
|
|
},
|
|
{
|
|
"current_steps": 318,
|
|
"loss": 0.6897,
|
|
"learning_rate": 5.347826086956521e-07,
|
|
"epoch": 2.8482142857142856
|
|
},
|
|
{
|
|
"current_steps": 319,
|
|
"loss": 0.634,
|
|
"learning_rate": 5.32608695652174e-07,
|
|
"epoch": 2.857142857142857
|
|
},
|
|
{
|
|
"current_steps": 320,
|
|
"loss": 0.7684,
|
|
"learning_rate": 5.304347826086957e-07,
|
|
"epoch": 2.866071428571429
|
|
},
|
|
{
|
|
"current_steps": 321,
|
|
"loss": 0.5758,
|
|
"learning_rate": 5.282608695652173e-07,
|
|
"epoch": 2.875
|
|
},
|
|
{
|
|
"current_steps": 322,
|
|
"loss": 0.687,
|
|
"learning_rate": 5.260869565217391e-07,
|
|
"epoch": 2.883928571428571
|
|
},
|
|
{
|
|
"current_steps": 323,
|
|
"loss": 0.6942,
|
|
"learning_rate": 5.239130434782608e-07,
|
|
"epoch": 2.892857142857143
|
|
},
|
|
{
|
|
"current_steps": 324,
|
|
"loss": 0.7698,
|
|
"learning_rate": 5.217391304347825e-07,
|
|
"epoch": 2.9017857142857144
|
|
},
|
|
{
|
|
"current_steps": 325,
|
|
"loss": 0.815,
|
|
"learning_rate": 5.195652173913043e-07,
|
|
"epoch": 2.9107142857142856
|
|
},
|
|
{
|
|
"current_steps": 326,
|
|
"loss": 0.6837,
|
|
"learning_rate": 5.173913043478261e-07,
|
|
"epoch": 2.919642857142857
|
|
},
|
|
{
|
|
"current_steps": 327,
|
|
"loss": 0.7103,
|
|
"learning_rate": 5.152173913043479e-07,
|
|
"epoch": 2.928571428571429
|
|
},
|
|
{
|
|
"current_steps": 328,
|
|
"loss": 0.6798,
|
|
"learning_rate": 5.130434782608696e-07,
|
|
"epoch": 2.9375
|
|
},
|
|
{
|
|
"current_steps": 329,
|
|
"loss": 0.767,
|
|
"learning_rate": 5.108695652173913e-07,
|
|
"epoch": 2.946428571428571
|
|
},
|
|
{
|
|
"current_steps": 330,
|
|
"loss": 0.6161,
|
|
"learning_rate": 5.08695652173913e-07,
|
|
"epoch": 2.955357142857143
|
|
},
|
|
{
|
|
"current_steps": 331,
|
|
"loss": 0.6607,
|
|
"learning_rate": 5.065217391304347e-07,
|
|
"epoch": 2.9642857142857144
|
|
},
|
|
{
|
|
"current_steps": 332,
|
|
"loss": 0.6875,
|
|
"learning_rate": 5.043478260869564e-07,
|
|
"epoch": 2.9732142857142856
|
|
},
|
|
{
|
|
"current_steps": 333,
|
|
"loss": 0.746,
|
|
"learning_rate": 5.021739130434783e-07,
|
|
"epoch": 2.982142857142857
|
|
},
|
|
{
|
|
"current_steps": 334,
|
|
"loss": 0.6093,
|
|
"learning_rate": 5e-07,
|
|
"epoch": 2.991071428571429
|
|
},
|
|
{
|
|
"current_steps": 335,
|
|
"loss": 0.5599,
|
|
"learning_rate": 4.978260869565217e-07,
|
|
"epoch": 3.0
|
|
},
|
|
{
|
|
"current_steps": 336,
|
|
"loss": 0.5985,
|
|
"learning_rate": 4.956521739130435e-07,
|
|
"epoch": 3.0089285714285716
|
|
},
|
|
{
|
|
"current_steps": 337,
|
|
"loss": 0.6692,
|
|
"learning_rate": 4.934782608695652e-07,
|
|
"epoch": 3.017857142857143
|
|
},
|
|
{
|
|
"current_steps": 338,
|
|
"loss": 0.5887,
|
|
"learning_rate": 4.913043478260869e-07,
|
|
"epoch": 3.0267857142857144
|
|
},
|
|
{
|
|
"current_steps": 339,
|
|
"loss": 0.5831,
|
|
"learning_rate": 4.891304347826087e-07,
|
|
"epoch": 3.0357142857142856
|
|
},
|
|
{
|
|
"current_steps": 340,
|
|
"loss": 0.5424,
|
|
"learning_rate": 4.869565217391305e-07,
|
|
"epoch": 3.044642857142857
|
|
},
|
|
{
|
|
"current_steps": 341,
|
|
"loss": 1.0041,
|
|
"learning_rate": 4.847826086956521e-07,
|
|
"epoch": 3.0535714285714284
|
|
},
|
|
{
|
|
"current_steps": 342,
|
|
"loss": 0.6989,
|
|
"learning_rate": 4.826086956521739e-07,
|
|
"epoch": 3.0625
|
|
},
|
|
{
|
|
"current_steps": 343,
|
|
"loss": 0.7104,
|
|
"learning_rate": 4.804347826086956e-07,
|
|
"epoch": 3.0714285714285716
|
|
},
|
|
{
|
|
"current_steps": 344,
|
|
"loss": 0.6493,
|
|
"learning_rate": 4.782608695652174e-07,
|
|
"epoch": 3.080357142857143
|
|
},
|
|
{
|
|
"current_steps": 345,
|
|
"loss": 0.8018,
|
|
"learning_rate": 4.7608695652173915e-07,
|
|
"epoch": 3.0892857142857144
|
|
},
|
|
{
|
|
"current_steps": 346,
|
|
"loss": 0.638,
|
|
"learning_rate": 4.739130434782608e-07,
|
|
"epoch": 3.0982142857142856
|
|
},
|
|
{
|
|
"current_steps": 347,
|
|
"loss": 0.7714,
|
|
"learning_rate": 4.717391304347826e-07,
|
|
"epoch": 3.107142857142857
|
|
},
|
|
{
|
|
"current_steps": 348,
|
|
"loss": 0.7103,
|
|
"learning_rate": 4.6956521739130434e-07,
|
|
"epoch": 3.1160714285714284
|
|
},
|
|
{
|
|
"current_steps": 349,
|
|
"loss": 0.5937,
|
|
"learning_rate": 4.673913043478261e-07,
|
|
"epoch": 3.125
|
|
},
|
|
{
|
|
"current_steps": 350,
|
|
"loss": 0.7256,
|
|
"learning_rate": 4.6521739130434777e-07,
|
|
"epoch": 3.1339285714285716
|
|
},
|
|
{
|
|
"current_steps": 351,
|
|
"loss": 0.864,
|
|
"learning_rate": 4.6304347826086954e-07,
|
|
"epoch": 3.142857142857143
|
|
},
|
|
{
|
|
"current_steps": 352,
|
|
"loss": 0.7429,
|
|
"learning_rate": 4.608695652173913e-07,
|
|
"epoch": 3.1517857142857144
|
|
},
|
|
{
|
|
"current_steps": 353,
|
|
"loss": 0.6658,
|
|
"learning_rate": 4.58695652173913e-07,
|
|
"epoch": 3.1607142857142856
|
|
},
|
|
{
|
|
"current_steps": 354,
|
|
"loss": 0.647,
|
|
"learning_rate": 4.5652173913043473e-07,
|
|
"epoch": 3.169642857142857
|
|
},
|
|
{
|
|
"current_steps": 355,
|
|
"loss": 0.7772,
|
|
"learning_rate": 4.543478260869565e-07,
|
|
"epoch": 3.1785714285714284
|
|
},
|
|
{
|
|
"current_steps": 356,
|
|
"loss": 0.6939,
|
|
"learning_rate": 4.521739130434782e-07,
|
|
"epoch": 3.1875
|
|
},
|
|
{
|
|
"current_steps": 357,
|
|
"loss": 0.5744,
|
|
"learning_rate": 4.5e-07,
|
|
"epoch": 3.1964285714285716
|
|
},
|
|
{
|
|
"current_steps": 358,
|
|
"loss": 0.7193,
|
|
"learning_rate": 4.4782608695652175e-07,
|
|
"epoch": 3.205357142857143
|
|
},
|
|
{
|
|
"current_steps": 359,
|
|
"loss": 0.667,
|
|
"learning_rate": 4.4565217391304346e-07,
|
|
"epoch": 3.2142857142857144
|
|
},
|
|
{
|
|
"current_steps": 360,
|
|
"loss": 0.6671,
|
|
"learning_rate": 4.434782608695652e-07,
|
|
"epoch": 3.2232142857142856
|
|
},
|
|
{
|
|
"current_steps": 361,
|
|
"loss": 0.8531,
|
|
"learning_rate": 4.4130434782608694e-07,
|
|
"epoch": 3.232142857142857
|
|
},
|
|
{
|
|
"current_steps": 362,
|
|
"loss": 0.6706,
|
|
"learning_rate": 4.391304347826087e-07,
|
|
"epoch": 3.2410714285714284
|
|
},
|
|
{
|
|
"current_steps": 363,
|
|
"loss": 0.8786,
|
|
"learning_rate": 4.3695652173913037e-07,
|
|
"epoch": 3.25
|
|
},
|
|
{
|
|
"current_steps": 364,
|
|
"loss": 0.6281,
|
|
"learning_rate": 4.3478260869565214e-07,
|
|
"epoch": 3.2589285714285716
|
|
},
|
|
{
|
|
"current_steps": 365,
|
|
"loss": 0.8648,
|
|
"learning_rate": 4.326086956521739e-07,
|
|
"epoch": 3.267857142857143
|
|
},
|
|
{
|
|
"current_steps": 366,
|
|
"loss": 0.5872,
|
|
"learning_rate": 4.3043478260869567e-07,
|
|
"epoch": 3.2767857142857144
|
|
},
|
|
{
|
|
"current_steps": 367,
|
|
"loss": 0.5874,
|
|
"learning_rate": 4.282608695652174e-07,
|
|
"epoch": 3.2857142857142856
|
|
},
|
|
{
|
|
"current_steps": 368,
|
|
"loss": 0.7057,
|
|
"learning_rate": 4.260869565217391e-07,
|
|
"epoch": 3.294642857142857
|
|
},
|
|
{
|
|
"current_steps": 369,
|
|
"loss": 0.6076,
|
|
"learning_rate": 4.2391304347826086e-07,
|
|
"epoch": 3.3035714285714284
|
|
},
|
|
{
|
|
"current_steps": 370,
|
|
"loss": 0.7514,
|
|
"learning_rate": 4.217391304347826e-07,
|
|
"epoch": 3.3125
|
|
},
|
|
{
|
|
"current_steps": 371,
|
|
"loss": 0.689,
|
|
"learning_rate": 4.1956521739130434e-07,
|
|
"epoch": 3.3214285714285716
|
|
},
|
|
{
|
|
"current_steps": 372,
|
|
"loss": 0.7074,
|
|
"learning_rate": 4.1739130434782606e-07,
|
|
"epoch": 3.330357142857143
|
|
},
|
|
{
|
|
"current_steps": 373,
|
|
"loss": 0.6425,
|
|
"learning_rate": 4.1521739130434777e-07,
|
|
"epoch": 3.3392857142857144
|
|
},
|
|
{
|
|
"current_steps": 374,
|
|
"loss": 0.5247,
|
|
"learning_rate": 4.1304347826086954e-07,
|
|
"epoch": 3.3482142857142856
|
|
},
|
|
{
|
|
"current_steps": 375,
|
|
"loss": 0.7755,
|
|
"learning_rate": 4.108695652173913e-07,
|
|
"epoch": 3.357142857142857
|
|
},
|
|
{
|
|
"current_steps": 376,
|
|
"loss": 0.7774,
|
|
"learning_rate": 4.0869565217391307e-07,
|
|
"epoch": 3.3660714285714284
|
|
},
|
|
{
|
|
"current_steps": 377,
|
|
"loss": 0.6871,
|
|
"learning_rate": 4.0652173913043473e-07,
|
|
"epoch": 3.375
|
|
},
|
|
{
|
|
"current_steps": 378,
|
|
"loss": 0.566,
|
|
"learning_rate": 4.043478260869565e-07,
|
|
"epoch": 3.3839285714285716
|
|
},
|
|
{
|
|
"current_steps": 379,
|
|
"loss": 1.0922,
|
|
"learning_rate": 4.0217391304347827e-07,
|
|
"epoch": 3.392857142857143
|
|
},
|
|
{
|
|
"current_steps": 380,
|
|
"loss": 0.5958,
|
|
"learning_rate": 4e-07,
|
|
"epoch": 3.4017857142857144
|
|
},
|
|
{
|
|
"current_steps": 381,
|
|
"loss": 0.9182,
|
|
"learning_rate": 3.978260869565217e-07,
|
|
"epoch": 3.4107142857142856
|
|
},
|
|
{
|
|
"current_steps": 382,
|
|
"loss": 0.7356,
|
|
"learning_rate": 3.9565217391304346e-07,
|
|
"epoch": 3.419642857142857
|
|
},
|
|
{
|
|
"current_steps": 383,
|
|
"loss": 0.8677,
|
|
"learning_rate": 3.9347826086956523e-07,
|
|
"epoch": 3.4285714285714284
|
|
},
|
|
{
|
|
"current_steps": 384,
|
|
"loss": 0.6885,
|
|
"learning_rate": 3.9130434782608694e-07,
|
|
"epoch": 3.4375
|
|
},
|
|
{
|
|
"current_steps": 385,
|
|
"loss": 0.7982,
|
|
"learning_rate": 3.8913043478260866e-07,
|
|
"epoch": 3.4464285714285716
|
|
},
|
|
{
|
|
"current_steps": 386,
|
|
"loss": 0.8466,
|
|
"learning_rate": 3.869565217391304e-07,
|
|
"epoch": 3.455357142857143
|
|
},
|
|
{
|
|
"current_steps": 387,
|
|
"loss": 0.4563,
|
|
"learning_rate": 3.8478260869565214e-07,
|
|
"epoch": 3.4642857142857144
|
|
},
|
|
{
|
|
"current_steps": 388,
|
|
"loss": 0.7675,
|
|
"learning_rate": 3.826086956521739e-07,
|
|
"epoch": 3.4732142857142856
|
|
},
|
|
{
|
|
"current_steps": 389,
|
|
"loss": 0.7642,
|
|
"learning_rate": 3.8043478260869567e-07,
|
|
"epoch": 3.482142857142857
|
|
},
|
|
{
|
|
"current_steps": 390,
|
|
"loss": 0.6065,
|
|
"learning_rate": 3.7826086956521733e-07,
|
|
"epoch": 3.4910714285714284
|
|
},
|
|
{
|
|
"current_steps": 391,
|
|
"loss": 0.6121,
|
|
"learning_rate": 3.760869565217391e-07,
|
|
"epoch": 3.5
|
|
},
|
|
{
|
|
"current_steps": 392,
|
|
"loss": 0.8562,
|
|
"learning_rate": 3.7391304347826087e-07,
|
|
"epoch": 3.508928571428571
|
|
},
|
|
{
|
|
"current_steps": 393,
|
|
"loss": 0.8169,
|
|
"learning_rate": 3.7173913043478263e-07,
|
|
"epoch": 3.517857142857143
|
|
},
|
|
{
|
|
"current_steps": 394,
|
|
"loss": 0.7264,
|
|
"learning_rate": 3.695652173913043e-07,
|
|
"epoch": 3.5267857142857144
|
|
},
|
|
{
|
|
"current_steps": 395,
|
|
"loss": 0.6761,
|
|
"learning_rate": 3.6739130434782606e-07,
|
|
"epoch": 3.5357142857142856
|
|
},
|
|
{
|
|
"current_steps": 396,
|
|
"loss": 0.485,
|
|
"learning_rate": 3.6521739130434783e-07,
|
|
"epoch": 3.544642857142857
|
|
},
|
|
{
|
|
"current_steps": 397,
|
|
"loss": 0.6992,
|
|
"learning_rate": 3.6304347826086954e-07,
|
|
"epoch": 3.553571428571429
|
|
},
|
|
{
|
|
"current_steps": 398,
|
|
"loss": 0.6543,
|
|
"learning_rate": 3.608695652173913e-07,
|
|
"epoch": 3.5625
|
|
},
|
|
{
|
|
"current_steps": 399,
|
|
"loss": 0.6019,
|
|
"learning_rate": 3.58695652173913e-07,
|
|
"epoch": 3.571428571428571
|
|
},
|
|
{
|
|
"current_steps": 400,
|
|
"loss": 0.8135,
|
|
"learning_rate": 3.5652173913043474e-07,
|
|
"epoch": 3.580357142857143
|
|
},
|
|
{
|
|
"current_steps": 401,
|
|
"loss": 0.5053,
|
|
"learning_rate": 3.543478260869565e-07,
|
|
"epoch": 3.5892857142857144
|
|
},
|
|
{
|
|
"current_steps": 402,
|
|
"loss": 0.6121,
|
|
"learning_rate": 3.5217391304347827e-07,
|
|
"epoch": 3.5982142857142856
|
|
},
|
|
{
|
|
"current_steps": 403,
|
|
"loss": 0.5648,
|
|
"learning_rate": 3.5e-07,
|
|
"epoch": 3.607142857142857
|
|
},
|
|
{
|
|
"current_steps": 404,
|
|
"loss": 0.6023,
|
|
"learning_rate": 3.478260869565217e-07,
|
|
"epoch": 3.616071428571429
|
|
},
|
|
{
|
|
"current_steps": 405,
|
|
"loss": 0.7843,
|
|
"learning_rate": 3.4565217391304346e-07,
|
|
"epoch": 3.625
|
|
},
|
|
{
|
|
"current_steps": 406,
|
|
"loss": 0.6902,
|
|
"learning_rate": 3.4347826086956523e-07,
|
|
"epoch": 3.633928571428571
|
|
},
|
|
{
|
|
"current_steps": 407,
|
|
"loss": 0.6103,
|
|
"learning_rate": 3.413043478260869e-07,
|
|
"epoch": 3.642857142857143
|
|
},
|
|
{
|
|
"current_steps": 408,
|
|
"loss": 0.759,
|
|
"learning_rate": 3.3913043478260866e-07,
|
|
"epoch": 3.6517857142857144
|
|
},
|
|
{
|
|
"current_steps": 409,
|
|
"loss": 0.7823,
|
|
"learning_rate": 3.369565217391304e-07,
|
|
"epoch": 3.6607142857142856
|
|
},
|
|
{
|
|
"current_steps": 410,
|
|
"loss": 0.8021,
|
|
"learning_rate": 3.347826086956522e-07,
|
|
"epoch": 3.669642857142857
|
|
},
|
|
{
|
|
"current_steps": 411,
|
|
"loss": 0.5927,
|
|
"learning_rate": 3.326086956521739e-07,
|
|
"epoch": 3.678571428571429
|
|
},
|
|
{
|
|
"current_steps": 412,
|
|
"loss": 0.6503,
|
|
"learning_rate": 3.304347826086956e-07,
|
|
"epoch": 3.6875
|
|
},
|
|
{
|
|
"current_steps": 413,
|
|
"loss": 0.886,
|
|
"learning_rate": 3.282608695652174e-07,
|
|
"epoch": 3.696428571428571
|
|
},
|
|
{
|
|
"current_steps": 414,
|
|
"loss": 0.6331,
|
|
"learning_rate": 3.260869565217391e-07,
|
|
"epoch": 3.705357142857143
|
|
},
|
|
{
|
|
"current_steps": 415,
|
|
"loss": 0.7633,
|
|
"learning_rate": 3.2391304347826087e-07,
|
|
"epoch": 3.7142857142857144
|
|
},
|
|
{
|
|
"current_steps": 416,
|
|
"loss": 0.6538,
|
|
"learning_rate": 3.217391304347826e-07,
|
|
"epoch": 3.7232142857142856
|
|
},
|
|
{
|
|
"current_steps": 417,
|
|
"loss": 0.6156,
|
|
"learning_rate": 3.195652173913043e-07,
|
|
"epoch": 3.732142857142857
|
|
},
|
|
{
|
|
"current_steps": 418,
|
|
"loss": 0.6973,
|
|
"learning_rate": 3.1739130434782606e-07,
|
|
"epoch": 3.741071428571429
|
|
},
|
|
{
|
|
"current_steps": 419,
|
|
"loss": 0.6521,
|
|
"learning_rate": 3.1521739130434783e-07,
|
|
"epoch": 3.75
|
|
},
|
|
{
|
|
"current_steps": 420,
|
|
"loss": 0.6931,
|
|
"learning_rate": 3.130434782608696e-07,
|
|
"epoch": 3.758928571428571
|
|
},
|
|
{
|
|
"current_steps": 421,
|
|
"loss": 0.8192,
|
|
"learning_rate": 3.1086956521739126e-07,
|
|
"epoch": 3.767857142857143
|
|
},
|
|
{
|
|
"current_steps": 422,
|
|
"loss": 0.5986,
|
|
"learning_rate": 3.08695652173913e-07,
|
|
"epoch": 3.7767857142857144
|
|
},
|
|
{
|
|
"current_steps": 423,
|
|
"loss": 0.9986,
|
|
"learning_rate": 3.065217391304348e-07,
|
|
"epoch": 3.7857142857142856
|
|
},
|
|
{
|
|
"current_steps": 424,
|
|
"loss": 0.7645,
|
|
"learning_rate": 3.043478260869565e-07,
|
|
"epoch": 3.794642857142857
|
|
},
|
|
{
|
|
"current_steps": 425,
|
|
"loss": 0.6489,
|
|
"learning_rate": 3.021739130434782e-07,
|
|
"epoch": 3.803571428571429
|
|
},
|
|
{
|
|
"current_steps": 426,
|
|
"loss": 0.5974,
|
|
"learning_rate": 3e-07,
|
|
"epoch": 3.8125
|
|
},
|
|
{
|
|
"current_steps": 427,
|
|
"loss": 0.7392,
|
|
"learning_rate": 2.9782608695652175e-07,
|
|
"epoch": 3.821428571428571
|
|
},
|
|
{
|
|
"current_steps": 428,
|
|
"loss": 0.7813,
|
|
"learning_rate": 2.9565217391304347e-07,
|
|
"epoch": 3.830357142857143
|
|
},
|
|
{
|
|
"current_steps": 429,
|
|
"loss": 0.7818,
|
|
"learning_rate": 2.9347826086956523e-07,
|
|
"epoch": 3.8392857142857144
|
|
},
|
|
{
|
|
"current_steps": 430,
|
|
"loss": 1.0693,
|
|
"learning_rate": 2.9130434782608695e-07,
|
|
"epoch": 3.8482142857142856
|
|
},
|
|
{
|
|
"current_steps": 431,
|
|
"loss": 0.6324,
|
|
"learning_rate": 2.8913043478260866e-07,
|
|
"epoch": 3.857142857142857
|
|
},
|
|
{
|
|
"current_steps": 432,
|
|
"loss": 0.5228,
|
|
"learning_rate": 2.8695652173913043e-07,
|
|
"epoch": 3.866071428571429
|
|
},
|
|
{
|
|
"current_steps": 433,
|
|
"loss": 0.6631,
|
|
"learning_rate": 2.847826086956522e-07,
|
|
"epoch": 3.875
|
|
},
|
|
{
|
|
"current_steps": 434,
|
|
"loss": 0.6685,
|
|
"learning_rate": 2.8260869565217386e-07,
|
|
"epoch": 3.883928571428571
|
|
},
|
|
{
|
|
"current_steps": 435,
|
|
"loss": 0.6566,
|
|
"learning_rate": 2.804347826086956e-07,
|
|
"epoch": 3.892857142857143
|
|
},
|
|
{
|
|
"current_steps": 436,
|
|
"loss": 0.6169,
|
|
"learning_rate": 2.782608695652174e-07,
|
|
"epoch": 3.9017857142857144
|
|
},
|
|
{
|
|
"current_steps": 437,
|
|
"loss": 0.5012,
|
|
"learning_rate": 2.7608695652173916e-07,
|
|
"epoch": 3.9107142857142856
|
|
},
|
|
{
|
|
"current_steps": 438,
|
|
"loss": 0.637,
|
|
"learning_rate": 2.739130434782608e-07,
|
|
"epoch": 3.919642857142857
|
|
},
|
|
{
|
|
"current_steps": 439,
|
|
"loss": 0.7777,
|
|
"learning_rate": 2.717391304347826e-07,
|
|
"epoch": 3.928571428571429
|
|
},
|
|
{
|
|
"current_steps": 440,
|
|
"loss": 0.6963,
|
|
"learning_rate": 2.6956521739130435e-07,
|
|
"epoch": 3.9375
|
|
},
|
|
{
|
|
"current_steps": 441,
|
|
"loss": 0.5398,
|
|
"learning_rate": 2.6739130434782607e-07,
|
|
"epoch": 3.946428571428571
|
|
},
|
|
{
|
|
"current_steps": 442,
|
|
"loss": 1.0029,
|
|
"learning_rate": 2.6521739130434783e-07,
|
|
"epoch": 3.955357142857143
|
|
},
|
|
{
|
|
"current_steps": 443,
|
|
"loss": 0.8166,
|
|
"learning_rate": 2.6304347826086955e-07,
|
|
"epoch": 3.9642857142857144
|
|
},
|
|
{
|
|
"current_steps": 444,
|
|
"loss": 0.8981,
|
|
"learning_rate": 2.6086956521739126e-07,
|
|
"epoch": 3.9732142857142856
|
|
},
|
|
{
|
|
"current_steps": 445,
|
|
"loss": 0.536,
|
|
"learning_rate": 2.5869565217391303e-07,
|
|
"epoch": 3.982142857142857
|
|
},
|
|
{
|
|
"current_steps": 446,
|
|
"loss": 0.7719,
|
|
"learning_rate": 2.565217391304348e-07,
|
|
"epoch": 3.991071428571429
|
|
},
|
|
{
|
|
"current_steps": 447,
|
|
"loss": 3.9574,
|
|
"learning_rate": 2.565217391304348e-07,
|
|
"epoch": 4.0
|
|
},
|
|
{
|
|
"current_steps": 448,
|
|
"loss": 0.6567,
|
|
"learning_rate": 2.543478260869565e-07,
|
|
"epoch": 4.008928571428571
|
|
},
|
|
{
|
|
"current_steps": 449,
|
|
"loss": 0.8622,
|
|
"learning_rate": 2.521739130434782e-07,
|
|
"epoch": 4.017857142857143
|
|
},
|
|
{
|
|
"current_steps": 450,
|
|
"loss": 0.5737,
|
|
"learning_rate": 2.5e-07,
|
|
"epoch": 4.026785714285714
|
|
},
|
|
{
|
|
"current_steps": 451,
|
|
"loss": 0.736,
|
|
"learning_rate": 2.4782608695652176e-07,
|
|
"epoch": 4.035714285714286
|
|
},
|
|
{
|
|
"current_steps": 452,
|
|
"loss": 0.8457,
|
|
"learning_rate": 2.4565217391304347e-07,
|
|
"epoch": 4.044642857142857
|
|
},
|
|
{
|
|
"current_steps": 453,
|
|
"loss": 0.7416,
|
|
"learning_rate": 2.4347826086956524e-07,
|
|
"epoch": 4.053571428571429
|
|
},
|
|
{
|
|
"current_steps": 454,
|
|
"loss": 1.0355,
|
|
"learning_rate": 2.4130434782608695e-07,
|
|
"epoch": 4.0625
|
|
},
|
|
{
|
|
"current_steps": 455,
|
|
"loss": 0.7162,
|
|
"learning_rate": 2.391304347826087e-07,
|
|
"epoch": 4.071428571428571
|
|
},
|
|
{
|
|
"current_steps": 456,
|
|
"loss": 0.8163,
|
|
"learning_rate": 2.369565217391304e-07,
|
|
"epoch": 4.080357142857143
|
|
},
|
|
{
|
|
"current_steps": 457,
|
|
"loss": 0.5188,
|
|
"learning_rate": 2.3478260869565217e-07,
|
|
"epoch": 4.089285714285714
|
|
},
|
|
{
|
|
"current_steps": 458,
|
|
"loss": 0.9544,
|
|
"learning_rate": 2.3260869565217389e-07,
|
|
"epoch": 4.098214285714286
|
|
},
|
|
{
|
|
"current_steps": 459,
|
|
"loss": 0.6205,
|
|
"learning_rate": 2.3043478260869565e-07,
|
|
"epoch": 4.107142857142857
|
|
},
|
|
{
|
|
"current_steps": 460,
|
|
"loss": 0.6643,
|
|
"learning_rate": 2.2826086956521737e-07,
|
|
"epoch": 4.116071428571429
|
|
},
|
|
{
|
|
"current_steps": 461,
|
|
"loss": 0.6465,
|
|
"learning_rate": 2.260869565217391e-07,
|
|
"epoch": 4.125
|
|
},
|
|
{
|
|
"current_steps": 462,
|
|
"loss": 0.6697,
|
|
"learning_rate": 2.2391304347826087e-07,
|
|
"epoch": 4.133928571428571
|
|
},
|
|
{
|
|
"current_steps": 463,
|
|
"loss": 0.7041,
|
|
"learning_rate": 2.217391304347826e-07,
|
|
"epoch": 4.142857142857143
|
|
},
|
|
{
|
|
"current_steps": 464,
|
|
"loss": 0.802,
|
|
"learning_rate": 2.1956521739130435e-07,
|
|
"epoch": 4.151785714285714
|
|
},
|
|
{
|
|
"current_steps": 465,
|
|
"loss": 0.623,
|
|
"learning_rate": 2.1739130434782607e-07,
|
|
"epoch": 4.160714285714286
|
|
},
|
|
{
|
|
"current_steps": 466,
|
|
"loss": 0.6071,
|
|
"learning_rate": 2.1521739130434783e-07,
|
|
"epoch": 4.169642857142857
|
|
},
|
|
{
|
|
"current_steps": 467,
|
|
"loss": 0.718,
|
|
"learning_rate": 2.1304347826086955e-07,
|
|
"epoch": 4.178571428571429
|
|
},
|
|
{
|
|
"current_steps": 468,
|
|
"loss": 0.6337,
|
|
"learning_rate": 2.108695652173913e-07,
|
|
"epoch": 4.1875
|
|
},
|
|
{
|
|
"current_steps": 469,
|
|
"loss": 0.5689,
|
|
"learning_rate": 2.0869565217391303e-07,
|
|
"epoch": 4.196428571428571
|
|
},
|
|
{
|
|
"current_steps": 470,
|
|
"loss": 0.62,
|
|
"learning_rate": 2.0652173913043477e-07,
|
|
"epoch": 4.205357142857143
|
|
},
|
|
{
|
|
"current_steps": 471,
|
|
"loss": 1.0191,
|
|
"learning_rate": 2.0434782608695654e-07,
|
|
"epoch": 4.214285714285714
|
|
},
|
|
{
|
|
"current_steps": 472,
|
|
"loss": 0.6678,
|
|
"learning_rate": 2.0217391304347825e-07,
|
|
"epoch": 4.223214285714286
|
|
},
|
|
{
|
|
"current_steps": 473,
|
|
"loss": 0.6296,
|
|
"learning_rate": 2e-07,
|
|
"epoch": 4.232142857142857
|
|
},
|
|
{
|
|
"current_steps": 474,
|
|
"loss": 0.884,
|
|
"learning_rate": 1.9782608695652173e-07,
|
|
"epoch": 4.241071428571429
|
|
},
|
|
{
|
|
"current_steps": 475,
|
|
"loss": 0.7207,
|
|
"learning_rate": 1.9565217391304347e-07,
|
|
"epoch": 4.25
|
|
},
|
|
{
|
|
"current_steps": 476,
|
|
"loss": 0.6856,
|
|
"learning_rate": 1.934782608695652e-07,
|
|
"epoch": 4.258928571428571
|
|
},
|
|
{
|
|
"current_steps": 477,
|
|
"loss": 0.6314,
|
|
"learning_rate": 1.9130434782608695e-07,
|
|
"epoch": 4.267857142857143
|
|
},
|
|
{
|
|
"current_steps": 478,
|
|
"loss": 0.5759,
|
|
"learning_rate": 1.8913043478260867e-07,
|
|
"epoch": 4.276785714285714
|
|
},
|
|
{
|
|
"current_steps": 479,
|
|
"loss": 0.6925,
|
|
"learning_rate": 1.8695652173913043e-07,
|
|
"epoch": 4.285714285714286
|
|
},
|
|
{
|
|
"current_steps": 480,
|
|
"loss": 0.6237,
|
|
"learning_rate": 1.8478260869565215e-07,
|
|
"epoch": 4.294642857142857
|
|
},
|
|
{
|
|
"current_steps": 481,
|
|
"loss": 0.6666,
|
|
"learning_rate": 1.8260869565217391e-07,
|
|
"epoch": 4.303571428571429
|
|
},
|
|
{
|
|
"current_steps": 482,
|
|
"loss": 0.709,
|
|
"learning_rate": 1.8043478260869565e-07,
|
|
"epoch": 4.3125
|
|
},
|
|
{
|
|
"current_steps": 483,
|
|
"loss": 0.8078,
|
|
"learning_rate": 1.7826086956521737e-07,
|
|
"epoch": 4.321428571428571
|
|
},
|
|
{
|
|
"current_steps": 484,
|
|
"loss": 0.7355,
|
|
"learning_rate": 1.7608695652173914e-07,
|
|
"epoch": 4.330357142857143
|
|
},
|
|
{
|
|
"current_steps": 485,
|
|
"loss": 0.8901,
|
|
"learning_rate": 1.7391304347826085e-07,
|
|
"epoch": 4.339285714285714
|
|
},
|
|
{
|
|
"current_steps": 486,
|
|
"loss": 0.565,
|
|
"learning_rate": 1.7173913043478262e-07,
|
|
"epoch": 4.348214285714286
|
|
},
|
|
{
|
|
"current_steps": 487,
|
|
"loss": 0.6396,
|
|
"learning_rate": 1.6956521739130433e-07,
|
|
"epoch": 4.357142857142857
|
|
},
|
|
{
|
|
"current_steps": 488,
|
|
"loss": 0.531,
|
|
"learning_rate": 1.673913043478261e-07,
|
|
"epoch": 4.366071428571429
|
|
},
|
|
{
|
|
"current_steps": 489,
|
|
"loss": 0.5726,
|
|
"learning_rate": 1.652173913043478e-07,
|
|
"epoch": 4.375
|
|
},
|
|
{
|
|
"current_steps": 490,
|
|
"loss": 0.602,
|
|
"learning_rate": 1.6304347826086955e-07,
|
|
"epoch": 4.383928571428571
|
|
},
|
|
{
|
|
"current_steps": 491,
|
|
"loss": 0.7032,
|
|
"learning_rate": 1.608695652173913e-07,
|
|
"epoch": 4.392857142857143
|
|
},
|
|
{
|
|
"current_steps": 492,
|
|
"loss": 0.8984,
|
|
"learning_rate": 1.5869565217391303e-07,
|
|
"epoch": 4.401785714285714
|
|
},
|
|
{
|
|
"current_steps": 493,
|
|
"loss": 0.5913,
|
|
"learning_rate": 1.565217391304348e-07,
|
|
"epoch": 4.410714285714286
|
|
},
|
|
{
|
|
"current_steps": 494,
|
|
"loss": 0.6021,
|
|
"learning_rate": 1.543478260869565e-07,
|
|
"epoch": 4.419642857142857
|
|
},
|
|
{
|
|
"current_steps": 495,
|
|
"loss": 0.7554,
|
|
"learning_rate": 1.5217391304347825e-07,
|
|
"epoch": 4.428571428571429
|
|
},
|
|
{
|
|
"current_steps": 496,
|
|
"loss": 0.8683,
|
|
"learning_rate": 1.5e-07,
|
|
"epoch": 4.4375
|
|
},
|
|
{
|
|
"current_steps": 497,
|
|
"loss": 0.5465,
|
|
"learning_rate": 1.4782608695652173e-07,
|
|
"epoch": 4.446428571428571
|
|
},
|
|
{
|
|
"current_steps": 498,
|
|
"loss": 0.6903,
|
|
"learning_rate": 1.4565217391304347e-07,
|
|
"epoch": 4.455357142857143
|
|
},
|
|
{
|
|
"current_steps": 499,
|
|
"loss": 0.4821,
|
|
"learning_rate": 1.4347826086956521e-07,
|
|
"epoch": 4.464285714285714
|
|
},
|
|
{
|
|
"current_steps": 500,
|
|
"loss": 0.6731,
|
|
"learning_rate": 1.4130434782608693e-07,
|
|
"epoch": 4.473214285714286
|
|
},
|
|
{
|
|
"current_steps": 501,
|
|
"loss": 0.7423,
|
|
"learning_rate": 1.391304347826087e-07,
|
|
"epoch": 4.482142857142857
|
|
},
|
|
{
|
|
"current_steps": 502,
|
|
"loss": 0.6967,
|
|
"learning_rate": 1.369565217391304e-07,
|
|
"epoch": 4.491071428571429
|
|
},
|
|
{
|
|
"current_steps": 503,
|
|
"loss": 0.5918,
|
|
"learning_rate": 1.3478260869565218e-07,
|
|
"epoch": 4.5
|
|
},
|
|
{
|
|
"current_steps": 504,
|
|
"loss": 0.8028,
|
|
"learning_rate": 1.3260869565217392e-07,
|
|
"epoch": 4.508928571428571
|
|
},
|
|
{
|
|
"current_steps": 505,
|
|
"loss": 0.9578,
|
|
"learning_rate": 1.3043478260869563e-07,
|
|
"epoch": 4.517857142857143
|
|
},
|
|
{
|
|
"current_steps": 506,
|
|
"loss": 0.6187,
|
|
"learning_rate": 1.282608695652174e-07,
|
|
"epoch": 4.526785714285714
|
|
},
|
|
{
|
|
"current_steps": 507,
|
|
"loss": 0.6426,
|
|
"learning_rate": 1.260869565217391e-07,
|
|
"epoch": 4.535714285714286
|
|
},
|
|
{
|
|
"current_steps": 508,
|
|
"loss": 0.5835,
|
|
"learning_rate": 1.2391304347826088e-07,
|
|
"epoch": 4.544642857142857
|
|
},
|
|
{
|
|
"current_steps": 509,
|
|
"loss": 0.7218,
|
|
"learning_rate": 1.2173913043478262e-07,
|
|
"epoch": 4.553571428571429
|
|
},
|
|
{
|
|
"current_steps": 510,
|
|
"loss": 0.812,
|
|
"learning_rate": 1.1956521739130436e-07,
|
|
"epoch": 4.5625
|
|
},
|
|
{
|
|
"current_steps": 511,
|
|
"loss": 0.5526,
|
|
"learning_rate": 1.1739130434782609e-07,
|
|
"epoch": 4.571428571428571
|
|
},
|
|
{
|
|
"current_steps": 512,
|
|
"loss": 0.8554,
|
|
"learning_rate": 1.1521739130434783e-07,
|
|
"epoch": 4.580357142857143
|
|
},
|
|
{
|
|
"current_steps": 513,
|
|
"loss": 0.7209,
|
|
"learning_rate": 1.1304347826086955e-07,
|
|
"epoch": 4.589285714285714
|
|
},
|
|
{
|
|
"current_steps": 514,
|
|
"loss": 0.7154,
|
|
"learning_rate": 1.108695652173913e-07,
|
|
"epoch": 4.598214285714286
|
|
},
|
|
{
|
|
"current_steps": 515,
|
|
"loss": 0.7147,
|
|
"learning_rate": 1.0869565217391303e-07,
|
|
"epoch": 4.607142857142857
|
|
},
|
|
{
|
|
"current_steps": 516,
|
|
"loss": 0.6997,
|
|
"learning_rate": 1.0652173913043477e-07,
|
|
"epoch": 4.616071428571429
|
|
},
|
|
{
|
|
"current_steps": 517,
|
|
"loss": 0.6283,
|
|
"learning_rate": 1.0434782608695651e-07,
|
|
"epoch": 4.625
|
|
},
|
|
{
|
|
"current_steps": 518,
|
|
"loss": 0.6279,
|
|
"learning_rate": 1.0217391304347827e-07,
|
|
"epoch": 4.633928571428571
|
|
},
|
|
{
|
|
"current_steps": 519,
|
|
"loss": 0.8152,
|
|
"learning_rate": 1e-07,
|
|
"epoch": 4.642857142857143
|
|
},
|
|
{
|
|
"current_steps": 520,
|
|
"loss": 0.6155,
|
|
"learning_rate": 9.782608695652174e-08,
|
|
"epoch": 4.651785714285714
|
|
},
|
|
{
|
|
"current_steps": 521,
|
|
"loss": 0.4727,
|
|
"learning_rate": 9.565217391304348e-08,
|
|
"epoch": 4.660714285714286
|
|
},
|
|
{
|
|
"current_steps": 522,
|
|
"loss": 0.7457,
|
|
"learning_rate": 9.347826086956522e-08,
|
|
"epoch": 4.669642857142857
|
|
},
|
|
{
|
|
"current_steps": 523,
|
|
"loss": 0.9712,
|
|
"learning_rate": 9.130434782608696e-08,
|
|
"epoch": 4.678571428571429
|
|
},
|
|
{
|
|
"current_steps": 524,
|
|
"loss": 0.7759,
|
|
"learning_rate": 8.913043478260868e-08,
|
|
"epoch": 4.6875
|
|
},
|
|
{
|
|
"current_steps": 525,
|
|
"loss": 0.6597,
|
|
"learning_rate": 8.695652173913042e-08,
|
|
"epoch": 4.696428571428571
|
|
},
|
|
{
|
|
"current_steps": 526,
|
|
"loss": 0.6258,
|
|
"learning_rate": 8.478260869565216e-08,
|
|
"epoch": 4.705357142857143
|
|
},
|
|
{
|
|
"current_steps": 527,
|
|
"loss": 0.6443,
|
|
"learning_rate": 8.26086956521739e-08,
|
|
"epoch": 4.714285714285714
|
|
},
|
|
{
|
|
"current_steps": 528,
|
|
"loss": 0.5547,
|
|
"learning_rate": 8.043478260869565e-08,
|
|
"epoch": 4.723214285714286
|
|
},
|
|
{
|
|
"current_steps": 529,
|
|
"loss": 0.7149,
|
|
"learning_rate": 7.82608695652174e-08,
|
|
"epoch": 4.732142857142857
|
|
},
|
|
{
|
|
"current_steps": 530,
|
|
"loss": 0.6138,
|
|
"learning_rate": 7.608695652173913e-08,
|
|
"epoch": 4.741071428571429
|
|
},
|
|
{
|
|
"current_steps": 531,
|
|
"loss": 0.8032,
|
|
"learning_rate": 7.391304347826087e-08,
|
|
"epoch": 4.75
|
|
},
|
|
{
|
|
"current_steps": 532,
|
|
"loss": 0.7141,
|
|
"learning_rate": 7.173913043478261e-08,
|
|
"epoch": 4.758928571428571
|
|
},
|
|
{
|
|
"current_steps": 533,
|
|
"loss": 0.724,
|
|
"learning_rate": 6.956521739130435e-08,
|
|
"epoch": 4.767857142857143
|
|
},
|
|
{
|
|
"current_steps": 534,
|
|
"loss": 0.7707,
|
|
"learning_rate": 6.739130434782609e-08,
|
|
"epoch": 4.776785714285714
|
|
},
|
|
{
|
|
"current_steps": 535,
|
|
"loss": 0.6754,
|
|
"learning_rate": 6.521739130434782e-08,
|
|
"epoch": 4.785714285714286
|
|
},
|
|
{
|
|
"current_steps": 536,
|
|
"loss": 0.5861,
|
|
"learning_rate": 6.304347826086956e-08,
|
|
"epoch": 4.794642857142857
|
|
},
|
|
{
|
|
"current_steps": 537,
|
|
"loss": 0.8395,
|
|
"learning_rate": 6.086956521739131e-08,
|
|
"epoch": 4.803571428571429
|
|
},
|
|
{
|
|
"current_steps": 538,
|
|
"loss": 0.7642,
|
|
"learning_rate": 5.869565217391304e-08,
|
|
"epoch": 4.8125
|
|
},
|
|
{
|
|
"current_steps": 539,
|
|
"loss": 0.735,
|
|
"learning_rate": 5.6521739130434777e-08,
|
|
"epoch": 4.821428571428571
|
|
},
|
|
{
|
|
"current_steps": 540,
|
|
"loss": 0.6153,
|
|
"learning_rate": 5.434782608695652e-08,
|
|
"epoch": 4.830357142857143
|
|
},
|
|
{
|
|
"current_steps": 541,
|
|
"loss": 0.6299,
|
|
"learning_rate": 5.217391304347826e-08,
|
|
"epoch": 4.839285714285714
|
|
},
|
|
{
|
|
"current_steps": 542,
|
|
"loss": 1.078,
|
|
"learning_rate": 5e-08,
|
|
"epoch": 4.848214285714286
|
|
},
|
|
{
|
|
"current_steps": 543,
|
|
"loss": 0.7314,
|
|
"learning_rate": 4.782608695652174e-08,
|
|
"epoch": 4.857142857142857
|
|
},
|
|
{
|
|
"current_steps": 544,
|
|
"loss": 0.8515,
|
|
"learning_rate": 4.565217391304348e-08,
|
|
"epoch": 4.866071428571429
|
|
},
|
|
{
|
|
"current_steps": 545,
|
|
"loss": 0.5401,
|
|
"learning_rate": 4.347826086956521e-08,
|
|
"epoch": 4.875
|
|
},
|
|
{
|
|
"current_steps": 546,
|
|
"loss": 0.7315,
|
|
"learning_rate": 4.130434782608695e-08,
|
|
"epoch": 4.883928571428571
|
|
},
|
|
{
|
|
"current_steps": 547,
|
|
"loss": 0.6113,
|
|
"learning_rate": 3.91304347826087e-08,
|
|
"epoch": 4.892857142857143
|
|
},
|
|
{
|
|
"current_steps": 548,
|
|
"loss": 0.6239,
|
|
"learning_rate": 3.6956521739130433e-08,
|
|
"epoch": 4.901785714285714
|
|
},
|
|
{
|
|
"current_steps": 549,
|
|
"loss": 0.7292,
|
|
"learning_rate": 3.4782608695652174e-08,
|
|
"epoch": 4.910714285714286
|
|
},
|
|
{
|
|
"current_steps": 550,
|
|
"loss": 0.5297,
|
|
"learning_rate": 3.260869565217391e-08,
|
|
"epoch": 4.919642857142857
|
|
},
|
|
{
|
|
"current_steps": 551,
|
|
"loss": 0.6269,
|
|
"learning_rate": 3.0434782608695655e-08,
|
|
"epoch": 4.928571428571429
|
|
},
|
|
{
|
|
"current_steps": 552,
|
|
"loss": 0.6724,
|
|
"learning_rate": 2.8260869565217388e-08,
|
|
"epoch": 4.9375
|
|
},
|
|
{
|
|
"current_steps": 553,
|
|
"loss": 0.5109,
|
|
"learning_rate": 2.608695652173913e-08,
|
|
"epoch": 4.946428571428571
|
|
},
|
|
{
|
|
"current_steps": 554,
|
|
"loss": 0.9446,
|
|
"learning_rate": 2.391304347826087e-08,
|
|
"epoch": 4.955357142857143
|
|
},
|
|
{
|
|
"current_steps": 555,
|
|
"loss": 0.6897,
|
|
"learning_rate": 2.1739130434782606e-08,
|
|
"epoch": 4.964285714285714
|
|
},
|
|
{
|
|
"current_steps": 556,
|
|
"loss": 0.5511,
|
|
"learning_rate": 1.956521739130435e-08,
|
|
"epoch": 4.973214285714286
|
|
},
|
|
{
|
|
"current_steps": 557,
|
|
"loss": 0.7246,
|
|
"learning_rate": 1.7391304347826087e-08,
|
|
"epoch": 4.982142857142857
|
|
},
|
|
{
|
|
"current_steps": 558,
|
|
"loss": 0.6332,
|
|
"learning_rate": 1.5217391304347827e-08,
|
|
"epoch": 4.991071428571429
|
|
},
|
|
{
|
|
"current_steps": 559,
|
|
"loss": 1.0499,
|
|
"learning_rate": 1.3043478260869564e-08,
|
|
"epoch": 5.0
|
|
},
|
|
{
|
|
"current_steps": 559,
|
|
"loss": 1.0499,
|
|
"learning_rate": 1.3043478260869564e-08,
|
|
"epoch": 5.0
|
|
}
|
|
] |