|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 17.142857142857142, |
|
"eval_steps": 500, |
|
"global_step": 600, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.000999404761904762, |
|
"loss": 3.0944, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0009988095238095238, |
|
"loss": 2.7802, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0009982142857142857, |
|
"loss": 2.7798, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0009976190476190477, |
|
"loss": 2.6729, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0009970238095238096, |
|
"loss": 2.7544, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0009964285714285715, |
|
"loss": 2.7115, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0009958333333333334, |
|
"loss": 2.7491, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0009952380952380953, |
|
"loss": 2.681, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0009946428571428571, |
|
"loss": 2.7396, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.000994047619047619, |
|
"loss": 2.6911, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0009934523809523809, |
|
"loss": 2.6725, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.000992857142857143, |
|
"loss": 2.6951, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0009922619047619049, |
|
"loss": 2.6741, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.0009916666666666667, |
|
"loss": 2.6623, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0009910714285714286, |
|
"loss": 2.6268, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.0009904761904761905, |
|
"loss": 2.6727, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.0009898809523809524, |
|
"loss": 2.6174, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.0009892857142857142, |
|
"loss": 2.6544, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0009886904761904763, |
|
"loss": 2.5404, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 0.0009880952380952382, |
|
"loss": 2.6286, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 0.0009875, |
|
"loss": 2.6824, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 0.000986904761904762, |
|
"loss": 2.6012, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 0.0009863095238095239, |
|
"loss": 2.6446, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 0.0009857142857142857, |
|
"loss": 2.6437, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 0.0009851190476190476, |
|
"loss": 2.6596, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 0.0009845238095238097, |
|
"loss": 2.6554, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 0.0009839285714285714, |
|
"loss": 2.6064, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 0.0009833333333333332, |
|
"loss": 2.5994, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 0.0009827380952380951, |
|
"loss": 2.545, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 0.0009821428571428572, |
|
"loss": 2.5704, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 0.000981547619047619, |
|
"loss": 2.6461, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 0.000980952380952381, |
|
"loss": 2.631, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 0.0009803571428571428, |
|
"loss": 2.6678, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 0.0009797619047619047, |
|
"loss": 2.5964, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.0009791666666666666, |
|
"loss": 2.6566, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 0.0009785714285714285, |
|
"loss": 2.4962, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 0.0009779761904761906, |
|
"loss": 2.4815, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 0.0009773809523809524, |
|
"loss": 2.4172, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 0.0009767857142857143, |
|
"loss": 2.4641, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 0.0009761904761904762, |
|
"loss": 2.4875, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 0.0009755952380952381, |
|
"loss": 2.4486, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 0.000975, |
|
"loss": 2.4463, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 0.0009744047619047619, |
|
"loss": 2.5063, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 0.0009738095238095238, |
|
"loss": 2.4506, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 0.0009732142857142857, |
|
"loss": 2.3862, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 0.0009726190476190476, |
|
"loss": 2.4925, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 0.0009720238095238096, |
|
"loss": 2.5, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 0.0009714285714285714, |
|
"loss": 2.4658, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 0.0009708333333333333, |
|
"loss": 2.3983, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 0.0009702380952380953, |
|
"loss": 2.369, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 0.0009696428571428572, |
|
"loss": 2.4849, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 0.0009690476190476191, |
|
"loss": 2.4106, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 0.0009684523809523809, |
|
"loss": 2.4363, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 0.0009678571428571429, |
|
"loss": 2.4935, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 0.0009672619047619048, |
|
"loss": 2.479, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 0.0009666666666666667, |
|
"loss": 2.4299, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 0.0009660714285714285, |
|
"loss": 2.3964, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 0.0009654761904761905, |
|
"loss": 2.4865, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 0.0009648809523809524, |
|
"loss": 2.3831, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 0.0009642857142857143, |
|
"loss": 2.4304, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 0.0009636904761904763, |
|
"loss": 2.4273, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 0.0009630952380952382, |
|
"loss": 2.4427, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 0.0009625, |
|
"loss": 2.4191, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 0.0009619047619047619, |
|
"loss": 2.3902, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 0.0009613095238095239, |
|
"loss": 2.4699, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 0.0009607142857142858, |
|
"loss": 2.3978, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 0.0009601190476190476, |
|
"loss": 2.4128, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 0.0009595238095238095, |
|
"loss": 2.4585, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 0.0009589285714285715, |
|
"loss": 2.4476, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 0.0009583333333333334, |
|
"loss": 2.4231, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 0.0009577380952380953, |
|
"loss": 2.2655, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 0.0009571428571428573, |
|
"loss": 2.208, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 0.0009565476190476191, |
|
"loss": 2.1758, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 0.000955952380952381, |
|
"loss": 2.3113, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 0.0009553571428571429, |
|
"loss": 2.1739, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 0.0009547619047619049, |
|
"loss": 2.2234, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 0.0009541666666666667, |
|
"loss": 2.21, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 0.0009535714285714286, |
|
"loss": 2.2316, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 0.0009529761904761904, |
|
"loss": 2.2044, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 0.0009523809523809524, |
|
"loss": 2.1784, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 0.0009517857142857143, |
|
"loss": 2.2489, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 0.0009511904761904761, |
|
"loss": 2.2003, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 0.0009505952380952381, |
|
"loss": 2.169, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 0.00095, |
|
"loss": 2.2303, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 0.0009494047619047619, |
|
"loss": 2.1744, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 0.0009488095238095238, |
|
"loss": 2.1904, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 0.0009482142857142857, |
|
"loss": 2.222, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 0.0009476190476190476, |
|
"loss": 2.2467, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 0.0009470238095238095, |
|
"loss": 2.2241, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 0.0009464285714285714, |
|
"loss": 2.2339, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 0.0009458333333333334, |
|
"loss": 2.2339, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 0.0009452380952380952, |
|
"loss": 2.1653, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 0.0009446428571428571, |
|
"loss": 2.2954, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 0.0009440476190476191, |
|
"loss": 2.2859, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 0.000943452380952381, |
|
"loss": 2.2107, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 0.0009428571428571429, |
|
"loss": 2.2142, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 0.0009422619047619047, |
|
"loss": 2.2433, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 0.0009416666666666667, |
|
"loss": 2.2417, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 0.0009410714285714286, |
|
"loss": 2.2045, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 0.0009404761904761905, |
|
"loss": 2.293, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 0.0009398809523809523, |
|
"loss": 2.2051, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 0.0009392857142857143, |
|
"loss": 2.2889, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 0.0009386904761904762, |
|
"loss": 2.2, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 0.0009380952380952381, |
|
"loss": 2.2298, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 0.0009375, |
|
"loss": 2.2722, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 0.000936904761904762, |
|
"loss": 2.0032, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 0.0009363095238095238, |
|
"loss": 1.9269, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 0.0009357142857142857, |
|
"loss": 1.916, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 0.0009351190476190477, |
|
"loss": 1.9165, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 0.0009345238095238096, |
|
"loss": 1.9296, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 0.0009339285714285714, |
|
"loss": 1.954, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 0.0009333333333333333, |
|
"loss": 1.9457, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 0.0009327380952380953, |
|
"loss": 1.9135, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 0.0009321428571428572, |
|
"loss": 1.9383, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 0.0009315476190476191, |
|
"loss": 1.9057, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 0.0009309523809523809, |
|
"loss": 1.9541, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 0.0009303571428571429, |
|
"loss": 1.8827, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 0.0009297619047619048, |
|
"loss": 1.899, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 0.0009291666666666667, |
|
"loss": 1.9095, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 0.0009285714285714287, |
|
"loss": 1.936, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 0.0009279761904761905, |
|
"loss": 1.9224, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"learning_rate": 0.0009273809523809524, |
|
"loss": 1.9547, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 0.0009267857142857143, |
|
"loss": 1.9323, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 0.0009261904761904763, |
|
"loss": 1.9938, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 0.0009255952380952382, |
|
"loss": 1.8674, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"learning_rate": 0.000925, |
|
"loss": 1.95, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"learning_rate": 0.0009244047619047619, |
|
"loss": 1.9374, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"learning_rate": 0.0009238095238095239, |
|
"loss": 1.9456, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"learning_rate": 0.0009232142857142858, |
|
"loss": 1.9791, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 0.0009226190476190477, |
|
"loss": 2.0047, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"learning_rate": 0.0009220238095238096, |
|
"loss": 1.9971, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 0.0009214285714285714, |
|
"loss": 2.0267, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"learning_rate": 0.0009208333333333333, |
|
"loss": 1.9374, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"learning_rate": 0.0009202380952380952, |
|
"loss": 1.9793, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 0.0009196428571428572, |
|
"loss": 2.0483, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 0.000919047619047619, |
|
"loss": 2.0003, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 0.0009184523809523809, |
|
"loss": 2.0185, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"learning_rate": 0.0009178571428571428, |
|
"loss": 2.0517, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"learning_rate": 0.0009172619047619048, |
|
"loss": 1.9824, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 0.0009166666666666666, |
|
"loss": 2.0383, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 0.0009160714285714285, |
|
"loss": 1.6818, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 4.06, |
|
"learning_rate": 0.0009154761904761905, |
|
"loss": 1.6208, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"learning_rate": 0.0009148809523809524, |
|
"loss": 1.6843, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 4.11, |
|
"learning_rate": 0.0009142857142857143, |
|
"loss": 1.5885, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"learning_rate": 0.0009136904761904761, |
|
"loss": 1.5799, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 4.17, |
|
"learning_rate": 0.0009130952380952381, |
|
"loss": 1.6334, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 4.2, |
|
"learning_rate": 0.0009125, |
|
"loss": 1.6297, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 4.23, |
|
"learning_rate": 0.0009119047619047619, |
|
"loss": 1.5929, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 4.26, |
|
"learning_rate": 0.0009113095238095238, |
|
"loss": 1.6621, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 4.29, |
|
"learning_rate": 0.0009107142857142857, |
|
"loss": 1.626, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 4.31, |
|
"learning_rate": 0.0009101190476190476, |
|
"loss": 1.6138, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 4.34, |
|
"learning_rate": 0.0009095238095238095, |
|
"loss": 1.6465, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"learning_rate": 0.0009089285714285715, |
|
"loss": 1.6622, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 4.4, |
|
"learning_rate": 0.0009083333333333334, |
|
"loss": 1.6662, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 4.43, |
|
"learning_rate": 0.0009077380952380952, |
|
"loss": 1.6348, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 4.46, |
|
"learning_rate": 0.0009071428571428571, |
|
"loss": 1.6196, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 4.49, |
|
"learning_rate": 0.0009065476190476191, |
|
"loss": 1.6766, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 4.51, |
|
"learning_rate": 0.000905952380952381, |
|
"loss": 1.7069, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 4.54, |
|
"learning_rate": 0.0009053571428571429, |
|
"loss": 1.6848, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 4.57, |
|
"learning_rate": 0.0009047619047619047, |
|
"loss": 1.6884, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 4.6, |
|
"learning_rate": 0.0009041666666666667, |
|
"loss": 1.6721, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 4.63, |
|
"learning_rate": 0.0009035714285714286, |
|
"loss": 1.7116, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 4.66, |
|
"learning_rate": 0.0009029761904761905, |
|
"loss": 1.693, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 4.69, |
|
"learning_rate": 0.0009023809523809525, |
|
"loss": 1.6826, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 4.71, |
|
"learning_rate": 0.0009017857142857143, |
|
"loss": 1.7061, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 4.74, |
|
"learning_rate": 0.0009011904761904762, |
|
"loss": 1.6964, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 4.77, |
|
"learning_rate": 0.0009005952380952381, |
|
"loss": 1.7903, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 4.8, |
|
"learning_rate": 0.0009000000000000001, |
|
"loss": 1.6829, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 4.83, |
|
"learning_rate": 0.000899404761904762, |
|
"loss": 1.7047, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"learning_rate": 0.0008988095238095238, |
|
"loss": 1.7671, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 0.0008982142857142857, |
|
"loss": 1.7184, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 4.91, |
|
"learning_rate": 0.0008976190476190477, |
|
"loss": 1.8213, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 4.94, |
|
"learning_rate": 0.0008970238095238096, |
|
"loss": 1.7688, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 4.97, |
|
"learning_rate": 0.0008964285714285715, |
|
"loss": 1.7522, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 0.0008958333333333334, |
|
"loss": 1.7862, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 0.0008952380952380953, |
|
"loss": 1.3871, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 5.06, |
|
"learning_rate": 0.0008946428571428572, |
|
"loss": 1.3491, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 5.09, |
|
"learning_rate": 0.0008940476190476191, |
|
"loss": 1.3399, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 5.11, |
|
"learning_rate": 0.0008934523809523811, |
|
"loss": 1.3569, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 5.14, |
|
"learning_rate": 0.0008928571428571429, |
|
"loss": 1.3734, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 5.17, |
|
"learning_rate": 0.0008922619047619048, |
|
"loss": 1.3151, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 5.2, |
|
"learning_rate": 0.0008916666666666667, |
|
"loss": 1.3243, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 5.23, |
|
"learning_rate": 0.0008910714285714287, |
|
"loss": 1.342, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 5.26, |
|
"learning_rate": 0.0008904761904761904, |
|
"loss": 1.3664, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 5.29, |
|
"learning_rate": 0.0008898809523809523, |
|
"loss": 1.3493, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 5.31, |
|
"learning_rate": 0.0008892857142857142, |
|
"loss": 1.32, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 5.34, |
|
"learning_rate": 0.0008886904761904762, |
|
"loss": 1.3978, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 5.37, |
|
"learning_rate": 0.0008880952380952381, |
|
"loss": 1.3762, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 5.4, |
|
"learning_rate": 0.0008874999999999999, |
|
"loss": 1.4172, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 5.43, |
|
"learning_rate": 0.0008869047619047619, |
|
"loss": 1.3817, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 5.46, |
|
"learning_rate": 0.0008863095238095238, |
|
"loss": 1.3779, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 5.49, |
|
"learning_rate": 0.0008857142857142857, |
|
"loss": 1.378, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 5.51, |
|
"learning_rate": 0.0008851190476190476, |
|
"loss": 1.4245, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 5.54, |
|
"learning_rate": 0.0008845238095238095, |
|
"loss": 1.4425, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 5.57, |
|
"learning_rate": 0.0008839285714285714, |
|
"loss": 1.4324, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 5.6, |
|
"learning_rate": 0.0008833333333333333, |
|
"loss": 1.4264, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 5.63, |
|
"learning_rate": 0.0008827380952380952, |
|
"loss": 1.4395, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 5.66, |
|
"learning_rate": 0.0008821428571428572, |
|
"loss": 1.4549, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"learning_rate": 0.000881547619047619, |
|
"loss": 1.501, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 5.71, |
|
"learning_rate": 0.0008809523809523809, |
|
"loss": 1.432, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 5.74, |
|
"learning_rate": 0.0008803571428571429, |
|
"loss": 1.4922, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 5.77, |
|
"learning_rate": 0.0008797619047619048, |
|
"loss": 1.4622, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 5.8, |
|
"learning_rate": 0.0008791666666666667, |
|
"loss": 1.4794, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 5.83, |
|
"learning_rate": 0.0008785714285714285, |
|
"loss": 1.4938, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 5.86, |
|
"learning_rate": 0.0008779761904761905, |
|
"loss": 1.4792, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 5.89, |
|
"learning_rate": 0.0008773809523809524, |
|
"loss": 1.5192, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 5.91, |
|
"learning_rate": 0.0008767857142857143, |
|
"loss": 1.5055, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 5.94, |
|
"learning_rate": 0.0008761904761904762, |
|
"loss": 1.5484, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 5.97, |
|
"learning_rate": 0.0008755952380952381, |
|
"loss": 1.5096, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 0.000875, |
|
"loss": 1.5298, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 0.0008744047619047619, |
|
"loss": 1.1704, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 6.06, |
|
"learning_rate": 0.0008738095238095239, |
|
"loss": 1.1261, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 6.09, |
|
"learning_rate": 0.0008732142857142858, |
|
"loss": 1.1144, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 6.11, |
|
"learning_rate": 0.0008726190476190476, |
|
"loss": 1.0984, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 6.14, |
|
"learning_rate": 0.0008720238095238095, |
|
"loss": 1.0704, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 6.17, |
|
"learning_rate": 0.0008714285714285715, |
|
"loss": 1.0655, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 6.2, |
|
"learning_rate": 0.0008708333333333334, |
|
"loss": 1.09, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 6.23, |
|
"learning_rate": 0.0008702380952380953, |
|
"loss": 1.0619, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 6.26, |
|
"learning_rate": 0.0008696428571428571, |
|
"loss": 1.1633, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 6.29, |
|
"learning_rate": 0.0008690476190476191, |
|
"loss": 1.1022, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 6.31, |
|
"learning_rate": 0.000868452380952381, |
|
"loss": 1.1057, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 6.34, |
|
"learning_rate": 0.0008678571428571429, |
|
"loss": 1.1279, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 6.37, |
|
"learning_rate": 0.0008672619047619049, |
|
"loss": 1.0915, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 6.4, |
|
"learning_rate": 0.0008666666666666667, |
|
"loss": 1.1731, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 6.43, |
|
"learning_rate": 0.0008660714285714286, |
|
"loss": 1.1352, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 6.46, |
|
"learning_rate": 0.0008654761904761905, |
|
"loss": 1.1632, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 6.49, |
|
"learning_rate": 0.0008648809523809525, |
|
"loss": 1.1691, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 6.51, |
|
"learning_rate": 0.0008642857142857144, |
|
"loss": 1.181, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 6.54, |
|
"learning_rate": 0.0008636904761904762, |
|
"loss": 1.1635, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 6.57, |
|
"learning_rate": 0.0008630952380952381, |
|
"loss": 1.1802, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 6.6, |
|
"learning_rate": 0.0008625000000000001, |
|
"loss": 1.2111, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 6.63, |
|
"learning_rate": 0.000861904761904762, |
|
"loss": 1.2503, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 6.66, |
|
"learning_rate": 0.0008613095238095238, |
|
"loss": 1.2305, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 6.69, |
|
"learning_rate": 0.0008607142857142858, |
|
"loss": 1.2446, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 6.71, |
|
"learning_rate": 0.0008601190476190477, |
|
"loss": 1.263, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 6.74, |
|
"learning_rate": 0.0008595238095238096, |
|
"loss": 1.2407, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 6.77, |
|
"learning_rate": 0.0008589285714285714, |
|
"loss": 1.303, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 6.8, |
|
"learning_rate": 0.0008583333333333333, |
|
"loss": 1.2309, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 6.83, |
|
"learning_rate": 0.0008577380952380952, |
|
"loss": 1.2669, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 6.86, |
|
"learning_rate": 0.0008571428571428571, |
|
"loss": 1.226, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 6.89, |
|
"learning_rate": 0.000856547619047619, |
|
"loss": 1.2862, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 6.91, |
|
"learning_rate": 0.000855952380952381, |
|
"loss": 1.2472, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 6.94, |
|
"learning_rate": 0.0008553571428571428, |
|
"loss": 1.2928, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 6.97, |
|
"learning_rate": 0.0008547619047619047, |
|
"loss": 1.2427, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 0.0008541666666666666, |
|
"loss": 1.3195, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 0.0008535714285714286, |
|
"loss": 0.8949, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 7.06, |
|
"learning_rate": 0.0008529761904761905, |
|
"loss": 0.8907, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 7.09, |
|
"learning_rate": 0.0008523809523809523, |
|
"loss": 0.8813, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 7.11, |
|
"learning_rate": 0.0008517857142857143, |
|
"loss": 0.8702, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 7.14, |
|
"learning_rate": 0.0008511904761904762, |
|
"loss": 0.9105, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 7.17, |
|
"learning_rate": 0.0008505952380952381, |
|
"loss": 0.9096, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 7.2, |
|
"learning_rate": 0.00085, |
|
"loss": 0.9121, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 7.23, |
|
"learning_rate": 0.0008494047619047619, |
|
"loss": 0.9063, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 7.26, |
|
"learning_rate": 0.0008488095238095238, |
|
"loss": 0.8976, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 7.29, |
|
"learning_rate": 0.0008482142857142857, |
|
"loss": 0.9283, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 7.31, |
|
"learning_rate": 0.0008476190476190476, |
|
"loss": 0.9409, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 7.34, |
|
"learning_rate": 0.0008470238095238096, |
|
"loss": 0.9311, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 7.37, |
|
"learning_rate": 0.0008464285714285714, |
|
"loss": 0.926, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 7.4, |
|
"learning_rate": 0.0008458333333333333, |
|
"loss": 0.9704, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 7.43, |
|
"learning_rate": 0.0008452380952380953, |
|
"loss": 0.9515, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 7.46, |
|
"learning_rate": 0.0008446428571428572, |
|
"loss": 0.9069, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 7.49, |
|
"learning_rate": 0.000844047619047619, |
|
"loss": 0.9359, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 7.51, |
|
"learning_rate": 0.0008434523809523809, |
|
"loss": 0.9482, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 7.54, |
|
"learning_rate": 0.0008428571428571429, |
|
"loss": 0.9717, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 7.57, |
|
"learning_rate": 0.0008422619047619048, |
|
"loss": 0.9869, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 7.6, |
|
"learning_rate": 0.0008416666666666667, |
|
"loss": 0.9728, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 7.63, |
|
"learning_rate": 0.0008410714285714285, |
|
"loss": 0.9516, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 7.66, |
|
"learning_rate": 0.0008404761904761905, |
|
"loss": 0.9838, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 7.69, |
|
"learning_rate": 0.0008398809523809524, |
|
"loss": 1.0044, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 7.71, |
|
"learning_rate": 0.0008392857142857143, |
|
"loss": 1.0153, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 7.74, |
|
"learning_rate": 0.0008386904761904763, |
|
"loss": 1.0382, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 7.77, |
|
"learning_rate": 0.0008380952380952382, |
|
"loss": 1.0109, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 7.8, |
|
"learning_rate": 0.0008375, |
|
"loss": 0.9989, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 7.83, |
|
"learning_rate": 0.0008369047619047619, |
|
"loss": 1.0631, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 7.86, |
|
"learning_rate": 0.0008363095238095239, |
|
"loss": 1.0546, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 7.89, |
|
"learning_rate": 0.0008357142857142858, |
|
"loss": 1.0827, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 7.91, |
|
"learning_rate": 0.0008351190476190476, |
|
"loss": 1.087, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 7.94, |
|
"learning_rate": 0.0008345238095238095, |
|
"loss": 1.041, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 7.97, |
|
"learning_rate": 0.0008339285714285715, |
|
"loss": 1.0633, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 0.0008333333333333334, |
|
"loss": 1.0709, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 0.0008327380952380953, |
|
"loss": 0.7273, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 8.06, |
|
"learning_rate": 0.0008321428571428573, |
|
"loss": 0.726, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 8.09, |
|
"learning_rate": 0.0008315476190476191, |
|
"loss": 0.6943, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 8.11, |
|
"learning_rate": 0.000830952380952381, |
|
"loss": 0.7127, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 8.14, |
|
"learning_rate": 0.0008303571428571429, |
|
"loss": 0.6915, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 8.17, |
|
"learning_rate": 0.0008297619047619049, |
|
"loss": 0.7138, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 8.2, |
|
"learning_rate": 0.0008291666666666667, |
|
"loss": 0.7356, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 8.23, |
|
"learning_rate": 0.0008285714285714286, |
|
"loss": 0.678, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 8.26, |
|
"learning_rate": 0.0008279761904761904, |
|
"loss": 0.7375, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 8.29, |
|
"learning_rate": 0.0008273809523809524, |
|
"loss": 0.7284, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 8.31, |
|
"learning_rate": 0.0008267857142857143, |
|
"loss": 0.7304, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 8.34, |
|
"learning_rate": 0.0008261904761904761, |
|
"loss": 0.7633, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 8.37, |
|
"learning_rate": 0.0008255952380952381, |
|
"loss": 0.7416, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 8.4, |
|
"learning_rate": 0.000825, |
|
"loss": 0.7895, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 8.43, |
|
"learning_rate": 0.0008244047619047619, |
|
"loss": 0.8037, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 8.46, |
|
"learning_rate": 0.0008238095238095238, |
|
"loss": 0.7736, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 8.49, |
|
"learning_rate": 0.0008232142857142857, |
|
"loss": 0.778, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 8.51, |
|
"learning_rate": 0.0008226190476190476, |
|
"loss": 0.7644, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 8.54, |
|
"learning_rate": 0.0008220238095238095, |
|
"loss": 0.7942, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 8.57, |
|
"learning_rate": 0.0008214285714285714, |
|
"loss": 0.7715, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 8.6, |
|
"learning_rate": 0.0008208333333333334, |
|
"loss": 0.8288, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 8.63, |
|
"learning_rate": 0.0008202380952380952, |
|
"loss": 0.8263, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 8.66, |
|
"learning_rate": 0.0008196428571428571, |
|
"loss": 0.7923, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 8.69, |
|
"learning_rate": 0.0008190476190476191, |
|
"loss": 0.8063, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 8.71, |
|
"learning_rate": 0.000818452380952381, |
|
"loss": 0.8016, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 8.74, |
|
"learning_rate": 0.0008178571428571428, |
|
"loss": 0.8467, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 8.77, |
|
"learning_rate": 0.0008172619047619047, |
|
"loss": 0.8353, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 8.8, |
|
"learning_rate": 0.0008166666666666667, |
|
"loss": 0.8272, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 8.83, |
|
"learning_rate": 0.0008160714285714286, |
|
"loss": 0.8852, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 8.86, |
|
"learning_rate": 0.0008154761904761905, |
|
"loss": 0.8541, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 8.89, |
|
"learning_rate": 0.0008148809523809523, |
|
"loss": 0.8236, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 8.91, |
|
"learning_rate": 0.0008142857142857143, |
|
"loss": 0.8609, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 8.94, |
|
"learning_rate": 0.0008136904761904762, |
|
"loss": 0.8802, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 8.97, |
|
"learning_rate": 0.0008130952380952381, |
|
"loss": 0.8615, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 0.0008125000000000001, |
|
"loss": 0.8514, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 0.000811904761904762, |
|
"loss": 0.5529, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 9.06, |
|
"learning_rate": 0.0008113095238095238, |
|
"loss": 0.5736, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 9.09, |
|
"learning_rate": 0.0008107142857142857, |
|
"loss": 0.5647, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 9.11, |
|
"learning_rate": 0.0008101190476190477, |
|
"loss": 0.5677, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 9.14, |
|
"learning_rate": 0.0008095238095238096, |
|
"loss": 0.5991, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 9.17, |
|
"learning_rate": 0.0008089285714285714, |
|
"loss": 0.5666, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 9.2, |
|
"learning_rate": 0.0008083333333333333, |
|
"loss": 0.5902, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 9.23, |
|
"learning_rate": 0.0008077380952380953, |
|
"loss": 0.5961, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 9.26, |
|
"learning_rate": 0.0008071428571428572, |
|
"loss": 0.5684, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 9.29, |
|
"learning_rate": 0.0008065476190476191, |
|
"loss": 0.5976, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 9.31, |
|
"learning_rate": 0.0008059523809523809, |
|
"loss": 0.6033, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 9.34, |
|
"learning_rate": 0.0008053571428571429, |
|
"loss": 0.5877, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 9.37, |
|
"learning_rate": 0.0008047619047619048, |
|
"loss": 0.5943, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 9.4, |
|
"learning_rate": 0.0008041666666666667, |
|
"loss": 0.6176, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 9.43, |
|
"learning_rate": 0.0008035714285714287, |
|
"loss": 0.6143, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 9.46, |
|
"learning_rate": 0.0008029761904761905, |
|
"loss": 0.597, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 9.49, |
|
"learning_rate": 0.0008023809523809524, |
|
"loss": 0.604, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 9.51, |
|
"learning_rate": 0.0008017857142857143, |
|
"loss": 0.6036, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 9.54, |
|
"learning_rate": 0.0008011904761904763, |
|
"loss": 0.6243, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 9.57, |
|
"learning_rate": 0.0008005952380952382, |
|
"loss": 0.6301, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 9.6, |
|
"learning_rate": 0.0008, |
|
"loss": 0.6271, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 9.63, |
|
"learning_rate": 0.0007994047619047619, |
|
"loss": 0.6246, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 9.66, |
|
"learning_rate": 0.0007988095238095239, |
|
"loss": 0.6597, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 9.69, |
|
"learning_rate": 0.0007982142857142858, |
|
"loss": 0.6517, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 9.71, |
|
"learning_rate": 0.0007976190476190477, |
|
"loss": 0.6645, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 9.74, |
|
"learning_rate": 0.0007970238095238096, |
|
"loss": 0.6542, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 9.77, |
|
"learning_rate": 0.0007964285714285714, |
|
"loss": 0.6496, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 9.8, |
|
"learning_rate": 0.0007958333333333333, |
|
"loss": 0.6309, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 9.83, |
|
"learning_rate": 0.0007952380952380952, |
|
"loss": 0.6668, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 9.86, |
|
"learning_rate": 0.0007946428571428572, |
|
"loss": 0.6841, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 9.89, |
|
"learning_rate": 0.000794047619047619, |
|
"loss": 0.6958, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 9.91, |
|
"learning_rate": 0.0007934523809523809, |
|
"loss": 0.6592, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 9.94, |
|
"learning_rate": 0.0007928571428571428, |
|
"loss": 0.6968, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 9.97, |
|
"learning_rate": 0.0007922619047619048, |
|
"loss": 0.6916, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 0.0007916666666666666, |
|
"loss": 0.7155, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 0.0007910714285714285, |
|
"loss": 0.4288, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 10.06, |
|
"learning_rate": 0.0007904761904761905, |
|
"loss": 0.4493, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 10.09, |
|
"learning_rate": 0.0007898809523809524, |
|
"loss": 0.4152, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 10.11, |
|
"learning_rate": 0.0007892857142857143, |
|
"loss": 0.4324, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 10.14, |
|
"learning_rate": 0.0007886904761904761, |
|
"loss": 0.4334, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 10.17, |
|
"learning_rate": 0.0007880952380952381, |
|
"loss": 0.4479, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 10.2, |
|
"learning_rate": 0.0007875, |
|
"loss": 0.4391, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 10.23, |
|
"learning_rate": 0.0007869047619047619, |
|
"loss": 0.4534, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 10.26, |
|
"learning_rate": 0.0007863095238095238, |
|
"loss": 0.4494, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 10.29, |
|
"learning_rate": 0.0007857142857142857, |
|
"loss": 0.4519, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 10.31, |
|
"learning_rate": 0.0007851190476190476, |
|
"loss": 0.4673, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 10.34, |
|
"learning_rate": 0.0007845238095238095, |
|
"loss": 0.4628, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 10.37, |
|
"learning_rate": 0.0007839285714285715, |
|
"loss": 0.4608, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 10.4, |
|
"learning_rate": 0.0007833333333333334, |
|
"loss": 0.4755, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 10.43, |
|
"learning_rate": 0.0007827380952380952, |
|
"loss": 0.4771, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 10.46, |
|
"learning_rate": 0.0007821428571428571, |
|
"loss": 0.4679, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 10.49, |
|
"learning_rate": 0.0007815476190476191, |
|
"loss": 0.4985, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 10.51, |
|
"learning_rate": 0.000780952380952381, |
|
"loss": 0.5242, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 10.54, |
|
"learning_rate": 0.0007803571428571429, |
|
"loss": 0.478, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 10.57, |
|
"learning_rate": 0.0007797619047619047, |
|
"loss": 0.5072, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 10.6, |
|
"learning_rate": 0.0007791666666666667, |
|
"loss": 0.5001, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 10.63, |
|
"learning_rate": 0.0007785714285714286, |
|
"loss": 0.5119, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 10.66, |
|
"learning_rate": 0.0007779761904761905, |
|
"loss": 0.5212, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 10.69, |
|
"learning_rate": 0.0007773809523809525, |
|
"loss": 0.5073, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 10.71, |
|
"learning_rate": 0.0007767857142857143, |
|
"loss": 0.5089, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 10.74, |
|
"learning_rate": 0.0007761904761904762, |
|
"loss": 0.5161, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 10.77, |
|
"learning_rate": 0.0007755952380952381, |
|
"loss": 0.4861, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 10.8, |
|
"learning_rate": 0.0007750000000000001, |
|
"loss": 0.531, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 10.83, |
|
"learning_rate": 0.000774404761904762, |
|
"loss": 0.5244, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 10.86, |
|
"learning_rate": 0.0007738095238095238, |
|
"loss": 0.5446, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 10.89, |
|
"learning_rate": 0.0007732142857142857, |
|
"loss": 0.5515, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 10.91, |
|
"learning_rate": 0.0007726190476190477, |
|
"loss": 0.5345, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 10.94, |
|
"learning_rate": 0.0007720238095238096, |
|
"loss": 0.537, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 10.97, |
|
"learning_rate": 0.0007714285714285715, |
|
"loss": 0.5589, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"learning_rate": 0.0007708333333333334, |
|
"loss": 0.5459, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 11.03, |
|
"learning_rate": 0.0007702380952380953, |
|
"loss": 0.3344, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 11.06, |
|
"learning_rate": 0.0007696428571428572, |
|
"loss": 0.3352, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 11.09, |
|
"learning_rate": 0.0007690476190476191, |
|
"loss": 0.3263, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 11.11, |
|
"learning_rate": 0.0007684523809523811, |
|
"loss": 0.3501, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 11.14, |
|
"learning_rate": 0.0007678571428571429, |
|
"loss": 0.3523, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 11.17, |
|
"learning_rate": 0.0007672619047619048, |
|
"loss": 0.3379, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 11.2, |
|
"learning_rate": 0.0007666666666666667, |
|
"loss": 0.3456, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 11.23, |
|
"learning_rate": 0.0007660714285714287, |
|
"loss": 0.347, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 11.26, |
|
"learning_rate": 0.0007654761904761904, |
|
"loss": 0.3622, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 11.29, |
|
"learning_rate": 0.0007648809523809523, |
|
"loss": 0.3612, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 11.31, |
|
"learning_rate": 0.0007642857142857142, |
|
"loss": 0.3789, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 11.34, |
|
"learning_rate": 0.0007636904761904762, |
|
"loss": 0.3491, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 11.37, |
|
"learning_rate": 0.0007630952380952381, |
|
"loss": 0.3578, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 11.4, |
|
"learning_rate": 0.0007624999999999999, |
|
"loss": 0.3524, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 11.43, |
|
"learning_rate": 0.0007619047619047619, |
|
"loss": 0.3671, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 11.46, |
|
"learning_rate": 0.0007613095238095238, |
|
"loss": 0.374, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 11.49, |
|
"learning_rate": 0.0007607142857142857, |
|
"loss": 0.3872, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 11.51, |
|
"learning_rate": 0.0007601190476190476, |
|
"loss": 0.3801, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 11.54, |
|
"learning_rate": 0.0007595238095238095, |
|
"loss": 0.3618, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 11.57, |
|
"learning_rate": 0.0007589285714285714, |
|
"loss": 0.3929, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 11.6, |
|
"learning_rate": 0.0007583333333333333, |
|
"loss": 0.4099, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 11.63, |
|
"learning_rate": 0.0007577380952380952, |
|
"loss": 0.3778, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 11.66, |
|
"learning_rate": 0.0007571428571428572, |
|
"loss": 0.3965, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 11.69, |
|
"learning_rate": 0.000756547619047619, |
|
"loss": 0.4013, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 11.71, |
|
"learning_rate": 0.0007559523809523809, |
|
"loss": 0.4031, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 11.74, |
|
"learning_rate": 0.0007553571428571429, |
|
"loss": 0.398, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 11.77, |
|
"learning_rate": 0.0007547619047619048, |
|
"loss": 0.4106, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 11.8, |
|
"learning_rate": 0.0007541666666666667, |
|
"loss": 0.4031, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 11.83, |
|
"learning_rate": 0.0007535714285714285, |
|
"loss": 0.4199, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 11.86, |
|
"learning_rate": 0.0007529761904761905, |
|
"loss": 0.4012, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 11.89, |
|
"learning_rate": 0.0007523809523809524, |
|
"loss": 0.4096, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 11.91, |
|
"learning_rate": 0.0007517857142857143, |
|
"loss": 0.4237, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 11.94, |
|
"learning_rate": 0.0007511904761904762, |
|
"loss": 0.4115, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 11.97, |
|
"learning_rate": 0.0007505952380952381, |
|
"loss": 0.4459, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"learning_rate": 0.00075, |
|
"loss": 0.4406, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 12.03, |
|
"learning_rate": 0.0007494047619047619, |
|
"loss": 0.2561, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 12.06, |
|
"learning_rate": 0.0007488095238095239, |
|
"loss": 0.2662, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 12.09, |
|
"learning_rate": 0.0007482142857142858, |
|
"loss": 0.2538, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 12.11, |
|
"learning_rate": 0.0007476190476190476, |
|
"loss": 0.2571, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 12.14, |
|
"learning_rate": 0.0007470238095238095, |
|
"loss": 0.2527, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 12.17, |
|
"learning_rate": 0.0007464285714285715, |
|
"loss": 0.2739, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 12.2, |
|
"learning_rate": 0.0007458333333333334, |
|
"loss": 0.2642, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 12.23, |
|
"learning_rate": 0.0007452380952380953, |
|
"loss": 0.2736, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 12.26, |
|
"learning_rate": 0.0007446428571428571, |
|
"loss": 0.2774, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 12.29, |
|
"learning_rate": 0.0007440476190476191, |
|
"loss": 0.2684, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 12.31, |
|
"learning_rate": 0.000743452380952381, |
|
"loss": 0.2727, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 12.34, |
|
"learning_rate": 0.0007428571428571429, |
|
"loss": 0.2801, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 12.37, |
|
"learning_rate": 0.0007422619047619049, |
|
"loss": 0.2685, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 12.4, |
|
"learning_rate": 0.0007416666666666667, |
|
"loss": 0.2958, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 12.43, |
|
"learning_rate": 0.0007410714285714286, |
|
"loss": 0.2868, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 12.46, |
|
"learning_rate": 0.0007404761904761905, |
|
"loss": 0.2838, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 12.49, |
|
"learning_rate": 0.0007398809523809525, |
|
"loss": 0.2976, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 12.51, |
|
"learning_rate": 0.0007392857142857144, |
|
"loss": 0.2963, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 12.54, |
|
"learning_rate": 0.0007386904761904762, |
|
"loss": 0.2891, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 12.57, |
|
"learning_rate": 0.0007380952380952381, |
|
"loss": 0.2897, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 12.6, |
|
"learning_rate": 0.0007375000000000001, |
|
"loss": 0.3013, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 12.63, |
|
"learning_rate": 0.000736904761904762, |
|
"loss": 0.2977, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 12.66, |
|
"learning_rate": 0.0007363095238095238, |
|
"loss": 0.2985, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 12.69, |
|
"learning_rate": 0.0007357142857142858, |
|
"loss": 0.3097, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 12.71, |
|
"learning_rate": 0.0007351190476190477, |
|
"loss": 0.3183, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 12.74, |
|
"learning_rate": 0.0007345238095238096, |
|
"loss": 0.3192, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 12.77, |
|
"learning_rate": 0.0007339285714285714, |
|
"loss": 0.3265, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 12.8, |
|
"learning_rate": 0.0007333333333333333, |
|
"loss": 0.3033, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 12.83, |
|
"learning_rate": 0.0007327380952380952, |
|
"loss": 0.313, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 12.86, |
|
"learning_rate": 0.0007321428571428571, |
|
"loss": 0.3186, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 12.89, |
|
"learning_rate": 0.000731547619047619, |
|
"loss": 0.3275, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 12.91, |
|
"learning_rate": 0.000730952380952381, |
|
"loss": 0.3272, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 12.94, |
|
"learning_rate": 0.0007303571428571428, |
|
"loss": 0.3336, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 12.97, |
|
"learning_rate": 0.0007297619047619047, |
|
"loss": 0.3244, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"learning_rate": 0.0007291666666666666, |
|
"loss": 0.3403, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 13.03, |
|
"learning_rate": 0.0007285714285714286, |
|
"loss": 0.1931, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 13.06, |
|
"learning_rate": 0.0007279761904761905, |
|
"loss": 0.2104, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 13.09, |
|
"learning_rate": 0.0007273809523809523, |
|
"loss": 0.2025, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 13.11, |
|
"learning_rate": 0.0007267857142857143, |
|
"loss": 0.1933, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 13.14, |
|
"learning_rate": 0.0007261904761904762, |
|
"loss": 0.2002, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 13.17, |
|
"learning_rate": 0.0007255952380952381, |
|
"loss": 0.2097, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 13.2, |
|
"learning_rate": 0.000725, |
|
"loss": 0.2097, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 13.23, |
|
"learning_rate": 0.0007244047619047619, |
|
"loss": 0.2061, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 13.26, |
|
"learning_rate": 0.0007238095238095238, |
|
"loss": 0.2, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 13.29, |
|
"learning_rate": 0.0007232142857142857, |
|
"loss": 0.2082, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 13.31, |
|
"learning_rate": 0.0007226190476190476, |
|
"loss": 0.209, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 13.34, |
|
"learning_rate": 0.0007220238095238096, |
|
"loss": 0.2083, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 13.37, |
|
"learning_rate": 0.0007214285714285714, |
|
"loss": 0.215, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 13.4, |
|
"learning_rate": 0.0007208333333333333, |
|
"loss": 0.2113, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 13.43, |
|
"learning_rate": 0.0007202380952380953, |
|
"loss": 0.2186, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 13.46, |
|
"learning_rate": 0.0007196428571428572, |
|
"loss": 0.2164, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 13.49, |
|
"learning_rate": 0.000719047619047619, |
|
"loss": 0.2197, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 13.51, |
|
"learning_rate": 0.0007184523809523809, |
|
"loss": 0.2256, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 13.54, |
|
"learning_rate": 0.0007178571428571429, |
|
"loss": 0.2165, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 13.57, |
|
"learning_rate": 0.0007172619047619048, |
|
"loss": 0.2227, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 13.6, |
|
"learning_rate": 0.0007166666666666667, |
|
"loss": 0.2245, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 13.63, |
|
"learning_rate": 0.0007160714285714285, |
|
"loss": 0.2239, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 13.66, |
|
"learning_rate": 0.0007154761904761905, |
|
"loss": 0.2251, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 13.69, |
|
"learning_rate": 0.0007148809523809524, |
|
"loss": 0.2305, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 13.71, |
|
"learning_rate": 0.0007142857142857143, |
|
"loss": 0.2334, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 13.74, |
|
"learning_rate": 0.0007136904761904763, |
|
"loss": 0.229, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 13.77, |
|
"learning_rate": 0.0007130952380952381, |
|
"loss": 0.2363, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 13.8, |
|
"learning_rate": 0.0007125, |
|
"loss": 0.2382, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 13.83, |
|
"learning_rate": 0.0007119047619047619, |
|
"loss": 0.2374, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 13.86, |
|
"learning_rate": 0.0007113095238095239, |
|
"loss": 0.2433, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 13.89, |
|
"learning_rate": 0.0007107142857142858, |
|
"loss": 0.2349, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 13.91, |
|
"learning_rate": 0.0007101190476190476, |
|
"loss": 0.2549, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 13.94, |
|
"learning_rate": 0.0007095238095238095, |
|
"loss": 0.2481, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 13.97, |
|
"learning_rate": 0.0007089285714285715, |
|
"loss": 0.2504, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"learning_rate": 0.0007083333333333334, |
|
"loss": 0.2521, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 0.0007077380952380953, |
|
"loss": 0.1489, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 14.06, |
|
"learning_rate": 0.0007071428571428572, |
|
"loss": 0.15, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 14.09, |
|
"learning_rate": 0.0007065476190476191, |
|
"loss": 0.1499, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 14.11, |
|
"learning_rate": 0.000705952380952381, |
|
"loss": 0.1515, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 14.14, |
|
"learning_rate": 0.0007053571428571429, |
|
"loss": 0.1542, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 14.17, |
|
"learning_rate": 0.0007047619047619049, |
|
"loss": 0.1523, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 14.2, |
|
"learning_rate": 0.0007041666666666667, |
|
"loss": 0.1605, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 14.23, |
|
"learning_rate": 0.0007035714285714286, |
|
"loss": 0.1572, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 14.26, |
|
"learning_rate": 0.0007029761904761904, |
|
"loss": 0.1634, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 14.29, |
|
"learning_rate": 0.0007023809523809524, |
|
"loss": 0.1548, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 14.31, |
|
"learning_rate": 0.0007017857142857143, |
|
"loss": 0.1521, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 14.34, |
|
"learning_rate": 0.0007011904761904761, |
|
"loss": 0.1644, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 14.37, |
|
"learning_rate": 0.0007005952380952381, |
|
"loss": 0.155, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 14.4, |
|
"learning_rate": 0.0007, |
|
"loss": 0.1674, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 14.43, |
|
"learning_rate": 0.0006994047619047619, |
|
"loss": 0.1619, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 14.46, |
|
"learning_rate": 0.0006988095238095237, |
|
"loss": 0.1644, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 14.49, |
|
"learning_rate": 0.0006982142857142857, |
|
"loss": 0.1723, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 14.51, |
|
"learning_rate": 0.0006976190476190476, |
|
"loss": 0.1621, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 14.54, |
|
"learning_rate": 0.0006970238095238095, |
|
"loss": 0.1647, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 14.57, |
|
"learning_rate": 0.0006964285714285714, |
|
"loss": 0.1741, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 14.6, |
|
"learning_rate": 0.0006958333333333334, |
|
"loss": 0.1673, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 14.63, |
|
"learning_rate": 0.0006952380952380952, |
|
"loss": 0.1752, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 14.66, |
|
"learning_rate": 0.0006946428571428571, |
|
"loss": 0.167, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 14.69, |
|
"learning_rate": 0.0006940476190476191, |
|
"loss": 0.1718, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 14.71, |
|
"learning_rate": 0.000693452380952381, |
|
"loss": 0.1787, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 14.74, |
|
"learning_rate": 0.0006928571428571428, |
|
"loss": 0.1747, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 14.77, |
|
"learning_rate": 0.0006922619047619047, |
|
"loss": 0.1766, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 14.8, |
|
"learning_rate": 0.0006916666666666667, |
|
"loss": 0.1782, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 14.83, |
|
"learning_rate": 0.0006910714285714286, |
|
"loss": 0.1799, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 14.86, |
|
"learning_rate": 0.0006904761904761905, |
|
"loss": 0.169, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 14.89, |
|
"learning_rate": 0.0006898809523809523, |
|
"loss": 0.1802, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 14.91, |
|
"learning_rate": 0.0006892857142857143, |
|
"loss": 0.18, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 14.94, |
|
"learning_rate": 0.0006886904761904762, |
|
"loss": 0.1823, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 14.97, |
|
"learning_rate": 0.0006880952380952381, |
|
"loss": 0.1829, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"learning_rate": 0.0006875, |
|
"loss": 0.1885, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 0.000686904761904762, |
|
"loss": 0.1171, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 15.06, |
|
"learning_rate": 0.0006863095238095238, |
|
"loss": 0.1126, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 15.09, |
|
"learning_rate": 0.0006857142857142857, |
|
"loss": 0.1186, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 15.11, |
|
"learning_rate": 0.0006851190476190477, |
|
"loss": 0.1164, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 15.14, |
|
"learning_rate": 0.0006845238095238096, |
|
"loss": 0.1163, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 15.17, |
|
"learning_rate": 0.0006839285714285714, |
|
"loss": 0.1204, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 15.2, |
|
"learning_rate": 0.0006833333333333333, |
|
"loss": 0.1212, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 15.23, |
|
"learning_rate": 0.0006827380952380953, |
|
"loss": 0.1188, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 15.26, |
|
"learning_rate": 0.0006821428571428572, |
|
"loss": 0.1154, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 15.29, |
|
"learning_rate": 0.0006815476190476191, |
|
"loss": 0.1244, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 15.31, |
|
"learning_rate": 0.0006809523809523809, |
|
"loss": 0.1214, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 15.34, |
|
"learning_rate": 0.0006803571428571429, |
|
"loss": 0.1294, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 15.37, |
|
"learning_rate": 0.0006797619047619048, |
|
"loss": 0.1232, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 15.4, |
|
"learning_rate": 0.0006791666666666667, |
|
"loss": 0.1262, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 15.43, |
|
"learning_rate": 0.0006785714285714287, |
|
"loss": 0.1195, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 15.46, |
|
"learning_rate": 0.0006779761904761905, |
|
"loss": 0.123, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 15.49, |
|
"learning_rate": 0.0006773809523809524, |
|
"loss": 0.1266, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 15.51, |
|
"learning_rate": 0.0006767857142857143, |
|
"loss": 0.1345, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 15.54, |
|
"learning_rate": 0.0006761904761904763, |
|
"loss": 0.1174, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 15.57, |
|
"learning_rate": 0.0006755952380952382, |
|
"loss": 0.1293, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 15.6, |
|
"learning_rate": 0.000675, |
|
"loss": 0.1293, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 15.63, |
|
"learning_rate": 0.0006744047619047619, |
|
"loss": 0.1269, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 15.66, |
|
"learning_rate": 0.0006738095238095239, |
|
"loss": 0.1321, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 15.69, |
|
"learning_rate": 0.0006732142857142858, |
|
"loss": 0.1318, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 15.71, |
|
"learning_rate": 0.0006726190476190477, |
|
"loss": 0.1283, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 15.74, |
|
"learning_rate": 0.0006720238095238096, |
|
"loss": 0.128, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 15.77, |
|
"learning_rate": 0.0006714285714285714, |
|
"loss": 0.1295, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 15.8, |
|
"learning_rate": 0.0006708333333333333, |
|
"loss": 0.1323, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 15.83, |
|
"learning_rate": 0.0006702380952380952, |
|
"loss": 0.1348, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 15.86, |
|
"learning_rate": 0.0006696428571428571, |
|
"loss": 0.1276, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 15.89, |
|
"learning_rate": 0.000669047619047619, |
|
"loss": 0.1356, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 15.91, |
|
"learning_rate": 0.0006684523809523809, |
|
"loss": 0.1404, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 15.94, |
|
"learning_rate": 0.0006678571428571428, |
|
"loss": 0.1311, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 15.97, |
|
"learning_rate": 0.0006672619047619048, |
|
"loss": 0.1401, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"learning_rate": 0.0006666666666666666, |
|
"loss": 0.1411, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 16.03, |
|
"learning_rate": 0.0006660714285714285, |
|
"loss": 0.0862, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 16.06, |
|
"learning_rate": 0.0006654761904761905, |
|
"loss": 0.0902, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 16.09, |
|
"learning_rate": 0.0006648809523809524, |
|
"loss": 0.0871, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 16.11, |
|
"learning_rate": 0.0006642857142857143, |
|
"loss": 0.0906, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 16.14, |
|
"learning_rate": 0.0006636904761904761, |
|
"loss": 0.0891, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 16.17, |
|
"learning_rate": 0.0006630952380952381, |
|
"loss": 0.0917, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 16.2, |
|
"learning_rate": 0.0006625, |
|
"loss": 0.0906, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 16.23, |
|
"learning_rate": 0.0006619047619047619, |
|
"loss": 0.0927, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 16.26, |
|
"learning_rate": 0.0006613095238095238, |
|
"loss": 0.0927, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 16.29, |
|
"learning_rate": 0.0006607142857142857, |
|
"loss": 0.0934, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 16.31, |
|
"learning_rate": 0.0006601190476190476, |
|
"loss": 0.0956, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 16.34, |
|
"learning_rate": 0.0006595238095238095, |
|
"loss": 0.0933, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 16.37, |
|
"learning_rate": 0.0006589285714285715, |
|
"loss": 0.0993, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 16.4, |
|
"learning_rate": 0.0006583333333333334, |
|
"loss": 0.095, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 16.43, |
|
"learning_rate": 0.0006577380952380952, |
|
"loss": 0.0963, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 16.46, |
|
"learning_rate": 0.0006571428571428571, |
|
"loss": 0.0948, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 16.49, |
|
"learning_rate": 0.0006565476190476191, |
|
"loss": 0.0952, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 16.51, |
|
"learning_rate": 0.000655952380952381, |
|
"loss": 0.1001, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 16.54, |
|
"learning_rate": 0.0006553571428571429, |
|
"loss": 0.0924, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 16.57, |
|
"learning_rate": 0.0006547619047619047, |
|
"loss": 0.0962, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 16.6, |
|
"learning_rate": 0.0006541666666666667, |
|
"loss": 0.0949, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 16.63, |
|
"learning_rate": 0.0006535714285714286, |
|
"loss": 0.1, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 16.66, |
|
"learning_rate": 0.0006529761904761905, |
|
"loss": 0.1009, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 16.69, |
|
"learning_rate": 0.0006523809523809525, |
|
"loss": 0.1023, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 16.71, |
|
"learning_rate": 0.0006517857142857143, |
|
"loss": 0.0995, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 16.74, |
|
"learning_rate": 0.0006511904761904762, |
|
"loss": 0.1015, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 16.77, |
|
"learning_rate": 0.0006505952380952381, |
|
"loss": 0.0966, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 16.8, |
|
"learning_rate": 0.0006500000000000001, |
|
"loss": 0.1019, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 16.83, |
|
"learning_rate": 0.000649404761904762, |
|
"loss": 0.0996, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 16.86, |
|
"learning_rate": 0.0006488095238095238, |
|
"loss": 0.103, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 16.89, |
|
"learning_rate": 0.0006482142857142857, |
|
"loss": 0.1042, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 16.91, |
|
"learning_rate": 0.0006476190476190477, |
|
"loss": 0.1039, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 16.94, |
|
"learning_rate": 0.0006470238095238096, |
|
"loss": 0.1058, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 16.97, |
|
"learning_rate": 0.0006464285714285715, |
|
"loss": 0.0994, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"learning_rate": 0.0006458333333333334, |
|
"loss": 0.1062, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 17.03, |
|
"learning_rate": 0.0006452380952380953, |
|
"loss": 0.0709, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 17.06, |
|
"learning_rate": 0.0006446428571428572, |
|
"loss": 0.0733, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 17.09, |
|
"learning_rate": 0.0006440476190476191, |
|
"loss": 0.0724, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 17.11, |
|
"learning_rate": 0.0006434523809523811, |
|
"loss": 0.0733, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 17.14, |
|
"learning_rate": 0.0006428571428571429, |
|
"loss": 0.0741, |
|
"step": 600 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 1680, |
|
"num_train_epochs": 48, |
|
"save_steps": 100, |
|
"total_flos": 3.463615849187021e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|