|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 0, |
|
"global_step": 495, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.00202020202020202, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 9.97979797979798e-06, |
|
"loss": 1.793, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.00404040404040404, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 9.95959595959596e-06, |
|
"loss": 1.8337, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.006060606060606061, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 9.939393939393939e-06, |
|
"loss": 1.829, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.00808080808080808, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 9.91919191919192e-06, |
|
"loss": 1.7952, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.010101010101010102, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 9.8989898989899e-06, |
|
"loss": 1.8804, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.012121212121212121, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 9.87878787878788e-06, |
|
"loss": 1.8649, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.014141414141414142, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 9.85858585858586e-06, |
|
"loss": 1.8901, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.01616161616161616, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 9.838383838383839e-06, |
|
"loss": 1.7955, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.01818181818181818, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 9.81818181818182e-06, |
|
"loss": 1.8151, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.020202020202020204, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 9.797979797979798e-06, |
|
"loss": 1.7557, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.022222222222222223, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 9.777777777777779e-06, |
|
"loss": 1.7174, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.024242424242424242, |
|
"grad_norm": 0.287109375, |
|
"learning_rate": 9.757575757575758e-06, |
|
"loss": 1.7118, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.026262626262626262, |
|
"grad_norm": 0.279296875, |
|
"learning_rate": 9.737373737373738e-06, |
|
"loss": 1.5918, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.028282828282828285, |
|
"grad_norm": 0.302734375, |
|
"learning_rate": 9.717171717171719e-06, |
|
"loss": 1.6372, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.030303030303030304, |
|
"grad_norm": 0.283203125, |
|
"learning_rate": 9.696969696969698e-06, |
|
"loss": 1.6647, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.03232323232323232, |
|
"grad_norm": 0.296875, |
|
"learning_rate": 9.676767676767678e-06, |
|
"loss": 1.6657, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.03434343434343434, |
|
"grad_norm": 0.283203125, |
|
"learning_rate": 9.656565656565657e-06, |
|
"loss": 1.4841, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.03636363636363636, |
|
"grad_norm": 0.2890625, |
|
"learning_rate": 9.636363636363638e-06, |
|
"loss": 1.5899, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.03838383838383838, |
|
"grad_norm": 0.29296875, |
|
"learning_rate": 9.616161616161616e-06, |
|
"loss": 1.6603, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.04040404040404041, |
|
"grad_norm": 0.275390625, |
|
"learning_rate": 9.595959595959597e-06, |
|
"loss": 1.5316, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.04242424242424243, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 9.575757575757576e-06, |
|
"loss": 1.6512, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.044444444444444446, |
|
"grad_norm": 0.25390625, |
|
"learning_rate": 9.555555555555556e-06, |
|
"loss": 1.499, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.046464646464646465, |
|
"grad_norm": 0.263671875, |
|
"learning_rate": 9.535353535353537e-06, |
|
"loss": 1.5049, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.048484848484848485, |
|
"grad_norm": 0.25, |
|
"learning_rate": 9.515151515151516e-06, |
|
"loss": 1.5328, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.050505050505050504, |
|
"grad_norm": 0.240234375, |
|
"learning_rate": 9.494949494949497e-06, |
|
"loss": 1.5673, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.052525252525252523, |
|
"grad_norm": 0.259765625, |
|
"learning_rate": 9.474747474747475e-06, |
|
"loss": 1.5331, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.05454545454545454, |
|
"grad_norm": 0.2373046875, |
|
"learning_rate": 9.454545454545456e-06, |
|
"loss": 1.5281, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.05656565656565657, |
|
"grad_norm": 0.220703125, |
|
"learning_rate": 9.434343434343435e-06, |
|
"loss": 1.4541, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.05858585858585859, |
|
"grad_norm": 0.2431640625, |
|
"learning_rate": 9.414141414141414e-06, |
|
"loss": 1.5046, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.06060606060606061, |
|
"grad_norm": 0.234375, |
|
"learning_rate": 9.393939393939396e-06, |
|
"loss": 1.4657, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.06262626262626263, |
|
"grad_norm": 0.2275390625, |
|
"learning_rate": 9.373737373737375e-06, |
|
"loss": 1.3571, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.06464646464646465, |
|
"grad_norm": 0.296875, |
|
"learning_rate": 9.353535353535354e-06, |
|
"loss": 1.3992, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.06666666666666667, |
|
"grad_norm": 0.287109375, |
|
"learning_rate": 9.333333333333334e-06, |
|
"loss": 1.4552, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.06868686868686869, |
|
"grad_norm": 0.216796875, |
|
"learning_rate": 9.313131313131313e-06, |
|
"loss": 1.4571, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.0707070707070707, |
|
"grad_norm": 0.2265625, |
|
"learning_rate": 9.292929292929294e-06, |
|
"loss": 1.4078, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.07272727272727272, |
|
"grad_norm": 0.279296875, |
|
"learning_rate": 9.272727272727273e-06, |
|
"loss": 1.3755, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.07474747474747474, |
|
"grad_norm": 0.2021484375, |
|
"learning_rate": 9.252525252525253e-06, |
|
"loss": 1.3593, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.07676767676767676, |
|
"grad_norm": 0.1923828125, |
|
"learning_rate": 9.232323232323232e-06, |
|
"loss": 1.2517, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.07878787878787878, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 9.212121212121213e-06, |
|
"loss": 1.3804, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.08080808080808081, |
|
"grad_norm": 0.2001953125, |
|
"learning_rate": 9.191919191919193e-06, |
|
"loss": 1.3428, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.08282828282828283, |
|
"grad_norm": 0.2109375, |
|
"learning_rate": 9.171717171717172e-06, |
|
"loss": 1.3576, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.08484848484848485, |
|
"grad_norm": 0.197265625, |
|
"learning_rate": 9.151515151515153e-06, |
|
"loss": 1.3772, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.08686868686868687, |
|
"grad_norm": 0.1826171875, |
|
"learning_rate": 9.131313131313132e-06, |
|
"loss": 1.2995, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.08888888888888889, |
|
"grad_norm": 0.185546875, |
|
"learning_rate": 9.111111111111112e-06, |
|
"loss": 1.3324, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.09090909090909091, |
|
"grad_norm": 0.1953125, |
|
"learning_rate": 9.090909090909091e-06, |
|
"loss": 1.324, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.09292929292929293, |
|
"grad_norm": 0.1923828125, |
|
"learning_rate": 9.070707070707072e-06, |
|
"loss": 1.3671, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.09494949494949495, |
|
"grad_norm": 0.1826171875, |
|
"learning_rate": 9.050505050505052e-06, |
|
"loss": 1.3559, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.09696969696969697, |
|
"grad_norm": 0.1728515625, |
|
"learning_rate": 9.030303030303031e-06, |
|
"loss": 1.3197, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.09898989898989899, |
|
"grad_norm": 0.171875, |
|
"learning_rate": 9.010101010101012e-06, |
|
"loss": 1.2988, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.10101010101010101, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 8.98989898989899e-06, |
|
"loss": 1.29, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.10303030303030303, |
|
"grad_norm": 0.1787109375, |
|
"learning_rate": 8.969696969696971e-06, |
|
"loss": 1.2839, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.10505050505050505, |
|
"grad_norm": 0.1650390625, |
|
"learning_rate": 8.94949494949495e-06, |
|
"loss": 1.244, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.10707070707070707, |
|
"grad_norm": 0.1689453125, |
|
"learning_rate": 8.92929292929293e-06, |
|
"loss": 1.2877, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.10909090909090909, |
|
"grad_norm": 0.1787109375, |
|
"learning_rate": 8.90909090909091e-06, |
|
"loss": 1.269, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.1111111111111111, |
|
"grad_norm": 0.171875, |
|
"learning_rate": 8.888888888888888e-06, |
|
"loss": 1.3108, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.11313131313131314, |
|
"grad_norm": 0.16796875, |
|
"learning_rate": 8.86868686868687e-06, |
|
"loss": 1.3129, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.11515151515151516, |
|
"grad_norm": 0.1728515625, |
|
"learning_rate": 8.84848484848485e-06, |
|
"loss": 1.1976, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.11717171717171718, |
|
"grad_norm": 0.169921875, |
|
"learning_rate": 8.82828282828283e-06, |
|
"loss": 1.283, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.1191919191919192, |
|
"grad_norm": 0.15625, |
|
"learning_rate": 8.808080808080809e-06, |
|
"loss": 1.2786, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.12121212121212122, |
|
"grad_norm": 0.197265625, |
|
"learning_rate": 8.787878787878788e-06, |
|
"loss": 1.2718, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.12323232323232323, |
|
"grad_norm": 0.181640625, |
|
"learning_rate": 8.767676767676768e-06, |
|
"loss": 1.3529, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.12525252525252525, |
|
"grad_norm": 0.16015625, |
|
"learning_rate": 8.747474747474747e-06, |
|
"loss": 1.1866, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.12727272727272726, |
|
"grad_norm": 0.1630859375, |
|
"learning_rate": 8.727272727272728e-06, |
|
"loss": 1.2852, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.1292929292929293, |
|
"grad_norm": 0.171875, |
|
"learning_rate": 8.707070707070707e-06, |
|
"loss": 1.2611, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.13131313131313133, |
|
"grad_norm": 0.2451171875, |
|
"learning_rate": 8.686868686868687e-06, |
|
"loss": 1.1993, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.13333333333333333, |
|
"grad_norm": 0.1572265625, |
|
"learning_rate": 8.666666666666668e-06, |
|
"loss": 1.2068, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.13535353535353536, |
|
"grad_norm": 0.1611328125, |
|
"learning_rate": 8.646464646464647e-06, |
|
"loss": 1.2273, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.13737373737373737, |
|
"grad_norm": 0.17578125, |
|
"learning_rate": 8.626262626262627e-06, |
|
"loss": 1.1792, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.1393939393939394, |
|
"grad_norm": 0.177734375, |
|
"learning_rate": 8.606060606060606e-06, |
|
"loss": 1.2232, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.1414141414141414, |
|
"grad_norm": 0.1748046875, |
|
"learning_rate": 8.585858585858587e-06, |
|
"loss": 1.1552, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.14343434343434344, |
|
"grad_norm": 0.1533203125, |
|
"learning_rate": 8.565656565656566e-06, |
|
"loss": 1.1916, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.14545454545454545, |
|
"grad_norm": 0.1552734375, |
|
"learning_rate": 8.545454545454546e-06, |
|
"loss": 1.1712, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.14747474747474748, |
|
"grad_norm": 0.162109375, |
|
"learning_rate": 8.525252525252527e-06, |
|
"loss": 1.2027, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.1494949494949495, |
|
"grad_norm": 0.1787109375, |
|
"learning_rate": 8.505050505050506e-06, |
|
"loss": 1.2892, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.15151515151515152, |
|
"grad_norm": 0.150390625, |
|
"learning_rate": 8.484848484848486e-06, |
|
"loss": 1.1894, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.15353535353535352, |
|
"grad_norm": 0.158203125, |
|
"learning_rate": 8.464646464646465e-06, |
|
"loss": 1.1829, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.15555555555555556, |
|
"grad_norm": 0.1455078125, |
|
"learning_rate": 8.444444444444446e-06, |
|
"loss": 1.1127, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.15757575757575756, |
|
"grad_norm": 0.1533203125, |
|
"learning_rate": 8.424242424242425e-06, |
|
"loss": 1.2093, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.1595959595959596, |
|
"grad_norm": 0.150390625, |
|
"learning_rate": 8.404040404040405e-06, |
|
"loss": 1.1844, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.16161616161616163, |
|
"grad_norm": 0.1650390625, |
|
"learning_rate": 8.383838383838384e-06, |
|
"loss": 1.175, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.16363636363636364, |
|
"grad_norm": 0.1962890625, |
|
"learning_rate": 8.363636363636365e-06, |
|
"loss": 1.1569, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.16565656565656567, |
|
"grad_norm": 0.197265625, |
|
"learning_rate": 8.343434343434345e-06, |
|
"loss": 1.2603, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.16767676767676767, |
|
"grad_norm": 0.1484375, |
|
"learning_rate": 8.323232323232324e-06, |
|
"loss": 1.1733, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.1696969696969697, |
|
"grad_norm": 0.1591796875, |
|
"learning_rate": 8.303030303030305e-06, |
|
"loss": 1.2082, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.1717171717171717, |
|
"grad_norm": 0.1982421875, |
|
"learning_rate": 8.282828282828283e-06, |
|
"loss": 1.1991, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.17373737373737375, |
|
"grad_norm": 0.15234375, |
|
"learning_rate": 8.262626262626264e-06, |
|
"loss": 1.1648, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.17575757575757575, |
|
"grad_norm": 0.2216796875, |
|
"learning_rate": 8.242424242424243e-06, |
|
"loss": 1.1292, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.17777777777777778, |
|
"grad_norm": 0.1552734375, |
|
"learning_rate": 8.222222222222222e-06, |
|
"loss": 1.1936, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.1797979797979798, |
|
"grad_norm": 0.1728515625, |
|
"learning_rate": 8.202020202020202e-06, |
|
"loss": 1.1531, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.18181818181818182, |
|
"grad_norm": 0.1513671875, |
|
"learning_rate": 8.181818181818183e-06, |
|
"loss": 1.1727, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.18383838383838383, |
|
"grad_norm": 0.162109375, |
|
"learning_rate": 8.161616161616162e-06, |
|
"loss": 1.1873, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.18585858585858586, |
|
"grad_norm": 0.1953125, |
|
"learning_rate": 8.141414141414142e-06, |
|
"loss": 1.1735, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.18787878787878787, |
|
"grad_norm": 0.1923828125, |
|
"learning_rate": 8.121212121212121e-06, |
|
"loss": 1.1552, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.1898989898989899, |
|
"grad_norm": 0.1669921875, |
|
"learning_rate": 8.101010101010102e-06, |
|
"loss": 1.1441, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.1919191919191919, |
|
"grad_norm": 0.1728515625, |
|
"learning_rate": 8.08080808080808e-06, |
|
"loss": 1.164, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.19393939393939394, |
|
"grad_norm": 0.1552734375, |
|
"learning_rate": 8.060606060606061e-06, |
|
"loss": 1.1634, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.19595959595959597, |
|
"grad_norm": 0.224609375, |
|
"learning_rate": 8.04040404040404e-06, |
|
"loss": 1.1828, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.19797979797979798, |
|
"grad_norm": 0.181640625, |
|
"learning_rate": 8.02020202020202e-06, |
|
"loss": 1.1501, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.166015625, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 1.1776, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.20202020202020202, |
|
"grad_norm": 0.1533203125, |
|
"learning_rate": 7.97979797979798e-06, |
|
"loss": 1.1401, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.20404040404040405, |
|
"grad_norm": 0.15234375, |
|
"learning_rate": 7.95959595959596e-06, |
|
"loss": 1.1554, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.20606060606060606, |
|
"grad_norm": 0.166015625, |
|
"learning_rate": 7.93939393939394e-06, |
|
"loss": 1.2024, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.2080808080808081, |
|
"grad_norm": 0.154296875, |
|
"learning_rate": 7.91919191919192e-06, |
|
"loss": 1.141, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.2101010101010101, |
|
"grad_norm": 0.158203125, |
|
"learning_rate": 7.898989898989899e-06, |
|
"loss": 1.1776, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.21212121212121213, |
|
"grad_norm": 0.1640625, |
|
"learning_rate": 7.87878787878788e-06, |
|
"loss": 1.1885, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.21414141414141413, |
|
"grad_norm": 0.201171875, |
|
"learning_rate": 7.858585858585859e-06, |
|
"loss": 1.1369, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.21616161616161617, |
|
"grad_norm": 0.162109375, |
|
"learning_rate": 7.838383838383839e-06, |
|
"loss": 1.1101, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.21818181818181817, |
|
"grad_norm": 0.185546875, |
|
"learning_rate": 7.81818181818182e-06, |
|
"loss": 1.1358, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.2202020202020202, |
|
"grad_norm": 0.1689453125, |
|
"learning_rate": 7.797979797979799e-06, |
|
"loss": 1.1267, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.2222222222222222, |
|
"grad_norm": 0.1689453125, |
|
"learning_rate": 7.77777777777778e-06, |
|
"loss": 1.1339, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.22424242424242424, |
|
"grad_norm": 0.19140625, |
|
"learning_rate": 7.757575757575758e-06, |
|
"loss": 1.0777, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.22626262626262628, |
|
"grad_norm": 0.16015625, |
|
"learning_rate": 7.737373737373739e-06, |
|
"loss": 1.1315, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.22828282828282828, |
|
"grad_norm": 0.15234375, |
|
"learning_rate": 7.717171717171717e-06, |
|
"loss": 1.096, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.23030303030303031, |
|
"grad_norm": 0.1787109375, |
|
"learning_rate": 7.696969696969696e-06, |
|
"loss": 1.1127, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.23232323232323232, |
|
"grad_norm": 0.1904296875, |
|
"learning_rate": 7.676767676767677e-06, |
|
"loss": 1.181, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.23434343434343435, |
|
"grad_norm": 0.1630859375, |
|
"learning_rate": 7.656565656565658e-06, |
|
"loss": 1.1442, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.23636363636363636, |
|
"grad_norm": 0.208984375, |
|
"learning_rate": 7.636363636363638e-06, |
|
"loss": 1.1313, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.2383838383838384, |
|
"grad_norm": 0.1630859375, |
|
"learning_rate": 7.616161616161617e-06, |
|
"loss": 1.1727, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.2404040404040404, |
|
"grad_norm": 0.16015625, |
|
"learning_rate": 7.595959595959597e-06, |
|
"loss": 1.0959, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.24242424242424243, |
|
"grad_norm": 0.2197265625, |
|
"learning_rate": 7.5757575757575764e-06, |
|
"loss": 1.1908, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.24444444444444444, |
|
"grad_norm": 0.197265625, |
|
"learning_rate": 7.555555555555556e-06, |
|
"loss": 1.1209, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.24646464646464647, |
|
"grad_norm": 0.1591796875, |
|
"learning_rate": 7.535353535353536e-06, |
|
"loss": 1.1523, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.24848484848484848, |
|
"grad_norm": 0.1767578125, |
|
"learning_rate": 7.515151515151516e-06, |
|
"loss": 1.1379, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.2505050505050505, |
|
"grad_norm": 0.1865234375, |
|
"learning_rate": 7.494949494949496e-06, |
|
"loss": 1.1438, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.25252525252525254, |
|
"grad_norm": 0.16015625, |
|
"learning_rate": 7.474747474747476e-06, |
|
"loss": 1.1362, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.2545454545454545, |
|
"grad_norm": 0.1630859375, |
|
"learning_rate": 7.454545454545456e-06, |
|
"loss": 1.118, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.25656565656565655, |
|
"grad_norm": 0.17578125, |
|
"learning_rate": 7.434343434343435e-06, |
|
"loss": 1.0866, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.2585858585858586, |
|
"grad_norm": 0.2353515625, |
|
"learning_rate": 7.414141414141415e-06, |
|
"loss": 1.0837, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.2606060606060606, |
|
"grad_norm": 0.177734375, |
|
"learning_rate": 7.393939393939395e-06, |
|
"loss": 1.1266, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.26262626262626265, |
|
"grad_norm": 0.162109375, |
|
"learning_rate": 7.373737373737374e-06, |
|
"loss": 1.164, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.26464646464646463, |
|
"grad_norm": 0.177734375, |
|
"learning_rate": 7.353535353535353e-06, |
|
"loss": 1.071, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.26666666666666666, |
|
"grad_norm": 0.19140625, |
|
"learning_rate": 7.333333333333333e-06, |
|
"loss": 1.1683, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.2686868686868687, |
|
"grad_norm": 0.2177734375, |
|
"learning_rate": 7.3131313131313146e-06, |
|
"loss": 1.1239, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.27070707070707073, |
|
"grad_norm": 0.16015625, |
|
"learning_rate": 7.2929292929292934e-06, |
|
"loss": 1.1213, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.2727272727272727, |
|
"grad_norm": 0.158203125, |
|
"learning_rate": 7.272727272727273e-06, |
|
"loss": 1.0712, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.27474747474747474, |
|
"grad_norm": 0.1630859375, |
|
"learning_rate": 7.252525252525253e-06, |
|
"loss": 1.109, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.2767676767676768, |
|
"grad_norm": 0.75, |
|
"learning_rate": 7.232323232323233e-06, |
|
"loss": 1.113, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.2787878787878788, |
|
"grad_norm": 0.1640625, |
|
"learning_rate": 7.212121212121212e-06, |
|
"loss": 1.0855, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.2808080808080808, |
|
"grad_norm": 0.1630859375, |
|
"learning_rate": 7.191919191919192e-06, |
|
"loss": 1.1196, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.2828282828282828, |
|
"grad_norm": 0.1884765625, |
|
"learning_rate": 7.171717171717172e-06, |
|
"loss": 1.1559, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.28484848484848485, |
|
"grad_norm": 0.1953125, |
|
"learning_rate": 7.151515151515152e-06, |
|
"loss": 1.1031, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.2868686868686869, |
|
"grad_norm": 0.16015625, |
|
"learning_rate": 7.131313131313132e-06, |
|
"loss": 1.0808, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.28888888888888886, |
|
"grad_norm": 0.193359375, |
|
"learning_rate": 7.111111111111112e-06, |
|
"loss": 1.0975, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.2909090909090909, |
|
"grad_norm": 0.1845703125, |
|
"learning_rate": 7.0909090909090916e-06, |
|
"loss": 1.1093, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.29292929292929293, |
|
"grad_norm": 0.166015625, |
|
"learning_rate": 7.070707070707071e-06, |
|
"loss": 1.1066, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.29494949494949496, |
|
"grad_norm": 0.20703125, |
|
"learning_rate": 7.050505050505051e-06, |
|
"loss": 1.1233, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.296969696969697, |
|
"grad_norm": 0.1748046875, |
|
"learning_rate": 7.030303030303031e-06, |
|
"loss": 1.1179, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.298989898989899, |
|
"grad_norm": 0.177734375, |
|
"learning_rate": 7.0101010101010105e-06, |
|
"loss": 1.1154, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.301010101010101, |
|
"grad_norm": 0.1953125, |
|
"learning_rate": 6.98989898989899e-06, |
|
"loss": 1.0917, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.30303030303030304, |
|
"grad_norm": 0.19140625, |
|
"learning_rate": 6.969696969696971e-06, |
|
"loss": 1.0864, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.30505050505050507, |
|
"grad_norm": 0.1962890625, |
|
"learning_rate": 6.9494949494949505e-06, |
|
"loss": 1.1178, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.30707070707070705, |
|
"grad_norm": 0.1630859375, |
|
"learning_rate": 6.92929292929293e-06, |
|
"loss": 1.1425, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.3090909090909091, |
|
"grad_norm": 0.1728515625, |
|
"learning_rate": 6.90909090909091e-06, |
|
"loss": 1.0966, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.3111111111111111, |
|
"grad_norm": 0.25, |
|
"learning_rate": 6.88888888888889e-06, |
|
"loss": 1.0981, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.31313131313131315, |
|
"grad_norm": 0.1650390625, |
|
"learning_rate": 6.868686868686869e-06, |
|
"loss": 1.0782, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.3151515151515151, |
|
"grad_norm": 0.1875, |
|
"learning_rate": 6.848484848484849e-06, |
|
"loss": 1.1391, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.31717171717171716, |
|
"grad_norm": 0.1689453125, |
|
"learning_rate": 6.828282828282828e-06, |
|
"loss": 1.0521, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.3191919191919192, |
|
"grad_norm": 0.2119140625, |
|
"learning_rate": 6.808080808080809e-06, |
|
"loss": 1.0959, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.3212121212121212, |
|
"grad_norm": 0.1669921875, |
|
"learning_rate": 6.787878787878789e-06, |
|
"loss": 1.0973, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.32323232323232326, |
|
"grad_norm": 0.1630859375, |
|
"learning_rate": 6.767676767676769e-06, |
|
"loss": 1.0526, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.32525252525252524, |
|
"grad_norm": 0.201171875, |
|
"learning_rate": 6.747474747474749e-06, |
|
"loss": 1.0886, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.32727272727272727, |
|
"grad_norm": 0.1650390625, |
|
"learning_rate": 6.7272727272727275e-06, |
|
"loss": 1.0519, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.3292929292929293, |
|
"grad_norm": 0.224609375, |
|
"learning_rate": 6.707070707070707e-06, |
|
"loss": 1.0623, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.33131313131313134, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 6.686868686868687e-06, |
|
"loss": 1.0819, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.3333333333333333, |
|
"grad_norm": 0.205078125, |
|
"learning_rate": 6.666666666666667e-06, |
|
"loss": 1.1033, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.33535353535353535, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 6.646464646464646e-06, |
|
"loss": 1.108, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.3373737373737374, |
|
"grad_norm": 0.2421875, |
|
"learning_rate": 6.626262626262627e-06, |
|
"loss": 1.0995, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.3393939393939394, |
|
"grad_norm": 0.1630859375, |
|
"learning_rate": 6.606060606060607e-06, |
|
"loss": 1.066, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.3414141414141414, |
|
"grad_norm": 0.2197265625, |
|
"learning_rate": 6.585858585858586e-06, |
|
"loss": 1.113, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.3434343434343434, |
|
"grad_norm": 0.1923828125, |
|
"learning_rate": 6.565656565656566e-06, |
|
"loss": 1.1046, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.34545454545454546, |
|
"grad_norm": 0.1669921875, |
|
"learning_rate": 6.545454545454546e-06, |
|
"loss": 1.129, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.3474747474747475, |
|
"grad_norm": 0.1845703125, |
|
"learning_rate": 6.525252525252526e-06, |
|
"loss": 1.0712, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.34949494949494947, |
|
"grad_norm": 0.166015625, |
|
"learning_rate": 6.505050505050505e-06, |
|
"loss": 1.0549, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.3515151515151515, |
|
"grad_norm": 0.1953125, |
|
"learning_rate": 6.484848484848485e-06, |
|
"loss": 1.0983, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.35353535353535354, |
|
"grad_norm": 0.1689453125, |
|
"learning_rate": 6.464646464646466e-06, |
|
"loss": 1.1056, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.35555555555555557, |
|
"grad_norm": 0.177734375, |
|
"learning_rate": 6.444444444444445e-06, |
|
"loss": 1.1125, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.3575757575757576, |
|
"grad_norm": 0.1708984375, |
|
"learning_rate": 6.424242424242425e-06, |
|
"loss": 1.143, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.3595959595959596, |
|
"grad_norm": 0.265625, |
|
"learning_rate": 6.404040404040405e-06, |
|
"loss": 1.0382, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.3616161616161616, |
|
"grad_norm": 0.1767578125, |
|
"learning_rate": 6.3838383838383845e-06, |
|
"loss": 1.1309, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.36363636363636365, |
|
"grad_norm": 0.27734375, |
|
"learning_rate": 6.363636363636364e-06, |
|
"loss": 1.0753, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.3656565656565657, |
|
"grad_norm": 0.1728515625, |
|
"learning_rate": 6.343434343434344e-06, |
|
"loss": 1.0436, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.36767676767676766, |
|
"grad_norm": 0.189453125, |
|
"learning_rate": 6.323232323232324e-06, |
|
"loss": 1.0809, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.3696969696969697, |
|
"grad_norm": 0.1767578125, |
|
"learning_rate": 6.303030303030303e-06, |
|
"loss": 1.0613, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.3717171717171717, |
|
"grad_norm": 0.1826171875, |
|
"learning_rate": 6.282828282828284e-06, |
|
"loss": 1.0975, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.37373737373737376, |
|
"grad_norm": 0.1708984375, |
|
"learning_rate": 6.262626262626264e-06, |
|
"loss": 1.0774, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.37575757575757573, |
|
"grad_norm": 0.1884765625, |
|
"learning_rate": 6.2424242424242434e-06, |
|
"loss": 1.0811, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.37777777777777777, |
|
"grad_norm": 0.1845703125, |
|
"learning_rate": 6.222222222222223e-06, |
|
"loss": 1.1261, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.3797979797979798, |
|
"grad_norm": 0.1943359375, |
|
"learning_rate": 6.202020202020203e-06, |
|
"loss": 1.0866, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.38181818181818183, |
|
"grad_norm": 0.1728515625, |
|
"learning_rate": 6.181818181818182e-06, |
|
"loss": 1.0776, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.3838383838383838, |
|
"grad_norm": 0.205078125, |
|
"learning_rate": 6.1616161616161615e-06, |
|
"loss": 1.0764, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.38585858585858585, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 6.141414141414141e-06, |
|
"loss": 1.0798, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.3878787878787879, |
|
"grad_norm": 0.1806640625, |
|
"learning_rate": 6.121212121212121e-06, |
|
"loss": 0.987, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.3898989898989899, |
|
"grad_norm": 0.1923828125, |
|
"learning_rate": 6.1010101010101015e-06, |
|
"loss": 1.068, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.39191919191919194, |
|
"grad_norm": 0.1640625, |
|
"learning_rate": 6.080808080808081e-06, |
|
"loss": 1.0608, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.3939393939393939, |
|
"grad_norm": 0.1845703125, |
|
"learning_rate": 6.060606060606061e-06, |
|
"loss": 1.0785, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.39595959595959596, |
|
"grad_norm": 0.193359375, |
|
"learning_rate": 6.040404040404041e-06, |
|
"loss": 1.1016, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.397979797979798, |
|
"grad_norm": 0.169921875, |
|
"learning_rate": 6.0202020202020204e-06, |
|
"loss": 1.0731, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 1.75, |
|
"learning_rate": 6e-06, |
|
"loss": 1.0454, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.402020202020202, |
|
"grad_norm": 0.1865234375, |
|
"learning_rate": 5.97979797979798e-06, |
|
"loss": 1.0912, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.40404040404040403, |
|
"grad_norm": 0.1689453125, |
|
"learning_rate": 5.95959595959596e-06, |
|
"loss": 1.0457, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.40606060606060607, |
|
"grad_norm": 0.166015625, |
|
"learning_rate": 5.93939393939394e-06, |
|
"loss": 1.0382, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.4080808080808081, |
|
"grad_norm": 0.201171875, |
|
"learning_rate": 5.91919191919192e-06, |
|
"loss": 1.1034, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.4101010101010101, |
|
"grad_norm": 0.236328125, |
|
"learning_rate": 5.8989898989899e-06, |
|
"loss": 1.0624, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.4121212121212121, |
|
"grad_norm": 0.181640625, |
|
"learning_rate": 5.878787878787879e-06, |
|
"loss": 1.088, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.41414141414141414, |
|
"grad_norm": 0.1982421875, |
|
"learning_rate": 5.858585858585859e-06, |
|
"loss": 1.0434, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.4161616161616162, |
|
"grad_norm": 0.1728515625, |
|
"learning_rate": 5.838383838383839e-06, |
|
"loss": 1.0536, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.41818181818181815, |
|
"grad_norm": 0.2216796875, |
|
"learning_rate": 5.8181818181818185e-06, |
|
"loss": 1.0236, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.4202020202020202, |
|
"grad_norm": 0.1904296875, |
|
"learning_rate": 5.797979797979798e-06, |
|
"loss": 1.013, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.4222222222222222, |
|
"grad_norm": 0.1787109375, |
|
"learning_rate": 5.777777777777778e-06, |
|
"loss": 1.0562, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.42424242424242425, |
|
"grad_norm": 0.1806640625, |
|
"learning_rate": 5.7575757575757586e-06, |
|
"loss": 1.0232, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.4262626262626263, |
|
"grad_norm": 0.1787109375, |
|
"learning_rate": 5.737373737373738e-06, |
|
"loss": 1.0425, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.42828282828282827, |
|
"grad_norm": 0.173828125, |
|
"learning_rate": 5.717171717171718e-06, |
|
"loss": 1.0381, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.4303030303030303, |
|
"grad_norm": 0.197265625, |
|
"learning_rate": 5.696969696969698e-06, |
|
"loss": 1.0393, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.43232323232323233, |
|
"grad_norm": 0.26171875, |
|
"learning_rate": 5.6767676767676775e-06, |
|
"loss": 1.0934, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.43434343434343436, |
|
"grad_norm": 0.1708984375, |
|
"learning_rate": 5.656565656565657e-06, |
|
"loss": 1.0685, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.43636363636363634, |
|
"grad_norm": 0.236328125, |
|
"learning_rate": 5.636363636363636e-06, |
|
"loss": 1.0601, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.4383838383838384, |
|
"grad_norm": 0.1748046875, |
|
"learning_rate": 5.616161616161616e-06, |
|
"loss": 1.0923, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.4404040404040404, |
|
"grad_norm": 0.1845703125, |
|
"learning_rate": 5.595959595959597e-06, |
|
"loss": 1.0975, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.44242424242424244, |
|
"grad_norm": 0.25390625, |
|
"learning_rate": 5.575757575757577e-06, |
|
"loss": 1.098, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.4444444444444444, |
|
"grad_norm": 0.1943359375, |
|
"learning_rate": 5.555555555555557e-06, |
|
"loss": 1.0009, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.44646464646464645, |
|
"grad_norm": 0.17578125, |
|
"learning_rate": 5.5353535353535355e-06, |
|
"loss": 1.0145, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.4484848484848485, |
|
"grad_norm": 0.17578125, |
|
"learning_rate": 5.515151515151515e-06, |
|
"loss": 1.0652, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.4505050505050505, |
|
"grad_norm": 0.173828125, |
|
"learning_rate": 5.494949494949495e-06, |
|
"loss": 1.0448, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.45252525252525255, |
|
"grad_norm": 0.1728515625, |
|
"learning_rate": 5.474747474747475e-06, |
|
"loss": 1.0548, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.45454545454545453, |
|
"grad_norm": 0.2060546875, |
|
"learning_rate": 5.4545454545454545e-06, |
|
"loss": 1.0661, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.45656565656565656, |
|
"grad_norm": 0.1826171875, |
|
"learning_rate": 5.434343434343434e-06, |
|
"loss": 1.0833, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.4585858585858586, |
|
"grad_norm": 0.1787109375, |
|
"learning_rate": 5.414141414141415e-06, |
|
"loss": 1.0515, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.46060606060606063, |
|
"grad_norm": 0.1708984375, |
|
"learning_rate": 5.3939393939393945e-06, |
|
"loss": 1.0603, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.4626262626262626, |
|
"grad_norm": 0.1923828125, |
|
"learning_rate": 5.373737373737374e-06, |
|
"loss": 1.073, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.46464646464646464, |
|
"grad_norm": 0.1826171875, |
|
"learning_rate": 5.353535353535354e-06, |
|
"loss": 1.0903, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.4666666666666667, |
|
"grad_norm": 0.1845703125, |
|
"learning_rate": 5.333333333333334e-06, |
|
"loss": 1.054, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.4686868686868687, |
|
"grad_norm": 0.1787109375, |
|
"learning_rate": 5.313131313131313e-06, |
|
"loss": 1.0634, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.4707070707070707, |
|
"grad_norm": 0.2275390625, |
|
"learning_rate": 5.292929292929293e-06, |
|
"loss": 1.0463, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.4727272727272727, |
|
"grad_norm": 0.1884765625, |
|
"learning_rate": 5.272727272727273e-06, |
|
"loss": 1.0138, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.47474747474747475, |
|
"grad_norm": 0.17578125, |
|
"learning_rate": 5.252525252525253e-06, |
|
"loss": 0.9779, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.4767676767676768, |
|
"grad_norm": 0.1943359375, |
|
"learning_rate": 5.232323232323233e-06, |
|
"loss": 1.112, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.47878787878787876, |
|
"grad_norm": 0.21875, |
|
"learning_rate": 5.212121212121213e-06, |
|
"loss": 1.079, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.4808080808080808, |
|
"grad_norm": 0.1826171875, |
|
"learning_rate": 5.191919191919193e-06, |
|
"loss": 1.0357, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.48282828282828283, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 5.171717171717172e-06, |
|
"loss": 1.0545, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.48484848484848486, |
|
"grad_norm": 0.1787109375, |
|
"learning_rate": 5.151515151515152e-06, |
|
"loss": 1.0452, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.4868686868686869, |
|
"grad_norm": 0.181640625, |
|
"learning_rate": 5.131313131313132e-06, |
|
"loss": 1.0302, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.4888888888888889, |
|
"grad_norm": 0.1787109375, |
|
"learning_rate": 5.1111111111111115e-06, |
|
"loss": 1.03, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.4909090909090909, |
|
"grad_norm": 0.189453125, |
|
"learning_rate": 5.090909090909091e-06, |
|
"loss": 0.9899, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.49292929292929294, |
|
"grad_norm": 0.2080078125, |
|
"learning_rate": 5.070707070707072e-06, |
|
"loss": 1.0448, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.494949494949495, |
|
"grad_norm": 0.203125, |
|
"learning_rate": 5.0505050505050515e-06, |
|
"loss": 1.0938, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.49696969696969695, |
|
"grad_norm": 0.17578125, |
|
"learning_rate": 5.030303030303031e-06, |
|
"loss": 1.0318, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.498989898989899, |
|
"grad_norm": 0.1689453125, |
|
"learning_rate": 5.010101010101011e-06, |
|
"loss": 0.9649, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.501010101010101, |
|
"grad_norm": 0.1865234375, |
|
"learning_rate": 4.98989898989899e-06, |
|
"loss": 1.0887, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.503030303030303, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 4.9696969696969696e-06, |
|
"loss": 1.0543, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.5050505050505051, |
|
"grad_norm": 0.18359375, |
|
"learning_rate": 4.94949494949495e-06, |
|
"loss": 1.0052, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.5070707070707071, |
|
"grad_norm": 0.1904296875, |
|
"learning_rate": 4.92929292929293e-06, |
|
"loss": 1.0364, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.509090909090909, |
|
"grad_norm": 0.1796875, |
|
"learning_rate": 4.90909090909091e-06, |
|
"loss": 1.0756, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.5111111111111111, |
|
"grad_norm": 0.2041015625, |
|
"learning_rate": 4.888888888888889e-06, |
|
"loss": 1.0512, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.5131313131313131, |
|
"grad_norm": 0.201171875, |
|
"learning_rate": 4.868686868686869e-06, |
|
"loss": 1.0372, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.5151515151515151, |
|
"grad_norm": 0.28125, |
|
"learning_rate": 4.848484848484849e-06, |
|
"loss": 1.1119, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.5171717171717172, |
|
"grad_norm": 0.181640625, |
|
"learning_rate": 4.8282828282828285e-06, |
|
"loss": 1.077, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.5191919191919192, |
|
"grad_norm": 0.19140625, |
|
"learning_rate": 4.808080808080808e-06, |
|
"loss": 1.1184, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.5212121212121212, |
|
"grad_norm": 0.208984375, |
|
"learning_rate": 4.787878787878788e-06, |
|
"loss": 1.0886, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.5232323232323233, |
|
"grad_norm": 0.1875, |
|
"learning_rate": 4.7676767676767685e-06, |
|
"loss": 1.0095, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.5252525252525253, |
|
"grad_norm": 0.205078125, |
|
"learning_rate": 4.747474747474748e-06, |
|
"loss": 1.045, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.5272727272727272, |
|
"grad_norm": 0.1904296875, |
|
"learning_rate": 4.727272727272728e-06, |
|
"loss": 1.0381, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.5292929292929293, |
|
"grad_norm": 0.1826171875, |
|
"learning_rate": 4.707070707070707e-06, |
|
"loss": 1.0379, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.5313131313131313, |
|
"grad_norm": 0.1748046875, |
|
"learning_rate": 4.6868686868686874e-06, |
|
"loss": 1.0157, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.5333333333333333, |
|
"grad_norm": 0.18359375, |
|
"learning_rate": 4.666666666666667e-06, |
|
"loss": 1.0467, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.5353535353535354, |
|
"grad_norm": 0.1796875, |
|
"learning_rate": 4.646464646464647e-06, |
|
"loss": 1.045, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.5373737373737374, |
|
"grad_norm": 0.185546875, |
|
"learning_rate": 4.626262626262627e-06, |
|
"loss": 1.0524, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.5393939393939394, |
|
"grad_norm": 0.265625, |
|
"learning_rate": 4.606060606060606e-06, |
|
"loss": 1.0835, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.5414141414141415, |
|
"grad_norm": 0.1962890625, |
|
"learning_rate": 4.585858585858586e-06, |
|
"loss": 1.0561, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.5434343434343434, |
|
"grad_norm": 0.1748046875, |
|
"learning_rate": 4.565656565656566e-06, |
|
"loss": 1.0416, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.5454545454545454, |
|
"grad_norm": 0.2158203125, |
|
"learning_rate": 4.5454545454545455e-06, |
|
"loss": 1.0629, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.5474747474747474, |
|
"grad_norm": 0.2099609375, |
|
"learning_rate": 4.525252525252526e-06, |
|
"loss": 1.0639, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.5494949494949495, |
|
"grad_norm": 0.1904296875, |
|
"learning_rate": 4.505050505050506e-06, |
|
"loss": 1.0732, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.5515151515151515, |
|
"grad_norm": 0.1904296875, |
|
"learning_rate": 4.4848484848484855e-06, |
|
"loss": 1.1106, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.5535353535353535, |
|
"grad_norm": 0.185546875, |
|
"learning_rate": 4.464646464646465e-06, |
|
"loss": 1.0459, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.5555555555555556, |
|
"grad_norm": 0.1943359375, |
|
"learning_rate": 4.444444444444444e-06, |
|
"loss": 1.0037, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.5575757575757576, |
|
"grad_norm": 0.197265625, |
|
"learning_rate": 4.424242424242425e-06, |
|
"loss": 1.0369, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.5595959595959596, |
|
"grad_norm": 0.234375, |
|
"learning_rate": 4.4040404040404044e-06, |
|
"loss": 1.0844, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.5616161616161616, |
|
"grad_norm": 0.18359375, |
|
"learning_rate": 4.383838383838384e-06, |
|
"loss": 1.0097, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.5636363636363636, |
|
"grad_norm": 0.1787109375, |
|
"learning_rate": 4.363636363636364e-06, |
|
"loss": 1.0342, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.5656565656565656, |
|
"grad_norm": 0.2177734375, |
|
"learning_rate": 4.343434343434344e-06, |
|
"loss": 1.0167, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.5676767676767677, |
|
"grad_norm": 0.1796875, |
|
"learning_rate": 4.323232323232323e-06, |
|
"loss": 1.0203, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.5696969696969697, |
|
"grad_norm": 0.251953125, |
|
"learning_rate": 4.303030303030303e-06, |
|
"loss": 1.0382, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.5717171717171717, |
|
"grad_norm": 0.19921875, |
|
"learning_rate": 4.282828282828283e-06, |
|
"loss": 1.0358, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.5737373737373738, |
|
"grad_norm": 0.173828125, |
|
"learning_rate": 4.262626262626263e-06, |
|
"loss": 1.072, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.5757575757575758, |
|
"grad_norm": 0.19140625, |
|
"learning_rate": 4.242424242424243e-06, |
|
"loss": 1.0884, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.5777777777777777, |
|
"grad_norm": 0.1787109375, |
|
"learning_rate": 4.222222222222223e-06, |
|
"loss": 1.032, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.5797979797979798, |
|
"grad_norm": 0.1884765625, |
|
"learning_rate": 4.2020202020202026e-06, |
|
"loss": 1.039, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.5818181818181818, |
|
"grad_norm": 0.17578125, |
|
"learning_rate": 4.181818181818182e-06, |
|
"loss": 1.0533, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.5838383838383838, |
|
"grad_norm": 0.17578125, |
|
"learning_rate": 4.161616161616162e-06, |
|
"loss": 1.027, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.5858585858585859, |
|
"grad_norm": 0.2109375, |
|
"learning_rate": 4.141414141414142e-06, |
|
"loss": 1.046, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.5878787878787879, |
|
"grad_norm": 0.17578125, |
|
"learning_rate": 4.1212121212121215e-06, |
|
"loss": 1.0367, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.5898989898989899, |
|
"grad_norm": 0.2431640625, |
|
"learning_rate": 4.101010101010101e-06, |
|
"loss": 1.0158, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.591919191919192, |
|
"grad_norm": 0.189453125, |
|
"learning_rate": 4.080808080808081e-06, |
|
"loss": 0.9974, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.593939393939394, |
|
"grad_norm": 0.1982421875, |
|
"learning_rate": 4.060606060606061e-06, |
|
"loss": 0.9762, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.5959595959595959, |
|
"grad_norm": 0.177734375, |
|
"learning_rate": 4.04040404040404e-06, |
|
"loss": 1.0425, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.597979797979798, |
|
"grad_norm": 0.1708984375, |
|
"learning_rate": 4.02020202020202e-06, |
|
"loss": 0.9974, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.17578125, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 1.0481, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.602020202020202, |
|
"grad_norm": 0.2333984375, |
|
"learning_rate": 3.97979797979798e-06, |
|
"loss": 1.0446, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.604040404040404, |
|
"grad_norm": 0.17578125, |
|
"learning_rate": 3.95959595959596e-06, |
|
"loss": 1.0086, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.6060606060606061, |
|
"grad_norm": 0.298828125, |
|
"learning_rate": 3.93939393939394e-06, |
|
"loss": 1.1283, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.6080808080808081, |
|
"grad_norm": 0.1884765625, |
|
"learning_rate": 3.9191919191919196e-06, |
|
"loss": 1.0657, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.6101010101010101, |
|
"grad_norm": 0.208984375, |
|
"learning_rate": 3.898989898989899e-06, |
|
"loss": 0.9508, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.6121212121212121, |
|
"grad_norm": 0.1923828125, |
|
"learning_rate": 3.878787878787879e-06, |
|
"loss": 1.0725, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.6141414141414141, |
|
"grad_norm": 0.1953125, |
|
"learning_rate": 3.858585858585859e-06, |
|
"loss": 1.0317, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.6161616161616161, |
|
"grad_norm": 0.1826171875, |
|
"learning_rate": 3.8383838383838385e-06, |
|
"loss": 1.0252, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.6181818181818182, |
|
"grad_norm": 0.2001953125, |
|
"learning_rate": 3.818181818181819e-06, |
|
"loss": 0.9875, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.6202020202020202, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 3.7979797979797984e-06, |
|
"loss": 1.0311, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.6222222222222222, |
|
"grad_norm": 0.18359375, |
|
"learning_rate": 3.777777777777778e-06, |
|
"loss": 1.0357, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.6242424242424243, |
|
"grad_norm": 0.2060546875, |
|
"learning_rate": 3.757575757575758e-06, |
|
"loss": 1.019, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.6262626262626263, |
|
"grad_norm": 0.181640625, |
|
"learning_rate": 3.737373737373738e-06, |
|
"loss": 1.0169, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.6282828282828283, |
|
"grad_norm": 0.1845703125, |
|
"learning_rate": 3.7171717171717177e-06, |
|
"loss": 1.0063, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.6303030303030303, |
|
"grad_norm": 0.2392578125, |
|
"learning_rate": 3.6969696969696974e-06, |
|
"loss": 1.0391, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.6323232323232323, |
|
"grad_norm": 0.1943359375, |
|
"learning_rate": 3.6767676767676767e-06, |
|
"loss": 1.0459, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.6343434343434343, |
|
"grad_norm": 0.1806640625, |
|
"learning_rate": 3.6565656565656573e-06, |
|
"loss": 1.0434, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.6363636363636364, |
|
"grad_norm": 0.185546875, |
|
"learning_rate": 3.6363636363636366e-06, |
|
"loss": 0.9825, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.6383838383838384, |
|
"grad_norm": 0.1904296875, |
|
"learning_rate": 3.6161616161616163e-06, |
|
"loss": 1.0962, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.6404040404040404, |
|
"grad_norm": 0.240234375, |
|
"learning_rate": 3.595959595959596e-06, |
|
"loss": 1.0104, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.6424242424242425, |
|
"grad_norm": 0.1767578125, |
|
"learning_rate": 3.575757575757576e-06, |
|
"loss": 0.9901, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.6444444444444445, |
|
"grad_norm": 0.18359375, |
|
"learning_rate": 3.555555555555556e-06, |
|
"loss": 1.0571, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.6464646464646465, |
|
"grad_norm": 0.1962890625, |
|
"learning_rate": 3.5353535353535356e-06, |
|
"loss": 0.959, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.6484848484848484, |
|
"grad_norm": 0.1884765625, |
|
"learning_rate": 3.5151515151515154e-06, |
|
"loss": 1.0403, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.6505050505050505, |
|
"grad_norm": 0.1923828125, |
|
"learning_rate": 3.494949494949495e-06, |
|
"loss": 1.0236, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.6525252525252525, |
|
"grad_norm": 0.1923828125, |
|
"learning_rate": 3.4747474747474752e-06, |
|
"loss": 1.0587, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.6545454545454545, |
|
"grad_norm": 0.1962890625, |
|
"learning_rate": 3.454545454545455e-06, |
|
"loss": 1.0402, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.6565656565656566, |
|
"grad_norm": 0.18359375, |
|
"learning_rate": 3.4343434343434347e-06, |
|
"loss": 1.0342, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.6585858585858586, |
|
"grad_norm": 0.205078125, |
|
"learning_rate": 3.414141414141414e-06, |
|
"loss": 1.0657, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.6606060606060606, |
|
"grad_norm": 0.1806640625, |
|
"learning_rate": 3.3939393939393946e-06, |
|
"loss": 1.0749, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.6626262626262627, |
|
"grad_norm": 0.181640625, |
|
"learning_rate": 3.3737373737373743e-06, |
|
"loss": 0.9923, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.6646464646464646, |
|
"grad_norm": 0.1708984375, |
|
"learning_rate": 3.3535353535353536e-06, |
|
"loss": 1.0401, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.6666666666666666, |
|
"grad_norm": 0.18359375, |
|
"learning_rate": 3.3333333333333333e-06, |
|
"loss": 1.0066, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.6686868686868687, |
|
"grad_norm": 0.19921875, |
|
"learning_rate": 3.3131313131313135e-06, |
|
"loss": 1.0443, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.6707070707070707, |
|
"grad_norm": 0.287109375, |
|
"learning_rate": 3.292929292929293e-06, |
|
"loss": 1.0418, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.6727272727272727, |
|
"grad_norm": 0.17578125, |
|
"learning_rate": 3.272727272727273e-06, |
|
"loss": 1.0639, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.6747474747474748, |
|
"grad_norm": 0.2158203125, |
|
"learning_rate": 3.2525252525252527e-06, |
|
"loss": 0.9951, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.6767676767676768, |
|
"grad_norm": 0.2431640625, |
|
"learning_rate": 3.232323232323233e-06, |
|
"loss": 1.0334, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.6787878787878788, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 3.2121212121212125e-06, |
|
"loss": 1.0239, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.6808080808080809, |
|
"grad_norm": 0.26171875, |
|
"learning_rate": 3.1919191919191923e-06, |
|
"loss": 1.0718, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.6828282828282828, |
|
"grad_norm": 0.1982421875, |
|
"learning_rate": 3.171717171717172e-06, |
|
"loss": 1.0556, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.6848484848484848, |
|
"grad_norm": 0.19140625, |
|
"learning_rate": 3.1515151515151517e-06, |
|
"loss": 0.9692, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.6868686868686869, |
|
"grad_norm": 0.1943359375, |
|
"learning_rate": 3.131313131313132e-06, |
|
"loss": 1.0153, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.6888888888888889, |
|
"grad_norm": 0.201171875, |
|
"learning_rate": 3.1111111111111116e-06, |
|
"loss": 1.0404, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.6909090909090909, |
|
"grad_norm": 0.2001953125, |
|
"learning_rate": 3.090909090909091e-06, |
|
"loss": 0.9807, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.692929292929293, |
|
"grad_norm": 0.185546875, |
|
"learning_rate": 3.0707070707070706e-06, |
|
"loss": 1.0399, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.694949494949495, |
|
"grad_norm": 0.1953125, |
|
"learning_rate": 3.0505050505050508e-06, |
|
"loss": 1.0141, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.696969696969697, |
|
"grad_norm": 0.193359375, |
|
"learning_rate": 3.0303030303030305e-06, |
|
"loss": 1.0309, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.6989898989898989, |
|
"grad_norm": 0.1904296875, |
|
"learning_rate": 3.0101010101010102e-06, |
|
"loss": 1.0634, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.701010101010101, |
|
"grad_norm": 0.1923828125, |
|
"learning_rate": 2.98989898989899e-06, |
|
"loss": 1.0344, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.703030303030303, |
|
"grad_norm": 0.1826171875, |
|
"learning_rate": 2.96969696969697e-06, |
|
"loss": 1.0023, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.705050505050505, |
|
"grad_norm": 0.19921875, |
|
"learning_rate": 2.94949494949495e-06, |
|
"loss": 1.0559, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.7070707070707071, |
|
"grad_norm": 0.1982421875, |
|
"learning_rate": 2.9292929292929295e-06, |
|
"loss": 0.9863, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.7090909090909091, |
|
"grad_norm": 0.1845703125, |
|
"learning_rate": 2.9090909090909093e-06, |
|
"loss": 1.0229, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.7111111111111111, |
|
"grad_norm": 0.20703125, |
|
"learning_rate": 2.888888888888889e-06, |
|
"loss": 1.044, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.7131313131313132, |
|
"grad_norm": 0.1923828125, |
|
"learning_rate": 2.868686868686869e-06, |
|
"loss": 1.0886, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.7151515151515152, |
|
"grad_norm": 0.2177734375, |
|
"learning_rate": 2.848484848484849e-06, |
|
"loss": 1.0104, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.7171717171717171, |
|
"grad_norm": 0.2197265625, |
|
"learning_rate": 2.8282828282828286e-06, |
|
"loss": 1.0258, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.7191919191919192, |
|
"grad_norm": 0.1845703125, |
|
"learning_rate": 2.808080808080808e-06, |
|
"loss": 1.0074, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.7212121212121212, |
|
"grad_norm": 0.18359375, |
|
"learning_rate": 2.7878787878787885e-06, |
|
"loss": 1.0737, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.7232323232323232, |
|
"grad_norm": 0.197265625, |
|
"learning_rate": 2.7676767676767678e-06, |
|
"loss": 1.0597, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.7252525252525253, |
|
"grad_norm": 0.1875, |
|
"learning_rate": 2.7474747474747475e-06, |
|
"loss": 1.0295, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.7272727272727273, |
|
"grad_norm": 0.1826171875, |
|
"learning_rate": 2.7272727272727272e-06, |
|
"loss": 1.0305, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.7292929292929293, |
|
"grad_norm": 0.1845703125, |
|
"learning_rate": 2.7070707070707074e-06, |
|
"loss": 1.0508, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.7313131313131314, |
|
"grad_norm": 0.1953125, |
|
"learning_rate": 2.686868686868687e-06, |
|
"loss": 1.008, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.7333333333333333, |
|
"grad_norm": 0.1806640625, |
|
"learning_rate": 2.666666666666667e-06, |
|
"loss": 0.9775, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.7353535353535353, |
|
"grad_norm": 0.1953125, |
|
"learning_rate": 2.6464646464646466e-06, |
|
"loss": 1.033, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.7373737373737373, |
|
"grad_norm": 0.2099609375, |
|
"learning_rate": 2.6262626262626267e-06, |
|
"loss": 1.0351, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.7393939393939394, |
|
"grad_norm": 0.166015625, |
|
"learning_rate": 2.6060606060606064e-06, |
|
"loss": 0.9793, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.7414141414141414, |
|
"grad_norm": 0.2060546875, |
|
"learning_rate": 2.585858585858586e-06, |
|
"loss": 1.0528, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.7434343434343434, |
|
"grad_norm": 0.2099609375, |
|
"learning_rate": 2.565656565656566e-06, |
|
"loss": 1.03, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.7454545454545455, |
|
"grad_norm": 0.1826171875, |
|
"learning_rate": 2.5454545454545456e-06, |
|
"loss": 1.0614, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.7474747474747475, |
|
"grad_norm": 0.181640625, |
|
"learning_rate": 2.5252525252525258e-06, |
|
"loss": 1.0243, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.7494949494949495, |
|
"grad_norm": 0.2060546875, |
|
"learning_rate": 2.5050505050505055e-06, |
|
"loss": 1.0558, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.7515151515151515, |
|
"grad_norm": 0.18359375, |
|
"learning_rate": 2.4848484848484848e-06, |
|
"loss": 1.0429, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.7535353535353535, |
|
"grad_norm": 0.21484375, |
|
"learning_rate": 2.464646464646465e-06, |
|
"loss": 1.1272, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.7555555555555555, |
|
"grad_norm": 0.28515625, |
|
"learning_rate": 2.4444444444444447e-06, |
|
"loss": 1.0403, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.7575757575757576, |
|
"grad_norm": 0.1943359375, |
|
"learning_rate": 2.4242424242424244e-06, |
|
"loss": 1.0417, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.7595959595959596, |
|
"grad_norm": 0.177734375, |
|
"learning_rate": 2.404040404040404e-06, |
|
"loss": 1.0172, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.7616161616161616, |
|
"grad_norm": 0.18359375, |
|
"learning_rate": 2.3838383838383843e-06, |
|
"loss": 1.0566, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.7636363636363637, |
|
"grad_norm": 0.234375, |
|
"learning_rate": 2.363636363636364e-06, |
|
"loss": 1.0459, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.7656565656565657, |
|
"grad_norm": 0.17578125, |
|
"learning_rate": 2.3434343434343437e-06, |
|
"loss": 1.0115, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.7676767676767676, |
|
"grad_norm": 0.197265625, |
|
"learning_rate": 2.3232323232323234e-06, |
|
"loss": 1.0139, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.7696969696969697, |
|
"grad_norm": 0.1865234375, |
|
"learning_rate": 2.303030303030303e-06, |
|
"loss": 1.0304, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.7717171717171717, |
|
"grad_norm": 0.2470703125, |
|
"learning_rate": 2.282828282828283e-06, |
|
"loss": 0.9806, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.7737373737373737, |
|
"grad_norm": 0.21484375, |
|
"learning_rate": 2.262626262626263e-06, |
|
"loss": 1.0062, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.7757575757575758, |
|
"grad_norm": 0.1875, |
|
"learning_rate": 2.2424242424242428e-06, |
|
"loss": 1.0253, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.7777777777777778, |
|
"grad_norm": 0.2109375, |
|
"learning_rate": 2.222222222222222e-06, |
|
"loss": 1.0317, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.7797979797979798, |
|
"grad_norm": 0.18359375, |
|
"learning_rate": 2.2020202020202022e-06, |
|
"loss": 1.0146, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.7818181818181819, |
|
"grad_norm": 0.2001953125, |
|
"learning_rate": 2.181818181818182e-06, |
|
"loss": 0.9777, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.7838383838383839, |
|
"grad_norm": 0.25390625, |
|
"learning_rate": 2.1616161616161617e-06, |
|
"loss": 1.0182, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.7858585858585858, |
|
"grad_norm": 0.2060546875, |
|
"learning_rate": 2.1414141414141414e-06, |
|
"loss": 1.0215, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.7878787878787878, |
|
"grad_norm": 0.1904296875, |
|
"learning_rate": 2.1212121212121216e-06, |
|
"loss": 1.0146, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.7898989898989899, |
|
"grad_norm": 0.181640625, |
|
"learning_rate": 2.1010101010101013e-06, |
|
"loss": 1.002, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.7919191919191919, |
|
"grad_norm": 0.1748046875, |
|
"learning_rate": 2.080808080808081e-06, |
|
"loss": 1.0192, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.793939393939394, |
|
"grad_norm": 0.1806640625, |
|
"learning_rate": 2.0606060606060607e-06, |
|
"loss": 0.9827, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.795959595959596, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 2.0404040404040405e-06, |
|
"loss": 1.0718, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.797979797979798, |
|
"grad_norm": 0.1884765625, |
|
"learning_rate": 2.02020202020202e-06, |
|
"loss": 1.0095, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.21875, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 1.0553, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.802020202020202, |
|
"grad_norm": 0.2412109375, |
|
"learning_rate": 1.97979797979798e-06, |
|
"loss": 0.9963, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.804040404040404, |
|
"grad_norm": 0.18359375, |
|
"learning_rate": 1.9595959595959598e-06, |
|
"loss": 1.0107, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.806060606060606, |
|
"grad_norm": 0.25390625, |
|
"learning_rate": 1.9393939393939395e-06, |
|
"loss": 1.036, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.8080808080808081, |
|
"grad_norm": 0.1865234375, |
|
"learning_rate": 1.9191919191919192e-06, |
|
"loss": 1.0436, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.8101010101010101, |
|
"grad_norm": 0.1953125, |
|
"learning_rate": 1.8989898989898992e-06, |
|
"loss": 1.0567, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.8121212121212121, |
|
"grad_norm": 0.1806640625, |
|
"learning_rate": 1.878787878787879e-06, |
|
"loss": 0.975, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.8141414141414142, |
|
"grad_norm": 0.2158203125, |
|
"learning_rate": 1.8585858585858588e-06, |
|
"loss": 1.0012, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.8161616161616162, |
|
"grad_norm": 0.1943359375, |
|
"learning_rate": 1.8383838383838384e-06, |
|
"loss": 1.051, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.8181818181818182, |
|
"grad_norm": 0.26171875, |
|
"learning_rate": 1.8181818181818183e-06, |
|
"loss": 1.0274, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.8202020202020202, |
|
"grad_norm": 0.1982421875, |
|
"learning_rate": 1.797979797979798e-06, |
|
"loss": 0.9988, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.8222222222222222, |
|
"grad_norm": 0.2216796875, |
|
"learning_rate": 1.777777777777778e-06, |
|
"loss": 1.0229, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.8242424242424242, |
|
"grad_norm": 0.1787109375, |
|
"learning_rate": 1.7575757575757577e-06, |
|
"loss": 1.0392, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.8262626262626263, |
|
"grad_norm": 0.1796875, |
|
"learning_rate": 1.7373737373737376e-06, |
|
"loss": 1.002, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.8282828282828283, |
|
"grad_norm": 0.2099609375, |
|
"learning_rate": 1.7171717171717173e-06, |
|
"loss": 1.0471, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.8303030303030303, |
|
"grad_norm": 0.21484375, |
|
"learning_rate": 1.6969696969696973e-06, |
|
"loss": 1.0112, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.8323232323232324, |
|
"grad_norm": 0.2001953125, |
|
"learning_rate": 1.6767676767676768e-06, |
|
"loss": 1.0887, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.8343434343434344, |
|
"grad_norm": 0.181640625, |
|
"learning_rate": 1.6565656565656567e-06, |
|
"loss": 1.014, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.8363636363636363, |
|
"grad_norm": 0.25390625, |
|
"learning_rate": 1.6363636363636365e-06, |
|
"loss": 0.9878, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.8383838383838383, |
|
"grad_norm": 0.1796875, |
|
"learning_rate": 1.6161616161616164e-06, |
|
"loss": 0.9874, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.8404040404040404, |
|
"grad_norm": 0.224609375, |
|
"learning_rate": 1.5959595959595961e-06, |
|
"loss": 1.0532, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.8424242424242424, |
|
"grad_norm": 0.1826171875, |
|
"learning_rate": 1.5757575757575759e-06, |
|
"loss": 0.9955, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.8444444444444444, |
|
"grad_norm": 0.1748046875, |
|
"learning_rate": 1.5555555555555558e-06, |
|
"loss": 1.0148, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.8464646464646465, |
|
"grad_norm": 0.197265625, |
|
"learning_rate": 1.5353535353535353e-06, |
|
"loss": 1.0331, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.8484848484848485, |
|
"grad_norm": 0.1748046875, |
|
"learning_rate": 1.5151515151515152e-06, |
|
"loss": 1.0199, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.8505050505050505, |
|
"grad_norm": 0.185546875, |
|
"learning_rate": 1.494949494949495e-06, |
|
"loss": 0.9248, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.8525252525252526, |
|
"grad_norm": 0.18359375, |
|
"learning_rate": 1.474747474747475e-06, |
|
"loss": 1.0306, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.8545454545454545, |
|
"grad_norm": 0.1748046875, |
|
"learning_rate": 1.4545454545454546e-06, |
|
"loss": 1.0285, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.8565656565656565, |
|
"grad_norm": 0.2060546875, |
|
"learning_rate": 1.4343434343434346e-06, |
|
"loss": 1.0151, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.8585858585858586, |
|
"grad_norm": 0.1845703125, |
|
"learning_rate": 1.4141414141414143e-06, |
|
"loss": 1.0575, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.8606060606060606, |
|
"grad_norm": 0.197265625, |
|
"learning_rate": 1.3939393939393942e-06, |
|
"loss": 1.0269, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.8626262626262626, |
|
"grad_norm": 0.2216796875, |
|
"learning_rate": 1.3737373737373738e-06, |
|
"loss": 1.0597, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.8646464646464647, |
|
"grad_norm": 0.224609375, |
|
"learning_rate": 1.3535353535353537e-06, |
|
"loss": 0.997, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.8666666666666667, |
|
"grad_norm": 0.208984375, |
|
"learning_rate": 1.3333333333333334e-06, |
|
"loss": 1.0348, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.8686868686868687, |
|
"grad_norm": 0.1787109375, |
|
"learning_rate": 1.3131313131313134e-06, |
|
"loss": 0.978, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.8707070707070707, |
|
"grad_norm": 0.2470703125, |
|
"learning_rate": 1.292929292929293e-06, |
|
"loss": 1.0392, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.8727272727272727, |
|
"grad_norm": 0.1796875, |
|
"learning_rate": 1.2727272727272728e-06, |
|
"loss": 1.003, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.8747474747474747, |
|
"grad_norm": 0.189453125, |
|
"learning_rate": 1.2525252525252527e-06, |
|
"loss": 1.0488, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.8767676767676768, |
|
"grad_norm": 0.1875, |
|
"learning_rate": 1.2323232323232325e-06, |
|
"loss": 1.0134, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.8787878787878788, |
|
"grad_norm": 0.1943359375, |
|
"learning_rate": 1.2121212121212122e-06, |
|
"loss": 1.0099, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.8808080808080808, |
|
"grad_norm": 0.181640625, |
|
"learning_rate": 1.1919191919191921e-06, |
|
"loss": 1.0093, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.8828282828282829, |
|
"grad_norm": 0.197265625, |
|
"learning_rate": 1.1717171717171719e-06, |
|
"loss": 1.041, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.8848484848484849, |
|
"grad_norm": 0.1904296875, |
|
"learning_rate": 1.1515151515151516e-06, |
|
"loss": 1.0555, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.8868686868686869, |
|
"grad_norm": 0.1962890625, |
|
"learning_rate": 1.1313131313131315e-06, |
|
"loss": 0.996, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.8888888888888888, |
|
"grad_norm": 0.26953125, |
|
"learning_rate": 1.111111111111111e-06, |
|
"loss": 1.0446, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.8909090909090909, |
|
"grad_norm": 0.279296875, |
|
"learning_rate": 1.090909090909091e-06, |
|
"loss": 1.0465, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.8929292929292929, |
|
"grad_norm": 0.2294921875, |
|
"learning_rate": 1.0707070707070707e-06, |
|
"loss": 1.0364, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.8949494949494949, |
|
"grad_norm": 0.185546875, |
|
"learning_rate": 1.0505050505050506e-06, |
|
"loss": 1.0254, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.896969696969697, |
|
"grad_norm": 0.2138671875, |
|
"learning_rate": 1.0303030303030304e-06, |
|
"loss": 0.9752, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.898989898989899, |
|
"grad_norm": 0.2255859375, |
|
"learning_rate": 1.01010101010101e-06, |
|
"loss": 1.0568, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.901010101010101, |
|
"grad_norm": 0.181640625, |
|
"learning_rate": 9.8989898989899e-07, |
|
"loss": 1.0076, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.9030303030303031, |
|
"grad_norm": 0.189453125, |
|
"learning_rate": 9.696969696969698e-07, |
|
"loss": 1.0118, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.9050505050505051, |
|
"grad_norm": 0.294921875, |
|
"learning_rate": 9.494949494949496e-07, |
|
"loss": 1.0203, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.907070707070707, |
|
"grad_norm": 0.2138671875, |
|
"learning_rate": 9.292929292929294e-07, |
|
"loss": 1.0872, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.9090909090909091, |
|
"grad_norm": 0.1796875, |
|
"learning_rate": 9.090909090909091e-07, |
|
"loss": 1.0107, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.9111111111111111, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 8.88888888888889e-07, |
|
"loss": 0.9961, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.9131313131313131, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 8.686868686868688e-07, |
|
"loss": 0.9945, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.9151515151515152, |
|
"grad_norm": 0.185546875, |
|
"learning_rate": 8.484848484848486e-07, |
|
"loss": 1.0275, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.9171717171717172, |
|
"grad_norm": 0.22265625, |
|
"learning_rate": 8.282828282828284e-07, |
|
"loss": 1.0274, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.9191919191919192, |
|
"grad_norm": 0.1826171875, |
|
"learning_rate": 8.080808080808082e-07, |
|
"loss": 1.0346, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.9212121212121213, |
|
"grad_norm": 0.2314453125, |
|
"learning_rate": 7.878787878787879e-07, |
|
"loss": 1.098, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.9232323232323232, |
|
"grad_norm": 0.28125, |
|
"learning_rate": 7.676767676767677e-07, |
|
"loss": 0.9775, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.9252525252525252, |
|
"grad_norm": 0.1923828125, |
|
"learning_rate": 7.474747474747475e-07, |
|
"loss": 1.0375, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.9272727272727272, |
|
"grad_norm": 0.57421875, |
|
"learning_rate": 7.272727272727273e-07, |
|
"loss": 1.0435, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.9292929292929293, |
|
"grad_norm": 0.2099609375, |
|
"learning_rate": 7.070707070707071e-07, |
|
"loss": 0.9988, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.9313131313131313, |
|
"grad_norm": 0.2216796875, |
|
"learning_rate": 6.868686868686869e-07, |
|
"loss": 1.0125, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.9333333333333333, |
|
"grad_norm": 0.18359375, |
|
"learning_rate": 6.666666666666667e-07, |
|
"loss": 1.0182, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.9353535353535354, |
|
"grad_norm": 0.2041015625, |
|
"learning_rate": 6.464646464646465e-07, |
|
"loss": 0.9986, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.9373737373737374, |
|
"grad_norm": 0.1953125, |
|
"learning_rate": 6.262626262626264e-07, |
|
"loss": 1.0093, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.9393939393939394, |
|
"grad_norm": 0.189453125, |
|
"learning_rate": 6.060606060606061e-07, |
|
"loss": 0.9871, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.9414141414141414, |
|
"grad_norm": 0.255859375, |
|
"learning_rate": 5.858585858585859e-07, |
|
"loss": 1.0191, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.9434343434343434, |
|
"grad_norm": 0.265625, |
|
"learning_rate": 5.656565656565658e-07, |
|
"loss": 1.0001, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.9454545454545454, |
|
"grad_norm": 0.2001953125, |
|
"learning_rate": 5.454545454545455e-07, |
|
"loss": 0.9726, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.9474747474747475, |
|
"grad_norm": 0.1962890625, |
|
"learning_rate": 5.252525252525253e-07, |
|
"loss": 1.069, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.9494949494949495, |
|
"grad_norm": 0.2373046875, |
|
"learning_rate": 5.05050505050505e-07, |
|
"loss": 1.0334, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.9515151515151515, |
|
"grad_norm": 0.177734375, |
|
"learning_rate": 4.848484848484849e-07, |
|
"loss": 1.0442, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.9535353535353536, |
|
"grad_norm": 0.1806640625, |
|
"learning_rate": 4.646464646464647e-07, |
|
"loss": 1.0116, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.9555555555555556, |
|
"grad_norm": 0.69921875, |
|
"learning_rate": 4.444444444444445e-07, |
|
"loss": 0.9919, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.9575757575757575, |
|
"grad_norm": 0.1865234375, |
|
"learning_rate": 4.242424242424243e-07, |
|
"loss": 1.012, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.9595959595959596, |
|
"grad_norm": 0.19140625, |
|
"learning_rate": 4.040404040404041e-07, |
|
"loss": 0.9972, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.9616161616161616, |
|
"grad_norm": 0.216796875, |
|
"learning_rate": 3.838383838383838e-07, |
|
"loss": 1.0253, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.9636363636363636, |
|
"grad_norm": 0.1796875, |
|
"learning_rate": 3.6363636363636366e-07, |
|
"loss": 1.0101, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.9656565656565657, |
|
"grad_norm": 0.1796875, |
|
"learning_rate": 3.4343434343434344e-07, |
|
"loss": 0.9982, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.9676767676767677, |
|
"grad_norm": 0.1982421875, |
|
"learning_rate": 3.2323232323232327e-07, |
|
"loss": 1.0391, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.9696969696969697, |
|
"grad_norm": 0.1787109375, |
|
"learning_rate": 3.0303030303030305e-07, |
|
"loss": 1.0184, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.9717171717171718, |
|
"grad_norm": 0.181640625, |
|
"learning_rate": 2.828282828282829e-07, |
|
"loss": 1.0248, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 0.9737373737373738, |
|
"grad_norm": 0.177734375, |
|
"learning_rate": 2.6262626262626266e-07, |
|
"loss": 0.9759, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.9757575757575757, |
|
"grad_norm": 0.185546875, |
|
"learning_rate": 2.4242424242424244e-07, |
|
"loss": 1.0363, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.9777777777777777, |
|
"grad_norm": 0.177734375, |
|
"learning_rate": 2.2222222222222224e-07, |
|
"loss": 1.0131, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.9797979797979798, |
|
"grad_norm": 0.197265625, |
|
"learning_rate": 2.0202020202020205e-07, |
|
"loss": 1.0333, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.9818181818181818, |
|
"grad_norm": 0.1962890625, |
|
"learning_rate": 1.8181818181818183e-07, |
|
"loss": 1.057, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.9838383838383838, |
|
"grad_norm": 0.1865234375, |
|
"learning_rate": 1.6161616161616163e-07, |
|
"loss": 1.0465, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 0.9858585858585859, |
|
"grad_norm": 0.1943359375, |
|
"learning_rate": 1.4141414141414144e-07, |
|
"loss": 1.0177, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.9878787878787879, |
|
"grad_norm": 0.251953125, |
|
"learning_rate": 1.2121212121212122e-07, |
|
"loss": 1.0767, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.98989898989899, |
|
"grad_norm": 0.2099609375, |
|
"learning_rate": 1.0101010101010103e-07, |
|
"loss": 1.0518, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.9919191919191919, |
|
"grad_norm": 0.1796875, |
|
"learning_rate": 8.080808080808082e-08, |
|
"loss": 1.0433, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 0.9939393939393939, |
|
"grad_norm": 0.208984375, |
|
"learning_rate": 6.060606060606061e-08, |
|
"loss": 1.0402, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.9959595959595959, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 4.040404040404041e-08, |
|
"loss": 1.0026, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.997979797979798, |
|
"grad_norm": 0.228515625, |
|
"learning_rate": 2.0202020202020204e-08, |
|
"loss": 1.1265, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.1806640625, |
|
"learning_rate": 0.0, |
|
"loss": 1.0203, |
|
"step": 495 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 495, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 0, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.5810777051205468e+18, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|