|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 945, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 6.896551724137931e-07, |
|
"loss": 1.0054, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.3793103448275862e-06, |
|
"loss": 1.0132, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2.0689655172413796e-06, |
|
"loss": 1.0171, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2.7586206896551725e-06, |
|
"loss": 1.0454, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 3.448275862068966e-06, |
|
"loss": 1.0244, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.137931034482759e-06, |
|
"loss": 0.7725, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.8275862068965525e-06, |
|
"loss": 0.6753, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 5.517241379310345e-06, |
|
"loss": 0.5906, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 6.206896551724138e-06, |
|
"loss": 0.4258, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 6.896551724137932e-06, |
|
"loss": 0.4697, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 7.586206896551724e-06, |
|
"loss": 0.4119, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.275862068965518e-06, |
|
"loss": 0.3669, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.965517241379312e-06, |
|
"loss": 0.3694, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.655172413793105e-06, |
|
"loss": 0.3057, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.0344827586206898e-05, |
|
"loss": 0.306, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.103448275862069e-05, |
|
"loss": 0.386, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.1724137931034483e-05, |
|
"loss": 0.4041, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.2413793103448277e-05, |
|
"loss": 0.3486, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.310344827586207e-05, |
|
"loss": 0.3163, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.3793103448275863e-05, |
|
"loss": 0.3091, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.4482758620689657e-05, |
|
"loss": 0.292, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.5172413793103448e-05, |
|
"loss": 0.3126, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.586206896551724e-05, |
|
"loss": 0.3416, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.6551724137931037e-05, |
|
"loss": 0.3394, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.7241379310344828e-05, |
|
"loss": 0.3389, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.7931034482758623e-05, |
|
"loss": 0.2695, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.8620689655172415e-05, |
|
"loss": 0.2732, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.931034482758621e-05, |
|
"loss": 0.2657, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 2e-05, |
|
"loss": 0.2874, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.9999941186316965e-05, |
|
"loss": 0.2688, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.999976474595967e-05, |
|
"loss": 0.2732, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.999947068100353e-05, |
|
"loss": 0.254, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.9999058994907564e-05, |
|
"loss": 0.2668, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.9998529692514318e-05, |
|
"loss": 0.2683, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.9997882780049847e-05, |
|
"loss": 0.2405, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.9997118265123597e-05, |
|
"loss": 0.2194, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.999623615672837e-05, |
|
"loss": 0.2686, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.9995236465240168e-05, |
|
"loss": 0.2915, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.99941192024181e-05, |
|
"loss": 0.2526, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.999288438140423e-05, |
|
"loss": 0.2572, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.999153201672344e-05, |
|
"loss": 0.2291, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.9990062124283233e-05, |
|
"loss": 0.2449, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.998847472137357e-05, |
|
"loss": 0.274, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.998676982666665e-05, |
|
"loss": 0.2616, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.9984947460216708e-05, |
|
"loss": 0.2395, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.9983007643459757e-05, |
|
"loss": 0.2493, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.9980950399213344e-05, |
|
"loss": 0.2405, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.99787757516763e-05, |
|
"loss": 0.2522, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.9976483726428423e-05, |
|
"loss": 0.2219, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.997407435043021e-05, |
|
"loss": 0.2367, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.9971547652022512e-05, |
|
"loss": 0.2212, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.996890366092622e-05, |
|
"loss": 0.2114, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.99661424082419e-05, |
|
"loss": 0.2485, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.9963263926449446e-05, |
|
"loss": 0.2455, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.9960268249407674e-05, |
|
"loss": 0.2487, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.995715541235395e-05, |
|
"loss": 0.2037, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.9953925451903757e-05, |
|
"loss": 0.2269, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.9950578406050263e-05, |
|
"loss": 0.2332, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.9947114314163892e-05, |
|
"loss": 0.2271, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.9943533216991843e-05, |
|
"loss": 0.2351, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.9939835156657616e-05, |
|
"loss": 0.2352, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.9936020176660524e-05, |
|
"loss": 0.2251, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.9932088321875175e-05, |
|
"loss": 0.2469, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.9928039638550932e-05, |
|
"loss": 0.2332, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.9923874174311394e-05, |
|
"loss": 0.2192, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.9919591978153824e-05, |
|
"loss": 0.2198, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.991519310044857e-05, |
|
"loss": 0.2253, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.9910677592938458e-05, |
|
"loss": 0.2036, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.990604550873823e-05, |
|
"loss": 0.2185, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.990129690233386e-05, |
|
"loss": 0.2292, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.989643182958196e-05, |
|
"loss": 0.2478, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.98914503477091e-05, |
|
"loss": 0.2175, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.9886352515311134e-05, |
|
"loss": 0.2115, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.9881138392352528e-05, |
|
"loss": 0.1942, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.987580804016563e-05, |
|
"loss": 0.2463, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.9870361521449978e-05, |
|
"loss": 0.1982, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.986479890027153e-05, |
|
"loss": 0.1976, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.985912024206194e-05, |
|
"loss": 0.2089, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.985332561361776e-05, |
|
"loss": 0.2252, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.9847415083099685e-05, |
|
"loss": 0.2255, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9841388720031727e-05, |
|
"loss": 0.2118, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9835246595300405e-05, |
|
"loss": 0.2319, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9828988781153916e-05, |
|
"loss": 0.2213, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9822615351201283e-05, |
|
"loss": 0.2305, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9816126380411478e-05, |
|
"loss": 0.2321, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.980952194511256e-05, |
|
"loss": 0.2576, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9802802122990758e-05, |
|
"loss": 0.1926, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9795966993089574e-05, |
|
"loss": 0.23, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9789016635808836e-05, |
|
"loss": 0.2072, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.978195113290377e-05, |
|
"loss": 0.2241, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.977477056748402e-05, |
|
"loss": 0.2291, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9767475024012698e-05, |
|
"loss": 0.215, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9760064588305347e-05, |
|
"loss": 0.239, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9752539347528973e-05, |
|
"loss": 0.2006, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9744899390201006e-05, |
|
"loss": 0.1998, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9737144806188248e-05, |
|
"loss": 0.2152, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9729275686705832e-05, |
|
"loss": 0.202, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.972129212431613e-05, |
|
"loss": 0.226, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9713194212927694e-05, |
|
"loss": 0.2236, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.9704982047794118e-05, |
|
"loss": 0.1976, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.9696655725512933e-05, |
|
"loss": 0.2054, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.968821534402448e-05, |
|
"loss": 0.1965, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.9679661002610743e-05, |
|
"loss": 0.1757, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.9670992801894185e-05, |
|
"loss": 0.2014, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.9662210843836574e-05, |
|
"loss": 0.1932, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.965331523173776e-05, |
|
"loss": 0.189, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.9644306070234492e-05, |
|
"loss": 0.185, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.9635183465299157e-05, |
|
"loss": 0.1849, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9625947524238564e-05, |
|
"loss": 0.1971, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9616598355692644e-05, |
|
"loss": 0.1665, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9607136069633212e-05, |
|
"loss": 0.1899, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9597560777362645e-05, |
|
"loss": 0.1633, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9587872591512583e-05, |
|
"loss": 0.1892, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9578071626042602e-05, |
|
"loss": 0.1857, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9568157996238884e-05, |
|
"loss": 0.2025, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9558131818712838e-05, |
|
"loss": 0.1909, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9547993211399753e-05, |
|
"loss": 0.2058, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9537742293557397e-05, |
|
"loss": 0.2028, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.9527379185764613e-05, |
|
"loss": 0.1567, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.951690400991991e-05, |
|
"loss": 0.1813, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.9506316889240027e-05, |
|
"loss": 0.1808, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.9495617948258473e-05, |
|
"loss": 0.1801, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.9484807312824066e-05, |
|
"loss": 0.2242, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.947388511009947e-05, |
|
"loss": 0.1932, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.946285146855968e-05, |
|
"loss": 0.1923, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.9451706517990516e-05, |
|
"loss": 0.1837, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.944045038948709e-05, |
|
"loss": 0.1685, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.9429083215452276e-05, |
|
"loss": 0.2152, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.941760512959516e-05, |
|
"loss": 0.2018, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.940601626692943e-05, |
|
"loss": 0.1931, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.939431676377183e-05, |
|
"loss": 0.2258, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.9382506757740532e-05, |
|
"loss": 0.194, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.9370586387753532e-05, |
|
"loss": 0.2153, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.9358555794027e-05, |
|
"loss": 0.212, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.9346415118073634e-05, |
|
"loss": 0.2058, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.9334164502701017e-05, |
|
"loss": 0.173, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.932180409200991e-05, |
|
"loss": 0.1968, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.930933403139256e-05, |
|
"loss": 0.23, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.9296754467531015e-05, |
|
"loss": 0.1965, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.9284065548395362e-05, |
|
"loss": 0.1716, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.9271267423242028e-05, |
|
"loss": 0.2417, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.9258360242611973e-05, |
|
"loss": 0.2018, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.9245344158328975e-05, |
|
"loss": 0.1996, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.9232219323497797e-05, |
|
"loss": 0.1869, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.921898589250242e-05, |
|
"loss": 0.1926, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.9205644021004203e-05, |
|
"loss": 0.2506, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.919219386594007e-05, |
|
"loss": 0.2209, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9178635585520648e-05, |
|
"loss": 0.2094, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9164969339228424e-05, |
|
"loss": 0.2151, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.915119528781585e-05, |
|
"loss": 0.2075, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9137313593303465e-05, |
|
"loss": 0.202, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9123324418977982e-05, |
|
"loss": 0.1809, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9109227929390378e-05, |
|
"loss": 0.2255, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9095024290353948e-05, |
|
"loss": 0.2092, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9080713668942356e-05, |
|
"loss": 0.2036, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.906629623348767e-05, |
|
"loss": 0.1858, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.905177215357839e-05, |
|
"loss": 0.1843, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.903714160005744e-05, |
|
"loss": 0.2146, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.9022404745020165e-05, |
|
"loss": 0.2113, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.900756176181231e-05, |
|
"loss": 0.1903, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.8992612825027978e-05, |
|
"loss": 0.1765, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.897755811050758e-05, |
|
"loss": 0.191, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.896239779533575e-05, |
|
"loss": 0.1844, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.8947132057839287e-05, |
|
"loss": 0.1686, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.8931761077585037e-05, |
|
"loss": 0.1933, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.8916285035377794e-05, |
|
"loss": 0.1661, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.8900704113258166e-05, |
|
"loss": 0.1726, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.8885018494500436e-05, |
|
"loss": 0.176, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.8869228363610406e-05, |
|
"loss": 0.189, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.8853333906323226e-05, |
|
"loss": 0.1692, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.8837335309601214e-05, |
|
"loss": 0.1749, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.882123276163164e-05, |
|
"loss": 0.22, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.8805026451824547e-05, |
|
"loss": 0.1686, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.8788716570810478e-05, |
|
"loss": 0.1675, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.8772303310438275e-05, |
|
"loss": 0.1842, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.875578686377279e-05, |
|
"loss": 0.175, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.8739167425092644e-05, |
|
"loss": 0.1985, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.872244518988791e-05, |
|
"loss": 0.1766, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.8705620354857833e-05, |
|
"loss": 0.1706, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.868869311790852e-05, |
|
"loss": 0.2061, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.8671663678150605e-05, |
|
"loss": 0.1982, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.8654532235896897e-05, |
|
"loss": 0.1971, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.8637298992660042e-05, |
|
"loss": 0.2109, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.861996415115014e-05, |
|
"loss": 0.1826, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.860252791527236e-05, |
|
"loss": 0.1697, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.858499049012456e-05, |
|
"loss": 0.2126, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.8567352081994852e-05, |
|
"loss": 0.1738, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.8549612898359185e-05, |
|
"loss": 0.2177, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.8531773147878895e-05, |
|
"loss": 0.189, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.851383304039828e-05, |
|
"loss": 0.1908, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.849579278694209e-05, |
|
"loss": 0.1669, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.8477652599713082e-05, |
|
"loss": 0.1788, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.8459412692089497e-05, |
|
"loss": 0.2174, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.8441073278622562e-05, |
|
"loss": 0.188, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.842263457503397e-05, |
|
"loss": 0.1992, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.8404096798213334e-05, |
|
"loss": 0.1494, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.838546016621564e-05, |
|
"loss": 0.1837, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.8366724898258678e-05, |
|
"loss": 0.198, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.8347891214720477e-05, |
|
"loss": 0.1876, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.8328959337136692e-05, |
|
"loss": 0.1809, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.8309929488198012e-05, |
|
"loss": 0.1947, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.829080189174754e-05, |
|
"loss": 0.1687, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.8271576772778154e-05, |
|
"loss": 0.1722, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.825225435742986e-05, |
|
"loss": 0.1736, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.8232834872987147e-05, |
|
"loss": 0.1716, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.82133185478763e-05, |
|
"loss": 0.214, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.8193705611662697e-05, |
|
"loss": 0.1625, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.8173996295048147e-05, |
|
"loss": 0.2052, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.8154190829868152e-05, |
|
"loss": 0.174, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.8134289449089184e-05, |
|
"loss": 0.1893, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.8114292386805935e-05, |
|
"loss": 0.1875, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.809419987823859e-05, |
|
"loss": 0.2062, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.8074012159730034e-05, |
|
"loss": 0.2076, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.805372946874308e-05, |
|
"loss": 0.2224, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.8033352043857677e-05, |
|
"loss": 0.204, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.801288012476811e-05, |
|
"loss": 0.1863, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.7992313952280175e-05, |
|
"loss": 0.1793, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.7971653768308334e-05, |
|
"loss": 0.1749, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.7950899815872894e-05, |
|
"loss": 0.2104, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.793005233909713e-05, |
|
"loss": 0.2169, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.790911158320442e-05, |
|
"loss": 0.1751, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.7888077794515362e-05, |
|
"loss": 0.1779, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.786695122044487e-05, |
|
"loss": 0.1709, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.7845732109499275e-05, |
|
"loss": 0.2078, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.782442071127338e-05, |
|
"loss": 0.1628, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.7803017276447558e-05, |
|
"loss": 0.1854, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.778152205678477e-05, |
|
"loss": 0.1862, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.775993530512763e-05, |
|
"loss": 0.1888, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.7738257275395404e-05, |
|
"loss": 0.212, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.771648822258105e-05, |
|
"loss": 0.1711, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.7694628402748203e-05, |
|
"loss": 0.1914, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.7672678073028163e-05, |
|
"loss": 0.1647, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.765063749161688e-05, |
|
"loss": 0.1597, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.7628506917771906e-05, |
|
"loss": 0.1624, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.7606286611809353e-05, |
|
"loss": 0.1767, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.758397683510083e-05, |
|
"loss": 0.1908, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.7561577850070355e-05, |
|
"loss": 0.2068, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.7539089920191298e-05, |
|
"loss": 0.174, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.7516513309983253e-05, |
|
"loss": 0.186, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.7493848285008938e-05, |
|
"loss": 0.1548, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.7471095111871076e-05, |
|
"loss": 0.1749, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.7448254058209244e-05, |
|
"loss": 0.1895, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.7425325392696742e-05, |
|
"loss": 0.1668, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.7402309385037423e-05, |
|
"loss": 0.1675, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.7379206305962525e-05, |
|
"loss": 0.1688, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.735601642722748e-05, |
|
"loss": 0.1871, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.7332740021608722e-05, |
|
"loss": 0.1651, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.7309377362900486e-05, |
|
"loss": 0.1945, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.7285928725911562e-05, |
|
"loss": 0.1787, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.7262394386462102e-05, |
|
"loss": 0.1766, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.7238774621380335e-05, |
|
"loss": 0.1692, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.721506970849934e-05, |
|
"loss": 0.1538, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.719127992665376e-05, |
|
"loss": 0.1746, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.7167405555676535e-05, |
|
"loss": 0.2094, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.71434468763956e-05, |
|
"loss": 0.2239, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.7119404170630594e-05, |
|
"loss": 0.1753, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.709527772118953e-05, |
|
"loss": 0.1757, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.7071067811865477e-05, |
|
"loss": 0.1679, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.7046774727433223e-05, |
|
"loss": 0.1985, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.7022398753645918e-05, |
|
"loss": 0.1835, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.6997940177231722e-05, |
|
"loss": 0.212, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.697339928589043e-05, |
|
"loss": 0.1696, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.6948776368290085e-05, |
|
"loss": 0.1902, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.6924071714063574e-05, |
|
"loss": 0.1707, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.6899285613805246e-05, |
|
"loss": 0.16, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.6874418359067463e-05, |
|
"loss": 0.179, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.6849470242357197e-05, |
|
"loss": 0.1711, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.682444155713257e-05, |
|
"loss": 0.1596, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.6799332597799413e-05, |
|
"loss": 0.1937, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.677414365970781e-05, |
|
"loss": 0.1669, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.6748875039148592e-05, |
|
"loss": 0.1915, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.67235270333499e-05, |
|
"loss": 0.1954, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.6698099940473644e-05, |
|
"loss": 0.1753, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.6672594059612018e-05, |
|
"loss": 0.1766, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.664700969078398e-05, |
|
"loss": 0.1864, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.662134713493173e-05, |
|
"loss": 0.163, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.659560669391714e-05, |
|
"loss": 0.1931, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.6569788670518253e-05, |
|
"loss": 0.1521, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.6543893368425664e-05, |
|
"loss": 0.1662, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.6517921092239002e-05, |
|
"loss": 0.1945, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.6491872147463307e-05, |
|
"loss": 0.1652, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.646574684050545e-05, |
|
"loss": 0.1874, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.6439545478670543e-05, |
|
"loss": 0.2133, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.64132683701583e-05, |
|
"loss": 0.1726, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.6386915824059427e-05, |
|
"loss": 0.1853, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.6360488150351984e-05, |
|
"loss": 0.1741, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.6333985659897737e-05, |
|
"loss": 0.1936, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.6307408664438496e-05, |
|
"loss": 0.1621, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.6280757476592467e-05, |
|
"loss": 0.1691, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.625403240985054e-05, |
|
"loss": 0.1737, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.622723377857265e-05, |
|
"loss": 0.1847, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.620036189798403e-05, |
|
"loss": 0.1832, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.6173417084171537e-05, |
|
"loss": 0.1951, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.6146399654079915e-05, |
|
"loss": 0.1627, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.611930992550808e-05, |
|
"loss": 0.1633, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.6092148217105372e-05, |
|
"loss": 0.1699, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.6064914848367818e-05, |
|
"loss": 0.1613, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.6037610139634358e-05, |
|
"loss": 0.1653, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.601023441208309e-05, |
|
"loss": 0.1856, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.598278798772748e-05, |
|
"loss": 0.1719, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.5955271189412596e-05, |
|
"loss": 0.2009, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.5927684340811292e-05, |
|
"loss": 0.1805, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.5900027766420396e-05, |
|
"loss": 0.1934, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.5872301791556904e-05, |
|
"loss": 0.2081, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.5844506742354163e-05, |
|
"loss": 0.1702, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.581664294575801e-05, |
|
"loss": 0.2058, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.5788710729522953e-05, |
|
"loss": 0.2079, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.576071042220829e-05, |
|
"loss": 0.1934, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.5732642353174257e-05, |
|
"loss": 0.149, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.5704506852578165e-05, |
|
"loss": 0.179, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.567630425137049e-05, |
|
"loss": 0.1692, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.5648034881291005e-05, |
|
"loss": 0.1747, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.5619699074864864e-05, |
|
"loss": 0.1747, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.5591297165398693e-05, |
|
"loss": 0.1605, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.5562829486976675e-05, |
|
"loss": 0.2003, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.553429637445661e-05, |
|
"loss": 0.1552, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.5505698163465986e-05, |
|
"loss": 0.1979, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.5477035190398028e-05, |
|
"loss": 0.1706, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.5448307792407737e-05, |
|
"loss": 0.1963, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.5419516307407925e-05, |
|
"loss": 0.1788, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.5390661074065257e-05, |
|
"loss": 0.1719, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.5361742431796235e-05, |
|
"loss": 0.1737, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.5332760720763232e-05, |
|
"loss": 0.1971, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.5303716281870485e-05, |
|
"loss": 0.1874, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.5274609456760073e-05, |
|
"loss": 0.1752, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.5245440587807917e-05, |
|
"loss": 0.1754, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.5216210018119735e-05, |
|
"loss": 0.1836, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.5186918091527025e-05, |
|
"loss": 0.1718, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.5157565152583002e-05, |
|
"loss": 0.1783, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.5128151546558554e-05, |
|
"loss": 0.1746, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.5098677619438183e-05, |
|
"loss": 0.1828, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.5069143717915928e-05, |
|
"loss": 0.1622, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.50395501893913e-05, |
|
"loss": 0.1609, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.5009897381965171e-05, |
|
"loss": 0.1647, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.4980185644435712e-05, |
|
"loss": 0.161, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.495041532629426e-05, |
|
"loss": 0.1456, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.4920586777721231e-05, |
|
"loss": 0.1805, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.4890700349581986e-05, |
|
"loss": 0.1694, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.4860756393422699e-05, |
|
"loss": 0.1953, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.4830755261466249e-05, |
|
"loss": 0.174, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.4800697306608043e-05, |
|
"loss": 0.1779, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.477058288241189e-05, |
|
"loss": 0.1992, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.4740412343105828e-05, |
|
"loss": 0.1655, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.4710186043577965e-05, |
|
"loss": 0.1842, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.4679904339372301e-05, |
|
"loss": 0.179, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.4649567586684548e-05, |
|
"loss": 0.1771, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.4619176142357936e-05, |
|
"loss": 0.17, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.4588730363879015e-05, |
|
"loss": 0.1736, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.4558230609373469e-05, |
|
"loss": 0.1572, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.4527677237601865e-05, |
|
"loss": 0.1661, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.4497070607955477e-05, |
|
"loss": 0.1768, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.4466411080452019e-05, |
|
"loss": 0.2008, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.4435699015731449e-05, |
|
"loss": 0.151, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.4404934775051682e-05, |
|
"loss": 0.1498, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.4374118720284388e-05, |
|
"loss": 0.1884, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.4343251213910698e-05, |
|
"loss": 0.1771, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.4312332619016964e-05, |
|
"loss": 0.1902, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.4281363299290466e-05, |
|
"loss": 0.1691, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.4250343619015162e-05, |
|
"loss": 0.1592, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.4219273943067375e-05, |
|
"loss": 0.1769, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.4188154636911524e-05, |
|
"loss": 0.1571, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.4156986066595807e-05, |
|
"loss": 0.1734, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.412576859874791e-05, |
|
"loss": 0.1779, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.4094502600570687e-05, |
|
"loss": 0.2015, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.4063188439837831e-05, |
|
"loss": 0.1682, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.4031826484889579e-05, |
|
"loss": 0.1337, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.4000417104628332e-05, |
|
"loss": 0.1711, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.396896066851437e-05, |
|
"loss": 0.1489, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.393745754656146e-05, |
|
"loss": 0.1701, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.3905908109332526e-05, |
|
"loss": 0.1627, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.3874312727935293e-05, |
|
"loss": 0.159, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.3842671774017905e-05, |
|
"loss": 0.1556, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.3810985619764573e-05, |
|
"loss": 0.1837, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.3779254637891181e-05, |
|
"loss": 0.1562, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.3747479201640914e-05, |
|
"loss": 0.1582, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.3715659684779857e-05, |
|
"loss": 0.153, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.3683796461592604e-05, |
|
"loss": 0.1954, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.3651889906877865e-05, |
|
"loss": 0.1959, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.3619940395944027e-05, |
|
"loss": 0.1877, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.3587948304604781e-05, |
|
"loss": 0.1806, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.3555914009174665e-05, |
|
"loss": 0.1613, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.3523837886464664e-05, |
|
"loss": 0.1575, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.3491720313777756e-05, |
|
"loss": 0.1968, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.3459561668904496e-05, |
|
"loss": 0.1797, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.3427362330118542e-05, |
|
"loss": 0.1943, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.3395122676172248e-05, |
|
"loss": 0.1663, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.336284308629216e-05, |
|
"loss": 0.1837, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.3330523940174603e-05, |
|
"loss": 0.1787, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.3298165617981171e-05, |
|
"loss": 0.1735, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.3265768500334286e-05, |
|
"loss": 0.1456, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.3233332968312715e-05, |
|
"loss": 0.1782, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.3200859403447072e-05, |
|
"loss": 0.1856, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.3168348187715353e-05, |
|
"loss": 0.1901, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.3135799703538418e-05, |
|
"loss": 0.1709, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.3103214333775522e-05, |
|
"loss": 0.1805, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.3070592461719778e-05, |
|
"loss": 0.1437, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.3037934471093683e-05, |
|
"loss": 0.1498, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.3005240746044572e-05, |
|
"loss": 0.2256, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.2972511671140127e-05, |
|
"loss": 0.16, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.293974763136383e-05, |
|
"loss": 0.1725, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.2906949012110456e-05, |
|
"loss": 0.1752, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.2874116199181522e-05, |
|
"loss": 0.1768, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.2841249578780756e-05, |
|
"loss": 0.1611, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.2808349537509564e-05, |
|
"loss": 0.1631, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.2775416462362458e-05, |
|
"loss": 0.1418, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.2742450740722532e-05, |
|
"loss": 0.1799, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.2709452760356884e-05, |
|
"loss": 0.1442, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.2676422909412067e-05, |
|
"loss": 0.1627, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.2643361576409517e-05, |
|
"loss": 0.1718, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.2610269150240985e-05, |
|
"loss": 0.1635, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.2577146020163968e-05, |
|
"loss": 0.1553, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.2543992575797118e-05, |
|
"loss": 0.1769, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.2510809207115666e-05, |
|
"loss": 0.1843, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.247759630444684e-05, |
|
"loss": 0.1547, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.2444354258465268e-05, |
|
"loss": 0.1702, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.2411083460188378e-05, |
|
"loss": 0.1331, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.2377784300971807e-05, |
|
"loss": 0.1664, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.2344457172504796e-05, |
|
"loss": 0.1741, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.2311102466805578e-05, |
|
"loss": 0.173, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.2277720576216772e-05, |
|
"loss": 0.1479, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.2244311893400761e-05, |
|
"loss": 0.1768, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.2210876811335089e-05, |
|
"loss": 0.1581, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.2177415723307808e-05, |
|
"loss": 0.1713, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.2143929022912895e-05, |
|
"loss": 0.1534, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.2110417104045575e-05, |
|
"loss": 0.1492, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.2076880360897737e-05, |
|
"loss": 0.173, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.2043319187953242e-05, |
|
"loss": 0.1793, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.2009733979983338e-05, |
|
"loss": 0.1532, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.1976125132041974e-05, |
|
"loss": 0.2076, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.1942493039461185e-05, |
|
"loss": 0.165, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.1908838097846404e-05, |
|
"loss": 0.1684, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.187516070307186e-05, |
|
"loss": 0.1588, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.1841461251275868e-05, |
|
"loss": 0.166, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.1807740138856203e-05, |
|
"loss": 0.1801, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.177399776246543e-05, |
|
"loss": 0.1755, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.1740234519006238e-05, |
|
"loss": 0.1556, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.1706450805626762e-05, |
|
"loss": 0.1544, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.1672647019715926e-05, |
|
"loss": 0.1785, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.1638823558898762e-05, |
|
"loss": 0.1554, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.1604980821031726e-05, |
|
"loss": 0.1469, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.1571119204198038e-05, |
|
"loss": 0.157, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.1537239106702968e-05, |
|
"loss": 0.1688, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.1503340927069189e-05, |
|
"loss": 0.1732, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.1469425064032051e-05, |
|
"loss": 0.189, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.1435491916534919e-05, |
|
"loss": 0.1651, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.1401541883724474e-05, |
|
"loss": 0.1724, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.1367575364946006e-05, |
|
"loss": 0.1705, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.133359275973873e-05, |
|
"loss": 0.1667, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.1299594467831079e-05, |
|
"loss": 0.181, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.1265580889136007e-05, |
|
"loss": 0.1566, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.1231552423746284e-05, |
|
"loss": 0.1652, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.1197509471929779e-05, |
|
"loss": 0.1476, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.1163452434124773e-05, |
|
"loss": 0.1862, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.1129381710935229e-05, |
|
"loss": 0.1567, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.1095297703126094e-05, |
|
"loss": 0.1669, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.1061200811618564e-05, |
|
"loss": 0.1525, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.1027091437485404e-05, |
|
"loss": 0.169, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.0992969981946192e-05, |
|
"loss": 0.1495, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.0958836846362622e-05, |
|
"loss": 0.1311, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.0924692432233781e-05, |
|
"loss": 0.1769, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.0890537141191417e-05, |
|
"loss": 0.148, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.0856371374995222e-05, |
|
"loss": 0.1639, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.0822195535528106e-05, |
|
"loss": 0.1526, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.0788010024791465e-05, |
|
"loss": 0.1486, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.0753815244900459e-05, |
|
"loss": 0.1348, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.0719611598079278e-05, |
|
"loss": 0.1766, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.0685399486656407e-05, |
|
"loss": 0.1902, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.0651179313059904e-05, |
|
"loss": 0.1559, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.0616951479812658e-05, |
|
"loss": 0.1741, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.0582716389527659e-05, |
|
"loss": 0.1879, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.0548474444903247e-05, |
|
"loss": 0.1678, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.051422604871841e-05, |
|
"loss": 0.1581, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.0479971603828001e-05, |
|
"loss": 0.17, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.0445711513158039e-05, |
|
"loss": 0.1495, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.0411446179700944e-05, |
|
"loss": 0.1721, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.0377176006510812e-05, |
|
"loss": 0.1772, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.0342901396698658e-05, |
|
"loss": 0.1719, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.0308622753427692e-05, |
|
"loss": 0.1616, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.0274340479908568e-05, |
|
"loss": 0.1441, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.024005497939464e-05, |
|
"loss": 0.1609, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.0205766655177217e-05, |
|
"loss": 0.1537, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.0171475910580824e-05, |
|
"loss": 0.1652, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.0137183148958462e-05, |
|
"loss": 0.1674, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.0102888773686852e-05, |
|
"loss": 0.1942, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.0068593188161698e-05, |
|
"loss": 0.1573, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.0034296795792936e-05, |
|
"loss": 0.1552, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1e-05, |
|
"loss": 0.1719, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.965703204207065e-06, |
|
"loss": 0.1786, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.931406811838307e-06, |
|
"loss": 0.1584, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.89711122631315e-06, |
|
"loss": 0.1495, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.862816851041541e-06, |
|
"loss": 0.1563, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.82852408941918e-06, |
|
"loss": 0.1406, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.79423334482279e-06, |
|
"loss": 0.1956, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.759945020605363e-06, |
|
"loss": 0.1658, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.725659520091433e-06, |
|
"loss": 0.1519, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.69137724657231e-06, |
|
"loss": 0.1797, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.657098603301347e-06, |
|
"loss": 0.1713, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.622823993489193e-06, |
|
"loss": 0.1469, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.588553820299057e-06, |
|
"loss": 0.1526, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.554288486841963e-06, |
|
"loss": 0.1721, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.520028396172002e-06, |
|
"loss": 0.1497, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.485773951281594e-06, |
|
"loss": 0.1597, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.451525555096753e-06, |
|
"loss": 0.1603, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.417283610472345e-06, |
|
"loss": 0.1558, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.383048520187344e-06, |
|
"loss": 0.1417, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.348820686940099e-06, |
|
"loss": 0.1705, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.314600513343595e-06, |
|
"loss": 0.1592, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.280388401920724e-06, |
|
"loss": 0.1918, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.246184755099545e-06, |
|
"loss": 0.1534, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.211989975208538e-06, |
|
"loss": 0.1559, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.177804464471897e-06, |
|
"loss": 0.1515, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.14362862500478e-06, |
|
"loss": 0.1519, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.109462858808586e-06, |
|
"loss": 0.1731, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.075307567766222e-06, |
|
"loss": 0.1337, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.041163153637382e-06, |
|
"loss": 0.1588, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.00703001805381e-06, |
|
"loss": 0.1674, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.9729085625146e-06, |
|
"loss": 0.1519, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.938799188381439e-06, |
|
"loss": 0.1472, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.904702296873913e-06, |
|
"loss": 0.2063, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.870618289064776e-06, |
|
"loss": 0.1545, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.836547565875227e-06, |
|
"loss": 0.1587, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.802490528070223e-06, |
|
"loss": 0.1602, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.76844757625372e-06, |
|
"loss": 0.1548, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.734419110863996e-06, |
|
"loss": 0.1608, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.700405532168921e-06, |
|
"loss": 0.1613, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.666407240261271e-06, |
|
"loss": 0.1769, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.632424635053997e-06, |
|
"loss": 0.1738, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.59845811627553e-06, |
|
"loss": 0.1274, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.56450808346508e-06, |
|
"loss": 0.1793, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.530574935967952e-06, |
|
"loss": 0.149, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.496659072930815e-06, |
|
"loss": 0.2003, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.462760893297033e-06, |
|
"loss": 0.1459, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.428880795801965e-06, |
|
"loss": 0.1364, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.395019178968274e-06, |
|
"loss": 0.1671, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.36117644110124e-06, |
|
"loss": 0.1881, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.327352980284076e-06, |
|
"loss": 0.1766, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.293549194373243e-06, |
|
"loss": 0.1605, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.259765480993767e-06, |
|
"loss": 0.1919, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.226002237534574e-06, |
|
"loss": 0.1732, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.1922598611438e-06, |
|
"loss": 0.1517, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.158538748724139e-06, |
|
"loss": 0.1475, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.124839296928144e-06, |
|
"loss": 0.1642, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.091161902153594e-06, |
|
"loss": 0.1441, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.057506960538818e-06, |
|
"loss": 0.1518, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.023874867958027e-06, |
|
"loss": 0.1406, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 7.990266020016667e-06, |
|
"loss": 0.1954, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 7.95668081204676e-06, |
|
"loss": 0.1654, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 7.923119639102268e-06, |
|
"loss": 0.1631, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 7.889582895954427e-06, |
|
"loss": 0.1586, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 7.85607097708711e-06, |
|
"loss": 0.1536, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 7.822584276692192e-06, |
|
"loss": 0.1939, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 7.789123188664914e-06, |
|
"loss": 0.1597, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.75568810659924e-06, |
|
"loss": 0.1479, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.722279423783231e-06, |
|
"loss": 0.1635, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.688897533194423e-06, |
|
"loss": 0.1781, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.655542827495205e-06, |
|
"loss": 0.1832, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.622215699028196e-06, |
|
"loss": 0.1431, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.588916539811626e-06, |
|
"loss": 0.1688, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.555645741534736e-06, |
|
"loss": 0.1573, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.52240369555316e-06, |
|
"loss": 0.1619, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.489190792884338e-06, |
|
"loss": 0.1605, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.456007424202886e-06, |
|
"loss": 0.1505, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.422853979836035e-06, |
|
"loss": 0.1549, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.389730849759018e-06, |
|
"loss": 0.1475, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.3566384235904855e-06, |
|
"loss": 0.1652, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.323577090587936e-06, |
|
"loss": 0.1417, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.290547239643117e-06, |
|
"loss": 0.1606, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.257549259277472e-06, |
|
"loss": 0.1558, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.224583537637544e-06, |
|
"loss": 0.1742, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.191650462490439e-06, |
|
"loss": 0.1355, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.158750421219245e-06, |
|
"loss": 0.1612, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.125883800818482e-06, |
|
"loss": 0.1493, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.093050987889547e-06, |
|
"loss": 0.1343, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.060252368636171e-06, |
|
"loss": 0.1364, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.027488328859876e-06, |
|
"loss": 0.1659, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 6.994759253955432e-06, |
|
"loss": 0.1554, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 6.962065528906321e-06, |
|
"loss": 0.1582, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 6.9294075382802215e-06, |
|
"loss": 0.1526, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 6.896785666224482e-06, |
|
"loss": 0.1365, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 6.864200296461585e-06, |
|
"loss": 0.1593, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 6.831651812284652e-06, |
|
"loss": 0.1712, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 6.799140596552929e-06, |
|
"loss": 0.1429, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 6.766667031687286e-06, |
|
"loss": 0.1566, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 6.734231499665717e-06, |
|
"loss": 0.1523, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 6.7018343820188324e-06, |
|
"loss": 0.1514, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 6.669476059825401e-06, |
|
"loss": 0.1373, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 6.637156913707839e-06, |
|
"loss": 0.1738, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 6.604877323827755e-06, |
|
"loss": 0.1703, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 6.572637669881459e-06, |
|
"loss": 0.1828, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 6.540438331095509e-06, |
|
"loss": 0.1624, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 6.508279686222243e-06, |
|
"loss": 0.1625, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 6.47616211353534e-06, |
|
"loss": 0.1369, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 6.444085990825338e-06, |
|
"loss": 0.155, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 6.412051695395225e-06, |
|
"loss": 0.1772, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 6.380059604055975e-06, |
|
"loss": 0.1478, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 6.34811009312214e-06, |
|
"loss": 0.1727, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 6.316203538407397e-06, |
|
"loss": 0.1654, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 6.284340315220146e-06, |
|
"loss": 0.1493, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 6.252520798359092e-06, |
|
"loss": 0.1693, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 6.22074536210882e-06, |
|
"loss": 0.1368, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.18901438023543e-06, |
|
"loss": 0.1303, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.157328225982098e-06, |
|
"loss": 0.1431, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.125687272064713e-06, |
|
"loss": 0.2003, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.094091890667473e-06, |
|
"loss": 0.1772, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.0625424534385425e-06, |
|
"loss": 0.1752, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.031039331485631e-06, |
|
"loss": 0.1427, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 5.999582895371669e-06, |
|
"loss": 0.179, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 5.968173515110427e-06, |
|
"loss": 0.1545, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 5.936811560162169e-06, |
|
"loss": 0.1549, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 5.905497399429316e-06, |
|
"loss": 0.1512, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 5.874231401252091e-06, |
|
"loss": 0.1317, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 5.843013933404197e-06, |
|
"loss": 0.1656, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 5.811845363088477e-06, |
|
"loss": 0.1354, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 5.78072605693263e-06, |
|
"loss": 0.176, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 5.7496563809848434e-06, |
|
"loss": 0.1751, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 5.718636700709538e-06, |
|
"loss": 0.161, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 5.687667380983037e-06, |
|
"loss": 0.1559, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 5.656748786089304e-06, |
|
"loss": 0.18, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 5.625881279715615e-06, |
|
"loss": 0.1296, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 5.59506522494832e-06, |
|
"loss": 0.1404, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 5.564300984268556e-06, |
|
"loss": 0.1471, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 5.533588919547979e-06, |
|
"loss": 0.1673, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 5.502929392044528e-06, |
|
"loss": 0.1345, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 5.472322762398139e-06, |
|
"loss": 0.1517, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 5.441769390626537e-06, |
|
"loss": 0.1605, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 5.411269636120984e-06, |
|
"loss": 0.1486, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 5.3808238576420694e-06, |
|
"loss": 0.161, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 5.350432413315455e-06, |
|
"loss": 0.1402, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.3200956606277006e-06, |
|
"loss": 0.1523, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.289813956422037e-06, |
|
"loss": 0.1819, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.259587656894174e-06, |
|
"loss": 0.1451, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.229417117588112e-06, |
|
"loss": 0.1317, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.199302693391958e-06, |
|
"loss": 0.1534, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.169244738533754e-06, |
|
"loss": 0.166, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.139243606577302e-06, |
|
"loss": 0.1554, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.109299650418018e-06, |
|
"loss": 0.1586, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.07941322227877e-06, |
|
"loss": 0.1608, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 5.049584673705742e-06, |
|
"loss": 0.1397, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 5.019814355564292e-06, |
|
"loss": 0.15, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 4.990102618034832e-06, |
|
"loss": 0.1719, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 4.960449810608705e-06, |
|
"loss": 0.1709, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 4.930856282084074e-06, |
|
"loss": 0.1503, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 4.90132238056182e-06, |
|
"loss": 0.1611, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 4.871848453441449e-06, |
|
"loss": 0.1637, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 4.842434847417001e-06, |
|
"loss": 0.1622, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 4.813081908472977e-06, |
|
"loss": 0.1268, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 4.783789981880267e-06, |
|
"loss": 0.1566, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.754559412192087e-06, |
|
"loss": 0.1494, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.7253905432399295e-06, |
|
"loss": 0.1556, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.696283718129517e-06, |
|
"loss": 0.1321, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.667239279236768e-06, |
|
"loss": 0.1418, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.6382575682037685e-06, |
|
"loss": 0.1356, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.609338925934743e-06, |
|
"loss": 0.1297, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.580483692592074e-06, |
|
"loss": 0.1531, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.551692207592265e-06, |
|
"loss": 0.1638, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.522964809601978e-06, |
|
"loss": 0.1347, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.494301836534016e-06, |
|
"loss": 0.1639, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.465703625543393e-06, |
|
"loss": 0.1619, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.437170513023328e-06, |
|
"loss": 0.1426, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.408702834601309e-06, |
|
"loss": 0.1478, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.380300925135138e-06, |
|
"loss": 0.1368, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.351965118708996e-06, |
|
"loss": 0.1509, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.3236957486295115e-06, |
|
"loss": 0.1674, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.295493147421836e-06, |
|
"loss": 0.1439, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.267357646825746e-06, |
|
"loss": 0.1643, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.239289577791712e-06, |
|
"loss": 0.1626, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.211289270477047e-06, |
|
"loss": 0.1414, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.183357054241989e-06, |
|
"loss": 0.1479, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.155493257645842e-06, |
|
"loss": 0.135, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.127698208443097e-06, |
|
"loss": 0.1652, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.099972233579608e-06, |
|
"loss": 0.1674, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.072315659188709e-06, |
|
"loss": 0.1458, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.044728810587406e-06, |
|
"loss": 0.1427, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.017212012272524e-06, |
|
"loss": 0.1591, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.989765587916914e-06, |
|
"loss": 0.1525, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.962389860365643e-06, |
|
"loss": 0.1439, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.935085151632185e-06, |
|
"loss": 0.1467, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.90785178289463e-06, |
|
"loss": 0.1445, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.8806900744919205e-06, |
|
"loss": 0.1448, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.853600345920087e-06, |
|
"loss": 0.1311, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.826582915828468e-06, |
|
"loss": 0.145, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.7996381020159733e-06, |
|
"loss": 0.1544, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.7727662214273496e-06, |
|
"loss": 0.1776, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.7459675901494587e-06, |
|
"loss": 0.1532, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.719242523407539e-06, |
|
"loss": 0.1525, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.6925913355615075e-06, |
|
"loss": 0.1479, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.6660143401022687e-06, |
|
"loss": 0.1343, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.6395118496480176e-06, |
|
"loss": 0.1765, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.6130841759405776e-06, |
|
"loss": 0.1525, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.5867316298417053e-06, |
|
"loss": 0.1207, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.5604545213294616e-06, |
|
"loss": 0.1422, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.5342531594945493e-06, |
|
"loss": 0.1448, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.508127852536698e-06, |
|
"loss": 0.1448, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.4820789077609997e-06, |
|
"loss": 0.1562, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.4561066315743365e-06, |
|
"loss": 0.1523, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.4302113294817505e-06, |
|
"loss": 0.1406, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.4043933060828606e-06, |
|
"loss": 0.1373, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.378652865068275e-06, |
|
"loss": 0.139, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.352990309216022e-06, |
|
"loss": 0.1824, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.3274059403879865e-06, |
|
"loss": 0.1463, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.3019000595263573e-06, |
|
"loss": 0.1438, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.2764729666501027e-06, |
|
"loss": 0.162, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.2511249608514083e-06, |
|
"loss": 0.1421, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.225856340292195e-06, |
|
"loss": 0.1581, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.2006674022005857e-06, |
|
"loss": 0.1207, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.1755584428674345e-06, |
|
"loss": 0.1577, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.1505297576428074e-06, |
|
"loss": 0.139, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.12558164093254e-06, |
|
"loss": 0.1411, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.1007143861947575e-06, |
|
"loss": 0.1487, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.075928285936427e-06, |
|
"loss": 0.1616, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.0512236317099173e-06, |
|
"loss": 0.1791, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.026600714109571e-06, |
|
"loss": 0.1325, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.0020598227682794e-06, |
|
"loss": 0.1549, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 2.977601246354086e-06, |
|
"loss": 0.1489, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 2.953225272566782e-06, |
|
"loss": 0.1686, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 2.9289321881345257e-06, |
|
"loss": 0.1417, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 2.9047222788104712e-06, |
|
"loss": 0.1448, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 2.880595829369406e-06, |
|
"loss": 0.1461, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 2.8565531236043997e-06, |
|
"loss": 0.1566, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 2.8325944443234674e-06, |
|
"loss": 0.124, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 2.8087200733462427e-06, |
|
"loss": 0.1511, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 2.7849302915006626e-06, |
|
"loss": 0.1261, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.7612253786196663e-06, |
|
"loss": 0.153, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.737605613537899e-06, |
|
"loss": 0.1627, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.714071274088438e-06, |
|
"loss": 0.1552, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.690622637099517e-06, |
|
"loss": 0.1519, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.667259978391281e-06, |
|
"loss": 0.1546, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.6439835727725217e-06, |
|
"loss": 0.1407, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.6207936940374767e-06, |
|
"loss": 0.1183, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.597690614962578e-06, |
|
"loss": 0.1262, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.5746746073032624e-06, |
|
"loss": 0.1283, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.5517459417907575e-06, |
|
"loss": 0.1537, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.5289048881289256e-06, |
|
"loss": 0.1547, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.506151714991062e-06, |
|
"loss": 0.1193, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.4834866900167478e-06, |
|
"loss": 0.1649, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.460910079808703e-06, |
|
"loss": 0.1245, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.4384221499296466e-06, |
|
"loss": 0.1282, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.416023164899175e-06, |
|
"loss": 0.1372, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.3937133881906484e-06, |
|
"loss": 0.161, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.371493082228098e-06, |
|
"loss": 0.1729, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.3493625083831217e-06, |
|
"loss": 0.1481, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.3273219269718384e-06, |
|
"loss": 0.1422, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.3053715972517998e-06, |
|
"loss": 0.1421, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.283511777418954e-06, |
|
"loss": 0.1369, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.2617427246045976e-06, |
|
"loss": 0.13, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.2400646948723724e-06, |
|
"loss": 0.1346, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.218477943215229e-06, |
|
"loss": 0.1328, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.1969827235524467e-06, |
|
"loss": 0.1382, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.1755792887266236e-06, |
|
"loss": 0.1667, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.1542678905007287e-06, |
|
"loss": 0.1215, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.133048779555129e-06, |
|
"loss": 0.1579, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.111922205484641e-06, |
|
"loss": 0.1506, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.090888416795582e-06, |
|
"loss": 0.1332, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.0699476609028714e-06, |
|
"loss": 0.1535, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.0491001841271073e-06, |
|
"loss": 0.1409, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.0283462316916712e-06, |
|
"loss": 0.1771, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.007686047719831e-06, |
|
"loss": 0.1565, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.987119875231891e-06, |
|
"loss": 0.145, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.9666479561423247e-06, |
|
"loss": 0.165, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.9462705312569265e-06, |
|
"loss": 0.1254, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.9259878402699704e-06, |
|
"loss": 0.15, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.9058001217614118e-06, |
|
"loss": 0.1693, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.8857076131940643e-06, |
|
"loss": 0.1694, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.8657105509108208e-06, |
|
"loss": 0.1464, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.8458091701318504e-06, |
|
"loss": 0.1712, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.8260037049518565e-06, |
|
"loss": 0.1452, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.8062943883373052e-06, |
|
"loss": 0.1636, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.7866814521237064e-06, |
|
"loss": 0.14, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.7671651270128531e-06, |
|
"loss": 0.1506, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.747745642570141e-06, |
|
"loss": 0.154, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.7284232272218503e-06, |
|
"loss": 0.1264, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.7091981082524633e-06, |
|
"loss": 0.1253, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.69007051180199e-06, |
|
"loss": 0.1353, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.6710406628633113e-06, |
|
"loss": 0.1415, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.652108785279526e-06, |
|
"loss": 0.1536, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.6332751017413218e-06, |
|
"loss": 0.1309, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.6145398337843654e-06, |
|
"loss": 0.1349, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.5959032017866705e-06, |
|
"loss": 0.1409, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.5773654249660342e-06, |
|
"loss": 0.1288, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.5589267213774394e-06, |
|
"loss": 0.1677, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.5405873079105083e-06, |
|
"loss": 0.1105, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.522347400286921e-06, |
|
"loss": 0.1428, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.5042072130579122e-06, |
|
"loss": 0.1635, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.4861669596017237e-06, |
|
"loss": 0.1631, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.4682268521211075e-06, |
|
"loss": 0.149, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.45038710164082e-06, |
|
"loss": 0.1434, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.43264791800515e-06, |
|
"loss": 0.1626, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.4150095098754402e-06, |
|
"loss": 0.1511, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.3974720847276412e-06, |
|
"loss": 0.1329, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.3800358488498645e-06, |
|
"loss": 0.1707, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.3627010073399605e-06, |
|
"loss": 0.17, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.3454677641031032e-06, |
|
"loss": 0.1479, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.3283363218493962e-06, |
|
"loss": 0.1275, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.3113068820914798e-06, |
|
"loss": 0.1429, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.2943796451421686e-06, |
|
"loss": 0.1609, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.2775548101120949e-06, |
|
"loss": 0.1356, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.2608325749073591e-06, |
|
"loss": 0.1253, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.2442131362272093e-06, |
|
"loss": 0.1437, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.2276966895617271e-06, |
|
"loss": 0.1503, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.2112834291895237e-06, |
|
"loss": 0.1301, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.1949735481754565e-06, |
|
"loss": 0.1578, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.1787672383683613e-06, |
|
"loss": 0.1425, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.1626646903987904e-06, |
|
"loss": 0.1453, |
|
"step": 803 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.146666093676776e-06, |
|
"loss": 0.1263, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.130771636389596e-06, |
|
"loss": 0.1403, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.1149815054995682e-06, |
|
"loss": 0.1499, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.0992958867418358e-06, |
|
"loss": 0.1264, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.083714964622208e-06, |
|
"loss": 0.1398, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.0682389224149648e-06, |
|
"loss": 0.1495, |
|
"step": 809 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.052867942160718e-06, |
|
"loss": 0.1531, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.037602204664252e-06, |
|
"loss": 0.1473, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.0224418894924226e-06, |
|
"loss": 0.1647, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.0073871749720221e-06, |
|
"loss": 0.1373, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.924382381876917e-07, |
|
"loss": 0.1507, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.775952549798406e-07, |
|
"loss": 0.1407, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.628583999425644e-07, |
|
"loss": 0.1623, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.482278464216121e-07, |
|
"loss": 0.1476, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.337037665123305e-07, |
|
"loss": 0.1503, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.192863310576472e-07, |
|
"loss": 0.1356, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.04975709646052e-07, |
|
"loss": 0.175, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 8.907720706096223e-07, |
|
"loss": 0.1301, |
|
"step": 821 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 8.766755810220195e-07, |
|
"loss": 0.1445, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 8.626864066965401e-07, |
|
"loss": 0.1525, |
|
"step": 823 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 8.488047121841525e-07, |
|
"loss": 0.1639, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 8.350306607715774e-07, |
|
"loss": 0.1667, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 8.21364414479352e-07, |
|
"loss": 0.1394, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.07806134059933e-07, |
|
"loss": 0.142, |
|
"step": 827 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 7.943559789957989e-07, |
|
"loss": 0.1265, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 7.810141074975819e-07, |
|
"loss": 0.1558, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 7.677806765022044e-07, |
|
"loss": 0.1281, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 7.546558416710292e-07, |
|
"loss": 0.1569, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 7.416397573880296e-07, |
|
"loss": 0.1552, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 7.287325767579756e-07, |
|
"loss": 0.1323, |
|
"step": 833 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 7.159344516046373e-07, |
|
"loss": 0.1512, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 7.032455324689902e-07, |
|
"loss": 0.1762, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 6.906659686074436e-07, |
|
"loss": 0.1467, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 6.781959079900958e-07, |
|
"loss": 0.1315, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 6.658354972989833e-07, |
|
"loss": 0.1303, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 6.535848819263679e-07, |
|
"loss": 0.1561, |
|
"step": 839 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 6.414442059730052e-07, |
|
"loss": 0.1328, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 6.294136122464701e-07, |
|
"loss": 0.1543, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 6.17493242259467e-07, |
|
"loss": 0.1628, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 6.056832362281728e-07, |
|
"loss": 0.1709, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 5.939837330705733e-07, |
|
"loss": 0.1422, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 5.823948704048443e-07, |
|
"loss": 0.1393, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 5.709167845477225e-07, |
|
"loss": 0.1646, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 5.595496105129139e-07, |
|
"loss": 0.1255, |
|
"step": 847 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 5.482934820094876e-07, |
|
"loss": 0.1604, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 5.371485314403202e-07, |
|
"loss": 0.1602, |
|
"step": 849 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 5.261148899005297e-07, |
|
"loss": 0.1641, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 5.15192687175935e-07, |
|
"loss": 0.1516, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 5.043820517415321e-07, |
|
"loss": 0.1536, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 4.936831107599749e-07, |
|
"loss": 0.1403, |
|
"step": 853 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 4.830959900800902e-07, |
|
"loss": 0.1242, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 4.7262081423538717e-07, |
|
"loss": 0.1349, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.6225770644260595e-07, |
|
"loss": 0.1512, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.520067886002488e-07, |
|
"loss": 0.135, |
|
"step": 857 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.418681812871639e-07, |
|
"loss": 0.1371, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.318420037611182e-07, |
|
"loss": 0.1512, |
|
"step": 859 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.2192837395739874e-07, |
|
"loss": 0.1597, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.121274084874194e-07, |
|
"loss": 0.1558, |
|
"step": 861 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.024392226373563e-07, |
|
"loss": 0.1432, |
|
"step": 862 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 3.9286393036678914e-07, |
|
"loss": 0.1299, |
|
"step": 863 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 3.834016443073574e-07, |
|
"loss": 0.1615, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.7405247576144055e-07, |
|
"loss": 0.1161, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.648165347008448e-07, |
|
"loss": 0.1371, |
|
"step": 866 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.556939297655115e-07, |
|
"loss": 0.1303, |
|
"step": 867 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.4668476826224205e-07, |
|
"loss": 0.1763, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.3778915616342947e-07, |
|
"loss": 0.1381, |
|
"step": 869 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.2900719810581406e-07, |
|
"loss": 0.1483, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.203389973892579e-07, |
|
"loss": 0.1479, |
|
"step": 871 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.117846559755211e-07, |
|
"loss": 0.1419, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.033442744870685e-07, |
|
"loss": 0.1454, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.95017952205886e-07, |
|
"loss": 0.1531, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.868057870723073e-07, |
|
"loss": 0.1406, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.7870787568387016e-07, |
|
"loss": 0.1427, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.707243132941717e-07, |
|
"loss": 0.1558, |
|
"step": 877 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.6285519381175407e-07, |
|
"loss": 0.1375, |
|
"step": 878 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.551006097989961e-07, |
|
"loss": 0.1409, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.474606524710288e-07, |
|
"loss": 0.1691, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.399354116946584e-07, |
|
"loss": 0.1522, |
|
"step": 881 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.3252497598730606e-07, |
|
"loss": 0.152, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.2522943251597874e-07, |
|
"loss": 0.1555, |
|
"step": 883 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.180488670962322e-07, |
|
"loss": 0.1644, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.1098336419116628e-07, |
|
"loss": 0.1636, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.0403300691042906e-07, |
|
"loss": 0.1588, |
|
"step": 886 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.9719787700924308e-07, |
|
"loss": 0.1331, |
|
"step": 887 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.9047805488744188e-07, |
|
"loss": 0.1278, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.8387361958852378e-07, |
|
"loss": 0.1511, |
|
"step": 889 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.7738464879871963e-07, |
|
"loss": 0.1613, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.710112188460844e-07, |
|
"loss": 0.1159, |
|
"step": 891 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.6475340469959689e-07, |
|
"loss": 0.1403, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.5861127996827597e-07, |
|
"loss": 0.1335, |
|
"step": 893 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.5258491690031796e-07, |
|
"loss": 0.1329, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.4667438638224064e-07, |
|
"loss": 0.1555, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.4087975793806275e-07, |
|
"loss": 0.1179, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.3520109972846918e-07, |
|
"loss": 0.156, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.2963847855002264e-07, |
|
"loss": 0.1517, |
|
"step": 898 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.2419195983436881e-07, |
|
"loss": 0.1242, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.1886160764747356e-07, |
|
"loss": 0.1666, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.1364748468886688e-07, |
|
"loss": 0.155, |
|
"step": 901 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.0854965229090352e-07, |
|
"loss": 0.1435, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.0356817041804246e-07, |
|
"loss": 0.1658, |
|
"step": 903 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.870309766614184e-08, |
|
"loss": 0.1481, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.395449126177291e-08, |
|
"loss": 0.1384, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 8.932240706154171e-08, |
|
"loss": 0.1556, |
|
"step": 906 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 8.480689955143395e-08, |
|
"loss": 0.1419, |
|
"step": 907 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 8.040802184617446e-08, |
|
"loss": 0.1354, |
|
"step": 908 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.61258256886055e-08, |
|
"loss": 0.1516, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.196036144907159e-08, |
|
"loss": 0.1555, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 6.791167812483013e-08, |
|
"loss": 0.1404, |
|
"step": 911 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 6.397982333947617e-08, |
|
"loss": 0.1223, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 6.016484334238515e-08, |
|
"loss": 0.1414, |
|
"step": 913 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 5.646678300816e-08, |
|
"loss": 0.1628, |
|
"step": 914 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 5.2885685836109316e-08, |
|
"loss": 0.1603, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.942159394973778e-08, |
|
"loss": 0.1467, |
|
"step": 916 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.6074548096244346e-08, |
|
"loss": 0.1398, |
|
"step": 917 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.284458764604926e-08, |
|
"loss": 0.1389, |
|
"step": 918 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 3.973175059232559e-08, |
|
"loss": 0.1497, |
|
"step": 919 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 3.6736073550556195e-08, |
|
"loss": 0.1467, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 3.385759175809966e-08, |
|
"loss": 0.1453, |
|
"step": 921 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 3.109633907378173e-08, |
|
"loss": 0.1378, |
|
"step": 922 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.845234797748897e-08, |
|
"loss": 0.1332, |
|
"step": 923 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.5925649569791266e-08, |
|
"loss": 0.1415, |
|
"step": 924 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.3516273571577708e-08, |
|
"loss": 0.1371, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.1224248323703512e-08, |
|
"loss": 0.1329, |
|
"step": 926 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.9049600786658073e-08, |
|
"loss": 0.1987, |
|
"step": 927 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.6992356540247446e-08, |
|
"loss": 0.1353, |
|
"step": 928 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.5052539783292353e-08, |
|
"loss": 0.1281, |
|
"step": 929 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.3230173333348417e-08, |
|
"loss": 0.1431, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.1525278626431935e-08, |
|
"loss": 0.1428, |
|
"step": 931 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.937875716770074e-09, |
|
"loss": 0.1397, |
|
"step": 932 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 8.467983276563285e-09, |
|
"loss": 0.1508, |
|
"step": 933 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 7.115618595772145e-09, |
|
"loss": 0.1465, |
|
"step": 934 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 5.880797581904185e-09, |
|
"loss": 0.1338, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 4.763534759835153e-09, |
|
"loss": 0.1458, |
|
"step": 936 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 3.763843271631373e-09, |
|
"loss": 0.1539, |
|
"step": 937 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.881734876403197e-09, |
|
"loss": 0.1475, |
|
"step": 938 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.1172199501573453e-09, |
|
"loss": 0.152, |
|
"step": 939 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.4703074856814437e-09, |
|
"loss": 0.1341, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.410050924374414e-10, |
|
"loss": 0.1522, |
|
"step": 941 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 5.293189964694634e-10, |
|
"loss": 0.1389, |
|
"step": 942 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.352540403327552e-10, |
|
"loss": 0.1511, |
|
"step": 943 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 5.881368303595202e-11, |
|
"loss": 0.122, |
|
"step": 944 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.0, |
|
"loss": 0.1313, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 945, |
|
"total_flos": 267616018546688.0, |
|
"train_loss": 0.1789411918196098, |
|
"train_runtime": 5663.5843, |
|
"train_samples_per_second": 32.028, |
|
"train_steps_per_second": 0.167 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 945, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 800, |
|
"total_flos": 267616018546688.0, |
|
"train_batch_size": 24, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|