|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0144827586206897, |
|
"eval_steps": 145, |
|
"global_step": 1500, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 5.000000000000001e-07, |
|
"loss": 1.4522, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"eval_loss": 1.6368393898010254, |
|
"eval_runtime": 52.4631, |
|
"eval_samples_per_second": 14.715, |
|
"eval_steps_per_second": 7.358, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"loss": 1.7542, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.5e-06, |
|
"loss": 1.4392, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 1.2693, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2.5e-06, |
|
"loss": 2.0255, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 3e-06, |
|
"loss": 1.8536, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 3.5e-06, |
|
"loss": 1.364, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 1.7387, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.5e-06, |
|
"loss": 1.6494, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 5e-06, |
|
"loss": 1.7298, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.999998522885947e-06, |
|
"loss": 1.1661, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.999994091545531e-06, |
|
"loss": 1.8255, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.999986705983991e-06, |
|
"loss": 0.8458, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.999976366210054e-06, |
|
"loss": 1.1704, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.999963072235937e-06, |
|
"loss": 1.8649, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.99994682407735e-06, |
|
"loss": 1.2619, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.999927621753493e-06, |
|
"loss": 1.9507, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.999905465287059e-06, |
|
"loss": 1.4712, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.999880354704227e-06, |
|
"loss": 1.4141, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.999852290034674e-06, |
|
"loss": 0.9806, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.9998212713115605e-06, |
|
"loss": 1.0796, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.999787298571541e-06, |
|
"loss": 1.8835, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.999750371854763e-06, |
|
"loss": 1.3385, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.9997104912048615e-06, |
|
"loss": 1.2733, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.999667656668961e-06, |
|
"loss": 1.1878, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.999621868297683e-06, |
|
"loss": 1.3143, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.999573126145132e-06, |
|
"loss": 0.9616, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.999521430268907e-06, |
|
"loss": 1.5245, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.999466780730097e-06, |
|
"loss": 1.135, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.999409177593279e-06, |
|
"loss": 1.2351, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.999348620926526e-06, |
|
"loss": 1.3666, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.999285110801393e-06, |
|
"loss": 1.6391, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.999218647292932e-06, |
|
"loss": 1.564, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.999149230479682e-06, |
|
"loss": 1.129, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.999076860443671e-06, |
|
"loss": 1.3177, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.999001537270419e-06, |
|
"loss": 1.2801, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.998923261048935e-06, |
|
"loss": 1.4633, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.998842031871717e-06, |
|
"loss": 1.4125, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.998757849834752e-06, |
|
"loss": 1.0231, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.998670715037518e-06, |
|
"loss": 1.4911, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.9985806275829815e-06, |
|
"loss": 1.4786, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.998487587577598e-06, |
|
"loss": 1.4127, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.9983915951313125e-06, |
|
"loss": 0.8299, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.998292650357558e-06, |
|
"loss": 1.4409, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.998190753373256e-06, |
|
"loss": 1.3596, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.998085904298817e-06, |
|
"loss": 1.0493, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.997978103258142e-06, |
|
"loss": 1.5774, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.997867350378618e-06, |
|
"loss": 0.9767, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.99775364579112e-06, |
|
"loss": 1.2138, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.997636989630012e-06, |
|
"loss": 1.416, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.997517382033145e-06, |
|
"loss": 1.017, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.9973948231418596e-06, |
|
"loss": 1.6522, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.997269313100981e-06, |
|
"loss": 1.0402, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.997140852058825e-06, |
|
"loss": 1.1969, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.997009440167192e-06, |
|
"loss": 1.863, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.99687507758137e-06, |
|
"loss": 1.3628, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.996737764460135e-06, |
|
"loss": 1.3267, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.996597500965747e-06, |
|
"loss": 1.8708, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.996454287263955e-06, |
|
"loss": 1.1691, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.996308123523996e-06, |
|
"loss": 1.5923, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.996159009918586e-06, |
|
"loss": 1.8901, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.996006946623933e-06, |
|
"loss": 1.6246, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.99585193381973e-06, |
|
"loss": 1.5248, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.995693971689153e-06, |
|
"loss": 1.405, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.995533060418865e-06, |
|
"loss": 1.6614, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.995369200199014e-06, |
|
"loss": 1.5643, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.995202391223231e-06, |
|
"loss": 1.2362, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.995032633688633e-06, |
|
"loss": 1.0164, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.9948599277958224e-06, |
|
"loss": 1.3667, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.994684273748881e-06, |
|
"loss": 1.141, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.9945056717553816e-06, |
|
"loss": 0.8697, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.994324122026374e-06, |
|
"loss": 1.1478, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.994139624776394e-06, |
|
"loss": 0.9953, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.993952180223461e-06, |
|
"loss": 0.9585, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.993761788589078e-06, |
|
"loss": 1.583, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.993568450098227e-06, |
|
"loss": 1.7872, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.993372164979375e-06, |
|
"loss": 1.6221, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.99317293346447e-06, |
|
"loss": 1.5637, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.992970755788944e-06, |
|
"loss": 0.9444, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.992765632191706e-06, |
|
"loss": 1.3295, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.992557562915152e-06, |
|
"loss": 1.2464, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.992346548205152e-06, |
|
"loss": 1.0352, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.992132588311062e-06, |
|
"loss": 1.0081, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.9919156834857165e-06, |
|
"loss": 1.4097, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.991695833985429e-06, |
|
"loss": 1.3536, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.991473040069996e-06, |
|
"loss": 1.4188, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.99124730200269e-06, |
|
"loss": 0.9988, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.991018620050263e-06, |
|
"loss": 1.4854, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.990786994482947e-06, |
|
"loss": 1.3899, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.990552425574452e-06, |
|
"loss": 1.1053, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.990314913601966e-06, |
|
"loss": 1.0796, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.990074458846155e-06, |
|
"loss": 1.3599, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.989831061591161e-06, |
|
"loss": 1.2571, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.989584722124606e-06, |
|
"loss": 1.6078, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.989335440737587e-06, |
|
"loss": 1.09, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.989083217724676e-06, |
|
"loss": 1.3914, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.988828053383925e-06, |
|
"loss": 1.0266, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.988569948016858e-06, |
|
"loss": 1.5676, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.988308901928477e-06, |
|
"loss": 1.0114, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.988044915427257e-06, |
|
"loss": 1.0441, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.987777988825147e-06, |
|
"loss": 1.1977, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.987508122437575e-06, |
|
"loss": 1.0235, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.987235316583439e-06, |
|
"loss": 1.886, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.986959571585109e-06, |
|
"loss": 0.8832, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.986680887768433e-06, |
|
"loss": 2.0845, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.986399265462728e-06, |
|
"loss": 1.2562, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.986114705000785e-06, |
|
"loss": 0.951, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.985827206718866e-06, |
|
"loss": 0.9751, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.985536770956706e-06, |
|
"loss": 1.048, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.9852433980575095e-06, |
|
"loss": 1.5994, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.984947088367954e-06, |
|
"loss": 1.2163, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.984647842238185e-06, |
|
"loss": 1.2535, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.984345660021819e-06, |
|
"loss": 1.0773, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.984040542075942e-06, |
|
"loss": 1.7597, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.98373248876111e-06, |
|
"loss": 0.9974, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.9834215004413474e-06, |
|
"loss": 1.6053, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.983107577484145e-06, |
|
"loss": 1.0382, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.982790720260463e-06, |
|
"loss": 1.4656, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.982470929144729e-06, |
|
"loss": 1.187, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.982148204514838e-06, |
|
"loss": 0.9327, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.98182254675215e-06, |
|
"loss": 1.168, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.981493956241492e-06, |
|
"loss": 1.1143, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.981162433371157e-06, |
|
"loss": 1.6322, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.980827978532903e-06, |
|
"loss": 1.4653, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.980490592121951e-06, |
|
"loss": 1.4012, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.980150274536989e-06, |
|
"loss": 1.7911, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.9798070261801665e-06, |
|
"loss": 1.5263, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.979460847457097e-06, |
|
"loss": 1.3853, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.9791117387768575e-06, |
|
"loss": 1.3603, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.978759700551986e-06, |
|
"loss": 1.1095, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.978404733198483e-06, |
|
"loss": 0.8909, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.9780468371358116e-06, |
|
"loss": 1.0373, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.977686012786893e-06, |
|
"loss": 0.9252, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.97732226057811e-06, |
|
"loss": 1.7641, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.976955580939306e-06, |
|
"loss": 1.4047, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.976585974303785e-06, |
|
"loss": 1.0995, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.976213441108304e-06, |
|
"loss": 1.1907, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.975837981793085e-06, |
|
"loss": 1.0744, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.975459596801805e-06, |
|
"loss": 1.097, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.975078286581596e-06, |
|
"loss": 0.8977, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.9746940515830514e-06, |
|
"loss": 0.5722, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.974306892260217e-06, |
|
"loss": 1.455, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.973916809070597e-06, |
|
"loss": 1.6582, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.973523802475147e-06, |
|
"loss": 1.1232, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.9731278729382814e-06, |
|
"loss": 1.1224, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"eval_loss": 1.3059237003326416, |
|
"eval_runtime": 53.7489, |
|
"eval_samples_per_second": 14.363, |
|
"eval_steps_per_second": 7.182, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.972729020927866e-06, |
|
"loss": 1.1827, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.97232724691522e-06, |
|
"loss": 0.8917, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.971922551375117e-06, |
|
"loss": 0.8041, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.971514934785783e-06, |
|
"loss": 1.5842, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.9711043976288935e-06, |
|
"loss": 1.5451, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.970690940389576e-06, |
|
"loss": 0.7718, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.970274563556412e-06, |
|
"loss": 1.3965, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.969855267621428e-06, |
|
"loss": 0.7756, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.969433053080105e-06, |
|
"loss": 1.5195, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.969007920431367e-06, |
|
"loss": 1.2027, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.968579870177591e-06, |
|
"loss": 1.7006, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.9681489028246e-06, |
|
"loss": 1.8294, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.9677150188816655e-06, |
|
"loss": 1.1739, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.967278218861502e-06, |
|
"loss": 1.4111, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.966838503280275e-06, |
|
"loss": 1.4639, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.966395872657591e-06, |
|
"loss": 1.4884, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.965950327516502e-06, |
|
"loss": 1.5146, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.965501868383507e-06, |
|
"loss": 1.2453, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.965050495788544e-06, |
|
"loss": 1.4534, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.964596210264997e-06, |
|
"loss": 1.0392, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.9641390123496926e-06, |
|
"loss": 1.6082, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.9636789025828945e-06, |
|
"loss": 1.2436, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.9632158815083135e-06, |
|
"loss": 1.4229, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.9627499496730945e-06, |
|
"loss": 1.5076, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.962281107627828e-06, |
|
"loss": 1.1288, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.961809355926539e-06, |
|
"loss": 0.9409, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.961334695126693e-06, |
|
"loss": 0.944, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.960857125789191e-06, |
|
"loss": 1.1763, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.9603766484783745e-06, |
|
"loss": 1.0742, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.9598932637620175e-06, |
|
"loss": 1.4498, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.959406972211334e-06, |
|
"loss": 1.8762, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.958917774400967e-06, |
|
"loss": 0.7468, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.958425670909001e-06, |
|
"loss": 1.8263, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.957930662316948e-06, |
|
"loss": 0.9994, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.957432749209755e-06, |
|
"loss": 1.1363, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.956931932175803e-06, |
|
"loss": 1.0489, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.956428211806903e-06, |
|
"loss": 0.9684, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.955921588698295e-06, |
|
"loss": 1.2817, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.955412063448654e-06, |
|
"loss": 1.5433, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.95489963666008e-06, |
|
"loss": 1.243, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.954384308938102e-06, |
|
"loss": 1.4317, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.953866080891682e-06, |
|
"loss": 0.8239, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.953344953133201e-06, |
|
"loss": 1.2553, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.952820926278474e-06, |
|
"loss": 1.1115, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.952294000946738e-06, |
|
"loss": 1.1013, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.951764177760657e-06, |
|
"loss": 0.9678, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.951231457346317e-06, |
|
"loss": 0.8286, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.95069584033323e-06, |
|
"loss": 1.0927, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.9501573273543295e-06, |
|
"loss": 1.046, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.949615919045972e-06, |
|
"loss": 1.0288, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.949071616047935e-06, |
|
"loss": 1.5807, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.948524419003415e-06, |
|
"loss": 1.2632, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.947974328559032e-06, |
|
"loss": 1.5379, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.947421345364822e-06, |
|
"loss": 1.5668, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.946865470074241e-06, |
|
"loss": 1.1485, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.946306703344161e-06, |
|
"loss": 1.1301, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.9457450458348735e-06, |
|
"loss": 1.9502, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.945180498210083e-06, |
|
"loss": 1.056, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.94461306113691e-06, |
|
"loss": 1.1029, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.944042735285891e-06, |
|
"loss": 0.9279, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.943469521330975e-06, |
|
"loss": 1.1618, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.942893419949523e-06, |
|
"loss": 1.1059, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.9423144318223085e-06, |
|
"loss": 2.1417, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.941732557633519e-06, |
|
"loss": 1.2264, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.941147798070748e-06, |
|
"loss": 1.1846, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.940560153825003e-06, |
|
"loss": 1.1038, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.939969625590696e-06, |
|
"loss": 1.6479, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.9393762140656484e-06, |
|
"loss": 1.4235, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.938779919951092e-06, |
|
"loss": 1.0992, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.938180743951661e-06, |
|
"loss": 1.2565, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.937578686775396e-06, |
|
"loss": 1.3175, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.936973749133743e-06, |
|
"loss": 0.8759, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.936365931741553e-06, |
|
"loss": 0.8367, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.9357552353170755e-06, |
|
"loss": 1.0833, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.935141660581967e-06, |
|
"loss": 1.2446, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.934525208261284e-06, |
|
"loss": 1.5117, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.933905879083481e-06, |
|
"loss": 1.0355, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.933283673780416e-06, |
|
"loss": 0.8376, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.932658593087341e-06, |
|
"loss": 1.0228, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.93203063774291e-06, |
|
"loss": 1.3021, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.9313998084891734e-06, |
|
"loss": 1.3589, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.930766106071574e-06, |
|
"loss": 1.1009, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.930129531238955e-06, |
|
"loss": 1.0263, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.929490084743549e-06, |
|
"loss": 1.0732, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.928847767340985e-06, |
|
"loss": 1.6764, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.928202579790285e-06, |
|
"loss": 1.3953, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.92755452285386e-06, |
|
"loss": 1.3431, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.926903597297514e-06, |
|
"loss": 1.4097, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.926249803890441e-06, |
|
"loss": 1.0773, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.925593143405221e-06, |
|
"loss": 1.1947, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.924933616617824e-06, |
|
"loss": 1.6708, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.924271224307609e-06, |
|
"loss": 1.5323, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.9236059672573175e-06, |
|
"loss": 1.4578, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.922937846253078e-06, |
|
"loss": 0.4763, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.922266862084405e-06, |
|
"loss": 0.9389, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.921593015544193e-06, |
|
"loss": 1.1362, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.9209163074287195e-06, |
|
"loss": 1.3309, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.920236738537647e-06, |
|
"loss": 1.3693, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.919554309674015e-06, |
|
"loss": 1.0019, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.918869021644243e-06, |
|
"loss": 0.9216, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.9181808752581316e-06, |
|
"loss": 1.089, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.917489871328856e-06, |
|
"loss": 2.1136, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.916796010672969e-06, |
|
"loss": 1.3636, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.916099294110401e-06, |
|
"loss": 1.6074, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.915399722464455e-06, |
|
"loss": 1.046, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.91469729656181e-06, |
|
"loss": 1.705, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.913992017232515e-06, |
|
"loss": 0.9568, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.913283885309994e-06, |
|
"loss": 1.0197, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.912572901631038e-06, |
|
"loss": 1.2031, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.911859067035812e-06, |
|
"loss": 1.5773, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.911142382367847e-06, |
|
"loss": 0.8763, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.910422848474044e-06, |
|
"loss": 2.1782, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.90970046620467e-06, |
|
"loss": 1.8428, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.908975236413356e-06, |
|
"loss": 0.7872, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.908247159957102e-06, |
|
"loss": 1.5376, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.907516237696268e-06, |
|
"loss": 1.1867, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.906782470494579e-06, |
|
"loss": 0.8102, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.906045859219122e-06, |
|
"loss": 0.8138, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.905306404740342e-06, |
|
"loss": 1.2403, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.904564107932048e-06, |
|
"loss": 1.4051, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.903818969671404e-06, |
|
"loss": 1.4872, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.903070990838936e-06, |
|
"loss": 1.8003, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.902320172318521e-06, |
|
"loss": 1.0195, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.901566514997396e-06, |
|
"loss": 0.814, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.900810019766151e-06, |
|
"loss": 0.9092, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.90005068751873e-06, |
|
"loss": 1.1939, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.89928851915243e-06, |
|
"loss": 0.939, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.898523515567897e-06, |
|
"loss": 1.079, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.897755677669131e-06, |
|
"loss": 0.5944, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.896985006363477e-06, |
|
"loss": 1.3437, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.896211502561633e-06, |
|
"loss": 1.5517, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.89543516717764e-06, |
|
"loss": 1.1935, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.894656001128887e-06, |
|
"loss": 1.1588, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.893874005336108e-06, |
|
"loss": 1.5299, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.89308918072338e-06, |
|
"loss": 1.9043, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.892301528218124e-06, |
|
"loss": 1.2985, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.891511048751102e-06, |
|
"loss": 1.4682, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.8907177432564165e-06, |
|
"loss": 1.3038, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.889921612671511e-06, |
|
"loss": 1.5238, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.889122657937163e-06, |
|
"loss": 1.4669, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.888320879997493e-06, |
|
"loss": 1.3409, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.8875162797999545e-06, |
|
"loss": 1.1065, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.886708858295335e-06, |
|
"loss": 1.2819, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.885898616437759e-06, |
|
"loss": 1.1449, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.885085555184681e-06, |
|
"loss": 1.3072, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"eval_loss": 1.2836387157440186, |
|
"eval_runtime": 53.6026, |
|
"eval_samples_per_second": 14.402, |
|
"eval_steps_per_second": 7.201, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.88426967549689e-06, |
|
"loss": 0.6308, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.883450978338503e-06, |
|
"loss": 1.0291, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.882629464676967e-06, |
|
"loss": 1.1062, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.881805135483057e-06, |
|
"loss": 1.2533, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.880977991730877e-06, |
|
"loss": 1.2679, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.880148034397854e-06, |
|
"loss": 1.4962, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.879315264464744e-06, |
|
"loss": 1.0716, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.87847968291562e-06, |
|
"loss": 1.023, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.8776412907378845e-06, |
|
"loss": 1.3586, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.876800088922257e-06, |
|
"loss": 1.0656, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.875956078462778e-06, |
|
"loss": 1.8788, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.875109260356809e-06, |
|
"loss": 1.0744, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.874259635605025e-06, |
|
"loss": 0.928, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.8734072052114215e-06, |
|
"loss": 0.9866, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.872551970183309e-06, |
|
"loss": 1.0274, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.871693931531309e-06, |
|
"loss": 1.0472, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.87083309026936e-06, |
|
"loss": 0.915, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.86996944741471e-06, |
|
"loss": 1.5422, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.869103003987919e-06, |
|
"loss": 1.2316, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.868233761012853e-06, |
|
"loss": 1.7933, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.867361719516692e-06, |
|
"loss": 0.7911, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.866486880529918e-06, |
|
"loss": 1.6406, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.865609245086321e-06, |
|
"loss": 1.2438, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.864728814222996e-06, |
|
"loss": 1.3507, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.863845588980338e-06, |
|
"loss": 1.355, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.86295957040205e-06, |
|
"loss": 0.9027, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.862070759535129e-06, |
|
"loss": 1.2907, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.861179157429877e-06, |
|
"loss": 1.9344, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.860284765139892e-06, |
|
"loss": 1.7715, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.859387583722069e-06, |
|
"loss": 1.1291, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.8584876142366e-06, |
|
"loss": 0.9811, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.8575848577469716e-06, |
|
"loss": 1.0526, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.856679315319962e-06, |
|
"loss": 0.8678, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.855770988025644e-06, |
|
"loss": 1.5962, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.854859876937379e-06, |
|
"loss": 1.9452, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.853945983131819e-06, |
|
"loss": 1.3036, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.853029307688905e-06, |
|
"loss": 1.0005, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.852109851691864e-06, |
|
"loss": 1.3575, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.851187616227209e-06, |
|
"loss": 0.9331, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.850262602384738e-06, |
|
"loss": 1.049, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.8493348112575315e-06, |
|
"loss": 1.3, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.848404243941951e-06, |
|
"loss": 1.3482, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.847470901537642e-06, |
|
"loss": 0.8853, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.846534785147526e-06, |
|
"loss": 0.9279, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.845595895877802e-06, |
|
"loss": 1.2538, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.8446542348379495e-06, |
|
"loss": 0.9986, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.84370980314072e-06, |
|
"loss": 1.4702, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.842762601902141e-06, |
|
"loss": 1.3831, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.84181263224151e-06, |
|
"loss": 1.1784, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.840859895281401e-06, |
|
"loss": 1.5826, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.839904392147652e-06, |
|
"loss": 0.8636, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.838946123969373e-06, |
|
"loss": 1.5024, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.8379850918789435e-06, |
|
"loss": 0.9235, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.837021297012004e-06, |
|
"loss": 1.5189, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.836054740507464e-06, |
|
"loss": 1.3184, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.8350854235074945e-06, |
|
"loss": 1.0936, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.834113347157529e-06, |
|
"loss": 1.2336, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.83313851260626e-06, |
|
"loss": 1.4692, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.832160921005644e-06, |
|
"loss": 1.3801, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.83118057351089e-06, |
|
"loss": 1.228, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.830197471280467e-06, |
|
"loss": 1.0152, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.829211615476098e-06, |
|
"loss": 0.9074, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.82822300726276e-06, |
|
"loss": 1.2977, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.827231647808683e-06, |
|
"loss": 1.1979, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.8262375382853485e-06, |
|
"loss": 1.011, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.825240679867486e-06, |
|
"loss": 1.5818, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.8242410737330745e-06, |
|
"loss": 1.1383, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.82323872106334e-06, |
|
"loss": 1.4268, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.822233623042754e-06, |
|
"loss": 0.779, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.821225780859032e-06, |
|
"loss": 1.3415, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.8202151957031316e-06, |
|
"loss": 0.931, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.819201868769253e-06, |
|
"loss": 1.416, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.818185801254837e-06, |
|
"loss": 1.5162, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.817166994360559e-06, |
|
"loss": 1.4471, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.816145449290336e-06, |
|
"loss": 1.5774, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.815121167251318e-06, |
|
"loss": 1.3644, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.814094149453891e-06, |
|
"loss": 1.7256, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.813064397111673e-06, |
|
"loss": 1.6978, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.812031911441512e-06, |
|
"loss": 1.3279, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.8109966936634885e-06, |
|
"loss": 1.4012, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.80995874500091e-06, |
|
"loss": 1.2934, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.808918066680311e-06, |
|
"loss": 1.1003, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.807874659931452e-06, |
|
"loss": 1.2407, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.806828525987319e-06, |
|
"loss": 0.7935, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.8057796660841175e-06, |
|
"loss": 1.0551, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.804728081461276e-06, |
|
"loss": 1.3337, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.803673773361444e-06, |
|
"loss": 1.2896, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.802616743030488e-06, |
|
"loss": 1.2686, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.8015569917174905e-06, |
|
"loss": 1.81, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.800494520674752e-06, |
|
"loss": 0.8058, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.7994293311577835e-06, |
|
"loss": 1.3053, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.798361424425312e-06, |
|
"loss": 1.3472, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.7972908017392715e-06, |
|
"loss": 1.6796, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.796217464364808e-06, |
|
"loss": 1.008, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.795141413570275e-06, |
|
"loss": 0.7311, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.794062650627234e-06, |
|
"loss": 1.1742, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.792981176810447e-06, |
|
"loss": 1.1034, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.791896993397885e-06, |
|
"loss": 0.8392, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.790810101670715e-06, |
|
"loss": 1.1105, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.789720502913308e-06, |
|
"loss": 1.3574, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.788628198413236e-06, |
|
"loss": 1.4231, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.787533189461263e-06, |
|
"loss": 1.0777, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.786435477351353e-06, |
|
"loss": 0.8623, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.785335063380662e-06, |
|
"loss": 1.678, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.7842319488495395e-06, |
|
"loss": 1.1619, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.783126135061526e-06, |
|
"loss": 0.8736, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.7820176233233525e-06, |
|
"loss": 0.9965, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.780906414944938e-06, |
|
"loss": 1.3315, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.779792511239388e-06, |
|
"loss": 1.3204, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.77867591352299e-06, |
|
"loss": 1.8712, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.7775566231152216e-06, |
|
"loss": 1.3979, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.7764346413387356e-06, |
|
"loss": 1.1211, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.775309969519369e-06, |
|
"loss": 0.7957, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.774182608986137e-06, |
|
"loss": 1.152, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.773052561071232e-06, |
|
"loss": 1.5962, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.771919827110021e-06, |
|
"loss": 0.9141, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.7707844084410444e-06, |
|
"loss": 1.4446, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.769646306406019e-06, |
|
"loss": 1.1342, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.76850552234983e-06, |
|
"loss": 0.8174, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.76736205762053e-06, |
|
"loss": 1.8466, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.766215913569342e-06, |
|
"loss": 0.885, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.765067091550655e-06, |
|
"loss": 0.9354, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.763915592922022e-06, |
|
"loss": 1.4256, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.762761419044157e-06, |
|
"loss": 1.0847, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.761604571280939e-06, |
|
"loss": 1.0844, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.760445050999405e-06, |
|
"loss": 1.2039, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.759282859569749e-06, |
|
"loss": 1.2732, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.7581179983653224e-06, |
|
"loss": 1.2579, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.756950468762632e-06, |
|
"loss": 0.8733, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.755780272141338e-06, |
|
"loss": 1.8547, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.75460740988425e-06, |
|
"loss": 0.8285, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.753431883377329e-06, |
|
"loss": 1.0063, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.752253694009686e-06, |
|
"loss": 1.2905, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.7510728431735765e-06, |
|
"loss": 1.2867, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.749889332264401e-06, |
|
"loss": 1.15, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.748703162680704e-06, |
|
"loss": 1.2434, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.747514335824172e-06, |
|
"loss": 1.1005, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.746322853099631e-06, |
|
"loss": 0.7872, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.745128715915047e-06, |
|
"loss": 1.2451, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.743931925681519e-06, |
|
"loss": 0.8286, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.742732483813286e-06, |
|
"loss": 0.8608, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.741530391727716e-06, |
|
"loss": 0.7686, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.7403256508453125e-06, |
|
"loss": 0.9078, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.739118262589706e-06, |
|
"loss": 1.4336, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.737908228387656e-06, |
|
"loss": 0.7947, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"eval_loss": 1.2669733762741089, |
|
"eval_runtime": 53.5588, |
|
"eval_samples_per_second": 14.414, |
|
"eval_steps_per_second": 7.207, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.736695549669052e-06, |
|
"loss": 1.4744, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.735480227866902e-06, |
|
"loss": 1.2414, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.734262264417345e-06, |
|
"loss": 0.9221, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.733041660759635e-06, |
|
"loss": 1.1588, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.73181841833615e-06, |
|
"loss": 1.019, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.730592538592384e-06, |
|
"loss": 1.3191, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.72936402297695e-06, |
|
"loss": 0.9169, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.728132872941572e-06, |
|
"loss": 0.9911, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.726899089941091e-06, |
|
"loss": 1.1003, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.725662675433455e-06, |
|
"loss": 1.4743, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.724423630879727e-06, |
|
"loss": 0.9946, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.723181957744075e-06, |
|
"loss": 1.0176, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.721937657493771e-06, |
|
"loss": 1.0483, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.720690731599195e-06, |
|
"loss": 1.6443, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.71944118153383e-06, |
|
"loss": 1.2673, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.718189008774256e-06, |
|
"loss": 1.1349, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.716934214800155e-06, |
|
"loss": 1.5773, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.7156768010943066e-06, |
|
"loss": 1.4548, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.714416769142586e-06, |
|
"loss": 1.1942, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.71315412043396e-06, |
|
"loss": 1.4373, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.711888856460493e-06, |
|
"loss": 1.5013, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.710620978717333e-06, |
|
"loss": 1.3288, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.709350488702721e-06, |
|
"loss": 0.9713, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.708077387917984e-06, |
|
"loss": 1.3237, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.706801677867535e-06, |
|
"loss": 1.3598, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.705523360058868e-06, |
|
"loss": 1.0094, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.7042424360025606e-06, |
|
"loss": 1.5585, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.70295890721227e-06, |
|
"loss": 1.5537, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.70167277520473e-06, |
|
"loss": 0.9304, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.700384041499752e-06, |
|
"loss": 0.9573, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.699092707620222e-06, |
|
"loss": 1.66, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.697798775092096e-06, |
|
"loss": 0.9916, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.6965022454444046e-06, |
|
"loss": 1.482, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.695203120209245e-06, |
|
"loss": 1.0236, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.693901400921782e-06, |
|
"loss": 1.093, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.692597089120246e-06, |
|
"loss": 1.2824, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.691290186345931e-06, |
|
"loss": 1.7949, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.689980694143192e-06, |
|
"loss": 1.4492, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.688668614059445e-06, |
|
"loss": 1.6656, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.687353947645162e-06, |
|
"loss": 1.5922, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.6860366964538765e-06, |
|
"loss": 1.1517, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.6847168620421685e-06, |
|
"loss": 1.719, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.6833944459696765e-06, |
|
"loss": 1.0782, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.682069449799089e-06, |
|
"loss": 1.3512, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.680741875096141e-06, |
|
"loss": 1.9916, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.679411723429617e-06, |
|
"loss": 1.7266, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.678078996371344e-06, |
|
"loss": 1.1965, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.6767436954961956e-06, |
|
"loss": 1.8546, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.675405822382083e-06, |
|
"loss": 0.9484, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.674065378609962e-06, |
|
"loss": 1.787, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.672722365763821e-06, |
|
"loss": 1.5217, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.671376785430688e-06, |
|
"loss": 0.8603, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.670028639200622e-06, |
|
"loss": 0.9619, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.668677928666717e-06, |
|
"loss": 1.5593, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.667324655425095e-06, |
|
"loss": 0.9245, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.665968821074907e-06, |
|
"loss": 0.8437, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.66461042721833e-06, |
|
"loss": 1.0066, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.663249475460569e-06, |
|
"loss": 0.8552, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.661885967409846e-06, |
|
"loss": 0.7843, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.660519904677407e-06, |
|
"loss": 1.0128, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.659151288877517e-06, |
|
"loss": 1.1489, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.6577801216274565e-06, |
|
"loss": 1.3491, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.656406404547523e-06, |
|
"loss": 0.8747, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.6550301392610256e-06, |
|
"loss": 0.9925, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.653651327394284e-06, |
|
"loss": 1.6417, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.652269970576628e-06, |
|
"loss": 1.7158, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.6508860704403965e-06, |
|
"loss": 1.4892, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.649499628620931e-06, |
|
"loss": 0.9178, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.648110646756579e-06, |
|
"loss": 0.8608, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.6467191264886855e-06, |
|
"loss": 0.6383, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.6453250694616e-06, |
|
"loss": 1.1731, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.643928477322667e-06, |
|
"loss": 1.1921, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.642529351722227e-06, |
|
"loss": 1.4647, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.641127694313615e-06, |
|
"loss": 1.8037, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.6397235067531564e-06, |
|
"loss": 1.3182, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.638316790700168e-06, |
|
"loss": 1.6523, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.636907547816953e-06, |
|
"loss": 1.6149, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.635495779768803e-06, |
|
"loss": 0.9444, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.63408148822399e-06, |
|
"loss": 0.9974, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.632664674853772e-06, |
|
"loss": 0.9567, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.631245341332383e-06, |
|
"loss": 1.2978, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.629823489337039e-06, |
|
"loss": 1.3438, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.628399120547928e-06, |
|
"loss": 1.3149, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.626972236648214e-06, |
|
"loss": 1.098, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.625542839324036e-06, |
|
"loss": 1.7058, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.624110930264498e-06, |
|
"loss": 0.9291, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.622676511161675e-06, |
|
"loss": 0.846, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.6212395837106065e-06, |
|
"loss": 1.5026, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.619800149609298e-06, |
|
"loss": 0.9427, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.618358210558717e-06, |
|
"loss": 0.7821, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.616913768262789e-06, |
|
"loss": 1.4015, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.6154668244284e-06, |
|
"loss": 1.6013, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.614017380765389e-06, |
|
"loss": 1.6039, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.612565438986553e-06, |
|
"loss": 0.8531, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.611111000807637e-06, |
|
"loss": 1.149, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.60965406794734e-06, |
|
"loss": 1.6213, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.608194642127304e-06, |
|
"loss": 0.8759, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.606732725072122e-06, |
|
"loss": 1.0741, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.605268318509327e-06, |
|
"loss": 1.4488, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.603801424169397e-06, |
|
"loss": 1.1136, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.602332043785748e-06, |
|
"loss": 1.0564, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.600860179094732e-06, |
|
"loss": 1.7812, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.599385831835641e-06, |
|
"loss": 1.2911, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.597909003750697e-06, |
|
"loss": 1.0261, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.596429696585054e-06, |
|
"loss": 0.9223, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.594947912086799e-06, |
|
"loss": 0.936, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.593463652006941e-06, |
|
"loss": 0.8679, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.5919769180994185e-06, |
|
"loss": 1.05, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.590487712121092e-06, |
|
"loss": 0.8358, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.588996035831742e-06, |
|
"loss": 1.4855, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.587501890994071e-06, |
|
"loss": 1.3494, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.586005279373696e-06, |
|
"loss": 1.1508, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.58450620273915e-06, |
|
"loss": 0.9384, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.583004662861877e-06, |
|
"loss": 1.3447, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.581500661516237e-06, |
|
"loss": 1.4123, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.579994200479493e-06, |
|
"loss": 1.1999, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.578485281531817e-06, |
|
"loss": 1.0577, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.576973906456286e-06, |
|
"loss": 0.7701, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.575460077038877e-06, |
|
"loss": 1.0168, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.573943795068471e-06, |
|
"loss": 1.0581, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.572425062336844e-06, |
|
"loss": 1.2371, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.570903880638669e-06, |
|
"loss": 1.027, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.569380251771514e-06, |
|
"loss": 1.555, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.567854177535837e-06, |
|
"loss": 1.5865, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.566325659734987e-06, |
|
"loss": 1.59, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.564794700175202e-06, |
|
"loss": 0.7054, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.563261300665599e-06, |
|
"loss": 1.8609, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.561725463018186e-06, |
|
"loss": 1.3677, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.560187189047849e-06, |
|
"loss": 1.3487, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.558646480572352e-06, |
|
"loss": 1.2953, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.557103339412337e-06, |
|
"loss": 1.4691, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.555557767391319e-06, |
|
"loss": 1.4562, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.5540097663356885e-06, |
|
"loss": 1.0457, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.5524593380747055e-06, |
|
"loss": 0.9963, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.550906484440495e-06, |
|
"loss": 1.1226, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.5493512072680535e-06, |
|
"loss": 1.4231, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.547793508395235e-06, |
|
"loss": 1.1163, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.546233389662763e-06, |
|
"loss": 0.9493, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.5446708529142116e-06, |
|
"loss": 1.4465, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.54310589999602e-06, |
|
"loss": 0.8873, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.541538532757478e-06, |
|
"loss": 1.456, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.53996875305073e-06, |
|
"loss": 0.9274, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.5383965627307705e-06, |
|
"loss": 1.3981, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.536821963655443e-06, |
|
"loss": 1.3882, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.535244957685438e-06, |
|
"loss": 1.3814, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"eval_loss": 1.2516705989837646, |
|
"eval_runtime": 53.6602, |
|
"eval_samples_per_second": 14.387, |
|
"eval_steps_per_second": 7.193, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.5336655466842904e-06, |
|
"loss": 1.1422, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.532083732518375e-06, |
|
"loss": 1.042, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.530499517056909e-06, |
|
"loss": 1.1034, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.528912902171944e-06, |
|
"loss": 0.7851, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.52732388973837e-06, |
|
"loss": 1.4648, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.52573248163391e-06, |
|
"loss": 0.8926, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.524138679739115e-06, |
|
"loss": 1.1289, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.522542485937369e-06, |
|
"loss": 1.074, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.520943902114878e-06, |
|
"loss": 1.179, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.519342930160678e-06, |
|
"loss": 1.0436, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.517739571966618e-06, |
|
"loss": 0.9787, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.516133829427377e-06, |
|
"loss": 1.4049, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.514525704440446e-06, |
|
"loss": 1.0072, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.512915198906131e-06, |
|
"loss": 0.9599, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.511302314727552e-06, |
|
"loss": 0.9298, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.509687053810643e-06, |
|
"loss": 0.885, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.508069418064141e-06, |
|
"loss": 0.7679, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.506449409399592e-06, |
|
"loss": 1.0181, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.504827029731347e-06, |
|
"loss": 1.3923, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.503202280976557e-06, |
|
"loss": 1.4239, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.5015751650551756e-06, |
|
"loss": 1.1096, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.499945683889948e-06, |
|
"loss": 1.6356, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.4983138394064205e-06, |
|
"loss": 1.4299, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.496679633532929e-06, |
|
"loss": 0.624, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.4950430682005995e-06, |
|
"loss": 1.2829, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.493404145343347e-06, |
|
"loss": 1.2277, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.491762866897872e-06, |
|
"loss": 1.4662, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.49011923480366e-06, |
|
"loss": 1.0535, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.488473251002976e-06, |
|
"loss": 1.3935, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.486824917440865e-06, |
|
"loss": 0.6957, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.485174236065149e-06, |
|
"loss": 1.4466, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.483521208826421e-06, |
|
"loss": 0.655, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.4818658376780514e-06, |
|
"loss": 1.1986, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.4802081245761765e-06, |
|
"loss": 1.8627, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.478548071479702e-06, |
|
"loss": 0.998, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.476885680350298e-06, |
|
"loss": 1.7648, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.475220953152397e-06, |
|
"loss": 1.2204, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.473553891853194e-06, |
|
"loss": 1.2936, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.471884498422639e-06, |
|
"loss": 1.1346, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.4702127748334406e-06, |
|
"loss": 1.2736, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.46853872306106e-06, |
|
"loss": 1.1092, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.4668623450837085e-06, |
|
"loss": 1.0157, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.465183642882348e-06, |
|
"loss": 1.3962, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.463502618440687e-06, |
|
"loss": 1.6577, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.461819273745176e-06, |
|
"loss": 1.1963, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.460133610785009e-06, |
|
"loss": 0.9, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.458445631552119e-06, |
|
"loss": 0.961, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.456755338041178e-06, |
|
"loss": 1.7704, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.455062732249589e-06, |
|
"loss": 1.36, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.4533678161774895e-06, |
|
"loss": 0.8109, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.451670591827749e-06, |
|
"loss": 1.3852, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.44997106120596e-06, |
|
"loss": 0.8871, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.4482692263204455e-06, |
|
"loss": 1.3472, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.4465650891822466e-06, |
|
"loss": 1.0071, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.444858651805128e-06, |
|
"loss": 1.2652, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.4431499162055724e-06, |
|
"loss": 1.4467, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.441438884402777e-06, |
|
"loss": 1.472, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.4397255584186534e-06, |
|
"loss": 1.4468, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.438009940277825e-06, |
|
"loss": 1.1595, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.436292032007621e-06, |
|
"loss": 1.4476, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.4345718356380795e-06, |
|
"loss": 1.299, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.432849353201941e-06, |
|
"loss": 0.6383, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.431124586734648e-06, |
|
"loss": 1.1705, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.429397538274342e-06, |
|
"loss": 1.2946, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.4276682098618604e-06, |
|
"loss": 0.9894, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.4259366035407375e-06, |
|
"loss": 0.8213, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.4242027213571954e-06, |
|
"loss": 1.7667, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.4224665653601476e-06, |
|
"loss": 0.7915, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.420728137601196e-06, |
|
"loss": 1.249, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.418987440134624e-06, |
|
"loss": 1.3703, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.417244475017398e-06, |
|
"loss": 1.3207, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.415499244309166e-06, |
|
"loss": 1.793, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.4137517500722516e-06, |
|
"loss": 1.4352, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.412001994371653e-06, |
|
"loss": 1.8998, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.410249979275042e-06, |
|
"loss": 1.3503, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.408495706852758e-06, |
|
"loss": 0.7287, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.4067391791778104e-06, |
|
"loss": 1.3658, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.404980398325873e-06, |
|
"loss": 0.9159, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.40321936637528e-06, |
|
"loss": 1.7305, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.40145608540703e-06, |
|
"loss": 0.8456, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.399690557504774e-06, |
|
"loss": 0.8557, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.397922784754824e-06, |
|
"loss": 0.8167, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.396152769246137e-06, |
|
"loss": 1.002, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.39438051307033e-06, |
|
"loss": 1.3232, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.392606018321659e-06, |
|
"loss": 1.5761, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.390829287097029e-06, |
|
"loss": 1.1345, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.389050321495991e-06, |
|
"loss": 1.6143, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.38726912362073e-06, |
|
"loss": 0.9397, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.385485695576071e-06, |
|
"loss": 1.3128, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.38370003946948e-06, |
|
"loss": 1.3883, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.381912157411046e-06, |
|
"loss": 0.8553, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.380122051513497e-06, |
|
"loss": 1.1036, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.378329723892184e-06, |
|
"loss": 1.2276, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.376535176665085e-06, |
|
"loss": 1.4422, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.374738411952801e-06, |
|
"loss": 1.1659, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.372939431878552e-06, |
|
"loss": 1.5647, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.371138238568179e-06, |
|
"loss": 1.3946, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.369334834150136e-06, |
|
"loss": 0.7131, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.367529220755489e-06, |
|
"loss": 1.0276, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.365721400517916e-06, |
|
"loss": 1.437, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.363911375573702e-06, |
|
"loss": 0.9089, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.362099148061739e-06, |
|
"loss": 1.0697, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.360284720123519e-06, |
|
"loss": 1.4068, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.358468093903136e-06, |
|
"loss": 0.8719, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.356649271547282e-06, |
|
"loss": 0.9841, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.354828255205242e-06, |
|
"loss": 1.541, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.3530050470288955e-06, |
|
"loss": 1.0059, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.3511796491727135e-06, |
|
"loss": 1.1743, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.34935206379375e-06, |
|
"loss": 1.5674, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.3475222930516484e-06, |
|
"loss": 1.3795, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.3456903391086315e-06, |
|
"loss": 0.824, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.3438562041295045e-06, |
|
"loss": 0.8724, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.342019890281648e-06, |
|
"loss": 1.026, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.340181399735018e-06, |
|
"loss": 0.9268, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.338340734662143e-06, |
|
"loss": 1.4699, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.33649789723812e-06, |
|
"loss": 1.5072, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.334652889640615e-06, |
|
"loss": 1.0228, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.332805714049856e-06, |
|
"loss": 0.9188, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.330956372648635e-06, |
|
"loss": 1.0169, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.329104867622302e-06, |
|
"loss": 0.7205, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.3272512011587665e-06, |
|
"loss": 1.1265, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.325395375448487e-06, |
|
"loss": 1.3212, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.323537392684477e-06, |
|
"loss": 0.8664, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.3216772550623e-06, |
|
"loss": 1.3096, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.319814964780063e-06, |
|
"loss": 1.6955, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.317950524038419e-06, |
|
"loss": 1.4587, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.316083935040561e-06, |
|
"loss": 1.2825, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.31421519999222e-06, |
|
"loss": 1.1505, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.3123443211016655e-06, |
|
"loss": 1.1944, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.310471300579697e-06, |
|
"loss": 1.0146, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.308596140639647e-06, |
|
"loss": 1.1227, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.3067188434973765e-06, |
|
"loss": 1.0811, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.3048394113712696e-06, |
|
"loss": 0.9842, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.302957846482236e-06, |
|
"loss": 1.7324, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.3010741510537046e-06, |
|
"loss": 1.1923, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.299188327311621e-06, |
|
"loss": 1.1261, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.2973003774844464e-06, |
|
"loss": 1.5921, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.295410303803155e-06, |
|
"loss": 0.9543, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.2935181085012306e-06, |
|
"loss": 1.0813, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.2916237938146635e-06, |
|
"loss": 1.7461, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.2897273619819494e-06, |
|
"loss": 1.2278, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.287828815244083e-06, |
|
"loss": 0.8105, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.2859281558445635e-06, |
|
"loss": 0.8013, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.284025386029381e-06, |
|
"loss": 0.977, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.282120508047023e-06, |
|
"loss": 1.7712, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"eval_loss": 1.2422988414764404, |
|
"eval_runtime": 53.6577, |
|
"eval_samples_per_second": 14.387, |
|
"eval_steps_per_second": 7.194, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.2802135241484676e-06, |
|
"loss": 0.9182, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.27830443658718e-06, |
|
"loss": 0.9902, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.2763932476191125e-06, |
|
"loss": 1.3087, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.274479959502701e-06, |
|
"loss": 0.7967, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.27256457449886e-06, |
|
"loss": 1.2144, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.270647094870985e-06, |
|
"loss": 0.8462, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.268727522884942e-06, |
|
"loss": 1.4271, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.266805860809076e-06, |
|
"loss": 1.2957, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.264882110914196e-06, |
|
"loss": 1.1985, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.262956275473581e-06, |
|
"loss": 1.2892, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.261028356762973e-06, |
|
"loss": 1.0717, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.259098357060578e-06, |
|
"loss": 0.948, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.25716627864706e-06, |
|
"loss": 0.9578, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.255232123805537e-06, |
|
"loss": 1.4776, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.253295894821585e-06, |
|
"loss": 0.6842, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.251357593983228e-06, |
|
"loss": 1.3206, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.249417223580939e-06, |
|
"loss": 1.1884, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.247474785907636e-06, |
|
"loss": 1.0852, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.245530283258682e-06, |
|
"loss": 0.9386, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.24358371793188e-06, |
|
"loss": 1.8676, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.241635092227465e-06, |
|
"loss": 0.8333, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.239684408448114e-06, |
|
"loss": 1.4695, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.2377316688989326e-06, |
|
"loss": 1.4006, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.235776875887455e-06, |
|
"loss": 1.5704, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.233820031723644e-06, |
|
"loss": 0.989, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.231861138719886e-06, |
|
"loss": 0.8683, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.229900199190985e-06, |
|
"loss": 1.1916, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.227937215454169e-06, |
|
"loss": 1.6889, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.225972189829076e-06, |
|
"loss": 0.9639, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.224005124637762e-06, |
|
"loss": 1.4438, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.222036022204689e-06, |
|
"loss": 0.6203, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.220064884856729e-06, |
|
"loss": 0.8918, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.218091714923157e-06, |
|
"loss": 0.9794, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.216116514735652e-06, |
|
"loss": 1.5727, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.21413928662829e-06, |
|
"loss": 1.9055, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.212160032937543e-06, |
|
"loss": 1.7443, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.210178756002279e-06, |
|
"loss": 1.2116, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.208195458163755e-06, |
|
"loss": 0.892, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.206210141765618e-06, |
|
"loss": 1.3735, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.204222809153897e-06, |
|
"loss": 1.1583, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.202233462677008e-06, |
|
"loss": 1.0731, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.200242104685742e-06, |
|
"loss": 0.9748, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.198248737533271e-06, |
|
"loss": 1.2398, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.1962533635751365e-06, |
|
"loss": 1.5763, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.194255985169259e-06, |
|
"loss": 0.7805, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.19225660467592e-06, |
|
"loss": 1.4626, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.19025522445777e-06, |
|
"loss": 0.7902, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.1882518468798235e-06, |
|
"loss": 1.3892, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.186246474309455e-06, |
|
"loss": 0.9948, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.184239109116393e-06, |
|
"loss": 1.5633, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.182229753672726e-06, |
|
"loss": 0.8679, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.18021841035289e-06, |
|
"loss": 0.7651, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.1782050815336725e-06, |
|
"loss": 1.3411, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.1761897695942076e-06, |
|
"loss": 1.6998, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.17417247691597e-06, |
|
"loss": 1.1193, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.172153205882777e-06, |
|
"loss": 1.437, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.170131958880784e-06, |
|
"loss": 1.0518, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.168108738298481e-06, |
|
"loss": 0.9892, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.166083546526689e-06, |
|
"loss": 1.4121, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.164056385958561e-06, |
|
"loss": 1.8558, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.162027258989575e-06, |
|
"loss": 1.6905, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.159996168017531e-06, |
|
"loss": 1.3669, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.157963115442552e-06, |
|
"loss": 0.9603, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.1559281036670785e-06, |
|
"loss": 0.7815, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.153891135095866e-06, |
|
"loss": 1.6807, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.151852212135984e-06, |
|
"loss": 0.5376, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.149811337196808e-06, |
|
"loss": 1.0208, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.1477685126900215e-06, |
|
"loss": 1.3474, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.145723741029613e-06, |
|
"loss": 1.1992, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.1436770246318725e-06, |
|
"loss": 1.1475, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.141628365915386e-06, |
|
"loss": 1.0059, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.139577767301035e-06, |
|
"loss": 0.8612, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.137525231211995e-06, |
|
"loss": 1.7248, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.1354707600737286e-06, |
|
"loss": 1.5388, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.133414356313988e-06, |
|
"loss": 1.2048, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.131356022362807e-06, |
|
"loss": 0.9788, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.129295760652499e-06, |
|
"loss": 1.2049, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.12723357361766e-06, |
|
"loss": 1.7343, |
|
"step": 803 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.125169463695156e-06, |
|
"loss": 0.7971, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.1231034333241295e-06, |
|
"loss": 0.8876, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.12103548494599e-06, |
|
"loss": 1.6504, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.118965621004413e-06, |
|
"loss": 1.1006, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.116893843945339e-06, |
|
"loss": 0.9648, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.114820156216969e-06, |
|
"loss": 0.859, |
|
"step": 809 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.112744560269762e-06, |
|
"loss": 1.0627, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.11066705855643e-06, |
|
"loss": 1.1119, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.108587653531942e-06, |
|
"loss": 1.1377, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.106506347653509e-06, |
|
"loss": 0.7878, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.104423143380594e-06, |
|
"loss": 1.485, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.1023380431749e-06, |
|
"loss": 1.3263, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.100251049500373e-06, |
|
"loss": 1.4766, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.098162164823194e-06, |
|
"loss": 1.094, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.096071391611781e-06, |
|
"loss": 0.9679, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.093978732336781e-06, |
|
"loss": 0.9718, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.091884189471071e-06, |
|
"loss": 1.0933, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.089787765489757e-06, |
|
"loss": 1.2173, |
|
"step": 821 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.087689462870161e-06, |
|
"loss": 1.3192, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.08558928409183e-06, |
|
"loss": 1.3672, |
|
"step": 823 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.083487231636528e-06, |
|
"loss": 0.9721, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.081383307988232e-06, |
|
"loss": 0.8727, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.079277515633127e-06, |
|
"loss": 1.194, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.077169857059614e-06, |
|
"loss": 0.7615, |
|
"step": 827 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.0750603347582915e-06, |
|
"loss": 1.2318, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.072948951221965e-06, |
|
"loss": 0.8819, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.070835708945637e-06, |
|
"loss": 1.4115, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.068720610426508e-06, |
|
"loss": 0.7792, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.066603658163972e-06, |
|
"loss": 0.8577, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.064484854659611e-06, |
|
"loss": 1.3564, |
|
"step": 833 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.062364202417198e-06, |
|
"loss": 1.0379, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.060241703942689e-06, |
|
"loss": 1.089, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.058117361744222e-06, |
|
"loss": 0.9309, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.055991178332113e-06, |
|
"loss": 1.0059, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.053863156218855e-06, |
|
"loss": 0.98, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.0517332979191126e-06, |
|
"loss": 1.2531, |
|
"step": 839 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.04960160594972e-06, |
|
"loss": 1.6179, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.0474680828296805e-06, |
|
"loss": 0.9697, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.045332731080159e-06, |
|
"loss": 1.9022, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.043195553224482e-06, |
|
"loss": 1.8396, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.041056551788134e-06, |
|
"loss": 1.2137, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.038915729298755e-06, |
|
"loss": 1.0898, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.0367730882861335e-06, |
|
"loss": 0.9471, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.0346286312822116e-06, |
|
"loss": 0.891, |
|
"step": 847 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.032482360821076e-06, |
|
"loss": 1.0731, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.030334279438954e-06, |
|
"loss": 1.1644, |
|
"step": 849 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.028184389674216e-06, |
|
"loss": 1.1629, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.0260326940673665e-06, |
|
"loss": 1.4033, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.023879195161047e-06, |
|
"loss": 0.9199, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.021723895500027e-06, |
|
"loss": 1.6712, |
|
"step": 853 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.019566797631206e-06, |
|
"loss": 1.7075, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.017407904103607e-06, |
|
"loss": 1.2572, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.015247217468377e-06, |
|
"loss": 1.4046, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.013084740278778e-06, |
|
"loss": 1.5426, |
|
"step": 857 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.010920475090192e-06, |
|
"loss": 1.4335, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.008754424460112e-06, |
|
"loss": 0.8425, |
|
"step": 859 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.006586590948141e-06, |
|
"loss": 1.1965, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.004416977115989e-06, |
|
"loss": 0.8491, |
|
"step": 861 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.00224558552747e-06, |
|
"loss": 0.9847, |
|
"step": 862 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.000072418748498e-06, |
|
"loss": 1.106, |
|
"step": 863 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.9978974793470855e-06, |
|
"loss": 1.0784, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.9957207698933385e-06, |
|
"loss": 1.0359, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.993542292959456e-06, |
|
"loss": 0.9047, |
|
"step": 866 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.991362051119725e-06, |
|
"loss": 0.7676, |
|
"step": 867 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.989180046950518e-06, |
|
"loss": 0.6939, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.986996283030291e-06, |
|
"loss": 1.286, |
|
"step": 869 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.984810761939578e-06, |
|
"loss": 1.2697, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"eval_loss": 1.2340365648269653, |
|
"eval_runtime": 54.4194, |
|
"eval_samples_per_second": 14.186, |
|
"eval_steps_per_second": 7.093, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.98262348626099e-06, |
|
"loss": 1.2608, |
|
"step": 871 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.980434458579213e-06, |
|
"loss": 1.5733, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.978243681481e-06, |
|
"loss": 0.756, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.976051157555174e-06, |
|
"loss": 1.592, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.973856889392622e-06, |
|
"loss": 1.2678, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.97166087958629e-06, |
|
"loss": 1.2461, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.969463130731183e-06, |
|
"loss": 1.2391, |
|
"step": 877 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.967263645424364e-06, |
|
"loss": 1.6344, |
|
"step": 878 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.965062426264944e-06, |
|
"loss": 1.6889, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.962859475854085e-06, |
|
"loss": 0.9765, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.960654796794993e-06, |
|
"loss": 0.7538, |
|
"step": 881 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.9584483916929195e-06, |
|
"loss": 1.0218, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.956240263155153e-06, |
|
"loss": 1.0937, |
|
"step": 883 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.954030413791021e-06, |
|
"loss": 1.7373, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.951818846211882e-06, |
|
"loss": 0.9229, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.949605563031126e-06, |
|
"loss": 0.766, |
|
"step": 886 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.9473905668641705e-06, |
|
"loss": 1.7482, |
|
"step": 887 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.945173860328457e-06, |
|
"loss": 0.855, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.942955446043448e-06, |
|
"loss": 1.6474, |
|
"step": 889 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.940735326630626e-06, |
|
"loss": 1.6998, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.938513504713485e-06, |
|
"loss": 1.1006, |
|
"step": 891 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.936289982917532e-06, |
|
"loss": 0.9537, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.9340647638702835e-06, |
|
"loss": 1.0979, |
|
"step": 893 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.9318378502012636e-06, |
|
"loss": 1.5734, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.929609244541994e-06, |
|
"loss": 1.5724, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.9273789495259995e-06, |
|
"loss": 1.1793, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.925146967788801e-06, |
|
"loss": 1.1859, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.92291330196791e-06, |
|
"loss": 1.4159, |
|
"step": 898 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.9206779547028305e-06, |
|
"loss": 1.6399, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.9184409286350535e-06, |
|
"loss": 1.4543, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.916202226408053e-06, |
|
"loss": 0.7884, |
|
"step": 901 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.913961850667283e-06, |
|
"loss": 1.5351, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.911719804060177e-06, |
|
"loss": 0.8836, |
|
"step": 903 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.9094760892361405e-06, |
|
"loss": 0.8275, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.907230708846552e-06, |
|
"loss": 1.0969, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.904983665544759e-06, |
|
"loss": 1.7432, |
|
"step": 906 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.9027349619860726e-06, |
|
"loss": 0.8615, |
|
"step": 907 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.900484600827765e-06, |
|
"loss": 1.4718, |
|
"step": 908 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.898232584729069e-06, |
|
"loss": 1.5248, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.895978916351172e-06, |
|
"loss": 1.6631, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.893723598357214e-06, |
|
"loss": 1.0259, |
|
"step": 911 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.891466633412285e-06, |
|
"loss": 1.6557, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.889208024183422e-06, |
|
"loss": 1.5224, |
|
"step": 913 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.886947773339601e-06, |
|
"loss": 0.9531, |
|
"step": 914 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.884685883551742e-06, |
|
"loss": 1.0861, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.882422357492701e-06, |
|
"loss": 0.8796, |
|
"step": 916 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.880157197837267e-06, |
|
"loss": 1.6122, |
|
"step": 917 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.877890407262157e-06, |
|
"loss": 1.6063, |
|
"step": 918 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.875621988446019e-06, |
|
"loss": 0.8907, |
|
"step": 919 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.873351944069425e-06, |
|
"loss": 0.834, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.8710802768148644e-06, |
|
"loss": 0.9929, |
|
"step": 921 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.868806989366748e-06, |
|
"loss": 1.2989, |
|
"step": 922 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.8665320844114e-06, |
|
"loss": 1.3995, |
|
"step": 923 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.864255564637053e-06, |
|
"loss": 1.8592, |
|
"step": 924 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.861977432733853e-06, |
|
"loss": 0.8865, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.859697691393849e-06, |
|
"loss": 1.1653, |
|
"step": 926 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.857416343310989e-06, |
|
"loss": 1.4607, |
|
"step": 927 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.855133391181124e-06, |
|
"loss": 1.2872, |
|
"step": 928 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.852848837701997e-06, |
|
"loss": 0.947, |
|
"step": 929 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.850562685573246e-06, |
|
"loss": 0.7027, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.848274937496397e-06, |
|
"loss": 0.9595, |
|
"step": 931 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.845985596174861e-06, |
|
"loss": 1.1452, |
|
"step": 932 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.843694664313932e-06, |
|
"loss": 1.1997, |
|
"step": 933 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.8414021446207866e-06, |
|
"loss": 1.6929, |
|
"step": 934 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.839108039804473e-06, |
|
"loss": 1.2888, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.836812352575916e-06, |
|
"loss": 1.4976, |
|
"step": 936 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.834515085647909e-06, |
|
"loss": 1.5642, |
|
"step": 937 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.832216241735111e-06, |
|
"loss": 1.3293, |
|
"step": 938 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.8299158235540475e-06, |
|
"loss": 1.3143, |
|
"step": 939 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.827613833823101e-06, |
|
"loss": 1.5578, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.825310275262513e-06, |
|
"loss": 0.7517, |
|
"step": 941 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.823005150594379e-06, |
|
"loss": 0.9068, |
|
"step": 942 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.820698462542644e-06, |
|
"loss": 1.0884, |
|
"step": 943 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.818390213833102e-06, |
|
"loss": 0.7859, |
|
"step": 944 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.81608040719339e-06, |
|
"loss": 1.4531, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.8137690453529854e-06, |
|
"loss": 1.4531, |
|
"step": 946 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.811456131043205e-06, |
|
"loss": 1.3391, |
|
"step": 947 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.8091416669971994e-06, |
|
"loss": 1.247, |
|
"step": 948 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.8068256559499493e-06, |
|
"loss": 1.1898, |
|
"step": 949 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.8045081006382668e-06, |
|
"loss": 1.0309, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.8021890038007848e-06, |
|
"loss": 0.9196, |
|
"step": 951 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.79986836817796e-06, |
|
"loss": 1.236, |
|
"step": 952 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.79754619651207e-06, |
|
"loss": 1.0914, |
|
"step": 953 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.7952224915472003e-06, |
|
"loss": 1.0782, |
|
"step": 954 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.7928972560292556e-06, |
|
"loss": 0.898, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.790570492705945e-06, |
|
"loss": 1.2418, |
|
"step": 956 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.788242204326785e-06, |
|
"loss": 0.5817, |
|
"step": 957 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.785912393643094e-06, |
|
"loss": 1.2435, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.7835810634079888e-06, |
|
"loss": 1.692, |
|
"step": 959 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.7812482163763807e-06, |
|
"loss": 1.5054, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.7789138553049767e-06, |
|
"loss": 0.7875, |
|
"step": 961 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.7765779829522674e-06, |
|
"loss": 1.6516, |
|
"step": 962 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.774240602078536e-06, |
|
"loss": 0.909, |
|
"step": 963 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.771901715445845e-06, |
|
"loss": 1.5424, |
|
"step": 964 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.769561325818033e-06, |
|
"loss": 1.1065, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.7672194359607213e-06, |
|
"loss": 1.0141, |
|
"step": 966 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.7648760486412993e-06, |
|
"loss": 1.173, |
|
"step": 967 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.762531166628927e-06, |
|
"loss": 1.3394, |
|
"step": 968 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.7601847926945318e-06, |
|
"loss": 1.8383, |
|
"step": 969 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.757836929610803e-06, |
|
"loss": 0.7631, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.7554875801521882e-06, |
|
"loss": 0.7929, |
|
"step": 971 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.7531367470948953e-06, |
|
"loss": 1.2148, |
|
"step": 972 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.750784433216883e-06, |
|
"loss": 1.717, |
|
"step": 973 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.7484306412978583e-06, |
|
"loss": 1.0289, |
|
"step": 974 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.7460753741192777e-06, |
|
"loss": 1.4954, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.7437186344643395e-06, |
|
"loss": 1.3094, |
|
"step": 976 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.7413604251179826e-06, |
|
"loss": 1.6131, |
|
"step": 977 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.739000748866882e-06, |
|
"loss": 1.0538, |
|
"step": 978 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.736639608499448e-06, |
|
"loss": 1.4786, |
|
"step": 979 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.7342770068058165e-06, |
|
"loss": 1.5938, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.731912946577855e-06, |
|
"loss": 1.3353, |
|
"step": 981 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.729547430609153e-06, |
|
"loss": 0.8837, |
|
"step": 982 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.727180461695019e-06, |
|
"loss": 1.1192, |
|
"step": 983 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.724812042632481e-06, |
|
"loss": 0.8178, |
|
"step": 984 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.722442176220278e-06, |
|
"loss": 0.7357, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.7200708652588603e-06, |
|
"loss": 1.5317, |
|
"step": 986 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.7176981125503846e-06, |
|
"loss": 0.7846, |
|
"step": 987 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.7153239208987136e-06, |
|
"loss": 0.8197, |
|
"step": 988 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.7129482931094085e-06, |
|
"loss": 1.4203, |
|
"step": 989 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.710571231989727e-06, |
|
"loss": 0.6815, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.7081927403486217e-06, |
|
"loss": 1.0855, |
|
"step": 991 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.705812820996736e-06, |
|
"loss": 1.5208, |
|
"step": 992 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.7034314767463985e-06, |
|
"loss": 1.912, |
|
"step": 993 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.7010487104116242e-06, |
|
"loss": 0.7749, |
|
"step": 994 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.6986645248081066e-06, |
|
"loss": 1.3164, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.6962789227532165e-06, |
|
"loss": 1.1781, |
|
"step": 996 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.6938919070660003e-06, |
|
"loss": 1.2583, |
|
"step": 997 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.691503480567171e-06, |
|
"loss": 1.3813, |
|
"step": 998 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.6891136460791143e-06, |
|
"loss": 1.2476, |
|
"step": 999 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.6867224064258754e-06, |
|
"loss": 0.9396, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.6843297644331612e-06, |
|
"loss": 0.8906, |
|
"step": 1001 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.6819357229283346e-06, |
|
"loss": 0.8506, |
|
"step": 1002 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.6795402847404153e-06, |
|
"loss": 0.9023, |
|
"step": 1003 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.677143452700071e-06, |
|
"loss": 1.518, |
|
"step": 1004 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.674745229639617e-06, |
|
"loss": 1.354, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.6723456183930128e-06, |
|
"loss": 1.6563, |
|
"step": 1006 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.669944621795857e-06, |
|
"loss": 1.2295, |
|
"step": 1007 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.6675422426853883e-06, |
|
"loss": 1.6811, |
|
"step": 1008 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.665138483900475e-06, |
|
"loss": 1.1634, |
|
"step": 1009 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.662733348281619e-06, |
|
"loss": 1.6505, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.660326838670948e-06, |
|
"loss": 1.9781, |
|
"step": 1011 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.6579189579122115e-06, |
|
"loss": 1.4078, |
|
"step": 1012 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.6555097088507837e-06, |
|
"loss": 1.2795, |
|
"step": 1013 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.6530990943336515e-06, |
|
"loss": 0.8649, |
|
"step": 1014 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.6506871172094173e-06, |
|
"loss": 1.4015, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"eval_loss": 1.2239829301834106, |
|
"eval_runtime": 53.5468, |
|
"eval_samples_per_second": 14.417, |
|
"eval_steps_per_second": 7.209, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.648273780328293e-06, |
|
"loss": 0.8306, |
|
"step": 1016 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.6458590865420995e-06, |
|
"loss": 0.7629, |
|
"step": 1017 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.643443038704257e-06, |
|
"loss": 0.7034, |
|
"step": 1018 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.6410256396697895e-06, |
|
"loss": 1.8372, |
|
"step": 1019 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.638606892295315e-06, |
|
"loss": 1.2091, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.6361867994390478e-06, |
|
"loss": 0.8437, |
|
"step": 1021 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.633765363960789e-06, |
|
"loss": 0.8183, |
|
"step": 1022 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.6313425887219276e-06, |
|
"loss": 0.9, |
|
"step": 1023 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.628918476585437e-06, |
|
"loss": 1.1369, |
|
"step": 1024 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.6264930304158697e-06, |
|
"loss": 1.0595, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.6240662530793518e-06, |
|
"loss": 0.7621, |
|
"step": 1026 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.6216381474435867e-06, |
|
"loss": 1.5716, |
|
"step": 1027 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.6192087163778456e-06, |
|
"loss": 0.7501, |
|
"step": 1028 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.6167779627529654e-06, |
|
"loss": 0.8721, |
|
"step": 1029 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.6143458894413463e-06, |
|
"loss": 0.8237, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.611912499316948e-06, |
|
"loss": 1.267, |
|
"step": 1031 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.6094777952552864e-06, |
|
"loss": 1.206, |
|
"step": 1032 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.60704178013343e-06, |
|
"loss": 0.8337, |
|
"step": 1033 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.6046044568299973e-06, |
|
"loss": 1.7589, |
|
"step": 1034 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.6021658282251504e-06, |
|
"loss": 0.9766, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.5997258972005967e-06, |
|
"loss": 1.5177, |
|
"step": 1036 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.59728466663958e-06, |
|
"loss": 1.4695, |
|
"step": 1037 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.594842139426882e-06, |
|
"loss": 0.9031, |
|
"step": 1038 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.5923983184488167e-06, |
|
"loss": 1.1727, |
|
"step": 1039 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.5899532065932236e-06, |
|
"loss": 1.071, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.5875068067494716e-06, |
|
"loss": 1.2549, |
|
"step": 1041 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.5850591218084498e-06, |
|
"loss": 0.7977, |
|
"step": 1042 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.5826101546625653e-06, |
|
"loss": 0.8636, |
|
"step": 1043 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.580159908205742e-06, |
|
"loss": 1.8348, |
|
"step": 1044 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.5777083853334145e-06, |
|
"loss": 1.0609, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.575255588942525e-06, |
|
"loss": 0.9279, |
|
"step": 1046 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.5728015219315226e-06, |
|
"loss": 1.1864, |
|
"step": 1047 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.5703461872003566e-06, |
|
"loss": 1.1254, |
|
"step": 1048 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.5678895876504745e-06, |
|
"loss": 1.0267, |
|
"step": 1049 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.5654317261848182e-06, |
|
"loss": 1.3926, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.562972605707821e-06, |
|
"loss": 0.8434, |
|
"step": 1051 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.5605122291254048e-06, |
|
"loss": 1.1, |
|
"step": 1052 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.5580505993449744e-06, |
|
"loss": 0.8945, |
|
"step": 1053 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.555587719275416e-06, |
|
"loss": 0.8897, |
|
"step": 1054 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.5531235918270934e-06, |
|
"loss": 1.0842, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.5506582199118457e-06, |
|
"loss": 0.919, |
|
"step": 1056 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.5481916064429796e-06, |
|
"loss": 0.9278, |
|
"step": 1057 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.545723754335272e-06, |
|
"loss": 1.544, |
|
"step": 1058 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.5432546665049605e-06, |
|
"loss": 0.7575, |
|
"step": 1059 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.540784345869746e-06, |
|
"loss": 1.0788, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.538312795348785e-06, |
|
"loss": 0.7757, |
|
"step": 1061 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.5358400178626856e-06, |
|
"loss": 1.0101, |
|
"step": 1062 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.533366016333509e-06, |
|
"loss": 1.2068, |
|
"step": 1063 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.530890793684759e-06, |
|
"loss": 0.8573, |
|
"step": 1064 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.5284143528413866e-06, |
|
"loss": 1.2888, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.525936696729779e-06, |
|
"loss": 1.4604, |
|
"step": 1066 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.523457828277762e-06, |
|
"loss": 0.877, |
|
"step": 1067 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.520977750414591e-06, |
|
"loss": 1.7115, |
|
"step": 1068 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.518496466070954e-06, |
|
"loss": 1.3532, |
|
"step": 1069 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.5160139781789615e-06, |
|
"loss": 1.1231, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.5135302896721486e-06, |
|
"loss": 1.1528, |
|
"step": 1071 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.5110454034854684e-06, |
|
"loss": 0.8566, |
|
"step": 1072 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.5085593225552887e-06, |
|
"loss": 1.4349, |
|
"step": 1073 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.5060720498193895e-06, |
|
"loss": 1.0565, |
|
"step": 1074 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.50358358821696e-06, |
|
"loss": 0.8588, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.501093940688592e-06, |
|
"loss": 1.8337, |
|
"step": 1076 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.4986031101762806e-06, |
|
"loss": 0.9244, |
|
"step": 1077 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.4961110996234205e-06, |
|
"loss": 0.6638, |
|
"step": 1078 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.4936179119747958e-06, |
|
"loss": 1.3721, |
|
"step": 1079 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.491123550176586e-06, |
|
"loss": 1.8717, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.4886280171763563e-06, |
|
"loss": 1.4513, |
|
"step": 1081 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.486131315923057e-06, |
|
"loss": 1.0628, |
|
"step": 1082 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.483633449367017e-06, |
|
"loss": 1.0992, |
|
"step": 1083 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.481134420459944e-06, |
|
"loss": 0.6331, |
|
"step": 1084 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.4786342321549178e-06, |
|
"loss": 0.8425, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.47613288740639e-06, |
|
"loss": 0.9034, |
|
"step": 1086 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.4736303891701766e-06, |
|
"loss": 1.6633, |
|
"step": 1087 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.471126740403459e-06, |
|
"loss": 1.4944, |
|
"step": 1088 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.4686219440647774e-06, |
|
"loss": 1.9144, |
|
"step": 1089 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.4661160031140257e-06, |
|
"loss": 0.8263, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.463608920512454e-06, |
|
"loss": 0.8001, |
|
"step": 1091 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.4611006992226594e-06, |
|
"loss": 0.8235, |
|
"step": 1092 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.4585913422085848e-06, |
|
"loss": 0.9863, |
|
"step": 1093 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.4560808524355162e-06, |
|
"loss": 0.7479, |
|
"step": 1094 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.453569232870077e-06, |
|
"loss": 0.9936, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.4510564864802255e-06, |
|
"loss": 1.0067, |
|
"step": 1096 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.4485426162352525e-06, |
|
"loss": 0.7327, |
|
"step": 1097 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.446027625105776e-06, |
|
"loss": 1.2916, |
|
"step": 1098 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.44351151606374e-06, |
|
"loss": 1.382, |
|
"step": 1099 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.440994292082408e-06, |
|
"loss": 0.936, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.438475956136361e-06, |
|
"loss": 0.7885, |
|
"step": 1101 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.4359565112014957e-06, |
|
"loss": 0.7974, |
|
"step": 1102 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.433435960255017e-06, |
|
"loss": 1.2391, |
|
"step": 1103 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.430914306275438e-06, |
|
"loss": 0.7443, |
|
"step": 1104 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.4283915522425758e-06, |
|
"loss": 1.525, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.4258677011375473e-06, |
|
"loss": 0.9756, |
|
"step": 1106 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.4233427559427634e-06, |
|
"loss": 1.2988, |
|
"step": 1107 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.4208167196419312e-06, |
|
"loss": 1.3856, |
|
"step": 1108 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.4182895952200448e-06, |
|
"loss": 1.0378, |
|
"step": 1109 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.4157613856633847e-06, |
|
"loss": 1.2839, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.4132320939595156e-06, |
|
"loss": 0.8511, |
|
"step": 1111 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.4107017230972777e-06, |
|
"loss": 0.8881, |
|
"step": 1112 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.4081702760667894e-06, |
|
"loss": 0.9388, |
|
"step": 1113 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.405637755859439e-06, |
|
"loss": 1.3218, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.403104165467883e-06, |
|
"loss": 1.2651, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.4005695078860428e-06, |
|
"loss": 1.6755, |
|
"step": 1116 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.3980337861091028e-06, |
|
"loss": 1.5393, |
|
"step": 1117 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.395497003133501e-06, |
|
"loss": 1.2368, |
|
"step": 1118 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.392959161956933e-06, |
|
"loss": 1.2073, |
|
"step": 1119 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.3904202655783426e-06, |
|
"loss": 1.2589, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.3878803169979225e-06, |
|
"loss": 0.9366, |
|
"step": 1121 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.385339319217107e-06, |
|
"loss": 1.101, |
|
"step": 1122 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.3827972752385706e-06, |
|
"loss": 1.3452, |
|
"step": 1123 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.380254188066226e-06, |
|
"loss": 1.3305, |
|
"step": 1124 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.377710060705215e-06, |
|
"loss": 1.3394, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.375164896161912e-06, |
|
"loss": 0.9721, |
|
"step": 1126 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.3726186974439144e-06, |
|
"loss": 1.0626, |
|
"step": 1127 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.3700714675600443e-06, |
|
"loss": 1.1734, |
|
"step": 1128 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.36752320952034e-06, |
|
"loss": 1.2122, |
|
"step": 1129 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.364973926336057e-06, |
|
"loss": 1.5386, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.3624236210196594e-06, |
|
"loss": 1.0616, |
|
"step": 1131 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.359872296584821e-06, |
|
"loss": 1.0179, |
|
"step": 1132 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.35731995604642e-06, |
|
"loss": 0.9667, |
|
"step": 1133 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.354766602420534e-06, |
|
"loss": 0.9075, |
|
"step": 1134 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.3522122387244393e-06, |
|
"loss": 1.8509, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.3496568679766068e-06, |
|
"loss": 1.3092, |
|
"step": 1136 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.347100493196692e-06, |
|
"loss": 0.8243, |
|
"step": 1137 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.3445431174055436e-06, |
|
"loss": 1.0863, |
|
"step": 1138 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.3419847436251897e-06, |
|
"loss": 1.0101, |
|
"step": 1139 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.339425374878836e-06, |
|
"loss": 0.8136, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.3368650141908693e-06, |
|
"loss": 0.9144, |
|
"step": 1141 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.3343036645868432e-06, |
|
"loss": 1.6482, |
|
"step": 1142 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.331741329093483e-06, |
|
"loss": 1.1715, |
|
"step": 1143 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.329178010738678e-06, |
|
"loss": 0.9454, |
|
"step": 1144 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.3266137125514784e-06, |
|
"loss": 0.7856, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.324048437562094e-06, |
|
"loss": 1.0394, |
|
"step": 1146 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.321482188801888e-06, |
|
"loss": 0.8741, |
|
"step": 1147 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.3189149693033717e-06, |
|
"loss": 1.2134, |
|
"step": 1148 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.3163467821002082e-06, |
|
"loss": 1.1549, |
|
"step": 1149 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.3137776302272017e-06, |
|
"loss": 1.6155, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.3112075167202955e-06, |
|
"loss": 1.8034, |
|
"step": 1151 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.3086364446165714e-06, |
|
"loss": 0.9134, |
|
"step": 1152 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.306064416954241e-06, |
|
"loss": 0.8948, |
|
"step": 1153 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.3034914367726483e-06, |
|
"loss": 1.4639, |
|
"step": 1154 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.300917507112261e-06, |
|
"loss": 1.3632, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.2983426310146698e-06, |
|
"loss": 1.0861, |
|
"step": 1156 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.2957668115225828e-06, |
|
"loss": 1.1023, |
|
"step": 1157 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.2931900516798232e-06, |
|
"loss": 1.3424, |
|
"step": 1158 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.290612354531325e-06, |
|
"loss": 1.9575, |
|
"step": 1159 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.2880337231231312e-06, |
|
"loss": 1.9992, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"eval_loss": 1.2133097648620605, |
|
"eval_runtime": 53.8686, |
|
"eval_samples_per_second": 14.331, |
|
"eval_steps_per_second": 7.166, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.2854541605023882e-06, |
|
"loss": 1.2993, |
|
"step": 1161 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.282873669717342e-06, |
|
"loss": 1.3915, |
|
"step": 1162 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.280292253817336e-06, |
|
"loss": 1.1089, |
|
"step": 1163 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.2777099158528064e-06, |
|
"loss": 1.7346, |
|
"step": 1164 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.27512665887528e-06, |
|
"loss": 1.5882, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.272542485937369e-06, |
|
"loss": 1.3054, |
|
"step": 1166 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.269957400092767e-06, |
|
"loss": 1.148, |
|
"step": 1167 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.267371404396249e-06, |
|
"loss": 1.0103, |
|
"step": 1168 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.2647845019036622e-06, |
|
"loss": 1.2608, |
|
"step": 1169 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.2621966956719265e-06, |
|
"loss": 0.7083, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.2596079887590295e-06, |
|
"loss": 1.0038, |
|
"step": 1171 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.2570183842240255e-06, |
|
"loss": 0.9609, |
|
"step": 1172 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.254427885127025e-06, |
|
"loss": 0.8564, |
|
"step": 1173 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.2518364945292003e-06, |
|
"loss": 1.177, |
|
"step": 1174 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.249244215492774e-06, |
|
"loss": 1.707, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.2466510510810185e-06, |
|
"loss": 1.1875, |
|
"step": 1176 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.2440570043582554e-06, |
|
"loss": 1.245, |
|
"step": 1177 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.241462078389846e-06, |
|
"loss": 1.8574, |
|
"step": 1178 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.2388662762421914e-06, |
|
"loss": 1.4792, |
|
"step": 1179 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.236269600982729e-06, |
|
"loss": 1.2552, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.233672055679926e-06, |
|
"loss": 1.1056, |
|
"step": 1181 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.2310736434032797e-06, |
|
"loss": 1.524, |
|
"step": 1182 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.228474367223312e-06, |
|
"loss": 0.8447, |
|
"step": 1183 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.2258742302115632e-06, |
|
"loss": 1.8112, |
|
"step": 1184 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.2232732354405936e-06, |
|
"loss": 0.8166, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.2206713859839755e-06, |
|
"loss": 1.0981, |
|
"step": 1186 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.218068684916291e-06, |
|
"loss": 0.789, |
|
"step": 1187 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.2154651353131305e-06, |
|
"loss": 1.8448, |
|
"step": 1188 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.212860740251085e-06, |
|
"loss": 1.0422, |
|
"step": 1189 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.210255502807745e-06, |
|
"loss": 1.169, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.2076494260616982e-06, |
|
"loss": 1.2136, |
|
"step": 1191 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.20504251309252e-06, |
|
"loss": 1.3436, |
|
"step": 1192 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.202434766980779e-06, |
|
"loss": 0.7606, |
|
"step": 1193 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.199826190808026e-06, |
|
"loss": 0.987, |
|
"step": 1194 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.1972167876567907e-06, |
|
"loss": 1.3677, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.1946065606105832e-06, |
|
"loss": 1.3028, |
|
"step": 1196 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.1919955127538863e-06, |
|
"loss": 0.9541, |
|
"step": 1197 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.189383647172152e-06, |
|
"loss": 1.2335, |
|
"step": 1198 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.1867709669517983e-06, |
|
"loss": 1.3703, |
|
"step": 1199 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.184157475180208e-06, |
|
"loss": 1.3037, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.181543174945719e-06, |
|
"loss": 1.2641, |
|
"step": 1201 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.17892806933763e-06, |
|
"loss": 1.0842, |
|
"step": 1202 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.176312161446187e-06, |
|
"loss": 0.8864, |
|
"step": 1203 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.1736954543625843e-06, |
|
"loss": 1.7271, |
|
"step": 1204 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.1710779511789634e-06, |
|
"loss": 1.0044, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.1684596549884047e-06, |
|
"loss": 0.8295, |
|
"step": 1206 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.1658405688849246e-06, |
|
"loss": 0.737, |
|
"step": 1207 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.163220695963476e-06, |
|
"loss": 0.7618, |
|
"step": 1208 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.1606000393199393e-06, |
|
"loss": 1.5471, |
|
"step": 1209 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.1579786020511203e-06, |
|
"loss": 0.7496, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.15535638725475e-06, |
|
"loss": 1.707, |
|
"step": 1211 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.1527333980294748e-06, |
|
"loss": 1.1097, |
|
"step": 1212 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.15010963747486e-06, |
|
"loss": 1.6137, |
|
"step": 1213 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.1474851086913793e-06, |
|
"loss": 1.0059, |
|
"step": 1214 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.1448598147804162e-06, |
|
"loss": 1.6696, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.1422337588442577e-06, |
|
"loss": 0.8577, |
|
"step": 1216 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.1396069439860894e-06, |
|
"loss": 0.7763, |
|
"step": 1217 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.136979373309997e-06, |
|
"loss": 1.5286, |
|
"step": 1218 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.134351049920957e-06, |
|
"loss": 1.3097, |
|
"step": 1219 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.1317219769248373e-06, |
|
"loss": 1.4764, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.129092157428389e-06, |
|
"loss": 1.2568, |
|
"step": 1221 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.1264615945392473e-06, |
|
"loss": 1.2939, |
|
"step": 1222 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.1238302913659254e-06, |
|
"loss": 1.1002, |
|
"step": 1223 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.1211982510178114e-06, |
|
"loss": 1.2296, |
|
"step": 1224 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.1185654766051638e-06, |
|
"loss": 0.9342, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.115931971239109e-06, |
|
"loss": 1.8502, |
|
"step": 1226 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.1132977380316388e-06, |
|
"loss": 0.9097, |
|
"step": 1227 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.1106627800956014e-06, |
|
"loss": 1.0691, |
|
"step": 1228 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.1080271005447038e-06, |
|
"loss": 1.3972, |
|
"step": 1229 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.105390702493508e-06, |
|
"loss": 0.8666, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.1027535890574194e-06, |
|
"loss": 1.8229, |
|
"step": 1231 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.100115763352693e-06, |
|
"loss": 0.7901, |
|
"step": 1232 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.097477228496425e-06, |
|
"loss": 1.1264, |
|
"step": 1233 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.094837987606547e-06, |
|
"loss": 1.2298, |
|
"step": 1234 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.0921980438018293e-06, |
|
"loss": 1.4231, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.089557400201869e-06, |
|
"loss": 1.0321, |
|
"step": 1236 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.0869160599270915e-06, |
|
"loss": 0.8622, |
|
"step": 1237 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.0842740260987457e-06, |
|
"loss": 1.2512, |
|
"step": 1238 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.0816313018389e-06, |
|
"loss": 1.0031, |
|
"step": 1239 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.0789878902704385e-06, |
|
"loss": 0.9935, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.0763437945170576e-06, |
|
"loss": 1.0102, |
|
"step": 1241 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.0736990177032614e-06, |
|
"loss": 1.83, |
|
"step": 1242 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.0710535629543597e-06, |
|
"loss": 1.5906, |
|
"step": 1243 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.0684074333964642e-06, |
|
"loss": 1.0378, |
|
"step": 1244 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.065760632156482e-06, |
|
"loss": 0.8305, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.063113162362115e-06, |
|
"loss": 0.7797, |
|
"step": 1246 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.0604650271418556e-06, |
|
"loss": 1.111, |
|
"step": 1247 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.0578162296249812e-06, |
|
"loss": 0.7282, |
|
"step": 1248 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.055166772941554e-06, |
|
"loss": 1.1565, |
|
"step": 1249 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.052516660222412e-06, |
|
"loss": 1.269, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.049865894599172e-06, |
|
"loss": 1.3521, |
|
"step": 1251 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.0472144792042196e-06, |
|
"loss": 1.4313, |
|
"step": 1252 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.0445624171707085e-06, |
|
"loss": 0.766, |
|
"step": 1253 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.041909711632558e-06, |
|
"loss": 0.7068, |
|
"step": 1254 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.0392563657244472e-06, |
|
"loss": 0.7982, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.0366023825818107e-06, |
|
"loss": 0.925, |
|
"step": 1256 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.0339477653408382e-06, |
|
"loss": 0.8127, |
|
"step": 1257 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.031292517138468e-06, |
|
"loss": 1.2297, |
|
"step": 1258 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.0286366411123817e-06, |
|
"loss": 0.8651, |
|
"step": 1259 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.025980140401007e-06, |
|
"loss": 0.7399, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.023323018143506e-06, |
|
"loss": 0.8387, |
|
"step": 1261 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.0206652774797775e-06, |
|
"loss": 0.9132, |
|
"step": 1262 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.0180069215504505e-06, |
|
"loss": 1.276, |
|
"step": 1263 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.0153479534968792e-06, |
|
"loss": 1.1802, |
|
"step": 1264 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.0126883764611452e-06, |
|
"loss": 1.2264, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.0100281935860465e-06, |
|
"loss": 1.0068, |
|
"step": 1266 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.0073674080150977e-06, |
|
"loss": 1.4417, |
|
"step": 1267 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.0047060228925256e-06, |
|
"loss": 1.3083, |
|
"step": 1268 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.002044041363267e-06, |
|
"loss": 1.3119, |
|
"step": 1269 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.9993814665729605e-06, |
|
"loss": 1.5967, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.996718301667948e-06, |
|
"loss": 0.8924, |
|
"step": 1271 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.9940545497952687e-06, |
|
"loss": 1.01, |
|
"step": 1272 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.991390214102654e-06, |
|
"loss": 1.007, |
|
"step": 1273 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.9887252977385272e-06, |
|
"loss": 1.0263, |
|
"step": 1274 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.9860598038519954e-06, |
|
"loss": 0.8872, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.98339373559285e-06, |
|
"loss": 1.9292, |
|
"step": 1276 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.9807270961115603e-06, |
|
"loss": 1.2176, |
|
"step": 1277 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.978059888559271e-06, |
|
"loss": 1.0038, |
|
"step": 1278 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.9753921160877976e-06, |
|
"loss": 1.0877, |
|
"step": 1279 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.972723781849624e-06, |
|
"loss": 0.9189, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.970054888997897e-06, |
|
"loss": 0.7527, |
|
"step": 1281 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.9673854406864244e-06, |
|
"loss": 0.9119, |
|
"step": 1282 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.964715440069669e-06, |
|
"loss": 0.7733, |
|
"step": 1283 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.9620448903027484e-06, |
|
"loss": 0.7463, |
|
"step": 1284 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.9593737945414264e-06, |
|
"loss": 0.6778, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.9567021559421144e-06, |
|
"loss": 1.4462, |
|
"step": 1286 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.9540299776618637e-06, |
|
"loss": 1.5849, |
|
"step": 1287 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.9513572628583654e-06, |
|
"loss": 1.5328, |
|
"step": 1288 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.9486840146899414e-06, |
|
"loss": 0.8186, |
|
"step": 1289 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.946010236315546e-06, |
|
"loss": 1.144, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.943335930894762e-06, |
|
"loss": 0.7505, |
|
"step": 1291 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.940661101587789e-06, |
|
"loss": 1.2277, |
|
"step": 1292 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.937985751555452e-06, |
|
"loss": 1.2852, |
|
"step": 1293 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.935309883959188e-06, |
|
"loss": 1.6152, |
|
"step": 1294 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.9326335019610465e-06, |
|
"loss": 0.7664, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.9299566087236834e-06, |
|
"loss": 1.2087, |
|
"step": 1296 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.9272792074103616e-06, |
|
"loss": 1.4253, |
|
"step": 1297 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.9246013011849415e-06, |
|
"loss": 0.9122, |
|
"step": 1298 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.921922893211883e-06, |
|
"loss": 0.9546, |
|
"step": 1299 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.9192439866562363e-06, |
|
"loss": 1.6417, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.916564584683641e-06, |
|
"loss": 1.1093, |
|
"step": 1301 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.913884690460325e-06, |
|
"loss": 1.406, |
|
"step": 1302 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.9112043071530942e-06, |
|
"loss": 1.6273, |
|
"step": 1303 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.9085234379293353e-06, |
|
"loss": 1.4468, |
|
"step": 1304 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.905842085957007e-06, |
|
"loss": 0.8916, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"eval_loss": 1.2053191661834717, |
|
"eval_runtime": 53.6929, |
|
"eval_samples_per_second": 14.378, |
|
"eval_steps_per_second": 7.189, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.9031602544046395e-06, |
|
"loss": 1.1239, |
|
"step": 1306 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.9004779464413303e-06, |
|
"loss": 0.9879, |
|
"step": 1307 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.8977951652367382e-06, |
|
"loss": 0.7417, |
|
"step": 1308 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.895111913961083e-06, |
|
"loss": 1.4614, |
|
"step": 1309 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.8924281957851398e-06, |
|
"loss": 0.8142, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.8897440138802336e-06, |
|
"loss": 0.9211, |
|
"step": 1311 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.8870593714182394e-06, |
|
"loss": 1.1745, |
|
"step": 1312 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.8843742715715755e-06, |
|
"loss": 0.813, |
|
"step": 1313 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.8816887175132015e-06, |
|
"loss": 1.3786, |
|
"step": 1314 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.8790027124166114e-06, |
|
"loss": 1.2715, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.8763162594558364e-06, |
|
"loss": 1.0655, |
|
"step": 1316 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.8736293618054318e-06, |
|
"loss": 1.4229, |
|
"step": 1317 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.8709420226404828e-06, |
|
"loss": 0.8199, |
|
"step": 1318 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.8682542451365943e-06, |
|
"loss": 0.7968, |
|
"step": 1319 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.8655660324698887e-06, |
|
"loss": 2.126, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.8628773878170048e-06, |
|
"loss": 1.0297, |
|
"step": 1321 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.860188314355088e-06, |
|
"loss": 1.6878, |
|
"step": 1322 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.857498815261795e-06, |
|
"loss": 0.9881, |
|
"step": 1323 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.854808893715283e-06, |
|
"loss": 0.9988, |
|
"step": 1324 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.8521185528942086e-06, |
|
"loss": 1.1359, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.8494277959777233e-06, |
|
"loss": 0.9327, |
|
"step": 1326 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.8467366261454723e-06, |
|
"loss": 0.8055, |
|
"step": 1327 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.8440450465775853e-06, |
|
"loss": 1.164, |
|
"step": 1328 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.84135306045468e-06, |
|
"loss": 0.7184, |
|
"step": 1329 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.8386606709578535e-06, |
|
"loss": 1.9446, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.835967881268677e-06, |
|
"loss": 1.1398, |
|
"step": 1331 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.8332746945691976e-06, |
|
"loss": 1.3205, |
|
"step": 1332 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.8305811140419298e-06, |
|
"loss": 0.8334, |
|
"step": 1333 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.8278871428698544e-06, |
|
"loss": 0.9549, |
|
"step": 1334 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.8251927842364147e-06, |
|
"loss": 0.8567, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.822498041325509e-06, |
|
"loss": 0.8521, |
|
"step": 1336 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.8198029173214913e-06, |
|
"loss": 0.9907, |
|
"step": 1337 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.8171074154091677e-06, |
|
"loss": 1.3401, |
|
"step": 1338 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.814411538773788e-06, |
|
"loss": 1.7585, |
|
"step": 1339 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.811715290601046e-06, |
|
"loss": 0.8718, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.8090186740770752e-06, |
|
"loss": 1.3148, |
|
"step": 1341 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.806321692388443e-06, |
|
"loss": 1.5861, |
|
"step": 1342 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.8036243487221506e-06, |
|
"loss": 0.6952, |
|
"step": 1343 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.800926646265623e-06, |
|
"loss": 0.6928, |
|
"step": 1344 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.798228588206712e-06, |
|
"loss": 1.435, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.7955301777336926e-06, |
|
"loss": 1.2767, |
|
"step": 1346 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.792831418035248e-06, |
|
"loss": 0.8592, |
|
"step": 1347 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.7901323123004816e-06, |
|
"loss": 1.371, |
|
"step": 1348 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.7874328637189028e-06, |
|
"loss": 0.7175, |
|
"step": 1349 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.7847330754804257e-06, |
|
"loss": 1.0062, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.782032950775367e-06, |
|
"loss": 0.6584, |
|
"step": 1351 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.7793324927944398e-06, |
|
"loss": 1.0581, |
|
"step": 1352 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.776631704728752e-06, |
|
"loss": 1.4408, |
|
"step": 1353 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.7739305897698023e-06, |
|
"loss": 1.4966, |
|
"step": 1354 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.7712291511094725e-06, |
|
"loss": 0.9441, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.7685273919400306e-06, |
|
"loss": 0.8899, |
|
"step": 1356 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.765825315454121e-06, |
|
"loss": 1.4834, |
|
"step": 1357 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.7631229248447637e-06, |
|
"loss": 1.6439, |
|
"step": 1358 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.76042022330535e-06, |
|
"loss": 1.8315, |
|
"step": 1359 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.75771721402964e-06, |
|
"loss": 1.636, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.7550139002117543e-06, |
|
"loss": 1.1515, |
|
"step": 1361 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.752310285046176e-06, |
|
"loss": 1.0332, |
|
"step": 1362 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.749606371727744e-06, |
|
"loss": 2.0102, |
|
"step": 1363 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.7469021634516475e-06, |
|
"loss": 1.3126, |
|
"step": 1364 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.7441976634134275e-06, |
|
"loss": 0.9242, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.741492874808967e-06, |
|
"loss": 1.3781, |
|
"step": 1366 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.7387878008344902e-06, |
|
"loss": 0.7837, |
|
"step": 1367 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.7360824446865615e-06, |
|
"loss": 1.571, |
|
"step": 1368 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.7333768095620745e-06, |
|
"loss": 1.6734, |
|
"step": 1369 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.730670898658255e-06, |
|
"loss": 1.0149, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 2.727964715172656e-06, |
|
"loss": 1.0424, |
|
"step": 1371 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 2.725258262303148e-06, |
|
"loss": 1.8595, |
|
"step": 1372 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 2.7225515432479245e-06, |
|
"loss": 0.9829, |
|
"step": 1373 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 2.7198445612054914e-06, |
|
"loss": 1.548, |
|
"step": 1374 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 2.7171373193746653e-06, |
|
"loss": 0.9945, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 2.714429820954571e-06, |
|
"loss": 0.6336, |
|
"step": 1376 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 2.711722069144635e-06, |
|
"loss": 1.2451, |
|
"step": 1377 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 2.7090140671445836e-06, |
|
"loss": 1.7466, |
|
"step": 1378 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 2.7063058181544395e-06, |
|
"loss": 1.4181, |
|
"step": 1379 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 2.7035973253745167e-06, |
|
"loss": 1.3393, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 2.700888592005417e-06, |
|
"loss": 1.2597, |
|
"step": 1381 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 2.698179621248028e-06, |
|
"loss": 1.5118, |
|
"step": 1382 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 2.695470416303516e-06, |
|
"loss": 0.9365, |
|
"step": 1383 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 2.6927609803733245e-06, |
|
"loss": 0.9296, |
|
"step": 1384 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.6900513166591715e-06, |
|
"loss": 0.9604, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.6873414283630405e-06, |
|
"loss": 0.9642, |
|
"step": 1386 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.684631318687185e-06, |
|
"loss": 0.8974, |
|
"step": 1387 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.6819209908341176e-06, |
|
"loss": 1.0633, |
|
"step": 1388 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.679210448006607e-06, |
|
"loss": 1.8002, |
|
"step": 1389 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.6764996934076805e-06, |
|
"loss": 1.3501, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.6737887302406117e-06, |
|
"loss": 0.75, |
|
"step": 1391 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.671077561708922e-06, |
|
"loss": 0.7531, |
|
"step": 1392 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.668366191016376e-06, |
|
"loss": 0.7146, |
|
"step": 1393 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.6656546213669764e-06, |
|
"loss": 0.8559, |
|
"step": 1394 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.662942855964961e-06, |
|
"loss": 1.399, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.6602308980148007e-06, |
|
"loss": 1.1126, |
|
"step": 1396 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.65751875072119e-06, |
|
"loss": 1.6005, |
|
"step": 1397 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.6548064172890524e-06, |
|
"loss": 1.2356, |
|
"step": 1398 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.652093900923527e-06, |
|
"loss": 1.4258, |
|
"step": 1399 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.64938120482997e-06, |
|
"loss": 1.2329, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.6466683322139515e-06, |
|
"loss": 1.639, |
|
"step": 1401 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.6439552862812495e-06, |
|
"loss": 1.2395, |
|
"step": 1402 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.641242070237846e-06, |
|
"loss": 1.0329, |
|
"step": 1403 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.638528687289925e-06, |
|
"loss": 1.3817, |
|
"step": 1404 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.6358151406438675e-06, |
|
"loss": 1.3485, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.6331014335062477e-06, |
|
"loss": 0.7669, |
|
"step": 1406 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.63038756908383e-06, |
|
"loss": 0.7133, |
|
"step": 1407 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.6276735505835636e-06, |
|
"loss": 1.159, |
|
"step": 1408 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.62495938121258e-06, |
|
"loss": 0.7357, |
|
"step": 1409 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.6222450641781904e-06, |
|
"loss": 0.7507, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.6195306026878788e-06, |
|
"loss": 1.1725, |
|
"step": 1411 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.6168159999492997e-06, |
|
"loss": 1.224, |
|
"step": 1412 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.614101259170278e-06, |
|
"loss": 0.8986, |
|
"step": 1413 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.6113863835587965e-06, |
|
"loss": 1.2494, |
|
"step": 1414 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.608671376323002e-06, |
|
"loss": 1.0252, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.605956240671193e-06, |
|
"loss": 0.9566, |
|
"step": 1416 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.6032409798118226e-06, |
|
"loss": 0.9629, |
|
"step": 1417 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.60052559695349e-06, |
|
"loss": 0.6855, |
|
"step": 1418 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.5978100953049406e-06, |
|
"loss": 1.2674, |
|
"step": 1419 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.5950944780750577e-06, |
|
"loss": 0.8772, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.592378748472863e-06, |
|
"loss": 0.8036, |
|
"step": 1421 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.58966290970751e-06, |
|
"loss": 1.1676, |
|
"step": 1422 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.586946964988281e-06, |
|
"loss": 1.8266, |
|
"step": 1423 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.584230917524586e-06, |
|
"loss": 1.6392, |
|
"step": 1424 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.581514770525953e-06, |
|
"loss": 0.8578, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.57879852720203e-06, |
|
"loss": 0.7239, |
|
"step": 1426 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.5760821907625767e-06, |
|
"loss": 1.2385, |
|
"step": 1427 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.573365764417465e-06, |
|
"loss": 1.4696, |
|
"step": 1428 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.570649251376673e-06, |
|
"loss": 1.7116, |
|
"step": 1429 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.567932654850278e-06, |
|
"loss": 1.3335, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.5652159780484598e-06, |
|
"loss": 0.9493, |
|
"step": 1431 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.5624992241814922e-06, |
|
"loss": 1.4103, |
|
"step": 1432 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.559782396459738e-06, |
|
"loss": 1.6927, |
|
"step": 1433 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.5570654980936493e-06, |
|
"loss": 1.5447, |
|
"step": 1434 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.5543485322937617e-06, |
|
"loss": 0.7246, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.5516315022706893e-06, |
|
"loss": 0.86, |
|
"step": 1436 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.5489144112351234e-06, |
|
"loss": 1.8581, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.546197262397825e-06, |
|
"loss": 0.8806, |
|
"step": 1438 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.5434800589696263e-06, |
|
"loss": 0.8996, |
|
"step": 1439 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.5407628041614234e-06, |
|
"loss": 0.9795, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.538045501184172e-06, |
|
"loss": 0.8858, |
|
"step": 1441 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.535328153248884e-06, |
|
"loss": 1.1132, |
|
"step": 1442 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.532610763566628e-06, |
|
"loss": 1.5197, |
|
"step": 1443 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.5298933353485174e-06, |
|
"loss": 1.2823, |
|
"step": 1444 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.5271758718057138e-06, |
|
"loss": 2.0808, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.524458376149421e-06, |
|
"loss": 1.4084, |
|
"step": 1446 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.521740851590879e-06, |
|
"loss": 1.7921, |
|
"step": 1447 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.519023301341364e-06, |
|
"loss": 1.6181, |
|
"step": 1448 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.516305728612179e-06, |
|
"loss": 1.2924, |
|
"step": 1449 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.5135881366146585e-06, |
|
"loss": 0.7885, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 1.1991530656814575, |
|
"eval_runtime": 53.8047, |
|
"eval_samples_per_second": 14.348, |
|
"eval_steps_per_second": 7.174, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.5108705285601554e-06, |
|
"loss": 1.5017, |
|
"step": 1451 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.508152907660044e-06, |
|
"loss": 0.7704, |
|
"step": 1452 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.505435277125713e-06, |
|
"loss": 1.4009, |
|
"step": 1453 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.5027176401685628e-06, |
|
"loss": 1.5988, |
|
"step": 1454 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.5e-06, |
|
"loss": 1.0798, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.497282359831438e-06, |
|
"loss": 1.5089, |
|
"step": 1456 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.4945647228742883e-06, |
|
"loss": 1.5705, |
|
"step": 1457 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.491847092339957e-06, |
|
"loss": 1.012, |
|
"step": 1458 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.4891294714398455e-06, |
|
"loss": 0.7862, |
|
"step": 1459 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.4864118633853424e-06, |
|
"loss": 1.3208, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.483694271387821e-06, |
|
"loss": 1.3518, |
|
"step": 1461 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.480976698658637e-06, |
|
"loss": 1.2256, |
|
"step": 1462 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.478259148409121e-06, |
|
"loss": 1.5057, |
|
"step": 1463 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.4755416238505792e-06, |
|
"loss": 0.6883, |
|
"step": 1464 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.472824128194286e-06, |
|
"loss": 1.4382, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.4701066646514843e-06, |
|
"loss": 1.4933, |
|
"step": 1466 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.4673892364333736e-06, |
|
"loss": 1.3664, |
|
"step": 1467 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.4646718467511167e-06, |
|
"loss": 1.7627, |
|
"step": 1468 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.4619544988158293e-06, |
|
"loss": 1.3693, |
|
"step": 1469 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.459237195838577e-06, |
|
"loss": 0.9821, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.456519941030374e-06, |
|
"loss": 1.3148, |
|
"step": 1471 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.453802737602176e-06, |
|
"loss": 1.0727, |
|
"step": 1472 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.4510855887648775e-06, |
|
"loss": 0.9669, |
|
"step": 1473 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.4483684977293116e-06, |
|
"loss": 1.3828, |
|
"step": 1474 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.445651467706239e-06, |
|
"loss": 1.3555, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.4429345019063516e-06, |
|
"loss": 0.8507, |
|
"step": 1476 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.440217603540263e-06, |
|
"loss": 0.6893, |
|
"step": 1477 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.4375007758185086e-06, |
|
"loss": 1.532, |
|
"step": 1478 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.434784021951541e-06, |
|
"loss": 0.726, |
|
"step": 1479 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.432067345149723e-06, |
|
"loss": 0.5863, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.429350748623328e-06, |
|
"loss": 1.0244, |
|
"step": 1481 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.4266342355825354e-06, |
|
"loss": 0.719, |
|
"step": 1482 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.4239178092374237e-06, |
|
"loss": 0.7407, |
|
"step": 1483 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.421201472797971e-06, |
|
"loss": 1.3884, |
|
"step": 1484 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.4184852294740474e-06, |
|
"loss": 0.9381, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.415769082475414e-06, |
|
"loss": 0.6054, |
|
"step": 1486 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.4130530350117194e-06, |
|
"loss": 1.0417, |
|
"step": 1487 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.4103370902924916e-06, |
|
"loss": 0.645, |
|
"step": 1488 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.4076212515271384e-06, |
|
"loss": 0.4432, |
|
"step": 1489 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.4049055219249435e-06, |
|
"loss": 0.9322, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.4021899046950607e-06, |
|
"loss": 0.5862, |
|
"step": 1491 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.3994744030465106e-06, |
|
"loss": 1.3218, |
|
"step": 1492 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.3967590201881786e-06, |
|
"loss": 0.5756, |
|
"step": 1493 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.3940437593288075e-06, |
|
"loss": 1.3496, |
|
"step": 1494 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.3913286236769988e-06, |
|
"loss": 0.8932, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.388613616441204e-06, |
|
"loss": 0.8556, |
|
"step": 1496 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.3858987408297225e-06, |
|
"loss": 1.4038, |
|
"step": 1497 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.3831840000507007e-06, |
|
"loss": 1.1387, |
|
"step": 1498 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.3804693973121225e-06, |
|
"loss": 0.6889, |
|
"step": 1499 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.3777549358218105e-06, |
|
"loss": 0.6566, |
|
"step": 1500 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 2900, |
|
"num_train_epochs": 2, |
|
"save_steps": 500, |
|
"total_flos": 5.24254747557888e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|