|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.9989281886387996, |
|
"global_step": 233, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.4285714285714286e-06, |
|
"loss": 0.9647, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.8571428571428573e-06, |
|
"loss": 0.9612, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.2857142857142855e-06, |
|
"loss": 0.9788, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 5.7142857142857145e-06, |
|
"loss": 1.0404, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 7.1428571428571436e-06, |
|
"loss": 1.0253, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.571428571428571e-06, |
|
"loss": 0.9271, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1e-05, |
|
"loss": 0.9537, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.999516923531906e-06, |
|
"loss": 1.0348, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.998067787472772e-06, |
|
"loss": 1.0089, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.995652871840006e-06, |
|
"loss": 0.9514, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.992272643269181e-06, |
|
"loss": 0.9565, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.987927754923844e-06, |
|
"loss": 0.9587, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.982619046369321e-06, |
|
"loss": 0.9813, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.976347543410487e-06, |
|
"loss": 0.9689, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.96911445789354e-06, |
|
"loss": 1.0083, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.960921187471841e-06, |
|
"loss": 1.0022, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.951769315335843e-06, |
|
"loss": 1.0201, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.94166060990718e-06, |
|
"loss": 0.9267, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.930597024496933e-06, |
|
"loss": 0.9543, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.918580696928206e-06, |
|
"loss": 0.9814, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.905613949123036e-06, |
|
"loss": 0.986, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.891699286653714e-06, |
|
"loss": 0.9464, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.87683939825864e-06, |
|
"loss": 0.9442, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.861037155322777e-06, |
|
"loss": 1.0023, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.844295611322804e-06, |
|
"loss": 0.9524, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.826618001237101e-06, |
|
"loss": 1.0088, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.808007740920647e-06, |
|
"loss": 0.9542, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.788468426444968e-06, |
|
"loss": 0.9861, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.768003833403278e-06, |
|
"loss": 0.9533, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.746617916180906e-06, |
|
"loss": 0.9402, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.724314807191197e-06, |
|
"loss": 1.0455, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.701098816076995e-06, |
|
"loss": 0.9379, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.6769744288779e-06, |
|
"loss": 0.988, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.651946307163417e-06, |
|
"loss": 0.9998, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.626019287132202e-06, |
|
"loss": 0.9943, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.599198378677559e-06, |
|
"loss": 0.999, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.571488764419381e-06, |
|
"loss": 0.9608, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.542895798702702e-06, |
|
"loss": 0.9852, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.51342500656308e-06, |
|
"loss": 0.9252, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.483082082658984e-06, |
|
"loss": 0.9802, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.451872890171419e-06, |
|
"loss": 0.955, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.41980345967098e-06, |
|
"loss": 0.9292, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.386879987952549e-06, |
|
"loss": 0.9818, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.353108836837907e-06, |
|
"loss": 0.952, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.318496531946411e-06, |
|
"loss": 0.9525, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.283049761434059e-06, |
|
"loss": 0.9136, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.246775374701139e-06, |
|
"loss": 0.9882, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.209680381068698e-06, |
|
"loss": 0.9812, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.171771948424138e-06, |
|
"loss": 0.9652, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.13305740183616e-06, |
|
"loss": 0.9753, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.093544222139338e-06, |
|
"loss": 0.9142, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.053240044488587e-06, |
|
"loss": 1.0215, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.012152656883824e-06, |
|
"loss": 0.9178, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.970289998665083e-06, |
|
"loss": 0.9146, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.927660158978392e-06, |
|
"loss": 0.9399, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.884271375212714e-06, |
|
"loss": 0.9224, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.84013203140821e-06, |
|
"loss": 0.9368, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.795250656636207e-06, |
|
"loss": 0.9538, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.749635923351108e-06, |
|
"loss": 0.944, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.70329664571461e-06, |
|
"loss": 0.9462, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.656241777892544e-06, |
|
"loss": 0.9415, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.608480412324652e-06, |
|
"loss": 0.9376, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.56002177796765e-06, |
|
"loss": 0.9305, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.510875238511911e-06, |
|
"loss": 0.9203, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.461050290572114e-06, |
|
"loss": 0.9625, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.410556561852212e-06, |
|
"loss": 0.92, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.359403809285054e-06, |
|
"loss": 0.912, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.30760191714706e-06, |
|
"loss": 0.9316, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.255160895148263e-06, |
|
"loss": 0.9243, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.202090876498144e-06, |
|
"loss": 0.9209, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.14840211594757e-06, |
|
"loss": 0.937, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 8.09410498780727e-06, |
|
"loss": 0.93, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 8.039209983943201e-06, |
|
"loss": 0.9364, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.983727711749194e-06, |
|
"loss": 0.9497, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.927668892097288e-06, |
|
"loss": 0.9155, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.871044357266124e-06, |
|
"loss": 0.9404, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.81386504884782e-06, |
|
"loss": 0.9512, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.75614201563372e-06, |
|
"loss": 0.8969, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.697886411479422e-06, |
|
"loss": 0.9449, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.639109493149537e-06, |
|
"loss": 0.9162, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.579822618142505e-06, |
|
"loss": 0.9265, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.520037242496e-06, |
|
"loss": 0.9004, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.459764918573264e-06, |
|
"loss": 0.9159, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.399017292830848e-06, |
|
"loss": 0.9897, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.3378061035681415e-06, |
|
"loss": 0.8831, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.276143178659195e-06, |
|
"loss": 0.8804, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.2140404332671986e-06, |
|
"loss": 0.9042, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 7.1515098675421125e-06, |
|
"loss": 0.9084, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 7.088563564301874e-06, |
|
"loss": 0.9262, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 7.0252136866976205e-06, |
|
"loss": 0.9149, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.961472475863406e-06, |
|
"loss": 0.8765, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.897352248550828e-06, |
|
"loss": 0.9194, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.832865394749065e-06, |
|
"loss": 0.9056, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.768024375290747e-06, |
|
"loss": 0.886, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.702841719444141e-06, |
|
"loss": 0.8881, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.637330022492112e-06, |
|
"loss": 0.9144, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.571501943298335e-06, |
|
"loss": 0.9248, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.50537020186121e-06, |
|
"loss": 0.8927, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.4389475768559675e-06, |
|
"loss": 0.8876, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.372246903165445e-06, |
|
"loss": 0.9196, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.305281069399989e-06, |
|
"loss": 0.9386, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 6.238063015406982e-06, |
|
"loss": 0.911, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 6.17060572977047e-06, |
|
"loss": 0.8829, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 6.1029222473013705e-06, |
|
"loss": 0.914, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 6.035025646518747e-06, |
|
"loss": 0.9027, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.966929047122641e-06, |
|
"loss": 0.9088, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.898645607458941e-06, |
|
"loss": 0.8972, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.830188521976794e-06, |
|
"loss": 1.0138, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.761571018679025e-06, |
|
"loss": 0.8856, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.6928063565660955e-06, |
|
"loss": 0.9073, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.623907823074044e-06, |
|
"loss": 0.9249, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.5548887315069575e-06, |
|
"loss": 0.8923, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.48576241846443e-06, |
|
"loss": 0.9531, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.416542241264524e-06, |
|
"loss": 0.8698, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.347241575362729e-06, |
|
"loss": 0.97, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.277873811767415e-06, |
|
"loss": 0.9077, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.208452354452275e-06, |
|
"loss": 0.8709, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 5.1389906177662705e-06, |
|
"loss": 0.8485, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 5.069502023841576e-06, |
|
"loss": 0.9081, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 5e-06, |
|
"loss": 0.8596, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.9304979761584256e-06, |
|
"loss": 0.9091, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.86100938223373e-06, |
|
"loss": 0.9067, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.791547645547727e-06, |
|
"loss": 0.9067, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.722126188232586e-06, |
|
"loss": 0.9169, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.652758424637271e-06, |
|
"loss": 0.8661, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.583457758735477e-06, |
|
"loss": 0.8648, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.514237581535571e-06, |
|
"loss": 0.9384, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.4451112684930424e-06, |
|
"loss": 0.9088, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.3760921769259585e-06, |
|
"loss": 0.8867, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.307193643433907e-06, |
|
"loss": 0.9048, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.2384289813209754e-06, |
|
"loss": 0.879, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.1698114780232085e-06, |
|
"loss": 0.9342, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.101354392541061e-06, |
|
"loss": 0.8803, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.033070952877362e-06, |
|
"loss": 0.9021, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.964974353481254e-06, |
|
"loss": 0.8531, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.89707775269863e-06, |
|
"loss": 0.9417, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.829394270229531e-06, |
|
"loss": 0.895, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.7619369845930195e-06, |
|
"loss": 0.9225, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.694718930600012e-06, |
|
"loss": 0.9061, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.6277530968345552e-06, |
|
"loss": 0.9171, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.5610524231440324e-06, |
|
"loss": 0.8518, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.4946297981387913e-06, |
|
"loss": 0.9252, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.428498056701665e-06, |
|
"loss": 0.8954, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.3626699775078884e-06, |
|
"loss": 0.9131, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.2971582805558622e-06, |
|
"loss": 0.8737, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.2319756247092552e-06, |
|
"loss": 0.901, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.167134605250938e-06, |
|
"loss": 0.8971, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.102647751449174e-06, |
|
"loss": 0.9102, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.0385275241365965e-06, |
|
"loss": 0.8664, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.9747863133023803e-06, |
|
"loss": 0.8842, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.9114364356981274e-06, |
|
"loss": 0.8718, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.8484901324578883e-06, |
|
"loss": 0.8987, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.7859595667328027e-06, |
|
"loss": 0.8636, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.723856821340806e-06, |
|
"loss": 0.8753, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.6621938964318593e-06, |
|
"loss": 0.9054, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.600982707169154e-06, |
|
"loss": 0.8859, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.5402350814267364e-06, |
|
"loss": 0.8888, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.4799627575040014e-06, |
|
"loss": 0.8704, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.4201773818574956e-06, |
|
"loss": 0.8943, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.360890506850464e-06, |
|
"loss": 0.9088, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.302113588520578e-06, |
|
"loss": 0.8857, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.243857984366284e-06, |
|
"loss": 0.8732, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.1861349511521817e-06, |
|
"loss": 0.8909, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.128955642733877e-06, |
|
"loss": 0.8663, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.072331107902713e-06, |
|
"loss": 0.8845, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.0162722882508072e-06, |
|
"loss": 0.8974, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.960790016056801e-06, |
|
"loss": 0.9084, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.90589501219273e-06, |
|
"loss": 0.8749, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.8515978840524302e-06, |
|
"loss": 0.8915, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.7979091235018564e-06, |
|
"loss": 0.8876, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.7448391048517378e-06, |
|
"loss": 0.8762, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.6923980828529424e-06, |
|
"loss": 0.8846, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.640596190714947e-06, |
|
"loss": 0.9112, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.589443438147789e-06, |
|
"loss": 0.9316, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.5389497094278861e-06, |
|
"loss": 0.9098, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.4891247614880904e-06, |
|
"loss": 0.9075, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.4399782220323515e-06, |
|
"loss": 0.8749, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.3915195876753495e-06, |
|
"loss": 0.8767, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.3437582221074574e-06, |
|
"loss": 0.8622, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.2967033542853918e-06, |
|
"loss": 0.8908, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.250364076648894e-06, |
|
"loss": 0.8607, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.2047493433637935e-06, |
|
"loss": 0.9185, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.1598679685917901e-06, |
|
"loss": 0.8728, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.1157286247872873e-06, |
|
"loss": 0.8557, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.0723398410216085e-06, |
|
"loss": 0.8861, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.0297100013349181e-06, |
|
"loss": 0.8568, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.878473431161767e-07, |
|
"loss": 0.8527, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.467599555114137e-07, |
|
"loss": 0.8622, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.064557778606631e-07, |
|
"loss": 0.951, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.669425981638413e-07, |
|
"loss": 0.9074, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 8.282280515758639e-07, |
|
"loss": 0.8839, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.903196189313039e-07, |
|
"loss": 0.9193, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.532246252988617e-07, |
|
"loss": 0.8785, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.16950238565941e-07, |
|
"loss": 0.8461, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.815034680535915e-07, |
|
"loss": 0.8308, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.46891163162095e-07, |
|
"loss": 0.8792, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.131200120474512e-07, |
|
"loss": 0.9292, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.801965403290221e-07, |
|
"loss": 0.8785, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.481271098285818e-07, |
|
"loss": 0.8839, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 5.169179173410178e-07, |
|
"loss": 0.8629, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.865749934369224e-07, |
|
"loss": 0.8593, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.571042012972993e-07, |
|
"loss": 0.8764, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.2851123558061927e-07, |
|
"loss": 0.8521, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.008016213224408e-07, |
|
"loss": 0.8982, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.739807128677986e-07, |
|
"loss": 0.8703, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.480536928365824e-07, |
|
"loss": 0.8618, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.230255711220992e-07, |
|
"loss": 0.8876, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.9890118392300493e-07, |
|
"loss": 0.8856, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.756851928088056e-07, |
|
"loss": 0.8938, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.533820838190959e-07, |
|
"loss": 0.9029, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.3199616659672352e-07, |
|
"loss": 0.8963, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.1153157355503274e-07, |
|
"loss": 0.8669, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.9199225907935492e-07, |
|
"loss": 0.8846, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.7338199876289984e-07, |
|
"loss": 0.8796, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.5570438867719695e-07, |
|
"loss": 0.9144, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.3896284467722398e-07, |
|
"loss": 0.912, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.2316060174136e-07, |
|
"loss": 0.9186, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.0830071334628655e-07, |
|
"loss": 0.8841, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.43860508769645e-08, |
|
"loss": 0.8699, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 8.141930307179468e-08, |
|
"loss": 0.8885, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 6.940297550306895e-08, |
|
"loss": 0.8901, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 5.833939009282086e-08, |
|
"loss": 0.8852, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 4.823068466415615e-08, |
|
"loss": 0.9279, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 3.907881252816048e-08, |
|
"loss": 0.8549, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 3.088554210646133e-08, |
|
"loss": 0.9369, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.3652456589512983e-08, |
|
"loss": 0.9129, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.7380953630678488e-08, |
|
"loss": 0.8472, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.2072245076156786e-08, |
|
"loss": 0.8672, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 7.727356730820035e-09, |
|
"loss": 0.8965, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 4.347128159993829e-09, |
|
"loss": 0.8852, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.9322125272297488e-09, |
|
"loss": 0.9028, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 4.830764680946453e-10, |
|
"loss": 0.8491, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.0, |
|
"loss": 0.8746, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 233, |
|
"total_flos": 4.1801555584619643e+18, |
|
"train_loss": 0.9163404472907726, |
|
"train_runtime": 12375.9474, |
|
"train_samples_per_second": 4.824, |
|
"train_steps_per_second": 0.019 |
|
} |
|
], |
|
"max_steps": 233, |
|
"num_train_epochs": 1, |
|
"total_flos": 4.1801555584619643e+18, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|