|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 4.999168882978723, |
|
"global_step": 12030, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 6.492946524638607e-06, |
|
"loss": 1.601, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 7.470232356717817e-06, |
|
"loss": 1.4225, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.041907920970229e-06, |
|
"loss": 1.4184, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.447518188797028e-06, |
|
"loss": 1.3869, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 8.762133954878699e-06, |
|
"loss": 1.3759, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"eval_loss": 1.3726129531860352, |
|
"eval_runtime": 272.1193, |
|
"eval_samples_per_second": 18.617, |
|
"eval_steps_per_second": 2.33, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.019193753049438e-06, |
|
"loss": 1.3454, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.236534715583339e-06, |
|
"loss": 1.358, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.42480402087624e-06, |
|
"loss": 1.3487, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.59086931730185e-06, |
|
"loss": 1.3397, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.73941978695791e-06, |
|
"loss": 1.3515, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"eval_loss": 1.346224069595337, |
|
"eval_runtime": 272.1001, |
|
"eval_samples_per_second": 18.618, |
|
"eval_steps_per_second": 2.33, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.873800032579688e-06, |
|
"loss": 1.3469, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.99647958512865e-06, |
|
"loss": 1.3331, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.911332779163204e-06, |
|
"loss": 1.3336, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.81897109079154e-06, |
|
"loss": 1.3345, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.726609402419877e-06, |
|
"loss": 1.3207, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"eval_loss": 1.328342080116272, |
|
"eval_runtime": 272.126, |
|
"eval_samples_per_second": 18.616, |
|
"eval_steps_per_second": 2.33, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.634247714048214e-06, |
|
"loss": 1.3345, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 9.54188602567655e-06, |
|
"loss": 1.3151, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 9.449524337304887e-06, |
|
"loss": 1.3387, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.357162648933224e-06, |
|
"loss": 1.3183, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 9.264800960561559e-06, |
|
"loss": 1.3172, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"eval_loss": 1.3202550411224365, |
|
"eval_runtime": 272.1379, |
|
"eval_samples_per_second": 18.616, |
|
"eval_steps_per_second": 2.33, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.172439272189896e-06, |
|
"loss": 1.2997, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 9.080077583818233e-06, |
|
"loss": 1.3338, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 8.987715895446569e-06, |
|
"loss": 1.3199, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 8.895354207074907e-06, |
|
"loss": 1.3067, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 8.802992518703243e-06, |
|
"loss": 1.2836, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"eval_loss": 1.310795545578003, |
|
"eval_runtime": 272.0952, |
|
"eval_samples_per_second": 18.618, |
|
"eval_steps_per_second": 2.33, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 8.710630830331578e-06, |
|
"loss": 1.2834, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 8.618269141959917e-06, |
|
"loss": 1.2902, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 8.525907453588252e-06, |
|
"loss": 1.2755, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 8.43354576521659e-06, |
|
"loss": 1.2887, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 8.341184076844926e-06, |
|
"loss": 1.2734, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"eval_loss": 1.3082149028778076, |
|
"eval_runtime": 272.1154, |
|
"eval_samples_per_second": 18.617, |
|
"eval_steps_per_second": 2.33, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 8.248822388473262e-06, |
|
"loss": 1.282, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 8.156460700101599e-06, |
|
"loss": 1.2855, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 8.064099011729936e-06, |
|
"loss": 1.2793, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 7.971737323358271e-06, |
|
"loss": 1.2811, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 7.879375634986609e-06, |
|
"loss": 1.261, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"eval_loss": 1.3000493049621582, |
|
"eval_runtime": 272.083, |
|
"eval_samples_per_second": 18.619, |
|
"eval_steps_per_second": 2.33, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 7.787013946614946e-06, |
|
"loss": 1.2742, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 7.694652258243281e-06, |
|
"loss": 1.2837, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 7.602290569871617e-06, |
|
"loss": 1.2663, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 7.509928881499955e-06, |
|
"loss": 1.2719, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 7.4175671931282914e-06, |
|
"loss": 1.2703, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"eval_loss": 1.2987574338912964, |
|
"eval_runtime": 272.1184, |
|
"eval_samples_per_second": 18.617, |
|
"eval_steps_per_second": 2.33, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 7.325205504756628e-06, |
|
"loss": 1.2695, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 7.232843816384965e-06, |
|
"loss": 1.262, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 7.140482128013301e-06, |
|
"loss": 1.2584, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 7.048120439641637e-06, |
|
"loss": 1.2707, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 6.955758751269974e-06, |
|
"loss": 1.2579, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"eval_loss": 1.2950292825698853, |
|
"eval_runtime": 272.1448, |
|
"eval_samples_per_second": 18.615, |
|
"eval_steps_per_second": 2.33, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 6.8633970628983106e-06, |
|
"loss": 1.2784, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 6.771035374526647e-06, |
|
"loss": 1.2573, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 6.678673686154984e-06, |
|
"loss": 1.2722, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 6.58631199778332e-06, |
|
"loss": 1.2396, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 6.493950309411656e-06, |
|
"loss": 1.2447, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"eval_loss": 1.297192096710205, |
|
"eval_runtime": 272.1306, |
|
"eval_samples_per_second": 18.616, |
|
"eval_steps_per_second": 2.33, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 6.4015886210399935e-06, |
|
"loss": 1.2418, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 6.30922693266833e-06, |
|
"loss": 1.2523, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 6.216865244296666e-06, |
|
"loss": 1.2309, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 6.124503555925003e-06, |
|
"loss": 1.2442, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 6.032141867553339e-06, |
|
"loss": 1.2261, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"eval_loss": 1.2935367822647095, |
|
"eval_runtime": 272.1527, |
|
"eval_samples_per_second": 18.615, |
|
"eval_steps_per_second": 2.33, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 5.9397801791816755e-06, |
|
"loss": 1.2561, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 5.847418490810013e-06, |
|
"loss": 1.234, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 5.755056802438349e-06, |
|
"loss": 1.2314, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 5.662695114066685e-06, |
|
"loss": 1.24, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 5.570333425695022e-06, |
|
"loss": 1.2388, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"eval_loss": 1.2919647693634033, |
|
"eval_runtime": 272.1268, |
|
"eval_samples_per_second": 18.616, |
|
"eval_steps_per_second": 2.33, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 5.477971737323358e-06, |
|
"loss": 1.2416, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 5.385610048951695e-06, |
|
"loss": 1.2527, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 5.2932483605800326e-06, |
|
"loss": 1.2331, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 5.200886672208369e-06, |
|
"loss": 1.2483, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 5.108524983836704e-06, |
|
"loss": 1.2467, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"eval_loss": 1.2915822267532349, |
|
"eval_runtime": 272.1288, |
|
"eval_samples_per_second": 18.616, |
|
"eval_steps_per_second": 2.33, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 5.016163295465042e-06, |
|
"loss": 1.2528, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 4.923801607093378e-06, |
|
"loss": 1.2373, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 4.831439918721715e-06, |
|
"loss": 1.2357, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 4.739078230350051e-06, |
|
"loss": 1.2423, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 4.646716541978388e-06, |
|
"loss": 1.2358, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"eval_loss": 1.2892886400222778, |
|
"eval_runtime": 272.1494, |
|
"eval_samples_per_second": 18.615, |
|
"eval_steps_per_second": 2.33, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 4.554354853606724e-06, |
|
"loss": 1.2398, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 4.4619931652350604e-06, |
|
"loss": 1.2271, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 4.3696314768633975e-06, |
|
"loss": 1.2167, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 4.277269788491734e-06, |
|
"loss": 1.2298, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 4.18490810012007e-06, |
|
"loss": 1.2121, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"eval_loss": 1.2931495904922485, |
|
"eval_runtime": 272.1691, |
|
"eval_samples_per_second": 18.613, |
|
"eval_steps_per_second": 2.329, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 4.092546411748407e-06, |
|
"loss": 1.2265, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 4.000184723376744e-06, |
|
"loss": 1.216, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 3.9078230350050796e-06, |
|
"loss": 1.2243, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 3.815461346633417e-06, |
|
"loss": 1.2047, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 3.7230996582617533e-06, |
|
"loss": 1.2274, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"eval_loss": 1.2907688617706299, |
|
"eval_runtime": 272.1306, |
|
"eval_samples_per_second": 18.616, |
|
"eval_steps_per_second": 2.33, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 3.6307379698900895e-06, |
|
"loss": 1.2214, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 3.538376281518426e-06, |
|
"loss": 1.2215, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"learning_rate": 3.4460145931467633e-06, |
|
"loss": 1.2221, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"learning_rate": 3.353652904775099e-06, |
|
"loss": 1.2179, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 3.261291216403436e-06, |
|
"loss": 1.2229, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"eval_loss": 1.290418267250061, |
|
"eval_runtime": 272.1197, |
|
"eval_samples_per_second": 18.617, |
|
"eval_steps_per_second": 2.33, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 3.168929528031773e-06, |
|
"loss": 1.2174, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"learning_rate": 3.076567839660109e-06, |
|
"loss": 1.2101, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"learning_rate": 2.9842061512884458e-06, |
|
"loss": 1.2199, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"learning_rate": 2.8918444629167824e-06, |
|
"loss": 1.2365, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"learning_rate": 2.799482774545119e-06, |
|
"loss": 1.218, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"eval_loss": 1.2891569137573242, |
|
"eval_runtime": 272.1682, |
|
"eval_samples_per_second": 18.613, |
|
"eval_steps_per_second": 2.329, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"learning_rate": 2.7071210861734553e-06, |
|
"loss": 1.2228, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"learning_rate": 2.614759397801792e-06, |
|
"loss": 1.2309, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 2.5223977094301287e-06, |
|
"loss": 1.2271, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 2.430036021058465e-06, |
|
"loss": 1.2194, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"learning_rate": 2.3376743326868016e-06, |
|
"loss": 1.214, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"eval_loss": 1.2882676124572754, |
|
"eval_runtime": 272.1318, |
|
"eval_samples_per_second": 18.616, |
|
"eval_steps_per_second": 2.33, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 2.2453126443151382e-06, |
|
"loss": 1.2166, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 2.152950955943475e-06, |
|
"loss": 1.2048, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 4.07, |
|
"learning_rate": 2.0605892675718116e-06, |
|
"loss": 1.2171, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 4.11, |
|
"learning_rate": 1.968227579200148e-06, |
|
"loss": 1.2093, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 4.16, |
|
"learning_rate": 1.8758658908284845e-06, |
|
"loss": 1.1948, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 4.16, |
|
"eval_loss": 1.2930774688720703, |
|
"eval_runtime": 272.152, |
|
"eval_samples_per_second": 18.615, |
|
"eval_steps_per_second": 2.33, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 4.2, |
|
"learning_rate": 1.7835042024568211e-06, |
|
"loss": 1.2097, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 4.24, |
|
"learning_rate": 1.6911425140851576e-06, |
|
"loss": 1.2187, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 1.5987808257134942e-06, |
|
"loss": 1.1994, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 4.32, |
|
"learning_rate": 1.5064191373418307e-06, |
|
"loss": 1.2129, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 4.36, |
|
"learning_rate": 1.4140574489701671e-06, |
|
"loss": 1.2067, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 4.36, |
|
"eval_loss": 1.292514681816101, |
|
"eval_runtime": 272.1549, |
|
"eval_samples_per_second": 18.614, |
|
"eval_steps_per_second": 2.33, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 4.4, |
|
"learning_rate": 1.321695760598504e-06, |
|
"loss": 1.2075, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 4.45, |
|
"learning_rate": 1.2293340722268405e-06, |
|
"loss": 1.2224, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 4.49, |
|
"learning_rate": 1.1369723838551771e-06, |
|
"loss": 1.212, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 4.53, |
|
"learning_rate": 1.0446106954835136e-06, |
|
"loss": 1.2171, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 4.57, |
|
"learning_rate": 9.522490071118501e-07, |
|
"loss": 1.2106, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 4.57, |
|
"eval_loss": 1.2926254272460938, |
|
"eval_runtime": 272.1417, |
|
"eval_samples_per_second": 18.615, |
|
"eval_steps_per_second": 2.33, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 4.61, |
|
"learning_rate": 8.598873187401867e-07, |
|
"loss": 1.2025, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 4.65, |
|
"learning_rate": 7.675256303685231e-07, |
|
"loss": 1.2081, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 4.7, |
|
"learning_rate": 6.751639419968597e-07, |
|
"loss": 1.1977, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 4.74, |
|
"learning_rate": 5.828022536251963e-07, |
|
"loss": 1.2217, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 4.78, |
|
"learning_rate": 4.904405652535329e-07, |
|
"loss": 1.2026, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 4.78, |
|
"eval_loss": 1.2920366525650024, |
|
"eval_runtime": 272.1179, |
|
"eval_samples_per_second": 18.617, |
|
"eval_steps_per_second": 2.33, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 4.82, |
|
"learning_rate": 3.9807887688186943e-07, |
|
"loss": 1.214, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"learning_rate": 3.05717188510206e-07, |
|
"loss": 1.2025, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 4.9, |
|
"learning_rate": 2.1335550013854254e-07, |
|
"loss": 1.225, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"learning_rate": 1.209938117668791e-07, |
|
"loss": 1.2101, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 4.99, |
|
"learning_rate": 2.8632123395215668e-08, |
|
"loss": 1.1956, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 4.99, |
|
"eval_loss": 1.2920278310775757, |
|
"eval_runtime": 272.1447, |
|
"eval_samples_per_second": 18.615, |
|
"eval_steps_per_second": 2.33, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"step": 12030, |
|
"total_flos": 1.5667694787731915e+19, |
|
"train_loss": 1.2596230482321824, |
|
"train_runtime": 116835.9972, |
|
"train_samples_per_second": 4.119, |
|
"train_steps_per_second": 0.103 |
|
} |
|
], |
|
"max_steps": 12030, |
|
"num_train_epochs": 5, |
|
"total_flos": 1.5667694787731915e+19, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|