|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 10.0, |
|
"global_step": 77340, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.935350400827516e-05, |
|
"loss": 8.0382, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.87070080165503e-05, |
|
"loss": 7.6429, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.806051202482544e-05, |
|
"loss": 7.1341, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.74140160331006e-05, |
|
"loss": 6.9301, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.676752004137574e-05, |
|
"loss": 6.8441, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.61210240496509e-05, |
|
"loss": 6.7735, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.547452805792605e-05, |
|
"loss": 6.3512, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.482803206620118e-05, |
|
"loss": 5.6406, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.418153607447634e-05, |
|
"loss": 5.2743, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.35350400827515e-05, |
|
"loss": 5.0399, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 9.288854409102664e-05, |
|
"loss": 4.8407, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 9.224204809930179e-05, |
|
"loss": 4.6694, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 9.159555210757695e-05, |
|
"loss": 4.5262, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 9.094905611585208e-05, |
|
"loss": 4.4021, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 9.030256012412723e-05, |
|
"loss": 4.2888, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 4.0826334953308105, |
|
"eval_runtime": 475.8214, |
|
"eval_samples_per_second": 693.498, |
|
"eval_steps_per_second": 1.807, |
|
"step": 7734 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 8.965606413240239e-05, |
|
"loss": 4.1903, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 8.900956814067753e-05, |
|
"loss": 4.1066, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 8.836307214895269e-05, |
|
"loss": 4.0331, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 8.771657615722783e-05, |
|
"loss": 3.9644, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 8.707008016550297e-05, |
|
"loss": 3.9044, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 8.642358417377813e-05, |
|
"loss": 3.851, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 8.577708818205328e-05, |
|
"loss": 3.8044, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 8.513059219032842e-05, |
|
"loss": 3.761, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 8.448409619860357e-05, |
|
"loss": 3.7197, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 8.383760020687872e-05, |
|
"loss": 3.6803, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 8.319110421515386e-05, |
|
"loss": 3.6381, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 8.254460822342902e-05, |
|
"loss": 3.6119, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 8.189811223170418e-05, |
|
"loss": 3.577, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 8.12516162399793e-05, |
|
"loss": 3.5515, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 8.060512024825446e-05, |
|
"loss": 3.5228, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 3.361564874649048, |
|
"eval_runtime": 476.9803, |
|
"eval_samples_per_second": 691.813, |
|
"eval_steps_per_second": 1.803, |
|
"step": 15468 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 7.995862425652962e-05, |
|
"loss": 3.4922, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 7.931212826480476e-05, |
|
"loss": 3.4583, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 7.866563227307991e-05, |
|
"loss": 3.4363, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 7.801913628135506e-05, |
|
"loss": 3.4124, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 7.73726402896302e-05, |
|
"loss": 3.3911, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 7.672614429790535e-05, |
|
"loss": 3.3731, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 7.607964830618051e-05, |
|
"loss": 3.3487, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 7.543315231445565e-05, |
|
"loss": 3.333, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 7.478665632273081e-05, |
|
"loss": 3.3162, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 7.414016033100595e-05, |
|
"loss": 3.2947, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 7.349366433928109e-05, |
|
"loss": 3.2783, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 7.284716834755625e-05, |
|
"loss": 3.2641, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 7.22006723558314e-05, |
|
"loss": 3.2446, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 7.155417636410655e-05, |
|
"loss": 3.2326, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 7.090768037238169e-05, |
|
"loss": 3.2149, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 7.026118438065684e-05, |
|
"loss": 3.2034, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 3.070152997970581, |
|
"eval_runtime": 476.4651, |
|
"eval_samples_per_second": 692.561, |
|
"eval_steps_per_second": 1.805, |
|
"step": 23202 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 6.961468838893199e-05, |
|
"loss": 3.1826, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 6.896819239720714e-05, |
|
"loss": 3.1653, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 6.83216964054823e-05, |
|
"loss": 3.1519, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 6.767520041375744e-05, |
|
"loss": 3.1387, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 6.702870442203258e-05, |
|
"loss": 3.1317, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 6.638220843030774e-05, |
|
"loss": 3.1206, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 6.573571243858288e-05, |
|
"loss": 3.1074, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"learning_rate": 6.508921644685804e-05, |
|
"loss": 3.0959, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 6.444272045513318e-05, |
|
"loss": 3.0846, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"learning_rate": 6.379622446340833e-05, |
|
"loss": 3.0691, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"learning_rate": 6.314972847168348e-05, |
|
"loss": 3.0621, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 6.250323247995863e-05, |
|
"loss": 3.0496, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 6.185673648823377e-05, |
|
"loss": 3.0406, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 6.121024049650892e-05, |
|
"loss": 3.0332, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"learning_rate": 6.056374450478407e-05, |
|
"loss": 3.0235, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 2.890948534011841, |
|
"eval_runtime": 475.9375, |
|
"eval_samples_per_second": 693.328, |
|
"eval_steps_per_second": 1.807, |
|
"step": 30936 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 5.991724851305922e-05, |
|
"loss": 3.0144, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 4.07, |
|
"learning_rate": 5.927075252133437e-05, |
|
"loss": 2.9923, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"learning_rate": 5.8624256529609526e-05, |
|
"loss": 2.9902, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 4.2, |
|
"learning_rate": 5.7977760537884675e-05, |
|
"loss": 2.9746, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 4.27, |
|
"learning_rate": 5.733126454615981e-05, |
|
"loss": 2.9708, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"learning_rate": 5.6684768554434966e-05, |
|
"loss": 2.9586, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 4.4, |
|
"learning_rate": 5.6038272562710115e-05, |
|
"loss": 2.9556, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 4.46, |
|
"learning_rate": 5.5391776570985264e-05, |
|
"loss": 2.9471, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 4.53, |
|
"learning_rate": 5.474528057926042e-05, |
|
"loss": 2.941, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 4.59, |
|
"learning_rate": 5.4098784587535555e-05, |
|
"loss": 2.9353, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 4.65, |
|
"learning_rate": 5.3452288595810704e-05, |
|
"loss": 2.9263, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 4.72, |
|
"learning_rate": 5.280579260408586e-05, |
|
"loss": 2.9193, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 4.78, |
|
"learning_rate": 5.215929661236101e-05, |
|
"loss": 2.9101, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 4.85, |
|
"learning_rate": 5.151280062063616e-05, |
|
"loss": 2.9071, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 4.91, |
|
"learning_rate": 5.08663046289113e-05, |
|
"loss": 2.9019, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 4.98, |
|
"learning_rate": 5.021980863718645e-05, |
|
"loss": 2.89, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 2.768686294555664, |
|
"eval_runtime": 485.6271, |
|
"eval_samples_per_second": 679.495, |
|
"eval_steps_per_second": 1.771, |
|
"step": 38670 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 4.95733126454616e-05, |
|
"loss": 2.8791, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 5.11, |
|
"learning_rate": 4.892681665373675e-05, |
|
"loss": 2.87, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 5.17, |
|
"learning_rate": 4.8280320662011896e-05, |
|
"loss": 2.8664, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 5.24, |
|
"learning_rate": 4.7633824670287045e-05, |
|
"loss": 2.859, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 5.3, |
|
"learning_rate": 4.6987328678562193e-05, |
|
"loss": 2.8562, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 5.37, |
|
"learning_rate": 4.634083268683734e-05, |
|
"loss": 2.8489, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 5.43, |
|
"learning_rate": 4.569433669511249e-05, |
|
"loss": 2.846, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 5.5, |
|
"learning_rate": 4.504784070338764e-05, |
|
"loss": 2.8342, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 5.56, |
|
"learning_rate": 4.440134471166279e-05, |
|
"loss": 2.8313, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 5.62, |
|
"learning_rate": 4.375484871993794e-05, |
|
"loss": 2.8217, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"learning_rate": 4.310835272821309e-05, |
|
"loss": 2.8197, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 5.75, |
|
"learning_rate": 4.2461856736488236e-05, |
|
"loss": 2.8158, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 5.82, |
|
"learning_rate": 4.1815360744763385e-05, |
|
"loss": 2.8094, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 5.88, |
|
"learning_rate": 4.1168864753038534e-05, |
|
"loss": 2.8061, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 5.95, |
|
"learning_rate": 4.052236876131368e-05, |
|
"loss": 2.8032, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 2.681367874145508, |
|
"eval_runtime": 476.3627, |
|
"eval_samples_per_second": 692.71, |
|
"eval_steps_per_second": 1.805, |
|
"step": 46404 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 3.987587276958883e-05, |
|
"loss": 2.794, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 6.08, |
|
"learning_rate": 3.922937677786398e-05, |
|
"loss": 2.7858, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 6.14, |
|
"learning_rate": 3.858288078613912e-05, |
|
"loss": 2.7765, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 6.21, |
|
"learning_rate": 3.793638479441428e-05, |
|
"loss": 2.7752, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 6.27, |
|
"learning_rate": 3.728988880268943e-05, |
|
"loss": 2.7725, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 6.34, |
|
"learning_rate": 3.664339281096457e-05, |
|
"loss": 2.7696, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 6.4, |
|
"learning_rate": 3.5996896819239726e-05, |
|
"loss": 2.7665, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 6.46, |
|
"learning_rate": 3.5350400827514875e-05, |
|
"loss": 2.7622, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 6.53, |
|
"learning_rate": 3.470390483579002e-05, |
|
"loss": 2.7566, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 6.59, |
|
"learning_rate": 3.4057408844065166e-05, |
|
"loss": 2.7543, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 6.66, |
|
"learning_rate": 3.3410912852340315e-05, |
|
"loss": 2.7451, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 6.72, |
|
"learning_rate": 3.2764416860615464e-05, |
|
"loss": 2.7392, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 6.79, |
|
"learning_rate": 3.211792086889061e-05, |
|
"loss": 2.7384, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 6.85, |
|
"learning_rate": 3.147142487716576e-05, |
|
"loss": 2.7341, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 6.92, |
|
"learning_rate": 3.082492888544091e-05, |
|
"loss": 2.7317, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 6.98, |
|
"learning_rate": 3.0178432893716057e-05, |
|
"loss": 2.7308, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 2.6155083179473877, |
|
"eval_runtime": 476.4568, |
|
"eval_samples_per_second": 692.573, |
|
"eval_steps_per_second": 1.805, |
|
"step": 54138 |
|
}, |
|
{ |
|
"epoch": 7.05, |
|
"learning_rate": 2.953193690199121e-05, |
|
"loss": 2.7178, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 7.11, |
|
"learning_rate": 2.8885440910266358e-05, |
|
"loss": 2.712, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 7.18, |
|
"learning_rate": 2.8238944918541504e-05, |
|
"loss": 2.7152, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 7.24, |
|
"learning_rate": 2.7592448926816656e-05, |
|
"loss": 2.7076, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 7.31, |
|
"learning_rate": 2.6945952935091805e-05, |
|
"loss": 2.7048, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 7.37, |
|
"learning_rate": 2.629945694336695e-05, |
|
"loss": 2.7038, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 7.43, |
|
"learning_rate": 2.5652960951642103e-05, |
|
"loss": 2.6975, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"learning_rate": 2.500646495991725e-05, |
|
"loss": 2.696, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 7.56, |
|
"learning_rate": 2.4359968968192397e-05, |
|
"loss": 2.6941, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 7.63, |
|
"learning_rate": 2.3713472976467546e-05, |
|
"loss": 2.6884, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 7.69, |
|
"learning_rate": 2.3066976984742695e-05, |
|
"loss": 2.6844, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 7.76, |
|
"learning_rate": 2.2420480993017844e-05, |
|
"loss": 2.6791, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 7.82, |
|
"learning_rate": 2.1773985001292993e-05, |
|
"loss": 2.6759, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 7.89, |
|
"learning_rate": 2.1127489009568142e-05, |
|
"loss": 2.6779, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 7.95, |
|
"learning_rate": 2.0480993017843288e-05, |
|
"loss": 2.6749, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 2.561079502105713, |
|
"eval_runtime": 476.3937, |
|
"eval_samples_per_second": 692.665, |
|
"eval_steps_per_second": 1.805, |
|
"step": 61872 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 1.983449702611844e-05, |
|
"loss": 2.6705, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 8.08, |
|
"learning_rate": 1.918800103439359e-05, |
|
"loss": 2.6613, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 8.15, |
|
"learning_rate": 1.8541505042668735e-05, |
|
"loss": 2.6649, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 8.21, |
|
"learning_rate": 1.7895009050943884e-05, |
|
"loss": 2.6597, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 8.28, |
|
"learning_rate": 1.7248513059219036e-05, |
|
"loss": 2.6568, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 8.34, |
|
"learning_rate": 1.660201706749418e-05, |
|
"loss": 2.6551, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 8.4, |
|
"learning_rate": 1.595552107576933e-05, |
|
"loss": 2.6518, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 8.47, |
|
"learning_rate": 1.530902508404448e-05, |
|
"loss": 2.6501, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 8.53, |
|
"learning_rate": 1.466252909231963e-05, |
|
"loss": 2.6469, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 8.6, |
|
"learning_rate": 1.4016033100594778e-05, |
|
"loss": 2.6465, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 8.66, |
|
"learning_rate": 1.3369537108869925e-05, |
|
"loss": 2.6455, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 8.73, |
|
"learning_rate": 1.2723041117145074e-05, |
|
"loss": 2.6398, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 8.79, |
|
"learning_rate": 1.2076545125420223e-05, |
|
"loss": 2.6436, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 8.86, |
|
"learning_rate": 1.1430049133695372e-05, |
|
"loss": 2.6395, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 8.92, |
|
"learning_rate": 1.078355314197052e-05, |
|
"loss": 2.634, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 8.99, |
|
"learning_rate": 1.0137057150245668e-05, |
|
"loss": 2.6359, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 2.5289456844329834, |
|
"eval_runtime": 476.5662, |
|
"eval_samples_per_second": 692.414, |
|
"eval_steps_per_second": 1.805, |
|
"step": 69606 |
|
}, |
|
{ |
|
"epoch": 9.05, |
|
"learning_rate": 9.490561158520819e-06, |
|
"loss": 2.6261, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 9.12, |
|
"learning_rate": 8.844065166795966e-06, |
|
"loss": 2.6228, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 9.18, |
|
"learning_rate": 8.197569175071115e-06, |
|
"loss": 2.6303, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 9.24, |
|
"learning_rate": 7.551073183346263e-06, |
|
"loss": 2.622, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 9.31, |
|
"learning_rate": 6.904577191621413e-06, |
|
"loss": 2.6257, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 9.37, |
|
"learning_rate": 6.258081199896561e-06, |
|
"loss": 2.6244, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 9.44, |
|
"learning_rate": 5.611585208171709e-06, |
|
"loss": 2.6238, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 9.5, |
|
"learning_rate": 4.965089216446858e-06, |
|
"loss": 2.6207, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 9.57, |
|
"learning_rate": 4.318593224722007e-06, |
|
"loss": 2.6168, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 9.63, |
|
"learning_rate": 3.6720972329971556e-06, |
|
"loss": 2.619, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 9.7, |
|
"learning_rate": 3.025601241272304e-06, |
|
"loss": 2.6185, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 9.76, |
|
"learning_rate": 2.379105249547453e-06, |
|
"loss": 2.6173, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 9.83, |
|
"learning_rate": 1.7326092578226015e-06, |
|
"loss": 2.6114, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 9.89, |
|
"learning_rate": 1.0861132660977503e-06, |
|
"loss": 2.6091, |
|
"step": 76500 |
|
}, |
|
{ |
|
"epoch": 9.96, |
|
"learning_rate": 4.396172743728989e-07, |
|
"loss": 2.6133, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 2.5088417530059814, |
|
"eval_runtime": 484.8466, |
|
"eval_samples_per_second": 680.588, |
|
"eval_steps_per_second": 1.774, |
|
"step": 77340 |
|
} |
|
], |
|
"max_steps": 77340, |
|
"num_train_epochs": 10, |
|
"total_flos": 1.9542406291981978e+18, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|