|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 5.0, |
|
"global_step": 122720, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.991900260756193e-05, |
|
"loss": 0.87, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.983751629726206e-05, |
|
"loss": 0.7054, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.9756029986962192e-05, |
|
"loss": 0.661, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.9674543676662323e-05, |
|
"loss": 0.6353, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9593057366362454e-05, |
|
"loss": 0.626, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9511571056062585e-05, |
|
"loss": 0.6097, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.9430084745762713e-05, |
|
"loss": 0.5873, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9348598435462844e-05, |
|
"loss": 0.6014, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.9267275097783573e-05, |
|
"loss": 0.5693, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.9185951760104305e-05, |
|
"loss": 0.5796, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.9104465449804433e-05, |
|
"loss": 0.5727, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.9022979139504564e-05, |
|
"loss": 0.5765, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.8941492829204696e-05, |
|
"loss": 0.5676, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.8860495436766623e-05, |
|
"loss": 0.561, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.8779009126466758e-05, |
|
"loss": 0.55, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.8697522816166886e-05, |
|
"loss": 0.5409, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.8616036505867017e-05, |
|
"loss": 0.5422, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.8534550195567148e-05, |
|
"loss": 0.5615, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.8453063885267276e-05, |
|
"loss": 0.5539, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.8371577574967407e-05, |
|
"loss": 0.5484, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.8290091264667538e-05, |
|
"loss": 0.5325, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.820860495436767e-05, |
|
"loss": 0.5254, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.8127118644067797e-05, |
|
"loss": 0.5269, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.804563233376793e-05, |
|
"loss": 0.5443, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.796414602346806e-05, |
|
"loss": 0.5331, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.7882659713168187e-05, |
|
"loss": 0.5199, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.780117340286832e-05, |
|
"loss": 0.5298, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.771968709256845e-05, |
|
"loss": 0.5314, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.763820078226858e-05, |
|
"loss": 0.5179, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.755687744458931e-05, |
|
"loss": 0.5288, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.747539113428944e-05, |
|
"loss": 0.525, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.7393904823989572e-05, |
|
"loss": 0.5161, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.73124185136897e-05, |
|
"loss": 0.5243, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.723093220338983e-05, |
|
"loss": 0.5145, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.7149445893089963e-05, |
|
"loss": 0.5353, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.7068285528031293e-05, |
|
"loss": 0.5261, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.6986799217731424e-05, |
|
"loss": 0.5178, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.690531290743155e-05, |
|
"loss": 0.5108, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.6823826597131683e-05, |
|
"loss": 0.5169, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.6742340286831814e-05, |
|
"loss": 0.4949, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.6660853976531942e-05, |
|
"loss": 0.5012, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.6579367666232073e-05, |
|
"loss": 0.5169, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.6497881355932204e-05, |
|
"loss": 0.5201, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.6416395045632335e-05, |
|
"loss": 0.4954, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.6334908735332466e-05, |
|
"loss": 0.5005, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.6253422425032598e-05, |
|
"loss": 0.5063, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.6172099087353327e-05, |
|
"loss": 0.4957, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.6090612777053454e-05, |
|
"loss": 0.4903, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.6009126466753586e-05, |
|
"loss": 0.4883, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.5927640156453717e-05, |
|
"loss": 0.3715, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 1.5846153846153848e-05, |
|
"loss": 0.3628, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.5764830508474577e-05, |
|
"loss": 0.3589, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.5683344198174708e-05, |
|
"loss": 0.3532, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.560185788787484e-05, |
|
"loss": 0.3578, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.5520371577574967e-05, |
|
"loss": 0.3897, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.54390482398957e-05, |
|
"loss": 0.3671, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.535756192959583e-05, |
|
"loss": 0.3505, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.5276075619295962e-05, |
|
"loss": 0.3782, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.5194589308996091e-05, |
|
"loss": 0.3765, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.511310299869622e-05, |
|
"loss": 0.3658, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.503161668839635e-05, |
|
"loss": 0.3752, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.4950130378096481e-05, |
|
"loss": 0.3762, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.4868644067796611e-05, |
|
"loss": 0.3742, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.4787157757496742e-05, |
|
"loss": 0.3748, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.4705834419817473e-05, |
|
"loss": 0.3628, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 1.4624348109517604e-05, |
|
"loss": 0.3635, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.4542861799217733e-05, |
|
"loss": 0.3824, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.4461375488917863e-05, |
|
"loss": 0.3639, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.4380215123859193e-05, |
|
"loss": 0.3632, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.4298728813559324e-05, |
|
"loss": 0.3739, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.4217242503259454e-05, |
|
"loss": 0.3871, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.4135756192959585e-05, |
|
"loss": 0.3843, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.4054269882659714e-05, |
|
"loss": 0.3855, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.3972783572359845e-05, |
|
"loss": 0.3836, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 1.3891297262059975e-05, |
|
"loss": 0.3762, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 1.3809973924380706e-05, |
|
"loss": 0.3816, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 1.3728487614080837e-05, |
|
"loss": 0.3808, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 1.3647001303780966e-05, |
|
"loss": 0.3672, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 1.3565514993481097e-05, |
|
"loss": 0.3776, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 1.3484191655801826e-05, |
|
"loss": 0.3735, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 1.3402868318122557e-05, |
|
"loss": 0.3684, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 1.3321382007822688e-05, |
|
"loss": 0.3794, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 1.3239895697522818e-05, |
|
"loss": 0.3877, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 1.3158409387222949e-05, |
|
"loss": 0.369, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 1.3077249022164277e-05, |
|
"loss": 0.3797, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 1.2995762711864408e-05, |
|
"loss": 0.3814, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 1.2914276401564538e-05, |
|
"loss": 0.3805, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 1.2832790091264669e-05, |
|
"loss": 0.3756, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 1.2751303780964798e-05, |
|
"loss": 0.3884, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.266981747066493e-05, |
|
"loss": 0.3815, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 1.2588331160365059e-05, |
|
"loss": 0.3754, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.250684485006519e-05, |
|
"loss": 0.3738, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.242535853976532e-05, |
|
"loss": 0.3824, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.234387222946545e-05, |
|
"loss": 0.3752, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 1.226238591916558e-05, |
|
"loss": 0.3719, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 1.218089960886571e-05, |
|
"loss": 0.3749, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 1.2099413298565843e-05, |
|
"loss": 0.3674, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 1.2017926988265974e-05, |
|
"loss": 0.3789, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 1.1936440677966104e-05, |
|
"loss": 0.2588, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 1.1855117340286832e-05, |
|
"loss": 0.2309, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 1.1773794002607563e-05, |
|
"loss": 0.2401, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 1.1692470664928292e-05, |
|
"loss": 0.2278, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 1.1610984354628423e-05, |
|
"loss": 0.234, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 1.1529498044328553e-05, |
|
"loss": 0.2319, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 1.1448011734028684e-05, |
|
"loss": 0.2405, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 1.1366525423728813e-05, |
|
"loss": 0.2346, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 1.1285039113428945e-05, |
|
"loss": 0.2382, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 1.1203552803129076e-05, |
|
"loss": 0.2455, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 1.1122066492829207e-05, |
|
"loss": 0.2385, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 1.1040580182529336e-05, |
|
"loss": 0.2447, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 1.0959093872229468e-05, |
|
"loss": 0.244, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 1.0877607561929597e-05, |
|
"loss": 0.2324, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 1.0796121251629728e-05, |
|
"loss": 0.2533, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 1.0714634941329858e-05, |
|
"loss": 0.2406, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 1.0633311603650587e-05, |
|
"loss": 0.2474, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 1.0551825293350716e-05, |
|
"loss": 0.2397, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 1.0470338983050847e-05, |
|
"loss": 0.2433, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 1.038885267275098e-05, |
|
"loss": 0.2568, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 1.030752933507171e-05, |
|
"loss": 0.227, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 1.0226043024771839e-05, |
|
"loss": 0.2339, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 1.014455671447197e-05, |
|
"loss": 0.2414, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 1.00630704041721e-05, |
|
"loss": 0.2448, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 9.98158409387223e-06, |
|
"loss": 0.255, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 9.900097783572362e-06, |
|
"loss": 0.2332, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 9.81877444589309e-06, |
|
"loss": 0.2453, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 9.737288135593222e-06, |
|
"loss": 0.2319, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 9.655801825293351e-06, |
|
"loss": 0.2517, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 9.574315514993483e-06, |
|
"loss": 0.2449, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 9.492992177314211e-06, |
|
"loss": 0.247, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 9.411505867014343e-06, |
|
"loss": 0.2441, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 9.330019556714472e-06, |
|
"loss": 0.241, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 9.248533246414603e-06, |
|
"loss": 0.2583, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 9.167209908735332e-06, |
|
"loss": 0.2497, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 9.085723598435463e-06, |
|
"loss": 0.2426, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 9.004237288135595e-06, |
|
"loss": 0.2529, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 8.922750977835724e-06, |
|
"loss": 0.2416, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 8.841427640156455e-06, |
|
"loss": 0.2468, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 8.759941329856584e-06, |
|
"loss": 0.2612, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 8.678455019556715e-06, |
|
"loss": 0.2496, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 8.596968709256845e-06, |
|
"loss": 0.2644, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 8.515482398956976e-06, |
|
"loss": 0.2314, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 8.433996088657106e-06, |
|
"loss": 0.2427, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 8.352672750977836e-06, |
|
"loss": 0.2529, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 8.271186440677966e-06, |
|
"loss": 0.2463, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 8.189700130378097e-06, |
|
"loss": 0.2446, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 8.108213820078228e-06, |
|
"loss": 0.2457, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 8.026727509778358e-06, |
|
"loss": 0.2435, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 7.945241199478489e-06, |
|
"loss": 0.1715, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 7.863917861799218e-06, |
|
"loss": 0.1444, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 7.782431551499349e-06, |
|
"loss": 0.1493, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 7.700945241199478e-06, |
|
"loss": 0.1546, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 7.6194589308996095e-06, |
|
"loss": 0.1532, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 7.53797262059974e-06, |
|
"loss": 0.1717, |
|
"step": 76500 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 7.45664928292047e-06, |
|
"loss": 0.1646, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 7.375162972620601e-06, |
|
"loss": 0.1613, |
|
"step": 77500 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 7.293676662320731e-06, |
|
"loss": 0.1695, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 7.2121903520208615e-06, |
|
"loss": 0.1655, |
|
"step": 78500 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 7.130867014341591e-06, |
|
"loss": 0.1509, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 7.0493807040417216e-06, |
|
"loss": 0.1621, |
|
"step": 79500 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 6.967894393741852e-06, |
|
"loss": 0.1589, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 6.886408083441982e-06, |
|
"loss": 0.1693, |
|
"step": 80500 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 6.804921773142113e-06, |
|
"loss": 0.1702, |
|
"step": 81000 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 6.723598435462842e-06, |
|
"loss": 0.1749, |
|
"step": 81500 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 6.642112125162973e-06, |
|
"loss": 0.1603, |
|
"step": 82000 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 6.560625814863103e-06, |
|
"loss": 0.1671, |
|
"step": 82500 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 6.479139504563234e-06, |
|
"loss": 0.1714, |
|
"step": 83000 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 6.397816166883965e-06, |
|
"loss": 0.1722, |
|
"step": 83500 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 6.316329856584094e-06, |
|
"loss": 0.1686, |
|
"step": 84000 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 6.234843546284225e-06, |
|
"loss": 0.1553, |
|
"step": 84500 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 6.153357235984355e-06, |
|
"loss": 0.1584, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 6.072033898305086e-06, |
|
"loss": 0.1665, |
|
"step": 85500 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 5.990547588005216e-06, |
|
"loss": 0.1607, |
|
"step": 86000 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"learning_rate": 5.9090612777053454e-06, |
|
"loss": 0.1748, |
|
"step": 86500 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 5.827737940026076e-06, |
|
"loss": 0.1688, |
|
"step": 87000 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 5.746251629726206e-06, |
|
"loss": 0.1681, |
|
"step": 87500 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"learning_rate": 5.664765319426337e-06, |
|
"loss": 0.171, |
|
"step": 88000 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 5.583279009126468e-06, |
|
"loss": 0.1744, |
|
"step": 88500 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"learning_rate": 5.501792698826598e-06, |
|
"loss": 0.1648, |
|
"step": 89000 |
|
}, |
|
{ |
|
"epoch": 3.65, |
|
"learning_rate": 5.420306388526729e-06, |
|
"loss": 0.1645, |
|
"step": 89500 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"learning_rate": 5.338820078226858e-06, |
|
"loss": 0.1518, |
|
"step": 90000 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"learning_rate": 5.2573337679269884e-06, |
|
"loss": 0.1715, |
|
"step": 90500 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 5.175847457627119e-06, |
|
"loss": 0.1768, |
|
"step": 91000 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 5.094524119947849e-06, |
|
"loss": 0.1655, |
|
"step": 91500 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 5.013037809647979e-06, |
|
"loss": 0.1783, |
|
"step": 92000 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 4.93155149934811e-06, |
|
"loss": 0.1721, |
|
"step": 92500 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"learning_rate": 4.85006518904824e-06, |
|
"loss": 0.1683, |
|
"step": 93000 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 4.76874185136897e-06, |
|
"loss": 0.1725, |
|
"step": 93500 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"learning_rate": 4.6872555410691005e-06, |
|
"loss": 0.1744, |
|
"step": 94000 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"learning_rate": 4.605769230769231e-06, |
|
"loss": 0.1729, |
|
"step": 94500 |
|
}, |
|
{ |
|
"epoch": 3.87, |
|
"learning_rate": 4.524282920469362e-06, |
|
"loss": 0.1642, |
|
"step": 95000 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 4.442959582790092e-06, |
|
"loss": 0.1653, |
|
"step": 95500 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 4.361473272490222e-06, |
|
"loss": 0.1914, |
|
"step": 96000 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 4.2799869621903525e-06, |
|
"loss": 0.1626, |
|
"step": 96500 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"learning_rate": 4.198500651890483e-06, |
|
"loss": 0.1808, |
|
"step": 97000 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"learning_rate": 4.117014341590613e-06, |
|
"loss": 0.1617, |
|
"step": 97500 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 4.035691003911344e-06, |
|
"loss": 0.1755, |
|
"step": 98000 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 3.954204693611473e-06, |
|
"loss": 0.1162, |
|
"step": 98500 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 3.872881355932204e-06, |
|
"loss": 0.0986, |
|
"step": 99000 |
|
}, |
|
{ |
|
"epoch": 4.05, |
|
"learning_rate": 3.7913950456323338e-06, |
|
"loss": 0.0932, |
|
"step": 99500 |
|
}, |
|
{ |
|
"epoch": 4.07, |
|
"learning_rate": 3.709908735332464e-06, |
|
"loss": 0.0952, |
|
"step": 100000 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"learning_rate": 3.628422425032595e-06, |
|
"loss": 0.0933, |
|
"step": 100500 |
|
}, |
|
{ |
|
"epoch": 4.12, |
|
"learning_rate": 3.5469361147327252e-06, |
|
"loss": 0.1057, |
|
"step": 101000 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"learning_rate": 3.4654498044328556e-06, |
|
"loss": 0.0988, |
|
"step": 101500 |
|
}, |
|
{ |
|
"epoch": 4.16, |
|
"learning_rate": 3.383963494132986e-06, |
|
"loss": 0.0894, |
|
"step": 102000 |
|
}, |
|
{ |
|
"epoch": 4.18, |
|
"learning_rate": 3.3024771838331167e-06, |
|
"loss": 0.0959, |
|
"step": 102500 |
|
}, |
|
{ |
|
"epoch": 4.2, |
|
"learning_rate": 3.221316818774446e-06, |
|
"loss": 0.1033, |
|
"step": 103000 |
|
}, |
|
{ |
|
"epoch": 4.22, |
|
"learning_rate": 3.1398305084745766e-06, |
|
"loss": 0.1019, |
|
"step": 103500 |
|
}, |
|
{ |
|
"epoch": 4.24, |
|
"learning_rate": 3.058344198174707e-06, |
|
"loss": 0.0942, |
|
"step": 104000 |
|
}, |
|
{ |
|
"epoch": 4.26, |
|
"learning_rate": 2.9768578878748373e-06, |
|
"loss": 0.1, |
|
"step": 104500 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 2.8955345501955674e-06, |
|
"loss": 0.1174, |
|
"step": 105000 |
|
}, |
|
{ |
|
"epoch": 4.3, |
|
"learning_rate": 2.8140482398956974e-06, |
|
"loss": 0.0919, |
|
"step": 105500 |
|
}, |
|
{ |
|
"epoch": 4.32, |
|
"learning_rate": 2.732561929595828e-06, |
|
"loss": 0.1201, |
|
"step": 106000 |
|
}, |
|
{ |
|
"epoch": 4.34, |
|
"learning_rate": 2.6510756192959585e-06, |
|
"loss": 0.1025, |
|
"step": 106500 |
|
}, |
|
{ |
|
"epoch": 4.36, |
|
"learning_rate": 2.5697522816166886e-06, |
|
"loss": 0.0948, |
|
"step": 107000 |
|
}, |
|
{ |
|
"epoch": 4.38, |
|
"learning_rate": 2.488265971316819e-06, |
|
"loss": 0.092, |
|
"step": 107500 |
|
}, |
|
{ |
|
"epoch": 4.4, |
|
"learning_rate": 2.4067796610169493e-06, |
|
"loss": 0.1014, |
|
"step": 108000 |
|
}, |
|
{ |
|
"epoch": 4.42, |
|
"learning_rate": 2.3252933507170797e-06, |
|
"loss": 0.0994, |
|
"step": 108500 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"learning_rate": 2.24380704041721e-06, |
|
"loss": 0.0776, |
|
"step": 109000 |
|
}, |
|
{ |
|
"epoch": 4.46, |
|
"learning_rate": 2.1623207301173404e-06, |
|
"loss": 0.0908, |
|
"step": 109500 |
|
}, |
|
{ |
|
"epoch": 4.48, |
|
"learning_rate": 2.0808344198174707e-06, |
|
"loss": 0.0979, |
|
"step": 110000 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"learning_rate": 1.999511082138201e-06, |
|
"loss": 0.0908, |
|
"step": 110500 |
|
}, |
|
{ |
|
"epoch": 4.52, |
|
"learning_rate": 1.9180247718383312e-06, |
|
"loss": 0.1044, |
|
"step": 111000 |
|
}, |
|
{ |
|
"epoch": 4.54, |
|
"learning_rate": 1.8365384615384618e-06, |
|
"loss": 0.099, |
|
"step": 111500 |
|
}, |
|
{ |
|
"epoch": 4.56, |
|
"learning_rate": 1.7550521512385921e-06, |
|
"loss": 0.0967, |
|
"step": 112000 |
|
}, |
|
{ |
|
"epoch": 4.58, |
|
"learning_rate": 1.6735658409387223e-06, |
|
"loss": 0.0871, |
|
"step": 112500 |
|
}, |
|
{ |
|
"epoch": 4.6, |
|
"learning_rate": 1.5922425032594526e-06, |
|
"loss": 0.101, |
|
"step": 113000 |
|
}, |
|
{ |
|
"epoch": 4.62, |
|
"learning_rate": 1.510756192959583e-06, |
|
"loss": 0.0913, |
|
"step": 113500 |
|
}, |
|
{ |
|
"epoch": 4.64, |
|
"learning_rate": 1.4292698826597131e-06, |
|
"loss": 0.0887, |
|
"step": 114000 |
|
}, |
|
{ |
|
"epoch": 4.67, |
|
"learning_rate": 1.3477835723598437e-06, |
|
"loss": 0.089, |
|
"step": 114500 |
|
}, |
|
{ |
|
"epoch": 4.69, |
|
"learning_rate": 1.266297262059974e-06, |
|
"loss": 0.0973, |
|
"step": 115000 |
|
}, |
|
{ |
|
"epoch": 4.71, |
|
"learning_rate": 1.1848109517601044e-06, |
|
"loss": 0.0938, |
|
"step": 115500 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"learning_rate": 1.1033246414602347e-06, |
|
"loss": 0.1079, |
|
"step": 116000 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"learning_rate": 1.0220013037809649e-06, |
|
"loss": 0.1027, |
|
"step": 116500 |
|
}, |
|
{ |
|
"epoch": 4.77, |
|
"learning_rate": 9.405149934810953e-07, |
|
"loss": 0.0991, |
|
"step": 117000 |
|
}, |
|
{ |
|
"epoch": 4.79, |
|
"learning_rate": 8.590286831812256e-07, |
|
"loss": 0.1028, |
|
"step": 117500 |
|
}, |
|
{ |
|
"epoch": 4.81, |
|
"learning_rate": 7.775423728813559e-07, |
|
"loss": 0.0997, |
|
"step": 118000 |
|
}, |
|
{ |
|
"epoch": 4.83, |
|
"learning_rate": 6.962190352020862e-07, |
|
"loss": 0.0964, |
|
"step": 118500 |
|
}, |
|
{ |
|
"epoch": 4.85, |
|
"learning_rate": 6.147327249022164e-07, |
|
"loss": 0.0954, |
|
"step": 119000 |
|
}, |
|
{ |
|
"epoch": 4.87, |
|
"learning_rate": 5.332464146023469e-07, |
|
"loss": 0.0963, |
|
"step": 119500 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 4.517601043024772e-07, |
|
"loss": 0.0816, |
|
"step": 120000 |
|
}, |
|
{ |
|
"epoch": 4.91, |
|
"learning_rate": 3.702737940026076e-07, |
|
"loss": 0.1004, |
|
"step": 120500 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 2.889504563233377e-07, |
|
"loss": 0.1038, |
|
"step": 121000 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"learning_rate": 2.0746414602346806e-07, |
|
"loss": 0.0843, |
|
"step": 121500 |
|
}, |
|
{ |
|
"epoch": 4.97, |
|
"learning_rate": 1.2597783572359844e-07, |
|
"loss": 0.1104, |
|
"step": 122000 |
|
}, |
|
{ |
|
"epoch": 4.99, |
|
"learning_rate": 4.449152542372882e-08, |
|
"loss": 0.0853, |
|
"step": 122500 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"step": 122720, |
|
"total_flos": 1.2826585381404672e+17, |
|
"train_loss": 0.2862651685536918, |
|
"train_runtime": 5523.0542, |
|
"train_samples_per_second": 355.512, |
|
"train_steps_per_second": 22.22 |
|
} |
|
], |
|
"max_steps": 122720, |
|
"num_train_epochs": 5, |
|
"total_flos": 1.2826585381404672e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|