|
{ |
|
"best_metric": 0.3966299593448639, |
|
"best_model_checkpoint": "mikhail-panzo/ceb_b128_le4_s8000/checkpoint-2000", |
|
"epoch": 158.41584158415841, |
|
"eval_steps": 500, |
|
"global_step": 2000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 3.9603960396039604, |
|
"grad_norm": 2.95651912689209, |
|
"learning_rate": 2.5e-06, |
|
"loss": 0.7707, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 7.920792079207921, |
|
"grad_norm": 1.1273679733276367, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6934, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 11.881188118811881, |
|
"grad_norm": 2.1830034255981445, |
|
"learning_rate": 7.5e-06, |
|
"loss": 0.5983, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 15.841584158415841, |
|
"grad_norm": 1.0733364820480347, |
|
"learning_rate": 1e-05, |
|
"loss": 0.5094, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 19.801980198019802, |
|
"grad_norm": 1.2824387550354004, |
|
"learning_rate": 1.25e-05, |
|
"loss": 0.4913, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 23.762376237623762, |
|
"grad_norm": 0.9164418578147888, |
|
"learning_rate": 1.5e-05, |
|
"loss": 0.4764, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 27.722772277227723, |
|
"grad_norm": 0.8938279747962952, |
|
"learning_rate": 1.75e-05, |
|
"loss": 0.4648, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 31.683168316831683, |
|
"grad_norm": 0.7855221033096313, |
|
"learning_rate": 2e-05, |
|
"loss": 0.459, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 35.64356435643565, |
|
"grad_norm": 0.9247761964797974, |
|
"learning_rate": 2.25e-05, |
|
"loss": 0.4519, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 39.603960396039604, |
|
"grad_norm": 1.1086606979370117, |
|
"learning_rate": 2.5e-05, |
|
"loss": 0.4432, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 39.603960396039604, |
|
"eval_loss": 0.4058004915714264, |
|
"eval_runtime": 7.6054, |
|
"eval_samples_per_second": 23.667, |
|
"eval_steps_per_second": 3.024, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 43.56435643564357, |
|
"grad_norm": 1.3994853496551514, |
|
"learning_rate": 2.7500000000000004e-05, |
|
"loss": 0.4438, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 47.524752475247524, |
|
"grad_norm": 1.1104933023452759, |
|
"learning_rate": 3e-05, |
|
"loss": 0.4367, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 51.48514851485149, |
|
"grad_norm": 0.7456745505332947, |
|
"learning_rate": 3.2500000000000004e-05, |
|
"loss": 0.4315, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 55.445544554455445, |
|
"grad_norm": 1.0349308252334595, |
|
"learning_rate": 3.5e-05, |
|
"loss": 0.4294, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 59.40594059405941, |
|
"grad_norm": 0.5290626287460327, |
|
"learning_rate": 3.7500000000000003e-05, |
|
"loss": 0.4234, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 63.366336633663366, |
|
"grad_norm": 1.054995059967041, |
|
"learning_rate": 4e-05, |
|
"loss": 0.4231, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 67.32673267326733, |
|
"grad_norm": 1.1774096488952637, |
|
"learning_rate": 4.25e-05, |
|
"loss": 0.4198, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 71.2871287128713, |
|
"grad_norm": 0.7513113617897034, |
|
"learning_rate": 4.5e-05, |
|
"loss": 0.4182, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 75.24752475247524, |
|
"grad_norm": 0.8465940952301025, |
|
"learning_rate": 4.75e-05, |
|
"loss": 0.4179, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 79.20792079207921, |
|
"grad_norm": 1.601589322090149, |
|
"learning_rate": 5e-05, |
|
"loss": 0.4129, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 79.20792079207921, |
|
"eval_loss": 0.3995668292045593, |
|
"eval_runtime": 8.0229, |
|
"eval_samples_per_second": 22.436, |
|
"eval_steps_per_second": 2.867, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 83.16831683168317, |
|
"grad_norm": 2.049264669418335, |
|
"learning_rate": 5.25e-05, |
|
"loss": 0.4129, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 87.12871287128714, |
|
"grad_norm": 1.0590511560440063, |
|
"learning_rate": 5.500000000000001e-05, |
|
"loss": 0.4118, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 91.08910891089108, |
|
"grad_norm": 1.015751600265503, |
|
"learning_rate": 5.7499999999999995e-05, |
|
"loss": 0.4074, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 95.04950495049505, |
|
"grad_norm": 1.295837640762329, |
|
"learning_rate": 6e-05, |
|
"loss": 0.4056, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 99.00990099009901, |
|
"grad_norm": 1.49947988986969, |
|
"learning_rate": 6.25e-05, |
|
"loss": 0.4052, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 102.97029702970298, |
|
"grad_norm": 1.3930383920669556, |
|
"learning_rate": 6.500000000000001e-05, |
|
"loss": 0.4072, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 106.93069306930693, |
|
"grad_norm": 1.1130273342132568, |
|
"learning_rate": 6.750000000000001e-05, |
|
"loss": 0.4015, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 110.89108910891089, |
|
"grad_norm": 2.1075382232666016, |
|
"learning_rate": 7e-05, |
|
"loss": 0.3994, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 114.85148514851485, |
|
"grad_norm": 1.8470244407653809, |
|
"learning_rate": 7.25e-05, |
|
"loss": 0.397, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 118.81188118811882, |
|
"grad_norm": 2.0715270042419434, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 0.3992, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 118.81188118811882, |
|
"eval_loss": 0.39855822920799255, |
|
"eval_runtime": 7.8631, |
|
"eval_samples_per_second": 22.892, |
|
"eval_steps_per_second": 2.925, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 122.77227722772277, |
|
"grad_norm": 0.9546501636505127, |
|
"learning_rate": 7.75e-05, |
|
"loss": 0.3934, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 126.73267326732673, |
|
"grad_norm": 1.1335992813110352, |
|
"learning_rate": 8e-05, |
|
"loss": 0.3927, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 130.69306930693068, |
|
"grad_norm": 1.905771017074585, |
|
"learning_rate": 8.25e-05, |
|
"loss": 0.3895, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 134.65346534653466, |
|
"grad_norm": 1.129404067993164, |
|
"learning_rate": 8.5e-05, |
|
"loss": 0.3897, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 138.6138613861386, |
|
"grad_norm": 0.9640969038009644, |
|
"learning_rate": 8.75e-05, |
|
"loss": 0.3894, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 142.5742574257426, |
|
"grad_norm": 0.9840293526649475, |
|
"learning_rate": 9e-05, |
|
"loss": 0.3862, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 146.53465346534654, |
|
"grad_norm": 2.770336627960205, |
|
"learning_rate": 9.250000000000001e-05, |
|
"loss": 0.3861, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 150.4950495049505, |
|
"grad_norm": 1.7189514636993408, |
|
"learning_rate": 9.5e-05, |
|
"loss": 0.3914, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 154.45544554455446, |
|
"grad_norm": 2.454707622528076, |
|
"learning_rate": 9.75e-05, |
|
"loss": 0.3848, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 158.41584158415841, |
|
"grad_norm": 2.1555542945861816, |
|
"learning_rate": 0.0001, |
|
"loss": 0.3814, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 158.41584158415841, |
|
"eval_loss": 0.3966299593448639, |
|
"eval_runtime": 7.7159, |
|
"eval_samples_per_second": 23.329, |
|
"eval_steps_per_second": 2.981, |
|
"step": 2000 |
|
} |
|
], |
|
"logging_steps": 50, |
|
"max_steps": 8000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 667, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 4.322114324975938e+16, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|