|
{ |
|
"best_metric": 0.40509456396102905, |
|
"best_model_checkpoint": "mikhail_panzo/fil_b128_le4_s4000/checkpoint-1500", |
|
"epoch": 133.33333333333334, |
|
"eval_steps": 500, |
|
"global_step": 1500, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 4.444444444444445, |
|
"grad_norm": 2.152308940887451, |
|
"learning_rate": 2.5e-06, |
|
"loss": 0.7738, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 8.88888888888889, |
|
"grad_norm": 1.326554536819458, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6968, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 13.333333333333334, |
|
"grad_norm": 1.4387542009353638, |
|
"learning_rate": 7.5e-06, |
|
"loss": 0.636, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 17.77777777777778, |
|
"grad_norm": 1.033190131187439, |
|
"learning_rate": 1e-05, |
|
"loss": 0.5486, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 22.22222222222222, |
|
"grad_norm": 1.4383914470672607, |
|
"learning_rate": 1.25e-05, |
|
"loss": 0.516, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 26.666666666666668, |
|
"grad_norm": 1.8267806768417358, |
|
"learning_rate": 1.5e-05, |
|
"loss": 0.4989, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 31.11111111111111, |
|
"grad_norm": 1.2525017261505127, |
|
"learning_rate": 1.75e-05, |
|
"loss": 0.4908, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 35.55555555555556, |
|
"grad_norm": 1.1662768125534058, |
|
"learning_rate": 2e-05, |
|
"loss": 0.4766, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"grad_norm": 1.110721230506897, |
|
"learning_rate": 2.25e-05, |
|
"loss": 0.4716, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 44.44444444444444, |
|
"grad_norm": 1.6693321466445923, |
|
"learning_rate": 2.5e-05, |
|
"loss": 0.4635, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 44.44444444444444, |
|
"eval_loss": 0.42072969675064087, |
|
"eval_runtime": 9.5188, |
|
"eval_samples_per_second": 16.704, |
|
"eval_steps_per_second": 2.101, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 48.888888888888886, |
|
"grad_norm": 1.0238850116729736, |
|
"learning_rate": 2.7500000000000004e-05, |
|
"loss": 0.4585, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 53.333333333333336, |
|
"grad_norm": 0.8729687929153442, |
|
"learning_rate": 3e-05, |
|
"loss": 0.4572, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 57.77777777777778, |
|
"grad_norm": 1.4291303157806396, |
|
"learning_rate": 3.2500000000000004e-05, |
|
"loss": 0.452, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 62.22222222222222, |
|
"grad_norm": 1.3374826908111572, |
|
"learning_rate": 3.5e-05, |
|
"loss": 0.447, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 66.66666666666667, |
|
"grad_norm": 1.7066700458526611, |
|
"learning_rate": 3.7500000000000003e-05, |
|
"loss": 0.4424, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 71.11111111111111, |
|
"grad_norm": 2.1874194145202637, |
|
"learning_rate": 4e-05, |
|
"loss": 0.4443, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 75.55555555555556, |
|
"grad_norm": 1.4660208225250244, |
|
"learning_rate": 4.25e-05, |
|
"loss": 0.4346, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 80.0, |
|
"grad_norm": 1.2421706914901733, |
|
"learning_rate": 4.5e-05, |
|
"loss": 0.4328, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 84.44444444444444, |
|
"grad_norm": 0.8712289929389954, |
|
"learning_rate": 4.75e-05, |
|
"loss": 0.4335, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 88.88888888888889, |
|
"grad_norm": 1.161516547203064, |
|
"learning_rate": 5e-05, |
|
"loss": 0.4317, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 88.88888888888889, |
|
"eval_loss": 0.40811634063720703, |
|
"eval_runtime": 9.5158, |
|
"eval_samples_per_second": 16.709, |
|
"eval_steps_per_second": 2.102, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 93.33333333333333, |
|
"grad_norm": 4.811057090759277, |
|
"learning_rate": 5.25e-05, |
|
"loss": 0.4311, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 97.77777777777777, |
|
"grad_norm": 1.5318878889083862, |
|
"learning_rate": 5.500000000000001e-05, |
|
"loss": 0.4253, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 102.22222222222223, |
|
"grad_norm": 1.7489895820617676, |
|
"learning_rate": 5.7499999999999995e-05, |
|
"loss": 0.4229, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 106.66666666666667, |
|
"grad_norm": 1.162777304649353, |
|
"learning_rate": 6e-05, |
|
"loss": 0.4214, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 111.11111111111111, |
|
"grad_norm": 1.2216763496398926, |
|
"learning_rate": 6.25e-05, |
|
"loss": 0.4168, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 115.55555555555556, |
|
"grad_norm": 1.368053674697876, |
|
"learning_rate": 6.500000000000001e-05, |
|
"loss": 0.4197, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 120.0, |
|
"grad_norm": 1.3240705728530884, |
|
"learning_rate": 6.750000000000001e-05, |
|
"loss": 0.4143, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 124.44444444444444, |
|
"grad_norm": 3.4118428230285645, |
|
"learning_rate": 7e-05, |
|
"loss": 0.4159, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 128.88888888888889, |
|
"grad_norm": 2.0438427925109863, |
|
"learning_rate": 7.25e-05, |
|
"loss": 0.4135, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 133.33333333333334, |
|
"grad_norm": 3.8993186950683594, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 0.412, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 133.33333333333334, |
|
"eval_loss": 0.40509456396102905, |
|
"eval_runtime": 8.6764, |
|
"eval_samples_per_second": 18.326, |
|
"eval_steps_per_second": 2.305, |
|
"step": 1500 |
|
} |
|
], |
|
"logging_steps": 50, |
|
"max_steps": 4000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 364, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 4.122316318184064e+16, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|