|
{ |
|
"best_metric": 0.7326732673267327, |
|
"best_model_checkpoint": "distilhubert-finetuned-not-a-word2/run-3/checkpoint-8", |
|
"epoch": 10.0, |
|
"eval_steps": 500, |
|
"global_step": 80, |
|
"is_hyper_param_search": true, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.94398033618927, |
|
"learning_rate": 7.138113611405621e-06, |
|
"loss": 0.6981, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_f1": 0.7326732673267327, |
|
"eval_loss": 0.6864166259765625, |
|
"eval_runtime": 1.3249, |
|
"eval_samples_per_second": 48.305, |
|
"eval_steps_per_second": 6.038, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 1.5316386222839355, |
|
"learning_rate": 1.1103732284408743e-05, |
|
"loss": 0.683, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 0.5028849840164185, |
|
"learning_rate": 1.031060854980812e-05, |
|
"loss": 0.6564, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_f1": 0.7326732673267327, |
|
"eval_loss": 0.680694580078125, |
|
"eval_runtime": 1.3565, |
|
"eval_samples_per_second": 47.179, |
|
"eval_steps_per_second": 5.897, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 0.5775301456451416, |
|
"learning_rate": 9.517484815207495e-06, |
|
"loss": 0.6436, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_f1": 0.7326732673267327, |
|
"eval_loss": 0.6852684020996094, |
|
"eval_runtime": 1.3445, |
|
"eval_samples_per_second": 47.601, |
|
"eval_steps_per_second": 5.95, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"grad_norm": 0.4549338221549988, |
|
"learning_rate": 8.72436108060687e-06, |
|
"loss": 0.6098, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"grad_norm": 0.3741567134857178, |
|
"learning_rate": 7.931237346006246e-06, |
|
"loss": 0.609, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_f1": 0.7326732673267327, |
|
"eval_loss": 0.6952018737792969, |
|
"eval_runtime": 1.3401, |
|
"eval_samples_per_second": 47.757, |
|
"eval_steps_per_second": 5.97, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 4.38, |
|
"grad_norm": 0.38969138264656067, |
|
"learning_rate": 7.138113611405621e-06, |
|
"loss": 0.6091, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"grad_norm": 0.2588692009449005, |
|
"learning_rate": 6.344989876804997e-06, |
|
"loss": 0.6043, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_f1": 0.7326732673267327, |
|
"eval_loss": 0.7042198181152344, |
|
"eval_runtime": 1.3815, |
|
"eval_samples_per_second": 46.326, |
|
"eval_steps_per_second": 5.791, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 5.62, |
|
"grad_norm": 0.6705239415168762, |
|
"learning_rate": 5.551866142204372e-06, |
|
"loss": 0.6093, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_f1": 0.7326732673267327, |
|
"eval_loss": 0.7079658508300781, |
|
"eval_runtime": 1.3745, |
|
"eval_samples_per_second": 46.561, |
|
"eval_steps_per_second": 5.82, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"grad_norm": 0.4161934554576874, |
|
"learning_rate": 4.758742407603747e-06, |
|
"loss": 0.5737, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 6.88, |
|
"grad_norm": 0.3530200719833374, |
|
"learning_rate": 3.965618673003123e-06, |
|
"loss": 0.6355, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_f1": 0.7326732673267327, |
|
"eval_loss": 0.7088470458984375, |
|
"eval_runtime": 1.3736, |
|
"eval_samples_per_second": 46.592, |
|
"eval_steps_per_second": 5.824, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"grad_norm": 0.7337541580200195, |
|
"learning_rate": 3.1724949384024984e-06, |
|
"loss": 0.5738, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_f1": 0.7326732673267327, |
|
"eval_loss": 0.7090225219726562, |
|
"eval_runtime": 1.3562, |
|
"eval_samples_per_second": 47.192, |
|
"eval_steps_per_second": 5.899, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 8.12, |
|
"grad_norm": 0.3913467526435852, |
|
"learning_rate": 2.3793712038018737e-06, |
|
"loss": 0.5872, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"grad_norm": 0.6211589574813843, |
|
"learning_rate": 1.5862474692012492e-06, |
|
"loss": 0.5549, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_f1": 0.7326732673267327, |
|
"eval_loss": 0.7093772888183594, |
|
"eval_runtime": 1.3631, |
|
"eval_samples_per_second": 46.952, |
|
"eval_steps_per_second": 5.869, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 9.38, |
|
"grad_norm": 0.30739858746528625, |
|
"learning_rate": 7.931237346006246e-07, |
|
"loss": 0.6143, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"grad_norm": 0.41105917096138, |
|
"learning_rate": 0.0, |
|
"loss": 0.6183, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_f1": 0.7326732673267327, |
|
"eval_loss": 0.7090110778808594, |
|
"eval_runtime": 1.366, |
|
"eval_samples_per_second": 46.852, |
|
"eval_steps_per_second": 5.857, |
|
"step": 80 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 80, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 500, |
|
"total_flos": 1.0052017938770688e+16, |
|
"train_batch_size": 24, |
|
"trial_name": null, |
|
"trial_params": { |
|
"learning_rate": 1.1420981778248994e-05, |
|
"per_device_train_batch_size": 24 |
|
} |
|
} |
|
|