|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 5.590496156533892, |
|
"global_step": 8000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 87.2627, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 52.4123, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 6e-06, |
|
"loss": 41.3932, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 36.7185, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1e-05, |
|
"loss": 33.4597, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.2e-05, |
|
"loss": 31.1323, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.4000000000000001e-05, |
|
"loss": 28.9204, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 27.0128, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.8e-05, |
|
"loss": 25.1703, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2e-05, |
|
"loss": 23.5486, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.2000000000000003e-05, |
|
"loss": 21.8419, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 2.4e-05, |
|
"loss": 20.3387, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.6000000000000002e-05, |
|
"loss": 18.7216, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.8000000000000003e-05, |
|
"loss": 16.7862, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 3e-05, |
|
"loss": 15.034, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 3.2000000000000005e-05, |
|
"loss": 13.3747, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 3.4000000000000007e-05, |
|
"loss": 11.7581, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 3.6e-05, |
|
"loss": 9.9702, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 3.8e-05, |
|
"loss": 8.3363, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 4e-05, |
|
"loss": 6.9199, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"eval_bleu": 3.4801, |
|
"eval_em": 0.0, |
|
"eval_gen_len": 220.966, |
|
"eval_loss": 6.3751444816589355, |
|
"eval_runtime": 1604.7901, |
|
"eval_samples_per_second": 0.751, |
|
"eval_steps_per_second": 0.094, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 4.2e-05, |
|
"loss": 6.2524, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 4.4000000000000006e-05, |
|
"loss": 5.8029, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 4.600000000000001e-05, |
|
"loss": 5.606, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.8e-05, |
|
"loss": 5.3757, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 5e-05, |
|
"loss": 5.1043, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 4.944121591417077e-05, |
|
"loss": 4.9414, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 4.888243182834153e-05, |
|
"loss": 4.7381, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 4.8323647742512295e-05, |
|
"loss": 4.6214, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 4.776486365668306e-05, |
|
"loss": 4.2971, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 4.720607957085382e-05, |
|
"loss": 4.1602, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 4.664729548502459e-05, |
|
"loss": 4.0391, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 4.6088511399195353e-05, |
|
"loss": 3.9211, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 4.552972731336611e-05, |
|
"loss": 3.7642, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 4.497094322753688e-05, |
|
"loss": 3.6698, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 4.4412159141707646e-05, |
|
"loss": 3.5409, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 4.385337505587841e-05, |
|
"loss": 3.4016, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 4.329459097004918e-05, |
|
"loss": 3.2761, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 4.2735806884219945e-05, |
|
"loss": 3.1708, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 4.2177022798390704e-05, |
|
"loss": 3.0849, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 4.161823871256147e-05, |
|
"loss": 3.0222, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"eval_bleu": 27.8543, |
|
"eval_em": 0.0, |
|
"eval_gen_len": 36.8, |
|
"eval_loss": 2.8796441555023193, |
|
"eval_runtime": 501.7382, |
|
"eval_samples_per_second": 2.402, |
|
"eval_steps_per_second": 0.301, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 4.105945462673223e-05, |
|
"loss": 2.9079, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 4.0500670540903e-05, |
|
"loss": 2.8688, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 3.994188645507376e-05, |
|
"loss": 2.7398, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 3.938310236924452e-05, |
|
"loss": 2.5503, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 3.882431828341529e-05, |
|
"loss": 2.5121, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 3.8265534197586055e-05, |
|
"loss": 2.4625, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 3.7706750111756815e-05, |
|
"loss": 2.3833, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 3.714796602592758e-05, |
|
"loss": 2.3133, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 3.658918194009835e-05, |
|
"loss": 2.2152, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"learning_rate": 3.603039785426911e-05, |
|
"loss": 2.2304, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 3.5471613768439874e-05, |
|
"loss": 2.1563, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"learning_rate": 3.491282968261064e-05, |
|
"loss": 2.0991, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"learning_rate": 3.4354045596781406e-05, |
|
"loss": 2.0702, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 3.379526151095217e-05, |
|
"loss": 1.9914, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"learning_rate": 3.323647742512294e-05, |
|
"loss": 1.9235, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 3.26776933392937e-05, |
|
"loss": 1.8922, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 3.98, |
|
"learning_rate": 3.2118909253464465e-05, |
|
"loss": 1.8512, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 4.05, |
|
"learning_rate": 3.156012516763523e-05, |
|
"loss": 1.6966, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 4.12, |
|
"learning_rate": 3.100134108180599e-05, |
|
"loss": 1.6399, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"learning_rate": 3.0442556995976757e-05, |
|
"loss": 1.5982, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"eval_bleu": 56.0747, |
|
"eval_em": 0.0017, |
|
"eval_gen_len": 43.9021, |
|
"eval_loss": 1.7495189905166626, |
|
"eval_runtime": 423.2589, |
|
"eval_samples_per_second": 2.847, |
|
"eval_steps_per_second": 0.357, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 4.26, |
|
"learning_rate": 2.9883772910147524e-05, |
|
"loss": 1.5968, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"learning_rate": 2.9324988824318283e-05, |
|
"loss": 1.5813, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 4.4, |
|
"learning_rate": 2.876620473848905e-05, |
|
"loss": 1.5421, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 4.47, |
|
"learning_rate": 2.8207420652659816e-05, |
|
"loss": 1.4852, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 4.54, |
|
"learning_rate": 2.7648636566830576e-05, |
|
"loss": 1.5027, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 4.61, |
|
"learning_rate": 2.7089852481001342e-05, |
|
"loss": 1.4951, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 4.68, |
|
"learning_rate": 2.653106839517211e-05, |
|
"loss": 1.4803, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"learning_rate": 2.597228430934287e-05, |
|
"loss": 1.4127, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 4.82, |
|
"learning_rate": 2.5413500223513638e-05, |
|
"loss": 1.3896, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 2.48547161376844e-05, |
|
"loss": 1.3656, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 4.96, |
|
"learning_rate": 2.4295932051855164e-05, |
|
"loss": 1.3432, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 2.373714796602593e-05, |
|
"loss": 1.2224, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 5.1, |
|
"learning_rate": 2.3178363880196693e-05, |
|
"loss": 1.1396, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 5.17, |
|
"learning_rate": 2.2619579794367456e-05, |
|
"loss": 1.1475, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 5.24, |
|
"learning_rate": 2.206079570853822e-05, |
|
"loss": 1.0669, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 5.31, |
|
"learning_rate": 2.1502011622708985e-05, |
|
"loss": 1.1356, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 5.38, |
|
"learning_rate": 2.0943227536879752e-05, |
|
"loss": 1.0965, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 5.45, |
|
"learning_rate": 2.0384443451050515e-05, |
|
"loss": 1.1086, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 5.52, |
|
"learning_rate": 1.982565936522128e-05, |
|
"loss": 1.0642, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 5.59, |
|
"learning_rate": 1.9266875279392044e-05, |
|
"loss": 1.0717, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 5.59, |
|
"eval_bleu": 69.9606, |
|
"eval_em": 0.0199, |
|
"eval_gen_len": 46.0722, |
|
"eval_loss": 1.26251220703125, |
|
"eval_runtime": 355.6397, |
|
"eval_samples_per_second": 3.388, |
|
"eval_steps_per_second": 0.425, |
|
"step": 8000 |
|
} |
|
], |
|
"max_steps": 11448, |
|
"num_train_epochs": 8, |
|
"total_flos": 5257694441864832.0, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|