|
{ |
|
"best_metric": 0.8679605722427368, |
|
"best_model_checkpoint": "./checkpoints/unifiedqa-t5-v1-large/checkpoint-16167", |
|
"epoch": 17.0, |
|
"global_step": 16167, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.953125e-06, |
|
"loss": 14.5454, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 3.90625e-06, |
|
"loss": 8.6384, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 4.976226496650098e-06, |
|
"loss": 2.2394, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 4.9221958072185e-06, |
|
"loss": 1.6256, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 1.028218388557434, |
|
"eval_runtime": 5.4191, |
|
"eval_samples_per_second": 92.266, |
|
"eval_steps_per_second": 2.953, |
|
"step": 951 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 4.868165117786904e-06, |
|
"loss": 1.4601, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 4.814134428355306e-06, |
|
"loss": 1.3551, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 4.760103738923709e-06, |
|
"loss": 1.3451, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.706073049492112e-06, |
|
"loss": 1.3367, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 4.652042360060515e-06, |
|
"loss": 1.3039, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.9447770714759827, |
|
"eval_runtime": 5.412, |
|
"eval_samples_per_second": 92.387, |
|
"eval_steps_per_second": 2.956, |
|
"step": 1902 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 4.598011670628918e-06, |
|
"loss": 1.2325, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 4.543980981197321e-06, |
|
"loss": 1.2314, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 4.489950291765723e-06, |
|
"loss": 1.26, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 4.4359196023341265e-06, |
|
"loss": 1.2501, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 4.381888912902529e-06, |
|
"loss": 1.2664, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.9094200730323792, |
|
"eval_runtime": 5.4314, |
|
"eval_samples_per_second": 92.058, |
|
"eval_steps_per_second": 2.946, |
|
"step": 2853 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 4.327858223470932e-06, |
|
"loss": 1.2451, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 4.273827534039335e-06, |
|
"loss": 1.1615, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"learning_rate": 4.2197968446077375e-06, |
|
"loss": 1.1887, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"learning_rate": 4.165766155176141e-06, |
|
"loss": 1.2022, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 4.111735465744543e-06, |
|
"loss": 1.2318, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 0.8889561891555786, |
|
"eval_runtime": 5.4218, |
|
"eval_samples_per_second": 92.219, |
|
"eval_steps_per_second": 2.951, |
|
"step": 3804 |
|
}, |
|
{ |
|
"epoch": 4.21, |
|
"learning_rate": 4.057704776312946e-06, |
|
"loss": 1.2317, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 4.42, |
|
"learning_rate": 4.003674086881349e-06, |
|
"loss": 1.1517, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 4.63, |
|
"learning_rate": 3.949643397449752e-06, |
|
"loss": 1.1888, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 4.84, |
|
"learning_rate": 3.895612708018154e-06, |
|
"loss": 1.1919, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 0.8820498585700989, |
|
"eval_runtime": 5.3661, |
|
"eval_samples_per_second": 93.178, |
|
"eval_steps_per_second": 2.982, |
|
"step": 4755 |
|
}, |
|
{ |
|
"epoch": 5.05, |
|
"learning_rate": 3.841582018586558e-06, |
|
"loss": 1.1792, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 5.26, |
|
"learning_rate": 3.7875513291549603e-06, |
|
"loss": 1.1444, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 5.47, |
|
"learning_rate": 3.733520639723363e-06, |
|
"loss": 1.1541, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 5.68, |
|
"learning_rate": 3.679489950291766e-06, |
|
"loss": 1.1781, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 5.89, |
|
"learning_rate": 3.625459260860169e-06, |
|
"loss": 1.1604, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 0.8776991963386536, |
|
"eval_runtime": 5.4089, |
|
"eval_samples_per_second": 92.441, |
|
"eval_steps_per_second": 2.958, |
|
"step": 5706 |
|
}, |
|
{ |
|
"epoch": 6.1, |
|
"learning_rate": 3.5714285714285718e-06, |
|
"loss": 1.1692, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 6.31, |
|
"learning_rate": 3.5173978819969745e-06, |
|
"loss": 1.1379, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 6.52, |
|
"learning_rate": 3.4633671925653773e-06, |
|
"loss": 1.1367, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 6.73, |
|
"learning_rate": 3.4093365031337805e-06, |
|
"loss": 1.1134, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 6.94, |
|
"learning_rate": 3.3553058137021832e-06, |
|
"loss": 1.1304, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 0.8747878074645996, |
|
"eval_runtime": 5.3628, |
|
"eval_samples_per_second": 93.235, |
|
"eval_steps_per_second": 2.984, |
|
"step": 6657 |
|
}, |
|
{ |
|
"epoch": 7.15, |
|
"learning_rate": 3.301275124270586e-06, |
|
"loss": 1.1245, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 7.36, |
|
"learning_rate": 3.2472444348389887e-06, |
|
"loss": 1.1174, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 7.57, |
|
"learning_rate": 3.193213745407392e-06, |
|
"loss": 1.0833, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 7.78, |
|
"learning_rate": 3.1391830559757947e-06, |
|
"loss": 1.1269, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 7.99, |
|
"learning_rate": 3.0851523665441974e-06, |
|
"loss": 1.1597, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 0.8729026913642883, |
|
"eval_runtime": 5.3708, |
|
"eval_samples_per_second": 93.096, |
|
"eval_steps_per_second": 2.979, |
|
"step": 7608 |
|
}, |
|
{ |
|
"epoch": 8.2, |
|
"learning_rate": 3.0311216771126e-06, |
|
"loss": 1.1167, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 8.41, |
|
"learning_rate": 2.977090987681003e-06, |
|
"loss": 1.0435, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 8.62, |
|
"learning_rate": 2.923060298249406e-06, |
|
"loss": 1.1781, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 8.83, |
|
"learning_rate": 2.869029608817809e-06, |
|
"loss": 1.1124, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 0.8715777397155762, |
|
"eval_runtime": 5.3225, |
|
"eval_samples_per_second": 93.941, |
|
"eval_steps_per_second": 3.006, |
|
"step": 8559 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 2.8149989193862116e-06, |
|
"loss": 1.1026, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 9.25, |
|
"learning_rate": 2.7609682299546144e-06, |
|
"loss": 1.0928, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 9.46, |
|
"learning_rate": 2.7069375405230175e-06, |
|
"loss": 1.0899, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 9.67, |
|
"learning_rate": 2.6529068510914203e-06, |
|
"loss": 1.0554, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 9.88, |
|
"learning_rate": 2.598876161659823e-06, |
|
"loss": 1.1124, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 0.8696691989898682, |
|
"eval_runtime": 5.3401, |
|
"eval_samples_per_second": 93.632, |
|
"eval_steps_per_second": 2.996, |
|
"step": 9510 |
|
}, |
|
{ |
|
"epoch": 10.09, |
|
"learning_rate": 2.5448454722282258e-06, |
|
"loss": 1.078, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 10.3, |
|
"learning_rate": 2.4908147827966285e-06, |
|
"loss": 1.0773, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 10.52, |
|
"learning_rate": 2.4367840933650317e-06, |
|
"loss": 1.1118, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 10.73, |
|
"learning_rate": 2.3827534039334345e-06, |
|
"loss": 1.0603, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 10.94, |
|
"learning_rate": 2.3287227145018372e-06, |
|
"loss": 1.0923, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_loss": 0.8697531819343567, |
|
"eval_runtime": 5.3465, |
|
"eval_samples_per_second": 93.519, |
|
"eval_steps_per_second": 2.993, |
|
"step": 10461 |
|
}, |
|
{ |
|
"epoch": 11.15, |
|
"learning_rate": 2.27469202507024e-06, |
|
"loss": 1.0872, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 11.36, |
|
"learning_rate": 2.220661335638643e-06, |
|
"loss": 1.0683, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 11.57, |
|
"learning_rate": 2.166630646207046e-06, |
|
"loss": 1.0708, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 11.78, |
|
"learning_rate": 2.1125999567754487e-06, |
|
"loss": 1.0625, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 11.99, |
|
"learning_rate": 2.0585692673438514e-06, |
|
"loss": 1.0732, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_loss": 0.8692195415496826, |
|
"eval_runtime": 5.3316, |
|
"eval_samples_per_second": 93.78, |
|
"eval_steps_per_second": 3.001, |
|
"step": 11412 |
|
}, |
|
{ |
|
"epoch": 12.2, |
|
"learning_rate": 2.004538577912254e-06, |
|
"loss": 1.0802, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 12.41, |
|
"learning_rate": 1.9505078884806573e-06, |
|
"loss": 1.0365, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 12.62, |
|
"learning_rate": 1.8964771990490599e-06, |
|
"loss": 1.0738, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 12.83, |
|
"learning_rate": 1.8424465096174629e-06, |
|
"loss": 1.084, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_loss": 0.8688466548919678, |
|
"eval_runtime": 5.2515, |
|
"eval_samples_per_second": 95.211, |
|
"eval_steps_per_second": 3.047, |
|
"step": 12363 |
|
}, |
|
{ |
|
"epoch": 13.04, |
|
"learning_rate": 1.7884158201858656e-06, |
|
"loss": 1.0532, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 13.25, |
|
"learning_rate": 1.7343851307542686e-06, |
|
"loss": 1.0409, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 13.46, |
|
"learning_rate": 1.6803544413226713e-06, |
|
"loss": 1.0711, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 13.67, |
|
"learning_rate": 1.6263237518910743e-06, |
|
"loss": 1.02, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 13.88, |
|
"learning_rate": 1.572293062459477e-06, |
|
"loss": 1.0559, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_loss": 0.8682753443717957, |
|
"eval_runtime": 5.2361, |
|
"eval_samples_per_second": 95.491, |
|
"eval_steps_per_second": 3.056, |
|
"step": 13314 |
|
}, |
|
{ |
|
"epoch": 14.09, |
|
"learning_rate": 1.51826237302788e-06, |
|
"loss": 1.0679, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 14.3, |
|
"learning_rate": 1.4642316835962828e-06, |
|
"loss": 1.0495, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 14.51, |
|
"learning_rate": 1.4102009941646857e-06, |
|
"loss": 1.0302, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 14.72, |
|
"learning_rate": 1.3561703047330885e-06, |
|
"loss": 1.0653, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 14.93, |
|
"learning_rate": 1.3021396153014914e-06, |
|
"loss": 1.054, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_loss": 0.8680118918418884, |
|
"eval_runtime": 5.2347, |
|
"eval_samples_per_second": 95.516, |
|
"eval_steps_per_second": 3.057, |
|
"step": 14265 |
|
}, |
|
{ |
|
"epoch": 15.14, |
|
"learning_rate": 1.2481089258698942e-06, |
|
"loss": 1.0485, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 15.35, |
|
"learning_rate": 1.194078236438297e-06, |
|
"loss": 1.0391, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 15.56, |
|
"learning_rate": 1.1400475470067e-06, |
|
"loss": 1.0549, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 15.77, |
|
"learning_rate": 1.0860168575751027e-06, |
|
"loss": 1.0278, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 15.98, |
|
"learning_rate": 1.0319861681435056e-06, |
|
"loss": 1.0478, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_loss": 0.8681474328041077, |
|
"eval_runtime": 5.2378, |
|
"eval_samples_per_second": 95.459, |
|
"eval_steps_per_second": 3.055, |
|
"step": 15216 |
|
}, |
|
{ |
|
"epoch": 16.19, |
|
"learning_rate": 9.779554787119084e-07, |
|
"loss": 1.026, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 16.4, |
|
"learning_rate": 9.239247892803114e-07, |
|
"loss": 1.0371, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 16.61, |
|
"learning_rate": 8.698940998487142e-07, |
|
"loss": 1.0391, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 16.82, |
|
"learning_rate": 8.158634104171171e-07, |
|
"loss": 1.0406, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_loss": 0.8679605722427368, |
|
"eval_runtime": 5.2538, |
|
"eval_samples_per_second": 95.169, |
|
"eval_steps_per_second": 3.045, |
|
"step": 16167 |
|
} |
|
], |
|
"max_steps": 19020, |
|
"num_train_epochs": 20, |
|
"total_flos": 2.8000725823488e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|