|
{ |
|
"best_metric": 0.7326732673267327, |
|
"best_model_checkpoint": "distilhubert-finetuned-not-a-word2/run-4/checkpoint-96", |
|
"epoch": 9.0, |
|
"eval_steps": 500, |
|
"global_step": 432, |
|
"is_hyper_param_search": true, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 1.2733114957809448, |
|
"learning_rate": 2.2702186710865246e-07, |
|
"loss": 0.7025, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 1.243804931640625, |
|
"learning_rate": 4.5404373421730493e-07, |
|
"loss": 0.6974, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 1.7711552381515503, |
|
"learning_rate": 6.810656013259573e-07, |
|
"loss": 0.696, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 1.1453403234481812, |
|
"learning_rate": 9.080874684346099e-07, |
|
"loss": 0.6989, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 1.2729355096817017, |
|
"learning_rate": 1.1351093355432624e-06, |
|
"loss": 0.6968, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 1.1592165231704712, |
|
"learning_rate": 1.3621312026519146e-06, |
|
"loss": 0.6959, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 1.1798148155212402, |
|
"learning_rate": 1.589153069760567e-06, |
|
"loss": 0.6952, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 2.1216671466827393, |
|
"learning_rate": 1.8161749368692197e-06, |
|
"loss": 0.6886, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 1.3416370153427124, |
|
"learning_rate": 2.043196803977872e-06, |
|
"loss": 0.6864, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_f1": 0.72, |
|
"eval_loss": 0.688262939453125, |
|
"eval_runtime": 1.3468, |
|
"eval_samples_per_second": 47.521, |
|
"eval_steps_per_second": 5.94, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 2.1856281757354736, |
|
"learning_rate": 2.169320063482679e-06, |
|
"loss": 0.6917, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 1.4077153205871582, |
|
"learning_rate": 2.1440954115817176e-06, |
|
"loss": 0.6884, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 2.1792664527893066, |
|
"learning_rate": 2.1188707596807562e-06, |
|
"loss": 0.6668, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 1.0386197566986084, |
|
"learning_rate": 2.093646107779795e-06, |
|
"loss": 0.6694, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 2.0565919876098633, |
|
"learning_rate": 2.0684214558788335e-06, |
|
"loss": 0.6561, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 1.2978509664535522, |
|
"learning_rate": 2.043196803977872e-06, |
|
"loss": 0.6789, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"grad_norm": 2.058328628540039, |
|
"learning_rate": 2.0179721520769108e-06, |
|
"loss": 0.6633, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 0.6023226976394653, |
|
"learning_rate": 1.9927475001759494e-06, |
|
"loss": 0.6655, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 0.5510762929916382, |
|
"learning_rate": 1.967522848274988e-06, |
|
"loss": 0.6622, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 1.098602533340454, |
|
"learning_rate": 1.9422981963740267e-06, |
|
"loss": 0.6633, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_f1": 0.7326732673267327, |
|
"eval_loss": 0.6816024780273438, |
|
"eval_runtime": 1.3765, |
|
"eval_samples_per_second": 46.493, |
|
"eval_steps_per_second": 5.812, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 0.9589098691940308, |
|
"learning_rate": 1.9170735444730654e-06, |
|
"loss": 0.659, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"grad_norm": 1.070695161819458, |
|
"learning_rate": 1.8918488925721038e-06, |
|
"loss": 0.6313, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"grad_norm": 0.9913639426231384, |
|
"learning_rate": 1.8666242406711424e-06, |
|
"loss": 0.6652, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 1.0632878541946411, |
|
"learning_rate": 1.841399588770181e-06, |
|
"loss": 0.673, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 2.1036579608917236, |
|
"learning_rate": 1.8161749368692197e-06, |
|
"loss": 0.6451, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"grad_norm": 1.08384108543396, |
|
"learning_rate": 1.7909502849682583e-06, |
|
"loss": 0.6322, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"grad_norm": 0.9407000541687012, |
|
"learning_rate": 1.765725633067297e-06, |
|
"loss": 0.6755, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"grad_norm": 0.9016568660736084, |
|
"learning_rate": 1.7405009811663356e-06, |
|
"loss": 0.5985, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"grad_norm": 1.1134448051452637, |
|
"learning_rate": 1.7152763292653743e-06, |
|
"loss": 0.603, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_f1": 0.7326732673267327, |
|
"eval_loss": 0.6800689697265625, |
|
"eval_runtime": 1.3861, |
|
"eval_samples_per_second": 46.173, |
|
"eval_steps_per_second": 5.772, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"grad_norm": 0.7627719640731812, |
|
"learning_rate": 1.6900516773644127e-06, |
|
"loss": 0.6557, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"grad_norm": 0.9291415214538574, |
|
"learning_rate": 1.6648270254634511e-06, |
|
"loss": 0.6219, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"grad_norm": 0.9248765707015991, |
|
"learning_rate": 1.6396023735624898e-06, |
|
"loss": 0.6325, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"grad_norm": 0.9842573404312134, |
|
"learning_rate": 1.6143777216615284e-06, |
|
"loss": 0.6521, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"grad_norm": 0.8689214587211609, |
|
"learning_rate": 1.589153069760567e-06, |
|
"loss": 0.5929, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"grad_norm": 1.0012000799179077, |
|
"learning_rate": 1.5639284178596057e-06, |
|
"loss": 0.584, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 3.65, |
|
"grad_norm": 0.7438368797302246, |
|
"learning_rate": 1.5387037659586443e-06, |
|
"loss": 0.6813, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"grad_norm": 1.8603870868682861, |
|
"learning_rate": 1.513479114057683e-06, |
|
"loss": 0.6099, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"grad_norm": 0.9918416738510132, |
|
"learning_rate": 1.4882544621567216e-06, |
|
"loss": 0.6192, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"grad_norm": 1.9146322011947632, |
|
"learning_rate": 1.4630298102557603e-06, |
|
"loss": 0.6472, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_f1": 0.7326732673267327, |
|
"eval_loss": 0.6818161010742188, |
|
"eval_runtime": 1.3841, |
|
"eval_samples_per_second": 46.239, |
|
"eval_steps_per_second": 5.78, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 4.06, |
|
"grad_norm": 0.9502781629562378, |
|
"learning_rate": 1.437805158354799e-06, |
|
"loss": 0.6447, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 4.17, |
|
"grad_norm": 0.8570067286491394, |
|
"learning_rate": 1.4125805064538375e-06, |
|
"loss": 0.5306, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 4.27, |
|
"grad_norm": 0.8097484111785889, |
|
"learning_rate": 1.3873558545528762e-06, |
|
"loss": 0.6202, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 4.38, |
|
"grad_norm": 2.0106472969055176, |
|
"learning_rate": 1.3621312026519146e-06, |
|
"loss": 0.6705, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 4.48, |
|
"grad_norm": 1.090775489807129, |
|
"learning_rate": 1.3369065507509533e-06, |
|
"loss": 0.6297, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 4.58, |
|
"grad_norm": 0.8988145589828491, |
|
"learning_rate": 1.311681898849992e-06, |
|
"loss": 0.5896, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 4.69, |
|
"grad_norm": 0.9149978756904602, |
|
"learning_rate": 1.2864572469490305e-06, |
|
"loss": 0.6156, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 4.79, |
|
"grad_norm": 1.9398412704467773, |
|
"learning_rate": 1.2612325950480692e-06, |
|
"loss": 0.6305, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 4.9, |
|
"grad_norm": 0.9217966794967651, |
|
"learning_rate": 1.2360079431471078e-06, |
|
"loss": 0.5943, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"grad_norm": 0.9083653688430786, |
|
"learning_rate": 1.2107832912461465e-06, |
|
"loss": 0.6386, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_f1": 0.7326732673267327, |
|
"eval_loss": 0.6846389770507812, |
|
"eval_runtime": 1.4094, |
|
"eval_samples_per_second": 45.409, |
|
"eval_steps_per_second": 5.676, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 5.1, |
|
"grad_norm": 0.9323675036430359, |
|
"learning_rate": 1.1855586393451851e-06, |
|
"loss": 0.5779, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 5.21, |
|
"grad_norm": 0.7549787163734436, |
|
"learning_rate": 1.1603339874442238e-06, |
|
"loss": 0.5948, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 5.31, |
|
"grad_norm": 0.8535837531089783, |
|
"learning_rate": 1.1351093355432624e-06, |
|
"loss": 0.6928, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 5.42, |
|
"grad_norm": 1.2038137912750244, |
|
"learning_rate": 1.109884683642301e-06, |
|
"loss": 0.5887, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 5.52, |
|
"grad_norm": 0.9501279592514038, |
|
"learning_rate": 1.0846600317413395e-06, |
|
"loss": 0.5776, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 5.62, |
|
"grad_norm": 0.7421719431877136, |
|
"learning_rate": 1.0594353798403781e-06, |
|
"loss": 0.6734, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 5.73, |
|
"grad_norm": 0.8555863499641418, |
|
"learning_rate": 1.0342107279394168e-06, |
|
"loss": 0.6399, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 5.83, |
|
"grad_norm": 0.8841156363487244, |
|
"learning_rate": 1.0089860760384554e-06, |
|
"loss": 0.6173, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 5.94, |
|
"grad_norm": 0.8565478324890137, |
|
"learning_rate": 9.83761424137494e-07, |
|
"loss": 0.5537, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_f1": 0.7326732673267327, |
|
"eval_loss": 0.6864242553710938, |
|
"eval_runtime": 1.3622, |
|
"eval_samples_per_second": 46.983, |
|
"eval_steps_per_second": 5.873, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"grad_norm": 0.8750139474868774, |
|
"learning_rate": 9.585367722365327e-07, |
|
"loss": 0.5531, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 6.15, |
|
"grad_norm": 1.0445302724838257, |
|
"learning_rate": 9.333121203355712e-07, |
|
"loss": 0.638, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"grad_norm": 0.7958914637565613, |
|
"learning_rate": 9.080874684346099e-07, |
|
"loss": 0.547, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 6.35, |
|
"grad_norm": 0.9992254376411438, |
|
"learning_rate": 8.828628165336485e-07, |
|
"loss": 0.6425, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 6.46, |
|
"grad_norm": 0.8400682806968689, |
|
"learning_rate": 8.576381646326871e-07, |
|
"loss": 0.6955, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 6.56, |
|
"grad_norm": 0.742438793182373, |
|
"learning_rate": 8.324135127317256e-07, |
|
"loss": 0.6473, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 6.67, |
|
"grad_norm": 0.6693254113197327, |
|
"learning_rate": 8.071888608307642e-07, |
|
"loss": 0.603, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 6.77, |
|
"grad_norm": 1.0816401243209839, |
|
"learning_rate": 7.819642089298028e-07, |
|
"loss": 0.6053, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 6.88, |
|
"grad_norm": 0.7275277376174927, |
|
"learning_rate": 7.567395570288415e-07, |
|
"loss": 0.612, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 6.98, |
|
"grad_norm": 0.7834873795509338, |
|
"learning_rate": 7.315149051278801e-07, |
|
"loss": 0.55, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_f1": 0.7326732673267327, |
|
"eval_loss": 0.6889228820800781, |
|
"eval_runtime": 1.3769, |
|
"eval_samples_per_second": 46.483, |
|
"eval_steps_per_second": 5.81, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 7.08, |
|
"grad_norm": 1.24147367477417, |
|
"learning_rate": 7.062902532269188e-07, |
|
"loss": 0.508, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 7.19, |
|
"grad_norm": 1.8932181596755981, |
|
"learning_rate": 6.810656013259573e-07, |
|
"loss": 0.6358, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 7.29, |
|
"grad_norm": 1.861436128616333, |
|
"learning_rate": 6.55840949424996e-07, |
|
"loss": 0.5741, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 7.4, |
|
"grad_norm": 0.8429200053215027, |
|
"learning_rate": 6.306162975240346e-07, |
|
"loss": 0.5717, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"grad_norm": 1.8665741682052612, |
|
"learning_rate": 6.053916456230732e-07, |
|
"loss": 0.6992, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 7.6, |
|
"grad_norm": 2.312748908996582, |
|
"learning_rate": 5.801669937221119e-07, |
|
"loss": 0.6151, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 7.71, |
|
"grad_norm": 1.1628329753875732, |
|
"learning_rate": 5.549423418211505e-07, |
|
"loss": 0.5354, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 7.81, |
|
"grad_norm": 1.8674992322921753, |
|
"learning_rate": 5.297176899201891e-07, |
|
"loss": 0.6411, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 7.92, |
|
"grad_norm": 0.7112137675285339, |
|
"learning_rate": 5.044930380192277e-07, |
|
"loss": 0.6063, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_f1": 0.7326732673267327, |
|
"eval_loss": 0.6909217834472656, |
|
"eval_runtime": 1.3913, |
|
"eval_samples_per_second": 45.999, |
|
"eval_steps_per_second": 5.75, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"grad_norm": 1.201416015625, |
|
"learning_rate": 4.792683861182663e-07, |
|
"loss": 0.647, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 8.12, |
|
"grad_norm": 0.9348795413970947, |
|
"learning_rate": 4.5404373421730493e-07, |
|
"loss": 0.5356, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 8.23, |
|
"grad_norm": 1.895739197731018, |
|
"learning_rate": 4.2881908231634357e-07, |
|
"loss": 0.5384, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 8.33, |
|
"grad_norm": 1.8363467454910278, |
|
"learning_rate": 4.035944304153821e-07, |
|
"loss": 0.541, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 8.44, |
|
"grad_norm": 2.475804090499878, |
|
"learning_rate": 3.7836977851442075e-07, |
|
"loss": 0.5503, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 8.54, |
|
"grad_norm": 1.314663290977478, |
|
"learning_rate": 3.531451266134594e-07, |
|
"loss": 0.575, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 8.65, |
|
"grad_norm": 1.850918173789978, |
|
"learning_rate": 3.27920474712498e-07, |
|
"loss": 0.6658, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"grad_norm": 0.7412477135658264, |
|
"learning_rate": 3.026958228115366e-07, |
|
"loss": 0.5632, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 8.85, |
|
"grad_norm": 1.3282320499420166, |
|
"learning_rate": 2.7747117091057526e-07, |
|
"loss": 0.7303, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 8.96, |
|
"grad_norm": 1.019906759262085, |
|
"learning_rate": 2.5224651900961385e-07, |
|
"loss": 0.7438, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_f1": 0.7326732673267327, |
|
"eval_loss": 0.6923065185546875, |
|
"eval_runtime": 1.3679, |
|
"eval_samples_per_second": 46.786, |
|
"eval_steps_per_second": 5.848, |
|
"step": 432 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 480, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 500, |
|
"total_flos": 6549027725117760.0, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": { |
|
"learning_rate": 2.1794099242430636e-06, |
|
"per_device_train_batch_size": 4 |
|
} |
|
} |
|
|