|
{ |
|
"best_metric": 0.7130943536758423, |
|
"best_model_checkpoint": "/media/mldrive/kcardenas/limb_classification_person_crop/beit-large-patch16-384/8_5e-5_1e-4_0.05/checkpoint-1712", |
|
"epoch": 10.0, |
|
"eval_steps": 500, |
|
"global_step": 2140, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.11682242990654206, |
|
"grad_norm": 1340459.625, |
|
"learning_rate": 2.5e-06, |
|
"loss": 1.5852, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.2336448598130841, |
|
"grad_norm": 994590.0, |
|
"learning_rate": 5e-06, |
|
"loss": 1.5062, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.35046728971962615, |
|
"grad_norm": 835627.5, |
|
"learning_rate": 7.5e-06, |
|
"loss": 1.4702, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.4672897196261682, |
|
"grad_norm": 708100.8125, |
|
"learning_rate": 1e-05, |
|
"loss": 1.405, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.5841121495327103, |
|
"grad_norm": 19124280.0, |
|
"learning_rate": 1.25e-05, |
|
"loss": 1.3983, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.7009345794392523, |
|
"grad_norm": 281499.28125, |
|
"learning_rate": 1.5e-05, |
|
"loss": 1.3889, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.8177570093457944, |
|
"grad_norm": 727612.8125, |
|
"learning_rate": 1.75e-05, |
|
"loss": 1.3625, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.9345794392523364, |
|
"grad_norm": 628275.8125, |
|
"learning_rate": 2e-05, |
|
"loss": 1.3649, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.36401326699834163, |
|
"eval_loss": 1.5214860439300537, |
|
"eval_runtime": 15.5708, |
|
"eval_samples_per_second": 77.453, |
|
"eval_steps_per_second": 2.44, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 1.0514018691588785, |
|
"grad_norm": 456936.21875, |
|
"learning_rate": 2.25e-05, |
|
"loss": 1.3042, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 1.1682242990654206, |
|
"grad_norm": 314467.59375, |
|
"learning_rate": 2.5e-05, |
|
"loss": 1.275, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.2850467289719627, |
|
"grad_norm": 409308.0625, |
|
"learning_rate": 2.7500000000000004e-05, |
|
"loss": 1.2176, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 1.4018691588785046, |
|
"grad_norm": 343987.34375, |
|
"learning_rate": 3e-05, |
|
"loss": 1.2691, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.5186915887850467, |
|
"grad_norm": 415037.3125, |
|
"learning_rate": 3.2500000000000004e-05, |
|
"loss": 1.2335, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 1.6355140186915889, |
|
"grad_norm": 369134.59375, |
|
"learning_rate": 3.5e-05, |
|
"loss": 1.2058, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.7523364485981308, |
|
"grad_norm": 246889.234375, |
|
"learning_rate": 3.7500000000000003e-05, |
|
"loss": 1.2045, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 1.8691588785046729, |
|
"grad_norm": 388048.875, |
|
"learning_rate": 4e-05, |
|
"loss": 1.143, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.985981308411215, |
|
"grad_norm": 442224.96875, |
|
"learning_rate": 4.25e-05, |
|
"loss": 1.1297, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.6003316749585407, |
|
"eval_loss": 1.0013911724090576, |
|
"eval_runtime": 15.593, |
|
"eval_samples_per_second": 77.343, |
|
"eval_steps_per_second": 2.437, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 2.102803738317757, |
|
"grad_norm": 518648.0, |
|
"learning_rate": 4.5e-05, |
|
"loss": 1.1438, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 2.2196261682242993, |
|
"grad_norm": 496547.5625, |
|
"learning_rate": 4.75e-05, |
|
"loss": 1.1246, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 2.336448598130841, |
|
"grad_norm": 306726.9375, |
|
"learning_rate": 5e-05, |
|
"loss": 1.1061, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 2.453271028037383, |
|
"grad_norm": 327747.03125, |
|
"learning_rate": 4.923780487804878e-05, |
|
"loss": 1.1146, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 2.5700934579439254, |
|
"grad_norm": 250996.6875, |
|
"learning_rate": 4.847560975609756e-05, |
|
"loss": 1.0954, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 2.6869158878504673, |
|
"grad_norm": 292352.34375, |
|
"learning_rate": 4.771341463414634e-05, |
|
"loss": 1.129, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 2.803738317757009, |
|
"grad_norm": 299764.5, |
|
"learning_rate": 4.695121951219512e-05, |
|
"loss": 1.0878, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 2.9205607476635516, |
|
"grad_norm": 263247.34375, |
|
"learning_rate": 4.618902439024391e-05, |
|
"loss": 1.0881, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.6558872305140961, |
|
"eval_loss": 0.9018123745918274, |
|
"eval_runtime": 15.6284, |
|
"eval_samples_per_second": 77.167, |
|
"eval_steps_per_second": 2.431, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 3.0373831775700935, |
|
"grad_norm": 281239.8125, |
|
"learning_rate": 4.542682926829269e-05, |
|
"loss": 1.094, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 3.1542056074766354, |
|
"grad_norm": 263740.625, |
|
"learning_rate": 4.466463414634147e-05, |
|
"loss": 1.036, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 3.2710280373831777, |
|
"grad_norm": 345132.46875, |
|
"learning_rate": 4.390243902439025e-05, |
|
"loss": 1.0542, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 3.3878504672897196, |
|
"grad_norm": 281706.03125, |
|
"learning_rate": 4.314024390243903e-05, |
|
"loss": 1.0712, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 3.5046728971962615, |
|
"grad_norm": 256964.125, |
|
"learning_rate": 4.237804878048781e-05, |
|
"loss": 1.0417, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 3.621495327102804, |
|
"grad_norm": 426164.21875, |
|
"learning_rate": 4.161585365853659e-05, |
|
"loss": 1.0363, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 3.7383177570093458, |
|
"grad_norm": 366963.625, |
|
"learning_rate": 4.085365853658537e-05, |
|
"loss": 1.0382, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 3.8551401869158877, |
|
"grad_norm": 314130.34375, |
|
"learning_rate": 4.0091463414634153e-05, |
|
"loss": 1.0255, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 3.97196261682243, |
|
"grad_norm": 343246.59375, |
|
"learning_rate": 3.932926829268293e-05, |
|
"loss": 1.0065, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.599502487562189, |
|
"eval_loss": 0.9687954187393188, |
|
"eval_runtime": 15.5027, |
|
"eval_samples_per_second": 77.793, |
|
"eval_steps_per_second": 2.451, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 4.088785046728972, |
|
"grad_norm": 355268.8125, |
|
"learning_rate": 3.856707317073171e-05, |
|
"loss": 0.9995, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 4.205607476635514, |
|
"grad_norm": 266058.53125, |
|
"learning_rate": 3.780487804878049e-05, |
|
"loss": 0.9972, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 4.322429906542056, |
|
"grad_norm": 447425.875, |
|
"learning_rate": 3.704268292682927e-05, |
|
"loss": 1.004, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 4.4392523364485985, |
|
"grad_norm": 382122.4375, |
|
"learning_rate": 3.628048780487805e-05, |
|
"loss": 0.967, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 4.55607476635514, |
|
"grad_norm": 379657.3125, |
|
"learning_rate": 3.551829268292683e-05, |
|
"loss": 0.9769, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 4.672897196261682, |
|
"grad_norm": 281210.5625, |
|
"learning_rate": 3.475609756097561e-05, |
|
"loss": 0.9521, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 4.789719626168225, |
|
"grad_norm": 337339.375, |
|
"learning_rate": 3.399390243902439e-05, |
|
"loss": 0.9654, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 4.906542056074766, |
|
"grad_norm": 227255.0625, |
|
"learning_rate": 3.323170731707317e-05, |
|
"loss": 1.0028, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_accuracy": 0.7014925373134329, |
|
"eval_loss": 0.8239747285842896, |
|
"eval_runtime": 15.7023, |
|
"eval_samples_per_second": 76.804, |
|
"eval_steps_per_second": 2.42, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 5.0233644859813085, |
|
"grad_norm": 396188.9375, |
|
"learning_rate": 3.246951219512195e-05, |
|
"loss": 1.002, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 5.140186915887851, |
|
"grad_norm": 331960.46875, |
|
"learning_rate": 3.170731707317073e-05, |
|
"loss": 0.9421, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 5.257009345794392, |
|
"grad_norm": 390978.8125, |
|
"learning_rate": 3.094512195121951e-05, |
|
"loss": 0.9338, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 5.373831775700935, |
|
"grad_norm": 321306.375, |
|
"learning_rate": 3.0182926829268294e-05, |
|
"loss": 0.9425, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 5.490654205607477, |
|
"grad_norm": 363200.09375, |
|
"learning_rate": 2.9420731707317074e-05, |
|
"loss": 0.9054, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 5.607476635514018, |
|
"grad_norm": 315464.1875, |
|
"learning_rate": 2.8658536585365854e-05, |
|
"loss": 0.9164, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 5.724299065420561, |
|
"grad_norm": 257387.84375, |
|
"learning_rate": 2.7896341463414637e-05, |
|
"loss": 0.9579, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 5.841121495327103, |
|
"grad_norm": 268010.53125, |
|
"learning_rate": 2.7134146341463417e-05, |
|
"loss": 0.9108, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 5.957943925233645, |
|
"grad_norm": 313754.125, |
|
"learning_rate": 2.6371951219512197e-05, |
|
"loss": 0.9225, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_accuracy": 0.7520729684908789, |
|
"eval_loss": 0.735476553440094, |
|
"eval_runtime": 15.6884, |
|
"eval_samples_per_second": 76.872, |
|
"eval_steps_per_second": 2.422, |
|
"step": 1284 |
|
}, |
|
{ |
|
"epoch": 6.074766355140187, |
|
"grad_norm": 319647.28125, |
|
"learning_rate": 2.5609756097560977e-05, |
|
"loss": 0.8836, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 6.191588785046729, |
|
"grad_norm": 440299.0, |
|
"learning_rate": 2.4847560975609756e-05, |
|
"loss": 0.8731, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 6.308411214953271, |
|
"grad_norm": 294800.8125, |
|
"learning_rate": 2.4085365853658536e-05, |
|
"loss": 0.8843, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 6.425233644859813, |
|
"grad_norm": 390740.28125, |
|
"learning_rate": 2.332317073170732e-05, |
|
"loss": 0.8199, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 6.542056074766355, |
|
"grad_norm": 393681.1875, |
|
"learning_rate": 2.25609756097561e-05, |
|
"loss": 0.8497, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 6.658878504672897, |
|
"grad_norm": 378513.21875, |
|
"learning_rate": 2.179878048780488e-05, |
|
"loss": 0.8445, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 6.775700934579439, |
|
"grad_norm": 330039.0625, |
|
"learning_rate": 2.103658536585366e-05, |
|
"loss": 0.8562, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 6.892523364485982, |
|
"grad_norm": 318586.625, |
|
"learning_rate": 2.0274390243902442e-05, |
|
"loss": 0.8522, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_accuracy": 0.746268656716418, |
|
"eval_loss": 0.76929771900177, |
|
"eval_runtime": 15.5157, |
|
"eval_samples_per_second": 77.728, |
|
"eval_steps_per_second": 2.449, |
|
"step": 1498 |
|
}, |
|
{ |
|
"epoch": 7.009345794392523, |
|
"grad_norm": 331889.5, |
|
"learning_rate": 1.9512195121951222e-05, |
|
"loss": 0.8529, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 7.126168224299065, |
|
"grad_norm": 381441.40625, |
|
"learning_rate": 1.8750000000000002e-05, |
|
"loss": 0.7812, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 7.242990654205608, |
|
"grad_norm": 248858.21875, |
|
"learning_rate": 1.798780487804878e-05, |
|
"loss": 0.8792, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 7.359813084112149, |
|
"grad_norm": 339380.375, |
|
"learning_rate": 1.722560975609756e-05, |
|
"loss": 0.8191, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 7.4766355140186915, |
|
"grad_norm": 363421.21875, |
|
"learning_rate": 1.6463414634146345e-05, |
|
"loss": 0.7828, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 7.593457943925234, |
|
"grad_norm": 332095.78125, |
|
"learning_rate": 1.5701219512195124e-05, |
|
"loss": 0.8328, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 7.710280373831775, |
|
"grad_norm": 267183.625, |
|
"learning_rate": 1.4939024390243902e-05, |
|
"loss": 0.8255, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 7.827102803738318, |
|
"grad_norm": 463742.25, |
|
"learning_rate": 1.4176829268292682e-05, |
|
"loss": 0.792, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 7.94392523364486, |
|
"grad_norm": 254509.796875, |
|
"learning_rate": 1.3414634146341466e-05, |
|
"loss": 0.821, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_accuracy": 0.7678275290215588, |
|
"eval_loss": 0.7130943536758423, |
|
"eval_runtime": 15.4772, |
|
"eval_samples_per_second": 77.921, |
|
"eval_steps_per_second": 2.455, |
|
"step": 1712 |
|
}, |
|
{ |
|
"epoch": 8.060747663551401, |
|
"grad_norm": 432679.84375, |
|
"learning_rate": 1.2652439024390245e-05, |
|
"loss": 0.7849, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 8.177570093457945, |
|
"grad_norm": 361022.125, |
|
"learning_rate": 1.1890243902439025e-05, |
|
"loss": 0.7447, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 8.294392523364486, |
|
"grad_norm": 416379.84375, |
|
"learning_rate": 1.1128048780487805e-05, |
|
"loss": 0.7898, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 8.411214953271028, |
|
"grad_norm": 386266.90625, |
|
"learning_rate": 1.0365853658536585e-05, |
|
"loss": 0.742, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 8.52803738317757, |
|
"grad_norm": 321598.21875, |
|
"learning_rate": 9.603658536585366e-06, |
|
"loss": 0.7681, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 8.644859813084112, |
|
"grad_norm": 386202.53125, |
|
"learning_rate": 8.841463414634146e-06, |
|
"loss": 0.7725, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 8.761682242990654, |
|
"grad_norm": 319017.71875, |
|
"learning_rate": 8.079268292682928e-06, |
|
"loss": 0.7394, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 8.878504672897197, |
|
"grad_norm": 388873.65625, |
|
"learning_rate": 7.317073170731707e-06, |
|
"loss": 0.7435, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 8.995327102803738, |
|
"grad_norm": 302018.6875, |
|
"learning_rate": 6.554878048780488e-06, |
|
"loss": 0.735, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_accuracy": 0.7761194029850746, |
|
"eval_loss": 0.7315632104873657, |
|
"eval_runtime": 15.8083, |
|
"eval_samples_per_second": 76.289, |
|
"eval_steps_per_second": 2.404, |
|
"step": 1926 |
|
}, |
|
{ |
|
"epoch": 9.11214953271028, |
|
"grad_norm": 400649.4375, |
|
"learning_rate": 5.792682926829269e-06, |
|
"loss": 0.6635, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 9.228971962616823, |
|
"grad_norm": 330898.75, |
|
"learning_rate": 5.030487804878049e-06, |
|
"loss": 0.7556, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 9.345794392523365, |
|
"grad_norm": 298238.21875, |
|
"learning_rate": 4.26829268292683e-06, |
|
"loss": 0.6982, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 9.462616822429906, |
|
"grad_norm": 244765.453125, |
|
"learning_rate": 3.5060975609756102e-06, |
|
"loss": 0.657, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 9.57943925233645, |
|
"grad_norm": 357814.375, |
|
"learning_rate": 2.7439024390243905e-06, |
|
"loss": 0.7641, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 9.69626168224299, |
|
"grad_norm": 500304.03125, |
|
"learning_rate": 1.9817073170731707e-06, |
|
"loss": 0.7561, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 9.813084112149532, |
|
"grad_norm": 457811.4375, |
|
"learning_rate": 1.2195121951219514e-06, |
|
"loss": 0.7183, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 9.929906542056075, |
|
"grad_norm": 392788.8125, |
|
"learning_rate": 4.573170731707317e-07, |
|
"loss": 0.7123, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_accuracy": 0.7777777777777778, |
|
"eval_loss": 0.7300788760185242, |
|
"eval_runtime": 15.6127, |
|
"eval_samples_per_second": 77.245, |
|
"eval_steps_per_second": 2.434, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"step": 2140, |
|
"total_flos": 5.510586115727032e+19, |
|
"train_loss": 0.9844673860852964, |
|
"train_runtime": 1946.7718, |
|
"train_samples_per_second": 35.084, |
|
"train_steps_per_second": 1.099 |
|
} |
|
], |
|
"logging_steps": 25, |
|
"max_steps": 2140, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 5.510586115727032e+19, |
|
"train_batch_size": 32, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|