{ "best_metric": 1.040339469909668, "best_model_checkpoint": "/media/mldrive/kcardenas/limb_classification_person_crop/beit-large-patch16-384/8_5e-5_5e-4_0.15/checkpoint-428", "epoch": 10.0, "eval_steps": 500, "global_step": 2140, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.11682242990654206, "grad_norm": 1226482.25, "learning_rate": 2.5e-06, "loss": 1.5898, "step": 25 }, { "epoch": 0.2336448598130841, "grad_norm": 883115.375, "learning_rate": 5e-06, "loss": 1.5063, "step": 50 }, { "epoch": 0.35046728971962615, "grad_norm": 729111.75, "learning_rate": 7.5e-06, "loss": 1.4618, "step": 75 }, { "epoch": 0.4672897196261682, "grad_norm": 401205.5, "learning_rate": 1e-05, "loss": 1.4029, "step": 100 }, { "epoch": 0.5841121495327103, "grad_norm": 392407.21875, "learning_rate": 1.25e-05, "loss": 1.3853, "step": 125 }, { "epoch": 0.7009345794392523, "grad_norm": 296861.40625, "learning_rate": 1.5e-05, "loss": 1.3715, "step": 150 }, { "epoch": 0.8177570093457944, "grad_norm": 644126.5, "learning_rate": 1.75e-05, "loss": 1.344, "step": 175 }, { "epoch": 0.9345794392523364, "grad_norm": 561771.25, "learning_rate": 2e-05, "loss": 1.3404, "step": 200 }, { "epoch": 1.0, "eval_accuracy": 0.39800995024875624, "eval_loss": 1.4185985326766968, "eval_runtime": 15.0252, "eval_samples_per_second": 80.265, "eval_steps_per_second": 2.529, "step": 214 }, { "epoch": 1.0514018691588785, "grad_norm": 450167.59375, "learning_rate": 2.25e-05, "loss": 1.2961, "step": 225 }, { "epoch": 1.1682242990654206, "grad_norm": 275605.5625, "learning_rate": 2.5e-05, "loss": 1.2836, "step": 250 }, { "epoch": 1.2850467289719627, "grad_norm": 417004.96875, "learning_rate": 2.7500000000000004e-05, "loss": 1.2374, "step": 275 }, { "epoch": 1.4018691588785046, "grad_norm": 313505.09375, "learning_rate": 3e-05, "loss": 1.2688, "step": 300 }, { "epoch": 1.5186915887850467, "grad_norm": 345996.875, "learning_rate": 3.2500000000000004e-05, "loss": 1.2501, "step": 325 }, { "epoch": 1.6355140186915889, "grad_norm": 439750.0, "learning_rate": 3.5e-05, "loss": 1.2455, "step": 350 }, { "epoch": 1.7523364485981308, "grad_norm": 192930.5625, "learning_rate": 3.7500000000000003e-05, "loss": 1.2433, "step": 375 }, { "epoch": 1.8691588785046729, "grad_norm": 403192.8125, "learning_rate": 4e-05, "loss": 1.1843, "step": 400 }, { "epoch": 1.985981308411215, "grad_norm": 312608.1875, "learning_rate": 4.25e-05, "loss": 1.1899, "step": 425 }, { "epoch": 2.0, "eval_accuracy": 0.6442786069651741, "eval_loss": 1.040339469909668, "eval_runtime": 15.0124, "eval_samples_per_second": 80.333, "eval_steps_per_second": 2.531, "step": 428 }, { "epoch": 2.102803738317757, "grad_norm": 565702.9375, "learning_rate": 4.5e-05, "loss": 1.1768, "step": 450 }, { "epoch": 2.2196261682242993, "grad_norm": 228019.625, "learning_rate": 4.75e-05, "loss": 1.3692, "step": 475 }, { "epoch": 2.336448598130841, "grad_norm": 97234.1796875, "learning_rate": 5e-05, "loss": 1.4012, "step": 500 }, { "epoch": 2.453271028037383, "grad_norm": 198352.921875, "learning_rate": 4.923780487804878e-05, "loss": 1.392, "step": 525 }, { "epoch": 2.5700934579439254, "grad_norm": 124283.6953125, "learning_rate": 4.847560975609756e-05, "loss": 1.3919, "step": 550 }, { "epoch": 2.6869158878504673, "grad_norm": 120723.7578125, "learning_rate": 4.771341463414634e-05, "loss": 1.3871, "step": 575 }, { "epoch": 2.803738317757009, "grad_norm": 213803.59375, "learning_rate": 4.695121951219512e-05, "loss": 1.3944, "step": 600 }, { "epoch": 2.9205607476635516, "grad_norm": 286424.5625, "learning_rate": 4.618902439024391e-05, "loss": 1.3962, "step": 625 }, { "epoch": 3.0, "eval_accuracy": 0.2744610281923715, "eval_loss": 1.3858356475830078, "eval_runtime": 15.0104, "eval_samples_per_second": 80.344, "eval_steps_per_second": 2.532, "step": 642 }, { "epoch": 3.0373831775700935, "grad_norm": 211193.328125, "learning_rate": 4.542682926829269e-05, "loss": 1.3901, "step": 650 }, { "epoch": 3.1542056074766354, "grad_norm": 111855.8359375, "learning_rate": 4.466463414634147e-05, "loss": 1.3935, "step": 675 }, { "epoch": 3.2710280373831777, "grad_norm": 157092.0625, "learning_rate": 4.390243902439025e-05, "loss": 1.3811, "step": 700 }, { "epoch": 3.3878504672897196, "grad_norm": 127590.5703125, "learning_rate": 4.314024390243903e-05, "loss": 1.3921, "step": 725 }, { "epoch": 3.5046728971962615, "grad_norm": 100796.25, "learning_rate": 4.237804878048781e-05, "loss": 1.3852, "step": 750 }, { "epoch": 3.621495327102804, "grad_norm": 139129.4375, "learning_rate": 4.161585365853659e-05, "loss": 1.3856, "step": 775 }, { "epoch": 3.7383177570093458, "grad_norm": 176988.171875, "learning_rate": 4.085365853658537e-05, "loss": 1.3844, "step": 800 }, { "epoch": 3.8551401869158877, "grad_norm": 127246.34375, "learning_rate": 4.0091463414634153e-05, "loss": 1.3805, "step": 825 }, { "epoch": 3.97196261682243, "grad_norm": 212300.828125, "learning_rate": 3.932926829268293e-05, "loss": 1.3914, "step": 850 }, { "epoch": 4.0, "eval_accuracy": 0.2744610281923715, "eval_loss": 1.3880783319473267, "eval_runtime": 15.2034, "eval_samples_per_second": 79.324, "eval_steps_per_second": 2.499, "step": 856 }, { "epoch": 4.088785046728972, "grad_norm": 61669.88671875, "learning_rate": 3.856707317073171e-05, "loss": 1.3905, "step": 875 }, { "epoch": 4.205607476635514, "grad_norm": 223599.234375, "learning_rate": 3.780487804878049e-05, "loss": 1.3906, "step": 900 }, { "epoch": 4.322429906542056, "grad_norm": 247944.15625, "learning_rate": 3.704268292682927e-05, "loss": 1.3902, "step": 925 }, { "epoch": 4.4392523364485985, "grad_norm": 136334.375, "learning_rate": 3.628048780487805e-05, "loss": 1.3901, "step": 950 }, { "epoch": 4.55607476635514, "grad_norm": 161834.765625, "learning_rate": 3.551829268292683e-05, "loss": 1.3842, "step": 975 }, { "epoch": 4.672897196261682, "grad_norm": 160751.734375, "learning_rate": 3.475609756097561e-05, "loss": 1.3856, "step": 1000 }, { "epoch": 4.789719626168225, "grad_norm": 152146.84375, "learning_rate": 3.399390243902439e-05, "loss": 1.3823, "step": 1025 }, { "epoch": 4.906542056074766, "grad_norm": 193127.203125, "learning_rate": 3.323170731707317e-05, "loss": 1.3868, "step": 1050 }, { "epoch": 5.0, "eval_accuracy": 0.27611940298507465, "eval_loss": 1.3848323822021484, "eval_runtime": 15.4933, "eval_samples_per_second": 77.84, "eval_steps_per_second": 2.453, "step": 1070 }, { "epoch": 5.0233644859813085, "grad_norm": 238072.984375, "learning_rate": 3.246951219512195e-05, "loss": 1.3888, "step": 1075 }, { "epoch": 5.140186915887851, "grad_norm": 133656.28125, "learning_rate": 3.170731707317073e-05, "loss": 1.3863, "step": 1100 }, { "epoch": 5.257009345794392, "grad_norm": 120263.9453125, "learning_rate": 3.094512195121951e-05, "loss": 1.3856, "step": 1125 }, { "epoch": 5.373831775700935, "grad_norm": 167475.484375, "learning_rate": 3.0182926829268294e-05, "loss": 1.3885, "step": 1150 }, { "epoch": 5.490654205607477, "grad_norm": 197263.59375, "learning_rate": 2.9420731707317074e-05, "loss": 1.385, "step": 1175 }, { "epoch": 5.607476635514018, "grad_norm": 145337.453125, "learning_rate": 2.8658536585365854e-05, "loss": 1.3807, "step": 1200 }, { "epoch": 5.724299065420561, "grad_norm": 119471.03125, "learning_rate": 2.7896341463414637e-05, "loss": 1.3806, "step": 1225 }, { "epoch": 5.841121495327103, "grad_norm": 139454.765625, "learning_rate": 2.7134146341463417e-05, "loss": 1.3843, "step": 1250 }, { "epoch": 5.957943925233645, "grad_norm": 147728.828125, "learning_rate": 2.6371951219512197e-05, "loss": 1.3895, "step": 1275 }, { "epoch": 6.0, "eval_accuracy": 0.2744610281923715, "eval_loss": 1.386699914932251, "eval_runtime": 14.8197, "eval_samples_per_second": 81.378, "eval_steps_per_second": 2.564, "step": 1284 }, { "epoch": 6.074766355140187, "grad_norm": 217932.9375, "learning_rate": 2.5609756097560977e-05, "loss": 1.3858, "step": 1300 }, { "epoch": 6.191588785046729, "grad_norm": 103804.140625, "learning_rate": 2.4847560975609756e-05, "loss": 1.3839, "step": 1325 }, { "epoch": 6.308411214953271, "grad_norm": 144548.296875, "learning_rate": 2.4085365853658536e-05, "loss": 1.3757, "step": 1350 }, { "epoch": 6.425233644859813, "grad_norm": 116841.2421875, "learning_rate": 2.332317073170732e-05, "loss": 1.3832, "step": 1375 }, { "epoch": 6.542056074766355, "grad_norm": 60290.0859375, "learning_rate": 2.25609756097561e-05, "loss": 1.3866, "step": 1400 }, { "epoch": 6.658878504672897, "grad_norm": 142147.796875, "learning_rate": 2.179878048780488e-05, "loss": 1.3812, "step": 1425 }, { "epoch": 6.775700934579439, "grad_norm": 77919.5390625, "learning_rate": 2.103658536585366e-05, "loss": 1.3852, "step": 1450 }, { "epoch": 6.892523364485982, "grad_norm": 112284.5234375, "learning_rate": 2.0274390243902442e-05, "loss": 1.3871, "step": 1475 }, { "epoch": 7.0, "eval_accuracy": 0.2744610281923715, "eval_loss": 1.3830299377441406, "eval_runtime": 15.5048, "eval_samples_per_second": 77.782, "eval_steps_per_second": 2.451, "step": 1498 }, { "epoch": 7.009345794392523, "grad_norm": 219239.40625, "learning_rate": 1.9512195121951222e-05, "loss": 1.3871, "step": 1500 }, { "epoch": 7.126168224299065, "grad_norm": 139572.75, "learning_rate": 1.8750000000000002e-05, "loss": 1.3849, "step": 1525 }, { "epoch": 7.242990654205608, "grad_norm": 71637.71875, "learning_rate": 1.798780487804878e-05, "loss": 1.3832, "step": 1550 }, { "epoch": 7.359813084112149, "grad_norm": 134367.84375, "learning_rate": 1.722560975609756e-05, "loss": 1.3789, "step": 1575 }, { "epoch": 7.4766355140186915, "grad_norm": 146302.65625, "learning_rate": 1.6463414634146345e-05, "loss": 1.3838, "step": 1600 }, { "epoch": 7.593457943925234, "grad_norm": 65325.31640625, "learning_rate": 1.5701219512195124e-05, "loss": 1.3808, "step": 1625 }, { "epoch": 7.710280373831775, "grad_norm": 105734.6328125, "learning_rate": 1.4939024390243902e-05, "loss": 1.3811, "step": 1650 }, { "epoch": 7.827102803738318, "grad_norm": 158432.8125, "learning_rate": 1.4176829268292682e-05, "loss": 1.3782, "step": 1675 }, { "epoch": 7.94392523364486, "grad_norm": 109856.3671875, "learning_rate": 1.3414634146341466e-05, "loss": 1.3802, "step": 1700 }, { "epoch": 8.0, "eval_accuracy": 0.29270315091210614, "eval_loss": 1.378788709640503, "eval_runtime": 15.2702, "eval_samples_per_second": 78.977, "eval_steps_per_second": 2.489, "step": 1712 }, { "epoch": 8.060747663551401, "grad_norm": 131191.734375, "learning_rate": 1.2652439024390245e-05, "loss": 1.3796, "step": 1725 }, { "epoch": 8.177570093457945, "grad_norm": 111280.453125, "learning_rate": 1.1890243902439025e-05, "loss": 1.3764, "step": 1750 }, { "epoch": 8.294392523364486, "grad_norm": 53637.3828125, "learning_rate": 1.1128048780487805e-05, "loss": 1.3815, "step": 1775 }, { "epoch": 8.411214953271028, "grad_norm": 67744.1640625, "learning_rate": 1.0365853658536585e-05, "loss": 1.3667, "step": 1800 }, { "epoch": 8.52803738317757, "grad_norm": 88018.4921875, "learning_rate": 9.603658536585366e-06, "loss": 1.369, "step": 1825 }, { "epoch": 8.644859813084112, "grad_norm": 92662.328125, "learning_rate": 8.841463414634146e-06, "loss": 1.3732, "step": 1850 }, { "epoch": 8.761682242990654, "grad_norm": 146536.6875, "learning_rate": 8.079268292682928e-06, "loss": 1.3732, "step": 1875 }, { "epoch": 8.878504672897197, "grad_norm": 132991.578125, "learning_rate": 7.317073170731707e-06, "loss": 1.3685, "step": 1900 }, { "epoch": 8.995327102803738, "grad_norm": 134634.984375, "learning_rate": 6.554878048780488e-06, "loss": 1.3745, "step": 1925 }, { "epoch": 9.0, "eval_accuracy": 0.3200663349917081, "eval_loss": 1.367501974105835, "eval_runtime": 15.3432, "eval_samples_per_second": 78.602, "eval_steps_per_second": 2.477, "step": 1926 }, { "epoch": 9.11214953271028, "grad_norm": 111594.0234375, "learning_rate": 5.792682926829269e-06, "loss": 1.3629, "step": 1950 }, { "epoch": 9.228971962616823, "grad_norm": 260942.34375, "learning_rate": 5.030487804878049e-06, "loss": 1.3681, "step": 1975 }, { "epoch": 9.345794392523365, "grad_norm": 132676.8125, "learning_rate": 4.26829268292683e-06, "loss": 1.3568, "step": 2000 }, { "epoch": 9.462616822429906, "grad_norm": 137239.046875, "learning_rate": 3.5060975609756102e-06, "loss": 1.3566, "step": 2025 }, { "epoch": 9.57943925233645, "grad_norm": 167074.9375, "learning_rate": 2.7439024390243905e-06, "loss": 1.3556, "step": 2050 }, { "epoch": 9.69626168224299, "grad_norm": 89201.1015625, "learning_rate": 1.9817073170731707e-06, "loss": 1.3697, "step": 2075 }, { "epoch": 9.813084112149532, "grad_norm": 104514.7109375, "learning_rate": 1.2195121951219514e-06, "loss": 1.346, "step": 2100 }, { "epoch": 9.929906542056075, "grad_norm": 137668.4375, "learning_rate": 4.573170731707317e-07, "loss": 1.356, "step": 2125 }, { "epoch": 10.0, "eval_accuracy": 0.3383084577114428, "eval_loss": 1.3566746711730957, "eval_runtime": 14.6136, "eval_samples_per_second": 82.526, "eval_steps_per_second": 2.6, "step": 2140 }, { "epoch": 10.0, "step": 2140, "total_flos": 5.510586115727032e+19, "train_loss": 1.3681953902556518, "train_runtime": 1910.7588, "train_samples_per_second": 35.745, "train_steps_per_second": 1.12 } ], "logging_steps": 25, "max_steps": 2140, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.510586115727032e+19, "train_batch_size": 32, "trial_name": null, "trial_params": null }