|
{ |
|
"best_metric": 0.9396355353075171, |
|
"best_model_checkpoint": "swin-base-patch4-window7-224-in22k-finetuned-brain-tumor-final/checkpoint-495", |
|
"epoch": 19.87878787878788, |
|
"global_step": 820, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.2195121951219514e-06, |
|
"loss": 1.4212, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 2.4390243902439027e-06, |
|
"loss": 1.3801, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.6585365853658537e-06, |
|
"loss": 1.2755, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.8780487804878055e-06, |
|
"loss": 1.1562, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"eval_accuracy": 0.6378132118451025, |
|
"eval_f1_score": 0.6190569285283298, |
|
"eval_loss": 1.1378333568572998, |
|
"eval_precision": 0.6536626392431208, |
|
"eval_runtime": 96.8359, |
|
"eval_samples_per_second": 18.134, |
|
"eval_steps_per_second": 0.568, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 6.0975609756097564e-06, |
|
"loss": 0.9839, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 7.317073170731707e-06, |
|
"loss": 0.7914, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 8.536585365853658e-06, |
|
"loss": 0.6245, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 9.756097560975611e-06, |
|
"loss": 0.4878, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"eval_accuracy": 0.7591116173120729, |
|
"eval_f1_score": 0.7498580587869668, |
|
"eval_loss": 0.6476529836654663, |
|
"eval_precision": 0.7873955323748831, |
|
"eval_runtime": 96.8155, |
|
"eval_samples_per_second": 18.138, |
|
"eval_steps_per_second": 0.568, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 9.89159891598916e-06, |
|
"loss": 0.3834, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 9.756097560975611e-06, |
|
"loss": 0.3288, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 9.620596205962061e-06, |
|
"loss": 0.3111, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 9.485094850948512e-06, |
|
"loss": 0.2623, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"eval_accuracy": 0.8337129840546698, |
|
"eval_f1_score": 0.8310652064326258, |
|
"eval_loss": 0.44103795289993286, |
|
"eval_precision": 0.8487545531816145, |
|
"eval_runtime": 95.5234, |
|
"eval_samples_per_second": 18.383, |
|
"eval_steps_per_second": 0.576, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 9.34959349593496e-06, |
|
"loss": 0.2449, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 9.21409214092141e-06, |
|
"loss": 0.2058, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 9.07859078590786e-06, |
|
"loss": 0.2426, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 8.94308943089431e-06, |
|
"loss": 0.1985, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.8143507972665148, |
|
"eval_f1_score": 0.8114988597034715, |
|
"eval_loss": 0.465999573469162, |
|
"eval_precision": 0.8455116716314306, |
|
"eval_runtime": 95.3707, |
|
"eval_samples_per_second": 18.412, |
|
"eval_steps_per_second": 0.577, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 4.12, |
|
"learning_rate": 8.80758807588076e-06, |
|
"loss": 0.1852, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 4.36, |
|
"learning_rate": 8.67208672086721e-06, |
|
"loss": 0.1738, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 4.61, |
|
"learning_rate": 8.536585365853658e-06, |
|
"loss": 0.1806, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 4.85, |
|
"learning_rate": 8.401084010840109e-06, |
|
"loss": 0.1736, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 4.99, |
|
"eval_accuracy": 0.8775626423690205, |
|
"eval_f1_score": 0.8759706154432891, |
|
"eval_loss": 0.3229566812515259, |
|
"eval_precision": 0.8894198873971897, |
|
"eval_runtime": 96.4712, |
|
"eval_samples_per_second": 18.202, |
|
"eval_steps_per_second": 0.57, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 5.09, |
|
"learning_rate": 8.265582655826559e-06, |
|
"loss": 0.1677, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 5.33, |
|
"learning_rate": 8.130081300813009e-06, |
|
"loss": 0.1384, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 5.58, |
|
"learning_rate": 7.99457994579946e-06, |
|
"loss": 0.1319, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 5.82, |
|
"learning_rate": 7.859078590785908e-06, |
|
"loss": 0.124, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 5.99, |
|
"eval_accuracy": 0.9026195899772209, |
|
"eval_f1_score": 0.9013627262353661, |
|
"eval_loss": 0.2684251368045807, |
|
"eval_precision": 0.9089527019800868, |
|
"eval_runtime": 94.5477, |
|
"eval_samples_per_second": 18.573, |
|
"eval_steps_per_second": 0.582, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 6.06, |
|
"learning_rate": 7.723577235772358e-06, |
|
"loss": 0.1508, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 6.3, |
|
"learning_rate": 7.5880758807588085e-06, |
|
"loss": 0.1168, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 6.55, |
|
"learning_rate": 7.452574525745257e-06, |
|
"loss": 0.1043, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 6.79, |
|
"learning_rate": 7.317073170731707e-06, |
|
"loss": 0.1278, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 6.98, |
|
"eval_accuracy": 0.9179954441913439, |
|
"eval_f1_score": 0.9166179163753916, |
|
"eval_loss": 0.22102278470993042, |
|
"eval_precision": 0.9210303187193152, |
|
"eval_runtime": 95.7688, |
|
"eval_samples_per_second": 18.336, |
|
"eval_steps_per_second": 0.574, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 7.181571815718158e-06, |
|
"loss": 0.1122, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 7.27, |
|
"learning_rate": 7.046070460704607e-06, |
|
"loss": 0.1052, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 7.52, |
|
"learning_rate": 6.910569105691057e-06, |
|
"loss": 0.0965, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 7.76, |
|
"learning_rate": 6.775067750677508e-06, |
|
"loss": 0.1072, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 6.639566395663957e-06, |
|
"loss": 0.0959, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_accuracy": 0.9208428246013668, |
|
"eval_f1_score": 0.9195432105337242, |
|
"eval_loss": 0.21511134505271912, |
|
"eval_precision": 0.9259948741044411, |
|
"eval_runtime": 94.727, |
|
"eval_samples_per_second": 18.537, |
|
"eval_steps_per_second": 0.581, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 8.24, |
|
"learning_rate": 6.504065040650407e-06, |
|
"loss": 0.087, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 8.48, |
|
"learning_rate": 6.368563685636857e-06, |
|
"loss": 0.0788, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 8.73, |
|
"learning_rate": 6.233062330623306e-06, |
|
"loss": 0.0952, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 8.97, |
|
"learning_rate": 6.0975609756097564e-06, |
|
"loss": 0.0849, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 8.99, |
|
"eval_accuracy": 0.9219817767653758, |
|
"eval_f1_score": 0.9205449523633584, |
|
"eval_loss": 0.21544188261032104, |
|
"eval_precision": 0.9290889598643586, |
|
"eval_runtime": 95.0804, |
|
"eval_samples_per_second": 18.469, |
|
"eval_steps_per_second": 0.578, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 9.21, |
|
"learning_rate": 5.962059620596207e-06, |
|
"loss": 0.0954, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 9.45, |
|
"learning_rate": 5.826558265582656e-06, |
|
"loss": 0.0808, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 9.7, |
|
"learning_rate": 5.691056910569106e-06, |
|
"loss": 0.0703, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 9.94, |
|
"learning_rate": 5.555555555555557e-06, |
|
"loss": 0.0805, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 9.99, |
|
"eval_accuracy": 0.9191343963553531, |
|
"eval_f1_score": 0.9178865857712233, |
|
"eval_loss": 0.21121914684772491, |
|
"eval_precision": 0.9250862844253733, |
|
"eval_runtime": 94.2987, |
|
"eval_samples_per_second": 18.622, |
|
"eval_steps_per_second": 0.583, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 10.18, |
|
"learning_rate": 5.420054200542005e-06, |
|
"loss": 0.0521, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 10.42, |
|
"learning_rate": 5.2845528455284555e-06, |
|
"loss": 0.0681, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 10.67, |
|
"learning_rate": 5.149051490514906e-06, |
|
"loss": 0.0769, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 10.91, |
|
"learning_rate": 5.013550135501355e-06, |
|
"loss": 0.0682, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 10.98, |
|
"eval_accuracy": 0.9384965831435079, |
|
"eval_f1_score": 0.9369393287065592, |
|
"eval_loss": 0.15628309547901154, |
|
"eval_precision": 0.9402176891312171, |
|
"eval_runtime": 96.028, |
|
"eval_samples_per_second": 18.286, |
|
"eval_steps_per_second": 0.573, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 11.15, |
|
"learning_rate": 4.8780487804878055e-06, |
|
"loss": 0.0668, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 11.39, |
|
"learning_rate": 4.742547425474256e-06, |
|
"loss": 0.0657, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 11.64, |
|
"learning_rate": 4.607046070460705e-06, |
|
"loss": 0.0703, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 11.88, |
|
"learning_rate": 4.471544715447155e-06, |
|
"loss": 0.0624, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_accuracy": 0.9396355353075171, |
|
"eval_f1_score": 0.9384678994386635, |
|
"eval_loss": 0.1576806604862213, |
|
"eval_precision": 0.9408448811333167, |
|
"eval_runtime": 95.9898, |
|
"eval_samples_per_second": 18.294, |
|
"eval_steps_per_second": 0.573, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 12.12, |
|
"learning_rate": 4.336043360433605e-06, |
|
"loss": 0.0531, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 12.36, |
|
"learning_rate": 4.200542005420054e-06, |
|
"loss": 0.0579, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 12.61, |
|
"learning_rate": 4.0650406504065046e-06, |
|
"loss": 0.0657, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 12.85, |
|
"learning_rate": 3.929539295392954e-06, |
|
"loss": 0.0415, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 12.99, |
|
"eval_accuracy": 0.9305239179954442, |
|
"eval_f1_score": 0.9294348170687547, |
|
"eval_loss": 0.1836157888174057, |
|
"eval_precision": 0.9332309174336554, |
|
"eval_runtime": 99.8745, |
|
"eval_samples_per_second": 17.582, |
|
"eval_steps_per_second": 0.551, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 13.09, |
|
"learning_rate": 3.7940379403794043e-06, |
|
"loss": 0.0569, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 13.33, |
|
"learning_rate": 3.6585365853658537e-06, |
|
"loss": 0.0707, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 13.58, |
|
"learning_rate": 3.5230352303523035e-06, |
|
"loss": 0.0424, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 13.82, |
|
"learning_rate": 3.387533875338754e-06, |
|
"loss": 0.0465, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 13.99, |
|
"eval_accuracy": 0.9202733485193622, |
|
"eval_f1_score": 0.9191914508633344, |
|
"eval_loss": 0.2144739031791687, |
|
"eval_precision": 0.9251966092567651, |
|
"eval_runtime": 96.9873, |
|
"eval_samples_per_second": 18.105, |
|
"eval_steps_per_second": 0.567, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 14.06, |
|
"learning_rate": 3.2520325203252037e-06, |
|
"loss": 0.0483, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 14.3, |
|
"learning_rate": 3.116531165311653e-06, |
|
"loss": 0.0432, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 14.55, |
|
"learning_rate": 2.9810298102981034e-06, |
|
"loss": 0.046, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 14.79, |
|
"learning_rate": 2.845528455284553e-06, |
|
"loss": 0.056, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 14.98, |
|
"eval_accuracy": 0.9339407744874715, |
|
"eval_f1_score": 0.9324607544046006, |
|
"eval_loss": 0.17098096013069153, |
|
"eval_precision": 0.9369044224464911, |
|
"eval_runtime": 96.6806, |
|
"eval_samples_per_second": 18.163, |
|
"eval_steps_per_second": 0.569, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 2.7100271002710026e-06, |
|
"loss": 0.0447, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 15.27, |
|
"learning_rate": 2.574525745257453e-06, |
|
"loss": 0.0441, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 15.52, |
|
"learning_rate": 2.4390243902439027e-06, |
|
"loss": 0.0515, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 15.76, |
|
"learning_rate": 2.3035230352303526e-06, |
|
"loss": 0.0435, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"learning_rate": 2.1680216802168024e-06, |
|
"loss": 0.0545, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_accuracy": 0.9248291571753986, |
|
"eval_f1_score": 0.9235589091910614, |
|
"eval_loss": 0.20944343507289886, |
|
"eval_precision": 0.929753616084827, |
|
"eval_runtime": 97.9348, |
|
"eval_samples_per_second": 17.93, |
|
"eval_steps_per_second": 0.562, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 16.24, |
|
"learning_rate": 2.0325203252032523e-06, |
|
"loss": 0.0477, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 16.48, |
|
"learning_rate": 1.8970189701897021e-06, |
|
"loss": 0.0424, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 16.73, |
|
"learning_rate": 1.7615176151761518e-06, |
|
"loss": 0.0377, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 16.97, |
|
"learning_rate": 1.6260162601626018e-06, |
|
"loss": 0.0591, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 16.99, |
|
"eval_accuracy": 0.9316628701594533, |
|
"eval_f1_score": 0.930292082850388, |
|
"eval_loss": 0.1752157062292099, |
|
"eval_precision": 0.934052681687829, |
|
"eval_runtime": 98.3671, |
|
"eval_samples_per_second": 17.852, |
|
"eval_steps_per_second": 0.559, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 17.21, |
|
"learning_rate": 1.4905149051490517e-06, |
|
"loss": 0.0425, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 17.45, |
|
"learning_rate": 1.3550135501355013e-06, |
|
"loss": 0.0394, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 17.7, |
|
"learning_rate": 1.2195121951219514e-06, |
|
"loss": 0.0323, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 17.94, |
|
"learning_rate": 1.0840108401084012e-06, |
|
"loss": 0.0512, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 17.99, |
|
"eval_accuracy": 0.9310933940774487, |
|
"eval_f1_score": 0.9297072356118401, |
|
"eval_loss": 0.17810992896556854, |
|
"eval_precision": 0.9341920124235589, |
|
"eval_runtime": 99.187, |
|
"eval_samples_per_second": 17.704, |
|
"eval_steps_per_second": 0.555, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 18.18, |
|
"learning_rate": 9.485094850948511e-07, |
|
"loss": 0.0278, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 18.42, |
|
"learning_rate": 8.130081300813009e-07, |
|
"loss": 0.0338, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 18.67, |
|
"learning_rate": 6.775067750677507e-07, |
|
"loss": 0.0467, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 18.91, |
|
"learning_rate": 5.420054200542006e-07, |
|
"loss": 0.0424, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 18.98, |
|
"eval_accuracy": 0.9305239179954442, |
|
"eval_f1_score": 0.9292526869769537, |
|
"eval_loss": 0.18729418516159058, |
|
"eval_precision": 0.9338045676207426, |
|
"eval_runtime": 97.3576, |
|
"eval_samples_per_second": 18.037, |
|
"eval_steps_per_second": 0.565, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 19.15, |
|
"learning_rate": 4.0650406504065046e-07, |
|
"loss": 0.0333, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 19.39, |
|
"learning_rate": 2.710027100271003e-07, |
|
"loss": 0.039, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 19.64, |
|
"learning_rate": 1.3550135501355015e-07, |
|
"loss": 0.0344, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 19.88, |
|
"learning_rate": 0.0, |
|
"loss": 0.0438, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 19.88, |
|
"eval_accuracy": 0.9265375854214123, |
|
"eval_f1_score": 0.9252094972625854, |
|
"eval_loss": 0.19554641842842102, |
|
"eval_precision": 0.9307429809120586, |
|
"eval_runtime": 100.1651, |
|
"eval_samples_per_second": 17.531, |
|
"eval_steps_per_second": 0.549, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 19.88, |
|
"step": 820, |
|
"total_flos": 8.203992042011222e+18, |
|
"train_loss": 0.18675074555524965, |
|
"train_runtime": 20102.5048, |
|
"train_samples_per_second": 5.24, |
|
"train_steps_per_second": 0.041 |
|
} |
|
], |
|
"max_steps": 820, |
|
"num_train_epochs": 20, |
|
"total_flos": 8.203992042011222e+18, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|