|
{ |
|
"best_metric": 0.7303220035778175, |
|
"best_model_checkpoint": "swin-tiny-patch4-window7-224-img_orientation/checkpoint-942", |
|
"epoch": 2.9952305246422894, |
|
"eval_steps": 500, |
|
"global_step": 942, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.263157894736842e-06, |
|
"loss": 1.4216, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.0526315789473684e-05, |
|
"loss": 1.4124, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.5789473684210526e-05, |
|
"loss": 1.3621, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 2.105263157894737e-05, |
|
"loss": 1.3148, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 2.6315789473684212e-05, |
|
"loss": 1.2304, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 3.157894736842105e-05, |
|
"loss": 1.108, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 3.6842105263157895e-05, |
|
"loss": 1.0056, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.210526315789474e-05, |
|
"loss": 0.8832, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.736842105263158e-05, |
|
"loss": 0.8584, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.970484061393152e-05, |
|
"loss": 0.796, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.9114521841794566e-05, |
|
"loss": 0.7721, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.852420306965762e-05, |
|
"loss": 0.8128, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.793388429752066e-05, |
|
"loss": 0.7318, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.734356552538371e-05, |
|
"loss": 0.7374, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.675324675324675e-05, |
|
"loss": 0.7078, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.61629279811098e-05, |
|
"loss": 0.7153, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.557260920897285e-05, |
|
"loss": 0.6978, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.4982290436835895e-05, |
|
"loss": 0.6526, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.439197166469894e-05, |
|
"loss": 0.6883, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 4.3801652892561984e-05, |
|
"loss": 0.6491, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 4.321133412042503e-05, |
|
"loss": 0.677, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.262101534828808e-05, |
|
"loss": 0.6831, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.2030696576151126e-05, |
|
"loss": 0.6765, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 4.144037780401417e-05, |
|
"loss": 0.6734, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 4.0850059031877216e-05, |
|
"loss": 0.6574, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 4.025974025974026e-05, |
|
"loss": 0.6897, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.9669421487603306e-05, |
|
"loss": 0.6886, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.907910271546635e-05, |
|
"loss": 0.6466, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.8488783943329396e-05, |
|
"loss": 0.6309, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 3.789846517119244e-05, |
|
"loss": 0.6513, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 3.730814639905549e-05, |
|
"loss": 0.6197, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.7122093023255814, |
|
"eval_loss": 0.4981530010700226, |
|
"eval_runtime": 32.8108, |
|
"eval_samples_per_second": 136.296, |
|
"eval_steps_per_second": 4.267, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 3.671782762691854e-05, |
|
"loss": 0.6347, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 3.612750885478158e-05, |
|
"loss": 0.6022, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 3.553719008264463e-05, |
|
"loss": 0.6179, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 3.494687131050768e-05, |
|
"loss": 0.6227, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 3.4356552538370724e-05, |
|
"loss": 0.6276, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 3.376623376623377e-05, |
|
"loss": 0.6085, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 3.3175914994096814e-05, |
|
"loss": 0.6492, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 3.258559622195986e-05, |
|
"loss": 0.6094, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 3.199527744982291e-05, |
|
"loss": 0.609, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 3.1404958677685955e-05, |
|
"loss": 0.5871, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 3.0814639905549e-05, |
|
"loss": 0.6319, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 3.022432113341204e-05, |
|
"loss": 0.5786, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 2.9634002361275087e-05, |
|
"loss": 0.5804, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 2.9043683589138138e-05, |
|
"loss": 0.6038, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 2.8453364817001183e-05, |
|
"loss": 0.6018, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 2.7863046044864228e-05, |
|
"loss": 0.5901, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 2.7272727272727273e-05, |
|
"loss": 0.5967, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 2.668240850059032e-05, |
|
"loss": 0.6058, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 2.6092089728453366e-05, |
|
"loss": 0.5511, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 2.550177095631641e-05, |
|
"loss": 0.5863, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 2.4911452184179456e-05, |
|
"loss": 0.559, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 2.4321133412042505e-05, |
|
"loss": 0.5856, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 2.373081463990555e-05, |
|
"loss": 0.5901, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 2.3140495867768598e-05, |
|
"loss": 0.5703, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 2.2550177095631643e-05, |
|
"loss": 0.6038, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 2.1959858323494688e-05, |
|
"loss": 0.5987, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 2.1369539551357733e-05, |
|
"loss": 0.5859, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 2.077922077922078e-05, |
|
"loss": 0.6033, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 2.0188902007083826e-05, |
|
"loss": 0.5886, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 1.959858323494687e-05, |
|
"loss": 0.5741, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 1.900826446280992e-05, |
|
"loss": 0.5524, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.7186940966010733, |
|
"eval_loss": 0.4689517319202423, |
|
"eval_runtime": 32.5075, |
|
"eval_samples_per_second": 137.568, |
|
"eval_steps_per_second": 4.307, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 1.8417945690672964e-05, |
|
"loss": 0.5363, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 1.7827626918536012e-05, |
|
"loss": 0.5789, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 1.7237308146399057e-05, |
|
"loss": 0.5492, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 1.6646989374262102e-05, |
|
"loss": 0.5799, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 1.6056670602125147e-05, |
|
"loss": 0.5762, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 1.5466351829988192e-05, |
|
"loss": 0.5331, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 1.487603305785124e-05, |
|
"loss": 0.531, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 1.4285714285714285e-05, |
|
"loss": 0.5419, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 1.3695395513577334e-05, |
|
"loss": 0.5359, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 1.3105076741440379e-05, |
|
"loss": 0.5551, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 1.2514757969303425e-05, |
|
"loss": 0.5616, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 1.192443919716647e-05, |
|
"loss": 0.5382, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 1.1334120425029517e-05, |
|
"loss": 0.537, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 1.0743801652892564e-05, |
|
"loss": 0.5468, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 1.0153482880755608e-05, |
|
"loss": 0.5369, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 9.563164108618655e-06, |
|
"loss": 0.531, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 8.9728453364817e-06, |
|
"loss": 0.527, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 8.382526564344747e-06, |
|
"loss": 0.5555, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 7.792207792207792e-06, |
|
"loss": 0.5759, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 7.201889020070838e-06, |
|
"loss": 0.5326, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 6.611570247933885e-06, |
|
"loss": 0.5443, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 6.021251475796931e-06, |
|
"loss": 0.5503, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 5.430932703659976e-06, |
|
"loss": 0.5494, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 4.840613931523023e-06, |
|
"loss": 0.5507, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 4.250295159386069e-06, |
|
"loss": 0.5402, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 3.659976387249114e-06, |
|
"loss": 0.5223, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 3.0696576151121604e-06, |
|
"loss": 0.5563, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 2.4793388429752066e-06, |
|
"loss": 0.5068, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 1.8890200708382526e-06, |
|
"loss": 0.5425, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.2987012987012988e-06, |
|
"loss": 0.5568, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 7.083825265643448e-07, |
|
"loss": 0.5354, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 1.1806375442739079e-07, |
|
"loss": 0.5268, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.7303220035778175, |
|
"eval_loss": 0.45461490750312805, |
|
"eval_runtime": 31.991, |
|
"eval_samples_per_second": 139.789, |
|
"eval_steps_per_second": 4.376, |
|
"step": 942 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 942, |
|
"total_flos": 2.996656354433065e+18, |
|
"train_loss": 0.6572925214301755, |
|
"train_runtime": 1977.1874, |
|
"train_samples_per_second": 61.064, |
|
"train_steps_per_second": 0.476 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 942, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"total_flos": 2.996656354433065e+18, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|