|
{ |
|
"best_metric": 0.9691358024691358, |
|
"best_model_checkpoint": "swin-tiny-patch4-window7-224-finetuned-eurosat/checkpoint-342", |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 342, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.08771929824561403, |
|
"grad_norm": 4.784806251525879, |
|
"learning_rate": 1.4285714285714285e-05, |
|
"loss": 2.2537, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.17543859649122806, |
|
"grad_norm": 7.325711727142334, |
|
"learning_rate": 2.857142857142857e-05, |
|
"loss": 1.9328, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.2631578947368421, |
|
"grad_norm": 12.205706596374512, |
|
"learning_rate": 4.2857142857142856e-05, |
|
"loss": 1.3033, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.3508771929824561, |
|
"grad_norm": 21.546316146850586, |
|
"learning_rate": 4.9185667752443e-05, |
|
"loss": 0.7435, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.43859649122807015, |
|
"grad_norm": 11.416431427001953, |
|
"learning_rate": 4.755700325732899e-05, |
|
"loss": 0.5723, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.5263157894736842, |
|
"grad_norm": 13.339532852172852, |
|
"learning_rate": 4.592833876221499e-05, |
|
"loss": 0.4688, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.6140350877192983, |
|
"grad_norm": 18.226308822631836, |
|
"learning_rate": 4.429967426710098e-05, |
|
"loss": 0.4296, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.7017543859649122, |
|
"grad_norm": 24.142580032348633, |
|
"learning_rate": 4.2671009771986977e-05, |
|
"loss": 0.3557, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.7894736842105263, |
|
"grad_norm": 18.747182846069336, |
|
"learning_rate": 4.104234527687297e-05, |
|
"loss": 0.3364, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.8771929824561403, |
|
"grad_norm": 9.604598999023438, |
|
"learning_rate": 3.941368078175896e-05, |
|
"loss": 0.379, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.9649122807017544, |
|
"grad_norm": 9.6405668258667, |
|
"learning_rate": 3.778501628664495e-05, |
|
"loss": 0.2967, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.9271604938271605, |
|
"eval_loss": 0.20734398066997528, |
|
"eval_runtime": 4.5343, |
|
"eval_samples_per_second": 357.28, |
|
"eval_steps_per_second": 11.248, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 1.0526315789473684, |
|
"grad_norm": 10.606453895568848, |
|
"learning_rate": 3.615635179153095e-05, |
|
"loss": 0.2506, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.1403508771929824, |
|
"grad_norm": 13.219770431518555, |
|
"learning_rate": 3.452768729641694e-05, |
|
"loss": 0.2646, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.2280701754385965, |
|
"grad_norm": 14.424461364746094, |
|
"learning_rate": 3.289902280130293e-05, |
|
"loss": 0.3161, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.3157894736842106, |
|
"grad_norm": 16.468141555786133, |
|
"learning_rate": 3.127035830618892e-05, |
|
"loss": 0.2296, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.4035087719298245, |
|
"grad_norm": 18.474872589111328, |
|
"learning_rate": 2.9641693811074923e-05, |
|
"loss": 0.252, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.4912280701754386, |
|
"grad_norm": 8.512451171875, |
|
"learning_rate": 2.8013029315960915e-05, |
|
"loss": 0.2526, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.5789473684210527, |
|
"grad_norm": 10.639461517333984, |
|
"learning_rate": 2.6384364820846906e-05, |
|
"loss": 0.2218, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.6666666666666665, |
|
"grad_norm": 10.84925651550293, |
|
"learning_rate": 2.47557003257329e-05, |
|
"loss": 0.2246, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.7543859649122808, |
|
"grad_norm": 7.055878162384033, |
|
"learning_rate": 2.3127035830618892e-05, |
|
"loss": 0.2181, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.8421052631578947, |
|
"grad_norm": 10.151970863342285, |
|
"learning_rate": 2.149837133550489e-05, |
|
"loss": 0.2336, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.9298245614035088, |
|
"grad_norm": 10.90556812286377, |
|
"learning_rate": 1.986970684039088e-05, |
|
"loss": 0.2175, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.9617283950617284, |
|
"eval_loss": 0.1142377108335495, |
|
"eval_runtime": 4.4649, |
|
"eval_samples_per_second": 362.828, |
|
"eval_steps_per_second": 11.422, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 2.017543859649123, |
|
"grad_norm": 7.411689281463623, |
|
"learning_rate": 1.8241042345276872e-05, |
|
"loss": 0.2493, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 2.1052631578947367, |
|
"grad_norm": 6.981323719024658, |
|
"learning_rate": 1.6612377850162867e-05, |
|
"loss": 0.1832, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 2.192982456140351, |
|
"grad_norm": 17.770326614379883, |
|
"learning_rate": 1.4983713355048862e-05, |
|
"loss": 0.1862, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 2.280701754385965, |
|
"grad_norm": 14.688613891601562, |
|
"learning_rate": 1.3355048859934855e-05, |
|
"loss": 0.1725, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 2.3684210526315788, |
|
"grad_norm": 9.546713829040527, |
|
"learning_rate": 1.1726384364820847e-05, |
|
"loss": 0.2187, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 2.456140350877193, |
|
"grad_norm": 13.742387771606445, |
|
"learning_rate": 1.009771986970684e-05, |
|
"loss": 0.211, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 2.543859649122807, |
|
"grad_norm": 10.309969902038574, |
|
"learning_rate": 8.469055374592833e-06, |
|
"loss": 0.1776, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 2.6315789473684212, |
|
"grad_norm": 11.343367576599121, |
|
"learning_rate": 6.840390879478828e-06, |
|
"loss": 0.1828, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 2.719298245614035, |
|
"grad_norm": 12.15509033203125, |
|
"learning_rate": 5.211726384364822e-06, |
|
"loss": 0.1565, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 2.807017543859649, |
|
"grad_norm": 13.258170127868652, |
|
"learning_rate": 3.5830618892508147e-06, |
|
"loss": 0.1601, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 2.8947368421052633, |
|
"grad_norm": 13.760292053222656, |
|
"learning_rate": 1.9543973941368076e-06, |
|
"loss": 0.1777, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 2.982456140350877, |
|
"grad_norm": 10.578232765197754, |
|
"learning_rate": 3.2573289902280136e-07, |
|
"loss": 0.1881, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.9691358024691358, |
|
"eval_loss": 0.08843858540058136, |
|
"eval_runtime": 4.4533, |
|
"eval_samples_per_second": 363.776, |
|
"eval_steps_per_second": 11.452, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 342, |
|
"total_flos": 1.0874439903456461e+18, |
|
"train_loss": 0.41079250579340415, |
|
"train_runtime": 248.8888, |
|
"train_samples_per_second": 175.741, |
|
"train_steps_per_second": 1.374 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 342, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.0874439903456461e+18, |
|
"train_batch_size": 32, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|