|
{ |
|
"best_metric": 1.0, |
|
"best_model_checkpoint": "swin-tiny-patch4-window7-224-finetuned-eurosat/checkpoint-212", |
|
"epoch": 2.976303317535545, |
|
"eval_steps": 500, |
|
"global_step": 315, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0947867298578199, |
|
"grad_norm": 7.812982082366943, |
|
"learning_rate": 1.5625e-05, |
|
"loss": 6.4564, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.1895734597156398, |
|
"grad_norm": 8.798693656921387, |
|
"learning_rate": 3.125e-05, |
|
"loss": 6.366, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.2843601895734597, |
|
"grad_norm": 36.50003433227539, |
|
"learning_rate": 4.6875e-05, |
|
"loss": 5.9633, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.3791469194312796, |
|
"grad_norm": 34.34447479248047, |
|
"learning_rate": 4.858657243816255e-05, |
|
"loss": 4.1076, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.47393364928909953, |
|
"grad_norm": 188.84632873535156, |
|
"learning_rate": 4.6819787985865726e-05, |
|
"loss": 2.403, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.5687203791469194, |
|
"grad_norm": 55.603084564208984, |
|
"learning_rate": 4.50530035335689e-05, |
|
"loss": 1.5914, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.6635071090047393, |
|
"grad_norm": 48.88166427612305, |
|
"learning_rate": 4.328621908127209e-05, |
|
"loss": 0.8822, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.7582938388625592, |
|
"grad_norm": 34.93512725830078, |
|
"learning_rate": 4.1519434628975266e-05, |
|
"loss": 0.6805, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.8530805687203792, |
|
"grad_norm": 42.45512771606445, |
|
"learning_rate": 3.975265017667845e-05, |
|
"loss": 0.4599, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.9478672985781991, |
|
"grad_norm": 25.569683074951172, |
|
"learning_rate": 3.7985865724381624e-05, |
|
"loss": 0.3571, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.992, |
|
"eval_loss": 0.046910375356674194, |
|
"eval_runtime": 14.0627, |
|
"eval_samples_per_second": 106.665, |
|
"eval_steps_per_second": 3.342, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 1.037914691943128, |
|
"grad_norm": 22.71461296081543, |
|
"learning_rate": 3.621908127208481e-05, |
|
"loss": 0.339, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.132701421800948, |
|
"grad_norm": 39.072593688964844, |
|
"learning_rate": 3.445229681978799e-05, |
|
"loss": 0.674, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.2274881516587677, |
|
"grad_norm": 28.356447219848633, |
|
"learning_rate": 3.2685512367491163e-05, |
|
"loss": 0.4667, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.3222748815165877, |
|
"grad_norm": 27.163755416870117, |
|
"learning_rate": 3.0918727915194346e-05, |
|
"loss": 0.3833, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.4170616113744074, |
|
"grad_norm": 53.821319580078125, |
|
"learning_rate": 2.915194346289753e-05, |
|
"loss": 0.2353, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.5118483412322274, |
|
"grad_norm": 15.961763381958008, |
|
"learning_rate": 2.738515901060071e-05, |
|
"loss": 0.2735, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.6066350710900474, |
|
"grad_norm": 23.296354293823242, |
|
"learning_rate": 2.5618374558303885e-05, |
|
"loss": 0.2108, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.7014218009478674, |
|
"grad_norm": 14.222640037536621, |
|
"learning_rate": 2.3851590106007068e-05, |
|
"loss": 0.2073, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.7962085308056872, |
|
"grad_norm": 17.987783432006836, |
|
"learning_rate": 2.2084805653710246e-05, |
|
"loss": 0.1279, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.890995260663507, |
|
"grad_norm": 34.04901123046875, |
|
"learning_rate": 2.031802120141343e-05, |
|
"loss": 0.2281, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.985781990521327, |
|
"grad_norm": 13.81363296508789, |
|
"learning_rate": 1.855123674911661e-05, |
|
"loss": 0.1158, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 1.0, |
|
"eval_loss": 0.001326344907283783, |
|
"eval_runtime": 14.0164, |
|
"eval_samples_per_second": 107.018, |
|
"eval_steps_per_second": 3.353, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 2.075829383886256, |
|
"grad_norm": 41.3836784362793, |
|
"learning_rate": 1.678445229681979e-05, |
|
"loss": 0.1724, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 2.170616113744076, |
|
"grad_norm": 25.138639450073242, |
|
"learning_rate": 1.501766784452297e-05, |
|
"loss": 0.1802, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 2.265402843601896, |
|
"grad_norm": 5.671016216278076, |
|
"learning_rate": 1.3250883392226149e-05, |
|
"loss": 0.1261, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 2.360189573459716, |
|
"grad_norm": 32.47148895263672, |
|
"learning_rate": 1.148409893992933e-05, |
|
"loss": 0.1031, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 2.4549763033175354, |
|
"grad_norm": 22.735076904296875, |
|
"learning_rate": 9.717314487632508e-06, |
|
"loss": 0.1143, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 2.5497630331753554, |
|
"grad_norm": 20.33681297302246, |
|
"learning_rate": 7.950530035335689e-06, |
|
"loss": 0.2151, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 2.6445497630331753, |
|
"grad_norm": 31.386720657348633, |
|
"learning_rate": 6.18374558303887e-06, |
|
"loss": 0.1495, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 2.7393364928909953, |
|
"grad_norm": 0.9427671432495117, |
|
"learning_rate": 4.41696113074205e-06, |
|
"loss": 0.1053, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 2.834123222748815, |
|
"grad_norm": 12.637048721313477, |
|
"learning_rate": 2.65017667844523e-06, |
|
"loss": 0.089, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 2.9289099526066353, |
|
"grad_norm": 24.29412841796875, |
|
"learning_rate": 8.8339222614841e-07, |
|
"loss": 0.0934, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 2.976303317535545, |
|
"eval_accuracy": 1.0, |
|
"eval_loss": 0.0002889372408390045, |
|
"eval_runtime": 13.4256, |
|
"eval_samples_per_second": 111.727, |
|
"eval_steps_per_second": 3.501, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 2.976303317535545, |
|
"step": 315, |
|
"total_flos": 9.988967891346555e+17, |
|
"train_loss": 1.077162364172557, |
|
"train_runtime": 1497.697, |
|
"train_samples_per_second": 27.042, |
|
"train_steps_per_second": 0.21 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 315, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 9.988967891346555e+17, |
|
"train_batch_size": 32, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|