pylu5229's picture
End of training
2b5cc26 verified
{
"best_metric": 1.0,
"best_model_checkpoint": "swin-tiny-patch4-window7-224-finetuned-eurosat/checkpoint-212",
"epoch": 2.976303317535545,
"eval_steps": 500,
"global_step": 315,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0947867298578199,
"grad_norm": 7.812982082366943,
"learning_rate": 1.5625e-05,
"loss": 6.4564,
"step": 10
},
{
"epoch": 0.1895734597156398,
"grad_norm": 8.798693656921387,
"learning_rate": 3.125e-05,
"loss": 6.366,
"step": 20
},
{
"epoch": 0.2843601895734597,
"grad_norm": 36.50003433227539,
"learning_rate": 4.6875e-05,
"loss": 5.9633,
"step": 30
},
{
"epoch": 0.3791469194312796,
"grad_norm": 34.34447479248047,
"learning_rate": 4.858657243816255e-05,
"loss": 4.1076,
"step": 40
},
{
"epoch": 0.47393364928909953,
"grad_norm": 188.84632873535156,
"learning_rate": 4.6819787985865726e-05,
"loss": 2.403,
"step": 50
},
{
"epoch": 0.5687203791469194,
"grad_norm": 55.603084564208984,
"learning_rate": 4.50530035335689e-05,
"loss": 1.5914,
"step": 60
},
{
"epoch": 0.6635071090047393,
"grad_norm": 48.88166427612305,
"learning_rate": 4.328621908127209e-05,
"loss": 0.8822,
"step": 70
},
{
"epoch": 0.7582938388625592,
"grad_norm": 34.93512725830078,
"learning_rate": 4.1519434628975266e-05,
"loss": 0.6805,
"step": 80
},
{
"epoch": 0.8530805687203792,
"grad_norm": 42.45512771606445,
"learning_rate": 3.975265017667845e-05,
"loss": 0.4599,
"step": 90
},
{
"epoch": 0.9478672985781991,
"grad_norm": 25.569683074951172,
"learning_rate": 3.7985865724381624e-05,
"loss": 0.3571,
"step": 100
},
{
"epoch": 1.0,
"eval_accuracy": 0.992,
"eval_loss": 0.046910375356674194,
"eval_runtime": 14.0627,
"eval_samples_per_second": 106.665,
"eval_steps_per_second": 3.342,
"step": 106
},
{
"epoch": 1.037914691943128,
"grad_norm": 22.71461296081543,
"learning_rate": 3.621908127208481e-05,
"loss": 0.339,
"step": 110
},
{
"epoch": 1.132701421800948,
"grad_norm": 39.072593688964844,
"learning_rate": 3.445229681978799e-05,
"loss": 0.674,
"step": 120
},
{
"epoch": 1.2274881516587677,
"grad_norm": 28.356447219848633,
"learning_rate": 3.2685512367491163e-05,
"loss": 0.4667,
"step": 130
},
{
"epoch": 1.3222748815165877,
"grad_norm": 27.163755416870117,
"learning_rate": 3.0918727915194346e-05,
"loss": 0.3833,
"step": 140
},
{
"epoch": 1.4170616113744074,
"grad_norm": 53.821319580078125,
"learning_rate": 2.915194346289753e-05,
"loss": 0.2353,
"step": 150
},
{
"epoch": 1.5118483412322274,
"grad_norm": 15.961763381958008,
"learning_rate": 2.738515901060071e-05,
"loss": 0.2735,
"step": 160
},
{
"epoch": 1.6066350710900474,
"grad_norm": 23.296354293823242,
"learning_rate": 2.5618374558303885e-05,
"loss": 0.2108,
"step": 170
},
{
"epoch": 1.7014218009478674,
"grad_norm": 14.222640037536621,
"learning_rate": 2.3851590106007068e-05,
"loss": 0.2073,
"step": 180
},
{
"epoch": 1.7962085308056872,
"grad_norm": 17.987783432006836,
"learning_rate": 2.2084805653710246e-05,
"loss": 0.1279,
"step": 190
},
{
"epoch": 1.890995260663507,
"grad_norm": 34.04901123046875,
"learning_rate": 2.031802120141343e-05,
"loss": 0.2281,
"step": 200
},
{
"epoch": 1.985781990521327,
"grad_norm": 13.81363296508789,
"learning_rate": 1.855123674911661e-05,
"loss": 0.1158,
"step": 210
},
{
"epoch": 2.0,
"eval_accuracy": 1.0,
"eval_loss": 0.001326344907283783,
"eval_runtime": 14.0164,
"eval_samples_per_second": 107.018,
"eval_steps_per_second": 3.353,
"step": 212
},
{
"epoch": 2.075829383886256,
"grad_norm": 41.3836784362793,
"learning_rate": 1.678445229681979e-05,
"loss": 0.1724,
"step": 220
},
{
"epoch": 2.170616113744076,
"grad_norm": 25.138639450073242,
"learning_rate": 1.501766784452297e-05,
"loss": 0.1802,
"step": 230
},
{
"epoch": 2.265402843601896,
"grad_norm": 5.671016216278076,
"learning_rate": 1.3250883392226149e-05,
"loss": 0.1261,
"step": 240
},
{
"epoch": 2.360189573459716,
"grad_norm": 32.47148895263672,
"learning_rate": 1.148409893992933e-05,
"loss": 0.1031,
"step": 250
},
{
"epoch": 2.4549763033175354,
"grad_norm": 22.735076904296875,
"learning_rate": 9.717314487632508e-06,
"loss": 0.1143,
"step": 260
},
{
"epoch": 2.5497630331753554,
"grad_norm": 20.33681297302246,
"learning_rate": 7.950530035335689e-06,
"loss": 0.2151,
"step": 270
},
{
"epoch": 2.6445497630331753,
"grad_norm": 31.386720657348633,
"learning_rate": 6.18374558303887e-06,
"loss": 0.1495,
"step": 280
},
{
"epoch": 2.7393364928909953,
"grad_norm": 0.9427671432495117,
"learning_rate": 4.41696113074205e-06,
"loss": 0.1053,
"step": 290
},
{
"epoch": 2.834123222748815,
"grad_norm": 12.637048721313477,
"learning_rate": 2.65017667844523e-06,
"loss": 0.089,
"step": 300
},
{
"epoch": 2.9289099526066353,
"grad_norm": 24.29412841796875,
"learning_rate": 8.8339222614841e-07,
"loss": 0.0934,
"step": 310
},
{
"epoch": 2.976303317535545,
"eval_accuracy": 1.0,
"eval_loss": 0.0002889372408390045,
"eval_runtime": 13.4256,
"eval_samples_per_second": 111.727,
"eval_steps_per_second": 3.501,
"step": 315
},
{
"epoch": 2.976303317535545,
"step": 315,
"total_flos": 9.988967891346555e+17,
"train_loss": 1.077162364172557,
"train_runtime": 1497.697,
"train_samples_per_second": 27.042,
"train_steps_per_second": 0.21
}
],
"logging_steps": 10,
"max_steps": 315,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 9.988967891346555e+17,
"train_batch_size": 32,
"trial_name": null,
"trial_params": null
}