suredream's picture
End of training
244762e verified
raw
history blame
7.66 kB
{
"best_metric": 0.9691358024691358,
"best_model_checkpoint": "swin-tiny-patch4-window7-224-finetuned-eurosat/checkpoint-342",
"epoch": 3.0,
"eval_steps": 500,
"global_step": 342,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.08771929824561403,
"grad_norm": 4.784806251525879,
"learning_rate": 1.4285714285714285e-05,
"loss": 2.2537,
"step": 10
},
{
"epoch": 0.17543859649122806,
"grad_norm": 7.325711727142334,
"learning_rate": 2.857142857142857e-05,
"loss": 1.9328,
"step": 20
},
{
"epoch": 0.2631578947368421,
"grad_norm": 12.205706596374512,
"learning_rate": 4.2857142857142856e-05,
"loss": 1.3033,
"step": 30
},
{
"epoch": 0.3508771929824561,
"grad_norm": 21.546316146850586,
"learning_rate": 4.9185667752443e-05,
"loss": 0.7435,
"step": 40
},
{
"epoch": 0.43859649122807015,
"grad_norm": 11.416431427001953,
"learning_rate": 4.755700325732899e-05,
"loss": 0.5723,
"step": 50
},
{
"epoch": 0.5263157894736842,
"grad_norm": 13.339532852172852,
"learning_rate": 4.592833876221499e-05,
"loss": 0.4688,
"step": 60
},
{
"epoch": 0.6140350877192983,
"grad_norm": 18.226308822631836,
"learning_rate": 4.429967426710098e-05,
"loss": 0.4296,
"step": 70
},
{
"epoch": 0.7017543859649122,
"grad_norm": 24.142580032348633,
"learning_rate": 4.2671009771986977e-05,
"loss": 0.3557,
"step": 80
},
{
"epoch": 0.7894736842105263,
"grad_norm": 18.747182846069336,
"learning_rate": 4.104234527687297e-05,
"loss": 0.3364,
"step": 90
},
{
"epoch": 0.8771929824561403,
"grad_norm": 9.604598999023438,
"learning_rate": 3.941368078175896e-05,
"loss": 0.379,
"step": 100
},
{
"epoch": 0.9649122807017544,
"grad_norm": 9.6405668258667,
"learning_rate": 3.778501628664495e-05,
"loss": 0.2967,
"step": 110
},
{
"epoch": 1.0,
"eval_accuracy": 0.9271604938271605,
"eval_loss": 0.20734398066997528,
"eval_runtime": 4.5343,
"eval_samples_per_second": 357.28,
"eval_steps_per_second": 11.248,
"step": 114
},
{
"epoch": 1.0526315789473684,
"grad_norm": 10.606453895568848,
"learning_rate": 3.615635179153095e-05,
"loss": 0.2506,
"step": 120
},
{
"epoch": 1.1403508771929824,
"grad_norm": 13.219770431518555,
"learning_rate": 3.452768729641694e-05,
"loss": 0.2646,
"step": 130
},
{
"epoch": 1.2280701754385965,
"grad_norm": 14.424461364746094,
"learning_rate": 3.289902280130293e-05,
"loss": 0.3161,
"step": 140
},
{
"epoch": 1.3157894736842106,
"grad_norm": 16.468141555786133,
"learning_rate": 3.127035830618892e-05,
"loss": 0.2296,
"step": 150
},
{
"epoch": 1.4035087719298245,
"grad_norm": 18.474872589111328,
"learning_rate": 2.9641693811074923e-05,
"loss": 0.252,
"step": 160
},
{
"epoch": 1.4912280701754386,
"grad_norm": 8.512451171875,
"learning_rate": 2.8013029315960915e-05,
"loss": 0.2526,
"step": 170
},
{
"epoch": 1.5789473684210527,
"grad_norm": 10.639461517333984,
"learning_rate": 2.6384364820846906e-05,
"loss": 0.2218,
"step": 180
},
{
"epoch": 1.6666666666666665,
"grad_norm": 10.84925651550293,
"learning_rate": 2.47557003257329e-05,
"loss": 0.2246,
"step": 190
},
{
"epoch": 1.7543859649122808,
"grad_norm": 7.055878162384033,
"learning_rate": 2.3127035830618892e-05,
"loss": 0.2181,
"step": 200
},
{
"epoch": 1.8421052631578947,
"grad_norm": 10.151970863342285,
"learning_rate": 2.149837133550489e-05,
"loss": 0.2336,
"step": 210
},
{
"epoch": 1.9298245614035088,
"grad_norm": 10.90556812286377,
"learning_rate": 1.986970684039088e-05,
"loss": 0.2175,
"step": 220
},
{
"epoch": 2.0,
"eval_accuracy": 0.9617283950617284,
"eval_loss": 0.1142377108335495,
"eval_runtime": 4.4649,
"eval_samples_per_second": 362.828,
"eval_steps_per_second": 11.422,
"step": 228
},
{
"epoch": 2.017543859649123,
"grad_norm": 7.411689281463623,
"learning_rate": 1.8241042345276872e-05,
"loss": 0.2493,
"step": 230
},
{
"epoch": 2.1052631578947367,
"grad_norm": 6.981323719024658,
"learning_rate": 1.6612377850162867e-05,
"loss": 0.1832,
"step": 240
},
{
"epoch": 2.192982456140351,
"grad_norm": 17.770326614379883,
"learning_rate": 1.4983713355048862e-05,
"loss": 0.1862,
"step": 250
},
{
"epoch": 2.280701754385965,
"grad_norm": 14.688613891601562,
"learning_rate": 1.3355048859934855e-05,
"loss": 0.1725,
"step": 260
},
{
"epoch": 2.3684210526315788,
"grad_norm": 9.546713829040527,
"learning_rate": 1.1726384364820847e-05,
"loss": 0.2187,
"step": 270
},
{
"epoch": 2.456140350877193,
"grad_norm": 13.742387771606445,
"learning_rate": 1.009771986970684e-05,
"loss": 0.211,
"step": 280
},
{
"epoch": 2.543859649122807,
"grad_norm": 10.309969902038574,
"learning_rate": 8.469055374592833e-06,
"loss": 0.1776,
"step": 290
},
{
"epoch": 2.6315789473684212,
"grad_norm": 11.343367576599121,
"learning_rate": 6.840390879478828e-06,
"loss": 0.1828,
"step": 300
},
{
"epoch": 2.719298245614035,
"grad_norm": 12.15509033203125,
"learning_rate": 5.211726384364822e-06,
"loss": 0.1565,
"step": 310
},
{
"epoch": 2.807017543859649,
"grad_norm": 13.258170127868652,
"learning_rate": 3.5830618892508147e-06,
"loss": 0.1601,
"step": 320
},
{
"epoch": 2.8947368421052633,
"grad_norm": 13.760292053222656,
"learning_rate": 1.9543973941368076e-06,
"loss": 0.1777,
"step": 330
},
{
"epoch": 2.982456140350877,
"grad_norm": 10.578232765197754,
"learning_rate": 3.2573289902280136e-07,
"loss": 0.1881,
"step": 340
},
{
"epoch": 3.0,
"eval_accuracy": 0.9691358024691358,
"eval_loss": 0.08843858540058136,
"eval_runtime": 4.4533,
"eval_samples_per_second": 363.776,
"eval_steps_per_second": 11.452,
"step": 342
},
{
"epoch": 3.0,
"step": 342,
"total_flos": 1.0874439903456461e+18,
"train_loss": 0.41079250579340415,
"train_runtime": 248.8888,
"train_samples_per_second": 175.741,
"train_steps_per_second": 1.374
}
],
"logging_steps": 10,
"max_steps": 342,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.0874439903456461e+18,
"train_batch_size": 32,
"trial_name": null,
"trial_params": null
}