shotegni's picture
End of training
63cc48c verified
{
"best_metric": 0.969,
"best_model_checkpoint": "swin-tiny-patch4-window7-224-finetuned-eurosat/checkpoint-1053",
"epoch": 2.9936034115138592,
"eval_steps": 500,
"global_step": 1053,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.03,
"grad_norm": 5.63645601272583,
"learning_rate": 4.716981132075472e-06,
"loss": 2.3306,
"step": 10
},
{
"epoch": 0.06,
"grad_norm": 6.468899726867676,
"learning_rate": 9.433962264150944e-06,
"loss": 2.2678,
"step": 20
},
{
"epoch": 0.09,
"grad_norm": 6.23107385635376,
"learning_rate": 1.4150943396226415e-05,
"loss": 2.1477,
"step": 30
},
{
"epoch": 0.11,
"grad_norm": 6.676746368408203,
"learning_rate": 1.8867924528301888e-05,
"loss": 1.9591,
"step": 40
},
{
"epoch": 0.14,
"grad_norm": 10.20413875579834,
"learning_rate": 2.358490566037736e-05,
"loss": 1.6696,
"step": 50
},
{
"epoch": 0.17,
"grad_norm": 12.014700889587402,
"learning_rate": 2.830188679245283e-05,
"loss": 1.3329,
"step": 60
},
{
"epoch": 0.2,
"grad_norm": 13.840166091918945,
"learning_rate": 3.30188679245283e-05,
"loss": 1.1415,
"step": 70
},
{
"epoch": 0.23,
"grad_norm": 14.116910934448242,
"learning_rate": 3.7735849056603776e-05,
"loss": 1.024,
"step": 80
},
{
"epoch": 0.26,
"grad_norm": 10.790963172912598,
"learning_rate": 4.245283018867925e-05,
"loss": 0.9195,
"step": 90
},
{
"epoch": 0.28,
"grad_norm": 14.884149551391602,
"learning_rate": 4.716981132075472e-05,
"loss": 0.8209,
"step": 100
},
{
"epoch": 0.31,
"grad_norm": 14.443344116210938,
"learning_rate": 4.978880675818374e-05,
"loss": 0.7393,
"step": 110
},
{
"epoch": 0.34,
"grad_norm": 15.056041717529297,
"learning_rate": 4.9260823653643085e-05,
"loss": 0.7175,
"step": 120
},
{
"epoch": 0.37,
"grad_norm": 10.33287525177002,
"learning_rate": 4.8732840549102435e-05,
"loss": 0.7329,
"step": 130
},
{
"epoch": 0.4,
"grad_norm": 12.791647911071777,
"learning_rate": 4.820485744456177e-05,
"loss": 0.7075,
"step": 140
},
{
"epoch": 0.43,
"grad_norm": 12.029749870300293,
"learning_rate": 4.767687434002112e-05,
"loss": 0.6584,
"step": 150
},
{
"epoch": 0.45,
"grad_norm": 11.643570899963379,
"learning_rate": 4.7148891235480466e-05,
"loss": 0.6047,
"step": 160
},
{
"epoch": 0.48,
"grad_norm": 19.02515983581543,
"learning_rate": 4.662090813093981e-05,
"loss": 0.6052,
"step": 170
},
{
"epoch": 0.51,
"grad_norm": 10.113385200500488,
"learning_rate": 4.609292502639916e-05,
"loss": 0.6043,
"step": 180
},
{
"epoch": 0.54,
"grad_norm": 12.620966911315918,
"learning_rate": 4.55649419218585e-05,
"loss": 0.5803,
"step": 190
},
{
"epoch": 0.57,
"grad_norm": 8.98147201538086,
"learning_rate": 4.503695881731785e-05,
"loss": 0.5365,
"step": 200
},
{
"epoch": 0.6,
"grad_norm": 11.942453384399414,
"learning_rate": 4.45089757127772e-05,
"loss": 0.6061,
"step": 210
},
{
"epoch": 0.63,
"grad_norm": 8.406760215759277,
"learning_rate": 4.398099260823654e-05,
"loss": 0.4963,
"step": 220
},
{
"epoch": 0.65,
"grad_norm": 8.798361778259277,
"learning_rate": 4.3453009503695884e-05,
"loss": 0.552,
"step": 230
},
{
"epoch": 0.68,
"grad_norm": 8.614947319030762,
"learning_rate": 4.292502639915523e-05,
"loss": 0.5015,
"step": 240
},
{
"epoch": 0.71,
"grad_norm": 9.218803405761719,
"learning_rate": 4.239704329461457e-05,
"loss": 0.5246,
"step": 250
},
{
"epoch": 0.74,
"grad_norm": 9.763008117675781,
"learning_rate": 4.186906019007392e-05,
"loss": 0.5188,
"step": 260
},
{
"epoch": 0.77,
"grad_norm": 10.057807922363281,
"learning_rate": 4.1341077085533265e-05,
"loss": 0.4884,
"step": 270
},
{
"epoch": 0.8,
"grad_norm": 10.842576026916504,
"learning_rate": 4.081309398099261e-05,
"loss": 0.5345,
"step": 280
},
{
"epoch": 0.82,
"grad_norm": 8.137351036071777,
"learning_rate": 4.028511087645195e-05,
"loss": 0.5157,
"step": 290
},
{
"epoch": 0.85,
"grad_norm": 14.897061347961426,
"learning_rate": 3.97571277719113e-05,
"loss": 0.5212,
"step": 300
},
{
"epoch": 0.88,
"grad_norm": 11.816178321838379,
"learning_rate": 3.9229144667370646e-05,
"loss": 0.5337,
"step": 310
},
{
"epoch": 0.91,
"grad_norm": 10.712966918945312,
"learning_rate": 3.870116156282999e-05,
"loss": 0.5154,
"step": 320
},
{
"epoch": 0.94,
"grad_norm": 7.876273155212402,
"learning_rate": 3.817317845828934e-05,
"loss": 0.4833,
"step": 330
},
{
"epoch": 0.97,
"grad_norm": 14.542892456054688,
"learning_rate": 3.764519535374868e-05,
"loss": 0.5037,
"step": 340
},
{
"epoch": 1.0,
"grad_norm": 8.809218406677246,
"learning_rate": 3.711721224920803e-05,
"loss": 0.5592,
"step": 350
},
{
"epoch": 1.0,
"eval_accuracy": 0.9538,
"eval_loss": 0.14359258115291595,
"eval_runtime": 31.1413,
"eval_samples_per_second": 160.558,
"eval_steps_per_second": 5.042,
"step": 351
},
{
"epoch": 1.02,
"grad_norm": 8.733403205871582,
"learning_rate": 3.658922914466738e-05,
"loss": 0.4327,
"step": 360
},
{
"epoch": 1.05,
"grad_norm": 9.024731636047363,
"learning_rate": 3.6061246040126714e-05,
"loss": 0.4674,
"step": 370
},
{
"epoch": 1.08,
"grad_norm": 8.668274879455566,
"learning_rate": 3.5533262935586064e-05,
"loss": 0.4807,
"step": 380
},
{
"epoch": 1.11,
"grad_norm": 11.616405487060547,
"learning_rate": 3.500527983104541e-05,
"loss": 0.51,
"step": 390
},
{
"epoch": 1.14,
"grad_norm": 8.094477653503418,
"learning_rate": 3.447729672650475e-05,
"loss": 0.4959,
"step": 400
},
{
"epoch": 1.17,
"grad_norm": 11.38330364227295,
"learning_rate": 3.3949313621964095e-05,
"loss": 0.4142,
"step": 410
},
{
"epoch": 1.19,
"grad_norm": 13.148619651794434,
"learning_rate": 3.3421330517423445e-05,
"loss": 0.4217,
"step": 420
},
{
"epoch": 1.22,
"grad_norm": 20.912269592285156,
"learning_rate": 3.289334741288279e-05,
"loss": 0.4437,
"step": 430
},
{
"epoch": 1.25,
"grad_norm": 12.211991310119629,
"learning_rate": 3.236536430834213e-05,
"loss": 0.4687,
"step": 440
},
{
"epoch": 1.28,
"grad_norm": 11.126351356506348,
"learning_rate": 3.183738120380148e-05,
"loss": 0.4691,
"step": 450
},
{
"epoch": 1.31,
"grad_norm": 9.944380760192871,
"learning_rate": 3.130939809926082e-05,
"loss": 0.4593,
"step": 460
},
{
"epoch": 1.34,
"grad_norm": 10.881024360656738,
"learning_rate": 3.078141499472017e-05,
"loss": 0.4594,
"step": 470
},
{
"epoch": 1.36,
"grad_norm": 7.276984214782715,
"learning_rate": 3.0253431890179517e-05,
"loss": 0.4566,
"step": 480
},
{
"epoch": 1.39,
"grad_norm": 11.3895263671875,
"learning_rate": 2.972544878563886e-05,
"loss": 0.4061,
"step": 490
},
{
"epoch": 1.42,
"grad_norm": 8.26075267791748,
"learning_rate": 2.9197465681098207e-05,
"loss": 0.4233,
"step": 500
},
{
"epoch": 1.45,
"grad_norm": 9.681741714477539,
"learning_rate": 2.8669482576557548e-05,
"loss": 0.4564,
"step": 510
},
{
"epoch": 1.48,
"grad_norm": 12.245609283447266,
"learning_rate": 2.8141499472016898e-05,
"loss": 0.4561,
"step": 520
},
{
"epoch": 1.51,
"grad_norm": 10.530965805053711,
"learning_rate": 2.7613516367476245e-05,
"loss": 0.429,
"step": 530
},
{
"epoch": 1.54,
"grad_norm": 6.2105889320373535,
"learning_rate": 2.7085533262935585e-05,
"loss": 0.3995,
"step": 540
},
{
"epoch": 1.56,
"grad_norm": 9.41408920288086,
"learning_rate": 2.6557550158394935e-05,
"loss": 0.4159,
"step": 550
},
{
"epoch": 1.59,
"grad_norm": 10.506197929382324,
"learning_rate": 2.6029567053854276e-05,
"loss": 0.4396,
"step": 560
},
{
"epoch": 1.62,
"grad_norm": 9.805855751037598,
"learning_rate": 2.5501583949313622e-05,
"loss": 0.4258,
"step": 570
},
{
"epoch": 1.65,
"grad_norm": 6.874490261077881,
"learning_rate": 2.497360084477297e-05,
"loss": 0.4014,
"step": 580
},
{
"epoch": 1.68,
"grad_norm": 9.369732856750488,
"learning_rate": 2.4445617740232313e-05,
"loss": 0.4045,
"step": 590
},
{
"epoch": 1.71,
"grad_norm": 10.258170127868652,
"learning_rate": 2.391763463569166e-05,
"loss": 0.4133,
"step": 600
},
{
"epoch": 1.73,
"grad_norm": 9.34408950805664,
"learning_rate": 2.3389651531151003e-05,
"loss": 0.4196,
"step": 610
},
{
"epoch": 1.76,
"grad_norm": 9.928686141967773,
"learning_rate": 2.286166842661035e-05,
"loss": 0.4014,
"step": 620
},
{
"epoch": 1.79,
"grad_norm": 10.1293306350708,
"learning_rate": 2.2333685322069694e-05,
"loss": 0.427,
"step": 630
},
{
"epoch": 1.82,
"grad_norm": 10.649211883544922,
"learning_rate": 2.180570221752904e-05,
"loss": 0.3947,
"step": 640
},
{
"epoch": 1.85,
"grad_norm": 12.145565032958984,
"learning_rate": 2.1277719112988384e-05,
"loss": 0.4131,
"step": 650
},
{
"epoch": 1.88,
"grad_norm": 9.348143577575684,
"learning_rate": 2.074973600844773e-05,
"loss": 0.459,
"step": 660
},
{
"epoch": 1.9,
"grad_norm": 8.802924156188965,
"learning_rate": 2.0221752903907075e-05,
"loss": 0.3921,
"step": 670
},
{
"epoch": 1.93,
"grad_norm": 18.41295051574707,
"learning_rate": 1.9693769799366422e-05,
"loss": 0.3756,
"step": 680
},
{
"epoch": 1.96,
"grad_norm": 8.476001739501953,
"learning_rate": 1.9165786694825765e-05,
"loss": 0.3941,
"step": 690
},
{
"epoch": 1.99,
"grad_norm": 9.943327903747559,
"learning_rate": 1.863780359028511e-05,
"loss": 0.3338,
"step": 700
},
{
"epoch": 2.0,
"eval_accuracy": 0.9632,
"eval_loss": 0.11315808445215225,
"eval_runtime": 31.2355,
"eval_samples_per_second": 160.074,
"eval_steps_per_second": 5.026,
"step": 703
},
{
"epoch": 2.02,
"grad_norm": 10.112526893615723,
"learning_rate": 1.810982048574446e-05,
"loss": 0.3586,
"step": 710
},
{
"epoch": 2.05,
"grad_norm": 9.651534080505371,
"learning_rate": 1.7581837381203803e-05,
"loss": 0.3957,
"step": 720
},
{
"epoch": 2.08,
"grad_norm": 7.620846748352051,
"learning_rate": 1.7053854276663146e-05,
"loss": 0.402,
"step": 730
},
{
"epoch": 2.1,
"grad_norm": 10.414019584655762,
"learning_rate": 1.6525871172122493e-05,
"loss": 0.3809,
"step": 740
},
{
"epoch": 2.13,
"grad_norm": 9.351815223693848,
"learning_rate": 1.5997888067581837e-05,
"loss": 0.3745,
"step": 750
},
{
"epoch": 2.16,
"grad_norm": 10.798675537109375,
"learning_rate": 1.5469904963041184e-05,
"loss": 0.3507,
"step": 760
},
{
"epoch": 2.19,
"grad_norm": 10.772116661071777,
"learning_rate": 1.4941921858500529e-05,
"loss": 0.3589,
"step": 770
},
{
"epoch": 2.22,
"grad_norm": 9.006869316101074,
"learning_rate": 1.4413938753959874e-05,
"loss": 0.359,
"step": 780
},
{
"epoch": 2.25,
"grad_norm": 7.237900257110596,
"learning_rate": 1.388595564941922e-05,
"loss": 0.3636,
"step": 790
},
{
"epoch": 2.27,
"grad_norm": 8.02499008178711,
"learning_rate": 1.3357972544878563e-05,
"loss": 0.3817,
"step": 800
},
{
"epoch": 2.3,
"grad_norm": 7.373661041259766,
"learning_rate": 1.2829989440337912e-05,
"loss": 0.3812,
"step": 810
},
{
"epoch": 2.33,
"grad_norm": 9.66775131225586,
"learning_rate": 1.2302006335797255e-05,
"loss": 0.3681,
"step": 820
},
{
"epoch": 2.36,
"grad_norm": 9.115776062011719,
"learning_rate": 1.17740232312566e-05,
"loss": 0.4156,
"step": 830
},
{
"epoch": 2.39,
"grad_norm": 8.92050838470459,
"learning_rate": 1.1246040126715946e-05,
"loss": 0.3629,
"step": 840
},
{
"epoch": 2.42,
"grad_norm": 7.8812031745910645,
"learning_rate": 1.0718057022175291e-05,
"loss": 0.356,
"step": 850
},
{
"epoch": 2.44,
"grad_norm": 5.577208518981934,
"learning_rate": 1.0190073917634636e-05,
"loss": 0.3653,
"step": 860
},
{
"epoch": 2.47,
"grad_norm": 9.963196754455566,
"learning_rate": 9.662090813093982e-06,
"loss": 0.3648,
"step": 870
},
{
"epoch": 2.5,
"grad_norm": 10.441061973571777,
"learning_rate": 9.134107708553327e-06,
"loss": 0.3671,
"step": 880
},
{
"epoch": 2.53,
"grad_norm": 11.808416366577148,
"learning_rate": 8.606124604012672e-06,
"loss": 0.3012,
"step": 890
},
{
"epoch": 2.56,
"grad_norm": 7.2837138175964355,
"learning_rate": 8.078141499472017e-06,
"loss": 0.4002,
"step": 900
},
{
"epoch": 2.59,
"grad_norm": 7.841160297393799,
"learning_rate": 7.5501583949313625e-06,
"loss": 0.424,
"step": 910
},
{
"epoch": 2.62,
"grad_norm": 11.717653274536133,
"learning_rate": 7.022175290390708e-06,
"loss": 0.3432,
"step": 920
},
{
"epoch": 2.64,
"grad_norm": 7.797212600708008,
"learning_rate": 6.494192185850054e-06,
"loss": 0.3618,
"step": 930
},
{
"epoch": 2.67,
"grad_norm": 9.33782958984375,
"learning_rate": 5.966209081309398e-06,
"loss": 0.3559,
"step": 940
},
{
"epoch": 2.7,
"grad_norm": 8.307150840759277,
"learning_rate": 5.438225976768744e-06,
"loss": 0.3601,
"step": 950
},
{
"epoch": 2.73,
"grad_norm": 8.53368854522705,
"learning_rate": 4.910242872228089e-06,
"loss": 0.3319,
"step": 960
},
{
"epoch": 2.76,
"grad_norm": 8.487812042236328,
"learning_rate": 4.382259767687434e-06,
"loss": 0.3331,
"step": 970
},
{
"epoch": 2.79,
"grad_norm": 10.636713027954102,
"learning_rate": 3.854276663146779e-06,
"loss": 0.359,
"step": 980
},
{
"epoch": 2.81,
"grad_norm": 13.220087051391602,
"learning_rate": 3.326293558606125e-06,
"loss": 0.3598,
"step": 990
},
{
"epoch": 2.84,
"grad_norm": 7.285949230194092,
"learning_rate": 2.79831045406547e-06,
"loss": 0.3503,
"step": 1000
},
{
"epoch": 2.87,
"grad_norm": 8.310001373291016,
"learning_rate": 2.2703273495248154e-06,
"loss": 0.3566,
"step": 1010
},
{
"epoch": 2.9,
"grad_norm": 9.105748176574707,
"learning_rate": 1.7423442449841606e-06,
"loss": 0.3345,
"step": 1020
},
{
"epoch": 2.93,
"grad_norm": 7.4982008934021,
"learning_rate": 1.2143611404435059e-06,
"loss": 0.3398,
"step": 1030
},
{
"epoch": 2.96,
"grad_norm": 11.533191680908203,
"learning_rate": 6.863780359028511e-07,
"loss": 0.3275,
"step": 1040
},
{
"epoch": 2.99,
"grad_norm": 7.991899013519287,
"learning_rate": 1.5839493136219642e-07,
"loss": 0.3196,
"step": 1050
},
{
"epoch": 2.99,
"eval_accuracy": 0.969,
"eval_loss": 0.09590400755405426,
"eval_runtime": 31.5821,
"eval_samples_per_second": 158.318,
"eval_steps_per_second": 4.971,
"step": 1053
},
{
"epoch": 2.99,
"step": 1053,
"total_flos": 3.3497451642252165e+18,
"train_loss": 0.5486761254009924,
"train_runtime": 2094.2938,
"train_samples_per_second": 64.461,
"train_steps_per_second": 0.503
}
],
"logging_steps": 10,
"max_steps": 1053,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"total_flos": 3.3497451642252165e+18,
"train_batch_size": 32,
"trial_name": null,
"trial_params": null
}