c14kevincardenas's picture
End of training
28f1cb3 verified
{
"best_metric": 0.9358813166618347,
"best_model_checkpoint": "/media/mldrive/kcardenas/limb_classification_person_crop/beit-large-patch16-384/8_1e-5_5e-4_0.15/checkpoint-1926",
"epoch": 10.0,
"eval_steps": 500,
"global_step": 2140,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.11682242990654206,
"grad_norm": 1232665.625,
"learning_rate": 5.000000000000001e-07,
"loss": 1.603,
"step": 25
},
{
"epoch": 0.2336448598130841,
"grad_norm": 1125859.125,
"learning_rate": 1.0000000000000002e-06,
"loss": 1.5674,
"step": 50
},
{
"epoch": 0.35046728971962615,
"grad_norm": 1237422.625,
"learning_rate": 1.5e-06,
"loss": 1.5855,
"step": 75
},
{
"epoch": 0.4672897196261682,
"grad_norm": 1008492.875,
"learning_rate": 2.0000000000000003e-06,
"loss": 1.5179,
"step": 100
},
{
"epoch": 0.5841121495327103,
"grad_norm": 825557.9375,
"learning_rate": 2.5e-06,
"loss": 1.4596,
"step": 125
},
{
"epoch": 0.7009345794392523,
"grad_norm": 461811.875,
"learning_rate": 3e-06,
"loss": 1.4276,
"step": 150
},
{
"epoch": 0.8177570093457944,
"grad_norm": 785723.125,
"learning_rate": 3.5e-06,
"loss": 1.3977,
"step": 175
},
{
"epoch": 0.9345794392523364,
"grad_norm": 589829.1875,
"learning_rate": 4.000000000000001e-06,
"loss": 1.4012,
"step": 200
},
{
"epoch": 1.0,
"eval_accuracy": 0.2777777777777778,
"eval_loss": 1.3970400094985962,
"eval_runtime": 14.9894,
"eval_samples_per_second": 80.457,
"eval_steps_per_second": 2.535,
"step": 214
},
{
"epoch": 1.0514018691588785,
"grad_norm": 396606.21875,
"learning_rate": 4.5e-06,
"loss": 1.3881,
"step": 225
},
{
"epoch": 1.1682242990654206,
"grad_norm": 353690.125,
"learning_rate": 5e-06,
"loss": 1.3671,
"step": 250
},
{
"epoch": 1.2850467289719627,
"grad_norm": 419103.875,
"learning_rate": 5.500000000000001e-06,
"loss": 1.3582,
"step": 275
},
{
"epoch": 1.4018691588785046,
"grad_norm": 374088.375,
"learning_rate": 6e-06,
"loss": 1.3592,
"step": 300
},
{
"epoch": 1.5186915887850467,
"grad_norm": 600778.9375,
"learning_rate": 6.5000000000000004e-06,
"loss": 1.3563,
"step": 325
},
{
"epoch": 1.6355140186915889,
"grad_norm": 5453174.5,
"learning_rate": 7e-06,
"loss": 1.3424,
"step": 350
},
{
"epoch": 1.7523364485981308,
"grad_norm": 657756.375,
"learning_rate": 7.500000000000001e-06,
"loss": 1.355,
"step": 375
},
{
"epoch": 1.8691588785046729,
"grad_norm": 560229.8125,
"learning_rate": 8.000000000000001e-06,
"loss": 1.3133,
"step": 400
},
{
"epoch": 1.985981308411215,
"grad_norm": 596723.375,
"learning_rate": 8.5e-06,
"loss": 1.2956,
"step": 425
},
{
"epoch": 2.0,
"eval_accuracy": 0.4709784411276949,
"eval_loss": 1.2437535524368286,
"eval_runtime": 14.8418,
"eval_samples_per_second": 81.257,
"eval_steps_per_second": 2.56,
"step": 428
},
{
"epoch": 2.102803738317757,
"grad_norm": 661517.125,
"learning_rate": 9e-06,
"loss": 1.278,
"step": 450
},
{
"epoch": 2.2196261682242993,
"grad_norm": 499854.875,
"learning_rate": 9.5e-06,
"loss": 1.2474,
"step": 475
},
{
"epoch": 2.336448598130841,
"grad_norm": 353308.375,
"learning_rate": 1e-05,
"loss": 1.2492,
"step": 500
},
{
"epoch": 2.453271028037383,
"grad_norm": 568008.8125,
"learning_rate": 9.847560975609756e-06,
"loss": 1.2333,
"step": 525
},
{
"epoch": 2.5700934579439254,
"grad_norm": 377403.0625,
"learning_rate": 9.695121951219513e-06,
"loss": 1.2086,
"step": 550
},
{
"epoch": 2.6869158878504673,
"grad_norm": 376984.78125,
"learning_rate": 9.542682926829268e-06,
"loss": 1.2084,
"step": 575
},
{
"epoch": 2.803738317757009,
"grad_norm": 424205.8125,
"learning_rate": 9.390243902439025e-06,
"loss": 1.2112,
"step": 600
},
{
"epoch": 2.9205607476635516,
"grad_norm": 518474.75,
"learning_rate": 9.237804878048782e-06,
"loss": 1.1902,
"step": 625
},
{
"epoch": 3.0,
"eval_accuracy": 0.5895522388059702,
"eval_loss": 1.0810370445251465,
"eval_runtime": 15.4574,
"eval_samples_per_second": 78.021,
"eval_steps_per_second": 2.458,
"step": 642
},
{
"epoch": 3.0373831775700935,
"grad_norm": 363418.0625,
"learning_rate": 9.085365853658538e-06,
"loss": 1.1903,
"step": 650
},
{
"epoch": 3.1542056074766354,
"grad_norm": 472692.21875,
"learning_rate": 8.932926829268293e-06,
"loss": 1.1369,
"step": 675
},
{
"epoch": 3.2710280373831777,
"grad_norm": 389273.5,
"learning_rate": 8.78048780487805e-06,
"loss": 1.1591,
"step": 700
},
{
"epoch": 3.3878504672897196,
"grad_norm": 536079.4375,
"learning_rate": 8.628048780487805e-06,
"loss": 1.1742,
"step": 725
},
{
"epoch": 3.5046728971962615,
"grad_norm": 369198.15625,
"learning_rate": 8.475609756097562e-06,
"loss": 1.157,
"step": 750
},
{
"epoch": 3.621495327102804,
"grad_norm": 459633.09375,
"learning_rate": 8.323170731707317e-06,
"loss": 1.1494,
"step": 775
},
{
"epoch": 3.7383177570093458,
"grad_norm": 457643.6875,
"learning_rate": 8.170731707317073e-06,
"loss": 1.1559,
"step": 800
},
{
"epoch": 3.8551401869158877,
"grad_norm": 459124.59375,
"learning_rate": 8.01829268292683e-06,
"loss": 1.1376,
"step": 825
},
{
"epoch": 3.97196261682243,
"grad_norm": 416160.96875,
"learning_rate": 7.865853658536587e-06,
"loss": 1.1316,
"step": 850
},
{
"epoch": 4.0,
"eval_accuracy": 0.6144278606965174,
"eval_loss": 1.085984230041504,
"eval_runtime": 15.1709,
"eval_samples_per_second": 79.494,
"eval_steps_per_second": 2.505,
"step": 856
},
{
"epoch": 4.088785046728972,
"grad_norm": 415968.6875,
"learning_rate": 7.713414634146342e-06,
"loss": 1.1317,
"step": 875
},
{
"epoch": 4.205607476635514,
"grad_norm": 417537.21875,
"learning_rate": 7.560975609756098e-06,
"loss": 1.1403,
"step": 900
},
{
"epoch": 4.322429906542056,
"grad_norm": 560744.375,
"learning_rate": 7.408536585365854e-06,
"loss": 1.1016,
"step": 925
},
{
"epoch": 4.4392523364485985,
"grad_norm": 597802.5,
"learning_rate": 7.25609756097561e-06,
"loss": 1.0959,
"step": 950
},
{
"epoch": 4.55607476635514,
"grad_norm": 442672.09375,
"learning_rate": 7.1036585365853665e-06,
"loss": 1.103,
"step": 975
},
{
"epoch": 4.672897196261682,
"grad_norm": 427007.0625,
"learning_rate": 6.951219512195122e-06,
"loss": 1.092,
"step": 1000
},
{
"epoch": 4.789719626168225,
"grad_norm": 545753.3125,
"learning_rate": 6.798780487804879e-06,
"loss": 1.0949,
"step": 1025
},
{
"epoch": 4.906542056074766,
"grad_norm": 332691.28125,
"learning_rate": 6.646341463414635e-06,
"loss": 1.1133,
"step": 1050
},
{
"epoch": 5.0,
"eval_accuracy": 0.6475953565505804,
"eval_loss": 1.0152363777160645,
"eval_runtime": 15.3737,
"eval_samples_per_second": 78.446,
"eval_steps_per_second": 2.472,
"step": 1070
},
{
"epoch": 5.0233644859813085,
"grad_norm": 608418.625,
"learning_rate": 6.493902439024391e-06,
"loss": 1.1168,
"step": 1075
},
{
"epoch": 5.140186915887851,
"grad_norm": 500188.65625,
"learning_rate": 6.341463414634147e-06,
"loss": 1.0857,
"step": 1100
},
{
"epoch": 5.257009345794392,
"grad_norm": 472268.6875,
"learning_rate": 6.189024390243903e-06,
"loss": 1.0784,
"step": 1125
},
{
"epoch": 5.373831775700935,
"grad_norm": 512482.09375,
"learning_rate": 6.0365853658536585e-06,
"loss": 1.0912,
"step": 1150
},
{
"epoch": 5.490654205607477,
"grad_norm": 553851.8125,
"learning_rate": 5.884146341463415e-06,
"loss": 1.0615,
"step": 1175
},
{
"epoch": 5.607476635514018,
"grad_norm": 362454.03125,
"learning_rate": 5.731707317073171e-06,
"loss": 1.0757,
"step": 1200
},
{
"epoch": 5.724299065420561,
"grad_norm": 439604.0625,
"learning_rate": 5.579268292682928e-06,
"loss": 1.1072,
"step": 1225
},
{
"epoch": 5.841121495327103,
"grad_norm": 363162.625,
"learning_rate": 5.426829268292684e-06,
"loss": 1.0625,
"step": 1250
},
{
"epoch": 5.957943925233645,
"grad_norm": 481210.59375,
"learning_rate": 5.27439024390244e-06,
"loss": 1.0626,
"step": 1275
},
{
"epoch": 6.0,
"eval_accuracy": 0.6890547263681592,
"eval_loss": 0.9663113355636597,
"eval_runtime": 14.7419,
"eval_samples_per_second": 81.807,
"eval_steps_per_second": 2.578,
"step": 1284
},
{
"epoch": 6.074766355140187,
"grad_norm": 537217.875,
"learning_rate": 5.121951219512195e-06,
"loss": 1.0343,
"step": 1300
},
{
"epoch": 6.191588785046729,
"grad_norm": 622598.75,
"learning_rate": 4.9695121951219515e-06,
"loss": 1.0544,
"step": 1325
},
{
"epoch": 6.308411214953271,
"grad_norm": 461880.90625,
"learning_rate": 4.817073170731708e-06,
"loss": 1.0571,
"step": 1350
},
{
"epoch": 6.425233644859813,
"grad_norm": 483674.78125,
"learning_rate": 4.664634146341464e-06,
"loss": 1.0273,
"step": 1375
},
{
"epoch": 6.542056074766355,
"grad_norm": 507309.1875,
"learning_rate": 4.51219512195122e-06,
"loss": 1.0243,
"step": 1400
},
{
"epoch": 6.658878504672897,
"grad_norm": 572243.0,
"learning_rate": 4.359756097560976e-06,
"loss": 1.0237,
"step": 1425
},
{
"epoch": 6.775700934579439,
"grad_norm": 447492.46875,
"learning_rate": 4.207317073170732e-06,
"loss": 1.0387,
"step": 1450
},
{
"epoch": 6.892523364485982,
"grad_norm": 471134.8125,
"learning_rate": 4.054878048780488e-06,
"loss": 1.0039,
"step": 1475
},
{
"epoch": 7.0,
"eval_accuracy": 0.6923714759535655,
"eval_loss": 0.9811031818389893,
"eval_runtime": 14.7461,
"eval_samples_per_second": 81.784,
"eval_steps_per_second": 2.577,
"step": 1498
},
{
"epoch": 7.009345794392523,
"grad_norm": 484021.0,
"learning_rate": 3.902439024390244e-06,
"loss": 1.0382,
"step": 1500
},
{
"epoch": 7.126168224299065,
"grad_norm": 545706.5,
"learning_rate": 3.7500000000000005e-06,
"loss": 1.0079,
"step": 1525
},
{
"epoch": 7.242990654205608,
"grad_norm": 385661.21875,
"learning_rate": 3.5975609756097562e-06,
"loss": 1.0593,
"step": 1550
},
{
"epoch": 7.359813084112149,
"grad_norm": 389834.625,
"learning_rate": 3.4451219512195124e-06,
"loss": 1.0175,
"step": 1575
},
{
"epoch": 7.4766355140186915,
"grad_norm": 565496.25,
"learning_rate": 3.292682926829269e-06,
"loss": 1.0171,
"step": 1600
},
{
"epoch": 7.593457943925234,
"grad_norm": 588504.4375,
"learning_rate": 3.1402439024390246e-06,
"loss": 1.0401,
"step": 1625
},
{
"epoch": 7.710280373831775,
"grad_norm": 439602.84375,
"learning_rate": 2.9878048780487808e-06,
"loss": 1.0515,
"step": 1650
},
{
"epoch": 7.827102803738318,
"grad_norm": 750778.125,
"learning_rate": 2.8353658536585365e-06,
"loss": 1.0046,
"step": 1675
},
{
"epoch": 7.94392523364486,
"grad_norm": 399670.3125,
"learning_rate": 2.682926829268293e-06,
"loss": 1.0062,
"step": 1700
},
{
"epoch": 8.0,
"eval_accuracy": 0.7205638474295191,
"eval_loss": 0.9383352398872375,
"eval_runtime": 15.2156,
"eval_samples_per_second": 79.261,
"eval_steps_per_second": 2.497,
"step": 1712
},
{
"epoch": 8.060747663551401,
"grad_norm": 632884.5,
"learning_rate": 2.530487804878049e-06,
"loss": 0.9951,
"step": 1725
},
{
"epoch": 8.177570093457945,
"grad_norm": 539285.375,
"learning_rate": 2.378048780487805e-06,
"loss": 0.9834,
"step": 1750
},
{
"epoch": 8.294392523364486,
"grad_norm": 564263.875,
"learning_rate": 2.225609756097561e-06,
"loss": 1.0018,
"step": 1775
},
{
"epoch": 8.411214953271028,
"grad_norm": 510646.3125,
"learning_rate": 2.073170731707317e-06,
"loss": 0.9775,
"step": 1800
},
{
"epoch": 8.52803738317757,
"grad_norm": 443133.09375,
"learning_rate": 1.9207317073170733e-06,
"loss": 1.0046,
"step": 1825
},
{
"epoch": 8.644859813084112,
"grad_norm": 459672.0625,
"learning_rate": 1.7682926829268294e-06,
"loss": 0.9931,
"step": 1850
},
{
"epoch": 8.761682242990654,
"grad_norm": 494908.71875,
"learning_rate": 1.6158536585365855e-06,
"loss": 0.9988,
"step": 1875
},
{
"epoch": 8.878504672897197,
"grad_norm": 472669.21875,
"learning_rate": 1.4634146341463414e-06,
"loss": 0.9798,
"step": 1900
},
{
"epoch": 8.995327102803738,
"grad_norm": 473329.9375,
"learning_rate": 1.3109756097560978e-06,
"loss": 0.9948,
"step": 1925
},
{
"epoch": 9.0,
"eval_accuracy": 0.7131011608623549,
"eval_loss": 0.9358813166618347,
"eval_runtime": 14.7963,
"eval_samples_per_second": 81.507,
"eval_steps_per_second": 2.568,
"step": 1926
},
{
"epoch": 9.11214953271028,
"grad_norm": 530739.5625,
"learning_rate": 1.158536585365854e-06,
"loss": 0.966,
"step": 1950
},
{
"epoch": 9.228971962616823,
"grad_norm": 483324.6875,
"learning_rate": 1.0060975609756098e-06,
"loss": 1.015,
"step": 1975
},
{
"epoch": 9.345794392523365,
"grad_norm": 518854.625,
"learning_rate": 8.53658536585366e-07,
"loss": 0.985,
"step": 2000
},
{
"epoch": 9.462616822429906,
"grad_norm": 473431.46875,
"learning_rate": 7.012195121951221e-07,
"loss": 0.9399,
"step": 2025
},
{
"epoch": 9.57943925233645,
"grad_norm": 576602.0625,
"learning_rate": 5.487804878048781e-07,
"loss": 1.0286,
"step": 2050
},
{
"epoch": 9.69626168224299,
"grad_norm": 585901.25,
"learning_rate": 3.963414634146342e-07,
"loss": 1.0004,
"step": 2075
},
{
"epoch": 9.813084112149532,
"grad_norm": 543298.6875,
"learning_rate": 2.439024390243903e-07,
"loss": 0.9931,
"step": 2100
},
{
"epoch": 9.929906542056075,
"grad_norm": 524303.25,
"learning_rate": 9.146341463414634e-08,
"loss": 0.9825,
"step": 2125
},
{
"epoch": 10.0,
"eval_accuracy": 0.7072968490878938,
"eval_loss": 0.9387073516845703,
"eval_runtime": 14.9995,
"eval_samples_per_second": 80.403,
"eval_steps_per_second": 2.533,
"step": 2140
},
{
"epoch": 10.0,
"step": 2140,
"total_flos": 5.510586115727032e+19,
"train_loss": 1.1433691055975228,
"train_runtime": 1926.7329,
"train_samples_per_second": 35.449,
"train_steps_per_second": 1.111
}
],
"logging_steps": 25,
"max_steps": 2140,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.510586115727032e+19,
"train_batch_size": 32,
"trial_name": null,
"trial_params": null
}