pszemraj's picture
End of training
3739f98 verified
raw
history blame contribute delete
No virus
18.9 kB
{
"best_metric": 0.9883040935672515,
"best_model_checkpoint": "./outputs/convnextv2-nano-22k-384-boulderspot-vN/checkpoint-1015",
"epoch": 4.993849938499385,
"eval_steps": 500,
"global_step": 1015,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.05,
"grad_norm": 26.665199279785156,
"learning_rate": 3.92156862745098e-06,
"loss": 0.6334,
"step": 10
},
{
"epoch": 0.1,
"grad_norm": 7.943151473999023,
"learning_rate": 7.84313725490196e-06,
"loss": 0.365,
"step": 20
},
{
"epoch": 0.15,
"grad_norm": 16.875411987304688,
"learning_rate": 1.1764705882352942e-05,
"loss": 0.2382,
"step": 30
},
{
"epoch": 0.2,
"grad_norm": 5.587479114532471,
"learning_rate": 1.568627450980392e-05,
"loss": 0.1425,
"step": 40
},
{
"epoch": 0.25,
"grad_norm": 7.946935176849365,
"learning_rate": 1.9607843137254903e-05,
"loss": 0.1257,
"step": 50
},
{
"epoch": 0.3,
"grad_norm": 4.481125831604004,
"learning_rate": 1.9995698998770955e-05,
"loss": 0.0967,
"step": 60
},
{
"epoch": 0.34,
"grad_norm": 9.453625679016113,
"learning_rate": 1.998083609002402e-05,
"loss": 0.0829,
"step": 70
},
{
"epoch": 0.39,
"grad_norm": 13.032307624816895,
"learning_rate": 1.995537395500004e-05,
"loss": 0.1073,
"step": 80
},
{
"epoch": 0.44,
"grad_norm": 5.036722183227539,
"learning_rate": 1.9919339633410737e-05,
"loss": 0.0788,
"step": 90
},
{
"epoch": 0.49,
"grad_norm": 10.13275146484375,
"learning_rate": 1.9872771392183334e-05,
"loss": 0.1421,
"step": 100
},
{
"epoch": 0.54,
"grad_norm": 10.442366600036621,
"learning_rate": 1.981571868482269e-05,
"loss": 0.1035,
"step": 110
},
{
"epoch": 0.59,
"grad_norm": 7.753129959106445,
"learning_rate": 1.974824209889377e-05,
"loss": 0.1001,
"step": 120
},
{
"epoch": 0.64,
"grad_norm": 10.481592178344727,
"learning_rate": 1.9670413291680223e-05,
"loss": 0.1007,
"step": 130
},
{
"epoch": 0.69,
"grad_norm": 7.881091594696045,
"learning_rate": 1.9582314914087344e-05,
"loss": 0.0669,
"step": 140
},
{
"epoch": 0.74,
"grad_norm": 12.763803482055664,
"learning_rate": 1.9484040522870333e-05,
"loss": 0.0723,
"step": 150
},
{
"epoch": 0.79,
"grad_norm": 6.890717029571533,
"learning_rate": 1.9375694481280965e-05,
"loss": 0.0804,
"step": 160
},
{
"epoch": 0.84,
"grad_norm": 6.086670398712158,
"learning_rate": 1.9257391848238212e-05,
"loss": 0.1073,
"step": 170
},
{
"epoch": 0.89,
"grad_norm": 4.475172996520996,
"learning_rate": 1.9129258256140556e-05,
"loss": 0.0779,
"step": 180
},
{
"epoch": 0.93,
"grad_norm": 4.19266414642334,
"learning_rate": 1.8991429777449674e-05,
"loss": 0.0934,
"step": 190
},
{
"epoch": 0.98,
"grad_norm": 14.975569725036621,
"learning_rate": 1.884405278018722e-05,
"loss": 0.1102,
"step": 200
},
{
"epoch": 1.0,
"eval_accuracy": 0.9839181286549707,
"eval_f1": 0.9840087498605248,
"eval_loss": 0.0431428886950016,
"eval_matthews_correlation": 0.8589815510372486,
"eval_precision": 0.9841155979252388,
"eval_recall": 0.9839181286549707,
"eval_runtime": 4.9799,
"eval_samples_per_second": 137.351,
"eval_steps_per_second": 8.635,
"step": 203
},
{
"epoch": 1.03,
"grad_norm": 14.823522567749023,
"learning_rate": 1.8687283772498205e-05,
"loss": 0.0567,
"step": 210
},
{
"epoch": 1.08,
"grad_norm": 2.4599366188049316,
"learning_rate": 1.852128923644593e-05,
"loss": 0.08,
"step": 220
},
{
"epoch": 1.13,
"grad_norm": 1.6709442138671875,
"learning_rate": 1.8346245451215068e-05,
"loss": 0.056,
"step": 230
},
{
"epoch": 1.18,
"grad_norm": 6.953076362609863,
"learning_rate": 1.8162338305910636e-05,
"loss": 0.0683,
"step": 240
},
{
"epoch": 1.23,
"grad_norm": 3.510495901107788,
"learning_rate": 1.79697631021516e-05,
"loss": 0.0825,
"step": 250
},
{
"epoch": 1.28,
"grad_norm": 8.399980545043945,
"learning_rate": 1.776872434666882e-05,
"loss": 0.0584,
"step": 260
},
{
"epoch": 1.33,
"grad_norm": 14.62066650390625,
"learning_rate": 1.7559435534127534e-05,
"loss": 0.0998,
"step": 270
},
{
"epoch": 1.38,
"grad_norm": 6.287339210510254,
"learning_rate": 1.7342118920405035e-05,
"loss": 0.0813,
"step": 280
},
{
"epoch": 1.43,
"grad_norm": 7.23698091506958,
"learning_rate": 1.7117005286564344e-05,
"loss": 0.0973,
"step": 290
},
{
"epoch": 1.48,
"grad_norm": 12.766968727111816,
"learning_rate": 1.688433369377444e-05,
"loss": 0.0928,
"step": 300
},
{
"epoch": 1.53,
"grad_norm": 16.611492156982422,
"learning_rate": 1.6644351229437416e-05,
"loss": 0.08,
"step": 310
},
{
"epoch": 1.57,
"grad_norm": 1.000801920890808,
"learning_rate": 1.63973127447921e-05,
"loss": 0.0582,
"step": 320
},
{
"epoch": 1.62,
"grad_norm": 10.745514869689941,
"learning_rate": 1.6143480584272794e-05,
"loss": 0.0915,
"step": 330
},
{
"epoch": 1.67,
"grad_norm": 10.188589096069336,
"learning_rate": 1.5883124306910563e-05,
"loss": 0.0869,
"step": 340
},
{
"epoch": 1.72,
"grad_norm": 7.215929985046387,
"learning_rate": 1.5616520400072963e-05,
"loss": 0.1329,
"step": 350
},
{
"epoch": 1.77,
"grad_norm": 4.967738151550293,
"learning_rate": 1.5343951985846096e-05,
"loss": 0.0781,
"step": 360
},
{
"epoch": 1.82,
"grad_norm": 2.9467406272888184,
"learning_rate": 1.5065708520370943e-05,
"loss": 0.0582,
"step": 370
},
{
"epoch": 1.87,
"grad_norm": 6.54543399810791,
"learning_rate": 1.4782085486453155e-05,
"loss": 0.0626,
"step": 380
},
{
"epoch": 1.92,
"grad_norm": 7.8247833251953125,
"learning_rate": 1.4493384079772815e-05,
"loss": 0.081,
"step": 390
},
{
"epoch": 1.97,
"grad_norm": 8.889460563659668,
"learning_rate": 1.4199910889027335e-05,
"loss": 0.0559,
"step": 400
},
{
"epoch": 2.0,
"eval_accuracy": 0.9839181286549707,
"eval_f1": 0.9845065535493497,
"eval_loss": 0.04758350923657417,
"eval_matthews_correlation": 0.8709455436036694,
"eval_precision": 0.9858048061716136,
"eval_recall": 0.9839181286549707,
"eval_runtime": 5.1397,
"eval_samples_per_second": 133.081,
"eval_steps_per_second": 8.366,
"step": 406
},
{
"epoch": 2.02,
"grad_norm": 11.741949081420898,
"learning_rate": 1.390197757034721e-05,
"loss": 0.0806,
"step": 410
},
{
"epoch": 2.07,
"grad_norm": 8.677542686462402,
"learning_rate": 1.3599900516330382e-05,
"loss": 0.1033,
"step": 420
},
{
"epoch": 2.12,
"grad_norm": 1.3803766965866089,
"learning_rate": 1.3294000520046666e-05,
"loss": 0.0433,
"step": 430
},
{
"epoch": 2.16,
"grad_norm": 8.467944145202637,
"learning_rate": 1.2984602434369058e-05,
"loss": 0.0691,
"step": 440
},
{
"epoch": 2.21,
"grad_norm": 6.276478290557861,
"learning_rate": 1.2672034826993716e-05,
"loss": 0.0678,
"step": 450
},
{
"epoch": 2.26,
"grad_norm": 11.57728385925293,
"learning_rate": 1.235662963151493e-05,
"loss": 0.0804,
"step": 460
},
{
"epoch": 2.31,
"grad_norm": 0.9122611880302429,
"learning_rate": 1.2038721794925689e-05,
"loss": 0.0476,
"step": 470
},
{
"epoch": 2.36,
"grad_norm": 1.983340859413147,
"learning_rate": 1.1718648921918112e-05,
"loss": 0.0545,
"step": 480
},
{
"epoch": 2.41,
"grad_norm": 17.866886138916016,
"learning_rate": 1.1396750916361526e-05,
"loss": 0.0512,
"step": 490
},
{
"epoch": 2.46,
"grad_norm": 9.374113082885742,
"learning_rate": 1.1073369620338928e-05,
"loss": 0.0604,
"step": 500
},
{
"epoch": 2.51,
"grad_norm": 4.585148811340332,
"learning_rate": 1.074884845112512e-05,
"loss": 0.0629,
"step": 510
},
{
"epoch": 2.56,
"grad_norm": 9.917503356933594,
"learning_rate": 1.0423532036492077e-05,
"loss": 0.054,
"step": 520
},
{
"epoch": 2.61,
"grad_norm": 1.8329569101333618,
"learning_rate": 1.0097765848728825e-05,
"loss": 0.071,
"step": 530
},
{
"epoch": 2.66,
"grad_norm": 10.834183692932129,
"learning_rate": 9.771895837764438e-06,
"loss": 0.0642,
"step": 540
},
{
"epoch": 2.71,
"grad_norm": 5.503483295440674,
"learning_rate": 9.446268063783853e-06,
"loss": 0.0654,
"step": 550
},
{
"epoch": 2.76,
"grad_norm": 6.054934024810791,
"learning_rate": 9.121228329726563e-06,
"loss": 0.061,
"step": 560
},
{
"epoch": 2.8,
"grad_norm": 8.71143627166748,
"learning_rate": 8.797121814058502e-06,
"loss": 0.0759,
"step": 570
},
{
"epoch": 2.85,
"grad_norm": 4.225427150726318,
"learning_rate": 8.474292704207095e-06,
"loss": 0.0662,
"step": 580
},
{
"epoch": 2.9,
"grad_norm": 6.926836967468262,
"learning_rate": 8.153083831048772e-06,
"loss": 0.07,
"step": 590
},
{
"epoch": 2.95,
"grad_norm": 1.132318377494812,
"learning_rate": 7.833836304837022e-06,
"loss": 0.0402,
"step": 600
},
{
"epoch": 3.0,
"eval_accuracy": 0.9809941520467836,
"eval_f1": 0.9816895632855951,
"eval_loss": 0.04637432098388672,
"eval_matthews_correlation": 0.8468348607550005,
"eval_precision": 0.9830779451248668,
"eval_recall": 0.9809941520467836,
"eval_runtime": 5.1222,
"eval_samples_per_second": 133.536,
"eval_steps_per_second": 8.395,
"step": 609
},
{
"epoch": 3.0,
"grad_norm": 2.548271656036377,
"learning_rate": 7.516889152957744e-06,
"loss": 0.0706,
"step": 610
},
{
"epoch": 3.05,
"grad_norm": 17.59124755859375,
"learning_rate": 7.202578959896491e-06,
"loss": 0.0777,
"step": 620
},
{
"epoch": 3.1,
"grad_norm": 10.620137214660645,
"learning_rate": 6.891239509799932e-06,
"loss": 0.0443,
"step": 630
},
{
"epoch": 3.15,
"grad_norm": 20.15471839904785,
"learning_rate": 6.583201432011217e-06,
"loss": 0.0746,
"step": 640
},
{
"epoch": 3.2,
"grad_norm": 7.445899486541748,
"learning_rate": 6.278791849955583e-06,
"loss": 0.0468,
"step": 650
},
{
"epoch": 3.25,
"grad_norm": 10.309353828430176,
"learning_rate": 5.978334033749076e-06,
"loss": 0.0698,
"step": 660
},
{
"epoch": 3.3,
"grad_norm": 4.898565769195557,
"learning_rate": 5.682147056899361e-06,
"loss": 0.058,
"step": 670
},
{
"epoch": 3.35,
"grad_norm": 8.0520658493042,
"learning_rate": 5.390545457463134e-06,
"loss": 0.0468,
"step": 680
},
{
"epoch": 3.39,
"grad_norm": 3.444305419921875,
"learning_rate": 5.103838904019993e-06,
"loss": 0.0381,
"step": 690
},
{
"epoch": 3.44,
"grad_norm": 11.149181365966797,
"learning_rate": 4.822331866817478e-06,
"loss": 0.0693,
"step": 700
},
{
"epoch": 3.49,
"grad_norm": 12.743651390075684,
"learning_rate": 4.546323294436556e-06,
"loss": 0.0739,
"step": 710
},
{
"epoch": 3.54,
"grad_norm": 9.802193641662598,
"learning_rate": 4.276106296320828e-06,
"loss": 0.0655,
"step": 720
},
{
"epoch": 3.59,
"grad_norm": 6.496978759765625,
"learning_rate": 4.0119678315067025e-06,
"loss": 0.066,
"step": 730
},
{
"epoch": 3.64,
"grad_norm": 12.13724422454834,
"learning_rate": 3.754188403885013e-06,
"loss": 0.0483,
"step": 740
},
{
"epoch": 3.69,
"grad_norm": 11.302549362182617,
"learning_rate": 3.5030417643177416e-06,
"loss": 0.0727,
"step": 750
},
{
"epoch": 3.74,
"grad_norm": 3.291935443878174,
"learning_rate": 3.258794619926159e-06,
"loss": 0.0685,
"step": 760
},
{
"epoch": 3.79,
"grad_norm": 7.776196479797363,
"learning_rate": 3.021706350859147e-06,
"loss": 0.0482,
"step": 770
},
{
"epoch": 3.84,
"grad_norm": 7.8189568519592285,
"learning_rate": 2.792028734842418e-06,
"loss": 0.0504,
"step": 780
},
{
"epoch": 3.89,
"grad_norm": 1.9077569246292114,
"learning_rate": 2.5700056798012164e-06,
"loss": 0.0578,
"step": 790
},
{
"epoch": 3.94,
"grad_norm": 5.7739667892456055,
"learning_rate": 2.3558729648404065e-06,
"loss": 0.0712,
"step": 800
},
{
"epoch": 3.99,
"grad_norm": 6.1386027336120605,
"learning_rate": 2.1498579898570228e-06,
"loss": 0.0334,
"step": 810
},
{
"epoch": 4.0,
"eval_accuracy": 0.9868421052631579,
"eval_f1": 0.986916249885884,
"eval_loss": 0.03484980762004852,
"eval_matthews_correlation": 0.884636226651146,
"eval_precision": 0.9870070425284395,
"eval_recall": 0.9868421052631579,
"eval_runtime": 5.1774,
"eval_samples_per_second": 132.112,
"eval_steps_per_second": 8.305,
"step": 813
},
{
"epoch": 4.03,
"grad_norm": 2.534424304962158,
"learning_rate": 1.952179534051183e-06,
"loss": 0.0613,
"step": 820
},
{
"epoch": 4.08,
"grad_norm": 3.645782947540283,
"learning_rate": 1.763047523591831e-06,
"loss": 0.0817,
"step": 830
},
{
"epoch": 4.13,
"grad_norm": 10.930766105651855,
"learning_rate": 1.5826628086839968e-06,
"loss": 0.081,
"step": 840
},
{
"epoch": 4.18,
"grad_norm": 6.021533489227295,
"learning_rate": 1.41121695027438e-06,
"loss": 0.0522,
"step": 850
},
{
"epoch": 4.23,
"grad_norm": 6.584732532501221,
"learning_rate": 1.2488920166217034e-06,
"loss": 0.054,
"step": 860
},
{
"epoch": 4.28,
"grad_norm": 5.629753112792969,
"learning_rate": 1.095860389947928e-06,
"loss": 0.0475,
"step": 870
},
{
"epoch": 4.33,
"grad_norm": 4.272614479064941,
"learning_rate": 9.522845833756001e-07,
"loss": 0.0458,
"step": 880
},
{
"epoch": 4.38,
"grad_norm": 9.34067153930664,
"learning_rate": 8.183170683457986e-07,
"loss": 0.0746,
"step": 890
},
{
"epoch": 4.43,
"grad_norm": 2.2390189170837402,
"learning_rate": 6.941001126998892e-07,
"loss": 0.0509,
"step": 900
},
{
"epoch": 4.48,
"grad_norm": 1.508949637413025,
"learning_rate": 5.797656295970955e-07,
"loss": 0.0329,
"step": 910
},
{
"epoch": 4.53,
"grad_norm": 7.0009355545043945,
"learning_rate": 4.754350374283001e-07,
"loss": 0.0709,
"step": 920
},
{
"epoch": 4.58,
"grad_norm": 6.07660436630249,
"learning_rate": 3.8121913087483033e-07,
"loss": 0.0393,
"step": 930
},
{
"epoch": 4.62,
"grad_norm": 6.966567039489746,
"learning_rate": 2.972179632491989e-07,
"loss": 0.0464,
"step": 940
},
{
"epoch": 4.67,
"grad_norm": 3.8328962326049805,
"learning_rate": 2.23520740242712e-07,
"loss": 0.0457,
"step": 950
},
{
"epoch": 4.72,
"grad_norm": 3.535372018814087,
"learning_rate": 1.602057251927891e-07,
"loss": 0.0599,
"step": 960
},
{
"epoch": 4.77,
"grad_norm": 11.542516708374023,
"learning_rate": 1.0734015597060222e-07,
"loss": 0.0557,
"step": 970
},
{
"epoch": 4.82,
"grad_norm": 7.840158939361572,
"learning_rate": 6.498017357731035e-08,
"loss": 0.0564,
"step": 980
},
{
"epoch": 4.87,
"grad_norm": 3.7357852458953857,
"learning_rate": 3.317076252467133e-08,
"loss": 0.033,
"step": 990
},
{
"epoch": 4.92,
"grad_norm": 4.429032802581787,
"learning_rate": 1.1945703063402925e-08,
"loss": 0.0496,
"step": 1000
},
{
"epoch": 4.97,
"grad_norm": 4.065526485443115,
"learning_rate": 1.327535309979533e-09,
"loss": 0.0445,
"step": 1010
},
{
"epoch": 4.99,
"eval_accuracy": 0.9883040935672515,
"eval_f1": 0.9883040935672515,
"eval_loss": 0.03396734222769737,
"eval_matthews_correlation": 0.8962181845768691,
"eval_precision": 0.9883040935672515,
"eval_recall": 0.9883040935672515,
"eval_runtime": 5.2898,
"eval_samples_per_second": 129.305,
"eval_steps_per_second": 8.129,
"step": 1015
},
{
"epoch": 4.99,
"step": 1015,
"total_flos": 2.581447650539471e+18,
"train_loss": 0.08031210996247277,
"train_runtime": 701.9998,
"train_samples_per_second": 92.557,
"train_steps_per_second": 1.446
}
],
"logging_steps": 10,
"max_steps": 1015,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 500,
"total_flos": 2.581447650539471e+18,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}