Kushagra07's picture
Upload folder using huggingface_hub
7222601 verified
raw
history blame contribute delete
No virus
76 kB
{
"best_metric": 0.23535355925559998,
"best_model_checkpoint": "autotrain-vit-large-patch16-224/checkpoint-10234",
"epoch": 7.0,
"eval_steps": 500,
"global_step": 10234,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01709986320109439,
"grad_norm": 16.03294563293457,
"learning_rate": 8.207934336525308e-07,
"loss": 3.1283,
"step": 25
},
{
"epoch": 0.03419972640218878,
"grad_norm": 15.965583801269531,
"learning_rate": 1.6073871409028727e-06,
"loss": 2.7271,
"step": 50
},
{
"epoch": 0.05129958960328317,
"grad_norm": 10.72908878326416,
"learning_rate": 2.4623803009575924e-06,
"loss": 2.2885,
"step": 75
},
{
"epoch": 0.06839945280437756,
"grad_norm": 12.793851852416992,
"learning_rate": 3.3173734610123124e-06,
"loss": 1.9045,
"step": 100
},
{
"epoch": 0.08549931600547196,
"grad_norm": 11.025067329406738,
"learning_rate": 4.172366621067032e-06,
"loss": 1.4383,
"step": 125
},
{
"epoch": 0.10259917920656635,
"grad_norm": 13.093396186828613,
"learning_rate": 5.027359781121752e-06,
"loss": 1.3066,
"step": 150
},
{
"epoch": 0.11969904240766074,
"grad_norm": 8.261626243591309,
"learning_rate": 5.882352941176471e-06,
"loss": 1.1223,
"step": 175
},
{
"epoch": 0.13679890560875513,
"grad_norm": 9.678178787231445,
"learning_rate": 6.73734610123119e-06,
"loss": 1.1502,
"step": 200
},
{
"epoch": 0.1538987688098495,
"grad_norm": 13.990431785583496,
"learning_rate": 7.592339261285911e-06,
"loss": 0.9833,
"step": 225
},
{
"epoch": 0.17099863201094392,
"grad_norm": 11.807621002197266,
"learning_rate": 8.44733242134063e-06,
"loss": 0.8433,
"step": 250
},
{
"epoch": 0.1880984952120383,
"grad_norm": 21.150636672973633,
"learning_rate": 9.302325581395349e-06,
"loss": 0.8553,
"step": 275
},
{
"epoch": 0.2051983584131327,
"grad_norm": 16.56648826599121,
"learning_rate": 1.0157318741450068e-05,
"loss": 1.0502,
"step": 300
},
{
"epoch": 0.22229822161422708,
"grad_norm": 8.26397705078125,
"learning_rate": 1.1012311901504789e-05,
"loss": 0.8686,
"step": 325
},
{
"epoch": 0.2393980848153215,
"grad_norm": 7.628605842590332,
"learning_rate": 1.1867305061559508e-05,
"loss": 0.7838,
"step": 350
},
{
"epoch": 0.25649794801641584,
"grad_norm": 6.690649509429932,
"learning_rate": 1.2722298221614229e-05,
"loss": 0.7284,
"step": 375
},
{
"epoch": 0.27359781121751026,
"grad_norm": 9.397786140441895,
"learning_rate": 1.3577291381668946e-05,
"loss": 0.6408,
"step": 400
},
{
"epoch": 0.29069767441860467,
"grad_norm": 1.714457392692566,
"learning_rate": 1.4432284541723667e-05,
"loss": 0.7046,
"step": 425
},
{
"epoch": 0.307797537619699,
"grad_norm": 16.014860153198242,
"learning_rate": 1.5287277701778386e-05,
"loss": 0.7838,
"step": 450
},
{
"epoch": 0.32489740082079344,
"grad_norm": 10.126370429992676,
"learning_rate": 1.6142270861833107e-05,
"loss": 0.735,
"step": 475
},
{
"epoch": 0.34199726402188785,
"grad_norm": 10.995970726013184,
"learning_rate": 1.6997264021887825e-05,
"loss": 0.7523,
"step": 500
},
{
"epoch": 0.3590971272229822,
"grad_norm": 7.909482955932617,
"learning_rate": 1.7852257181942546e-05,
"loss": 0.7157,
"step": 525
},
{
"epoch": 0.3761969904240766,
"grad_norm": 10.751182556152344,
"learning_rate": 1.8707250341997263e-05,
"loss": 0.7966,
"step": 550
},
{
"epoch": 0.393296853625171,
"grad_norm": 8.872684478759766,
"learning_rate": 1.9562243502051984e-05,
"loss": 0.6923,
"step": 575
},
{
"epoch": 0.4103967168262654,
"grad_norm": 12.542756080627441,
"learning_rate": 2.0417236662106705e-05,
"loss": 0.8022,
"step": 600
},
{
"epoch": 0.4274965800273598,
"grad_norm": 4.314883708953857,
"learning_rate": 2.1272229822161423e-05,
"loss": 0.6959,
"step": 625
},
{
"epoch": 0.44459644322845415,
"grad_norm": 10.603111267089844,
"learning_rate": 2.2127222982216144e-05,
"loss": 0.7282,
"step": 650
},
{
"epoch": 0.46169630642954856,
"grad_norm": 12.411425590515137,
"learning_rate": 2.298221614227086e-05,
"loss": 0.8191,
"step": 675
},
{
"epoch": 0.478796169630643,
"grad_norm": 6.108178615570068,
"learning_rate": 2.3837209302325582e-05,
"loss": 0.507,
"step": 700
},
{
"epoch": 0.49589603283173733,
"grad_norm": 11.323492050170898,
"learning_rate": 2.46922024623803e-05,
"loss": 0.5333,
"step": 725
},
{
"epoch": 0.5129958960328317,
"grad_norm": 9.001029014587402,
"learning_rate": 2.554719562243502e-05,
"loss": 0.6816,
"step": 750
},
{
"epoch": 0.5300957592339262,
"grad_norm": 1.6873475313186646,
"learning_rate": 2.6402188782489738e-05,
"loss": 0.6621,
"step": 775
},
{
"epoch": 0.5471956224350205,
"grad_norm": 9.71667194366455,
"learning_rate": 2.7257181942544462e-05,
"loss": 0.5873,
"step": 800
},
{
"epoch": 0.5642954856361149,
"grad_norm": 5.7018866539001465,
"learning_rate": 2.811217510259918e-05,
"loss": 0.6119,
"step": 825
},
{
"epoch": 0.5813953488372093,
"grad_norm": 6.869658946990967,
"learning_rate": 2.8967168262653897e-05,
"loss": 0.8036,
"step": 850
},
{
"epoch": 0.5984952120383037,
"grad_norm": 10.419452667236328,
"learning_rate": 2.982216142270862e-05,
"loss": 0.6884,
"step": 875
},
{
"epoch": 0.615595075239398,
"grad_norm": 6.166721343994141,
"learning_rate": 3.067715458276334e-05,
"loss": 0.652,
"step": 900
},
{
"epoch": 0.6326949384404925,
"grad_norm": 10.317444801330566,
"learning_rate": 3.1532147742818057e-05,
"loss": 0.6249,
"step": 925
},
{
"epoch": 0.6497948016415869,
"grad_norm": 11.344230651855469,
"learning_rate": 3.2387140902872784e-05,
"loss": 0.6797,
"step": 950
},
{
"epoch": 0.6668946648426812,
"grad_norm": 8.010811805725098,
"learning_rate": 3.32421340629275e-05,
"loss": 0.8409,
"step": 975
},
{
"epoch": 0.6839945280437757,
"grad_norm": 8.49295425415039,
"learning_rate": 3.409712722298222e-05,
"loss": 0.5985,
"step": 1000
},
{
"epoch": 0.70109439124487,
"grad_norm": 14.201081275939941,
"learning_rate": 3.4952120383036933e-05,
"loss": 0.7377,
"step": 1025
},
{
"epoch": 0.7181942544459644,
"grad_norm": 12.270833015441895,
"learning_rate": 3.580711354309166e-05,
"loss": 0.5195,
"step": 1050
},
{
"epoch": 0.7352941176470589,
"grad_norm": 2.879152297973633,
"learning_rate": 3.6662106703146375e-05,
"loss": 0.6526,
"step": 1075
},
{
"epoch": 0.7523939808481532,
"grad_norm": 8.10263442993164,
"learning_rate": 3.7517099863201096e-05,
"loss": 0.7119,
"step": 1100
},
{
"epoch": 0.7694938440492476,
"grad_norm": 4.584366321563721,
"learning_rate": 3.837209302325582e-05,
"loss": 0.6515,
"step": 1125
},
{
"epoch": 0.786593707250342,
"grad_norm": 6.819866180419922,
"learning_rate": 3.922708618331054e-05,
"loss": 0.5304,
"step": 1150
},
{
"epoch": 0.8036935704514364,
"grad_norm": 9.715897560119629,
"learning_rate": 4.008207934336525e-05,
"loss": 0.5718,
"step": 1175
},
{
"epoch": 0.8207934336525308,
"grad_norm": 6.830766677856445,
"learning_rate": 4.093707250341998e-05,
"loss": 0.6228,
"step": 1200
},
{
"epoch": 0.8378932968536251,
"grad_norm": 8.660406112670898,
"learning_rate": 4.1792065663474694e-05,
"loss": 0.7842,
"step": 1225
},
{
"epoch": 0.8549931600547196,
"grad_norm": 8.981478691101074,
"learning_rate": 4.2647058823529415e-05,
"loss": 0.7702,
"step": 1250
},
{
"epoch": 0.872093023255814,
"grad_norm": 5.491425037384033,
"learning_rate": 4.3502051983584136e-05,
"loss": 0.6483,
"step": 1275
},
{
"epoch": 0.8891928864569083,
"grad_norm": 8.423704147338867,
"learning_rate": 4.435704514363886e-05,
"loss": 0.6889,
"step": 1300
},
{
"epoch": 0.9062927496580028,
"grad_norm": 11.879225730895996,
"learning_rate": 4.521203830369357e-05,
"loss": 0.7592,
"step": 1325
},
{
"epoch": 0.9233926128590971,
"grad_norm": 10.297099113464355,
"learning_rate": 4.606703146374829e-05,
"loss": 0.6994,
"step": 1350
},
{
"epoch": 0.9404924760601915,
"grad_norm": 34.574771881103516,
"learning_rate": 4.692202462380301e-05,
"loss": 0.6254,
"step": 1375
},
{
"epoch": 0.957592339261286,
"grad_norm": 7.829855442047119,
"learning_rate": 4.7777017783857733e-05,
"loss": 0.62,
"step": 1400
},
{
"epoch": 0.9746922024623803,
"grad_norm": 10.330737113952637,
"learning_rate": 4.863201094391245e-05,
"loss": 0.5799,
"step": 1425
},
{
"epoch": 0.9917920656634747,
"grad_norm": 7.56599235534668,
"learning_rate": 4.948700410396717e-05,
"loss": 0.6286,
"step": 1450
},
{
"epoch": 1.0,
"eval_accuracy": 0.8296346876066917,
"eval_f1_macro": 0.5237337030166633,
"eval_f1_micro": 0.8296346876066917,
"eval_f1_weighted": 0.818531665770605,
"eval_loss": 0.5858680605888367,
"eval_precision_macro": 0.6349808627313414,
"eval_precision_micro": 0.8296346876066917,
"eval_precision_weighted": 0.8428198229026794,
"eval_recall_macro": 0.5158734772675513,
"eval_recall_micro": 0.8296346876066917,
"eval_recall_weighted": 0.8296346876066917,
"eval_runtime": 29.6775,
"eval_samples_per_second": 98.694,
"eval_steps_per_second": 6.2,
"step": 1462
},
{
"epoch": 1.008891928864569,
"grad_norm": 7.980313301086426,
"learning_rate": 4.996200030399757e-05,
"loss": 0.6037,
"step": 1475
},
{
"epoch": 1.0259917920656634,
"grad_norm": 6.049495220184326,
"learning_rate": 4.9867001063991494e-05,
"loss": 0.6332,
"step": 1500
},
{
"epoch": 1.043091655266758,
"grad_norm": 11.885618209838867,
"learning_rate": 4.977200182398541e-05,
"loss": 0.761,
"step": 1525
},
{
"epoch": 1.0601915184678523,
"grad_norm": 14.27984619140625,
"learning_rate": 4.9677002583979335e-05,
"loss": 0.4568,
"step": 1550
},
{
"epoch": 1.0772913816689467,
"grad_norm": 6.687506198883057,
"learning_rate": 4.958200334397325e-05,
"loss": 0.6308,
"step": 1575
},
{
"epoch": 1.094391244870041,
"grad_norm": 6.140110492706299,
"learning_rate": 4.948700410396717e-05,
"loss": 0.5896,
"step": 1600
},
{
"epoch": 1.1114911080711354,
"grad_norm": 8.485834121704102,
"learning_rate": 4.939200486396109e-05,
"loss": 0.7167,
"step": 1625
},
{
"epoch": 1.1285909712722297,
"grad_norm": 8.836180686950684,
"learning_rate": 4.929700562395501e-05,
"loss": 0.5643,
"step": 1650
},
{
"epoch": 1.1456908344733243,
"grad_norm": 5.937542915344238,
"learning_rate": 4.9202006383948926e-05,
"loss": 0.5568,
"step": 1675
},
{
"epoch": 1.1627906976744187,
"grad_norm": 4.0352582931518555,
"learning_rate": 4.910700714394285e-05,
"loss": 0.7228,
"step": 1700
},
{
"epoch": 1.179890560875513,
"grad_norm": 6.907437801361084,
"learning_rate": 4.9012007903936766e-05,
"loss": 0.5816,
"step": 1725
},
{
"epoch": 1.1969904240766074,
"grad_norm": 6.376392841339111,
"learning_rate": 4.891700866393069e-05,
"loss": 0.5199,
"step": 1750
},
{
"epoch": 1.2140902872777017,
"grad_norm": 9.630695343017578,
"learning_rate": 4.882200942392461e-05,
"loss": 0.6383,
"step": 1775
},
{
"epoch": 1.231190150478796,
"grad_norm": 6.556370258331299,
"learning_rate": 4.872701018391853e-05,
"loss": 0.5977,
"step": 1800
},
{
"epoch": 1.2482900136798905,
"grad_norm": 8.746458053588867,
"learning_rate": 4.863201094391245e-05,
"loss": 0.5699,
"step": 1825
},
{
"epoch": 1.265389876880985,
"grad_norm": 6.0708537101745605,
"learning_rate": 4.853701170390637e-05,
"loss": 0.6479,
"step": 1850
},
{
"epoch": 1.2824897400820794,
"grad_norm": 7.421648979187012,
"learning_rate": 4.8442012463900295e-05,
"loss": 0.7258,
"step": 1875
},
{
"epoch": 1.2995896032831737,
"grad_norm": 11.180326461791992,
"learning_rate": 4.834701322389421e-05,
"loss": 0.5853,
"step": 1900
},
{
"epoch": 1.316689466484268,
"grad_norm": 12.808277130126953,
"learning_rate": 4.8252013983888135e-05,
"loss": 0.625,
"step": 1925
},
{
"epoch": 1.3337893296853625,
"grad_norm": 5.873989105224609,
"learning_rate": 4.815701474388205e-05,
"loss": 0.6044,
"step": 1950
},
{
"epoch": 1.350889192886457,
"grad_norm": 6.302676200866699,
"learning_rate": 4.8062015503875976e-05,
"loss": 0.7381,
"step": 1975
},
{
"epoch": 1.3679890560875512,
"grad_norm": 10.768912315368652,
"learning_rate": 4.796701626386989e-05,
"loss": 0.7416,
"step": 2000
},
{
"epoch": 1.3850889192886457,
"grad_norm": 3.4966821670532227,
"learning_rate": 4.7872017023863817e-05,
"loss": 0.5816,
"step": 2025
},
{
"epoch": 1.40218878248974,
"grad_norm": 20.856693267822266,
"learning_rate": 4.7777017783857733e-05,
"loss": 0.674,
"step": 2050
},
{
"epoch": 1.4192886456908345,
"grad_norm": 3.353882074356079,
"learning_rate": 4.768201854385165e-05,
"loss": 0.6317,
"step": 2075
},
{
"epoch": 1.4363885088919288,
"grad_norm": 4.5955939292907715,
"learning_rate": 4.758701930384557e-05,
"loss": 0.5058,
"step": 2100
},
{
"epoch": 1.4534883720930232,
"grad_norm": 3.8296780586242676,
"learning_rate": 4.749202006383949e-05,
"loss": 0.55,
"step": 2125
},
{
"epoch": 1.4705882352941178,
"grad_norm": 8.511550903320312,
"learning_rate": 4.739702082383341e-05,
"loss": 0.7234,
"step": 2150
},
{
"epoch": 1.487688098495212,
"grad_norm": 7.452313423156738,
"learning_rate": 4.730202158382733e-05,
"loss": 0.4657,
"step": 2175
},
{
"epoch": 1.5047879616963065,
"grad_norm": 5.583883762359619,
"learning_rate": 4.720702234382125e-05,
"loss": 0.5554,
"step": 2200
},
{
"epoch": 1.5218878248974008,
"grad_norm": 3.6462392807006836,
"learning_rate": 4.711202310381517e-05,
"loss": 0.6182,
"step": 2225
},
{
"epoch": 1.5389876880984952,
"grad_norm": 5.7832207679748535,
"learning_rate": 4.701702386380909e-05,
"loss": 0.4694,
"step": 2250
},
{
"epoch": 1.5560875512995898,
"grad_norm": 3.335848331451416,
"learning_rate": 4.692202462380301e-05,
"loss": 0.6423,
"step": 2275
},
{
"epoch": 1.573187414500684,
"grad_norm": 15.582513809204102,
"learning_rate": 4.682702538379693e-05,
"loss": 0.6912,
"step": 2300
},
{
"epoch": 1.5902872777017785,
"grad_norm": 8.966438293457031,
"learning_rate": 4.673202614379085e-05,
"loss": 0.4998,
"step": 2325
},
{
"epoch": 1.6073871409028728,
"grad_norm": 6.383381366729736,
"learning_rate": 4.663702690378477e-05,
"loss": 0.5879,
"step": 2350
},
{
"epoch": 1.6244870041039672,
"grad_norm": 3.314211845397949,
"learning_rate": 4.6542027663778694e-05,
"loss": 0.5865,
"step": 2375
},
{
"epoch": 1.6415868673050615,
"grad_norm": 7.680863380432129,
"learning_rate": 4.644702842377261e-05,
"loss": 0.5731,
"step": 2400
},
{
"epoch": 1.658686730506156,
"grad_norm": 15.084454536437988,
"learning_rate": 4.6352029183766534e-05,
"loss": 0.6018,
"step": 2425
},
{
"epoch": 1.6757865937072505,
"grad_norm": 4.016242027282715,
"learning_rate": 4.625702994376046e-05,
"loss": 0.4291,
"step": 2450
},
{
"epoch": 1.6928864569083446,
"grad_norm": 6.622522354125977,
"learning_rate": 4.6162030703754375e-05,
"loss": 0.5729,
"step": 2475
},
{
"epoch": 1.7099863201094392,
"grad_norm": 7.149597644805908,
"learning_rate": 4.606703146374829e-05,
"loss": 0.6435,
"step": 2500
},
{
"epoch": 1.7270861833105335,
"grad_norm": 4.473049163818359,
"learning_rate": 4.597203222374221e-05,
"loss": 0.5068,
"step": 2525
},
{
"epoch": 1.744186046511628,
"grad_norm": 8.462128639221191,
"learning_rate": 4.587703298373613e-05,
"loss": 0.4735,
"step": 2550
},
{
"epoch": 1.7612859097127223,
"grad_norm": 3.4091100692749023,
"learning_rate": 4.578203374373005e-05,
"loss": 0.591,
"step": 2575
},
{
"epoch": 1.7783857729138166,
"grad_norm": 4.581203937530518,
"learning_rate": 4.568703450372397e-05,
"loss": 0.4642,
"step": 2600
},
{
"epoch": 1.7954856361149112,
"grad_norm": 7.462313175201416,
"learning_rate": 4.559203526371789e-05,
"loss": 0.5036,
"step": 2625
},
{
"epoch": 1.8125854993160053,
"grad_norm": 7.391847610473633,
"learning_rate": 4.549703602371181e-05,
"loss": 0.5039,
"step": 2650
},
{
"epoch": 1.8296853625171,
"grad_norm": 16.374174118041992,
"learning_rate": 4.540203678370573e-05,
"loss": 0.6136,
"step": 2675
},
{
"epoch": 1.8467852257181943,
"grad_norm": 7.047588348388672,
"learning_rate": 4.5307037543699654e-05,
"loss": 0.4051,
"step": 2700
},
{
"epoch": 1.8638850889192886,
"grad_norm": 4.181390285491943,
"learning_rate": 4.521203830369357e-05,
"loss": 0.7065,
"step": 2725
},
{
"epoch": 1.8809849521203832,
"grad_norm": 2.601778984069824,
"learning_rate": 4.5117039063687494e-05,
"loss": 0.5517,
"step": 2750
},
{
"epoch": 1.8980848153214773,
"grad_norm": 2.9935567378997803,
"learning_rate": 4.502203982368141e-05,
"loss": 0.3861,
"step": 2775
},
{
"epoch": 1.915184678522572,
"grad_norm": 7.431847095489502,
"learning_rate": 4.4927040583675335e-05,
"loss": 0.5537,
"step": 2800
},
{
"epoch": 1.9322845417236663,
"grad_norm": 7.30590295791626,
"learning_rate": 4.483204134366925e-05,
"loss": 0.5082,
"step": 2825
},
{
"epoch": 1.9493844049247606,
"grad_norm": 6.986799240112305,
"learning_rate": 4.4737042103663176e-05,
"loss": 0.5727,
"step": 2850
},
{
"epoch": 1.966484268125855,
"grad_norm": 4.823264122009277,
"learning_rate": 4.464204286365709e-05,
"loss": 0.5773,
"step": 2875
},
{
"epoch": 1.9835841313269493,
"grad_norm": 0.19885371625423431,
"learning_rate": 4.4547043623651016e-05,
"loss": 0.5056,
"step": 2900
},
{
"epoch": 2.0,
"eval_accuracy": 0.9095254353021509,
"eval_f1_macro": 0.7148556374573719,
"eval_f1_micro": 0.9095254353021509,
"eval_f1_weighted": 0.906235846022486,
"eval_loss": 0.2884175181388855,
"eval_precision_macro": 0.779691628927798,
"eval_precision_micro": 0.9095254353021509,
"eval_precision_weighted": 0.9118121732094211,
"eval_recall_macro": 0.6904984139815875,
"eval_recall_micro": 0.9095254353021509,
"eval_recall_weighted": 0.9095254353021509,
"eval_runtime": 29.8929,
"eval_samples_per_second": 97.983,
"eval_steps_per_second": 6.155,
"step": 2924
},
{
"epoch": 2.000683994528044,
"grad_norm": 8.045063972473145,
"learning_rate": 4.445204438364493e-05,
"loss": 0.5348,
"step": 2925
},
{
"epoch": 2.017783857729138,
"grad_norm": 7.1167521476745605,
"learning_rate": 4.435704514363886e-05,
"loss": 0.4205,
"step": 2950
},
{
"epoch": 2.0348837209302326,
"grad_norm": 7.508376121520996,
"learning_rate": 4.4262045903632774e-05,
"loss": 0.5367,
"step": 2975
},
{
"epoch": 2.0519835841313268,
"grad_norm": 10.526673316955566,
"learning_rate": 4.416704666362669e-05,
"loss": 0.4326,
"step": 3000
},
{
"epoch": 2.0690834473324213,
"grad_norm": 7.46666145324707,
"learning_rate": 4.407204742362061e-05,
"loss": 0.6013,
"step": 3025
},
{
"epoch": 2.086183310533516,
"grad_norm": 3.7393107414245605,
"learning_rate": 4.397704818361453e-05,
"loss": 0.5544,
"step": 3050
},
{
"epoch": 2.10328317373461,
"grad_norm": 6.285150051116943,
"learning_rate": 4.3882048943608455e-05,
"loss": 0.5717,
"step": 3075
},
{
"epoch": 2.1203830369357046,
"grad_norm": 8.156254768371582,
"learning_rate": 4.378704970360237e-05,
"loss": 0.5008,
"step": 3100
},
{
"epoch": 2.1374829001367988,
"grad_norm": 10.280779838562012,
"learning_rate": 4.3692050463596295e-05,
"loss": 0.3871,
"step": 3125
},
{
"epoch": 2.1545827633378933,
"grad_norm": 11.541882514953613,
"learning_rate": 4.359705122359021e-05,
"loss": 0.4684,
"step": 3150
},
{
"epoch": 2.1716826265389875,
"grad_norm": 12.31938648223877,
"learning_rate": 4.3502051983584136e-05,
"loss": 0.5055,
"step": 3175
},
{
"epoch": 2.188782489740082,
"grad_norm": 6.534464359283447,
"learning_rate": 4.340705274357805e-05,
"loss": 0.682,
"step": 3200
},
{
"epoch": 2.2058823529411766,
"grad_norm": 6.284212112426758,
"learning_rate": 4.3312053503571976e-05,
"loss": 0.5189,
"step": 3225
},
{
"epoch": 2.2229822161422708,
"grad_norm": 4.796578884124756,
"learning_rate": 4.321705426356589e-05,
"loss": 0.5516,
"step": 3250
},
{
"epoch": 2.2400820793433653,
"grad_norm": 0.08828147500753403,
"learning_rate": 4.312205502355982e-05,
"loss": 0.4325,
"step": 3275
},
{
"epoch": 2.2571819425444595,
"grad_norm": 3.709010362625122,
"learning_rate": 4.3027055783553734e-05,
"loss": 0.524,
"step": 3300
},
{
"epoch": 2.274281805745554,
"grad_norm": 6.48209810256958,
"learning_rate": 4.293205654354766e-05,
"loss": 0.4304,
"step": 3325
},
{
"epoch": 2.2913816689466486,
"grad_norm": 7.461411952972412,
"learning_rate": 4.2837057303541574e-05,
"loss": 0.5894,
"step": 3350
},
{
"epoch": 2.3084815321477428,
"grad_norm": 6.687474727630615,
"learning_rate": 4.27420580635355e-05,
"loss": 0.4504,
"step": 3375
},
{
"epoch": 2.3255813953488373,
"grad_norm": 4.4408745765686035,
"learning_rate": 4.2647058823529415e-05,
"loss": 0.5091,
"step": 3400
},
{
"epoch": 2.3426812585499315,
"grad_norm": 7.8160552978515625,
"learning_rate": 4.255205958352333e-05,
"loss": 0.3612,
"step": 3425
},
{
"epoch": 2.359781121751026,
"grad_norm": 4.452754497528076,
"learning_rate": 4.2457060343517255e-05,
"loss": 0.6184,
"step": 3450
},
{
"epoch": 2.37688098495212,
"grad_norm": 12.292354583740234,
"learning_rate": 4.236206110351117e-05,
"loss": 0.5868,
"step": 3475
},
{
"epoch": 2.3939808481532148,
"grad_norm": 3.688055992126465,
"learning_rate": 4.226706186350509e-05,
"loss": 0.5325,
"step": 3500
},
{
"epoch": 2.4110807113543093,
"grad_norm": 5.77033805847168,
"learning_rate": 4.217206262349901e-05,
"loss": 0.3577,
"step": 3525
},
{
"epoch": 2.4281805745554035,
"grad_norm": 4.816909313201904,
"learning_rate": 4.207706338349293e-05,
"loss": 0.4681,
"step": 3550
},
{
"epoch": 2.445280437756498,
"grad_norm": 8.695076942443848,
"learning_rate": 4.198206414348685e-05,
"loss": 0.4151,
"step": 3575
},
{
"epoch": 2.462380300957592,
"grad_norm": 10.278604507446289,
"learning_rate": 4.188706490348077e-05,
"loss": 0.4165,
"step": 3600
},
{
"epoch": 2.4794801641586868,
"grad_norm": 17.01686668395996,
"learning_rate": 4.1792065663474694e-05,
"loss": 0.5273,
"step": 3625
},
{
"epoch": 2.496580027359781,
"grad_norm": 7.072406768798828,
"learning_rate": 4.170086639306886e-05,
"loss": 0.4383,
"step": 3650
},
{
"epoch": 2.5136798905608755,
"grad_norm": 5.283012866973877,
"learning_rate": 4.1605867153062774e-05,
"loss": 0.578,
"step": 3675
},
{
"epoch": 2.53077975376197,
"grad_norm": 3.6656084060668945,
"learning_rate": 4.15108679130567e-05,
"loss": 0.491,
"step": 3700
},
{
"epoch": 2.547879616963064,
"grad_norm": 9.410102844238281,
"learning_rate": 4.1415868673050615e-05,
"loss": 0.4382,
"step": 3725
},
{
"epoch": 2.5649794801641588,
"grad_norm": 8.41618824005127,
"learning_rate": 4.132086943304454e-05,
"loss": 0.5658,
"step": 3750
},
{
"epoch": 2.582079343365253,
"grad_norm": 1.3863807916641235,
"learning_rate": 4.1225870193038455e-05,
"loss": 0.5228,
"step": 3775
},
{
"epoch": 2.5991792065663475,
"grad_norm": 3.929469585418701,
"learning_rate": 4.113087095303238e-05,
"loss": 0.4993,
"step": 3800
},
{
"epoch": 2.616279069767442,
"grad_norm": 4.706065654754639,
"learning_rate": 4.1035871713026296e-05,
"loss": 0.4045,
"step": 3825
},
{
"epoch": 2.633378932968536,
"grad_norm": 2.805471658706665,
"learning_rate": 4.094087247302022e-05,
"loss": 0.4283,
"step": 3850
},
{
"epoch": 2.650478796169631,
"grad_norm": 8.297080993652344,
"learning_rate": 4.0845873233014136e-05,
"loss": 0.4499,
"step": 3875
},
{
"epoch": 2.667578659370725,
"grad_norm": 2.637622594833374,
"learning_rate": 4.075087399300806e-05,
"loss": 0.5087,
"step": 3900
},
{
"epoch": 2.6846785225718195,
"grad_norm": 2.461876392364502,
"learning_rate": 4.065587475300198e-05,
"loss": 0.5104,
"step": 3925
},
{
"epoch": 2.701778385772914,
"grad_norm": 6.3255486488342285,
"learning_rate": 4.05608755129959e-05,
"loss": 0.5861,
"step": 3950
},
{
"epoch": 2.718878248974008,
"grad_norm": 11.981715202331543,
"learning_rate": 4.046587627298982e-05,
"loss": 0.5593,
"step": 3975
},
{
"epoch": 2.7359781121751023,
"grad_norm": 7.206056594848633,
"learning_rate": 4.037087703298374e-05,
"loss": 0.4665,
"step": 4000
},
{
"epoch": 2.753077975376197,
"grad_norm": 1.9749879837036133,
"learning_rate": 4.027587779297766e-05,
"loss": 0.4984,
"step": 4025
},
{
"epoch": 2.7701778385772915,
"grad_norm": 4.363369464874268,
"learning_rate": 4.018087855297158e-05,
"loss": 0.5785,
"step": 4050
},
{
"epoch": 2.7872777017783856,
"grad_norm": 6.443023681640625,
"learning_rate": 4.00858793129655e-05,
"loss": 0.5287,
"step": 4075
},
{
"epoch": 2.80437756497948,
"grad_norm": 6.106603622436523,
"learning_rate": 3.999088007295942e-05,
"loss": 0.6016,
"step": 4100
},
{
"epoch": 2.8214774281805743,
"grad_norm": 9.416887283325195,
"learning_rate": 3.989588083295334e-05,
"loss": 0.5532,
"step": 4125
},
{
"epoch": 2.838577291381669,
"grad_norm": 2.5579991340637207,
"learning_rate": 3.9800881592947256e-05,
"loss": 0.4579,
"step": 4150
},
{
"epoch": 2.8556771545827635,
"grad_norm": 10.23747444152832,
"learning_rate": 3.970588235294117e-05,
"loss": 0.5402,
"step": 4175
},
{
"epoch": 2.8727770177838576,
"grad_norm": 9.435909271240234,
"learning_rate": 3.96108831129351e-05,
"loss": 0.4845,
"step": 4200
},
{
"epoch": 2.889876880984952,
"grad_norm": 3.5227017402648926,
"learning_rate": 3.9515883872929014e-05,
"loss": 0.3969,
"step": 4225
},
{
"epoch": 2.9069767441860463,
"grad_norm": 2.499969482421875,
"learning_rate": 3.942088463292294e-05,
"loss": 0.4392,
"step": 4250
},
{
"epoch": 2.924076607387141,
"grad_norm": 8.345962524414062,
"learning_rate": 3.9325885392916854e-05,
"loss": 0.4584,
"step": 4275
},
{
"epoch": 2.9411764705882355,
"grad_norm": 3.583355188369751,
"learning_rate": 3.923088615291078e-05,
"loss": 0.3816,
"step": 4300
},
{
"epoch": 2.9582763337893296,
"grad_norm": 1.7532494068145752,
"learning_rate": 3.91358869129047e-05,
"loss": 0.3497,
"step": 4325
},
{
"epoch": 2.975376196990424,
"grad_norm": 9.220611572265625,
"learning_rate": 3.904088767289862e-05,
"loss": 0.5747,
"step": 4350
},
{
"epoch": 2.9924760601915183,
"grad_norm": 1.715482234954834,
"learning_rate": 3.894588843289254e-05,
"loss": 0.4701,
"step": 4375
},
{
"epoch": 3.0,
"eval_accuracy": 0.9071355411403209,
"eval_f1_macro": 0.7615092412094518,
"eval_f1_micro": 0.9071355411403209,
"eval_f1_weighted": 0.9053250303728207,
"eval_loss": 0.2766139805316925,
"eval_precision_macro": 0.8223788212097681,
"eval_precision_micro": 0.9071355411403209,
"eval_precision_weighted": 0.9095949572268466,
"eval_recall_macro": 0.7322063659665187,
"eval_recall_micro": 0.9071355411403209,
"eval_recall_weighted": 0.9071355411403209,
"eval_runtime": 29.8386,
"eval_samples_per_second": 98.161,
"eval_steps_per_second": 6.167,
"step": 4386
},
{
"epoch": 3.009575923392613,
"grad_norm": 4.551536560058594,
"learning_rate": 3.885088919288646e-05,
"loss": 0.4776,
"step": 4400
},
{
"epoch": 3.026675786593707,
"grad_norm": 10.111888885498047,
"learning_rate": 3.875588995288038e-05,
"loss": 0.5966,
"step": 4425
},
{
"epoch": 3.0437756497948016,
"grad_norm": 13.998083114624023,
"learning_rate": 3.86608907128743e-05,
"loss": 0.4699,
"step": 4450
},
{
"epoch": 3.060875512995896,
"grad_norm": 10.541669845581055,
"learning_rate": 3.856589147286822e-05,
"loss": 0.5154,
"step": 4475
},
{
"epoch": 3.0779753761969904,
"grad_norm": 3.1978766918182373,
"learning_rate": 3.847089223286214e-05,
"loss": 0.4304,
"step": 4500
},
{
"epoch": 3.095075239398085,
"grad_norm": 5.206043720245361,
"learning_rate": 3.8375892992856064e-05,
"loss": 0.3223,
"step": 4525
},
{
"epoch": 3.112175102599179,
"grad_norm": 9.216475486755371,
"learning_rate": 3.828089375284998e-05,
"loss": 0.6146,
"step": 4550
},
{
"epoch": 3.1292749658002736,
"grad_norm": 4.771245002746582,
"learning_rate": 3.81858945128439e-05,
"loss": 0.3407,
"step": 4575
},
{
"epoch": 3.146374829001368,
"grad_norm": 8.766341209411621,
"learning_rate": 3.8090895272837814e-05,
"loss": 0.4866,
"step": 4600
},
{
"epoch": 3.1634746922024624,
"grad_norm": 6.949918270111084,
"learning_rate": 3.799589603283174e-05,
"loss": 0.4899,
"step": 4625
},
{
"epoch": 3.180574555403557,
"grad_norm": 1.0002543926239014,
"learning_rate": 3.7900896792825655e-05,
"loss": 0.481,
"step": 4650
},
{
"epoch": 3.197674418604651,
"grad_norm": 9.768813133239746,
"learning_rate": 3.780589755281958e-05,
"loss": 0.4338,
"step": 4675
},
{
"epoch": 3.2147742818057456,
"grad_norm": 4.305062770843506,
"learning_rate": 3.7710898312813495e-05,
"loss": 0.3783,
"step": 4700
},
{
"epoch": 3.23187414500684,
"grad_norm": 0.7493396997451782,
"learning_rate": 3.761589907280742e-05,
"loss": 0.5001,
"step": 4725
},
{
"epoch": 3.2489740082079344,
"grad_norm": 5.481766700744629,
"learning_rate": 3.7520899832801336e-05,
"loss": 0.4894,
"step": 4750
},
{
"epoch": 3.266073871409029,
"grad_norm": 3.8942747116088867,
"learning_rate": 3.742590059279526e-05,
"loss": 0.4761,
"step": 4775
},
{
"epoch": 3.283173734610123,
"grad_norm": 8.91220760345459,
"learning_rate": 3.7330901352789176e-05,
"loss": 0.5526,
"step": 4800
},
{
"epoch": 3.3002735978112177,
"grad_norm": 1.765817642211914,
"learning_rate": 3.72359021127831e-05,
"loss": 0.5665,
"step": 4825
},
{
"epoch": 3.317373461012312,
"grad_norm": 6.553985118865967,
"learning_rate": 3.714090287277702e-05,
"loss": 0.287,
"step": 4850
},
{
"epoch": 3.3344733242134064,
"grad_norm": 5.839998245239258,
"learning_rate": 3.704590363277094e-05,
"loss": 0.4927,
"step": 4875
},
{
"epoch": 3.3515731874145005,
"grad_norm": 8.003843307495117,
"learning_rate": 3.6950904392764864e-05,
"loss": 0.552,
"step": 4900
},
{
"epoch": 3.368673050615595,
"grad_norm": 3.3592958450317383,
"learning_rate": 3.685590515275878e-05,
"loss": 0.5703,
"step": 4925
},
{
"epoch": 3.3857729138166897,
"grad_norm": 4.747976779937744,
"learning_rate": 3.6760905912752705e-05,
"loss": 0.5218,
"step": 4950
},
{
"epoch": 3.402872777017784,
"grad_norm": 4.351706504821777,
"learning_rate": 3.666590667274662e-05,
"loss": 0.4671,
"step": 4975
},
{
"epoch": 3.4199726402188784,
"grad_norm": 5.302130222320557,
"learning_rate": 3.6570907432740545e-05,
"loss": 0.3913,
"step": 5000
},
{
"epoch": 3.4370725034199725,
"grad_norm": 1.5275038480758667,
"learning_rate": 3.647590819273446e-05,
"loss": 0.3149,
"step": 5025
},
{
"epoch": 3.454172366621067,
"grad_norm": 0.8008119463920593,
"learning_rate": 3.638090895272838e-05,
"loss": 0.5232,
"step": 5050
},
{
"epoch": 3.471272229822161,
"grad_norm": 6.361542701721191,
"learning_rate": 3.6285909712722296e-05,
"loss": 0.4149,
"step": 5075
},
{
"epoch": 3.488372093023256,
"grad_norm": 0.8227640390396118,
"learning_rate": 3.619091047271622e-05,
"loss": 0.5698,
"step": 5100
},
{
"epoch": 3.5054719562243504,
"grad_norm": 4.734671115875244,
"learning_rate": 3.609591123271014e-05,
"loss": 0.4835,
"step": 5125
},
{
"epoch": 3.5225718194254445,
"grad_norm": 6.9522576332092285,
"learning_rate": 3.600091199270406e-05,
"loss": 0.5387,
"step": 5150
},
{
"epoch": 3.539671682626539,
"grad_norm": 3.4246246814727783,
"learning_rate": 3.590591275269798e-05,
"loss": 0.4044,
"step": 5175
},
{
"epoch": 3.556771545827633,
"grad_norm": 6.183599472045898,
"learning_rate": 3.58109135126919e-05,
"loss": 0.4785,
"step": 5200
},
{
"epoch": 3.573871409028728,
"grad_norm": 3.738954782485962,
"learning_rate": 3.571591427268582e-05,
"loss": 0.5516,
"step": 5225
},
{
"epoch": 3.5909712722298224,
"grad_norm": 7.439993381500244,
"learning_rate": 3.562091503267974e-05,
"loss": 0.523,
"step": 5250
},
{
"epoch": 3.6080711354309165,
"grad_norm": 4.419105052947998,
"learning_rate": 3.552591579267366e-05,
"loss": 0.4118,
"step": 5275
},
{
"epoch": 3.625170998632011,
"grad_norm": 16.105833053588867,
"learning_rate": 3.543091655266758e-05,
"loss": 0.5083,
"step": 5300
},
{
"epoch": 3.6422708618331052,
"grad_norm": 5.370678424835205,
"learning_rate": 3.53359173126615e-05,
"loss": 0.4519,
"step": 5325
},
{
"epoch": 3.6593707250342,
"grad_norm": 11.25657844543457,
"learning_rate": 3.524091807265542e-05,
"loss": 0.3801,
"step": 5350
},
{
"epoch": 3.6764705882352944,
"grad_norm": 3.6341259479522705,
"learning_rate": 3.514591883264934e-05,
"loss": 0.3978,
"step": 5375
},
{
"epoch": 3.6935704514363885,
"grad_norm": 0.18588456511497498,
"learning_rate": 3.505091959264326e-05,
"loss": 0.4341,
"step": 5400
},
{
"epoch": 3.7106703146374826,
"grad_norm": 4.884586811065674,
"learning_rate": 3.495592035263718e-05,
"loss": 0.5095,
"step": 5425
},
{
"epoch": 3.7277701778385772,
"grad_norm": 5.724035739898682,
"learning_rate": 3.4860921112631104e-05,
"loss": 0.4625,
"step": 5450
},
{
"epoch": 3.744870041039672,
"grad_norm": 5.873854637145996,
"learning_rate": 3.476592187262502e-05,
"loss": 0.4019,
"step": 5475
},
{
"epoch": 3.761969904240766,
"grad_norm": 3.843010902404785,
"learning_rate": 3.467092263261894e-05,
"loss": 0.5835,
"step": 5500
},
{
"epoch": 3.7790697674418605,
"grad_norm": 5.200079441070557,
"learning_rate": 3.457592339261286e-05,
"loss": 0.4441,
"step": 5525
},
{
"epoch": 3.7961696306429547,
"grad_norm": 7.437492847442627,
"learning_rate": 3.448092415260678e-05,
"loss": 0.3881,
"step": 5550
},
{
"epoch": 3.8132694938440492,
"grad_norm": 5.660449504852295,
"learning_rate": 3.43859249126007e-05,
"loss": 0.446,
"step": 5575
},
{
"epoch": 3.830369357045144,
"grad_norm": 2.2128264904022217,
"learning_rate": 3.429092567259462e-05,
"loss": 0.4257,
"step": 5600
},
{
"epoch": 3.847469220246238,
"grad_norm": 5.218535423278809,
"learning_rate": 3.419592643258854e-05,
"loss": 0.4946,
"step": 5625
},
{
"epoch": 3.8645690834473325,
"grad_norm": 1.4989601373672485,
"learning_rate": 3.410092719258246e-05,
"loss": 0.3866,
"step": 5650
},
{
"epoch": 3.8816689466484267,
"grad_norm": 11.225085258483887,
"learning_rate": 3.400592795257638e-05,
"loss": 0.4764,
"step": 5675
},
{
"epoch": 3.8987688098495212,
"grad_norm": 5.653949737548828,
"learning_rate": 3.39109287125703e-05,
"loss": 0.4362,
"step": 5700
},
{
"epoch": 3.915868673050616,
"grad_norm": 10.87972354888916,
"learning_rate": 3.381592947256422e-05,
"loss": 0.4708,
"step": 5725
},
{
"epoch": 3.93296853625171,
"grad_norm": 5.110008239746094,
"learning_rate": 3.372093023255814e-05,
"loss": 0.4373,
"step": 5750
},
{
"epoch": 3.9500683994528045,
"grad_norm": 4.611896991729736,
"learning_rate": 3.3625930992552064e-05,
"loss": 0.4808,
"step": 5775
},
{
"epoch": 3.9671682626538987,
"grad_norm": 3.967752695083618,
"learning_rate": 3.353093175254598e-05,
"loss": 0.3827,
"step": 5800
},
{
"epoch": 3.9842681258549932,
"grad_norm": 7.738473892211914,
"learning_rate": 3.3435932512539904e-05,
"loss": 0.4564,
"step": 5825
},
{
"epoch": 4.0,
"eval_accuracy": 0.8968931375896211,
"eval_f1_macro": 0.6842607301838518,
"eval_f1_micro": 0.8968931375896211,
"eval_f1_weighted": 0.8922998232908407,
"eval_loss": 0.3079198896884918,
"eval_precision_macro": 0.7711077815931238,
"eval_precision_micro": 0.8968931375896211,
"eval_precision_weighted": 0.9026293299812516,
"eval_recall_macro": 0.6595431638738259,
"eval_recall_micro": 0.8968931375896211,
"eval_recall_weighted": 0.8968931375896211,
"eval_runtime": 29.6904,
"eval_samples_per_second": 98.651,
"eval_steps_per_second": 6.197,
"step": 5848
},
{
"epoch": 4.001367989056088,
"grad_norm": 12.561144828796387,
"learning_rate": 3.334093327253382e-05,
"loss": 0.285,
"step": 5850
},
{
"epoch": 4.0184678522571815,
"grad_norm": 4.646189212799072,
"learning_rate": 3.3245934032527745e-05,
"loss": 0.4118,
"step": 5875
},
{
"epoch": 4.035567715458276,
"grad_norm": 3.1821000576019287,
"learning_rate": 3.315093479252166e-05,
"loss": 0.5941,
"step": 5900
},
{
"epoch": 4.052667578659371,
"grad_norm": 6.912432670593262,
"learning_rate": 3.3055935552515586e-05,
"loss": 0.3519,
"step": 5925
},
{
"epoch": 4.069767441860465,
"grad_norm": 2.416905641555786,
"learning_rate": 3.29609363125095e-05,
"loss": 0.4982,
"step": 5950
},
{
"epoch": 4.08686730506156,
"grad_norm": 8.833620071411133,
"learning_rate": 3.286593707250342e-05,
"loss": 0.3608,
"step": 5975
},
{
"epoch": 4.1039671682626535,
"grad_norm": 0.06249883025884628,
"learning_rate": 3.2770937832497336e-05,
"loss": 0.4883,
"step": 6000
},
{
"epoch": 4.121067031463748,
"grad_norm": 2.5257112979888916,
"learning_rate": 3.267593859249126e-05,
"loss": 0.3371,
"step": 6025
},
{
"epoch": 4.138166894664843,
"grad_norm": 5.706219673156738,
"learning_rate": 3.258093935248518e-05,
"loss": 0.3998,
"step": 6050
},
{
"epoch": 4.155266757865937,
"grad_norm": 9.878058433532715,
"learning_rate": 3.24859401124791e-05,
"loss": 0.5082,
"step": 6075
},
{
"epoch": 4.172366621067032,
"grad_norm": 0.6408124566078186,
"learning_rate": 3.239094087247302e-05,
"loss": 0.2915,
"step": 6100
},
{
"epoch": 4.1894664842681255,
"grad_norm": 4.155546188354492,
"learning_rate": 3.229594163246694e-05,
"loss": 0.421,
"step": 6125
},
{
"epoch": 4.20656634746922,
"grad_norm": 2.4020016193389893,
"learning_rate": 3.2200942392460865e-05,
"loss": 0.3445,
"step": 6150
},
{
"epoch": 4.223666210670315,
"grad_norm": 9.44356632232666,
"learning_rate": 3.210594315245478e-05,
"loss": 0.478,
"step": 6175
},
{
"epoch": 4.240766073871409,
"grad_norm": 6.012132167816162,
"learning_rate": 3.2010943912448705e-05,
"loss": 0.4779,
"step": 6200
},
{
"epoch": 4.257865937072504,
"grad_norm": 5.681918621063232,
"learning_rate": 3.191594467244262e-05,
"loss": 0.5116,
"step": 6225
},
{
"epoch": 4.2749658002735975,
"grad_norm": 7.404222011566162,
"learning_rate": 3.1820945432436546e-05,
"loss": 0.3494,
"step": 6250
},
{
"epoch": 4.292065663474692,
"grad_norm": 14.846484184265137,
"learning_rate": 3.172594619243046e-05,
"loss": 0.5149,
"step": 6275
},
{
"epoch": 4.309165526675787,
"grad_norm": 1.3198221921920776,
"learning_rate": 3.1630946952424386e-05,
"loss": 0.4053,
"step": 6300
},
{
"epoch": 4.326265389876881,
"grad_norm": 3.9458107948303223,
"learning_rate": 3.15359477124183e-05,
"loss": 0.4499,
"step": 6325
},
{
"epoch": 4.343365253077975,
"grad_norm": 4.272012233734131,
"learning_rate": 3.144094847241223e-05,
"loss": 0.4298,
"step": 6350
},
{
"epoch": 4.3604651162790695,
"grad_norm": 8.574115753173828,
"learning_rate": 3.1345949232406144e-05,
"loss": 0.3282,
"step": 6375
},
{
"epoch": 4.377564979480164,
"grad_norm": 10.030281066894531,
"learning_rate": 3.125094999240006e-05,
"loss": 0.486,
"step": 6400
},
{
"epoch": 4.394664842681259,
"grad_norm": 8.065329551696777,
"learning_rate": 3.115595075239398e-05,
"loss": 0.332,
"step": 6425
},
{
"epoch": 4.411764705882353,
"grad_norm": 10.601835250854492,
"learning_rate": 3.10609515123879e-05,
"loss": 0.4674,
"step": 6450
},
{
"epoch": 4.428864569083447,
"grad_norm": 4.035892486572266,
"learning_rate": 3.096595227238182e-05,
"loss": 0.5078,
"step": 6475
},
{
"epoch": 4.4459644322845415,
"grad_norm": 6.8672685623168945,
"learning_rate": 3.087095303237574e-05,
"loss": 0.3255,
"step": 6500
},
{
"epoch": 4.463064295485636,
"grad_norm": 4.926382541656494,
"learning_rate": 3.077595379236966e-05,
"loss": 0.4204,
"step": 6525
},
{
"epoch": 4.480164158686731,
"grad_norm": 4.504785060882568,
"learning_rate": 3.068095455236358e-05,
"loss": 0.4257,
"step": 6550
},
{
"epoch": 4.497264021887825,
"grad_norm": 10.662145614624023,
"learning_rate": 3.05859553123575e-05,
"loss": 0.4991,
"step": 6575
},
{
"epoch": 4.514363885088919,
"grad_norm": 4.892814636230469,
"learning_rate": 3.0490956072351423e-05,
"loss": 0.3588,
"step": 6600
},
{
"epoch": 4.5314637482900135,
"grad_norm": 2.461637258529663,
"learning_rate": 3.039595683234534e-05,
"loss": 0.3509,
"step": 6625
},
{
"epoch": 4.548563611491108,
"grad_norm": 0.9665390253067017,
"learning_rate": 3.0300957592339263e-05,
"loss": 0.4587,
"step": 6650
},
{
"epoch": 4.565663474692203,
"grad_norm": 5.2920002937316895,
"learning_rate": 3.020595835233318e-05,
"loss": 0.3853,
"step": 6675
},
{
"epoch": 4.582763337893297,
"grad_norm": 6.542552471160889,
"learning_rate": 3.0110959112327104e-05,
"loss": 0.475,
"step": 6700
},
{
"epoch": 4.599863201094391,
"grad_norm": 3.7444801330566406,
"learning_rate": 3.0015959872321024e-05,
"loss": 0.3638,
"step": 6725
},
{
"epoch": 4.6169630642954855,
"grad_norm": 6.542828559875488,
"learning_rate": 2.992096063231494e-05,
"loss": 0.4892,
"step": 6750
},
{
"epoch": 4.63406292749658,
"grad_norm": 1.4131304025650024,
"learning_rate": 2.9825961392308865e-05,
"loss": 0.4617,
"step": 6775
},
{
"epoch": 4.651162790697675,
"grad_norm": 2.6974217891693115,
"learning_rate": 2.973096215230278e-05,
"loss": 0.3848,
"step": 6800
},
{
"epoch": 4.668262653898768,
"grad_norm": 5.639090061187744,
"learning_rate": 2.9635962912296705e-05,
"loss": 0.3584,
"step": 6825
},
{
"epoch": 4.685362517099863,
"grad_norm": 11.76826000213623,
"learning_rate": 2.9540963672290622e-05,
"loss": 0.4498,
"step": 6850
},
{
"epoch": 4.7024623803009575,
"grad_norm": 3.208690643310547,
"learning_rate": 2.9445964432284546e-05,
"loss": 0.3597,
"step": 6875
},
{
"epoch": 4.719562243502052,
"grad_norm": 5.359434604644775,
"learning_rate": 2.9350965192278463e-05,
"loss": 0.3621,
"step": 6900
},
{
"epoch": 4.736662106703147,
"grad_norm": 6.385953426361084,
"learning_rate": 2.9255965952272386e-05,
"loss": 0.4097,
"step": 6925
},
{
"epoch": 4.75376196990424,
"grad_norm": 10.398621559143066,
"learning_rate": 2.9160966712266303e-05,
"loss": 0.3647,
"step": 6950
},
{
"epoch": 4.770861833105335,
"grad_norm": 2.8369972705841064,
"learning_rate": 2.9065967472260224e-05,
"loss": 0.3396,
"step": 6975
},
{
"epoch": 4.7879616963064295,
"grad_norm": 6.925849437713623,
"learning_rate": 2.897096823225414e-05,
"loss": 0.3689,
"step": 7000
},
{
"epoch": 4.805061559507524,
"grad_norm": 3.137173652648926,
"learning_rate": 2.8875968992248064e-05,
"loss": 0.5454,
"step": 7025
},
{
"epoch": 4.822161422708619,
"grad_norm": 12.329541206359863,
"learning_rate": 2.878096975224198e-05,
"loss": 0.417,
"step": 7050
},
{
"epoch": 4.839261285909712,
"grad_norm": 1.9504649639129639,
"learning_rate": 2.8685970512235905e-05,
"loss": 0.3552,
"step": 7075
},
{
"epoch": 4.856361149110807,
"grad_norm": 6.352153778076172,
"learning_rate": 2.859097127222982e-05,
"loss": 0.4427,
"step": 7100
},
{
"epoch": 4.8734610123119015,
"grad_norm": 5.652337074279785,
"learning_rate": 2.8495972032223745e-05,
"loss": 0.4454,
"step": 7125
},
{
"epoch": 4.890560875512996,
"grad_norm": 4.495980262756348,
"learning_rate": 2.8400972792217662e-05,
"loss": 0.4396,
"step": 7150
},
{
"epoch": 4.907660738714091,
"grad_norm": 9.805567741394043,
"learning_rate": 2.8305973552211586e-05,
"loss": 0.4436,
"step": 7175
},
{
"epoch": 4.924760601915184,
"grad_norm": 2.151742696762085,
"learning_rate": 2.8210974312205503e-05,
"loss": 0.4238,
"step": 7200
},
{
"epoch": 4.941860465116279,
"grad_norm": 4.97649621963501,
"learning_rate": 2.8115975072199423e-05,
"loss": 0.4693,
"step": 7225
},
{
"epoch": 4.9589603283173735,
"grad_norm": 2.5066914558410645,
"learning_rate": 2.802097583219334e-05,
"loss": 0.4568,
"step": 7250
},
{
"epoch": 4.976060191518468,
"grad_norm": 13.676875114440918,
"learning_rate": 2.7925976592187263e-05,
"loss": 0.3527,
"step": 7275
},
{
"epoch": 4.993160054719562,
"grad_norm": 3.240598678588867,
"learning_rate": 2.783097735218118e-05,
"loss": 0.4515,
"step": 7300
},
{
"epoch": 5.0,
"eval_accuracy": 0.9166951177876408,
"eval_f1_macro": 0.7653415866627981,
"eval_f1_micro": 0.9166951177876408,
"eval_f1_weighted": 0.9141778991500009,
"eval_loss": 0.24330385029315948,
"eval_precision_macro": 0.794330339771912,
"eval_precision_micro": 0.9166951177876408,
"eval_precision_weighted": 0.9162095802352392,
"eval_recall_macro": 0.7496536933673316,
"eval_recall_micro": 0.9166951177876408,
"eval_recall_weighted": 0.9166951177876408,
"eval_runtime": 29.505,
"eval_samples_per_second": 99.271,
"eval_steps_per_second": 6.236,
"step": 7310
},
{
"epoch": 5.010259917920656,
"grad_norm": 3.2240583896636963,
"learning_rate": 2.7735978112175104e-05,
"loss": 0.406,
"step": 7325
},
{
"epoch": 5.027359781121751,
"grad_norm": 1.1182893514633179,
"learning_rate": 2.7640978872169028e-05,
"loss": 0.4151,
"step": 7350
},
{
"epoch": 5.0444596443228455,
"grad_norm": 3.5205605030059814,
"learning_rate": 2.7545979632162945e-05,
"loss": 0.4349,
"step": 7375
},
{
"epoch": 5.06155950752394,
"grad_norm": 5.898115634918213,
"learning_rate": 2.7450980392156865e-05,
"loss": 0.4127,
"step": 7400
},
{
"epoch": 5.078659370725034,
"grad_norm": 2.958627700805664,
"learning_rate": 2.7355981152150785e-05,
"loss": 0.4479,
"step": 7425
},
{
"epoch": 5.095759233926128,
"grad_norm": 4.672809600830078,
"learning_rate": 2.7260981912144705e-05,
"loss": 0.4976,
"step": 7450
},
{
"epoch": 5.112859097127223,
"grad_norm": 8.304091453552246,
"learning_rate": 2.7165982672138622e-05,
"loss": 0.4065,
"step": 7475
},
{
"epoch": 5.1299589603283176,
"grad_norm": 2.276463270187378,
"learning_rate": 2.7070983432132546e-05,
"loss": 0.3508,
"step": 7500
},
{
"epoch": 5.147058823529412,
"grad_norm": 0.10079041868448257,
"learning_rate": 2.6975984192126463e-05,
"loss": 0.4274,
"step": 7525
},
{
"epoch": 5.164158686730506,
"grad_norm": 5.3365936279296875,
"learning_rate": 2.6880984952120387e-05,
"loss": 0.2555,
"step": 7550
},
{
"epoch": 5.1812585499316,
"grad_norm": 0.33459940552711487,
"learning_rate": 2.6785985712114303e-05,
"loss": 0.4309,
"step": 7575
},
{
"epoch": 5.198358413132695,
"grad_norm": 6.256367206573486,
"learning_rate": 2.6690986472108227e-05,
"loss": 0.3263,
"step": 7600
},
{
"epoch": 5.2154582763337896,
"grad_norm": 2.592665672302246,
"learning_rate": 2.6595987232102144e-05,
"loss": 0.4089,
"step": 7625
},
{
"epoch": 5.232558139534884,
"grad_norm": 5.69423246383667,
"learning_rate": 2.6500987992096064e-05,
"loss": 0.4288,
"step": 7650
},
{
"epoch": 5.249658002735978,
"grad_norm": 7.463606834411621,
"learning_rate": 2.6405988752089985e-05,
"loss": 0.35,
"step": 7675
},
{
"epoch": 5.266757865937072,
"grad_norm": 4.466919898986816,
"learning_rate": 2.6310989512083905e-05,
"loss": 0.4364,
"step": 7700
},
{
"epoch": 5.283857729138167,
"grad_norm": 5.60076904296875,
"learning_rate": 2.621599027207782e-05,
"loss": 0.2817,
"step": 7725
},
{
"epoch": 5.300957592339262,
"grad_norm": 1.8708001375198364,
"learning_rate": 2.6120991032071745e-05,
"loss": 0.3567,
"step": 7750
},
{
"epoch": 5.318057455540355,
"grad_norm": 4.970040798187256,
"learning_rate": 2.6025991792065662e-05,
"loss": 0.2724,
"step": 7775
},
{
"epoch": 5.33515731874145,
"grad_norm": 1.4727894067764282,
"learning_rate": 2.5930992552059586e-05,
"loss": 0.3978,
"step": 7800
},
{
"epoch": 5.352257181942544,
"grad_norm": 4.697504043579102,
"learning_rate": 2.5835993312053503e-05,
"loss": 0.447,
"step": 7825
},
{
"epoch": 5.369357045143639,
"grad_norm": 2.3965351581573486,
"learning_rate": 2.5740994072047426e-05,
"loss": 0.3596,
"step": 7850
},
{
"epoch": 5.386456908344734,
"grad_norm": 8.727534294128418,
"learning_rate": 2.5645994832041343e-05,
"loss": 0.3787,
"step": 7875
},
{
"epoch": 5.403556771545827,
"grad_norm": 5.903003692626953,
"learning_rate": 2.5550995592035264e-05,
"loss": 0.3899,
"step": 7900
},
{
"epoch": 5.420656634746922,
"grad_norm": 4.570006847381592,
"learning_rate": 2.5455996352029187e-05,
"loss": 0.4558,
"step": 7925
},
{
"epoch": 5.437756497948016,
"grad_norm": 0.9773057699203491,
"learning_rate": 2.5360997112023104e-05,
"loss": 0.3345,
"step": 7950
},
{
"epoch": 5.454856361149111,
"grad_norm": 3.641374349594116,
"learning_rate": 2.5265997872017028e-05,
"loss": 0.4155,
"step": 7975
},
{
"epoch": 5.471956224350206,
"grad_norm": 5.382412433624268,
"learning_rate": 2.5170998632010945e-05,
"loss": 0.4111,
"step": 8000
},
{
"epoch": 5.489056087551299,
"grad_norm": 2.235764980316162,
"learning_rate": 2.507599939200487e-05,
"loss": 0.3047,
"step": 8025
},
{
"epoch": 5.506155950752394,
"grad_norm": 3.8063552379608154,
"learning_rate": 2.4981000151998785e-05,
"loss": 0.4702,
"step": 8050
},
{
"epoch": 5.523255813953488,
"grad_norm": 3.573058843612671,
"learning_rate": 2.4886000911992706e-05,
"loss": 0.4619,
"step": 8075
},
{
"epoch": 5.540355677154583,
"grad_norm": 7.515729904174805,
"learning_rate": 2.4791001671986626e-05,
"loss": 0.3179,
"step": 8100
},
{
"epoch": 5.557455540355678,
"grad_norm": 1.7556642293930054,
"learning_rate": 2.4696002431980546e-05,
"loss": 0.3382,
"step": 8125
},
{
"epoch": 5.574555403556771,
"grad_norm": 2.0541181564331055,
"learning_rate": 2.4601003191974463e-05,
"loss": 0.4592,
"step": 8150
},
{
"epoch": 5.591655266757866,
"grad_norm": 2.7318685054779053,
"learning_rate": 2.4506003951968383e-05,
"loss": 0.4169,
"step": 8175
},
{
"epoch": 5.60875512995896,
"grad_norm": 8.095924377441406,
"learning_rate": 2.4411004711962304e-05,
"loss": 0.4333,
"step": 8200
},
{
"epoch": 5.625854993160055,
"grad_norm": 2.2528128623962402,
"learning_rate": 2.4316005471956224e-05,
"loss": 0.402,
"step": 8225
},
{
"epoch": 5.642954856361149,
"grad_norm": 10.526260375976562,
"learning_rate": 2.4221006231950147e-05,
"loss": 0.3477,
"step": 8250
},
{
"epoch": 5.660054719562243,
"grad_norm": 5.1542487144470215,
"learning_rate": 2.4126006991944068e-05,
"loss": 0.3982,
"step": 8275
},
{
"epoch": 5.677154582763338,
"grad_norm": 5.238847255706787,
"learning_rate": 2.4031007751937988e-05,
"loss": 0.416,
"step": 8300
},
{
"epoch": 5.694254445964432,
"grad_norm": 8.261098861694336,
"learning_rate": 2.3936008511931908e-05,
"loss": 0.2776,
"step": 8325
},
{
"epoch": 5.711354309165527,
"grad_norm": 6.928776741027832,
"learning_rate": 2.3841009271925825e-05,
"loss": 0.3845,
"step": 8350
},
{
"epoch": 5.728454172366621,
"grad_norm": 9.97043514251709,
"learning_rate": 2.3746010031919745e-05,
"loss": 0.4559,
"step": 8375
},
{
"epoch": 5.745554035567715,
"grad_norm": 1.063759446144104,
"learning_rate": 2.3651010791913666e-05,
"loss": 0.2716,
"step": 8400
},
{
"epoch": 5.76265389876881,
"grad_norm": 2.098356246948242,
"learning_rate": 2.3556011551907586e-05,
"loss": 0.3414,
"step": 8425
},
{
"epoch": 5.779753761969904,
"grad_norm": 5.218485355377197,
"learning_rate": 2.3461012311901506e-05,
"loss": 0.3264,
"step": 8450
},
{
"epoch": 5.796853625170999,
"grad_norm": 5.245093822479248,
"learning_rate": 2.3366013071895427e-05,
"loss": 0.3575,
"step": 8475
},
{
"epoch": 5.813953488372093,
"grad_norm": 2.249113082885742,
"learning_rate": 2.3271013831889347e-05,
"loss": 0.3042,
"step": 8500
},
{
"epoch": 5.831053351573187,
"grad_norm": 8.331680297851562,
"learning_rate": 2.3176014591883267e-05,
"loss": 0.349,
"step": 8525
},
{
"epoch": 5.848153214774282,
"grad_norm": 5.0042805671691895,
"learning_rate": 2.3081015351877187e-05,
"loss": 0.4322,
"step": 8550
},
{
"epoch": 5.865253077975376,
"grad_norm": 15.829472541809082,
"learning_rate": 2.2986016111871104e-05,
"loss": 0.5301,
"step": 8575
},
{
"epoch": 5.882352941176471,
"grad_norm": 5.099222660064697,
"learning_rate": 2.2891016871865025e-05,
"loss": 0.4318,
"step": 8600
},
{
"epoch": 5.899452804377565,
"grad_norm": 3.355231285095215,
"learning_rate": 2.2796017631858945e-05,
"loss": 0.3114,
"step": 8625
},
{
"epoch": 5.916552667578659,
"grad_norm": 5.962894916534424,
"learning_rate": 2.2701018391852865e-05,
"loss": 0.389,
"step": 8650
},
{
"epoch": 5.933652530779754,
"grad_norm": 0.5913941860198975,
"learning_rate": 2.2606019151846785e-05,
"loss": 0.3845,
"step": 8675
},
{
"epoch": 5.950752393980848,
"grad_norm": 2.7873880863189697,
"learning_rate": 2.2511019911840706e-05,
"loss": 0.3045,
"step": 8700
},
{
"epoch": 5.967852257181942,
"grad_norm": 6.107980728149414,
"learning_rate": 2.2416020671834626e-05,
"loss": 0.4016,
"step": 8725
},
{
"epoch": 5.984952120383037,
"grad_norm": 2.5966835021972656,
"learning_rate": 2.2321021431828546e-05,
"loss": 0.3413,
"step": 8750
},
{
"epoch": 6.0,
"eval_accuracy": 0.9095254353021509,
"eval_f1_macro": 0.787176394659271,
"eval_f1_micro": 0.9095254353021509,
"eval_f1_weighted": 0.9072901390204282,
"eval_loss": 0.29091742634773254,
"eval_precision_macro": 0.8953297645243259,
"eval_precision_micro": 0.9095254353021509,
"eval_precision_weighted": 0.9174594106183492,
"eval_recall_macro": 0.7531104129954299,
"eval_recall_micro": 0.9095254353021509,
"eval_recall_weighted": 0.9095254353021509,
"eval_runtime": 29.8878,
"eval_samples_per_second": 98.0,
"eval_steps_per_second": 6.156,
"step": 8772
},
{
"epoch": 6.002051983584131,
"grad_norm": 6.595887660980225,
"learning_rate": 2.2226022191822466e-05,
"loss": 0.3606,
"step": 8775
},
{
"epoch": 6.019151846785226,
"grad_norm": 6.390087127685547,
"learning_rate": 2.2131022951816387e-05,
"loss": 0.4026,
"step": 8800
},
{
"epoch": 6.03625170998632,
"grad_norm": 6.15175199508667,
"learning_rate": 2.2036023711810304e-05,
"loss": 0.3281,
"step": 8825
},
{
"epoch": 6.053351573187414,
"grad_norm": 18.778564453125,
"learning_rate": 2.1941024471804227e-05,
"loss": 0.4304,
"step": 8850
},
{
"epoch": 6.070451436388509,
"grad_norm": 9.964593887329102,
"learning_rate": 2.1846025231798148e-05,
"loss": 0.2677,
"step": 8875
},
{
"epoch": 6.087551299589603,
"grad_norm": 10.952362060546875,
"learning_rate": 2.1751025991792068e-05,
"loss": 0.5143,
"step": 8900
},
{
"epoch": 6.104651162790698,
"grad_norm": 2.978898048400879,
"learning_rate": 2.1656026751785988e-05,
"loss": 0.3024,
"step": 8925
},
{
"epoch": 6.121751025991792,
"grad_norm": 3.3607687950134277,
"learning_rate": 2.156102751177991e-05,
"loss": 0.2684,
"step": 8950
},
{
"epoch": 6.138850889192886,
"grad_norm": 1.9624247550964355,
"learning_rate": 2.146602827177383e-05,
"loss": 0.3062,
"step": 8975
},
{
"epoch": 6.155950752393981,
"grad_norm": 1.7253330945968628,
"learning_rate": 2.137102903176775e-05,
"loss": 0.3242,
"step": 9000
},
{
"epoch": 6.173050615595075,
"grad_norm": 2.6368486881256104,
"learning_rate": 2.1276029791761666e-05,
"loss": 0.3678,
"step": 9025
},
{
"epoch": 6.19015047879617,
"grad_norm": 7.951999664306641,
"learning_rate": 2.1181030551755586e-05,
"loss": 0.3242,
"step": 9050
},
{
"epoch": 6.207250341997264,
"grad_norm": 13.932353019714355,
"learning_rate": 2.1086031311749506e-05,
"loss": 0.244,
"step": 9075
},
{
"epoch": 6.224350205198358,
"grad_norm": 0.4225591719150543,
"learning_rate": 2.0991032071743427e-05,
"loss": 0.503,
"step": 9100
},
{
"epoch": 6.241450068399453,
"grad_norm": 5.91692590713501,
"learning_rate": 2.0896032831737347e-05,
"loss": 0.3721,
"step": 9125
},
{
"epoch": 6.258549931600547,
"grad_norm": 7.2567009925842285,
"learning_rate": 2.0801033591731267e-05,
"loss": 0.3032,
"step": 9150
},
{
"epoch": 6.275649794801642,
"grad_norm": 12.250884056091309,
"learning_rate": 2.0706034351725188e-05,
"loss": 0.4468,
"step": 9175
},
{
"epoch": 6.292749658002736,
"grad_norm": 5.396862030029297,
"learning_rate": 2.0611035111719108e-05,
"loss": 0.357,
"step": 9200
},
{
"epoch": 6.30984952120383,
"grad_norm": 0.8270015120506287,
"learning_rate": 2.0516035871713028e-05,
"loss": 0.2127,
"step": 9225
},
{
"epoch": 6.326949384404925,
"grad_norm": 6.996820449829102,
"learning_rate": 2.042103663170695e-05,
"loss": 0.4843,
"step": 9250
},
{
"epoch": 6.344049247606019,
"grad_norm": 3.144995927810669,
"learning_rate": 2.0326037391700865e-05,
"loss": 0.4409,
"step": 9275
},
{
"epoch": 6.361149110807114,
"grad_norm": 10.23337459564209,
"learning_rate": 2.0231038151694785e-05,
"loss": 0.406,
"step": 9300
},
{
"epoch": 6.378248974008208,
"grad_norm": 3.1300835609436035,
"learning_rate": 2.0136038911688706e-05,
"loss": 0.2881,
"step": 9325
},
{
"epoch": 6.395348837209302,
"grad_norm": 3.322756290435791,
"learning_rate": 2.0041039671682626e-05,
"loss": 0.2679,
"step": 9350
},
{
"epoch": 6.412448700410397,
"grad_norm": 7.402932643890381,
"learning_rate": 1.9946040431676546e-05,
"loss": 0.3229,
"step": 9375
},
{
"epoch": 6.429548563611491,
"grad_norm": 2.389960289001465,
"learning_rate": 1.9851041191670467e-05,
"loss": 0.3176,
"step": 9400
},
{
"epoch": 6.446648426812586,
"grad_norm": 4.060113906860352,
"learning_rate": 1.9756041951664387e-05,
"loss": 0.4683,
"step": 9425
},
{
"epoch": 6.46374829001368,
"grad_norm": 3.9129064083099365,
"learning_rate": 1.966104271165831e-05,
"loss": 0.3247,
"step": 9450
},
{
"epoch": 6.480848153214774,
"grad_norm": 7.171699523925781,
"learning_rate": 1.9566043471652227e-05,
"loss": 0.2698,
"step": 9475
},
{
"epoch": 6.497948016415869,
"grad_norm": 5.123276233673096,
"learning_rate": 1.9471044231646148e-05,
"loss": 0.3883,
"step": 9500
},
{
"epoch": 6.515047879616963,
"grad_norm": 1.7513083219528198,
"learning_rate": 1.9376044991640068e-05,
"loss": 0.3488,
"step": 9525
},
{
"epoch": 6.532147742818058,
"grad_norm": 3.334690809249878,
"learning_rate": 1.9281045751633988e-05,
"loss": 0.4843,
"step": 9550
},
{
"epoch": 6.549247606019152,
"grad_norm": 2.4918205738067627,
"learning_rate": 1.918604651162791e-05,
"loss": 0.4017,
"step": 9575
},
{
"epoch": 6.566347469220246,
"grad_norm": 3.277615785598755,
"learning_rate": 1.909104727162183e-05,
"loss": 0.3807,
"step": 9600
},
{
"epoch": 6.583447332421341,
"grad_norm": 0.09262839704751968,
"learning_rate": 1.899604803161575e-05,
"loss": 0.3623,
"step": 9625
},
{
"epoch": 6.600547195622435,
"grad_norm": 9.153847694396973,
"learning_rate": 1.890104879160967e-05,
"loss": 0.3903,
"step": 9650
},
{
"epoch": 6.617647058823529,
"grad_norm": 4.24110221862793,
"learning_rate": 1.880604955160359e-05,
"loss": 0.3078,
"step": 9675
},
{
"epoch": 6.634746922024624,
"grad_norm": 5.625593185424805,
"learning_rate": 1.871105031159751e-05,
"loss": 0.3915,
"step": 9700
},
{
"epoch": 6.651846785225718,
"grad_norm": 6.439937591552734,
"learning_rate": 1.8616051071591427e-05,
"loss": 0.4783,
"step": 9725
},
{
"epoch": 6.668946648426813,
"grad_norm": 6.7055559158325195,
"learning_rate": 1.8521051831585347e-05,
"loss": 0.3364,
"step": 9750
},
{
"epoch": 6.686046511627907,
"grad_norm": 2.9428536891937256,
"learning_rate": 1.8426052591579267e-05,
"loss": 0.3409,
"step": 9775
},
{
"epoch": 6.703146374829001,
"grad_norm": 3.669163942337036,
"learning_rate": 1.8331053351573188e-05,
"loss": 0.2637,
"step": 9800
},
{
"epoch": 6.720246238030096,
"grad_norm": 2.797004461288452,
"learning_rate": 1.8236054111567108e-05,
"loss": 0.3721,
"step": 9825
},
{
"epoch": 6.73734610123119,
"grad_norm": 1.3925397396087646,
"learning_rate": 1.8141054871561028e-05,
"loss": 0.3227,
"step": 9850
},
{
"epoch": 6.754445964432285,
"grad_norm": 9.197000503540039,
"learning_rate": 1.804605563155495e-05,
"loss": 0.3789,
"step": 9875
},
{
"epoch": 6.771545827633379,
"grad_norm": 7.823335647583008,
"learning_rate": 1.795105639154887e-05,
"loss": 0.2718,
"step": 9900
},
{
"epoch": 6.788645690834473,
"grad_norm": 1.3368386030197144,
"learning_rate": 1.785605715154279e-05,
"loss": 0.3345,
"step": 9925
},
{
"epoch": 6.805745554035568,
"grad_norm": 4.986422061920166,
"learning_rate": 1.776105791153671e-05,
"loss": 0.3514,
"step": 9950
},
{
"epoch": 6.822845417236662,
"grad_norm": 3.647663116455078,
"learning_rate": 1.7666058671530626e-05,
"loss": 0.3476,
"step": 9975
},
{
"epoch": 6.839945280437757,
"grad_norm": 10.99355697631836,
"learning_rate": 1.7571059431524546e-05,
"loss": 0.4088,
"step": 10000
},
{
"epoch": 6.857045143638851,
"grad_norm": 6.52988862991333,
"learning_rate": 1.7476060191518467e-05,
"loss": 0.3857,
"step": 10025
},
{
"epoch": 6.874145006839945,
"grad_norm": 4.713931560516357,
"learning_rate": 1.738106095151239e-05,
"loss": 0.5173,
"step": 10050
},
{
"epoch": 6.89124487004104,
"grad_norm": 1.7083477973937988,
"learning_rate": 1.728606171150631e-05,
"loss": 0.2722,
"step": 10075
},
{
"epoch": 6.908344733242134,
"grad_norm": 2.890033483505249,
"learning_rate": 1.719106247150023e-05,
"loss": 0.346,
"step": 10100
},
{
"epoch": 6.925444596443229,
"grad_norm": 2.318939208984375,
"learning_rate": 1.709606323149415e-05,
"loss": 0.3563,
"step": 10125
},
{
"epoch": 6.942544459644322,
"grad_norm": 1.522857666015625,
"learning_rate": 1.700106399148807e-05,
"loss": 0.4441,
"step": 10150
},
{
"epoch": 6.959644322845417,
"grad_norm": 4.418458938598633,
"learning_rate": 1.690986472108223e-05,
"loss": 0.253,
"step": 10175
},
{
"epoch": 6.976744186046512,
"grad_norm": 7.665693759918213,
"learning_rate": 1.6814865481076152e-05,
"loss": 0.329,
"step": 10200
},
{
"epoch": 6.993844049247606,
"grad_norm": 2.388054609298706,
"learning_rate": 1.6719866241070072e-05,
"loss": 0.3777,
"step": 10225
},
{
"epoch": 7.0,
"eval_accuracy": 0.9214749061113008,
"eval_f1_macro": 0.833050479179196,
"eval_f1_micro": 0.9214749061113008,
"eval_f1_weighted": 0.9214057890123231,
"eval_loss": 0.23535355925559998,
"eval_precision_macro": 0.8898753259090769,
"eval_precision_micro": 0.9214749061113008,
"eval_precision_weighted": 0.9262844759199463,
"eval_recall_macro": 0.7987766214501736,
"eval_recall_micro": 0.9214749061113008,
"eval_recall_weighted": 0.9214749061113008,
"eval_runtime": 30.0564,
"eval_samples_per_second": 97.45,
"eval_steps_per_second": 6.122,
"step": 10234
}
],
"logging_steps": 25,
"max_steps": 14620,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 500,
"total_flos": 2.242292315610641e+19,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}