RodrigoSalazar-U's picture
Training in progress, step 7500, checkpoint
021f49d verified
raw
history blame
13.8 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 4.140215291195142,
"eval_steps": 500,
"global_step": 7500,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.05520287054926856,
"grad_norm": 1.2856197357177734,
"learning_rate": 1.1037527593818985e-05,
"loss": 0.7609,
"step": 100
},
{
"epoch": 0.11040574109853712,
"grad_norm": 1.0077331066131592,
"learning_rate": 2.207505518763797e-05,
"loss": 0.5552,
"step": 200
},
{
"epoch": 0.1656086116478057,
"grad_norm": 1.2774763107299805,
"learning_rate": 3.311258278145696e-05,
"loss": 0.5229,
"step": 300
},
{
"epoch": 0.22081148219707425,
"grad_norm": 1.3611174821853638,
"learning_rate": 4.415011037527594e-05,
"loss": 0.5191,
"step": 400
},
{
"epoch": 0.2760143527463428,
"grad_norm": 1.112617015838623,
"learning_rate": 5.518763796909493e-05,
"loss": 0.5068,
"step": 500
},
{
"epoch": 0.3312172232956114,
"grad_norm": 1.1730616092681885,
"learning_rate": 6.622516556291392e-05,
"loss": 0.4895,
"step": 600
},
{
"epoch": 0.3864200938448799,
"grad_norm": 1.1149307489395142,
"learning_rate": 7.726269315673289e-05,
"loss": 0.5044,
"step": 700
},
{
"epoch": 0.4416229643941485,
"grad_norm": 1.1326206922531128,
"learning_rate": 8.830022075055188e-05,
"loss": 0.5081,
"step": 800
},
{
"epoch": 0.49682583494341703,
"grad_norm": 1.1592661142349243,
"learning_rate": 9.933774834437086e-05,
"loss": 0.5024,
"step": 900
},
{
"epoch": 0.5520287054926856,
"grad_norm": 1.4300315380096436,
"learning_rate": 9.996717238759354e-05,
"loss": 0.5078,
"step": 1000
},
{
"epoch": 0.6072315760419542,
"grad_norm": 1.3841311931610107,
"learning_rate": 9.986022415440564e-05,
"loss": 0.5091,
"step": 1100
},
{
"epoch": 0.6624344465912227,
"grad_norm": 1.3825188875198364,
"learning_rate": 9.967918047007774e-05,
"loss": 0.4915,
"step": 1200
},
{
"epoch": 0.7176373171404913,
"grad_norm": 1.248660683631897,
"learning_rate": 9.942431037699172e-05,
"loss": 0.5049,
"step": 1300
},
{
"epoch": 0.7728401876897598,
"grad_norm": 1.2805066108703613,
"learning_rate": 9.909599262824883e-05,
"loss": 0.4787,
"step": 1400
},
{
"epoch": 0.8280430582390285,
"grad_norm": 1.5455862283706665,
"learning_rate": 9.869471512481871e-05,
"loss": 0.4844,
"step": 1500
},
{
"epoch": 0.883245928788297,
"grad_norm": 1.647222876548767,
"learning_rate": 9.822107419048758e-05,
"loss": 0.4732,
"step": 1600
},
{
"epoch": 0.9384487993375655,
"grad_norm": 1.128386378288269,
"learning_rate": 9.76757736856833e-05,
"loss": 0.4515,
"step": 1700
},
{
"epoch": 0.9936516698868341,
"grad_norm": 1.1730600595474243,
"learning_rate": 9.705962396149427e-05,
"loss": 0.4496,
"step": 1800
},
{
"epoch": 1.0488545404361027,
"grad_norm": 1.2323981523513794,
"learning_rate": 9.637354065543631e-05,
"loss": 0.3201,
"step": 1900
},
{
"epoch": 1.1040574109853711,
"grad_norm": 1.3514504432678223,
"learning_rate": 9.561854333075736e-05,
"loss": 0.3029,
"step": 2000
},
{
"epoch": 1.1592602815346398,
"grad_norm": 1.305432915687561,
"learning_rate": 9.479575396130191e-05,
"loss": 0.2974,
"step": 2100
},
{
"epoch": 1.2144631520839084,
"grad_norm": 1.2165225744247437,
"learning_rate": 9.390639526418681e-05,
"loss": 0.2999,
"step": 2200
},
{
"epoch": 1.2696660226331768,
"grad_norm": 1.5560479164123535,
"learning_rate": 9.295178888276614e-05,
"loss": 0.2957,
"step": 2300
},
{
"epoch": 1.3248688931824455,
"grad_norm": 1.236303448677063,
"learning_rate": 9.193335342258558e-05,
"loss": 0.3,
"step": 2400
},
{
"epoch": 1.380071763731714,
"grad_norm": 1.3867113590240479,
"learning_rate": 9.08526023432446e-05,
"loss": 0.3113,
"step": 2500
},
{
"epoch": 1.4352746342809826,
"grad_norm": 1.3542137145996094,
"learning_rate": 8.971114170929969e-05,
"loss": 0.2965,
"step": 2600
},
{
"epoch": 1.4904775048302512,
"grad_norm": 1.0694419145584106,
"learning_rate": 8.851066780355073e-05,
"loss": 0.2907,
"step": 2700
},
{
"epoch": 1.5456803753795199,
"grad_norm": 1.2589095830917358,
"learning_rate": 8.72529646062573e-05,
"loss": 0.2881,
"step": 2800
},
{
"epoch": 1.6008832459287883,
"grad_norm": 0.9907445311546326,
"learning_rate": 8.593990114403092e-05,
"loss": 0.2833,
"step": 2900
},
{
"epoch": 1.6560861164780567,
"grad_norm": 1.1165964603424072,
"learning_rate": 8.45734287123433e-05,
"loss": 0.279,
"step": 3000
},
{
"epoch": 1.7112889870273253,
"grad_norm": 1.4846047163009644,
"learning_rate": 8.315557797577755e-05,
"loss": 0.2788,
"step": 3100
},
{
"epoch": 1.766491857576594,
"grad_norm": 1.2695362567901611,
"learning_rate": 8.168845595033201e-05,
"loss": 0.2764,
"step": 3200
},
{
"epoch": 1.8216947281258626,
"grad_norm": 1.3499970436096191,
"learning_rate": 8.017424287226106e-05,
"loss": 0.2576,
"step": 3300
},
{
"epoch": 1.876897598675131,
"grad_norm": 1.3490862846374512,
"learning_rate": 7.861518895810596e-05,
"loss": 0.2529,
"step": 3400
},
{
"epoch": 1.9321004692243997,
"grad_norm": 1.3637194633483887,
"learning_rate": 7.701361106073043e-05,
"loss": 0.247,
"step": 3500
},
{
"epoch": 1.9873033397736681,
"grad_norm": 1.1327725648880005,
"learning_rate": 7.537188922633076e-05,
"loss": 0.2376,
"step": 3600
},
{
"epoch": 2.0425062103229368,
"grad_norm": 0.8765416741371155,
"learning_rate": 7.369246315753623e-05,
"loss": 0.1724,
"step": 3700
},
{
"epoch": 2.0977090808722054,
"grad_norm": 0.8936216831207275,
"learning_rate": 7.19778285878565e-05,
"loss": 0.1515,
"step": 3800
},
{
"epoch": 2.152911951421474,
"grad_norm": 0.7982431650161743,
"learning_rate": 7.023053357286366e-05,
"loss": 0.1542,
"step": 3900
},
{
"epoch": 2.2081148219707423,
"grad_norm": 0.7692477107048035,
"learning_rate": 6.845317470362014e-05,
"loss": 0.154,
"step": 4000
},
{
"epoch": 2.263317692520011,
"grad_norm": 0.7042354345321655,
"learning_rate": 6.664839324798001e-05,
"loss": 0.155,
"step": 4100
},
{
"epoch": 2.3185205630692796,
"grad_norm": 0.8404821157455444,
"learning_rate": 6.481887122549755e-05,
"loss": 0.1576,
"step": 4200
},
{
"epoch": 2.373723433618548,
"grad_norm": 0.8974793553352356,
"learning_rate": 6.296732742177644e-05,
"loss": 0.1553,
"step": 4300
},
{
"epoch": 2.428926304167817,
"grad_norm": 0.8145149350166321,
"learning_rate": 6.109651334818204e-05,
"loss": 0.1542,
"step": 4400
},
{
"epoch": 2.4841291747170855,
"grad_norm": 0.9313832521438599,
"learning_rate": 5.9209209152921384e-05,
"loss": 0.1541,
"step": 4500
},
{
"epoch": 2.5393320452663537,
"grad_norm": 0.9550230503082275,
"learning_rate": 5.730821948956665e-05,
"loss": 0.1556,
"step": 4600
},
{
"epoch": 2.5945349158156223,
"grad_norm": 0.6794816255569458,
"learning_rate": 5.539636934916247e-05,
"loss": 0.1529,
"step": 4700
},
{
"epoch": 2.649737786364891,
"grad_norm": 0.8268643021583557,
"learning_rate": 5.347649986211022e-05,
"loss": 0.1538,
"step": 4800
},
{
"epoch": 2.7049406569141596,
"grad_norm": 0.8783162236213684,
"learning_rate": 5.155146407606835e-05,
"loss": 0.1527,
"step": 4900
},
{
"epoch": 2.760143527463428,
"grad_norm": 0.788532555103302,
"learning_rate": 4.962412271614282e-05,
"loss": 0.146,
"step": 5000
},
{
"epoch": 2.8153463980126965,
"grad_norm": 0.671291172504425,
"learning_rate": 4.7697339933668414e-05,
"loss": 0.1483,
"step": 5100
},
{
"epoch": 2.870549268561965,
"grad_norm": 0.6203655004501343,
"learning_rate": 4.5773979049898455e-05,
"loss": 0.1451,
"step": 5200
},
{
"epoch": 2.9257521391112338,
"grad_norm": 0.8333371877670288,
"learning_rate": 4.385689830092801e-05,
"loss": 0.1437,
"step": 5300
},
{
"epoch": 2.9809550096605024,
"grad_norm": 0.725379467010498,
"learning_rate": 4.194894659017415e-05,
"loss": 0.1412,
"step": 5400
},
{
"epoch": 3.036157880209771,
"grad_norm": 0.4681724011898041,
"learning_rate": 4.005295925472484e-05,
"loss": 0.1206,
"step": 5500
},
{
"epoch": 3.0913607507590393,
"grad_norm": 0.6104450821876526,
"learning_rate": 3.817175385184861e-05,
"loss": 0.1074,
"step": 5600
},
{
"epoch": 3.146563621308308,
"grad_norm": 0.5122411847114563,
"learning_rate": 3.630812597192591e-05,
"loss": 0.1088,
"step": 5700
},
{
"epoch": 3.2017664918575766,
"grad_norm": 0.48834362626075745,
"learning_rate": 3.4464845084024766e-05,
"loss": 0.1076,
"step": 5800
},
{
"epoch": 3.256969362406845,
"grad_norm": 0.4744304120540619,
"learning_rate": 3.2644650420294285e-05,
"loss": 0.1087,
"step": 5900
},
{
"epoch": 3.312172232956114,
"grad_norm": 0.5677183270454407,
"learning_rate": 3.085024690529227e-05,
"loss": 0.1086,
"step": 6000
},
{
"epoch": 3.367375103505382,
"grad_norm": 0.5468078851699829,
"learning_rate": 2.908430113629592e-05,
"loss": 0.1083,
"step": 6100
},
{
"epoch": 3.4225779740546507,
"grad_norm": 0.49241673946380615,
"learning_rate": 2.734943742056943e-05,
"loss": 0.107,
"step": 6200
},
{
"epoch": 3.4777808446039193,
"grad_norm": 0.4840191900730133,
"learning_rate": 2.5648233875477157e-05,
"loss": 0.1081,
"step": 6300
},
{
"epoch": 3.532983715153188,
"grad_norm": 0.43311458826065063,
"learning_rate": 2.398321859723792e-05,
"loss": 0.1089,
"step": 6400
},
{
"epoch": 3.5881865857024566,
"grad_norm": 0.506381094455719,
"learning_rate": 2.2356865904013918e-05,
"loss": 0.1097,
"step": 6500
},
{
"epoch": 3.6433894562517253,
"grad_norm": 0.4998696446418762,
"learning_rate": 2.0771592658917193e-05,
"loss": 0.1074,
"step": 6600
},
{
"epoch": 3.698592326800994,
"grad_norm": 0.40012943744659424,
"learning_rate": 1.922975467839799e-05,
"loss": 0.1078,
"step": 6700
},
{
"epoch": 3.753795197350262,
"grad_norm": 0.3930221498012543,
"learning_rate": 1.773364323135227e-05,
"loss": 0.1054,
"step": 6800
},
{
"epoch": 3.8089980678995308,
"grad_norm": 0.6007972359657288,
"learning_rate": 1.6285481634151057e-05,
"loss": 0.1068,
"step": 6900
},
{
"epoch": 3.8642009384487994,
"grad_norm": 0.4134114384651184,
"learning_rate": 1.4887421946651437e-05,
"loss": 0.1058,
"step": 7000
},
{
"epoch": 3.919403808998068,
"grad_norm": 0.451486200094223,
"learning_rate": 1.354154177409932e-05,
"loss": 0.1057,
"step": 7100
},
{
"epoch": 3.9746066795473363,
"grad_norm": 0.4829927384853363,
"learning_rate": 1.224984117967648e-05,
"loss": 0.1034,
"step": 7200
},
{
"epoch": 4.029809550096605,
"grad_norm": 0.29708752036094666,
"learning_rate": 1.1014239712279945e-05,
"loss": 0.0977,
"step": 7300
},
{
"epoch": 4.0850124206458736,
"grad_norm": 0.4131404459476471,
"learning_rate": 9.83657355395079e-06,
"loss": 0.09,
"step": 7400
},
{
"epoch": 4.140215291195142,
"grad_norm": 0.3377688229084015,
"learning_rate": 8.718592791191299e-06,
"loss": 0.0907,
"step": 7500
}
],
"logging_steps": 100,
"max_steps": 9055,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 3.9786029952435487e+18,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}