3m-classical-maestro / trainer_state.json
hidude562's picture
Upload 12 files
5ea9c8f verified
raw
history blame contribute delete
No virus
14.9 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9977335435553804,
"eval_steps": 500,
"global_step": 40500,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.012317698068584943,
"grad_norm": 2.049720048904419,
"learning_rate": 4.938534686637761e-05,
"loss": 1.4868,
"step": 500
},
{
"epoch": 0.024635396137169887,
"grad_norm": 1.9058018922805786,
"learning_rate": 4.8769461962948364e-05,
"loss": 1.1702,
"step": 1000
},
{
"epoch": 0.03695309420575483,
"grad_norm": 1.4330534934997559,
"learning_rate": 4.815357705951912e-05,
"loss": 1.0684,
"step": 1500
},
{
"epoch": 0.049270792274339774,
"grad_norm": 2.9333412647247314,
"learning_rate": 4.753769215608987e-05,
"loss": 1.0331,
"step": 2000
},
{
"epoch": 0.06158849034292471,
"grad_norm": 1.674055576324463,
"learning_rate": 4.6921807252660625e-05,
"loss": 1.0139,
"step": 2500
},
{
"epoch": 0.07390618841150966,
"grad_norm": 1.7064428329467773,
"learning_rate": 4.630592234923138e-05,
"loss": 0.987,
"step": 3000
},
{
"epoch": 0.0862238864800946,
"grad_norm": 1.4718812704086304,
"learning_rate": 4.569003744580213e-05,
"loss": 0.9881,
"step": 3500
},
{
"epoch": 0.09854158454867955,
"grad_norm": 1.326068639755249,
"learning_rate": 4.507415254237288e-05,
"loss": 0.9511,
"step": 4000
},
{
"epoch": 0.11085928261726448,
"grad_norm": 1.724241852760315,
"learning_rate": 4.4458267638943634e-05,
"loss": 0.952,
"step": 4500
},
{
"epoch": 0.12317698068584942,
"grad_norm": 0.20040610432624817,
"learning_rate": 4.384361450532125e-05,
"loss": 0.9575,
"step": 5000
},
{
"epoch": 0.13549467875443438,
"grad_norm": 1.6827597618103027,
"learning_rate": 4.3227729601892e-05,
"loss": 0.9589,
"step": 5500
},
{
"epoch": 0.14781237682301931,
"grad_norm": 0.9528223276138306,
"learning_rate": 4.2611844698462756e-05,
"loss": 0.9464,
"step": 6000
},
{
"epoch": 0.16013007489160425,
"grad_norm": 1.1192400455474854,
"learning_rate": 4.199595979503351e-05,
"loss": 0.9395,
"step": 6500
},
{
"epoch": 0.1724477729601892,
"grad_norm": 1.5803415775299072,
"learning_rate": 4.138007489160426e-05,
"loss": 0.9376,
"step": 7000
},
{
"epoch": 0.18476547102877414,
"grad_norm": 1.0329970121383667,
"learning_rate": 4.076542175798187e-05,
"loss": 0.9247,
"step": 7500
},
{
"epoch": 0.1970831690973591,
"grad_norm": 1.0181173086166382,
"learning_rate": 4.014953685455262e-05,
"loss": 0.9469,
"step": 8000
},
{
"epoch": 0.20940086716594403,
"grad_norm": 0.837753176689148,
"learning_rate": 3.953365195112337e-05,
"loss": 0.9017,
"step": 8500
},
{
"epoch": 0.22171856523452896,
"grad_norm": 1.228780746459961,
"learning_rate": 3.891776704769413e-05,
"loss": 0.9006,
"step": 9000
},
{
"epoch": 0.23403626330311392,
"grad_norm": 1.6748690605163574,
"learning_rate": 3.830188214426488e-05,
"loss": 0.9312,
"step": 9500
},
{
"epoch": 0.24635396137169885,
"grad_norm": 2.5065925121307373,
"learning_rate": 3.7687229010642495e-05,
"loss": 0.9102,
"step": 10000
},
{
"epoch": 0.2586716594402838,
"grad_norm": 0.4429062008857727,
"learning_rate": 3.70725758770201e-05,
"loss": 0.9185,
"step": 10500
},
{
"epoch": 0.27098935750886877,
"grad_norm": 0.8518178462982178,
"learning_rate": 3.6456690973590856e-05,
"loss": 0.9158,
"step": 11000
},
{
"epoch": 0.28330705557745367,
"grad_norm": 1.952348232269287,
"learning_rate": 3.584080607016161e-05,
"loss": 0.8735,
"step": 11500
},
{
"epoch": 0.29562475364603863,
"grad_norm": 0.9509502053260803,
"learning_rate": 3.5224921166732364e-05,
"loss": 0.8721,
"step": 12000
},
{
"epoch": 0.3079424517146236,
"grad_norm": 0.9900819659233093,
"learning_rate": 3.460903626330311e-05,
"loss": 0.9015,
"step": 12500
},
{
"epoch": 0.3202601497832085,
"grad_norm": 1.6706575155258179,
"learning_rate": 3.399315135987387e-05,
"loss": 0.9053,
"step": 13000
},
{
"epoch": 0.33257784785179345,
"grad_norm": 2.245882272720337,
"learning_rate": 3.3377266456444625e-05,
"loss": 0.9008,
"step": 13500
},
{
"epoch": 0.3448955459203784,
"grad_norm": 1.2105801105499268,
"learning_rate": 3.276138155301537e-05,
"loss": 0.8859,
"step": 14000
},
{
"epoch": 0.35721324398896337,
"grad_norm": 1.136516809463501,
"learning_rate": 3.214672841939299e-05,
"loss": 0.8922,
"step": 14500
},
{
"epoch": 0.3695309420575483,
"grad_norm": 1.4451056718826294,
"learning_rate": 3.1530843515963734e-05,
"loss": 0.8907,
"step": 15000
},
{
"epoch": 0.38184864012613323,
"grad_norm": 0.8413623571395874,
"learning_rate": 3.091495861253449e-05,
"loss": 0.8929,
"step": 15500
},
{
"epoch": 0.3941663381947182,
"grad_norm": 1.7711334228515625,
"learning_rate": 3.029907370910524e-05,
"loss": 0.8792,
"step": 16000
},
{
"epoch": 0.4064840362633031,
"grad_norm": 1.3520532846450806,
"learning_rate": 2.9683188805676e-05,
"loss": 0.8809,
"step": 16500
},
{
"epoch": 0.41880173433188805,
"grad_norm": 1.8900405168533325,
"learning_rate": 2.906853567205361e-05,
"loss": 0.8834,
"step": 17000
},
{
"epoch": 0.431119432400473,
"grad_norm": 1.377387523651123,
"learning_rate": 2.845265076862436e-05,
"loss": 0.8902,
"step": 17500
},
{
"epoch": 0.4434371304690579,
"grad_norm": 1.2132008075714111,
"learning_rate": 2.7836765865195114e-05,
"loss": 0.8607,
"step": 18000
},
{
"epoch": 0.4557548285376429,
"grad_norm": 0.94814532995224,
"learning_rate": 2.7220880961765865e-05,
"loss": 0.88,
"step": 18500
},
{
"epoch": 0.46807252660622783,
"grad_norm": 0.8629463315010071,
"learning_rate": 2.6606227828143476e-05,
"loss": 0.8609,
"step": 19000
},
{
"epoch": 0.4803902246748128,
"grad_norm": 1.1995203495025635,
"learning_rate": 2.599034292471423e-05,
"loss": 0.8778,
"step": 19500
},
{
"epoch": 0.4927079227433977,
"grad_norm": 1.2757550477981567,
"learning_rate": 2.5374458021284987e-05,
"loss": 0.853,
"step": 20000
},
{
"epoch": 0.5050256208119827,
"grad_norm": 1.4950288534164429,
"learning_rate": 2.4758573117855734e-05,
"loss": 0.8688,
"step": 20500
},
{
"epoch": 0.5173433188805676,
"grad_norm": 0.997870147228241,
"learning_rate": 2.4143919984233348e-05,
"loss": 0.8738,
"step": 21000
},
{
"epoch": 0.5296610169491526,
"grad_norm": 1.2419720888137817,
"learning_rate": 2.3528035080804102e-05,
"loss": 0.8929,
"step": 21500
},
{
"epoch": 0.5419787150177375,
"grad_norm": 1.197637915611267,
"learning_rate": 2.2912150177374853e-05,
"loss": 0.8848,
"step": 22000
},
{
"epoch": 0.5542964130863224,
"grad_norm": 1.501338005065918,
"learning_rate": 2.2296265273945606e-05,
"loss": 0.8543,
"step": 22500
},
{
"epoch": 0.5666141111549073,
"grad_norm": 1.2377902269363403,
"learning_rate": 2.1681612140323217e-05,
"loss": 0.8402,
"step": 23000
},
{
"epoch": 0.5789318092234923,
"grad_norm": 0.6151154637336731,
"learning_rate": 2.106572723689397e-05,
"loss": 0.872,
"step": 23500
},
{
"epoch": 0.5912495072920773,
"grad_norm": 0.8272470235824585,
"learning_rate": 2.044984233346472e-05,
"loss": 0.8524,
"step": 24000
},
{
"epoch": 0.6035672053606622,
"grad_norm": 1.718947410583496,
"learning_rate": 1.9833957430035476e-05,
"loss": 0.8404,
"step": 24500
},
{
"epoch": 0.6158849034292472,
"grad_norm": 1.5924681425094604,
"learning_rate": 1.921807252660623e-05,
"loss": 0.868,
"step": 25000
},
{
"epoch": 0.6282026014978321,
"grad_norm": 1.3491289615631104,
"learning_rate": 1.860341939298384e-05,
"loss": 0.8521,
"step": 25500
},
{
"epoch": 0.640520299566417,
"grad_norm": 1.2171722650527954,
"learning_rate": 1.798753448955459e-05,
"loss": 0.8312,
"step": 26000
},
{
"epoch": 0.6528379976350019,
"grad_norm": 1.0879688262939453,
"learning_rate": 1.7371649586125348e-05,
"loss": 0.8471,
"step": 26500
},
{
"epoch": 0.6651556957035869,
"grad_norm": 0.9785297513008118,
"learning_rate": 1.67557646826961e-05,
"loss": 0.8571,
"step": 27000
},
{
"epoch": 0.6774733937721719,
"grad_norm": 2.27364444732666,
"learning_rate": 1.614111154907371e-05,
"loss": 0.8548,
"step": 27500
},
{
"epoch": 0.6897910918407568,
"grad_norm": 1.1297192573547363,
"learning_rate": 1.552645841545132e-05,
"loss": 0.8662,
"step": 28000
},
{
"epoch": 0.7021087899093418,
"grad_norm": 1.4344604015350342,
"learning_rate": 1.4910573512022074e-05,
"loss": 0.8609,
"step": 28500
},
{
"epoch": 0.7144264879779267,
"grad_norm": 0.8162903189659119,
"learning_rate": 1.4294688608592826e-05,
"loss": 0.8519,
"step": 29000
},
{
"epoch": 0.7267441860465116,
"grad_norm": 1.4598755836486816,
"learning_rate": 1.3678803705163579e-05,
"loss": 0.8376,
"step": 29500
},
{
"epoch": 0.7390618841150965,
"grad_norm": 0.9850552082061768,
"learning_rate": 1.3062918801734334e-05,
"loss": 0.842,
"step": 30000
},
{
"epoch": 0.7513795821836815,
"grad_norm": 1.2781943082809448,
"learning_rate": 1.2447033898305085e-05,
"loss": 0.8539,
"step": 30500
},
{
"epoch": 0.7636972802522665,
"grad_norm": 0.9877688884735107,
"learning_rate": 1.1831148994875837e-05,
"loss": 0.8439,
"step": 31000
},
{
"epoch": 0.7760149783208514,
"grad_norm": 1.0702171325683594,
"learning_rate": 1.121526409144659e-05,
"loss": 0.8492,
"step": 31500
},
{
"epoch": 0.7883326763894364,
"grad_norm": 1.0662543773651123,
"learning_rate": 1.0600610957824203e-05,
"loss": 0.8572,
"step": 32000
},
{
"epoch": 0.8006503744580212,
"grad_norm": 1.029024600982666,
"learning_rate": 9.985957824201814e-06,
"loss": 0.8347,
"step": 32500
},
{
"epoch": 0.8129680725266062,
"grad_norm": 0.965148389339447,
"learning_rate": 9.370072920772566e-06,
"loss": 0.8408,
"step": 33000
},
{
"epoch": 0.8252857705951911,
"grad_norm": 1.441133975982666,
"learning_rate": 8.754188017343319e-06,
"loss": 0.8385,
"step": 33500
},
{
"epoch": 0.8376034686637761,
"grad_norm": 1.6267979145050049,
"learning_rate": 8.138303113914072e-06,
"loss": 0.8426,
"step": 34000
},
{
"epoch": 0.8499211667323611,
"grad_norm": 1.0745171308517456,
"learning_rate": 7.522418210484825e-06,
"loss": 0.8509,
"step": 34500
},
{
"epoch": 0.862238864800946,
"grad_norm": 1.2497833967208862,
"learning_rate": 6.906533307055578e-06,
"loss": 0.868,
"step": 35000
},
{
"epoch": 0.874556562869531,
"grad_norm": 1.1674834489822388,
"learning_rate": 6.2918801734331886e-06,
"loss": 0.8548,
"step": 35500
},
{
"epoch": 0.8868742609381158,
"grad_norm": 1.369992733001709,
"learning_rate": 5.6759952700039425e-06,
"loss": 0.8662,
"step": 36000
},
{
"epoch": 0.8991919590067008,
"grad_norm": 2.269826889038086,
"learning_rate": 5.060110366574695e-06,
"loss": 0.8518,
"step": 36500
},
{
"epoch": 0.9115096570752858,
"grad_norm": 1.3908036947250366,
"learning_rate": 4.444225463145448e-06,
"loss": 0.8583,
"step": 37000
},
{
"epoch": 0.9238273551438707,
"grad_norm": 0.907351553440094,
"learning_rate": 3.8283405597162e-06,
"loss": 0.8568,
"step": 37500
},
{
"epoch": 0.9361450532124557,
"grad_norm": 1.2205787897109985,
"learning_rate": 3.2124556562869534e-06,
"loss": 0.86,
"step": 38000
},
{
"epoch": 0.9484627512810406,
"grad_norm": 0.8388875722885132,
"learning_rate": 2.596570752857706e-06,
"loss": 0.8228,
"step": 38500
},
{
"epoch": 0.9607804493496256,
"grad_norm": 1.2039369344711304,
"learning_rate": 1.980685849428459e-06,
"loss": 0.8462,
"step": 39000
},
{
"epoch": 0.9730981474182104,
"grad_norm": 1.1439008712768555,
"learning_rate": 1.3660327158060701e-06,
"loss": 0.8669,
"step": 39500
},
{
"epoch": 0.9854158454867954,
"grad_norm": 1.5806646347045898,
"learning_rate": 7.50147812376823e-07,
"loss": 0.8306,
"step": 40000
},
{
"epoch": 0.9977335435553804,
"grad_norm": 0.8370537757873535,
"learning_rate": 1.342629089475759e-07,
"loss": 0.8485,
"step": 40500
}
],
"logging_steps": 500,
"max_steps": 40592,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 74657562624000.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}