Mistral-DNA-v1-138M-hg38 / trainer_state.json
RaphaelMourad's picture
Upload 9 files
9ed7b83 verified
{
"best_metric": 6.946237564086914,
"best_model_checkpoint": "./results/models/checkpoint-45320",
"epoch": 5.0,
"eval_steps": 500,
"global_step": 45320,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.05516328331862312,
"grad_norm": 0.3359375,
"learning_rate": 0.004994483671668138,
"loss": 7.1096,
"step": 500
},
{
"epoch": 0.11032656663724624,
"grad_norm": 0.2890625,
"learning_rate": 0.004988967343336276,
"loss": 6.9824,
"step": 1000
},
{
"epoch": 0.16548984995586938,
"grad_norm": 0.369140625,
"learning_rate": 0.004983451015004413,
"loss": 6.971,
"step": 1500
},
{
"epoch": 0.22065313327449249,
"grad_norm": 0.8046875,
"learning_rate": 0.004977934686672551,
"loss": 7.0129,
"step": 2000
},
{
"epoch": 0.2758164165931156,
"grad_norm": 4.59375,
"learning_rate": 0.004972418358340689,
"loss": 7.0571,
"step": 2500
},
{
"epoch": 0.33097969991173876,
"grad_norm": 4.03125,
"learning_rate": 0.004966902030008827,
"loss": 7.0532,
"step": 3000
},
{
"epoch": 0.3861429832303619,
"grad_norm": 3.75,
"learning_rate": 0.0049613857016769635,
"loss": 7.0486,
"step": 3500
},
{
"epoch": 0.44130626654898497,
"grad_norm": 9.125,
"learning_rate": 0.004955869373345101,
"loss": 7.0385,
"step": 4000
},
{
"epoch": 0.4964695498676081,
"grad_norm": 9.1875,
"learning_rate": 0.00495035304501324,
"loss": 7.0304,
"step": 4500
},
{
"epoch": 0.5516328331862312,
"grad_norm": 12.625,
"learning_rate": 0.004944836716681377,
"loss": 7.023,
"step": 5000
},
{
"epoch": 0.6067961165048543,
"grad_norm": 8.6875,
"learning_rate": 0.0049393203883495145,
"loss": 7.0205,
"step": 5500
},
{
"epoch": 0.6619593998234775,
"grad_norm": 7.6875,
"learning_rate": 0.004933804060017652,
"loss": 7.025,
"step": 6000
},
{
"epoch": 0.7171226831421006,
"grad_norm": 4.96875,
"learning_rate": 0.00492828773168579,
"loss": 7.0285,
"step": 6500
},
{
"epoch": 0.7722859664607238,
"grad_norm": 5.46875,
"learning_rate": 0.004922771403353928,
"loss": 7.025,
"step": 7000
},
{
"epoch": 0.8274492497793469,
"grad_norm": 20.125,
"learning_rate": 0.0049172550750220655,
"loss": 7.0162,
"step": 7500
},
{
"epoch": 0.8826125330979699,
"grad_norm": 7.1875,
"learning_rate": 0.004911738746690203,
"loss": 7.0154,
"step": 8000
},
{
"epoch": 0.9377758164165931,
"grad_norm": 8.1875,
"learning_rate": 0.004906222418358341,
"loss": 7.0104,
"step": 8500
},
{
"epoch": 0.9929390997352162,
"grad_norm": 6.71875,
"learning_rate": 0.004900706090026478,
"loss": 7.0374,
"step": 9000
},
{
"epoch": 1.0,
"eval_loss": 7.0309014320373535,
"eval_runtime": 9.4135,
"eval_samples_per_second": 53.115,
"eval_steps_per_second": 1.7,
"step": 9064
},
{
"epoch": 1.0481023830538394,
"grad_norm": 3.5,
"learning_rate": 0.0048951897616946165,
"loss": 7.02,
"step": 9500
},
{
"epoch": 1.1032656663724625,
"grad_norm": 10.5,
"learning_rate": 0.004889673433362754,
"loss": 7.0297,
"step": 10000
},
{
"epoch": 1.1584289496910856,
"grad_norm": 8.875,
"learning_rate": 0.004884157105030891,
"loss": 7.0361,
"step": 10500
},
{
"epoch": 1.2135922330097086,
"grad_norm": 3.5625,
"learning_rate": 0.004878640776699029,
"loss": 7.015,
"step": 11000
},
{
"epoch": 1.268755516328332,
"grad_norm": 9.1875,
"learning_rate": 0.0048731244483671676,
"loss": 6.9996,
"step": 11500
},
{
"epoch": 1.323918799646955,
"grad_norm": 13.625,
"learning_rate": 0.0048676081200353044,
"loss": 6.9975,
"step": 12000
},
{
"epoch": 1.379082082965578,
"grad_norm": 6.34375,
"learning_rate": 0.004862091791703442,
"loss": 7.0081,
"step": 12500
},
{
"epoch": 1.4342453662842012,
"grad_norm": 14.375,
"learning_rate": 0.00485657546337158,
"loss": 7.0008,
"step": 13000
},
{
"epoch": 1.4894086496028245,
"grad_norm": 8.8125,
"learning_rate": 0.004851059135039718,
"loss": 6.9962,
"step": 13500
},
{
"epoch": 1.5445719329214476,
"grad_norm": 22.625,
"learning_rate": 0.0048455428067078555,
"loss": 6.9928,
"step": 14000
},
{
"epoch": 1.5997352162400706,
"grad_norm": 12.6875,
"learning_rate": 0.004840026478375993,
"loss": 6.9919,
"step": 14500
},
{
"epoch": 1.6548984995586937,
"grad_norm": 53.75,
"learning_rate": 0.004834510150044131,
"loss": 6.9923,
"step": 15000
},
{
"epoch": 1.7100617828773168,
"grad_norm": 9.9375,
"learning_rate": 0.004828993821712269,
"loss": 6.9966,
"step": 15500
},
{
"epoch": 1.7652250661959399,
"grad_norm": 13.0625,
"learning_rate": 0.004823477493380406,
"loss": 6.9908,
"step": 16000
},
{
"epoch": 1.820388349514563,
"grad_norm": 8.9375,
"learning_rate": 0.004817961165048544,
"loss": 6.9937,
"step": 16500
},
{
"epoch": 1.8755516328331863,
"grad_norm": 42.0,
"learning_rate": 0.004812444836716681,
"loss": 6.9894,
"step": 17000
},
{
"epoch": 1.9307149161518093,
"grad_norm": 17.5,
"learning_rate": 0.004806928508384819,
"loss": 6.9926,
"step": 17500
},
{
"epoch": 1.9858781994704324,
"grad_norm": 30.75,
"learning_rate": 0.004801412180052957,
"loss": 6.9989,
"step": 18000
},
{
"epoch": 2.0,
"eval_loss": 6.996521472930908,
"eval_runtime": 8.6549,
"eval_samples_per_second": 57.771,
"eval_steps_per_second": 1.849,
"step": 18128
},
{
"epoch": 2.0410414827890557,
"grad_norm": 12.125,
"learning_rate": 0.004795895851721094,
"loss": 6.9915,
"step": 18500
},
{
"epoch": 2.096204766107679,
"grad_norm": 18.625,
"learning_rate": 0.004790379523389232,
"loss": 6.9839,
"step": 19000
},
{
"epoch": 2.151368049426302,
"grad_norm": 21.25,
"learning_rate": 0.00478486319505737,
"loss": 6.9821,
"step": 19500
},
{
"epoch": 2.206531332744925,
"grad_norm": 21.375,
"learning_rate": 0.004779346866725508,
"loss": 6.9842,
"step": 20000
},
{
"epoch": 2.261694616063548,
"grad_norm": 16.625,
"learning_rate": 0.004773830538393645,
"loss": 6.9836,
"step": 20500
},
{
"epoch": 2.316857899382171,
"grad_norm": 17.125,
"learning_rate": 0.004768314210061783,
"loss": 6.9837,
"step": 21000
},
{
"epoch": 2.372021182700794,
"grad_norm": 8.9375,
"learning_rate": 0.004762797881729921,
"loss": 6.9822,
"step": 21500
},
{
"epoch": 2.4271844660194173,
"grad_norm": 8.25,
"learning_rate": 0.004757281553398059,
"loss": 6.9791,
"step": 22000
},
{
"epoch": 2.4823477493380404,
"grad_norm": 8.375,
"learning_rate": 0.0047517652250661955,
"loss": 6.986,
"step": 22500
},
{
"epoch": 2.537511032656664,
"grad_norm": 10.6875,
"learning_rate": 0.004746248896734333,
"loss": 6.9807,
"step": 23000
},
{
"epoch": 2.592674315975287,
"grad_norm": 14.375,
"learning_rate": 0.004740732568402472,
"loss": 6.9822,
"step": 23500
},
{
"epoch": 2.64783759929391,
"grad_norm": 10.625,
"learning_rate": 0.004735216240070609,
"loss": 6.9701,
"step": 24000
},
{
"epoch": 2.703000882612533,
"grad_norm": 11.75,
"learning_rate": 0.0047296999117387465,
"loss": 6.9707,
"step": 24500
},
{
"epoch": 2.758164165931156,
"grad_norm": 30.375,
"learning_rate": 0.004724183583406884,
"loss": 6.9792,
"step": 25000
},
{
"epoch": 2.8133274492497793,
"grad_norm": 17.25,
"learning_rate": 0.004718667255075022,
"loss": 6.9803,
"step": 25500
},
{
"epoch": 2.8684907325684024,
"grad_norm": 9.0,
"learning_rate": 0.00471315092674316,
"loss": 6.9752,
"step": 26000
},
{
"epoch": 2.9236540158870254,
"grad_norm": 6.0,
"learning_rate": 0.0047076345984112975,
"loss": 6.9687,
"step": 26500
},
{
"epoch": 2.978817299205649,
"grad_norm": 12.0625,
"learning_rate": 0.004702118270079435,
"loss": 6.9635,
"step": 27000
},
{
"epoch": 3.0,
"eval_loss": 6.960501194000244,
"eval_runtime": 8.9893,
"eval_samples_per_second": 55.622,
"eval_steps_per_second": 1.78,
"step": 27192
},
{
"epoch": 3.033980582524272,
"grad_norm": 15.25,
"learning_rate": 0.004696601941747573,
"loss": 6.9586,
"step": 27500
},
{
"epoch": 3.089143865842895,
"grad_norm": 12.1875,
"learning_rate": 0.004691085613415711,
"loss": 6.9644,
"step": 28000
},
{
"epoch": 3.144307149161518,
"grad_norm": 12.125,
"learning_rate": 0.0046855692850838486,
"loss": 6.955,
"step": 28500
},
{
"epoch": 3.1994704324801413,
"grad_norm": 11.3125,
"learning_rate": 0.004680052956751986,
"loss": 6.9499,
"step": 29000
},
{
"epoch": 3.2546337157987644,
"grad_norm": 9.0625,
"learning_rate": 0.004674536628420123,
"loss": 6.9522,
"step": 29500
},
{
"epoch": 3.3097969991173875,
"grad_norm": 10.25,
"learning_rate": 0.004669020300088262,
"loss": 6.952,
"step": 30000
},
{
"epoch": 3.3649602824360105,
"grad_norm": 15.6875,
"learning_rate": 0.0046635039717564,
"loss": 6.9524,
"step": 30500
},
{
"epoch": 3.4201235657546336,
"grad_norm": 18.875,
"learning_rate": 0.0046579876434245365,
"loss": 6.9502,
"step": 31000
},
{
"epoch": 3.4752868490732567,
"grad_norm": 8.6875,
"learning_rate": 0.004652471315092674,
"loss": 6.9465,
"step": 31500
},
{
"epoch": 3.5304501323918798,
"grad_norm": 16.375,
"learning_rate": 0.004646954986760812,
"loss": 6.9453,
"step": 32000
},
{
"epoch": 3.585613415710503,
"grad_norm": 20.5,
"learning_rate": 0.00464143865842895,
"loss": 6.9477,
"step": 32500
},
{
"epoch": 3.6407766990291264,
"grad_norm": 15.5625,
"learning_rate": 0.0046359223300970875,
"loss": 6.9507,
"step": 33000
},
{
"epoch": 3.6959399823477495,
"grad_norm": 47.5,
"learning_rate": 0.004630406001765225,
"loss": 6.9513,
"step": 33500
},
{
"epoch": 3.7511032656663725,
"grad_norm": 17.125,
"learning_rate": 0.004624889673433363,
"loss": 6.9489,
"step": 34000
},
{
"epoch": 3.8062665489849956,
"grad_norm": 190.0,
"learning_rate": 0.004619373345101501,
"loss": 6.9486,
"step": 34500
},
{
"epoch": 3.8614298323036187,
"grad_norm": 12.375,
"learning_rate": 0.0046138570167696385,
"loss": 6.954,
"step": 35000
},
{
"epoch": 3.9165931156222418,
"grad_norm": 12.25,
"learning_rate": 0.004608340688437776,
"loss": 6.945,
"step": 35500
},
{
"epoch": 3.971756398940865,
"grad_norm": 9.6875,
"learning_rate": 0.004602824360105914,
"loss": 6.9449,
"step": 36000
},
{
"epoch": 4.0,
"eval_loss": 6.9535441398620605,
"eval_runtime": 9.4952,
"eval_samples_per_second": 52.658,
"eval_steps_per_second": 1.685,
"step": 36256
},
{
"epoch": 4.026919682259488,
"grad_norm": 13.375,
"learning_rate": 0.004597308031774051,
"loss": 6.9449,
"step": 36500
},
{
"epoch": 4.0820829655781115,
"grad_norm": 12.0625,
"learning_rate": 0.0045917917034421895,
"loss": 6.9412,
"step": 37000
},
{
"epoch": 4.1372462488967345,
"grad_norm": 12.8125,
"learning_rate": 0.004586275375110326,
"loss": 6.9405,
"step": 37500
},
{
"epoch": 4.192409532215358,
"grad_norm": 10.6875,
"learning_rate": 0.004580759046778464,
"loss": 6.9472,
"step": 38000
},
{
"epoch": 4.247572815533981,
"grad_norm": 12.625,
"learning_rate": 0.004575242718446602,
"loss": 6.9417,
"step": 38500
},
{
"epoch": 4.302736098852604,
"grad_norm": 7.96875,
"learning_rate": 0.00456972639011474,
"loss": 6.9397,
"step": 39000
},
{
"epoch": 4.357899382171227,
"grad_norm": 9.5625,
"learning_rate": 0.004564210061782877,
"loss": 6.9399,
"step": 39500
},
{
"epoch": 4.41306266548985,
"grad_norm": 8.9375,
"learning_rate": 0.004558693733451015,
"loss": 6.9393,
"step": 40000
},
{
"epoch": 4.468225948808473,
"grad_norm": 12.875,
"learning_rate": 0.004553177405119153,
"loss": 6.9348,
"step": 40500
},
{
"epoch": 4.523389232127096,
"grad_norm": 20.125,
"learning_rate": 0.004547661076787291,
"loss": 6.94,
"step": 41000
},
{
"epoch": 4.578552515445719,
"grad_norm": 12.0,
"learning_rate": 0.0045421447484554275,
"loss": 6.9417,
"step": 41500
},
{
"epoch": 4.633715798764342,
"grad_norm": 14.6875,
"learning_rate": 0.004536628420123566,
"loss": 6.9362,
"step": 42000
},
{
"epoch": 4.688879082082965,
"grad_norm": 12.3125,
"learning_rate": 0.004531112091791704,
"loss": 6.9365,
"step": 42500
},
{
"epoch": 4.744042365401588,
"grad_norm": 9.1875,
"learning_rate": 0.004525595763459841,
"loss": 6.9306,
"step": 43000
},
{
"epoch": 4.7992056487202115,
"grad_norm": 36.5,
"learning_rate": 0.0045200794351279786,
"loss": 6.9657,
"step": 43500
},
{
"epoch": 4.854368932038835,
"grad_norm": 33.25,
"learning_rate": 0.004514563106796117,
"loss": 6.953,
"step": 44000
},
{
"epoch": 4.9095322153574585,
"grad_norm": 20.375,
"learning_rate": 0.004509046778464254,
"loss": 6.9506,
"step": 44500
},
{
"epoch": 4.964695498676081,
"grad_norm": 7.875,
"learning_rate": 0.004503530450132392,
"loss": 6.9484,
"step": 45000
},
{
"epoch": 5.0,
"eval_loss": 6.946237564086914,
"eval_runtime": 8.659,
"eval_samples_per_second": 57.744,
"eval_steps_per_second": 1.848,
"step": 45320
}
],
"logging_steps": 500,
"max_steps": 453200,
"num_input_tokens_seen": 0,
"num_train_epochs": 50,
"save_steps": 500,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 3,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 2.553070427093514e+18,
"train_batch_size": 32,
"trial_name": null,
"trial_params": null
}