{ "best_metric": 6.946237564086914, "best_model_checkpoint": "./results/models/checkpoint-45320", "epoch": 5.0, "eval_steps": 500, "global_step": 45320, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.05516328331862312, "grad_norm": 0.3359375, "learning_rate": 0.004994483671668138, "loss": 7.1096, "step": 500 }, { "epoch": 0.11032656663724624, "grad_norm": 0.2890625, "learning_rate": 0.004988967343336276, "loss": 6.9824, "step": 1000 }, { "epoch": 0.16548984995586938, "grad_norm": 0.369140625, "learning_rate": 0.004983451015004413, "loss": 6.971, "step": 1500 }, { "epoch": 0.22065313327449249, "grad_norm": 0.8046875, "learning_rate": 0.004977934686672551, "loss": 7.0129, "step": 2000 }, { "epoch": 0.2758164165931156, "grad_norm": 4.59375, "learning_rate": 0.004972418358340689, "loss": 7.0571, "step": 2500 }, { "epoch": 0.33097969991173876, "grad_norm": 4.03125, "learning_rate": 0.004966902030008827, "loss": 7.0532, "step": 3000 }, { "epoch": 0.3861429832303619, "grad_norm": 3.75, "learning_rate": 0.0049613857016769635, "loss": 7.0486, "step": 3500 }, { "epoch": 0.44130626654898497, "grad_norm": 9.125, "learning_rate": 0.004955869373345101, "loss": 7.0385, "step": 4000 }, { "epoch": 0.4964695498676081, "grad_norm": 9.1875, "learning_rate": 0.00495035304501324, "loss": 7.0304, "step": 4500 }, { "epoch": 0.5516328331862312, "grad_norm": 12.625, "learning_rate": 0.004944836716681377, "loss": 7.023, "step": 5000 }, { "epoch": 0.6067961165048543, "grad_norm": 8.6875, "learning_rate": 0.0049393203883495145, "loss": 7.0205, "step": 5500 }, { "epoch": 0.6619593998234775, "grad_norm": 7.6875, "learning_rate": 0.004933804060017652, "loss": 7.025, "step": 6000 }, { "epoch": 0.7171226831421006, "grad_norm": 4.96875, "learning_rate": 0.00492828773168579, "loss": 7.0285, "step": 6500 }, { "epoch": 0.7722859664607238, "grad_norm": 5.46875, "learning_rate": 0.004922771403353928, "loss": 7.025, "step": 7000 }, { "epoch": 0.8274492497793469, "grad_norm": 20.125, "learning_rate": 0.0049172550750220655, "loss": 7.0162, "step": 7500 }, { "epoch": 0.8826125330979699, "grad_norm": 7.1875, "learning_rate": 0.004911738746690203, "loss": 7.0154, "step": 8000 }, { "epoch": 0.9377758164165931, "grad_norm": 8.1875, "learning_rate": 0.004906222418358341, "loss": 7.0104, "step": 8500 }, { "epoch": 0.9929390997352162, "grad_norm": 6.71875, "learning_rate": 0.004900706090026478, "loss": 7.0374, "step": 9000 }, { "epoch": 1.0, "eval_loss": 7.0309014320373535, "eval_runtime": 9.4135, "eval_samples_per_second": 53.115, "eval_steps_per_second": 1.7, "step": 9064 }, { "epoch": 1.0481023830538394, "grad_norm": 3.5, "learning_rate": 0.0048951897616946165, "loss": 7.02, "step": 9500 }, { "epoch": 1.1032656663724625, "grad_norm": 10.5, "learning_rate": 0.004889673433362754, "loss": 7.0297, "step": 10000 }, { "epoch": 1.1584289496910856, "grad_norm": 8.875, "learning_rate": 0.004884157105030891, "loss": 7.0361, "step": 10500 }, { "epoch": 1.2135922330097086, "grad_norm": 3.5625, "learning_rate": 0.004878640776699029, "loss": 7.015, "step": 11000 }, { "epoch": 1.268755516328332, "grad_norm": 9.1875, "learning_rate": 0.0048731244483671676, "loss": 6.9996, "step": 11500 }, { "epoch": 1.323918799646955, "grad_norm": 13.625, "learning_rate": 0.0048676081200353044, "loss": 6.9975, "step": 12000 }, { "epoch": 1.379082082965578, "grad_norm": 6.34375, "learning_rate": 0.004862091791703442, "loss": 7.0081, "step": 12500 }, { "epoch": 1.4342453662842012, "grad_norm": 14.375, "learning_rate": 0.00485657546337158, "loss": 7.0008, "step": 13000 }, { "epoch": 1.4894086496028245, "grad_norm": 8.8125, "learning_rate": 0.004851059135039718, "loss": 6.9962, "step": 13500 }, { "epoch": 1.5445719329214476, "grad_norm": 22.625, "learning_rate": 0.0048455428067078555, "loss": 6.9928, "step": 14000 }, { "epoch": 1.5997352162400706, "grad_norm": 12.6875, "learning_rate": 0.004840026478375993, "loss": 6.9919, "step": 14500 }, { "epoch": 1.6548984995586937, "grad_norm": 53.75, "learning_rate": 0.004834510150044131, "loss": 6.9923, "step": 15000 }, { "epoch": 1.7100617828773168, "grad_norm": 9.9375, "learning_rate": 0.004828993821712269, "loss": 6.9966, "step": 15500 }, { "epoch": 1.7652250661959399, "grad_norm": 13.0625, "learning_rate": 0.004823477493380406, "loss": 6.9908, "step": 16000 }, { "epoch": 1.820388349514563, "grad_norm": 8.9375, "learning_rate": 0.004817961165048544, "loss": 6.9937, "step": 16500 }, { "epoch": 1.8755516328331863, "grad_norm": 42.0, "learning_rate": 0.004812444836716681, "loss": 6.9894, "step": 17000 }, { "epoch": 1.9307149161518093, "grad_norm": 17.5, "learning_rate": 0.004806928508384819, "loss": 6.9926, "step": 17500 }, { "epoch": 1.9858781994704324, "grad_norm": 30.75, "learning_rate": 0.004801412180052957, "loss": 6.9989, "step": 18000 }, { "epoch": 2.0, "eval_loss": 6.996521472930908, "eval_runtime": 8.6549, "eval_samples_per_second": 57.771, "eval_steps_per_second": 1.849, "step": 18128 }, { "epoch": 2.0410414827890557, "grad_norm": 12.125, "learning_rate": 0.004795895851721094, "loss": 6.9915, "step": 18500 }, { "epoch": 2.096204766107679, "grad_norm": 18.625, "learning_rate": 0.004790379523389232, "loss": 6.9839, "step": 19000 }, { "epoch": 2.151368049426302, "grad_norm": 21.25, "learning_rate": 0.00478486319505737, "loss": 6.9821, "step": 19500 }, { "epoch": 2.206531332744925, "grad_norm": 21.375, "learning_rate": 0.004779346866725508, "loss": 6.9842, "step": 20000 }, { "epoch": 2.261694616063548, "grad_norm": 16.625, "learning_rate": 0.004773830538393645, "loss": 6.9836, "step": 20500 }, { "epoch": 2.316857899382171, "grad_norm": 17.125, "learning_rate": 0.004768314210061783, "loss": 6.9837, "step": 21000 }, { "epoch": 2.372021182700794, "grad_norm": 8.9375, "learning_rate": 0.004762797881729921, "loss": 6.9822, "step": 21500 }, { "epoch": 2.4271844660194173, "grad_norm": 8.25, "learning_rate": 0.004757281553398059, "loss": 6.9791, "step": 22000 }, { "epoch": 2.4823477493380404, "grad_norm": 8.375, "learning_rate": 0.0047517652250661955, "loss": 6.986, "step": 22500 }, { "epoch": 2.537511032656664, "grad_norm": 10.6875, "learning_rate": 0.004746248896734333, "loss": 6.9807, "step": 23000 }, { "epoch": 2.592674315975287, "grad_norm": 14.375, "learning_rate": 0.004740732568402472, "loss": 6.9822, "step": 23500 }, { "epoch": 2.64783759929391, "grad_norm": 10.625, "learning_rate": 0.004735216240070609, "loss": 6.9701, "step": 24000 }, { "epoch": 2.703000882612533, "grad_norm": 11.75, "learning_rate": 0.0047296999117387465, "loss": 6.9707, "step": 24500 }, { "epoch": 2.758164165931156, "grad_norm": 30.375, "learning_rate": 0.004724183583406884, "loss": 6.9792, "step": 25000 }, { "epoch": 2.8133274492497793, "grad_norm": 17.25, "learning_rate": 0.004718667255075022, "loss": 6.9803, "step": 25500 }, { "epoch": 2.8684907325684024, "grad_norm": 9.0, "learning_rate": 0.00471315092674316, "loss": 6.9752, "step": 26000 }, { "epoch": 2.9236540158870254, "grad_norm": 6.0, "learning_rate": 0.0047076345984112975, "loss": 6.9687, "step": 26500 }, { "epoch": 2.978817299205649, "grad_norm": 12.0625, "learning_rate": 0.004702118270079435, "loss": 6.9635, "step": 27000 }, { "epoch": 3.0, "eval_loss": 6.960501194000244, "eval_runtime": 8.9893, "eval_samples_per_second": 55.622, "eval_steps_per_second": 1.78, "step": 27192 }, { "epoch": 3.033980582524272, "grad_norm": 15.25, "learning_rate": 0.004696601941747573, "loss": 6.9586, "step": 27500 }, { "epoch": 3.089143865842895, "grad_norm": 12.1875, "learning_rate": 0.004691085613415711, "loss": 6.9644, "step": 28000 }, { "epoch": 3.144307149161518, "grad_norm": 12.125, "learning_rate": 0.0046855692850838486, "loss": 6.955, "step": 28500 }, { "epoch": 3.1994704324801413, "grad_norm": 11.3125, "learning_rate": 0.004680052956751986, "loss": 6.9499, "step": 29000 }, { "epoch": 3.2546337157987644, "grad_norm": 9.0625, "learning_rate": 0.004674536628420123, "loss": 6.9522, "step": 29500 }, { "epoch": 3.3097969991173875, "grad_norm": 10.25, "learning_rate": 0.004669020300088262, "loss": 6.952, "step": 30000 }, { "epoch": 3.3649602824360105, "grad_norm": 15.6875, "learning_rate": 0.0046635039717564, "loss": 6.9524, "step": 30500 }, { "epoch": 3.4201235657546336, "grad_norm": 18.875, "learning_rate": 0.0046579876434245365, "loss": 6.9502, "step": 31000 }, { "epoch": 3.4752868490732567, "grad_norm": 8.6875, "learning_rate": 0.004652471315092674, "loss": 6.9465, "step": 31500 }, { "epoch": 3.5304501323918798, "grad_norm": 16.375, "learning_rate": 0.004646954986760812, "loss": 6.9453, "step": 32000 }, { "epoch": 3.585613415710503, "grad_norm": 20.5, "learning_rate": 0.00464143865842895, "loss": 6.9477, "step": 32500 }, { "epoch": 3.6407766990291264, "grad_norm": 15.5625, "learning_rate": 0.0046359223300970875, "loss": 6.9507, "step": 33000 }, { "epoch": 3.6959399823477495, "grad_norm": 47.5, "learning_rate": 0.004630406001765225, "loss": 6.9513, "step": 33500 }, { "epoch": 3.7511032656663725, "grad_norm": 17.125, "learning_rate": 0.004624889673433363, "loss": 6.9489, "step": 34000 }, { "epoch": 3.8062665489849956, "grad_norm": 190.0, "learning_rate": 0.004619373345101501, "loss": 6.9486, "step": 34500 }, { "epoch": 3.8614298323036187, "grad_norm": 12.375, "learning_rate": 0.0046138570167696385, "loss": 6.954, "step": 35000 }, { "epoch": 3.9165931156222418, "grad_norm": 12.25, "learning_rate": 0.004608340688437776, "loss": 6.945, "step": 35500 }, { "epoch": 3.971756398940865, "grad_norm": 9.6875, "learning_rate": 0.004602824360105914, "loss": 6.9449, "step": 36000 }, { "epoch": 4.0, "eval_loss": 6.9535441398620605, "eval_runtime": 9.4952, "eval_samples_per_second": 52.658, "eval_steps_per_second": 1.685, "step": 36256 }, { "epoch": 4.026919682259488, "grad_norm": 13.375, "learning_rate": 0.004597308031774051, "loss": 6.9449, "step": 36500 }, { "epoch": 4.0820829655781115, "grad_norm": 12.0625, "learning_rate": 0.0045917917034421895, "loss": 6.9412, "step": 37000 }, { "epoch": 4.1372462488967345, "grad_norm": 12.8125, "learning_rate": 0.004586275375110326, "loss": 6.9405, "step": 37500 }, { "epoch": 4.192409532215358, "grad_norm": 10.6875, "learning_rate": 0.004580759046778464, "loss": 6.9472, "step": 38000 }, { "epoch": 4.247572815533981, "grad_norm": 12.625, "learning_rate": 0.004575242718446602, "loss": 6.9417, "step": 38500 }, { "epoch": 4.302736098852604, "grad_norm": 7.96875, "learning_rate": 0.00456972639011474, "loss": 6.9397, "step": 39000 }, { "epoch": 4.357899382171227, "grad_norm": 9.5625, "learning_rate": 0.004564210061782877, "loss": 6.9399, "step": 39500 }, { "epoch": 4.41306266548985, "grad_norm": 8.9375, "learning_rate": 0.004558693733451015, "loss": 6.9393, "step": 40000 }, { "epoch": 4.468225948808473, "grad_norm": 12.875, "learning_rate": 0.004553177405119153, "loss": 6.9348, "step": 40500 }, { "epoch": 4.523389232127096, "grad_norm": 20.125, "learning_rate": 0.004547661076787291, "loss": 6.94, "step": 41000 }, { "epoch": 4.578552515445719, "grad_norm": 12.0, "learning_rate": 0.0045421447484554275, "loss": 6.9417, "step": 41500 }, { "epoch": 4.633715798764342, "grad_norm": 14.6875, "learning_rate": 0.004536628420123566, "loss": 6.9362, "step": 42000 }, { "epoch": 4.688879082082965, "grad_norm": 12.3125, "learning_rate": 0.004531112091791704, "loss": 6.9365, "step": 42500 }, { "epoch": 4.744042365401588, "grad_norm": 9.1875, "learning_rate": 0.004525595763459841, "loss": 6.9306, "step": 43000 }, { "epoch": 4.7992056487202115, "grad_norm": 36.5, "learning_rate": 0.0045200794351279786, "loss": 6.9657, "step": 43500 }, { "epoch": 4.854368932038835, "grad_norm": 33.25, "learning_rate": 0.004514563106796117, "loss": 6.953, "step": 44000 }, { "epoch": 4.9095322153574585, "grad_norm": 20.375, "learning_rate": 0.004509046778464254, "loss": 6.9506, "step": 44500 }, { "epoch": 4.964695498676081, "grad_norm": 7.875, "learning_rate": 0.004503530450132392, "loss": 6.9484, "step": 45000 }, { "epoch": 5.0, "eval_loss": 6.946237564086914, "eval_runtime": 8.659, "eval_samples_per_second": 57.744, "eval_steps_per_second": 1.848, "step": 45320 } ], "logging_steps": 500, "max_steps": 453200, "num_input_tokens_seen": 0, "num_train_epochs": 50, "save_steps": 500, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 3, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 0 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.553070427093514e+18, "train_batch_size": 32, "trial_name": null, "trial_params": null }