{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.995708154506438, "eval_steps": 500, "global_step": 349, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.04, "grad_norm": 3.838202126167235, "learning_rate": 9.997135147120633e-05, "loss": 2.1255, "step": 5 }, { "epoch": 0.09, "grad_norm": 2.1177204859561205, "learning_rate": 9.98854387143534e-05, "loss": 1.7295, "step": 10 }, { "epoch": 0.13, "grad_norm": 1.4704605855835866, "learning_rate": 9.974236018040474e-05, "loss": 1.4756, "step": 15 }, { "epoch": 0.17, "grad_norm": 1.4984947152628136, "learning_rate": 9.954227982894034e-05, "loss": 1.3201, "step": 20 }, { "epoch": 0.21, "grad_norm": 1.3923423053547814, "learning_rate": 9.928542694026862e-05, "loss": 1.28, "step": 25 }, { "epoch": 0.26, "grad_norm": 1.3739605464965448, "learning_rate": 9.897209585268458e-05, "loss": 1.2543, "step": 30 }, { "epoch": 0.3, "grad_norm": 1.275550734636967, "learning_rate": 9.86026456251757e-05, "loss": 1.2261, "step": 35 }, { "epoch": 0.34, "grad_norm": 1.3278212210523785, "learning_rate": 9.817749962596115e-05, "loss": 1.1967, "step": 40 }, { "epoch": 0.39, "grad_norm": 1.407416916788669, "learning_rate": 9.769714504733694e-05, "loss": 1.1876, "step": 45 }, { "epoch": 0.43, "grad_norm": 1.4541097126165785, "learning_rate": 9.716213234738215e-05, "loss": 1.1845, "step": 50 }, { "epoch": 0.47, "grad_norm": 1.4706662049445451, "learning_rate": 9.657307461916635e-05, "loss": 1.1681, "step": 55 }, { "epoch": 0.52, "grad_norm": 1.3719130683126355, "learning_rate": 9.59306468881811e-05, "loss": 1.1622, "step": 60 }, { "epoch": 0.56, "grad_norm": 1.3153375215490886, "learning_rate": 9.52355853388003e-05, "loss": 1.1298, "step": 65 }, { "epoch": 0.6, "grad_norm": 1.488930807431193, "learning_rate": 9.448868647065642e-05, "loss": 1.1598, "step": 70 }, { "epoch": 0.64, "grad_norm": 1.2649466524851178, "learning_rate": 9.369080618589864e-05, "loss": 1.135, "step": 75 }, { "epoch": 0.69, "grad_norm": 1.4161938356640078, "learning_rate": 9.284285880837946e-05, "loss": 1.1325, "step": 80 }, { "epoch": 0.73, "grad_norm": 1.418651947979571, "learning_rate": 9.194581603589328e-05, "loss": 1.1078, "step": 85 }, { "epoch": 0.77, "grad_norm": 1.2926106311909347, "learning_rate": 9.100070582666795e-05, "loss": 1.1134, "step": 90 }, { "epoch": 0.82, "grad_norm": 1.387359204219206, "learning_rate": 9.000861122138517e-05, "loss": 1.0808, "step": 95 }, { "epoch": 0.86, "grad_norm": 1.5961791185416505, "learning_rate": 8.897066910207958e-05, "loss": 1.0832, "step": 100 }, { "epoch": 0.9, "grad_norm": 1.476620650207239, "learning_rate": 8.788806888933881e-05, "loss": 1.0843, "step": 105 }, { "epoch": 0.94, "grad_norm": 1.270735111137924, "learning_rate": 8.676205117929752e-05, "loss": 1.1029, "step": 110 }, { "epoch": 0.99, "grad_norm": 1.425141501784203, "learning_rate": 8.559390632198723e-05, "loss": 1.101, "step": 115 }, { "epoch": 1.0, "eval_loss": 1.1002874374389648, "eval_runtime": 69.0354, "eval_samples_per_second": 4.346, "eval_steps_per_second": 0.55, "step": 116 }, { "epoch": 1.03, "grad_norm": 1.1791383340555186, "learning_rate": 8.438497294267117e-05, "loss": 1.0411, "step": 120 }, { "epoch": 1.07, "grad_norm": 1.333740137623119, "learning_rate": 8.313663640785839e-05, "loss": 1.0033, "step": 125 }, { "epoch": 1.12, "grad_norm": 1.4447991614503835, "learning_rate": 8.185032723775539e-05, "loss": 1.035, "step": 130 }, { "epoch": 1.16, "grad_norm": 1.2844125525377348, "learning_rate": 8.052751946697403e-05, "loss": 0.9736, "step": 135 }, { "epoch": 1.2, "grad_norm": 1.2700768647818643, "learning_rate": 7.916972895537471e-05, "loss": 1.0296, "step": 140 }, { "epoch": 1.24, "grad_norm": 1.3154218666274633, "learning_rate": 7.777851165098012e-05, "loss": 1.0291, "step": 145 }, { "epoch": 1.29, "grad_norm": 1.1964311130128484, "learning_rate": 7.635546180695038e-05, "loss": 0.9498, "step": 150 }, { "epoch": 1.33, "grad_norm": 1.5080490468085683, "learning_rate": 7.490221015466279e-05, "loss": 0.9916, "step": 155 }, { "epoch": 1.37, "grad_norm": 1.4371536477549405, "learning_rate": 7.342042203498951e-05, "loss": 1.0462, "step": 160 }, { "epoch": 1.42, "grad_norm": 1.633026684473331, "learning_rate": 7.191179548991507e-05, "loss": 0.9906, "step": 165 }, { "epoch": 1.46, "grad_norm": 1.5739785726895548, "learning_rate": 7.037805931668005e-05, "loss": 1.0287, "step": 170 }, { "epoch": 1.5, "grad_norm": 1.429154502412058, "learning_rate": 6.882097108668132e-05, "loss": 1.0154, "step": 175 }, { "epoch": 1.55, "grad_norm": 1.1785345522362023, "learning_rate": 6.724231513139852e-05, "loss": 1.0018, "step": 180 }, { "epoch": 1.59, "grad_norm": 1.6392707933966606, "learning_rate": 6.564390049765528e-05, "loss": 0.9824, "step": 185 }, { "epoch": 1.63, "grad_norm": 1.5381793424250634, "learning_rate": 6.402755887455792e-05, "loss": 0.9953, "step": 190 }, { "epoch": 1.67, "grad_norm": 1.2412583778357746, "learning_rate": 6.239514249448767e-05, "loss": 0.9974, "step": 195 }, { "epoch": 1.72, "grad_norm": 1.5172544424431558, "learning_rate": 6.0748522010551215e-05, "loss": 0.9531, "step": 200 }, { "epoch": 1.76, "grad_norm": 1.3293697921192489, "learning_rate": 5.908958435292241e-05, "loss": 0.9842, "step": 205 }, { "epoch": 1.8, "grad_norm": 1.4362543686797302, "learning_rate": 5.742023056653131e-05, "loss": 0.9993, "step": 210 }, { "epoch": 1.85, "grad_norm": 1.3658639776213064, "learning_rate": 5.574237363257858e-05, "loss": 0.9936, "step": 215 }, { "epoch": 1.89, "grad_norm": 1.583719578671925, "learning_rate": 5.4057936276371565e-05, "loss": 0.9686, "step": 220 }, { "epoch": 1.93, "grad_norm": 1.2596980255346535, "learning_rate": 5.236884876399429e-05, "loss": 1.0095, "step": 225 }, { "epoch": 1.97, "grad_norm": 1.3013094438190969, "learning_rate": 5.0677046690336096e-05, "loss": 0.999, "step": 230 }, { "epoch": 2.0, "eval_loss": 1.0417044162750244, "eval_runtime": 68.6388, "eval_samples_per_second": 4.371, "eval_steps_per_second": 0.554, "step": 233 }, { "epoch": 2.02, "grad_norm": 1.2842579006806363, "learning_rate": 4.898446876101379e-05, "loss": 0.9169, "step": 235 }, { "epoch": 2.06, "grad_norm": 1.3799527209106432, "learning_rate": 4.729305457072913e-05, "loss": 0.9044, "step": 240 }, { "epoch": 2.1, "grad_norm": 1.332914263778945, "learning_rate": 4.560474238060739e-05, "loss": 0.8905, "step": 245 }, { "epoch": 2.15, "grad_norm": 1.4483104158765132, "learning_rate": 4.392146689706425e-05, "loss": 0.895, "step": 250 }, { "epoch": 2.19, "grad_norm": 1.2300017765319582, "learning_rate": 4.224515705474603e-05, "loss": 0.9144, "step": 255 }, { "epoch": 2.23, "grad_norm": 1.5671642418549678, "learning_rate": 4.057773380608411e-05, "loss": 0.8692, "step": 260 }, { "epoch": 2.27, "grad_norm": 1.1553793609487584, "learning_rate": 3.892110791999649e-05, "loss": 0.9092, "step": 265 }, { "epoch": 2.32, "grad_norm": 1.3993702614119785, "learning_rate": 3.7277177792259114e-05, "loss": 0.8869, "step": 270 }, { "epoch": 2.36, "grad_norm": 1.512174508677905, "learning_rate": 3.5647827270055945e-05, "loss": 0.9104, "step": 275 }, { "epoch": 2.4, "grad_norm": 1.4126799043823064, "learning_rate": 3.403492349320101e-05, "loss": 0.8837, "step": 280 }, { "epoch": 2.45, "grad_norm": 1.3750228704249976, "learning_rate": 3.244031475450599e-05, "loss": 0.9164, "step": 285 }, { "epoch": 2.49, "grad_norm": 1.5268920387463116, "learning_rate": 3.086582838174551e-05, "loss": 0.8805, "step": 290 }, { "epoch": 2.53, "grad_norm": 1.4206059480865851, "learning_rate": 2.9313268643646986e-05, "loss": 0.8942, "step": 295 }, { "epoch": 2.58, "grad_norm": 1.4073831093700344, "learning_rate": 2.7784414682304832e-05, "loss": 0.9289, "step": 300 }, { "epoch": 2.62, "grad_norm": 1.4564634651821706, "learning_rate": 2.628101847438835e-05, "loss": 0.8864, "step": 305 }, { "epoch": 2.66, "grad_norm": 1.4217201527416836, "learning_rate": 2.4804802823479613e-05, "loss": 0.8947, "step": 310 }, { "epoch": 2.7, "grad_norm": 1.5100802994235183, "learning_rate": 2.3357459385841823e-05, "loss": 0.8829, "step": 315 }, { "epoch": 2.75, "grad_norm": 1.4578776901613366, "learning_rate": 2.194064673188089e-05, "loss": 0.922, "step": 320 }, { "epoch": 2.79, "grad_norm": 1.3183332342703558, "learning_rate": 2.055598844552129e-05, "loss": 0.8785, "step": 325 }, { "epoch": 2.83, "grad_norm": 1.3169540028278115, "learning_rate": 1.920507126367448e-05, "loss": 0.8853, "step": 330 }, { "epoch": 2.88, "grad_norm": 1.4997550219337101, "learning_rate": 1.7889443257931737e-05, "loss": 0.8932, "step": 335 }, { "epoch": 2.92, "grad_norm": 1.2868613064278502, "learning_rate": 1.6610612060565234e-05, "loss": 0.9047, "step": 340 }, { "epoch": 2.96, "grad_norm": 1.3566421396199244, "learning_rate": 1.5370043136870148e-05, "loss": 0.9106, "step": 345 }, { "epoch": 3.0, "eval_loss": 1.0241451263427734, "eval_runtime": 68.638, "eval_samples_per_second": 4.371, "eval_steps_per_second": 0.554, "step": 349 } ], "logging_steps": 5, "max_steps": 464, "num_input_tokens_seen": 0, "num_train_epochs": 4, "save_steps": 500, "total_flos": 2581092118822912.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }