|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.995708154506438, |
|
"eval_steps": 500, |
|
"global_step": 349, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 3.838202126167235, |
|
"learning_rate": 9.997135147120633e-05, |
|
"loss": 2.1255, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 2.1177204859561205, |
|
"learning_rate": 9.98854387143534e-05, |
|
"loss": 1.7295, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 1.4704605855835866, |
|
"learning_rate": 9.974236018040474e-05, |
|
"loss": 1.4756, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 1.4984947152628136, |
|
"learning_rate": 9.954227982894034e-05, |
|
"loss": 1.3201, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 1.3923423053547814, |
|
"learning_rate": 9.928542694026862e-05, |
|
"loss": 1.28, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 1.3739605464965448, |
|
"learning_rate": 9.897209585268458e-05, |
|
"loss": 1.2543, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 1.275550734636967, |
|
"learning_rate": 9.86026456251757e-05, |
|
"loss": 1.2261, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 1.3278212210523785, |
|
"learning_rate": 9.817749962596115e-05, |
|
"loss": 1.1967, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 1.407416916788669, |
|
"learning_rate": 9.769714504733694e-05, |
|
"loss": 1.1876, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 1.4541097126165785, |
|
"learning_rate": 9.716213234738215e-05, |
|
"loss": 1.1845, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 1.4706662049445451, |
|
"learning_rate": 9.657307461916635e-05, |
|
"loss": 1.1681, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 1.3719130683126355, |
|
"learning_rate": 9.59306468881811e-05, |
|
"loss": 1.1622, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 1.3153375215490886, |
|
"learning_rate": 9.52355853388003e-05, |
|
"loss": 1.1298, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 1.488930807431193, |
|
"learning_rate": 9.448868647065642e-05, |
|
"loss": 1.1598, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 1.2649466524851178, |
|
"learning_rate": 9.369080618589864e-05, |
|
"loss": 1.135, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 1.4161938356640078, |
|
"learning_rate": 9.284285880837946e-05, |
|
"loss": 1.1325, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 1.418651947979571, |
|
"learning_rate": 9.194581603589328e-05, |
|
"loss": 1.1078, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 1.2926106311909347, |
|
"learning_rate": 9.100070582666795e-05, |
|
"loss": 1.1134, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 1.387359204219206, |
|
"learning_rate": 9.000861122138517e-05, |
|
"loss": 1.0808, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 1.5961791185416505, |
|
"learning_rate": 8.897066910207958e-05, |
|
"loss": 1.0832, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 1.476620650207239, |
|
"learning_rate": 8.788806888933881e-05, |
|
"loss": 1.0843, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 1.270735111137924, |
|
"learning_rate": 8.676205117929752e-05, |
|
"loss": 1.1029, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 1.425141501784203, |
|
"learning_rate": 8.559390632198723e-05, |
|
"loss": 1.101, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 1.1002874374389648, |
|
"eval_runtime": 69.0354, |
|
"eval_samples_per_second": 4.346, |
|
"eval_steps_per_second": 0.55, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 1.1791383340555186, |
|
"learning_rate": 8.438497294267117e-05, |
|
"loss": 1.0411, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 1.333740137623119, |
|
"learning_rate": 8.313663640785839e-05, |
|
"loss": 1.0033, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 1.4447991614503835, |
|
"learning_rate": 8.185032723775539e-05, |
|
"loss": 1.035, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 1.2844125525377348, |
|
"learning_rate": 8.052751946697403e-05, |
|
"loss": 0.9736, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 1.2700768647818643, |
|
"learning_rate": 7.916972895537471e-05, |
|
"loss": 1.0296, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 1.3154218666274633, |
|
"learning_rate": 7.777851165098012e-05, |
|
"loss": 1.0291, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 1.1964311130128484, |
|
"learning_rate": 7.635546180695038e-05, |
|
"loss": 0.9498, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 1.5080490468085683, |
|
"learning_rate": 7.490221015466279e-05, |
|
"loss": 0.9916, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 1.4371536477549405, |
|
"learning_rate": 7.342042203498951e-05, |
|
"loss": 1.0462, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 1.633026684473331, |
|
"learning_rate": 7.191179548991507e-05, |
|
"loss": 0.9906, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 1.5739785726895548, |
|
"learning_rate": 7.037805931668005e-05, |
|
"loss": 1.0287, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 1.429154502412058, |
|
"learning_rate": 6.882097108668132e-05, |
|
"loss": 1.0154, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 1.1785345522362023, |
|
"learning_rate": 6.724231513139852e-05, |
|
"loss": 1.0018, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 1.6392707933966606, |
|
"learning_rate": 6.564390049765528e-05, |
|
"loss": 0.9824, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 1.5381793424250634, |
|
"learning_rate": 6.402755887455792e-05, |
|
"loss": 0.9953, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"grad_norm": 1.2412583778357746, |
|
"learning_rate": 6.239514249448767e-05, |
|
"loss": 0.9974, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 1.5172544424431558, |
|
"learning_rate": 6.0748522010551215e-05, |
|
"loss": 0.9531, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 1.3293697921192489, |
|
"learning_rate": 5.908958435292241e-05, |
|
"loss": 0.9842, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 1.4362543686797302, |
|
"learning_rate": 5.742023056653131e-05, |
|
"loss": 0.9993, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 1.3658639776213064, |
|
"learning_rate": 5.574237363257858e-05, |
|
"loss": 0.9936, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 1.583719578671925, |
|
"learning_rate": 5.4057936276371565e-05, |
|
"loss": 0.9686, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 1.2596980255346535, |
|
"learning_rate": 5.236884876399429e-05, |
|
"loss": 1.0095, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 1.3013094438190969, |
|
"learning_rate": 5.0677046690336096e-05, |
|
"loss": 0.999, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 1.0417044162750244, |
|
"eval_runtime": 68.6388, |
|
"eval_samples_per_second": 4.371, |
|
"eval_steps_per_second": 0.554, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"grad_norm": 1.2842579006806363, |
|
"learning_rate": 4.898446876101379e-05, |
|
"loss": 0.9169, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"grad_norm": 1.3799527209106432, |
|
"learning_rate": 4.729305457072913e-05, |
|
"loss": 0.9044, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"grad_norm": 1.332914263778945, |
|
"learning_rate": 4.560474238060739e-05, |
|
"loss": 0.8905, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"grad_norm": 1.4483104158765132, |
|
"learning_rate": 4.392146689706425e-05, |
|
"loss": 0.895, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"grad_norm": 1.2300017765319582, |
|
"learning_rate": 4.224515705474603e-05, |
|
"loss": 0.9144, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"grad_norm": 1.5671642418549678, |
|
"learning_rate": 4.057773380608411e-05, |
|
"loss": 0.8692, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"grad_norm": 1.1553793609487584, |
|
"learning_rate": 3.892110791999649e-05, |
|
"loss": 0.9092, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"grad_norm": 1.3993702614119785, |
|
"learning_rate": 3.7277177792259114e-05, |
|
"loss": 0.8869, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"grad_norm": 1.512174508677905, |
|
"learning_rate": 3.5647827270055945e-05, |
|
"loss": 0.9104, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 1.4126799043823064, |
|
"learning_rate": 3.403492349320101e-05, |
|
"loss": 0.8837, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"grad_norm": 1.3750228704249976, |
|
"learning_rate": 3.244031475450599e-05, |
|
"loss": 0.9164, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"grad_norm": 1.5268920387463116, |
|
"learning_rate": 3.086582838174551e-05, |
|
"loss": 0.8805, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"grad_norm": 1.4206059480865851, |
|
"learning_rate": 2.9313268643646986e-05, |
|
"loss": 0.8942, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"grad_norm": 1.4073831093700344, |
|
"learning_rate": 2.7784414682304832e-05, |
|
"loss": 0.9289, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"grad_norm": 1.4564634651821706, |
|
"learning_rate": 2.628101847438835e-05, |
|
"loss": 0.8864, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"grad_norm": 1.4217201527416836, |
|
"learning_rate": 2.4804802823479613e-05, |
|
"loss": 0.8947, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"grad_norm": 1.5100802994235183, |
|
"learning_rate": 2.3357459385841823e-05, |
|
"loss": 0.8829, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"grad_norm": 1.4578776901613366, |
|
"learning_rate": 2.194064673188089e-05, |
|
"loss": 0.922, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"grad_norm": 1.3183332342703558, |
|
"learning_rate": 2.055598844552129e-05, |
|
"loss": 0.8785, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"grad_norm": 1.3169540028278115, |
|
"learning_rate": 1.920507126367448e-05, |
|
"loss": 0.8853, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"grad_norm": 1.4997550219337101, |
|
"learning_rate": 1.7889443257931737e-05, |
|
"loss": 0.8932, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"grad_norm": 1.2868613064278502, |
|
"learning_rate": 1.6610612060565234e-05, |
|
"loss": 0.9047, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"grad_norm": 1.3566421396199244, |
|
"learning_rate": 1.5370043136870148e-05, |
|
"loss": 0.9106, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 1.0241451263427734, |
|
"eval_runtime": 68.638, |
|
"eval_samples_per_second": 4.371, |
|
"eval_steps_per_second": 0.554, |
|
"step": 349 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 464, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 500, |
|
"total_flos": 2581092118822912.0, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|