{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 300, "global_step": 1089, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0027548209366391185, "grad_norm": 10.8125, "learning_rate": 1e-06, "loss": 1.9168, "step": 1 }, { "epoch": 0.013774104683195593, "grad_norm": 8.4375, "learning_rate": 1.999962411893365e-06, "loss": 1.9099, "step": 5 }, { "epoch": 0.027548209366391185, "grad_norm": 4.90625, "learning_rate": 1.9997327170302815e-06, "loss": 1.8629, "step": 10 }, { "epoch": 0.04132231404958678, "grad_norm": 3.671875, "learning_rate": 1.99929425749243e-06, "loss": 1.8376, "step": 15 }, { "epoch": 0.05509641873278237, "grad_norm": 2.828125, "learning_rate": 1.998647124839145e-06, "loss": 1.8137, "step": 20 }, { "epoch": 0.06887052341597796, "grad_norm": 2.640625, "learning_rate": 1.997791454204984e-06, "loss": 1.8055, "step": 25 }, { "epoch": 0.08264462809917356, "grad_norm": 2.234375, "learning_rate": 1.9967274242715065e-06, "loss": 1.7913, "step": 30 }, { "epoch": 0.09641873278236915, "grad_norm": 2.21875, "learning_rate": 1.995455257229964e-06, "loss": 1.7897, "step": 35 }, { "epoch": 0.11019283746556474, "grad_norm": 1.8828125, "learning_rate": 1.9939752187349e-06, "loss": 1.7752, "step": 40 }, { "epoch": 0.12396694214876033, "grad_norm": 1.8671875, "learning_rate": 1.9922876178486764e-06, "loss": 1.7802, "step": 45 }, { "epoch": 0.13774104683195593, "grad_norm": 1.65625, "learning_rate": 1.9903928069769356e-06, "loss": 1.7686, "step": 50 }, { "epoch": 0.15151515151515152, "grad_norm": 1.765625, "learning_rate": 1.9882911817950105e-06, "loss": 1.7702, "step": 55 }, { "epoch": 0.1652892561983471, "grad_norm": 1.7265625, "learning_rate": 1.985983181165299e-06, "loss": 1.7618, "step": 60 }, { "epoch": 0.1790633608815427, "grad_norm": 1.609375, "learning_rate": 1.98346928704562e-06, "loss": 1.7627, "step": 65 }, { "epoch": 0.1928374655647383, "grad_norm": 1.5625, "learning_rate": 1.9807500243885744e-06, "loss": 1.7665, "step": 70 }, { "epoch": 0.2066115702479339, "grad_norm": 1.515625, "learning_rate": 1.9778259610319187e-06, "loss": 1.755, "step": 75 }, { "epoch": 0.22038567493112948, "grad_norm": 1.5234375, "learning_rate": 1.9746977075799933e-06, "loss": 1.7574, "step": 80 }, { "epoch": 0.23415977961432508, "grad_norm": 1.4609375, "learning_rate": 1.9713659172762126e-06, "loss": 1.7529, "step": 85 }, { "epoch": 0.24793388429752067, "grad_norm": 1.453125, "learning_rate": 1.9678312858666578e-06, "loss": 1.7417, "step": 90 }, { "epoch": 0.26170798898071623, "grad_norm": 1.421875, "learning_rate": 1.964094551454788e-06, "loss": 1.7509, "step": 95 }, { "epoch": 0.27548209366391185, "grad_norm": 1.5, "learning_rate": 1.960156494347309e-06, "loss": 1.7486, "step": 100 }, { "epoch": 0.2892561983471074, "grad_norm": 1.3984375, "learning_rate": 1.9560179368912327e-06, "loss": 1.7531, "step": 105 }, { "epoch": 0.30303030303030304, "grad_norm": 1.40625, "learning_rate": 1.95167974330215e-06, "loss": 1.7435, "step": 110 }, { "epoch": 0.3168044077134986, "grad_norm": 1.4140625, "learning_rate": 1.9471428194837667e-06, "loss": 1.7403, "step": 115 }, { "epoch": 0.3305785123966942, "grad_norm": 1.3984375, "learning_rate": 1.9424081128387337e-06, "loss": 1.7435, "step": 120 }, { "epoch": 0.3443526170798898, "grad_norm": 1.40625, "learning_rate": 1.9374766120708077e-06, "loss": 1.75, "step": 125 }, { "epoch": 0.3581267217630854, "grad_norm": 1.3828125, "learning_rate": 1.932349346978389e-06, "loss": 1.7469, "step": 130 }, { "epoch": 0.371900826446281, "grad_norm": 1.359375, "learning_rate": 1.927027388239482e-06, "loss": 1.734, "step": 135 }, { "epoch": 0.3856749311294766, "grad_norm": 1.40625, "learning_rate": 1.921511847188112e-06, "loss": 1.7385, "step": 140 }, { "epoch": 0.39944903581267216, "grad_norm": 1.390625, "learning_rate": 1.915803875582259e-06, "loss": 1.736, "step": 145 }, { "epoch": 0.4132231404958678, "grad_norm": 1.34375, "learning_rate": 1.9099046653633437e-06, "loss": 1.7336, "step": 150 }, { "epoch": 0.42699724517906334, "grad_norm": 1.3515625, "learning_rate": 1.9038154484073284e-06, "loss": 1.7324, "step": 155 }, { "epoch": 0.44077134986225897, "grad_norm": 1.3828125, "learning_rate": 1.8975374962674753e-06, "loss": 1.7406, "step": 160 }, { "epoch": 0.45454545454545453, "grad_norm": 1.3359375, "learning_rate": 1.8910721199088195e-06, "loss": 1.7461, "step": 165 }, { "epoch": 0.46831955922865015, "grad_norm": 1.3515625, "learning_rate": 1.8844206694344138e-06, "loss": 1.7296, "step": 170 }, { "epoch": 0.4820936639118457, "grad_norm": 1.390625, "learning_rate": 1.877584533803398e-06, "loss": 1.7286, "step": 175 }, { "epoch": 0.49586776859504134, "grad_norm": 1.3203125, "learning_rate": 1.8705651405409566e-06, "loss": 1.7225, "step": 180 }, { "epoch": 0.509641873278237, "grad_norm": 1.328125, "learning_rate": 1.8633639554402234e-06, "loss": 1.7366, "step": 185 }, { "epoch": 0.5234159779614325, "grad_norm": 1.3828125, "learning_rate": 1.8559824822561913e-06, "loss": 1.7353, "step": 190 }, { "epoch": 0.5371900826446281, "grad_norm": 1.3515625, "learning_rate": 1.8484222623917e-06, "loss": 1.7223, "step": 195 }, { "epoch": 0.5509641873278237, "grad_norm": 1.328125, "learning_rate": 1.8406848745755578e-06, "loss": 1.7256, "step": 200 }, { "epoch": 0.5647382920110193, "grad_norm": 1.3671875, "learning_rate": 1.832771934532872e-06, "loss": 1.7288, "step": 205 }, { "epoch": 0.5785123966942148, "grad_norm": 1.34375, "learning_rate": 1.8246850946476505e-06, "loss": 1.7247, "step": 210 }, { "epoch": 0.5922865013774105, "grad_norm": 1.3359375, "learning_rate": 1.8164260436177524e-06, "loss": 1.7221, "step": 215 }, { "epoch": 0.6060606060606061, "grad_norm": 1.3515625, "learning_rate": 1.8079965061022518e-06, "loss": 1.7224, "step": 220 }, { "epoch": 0.6198347107438017, "grad_norm": 1.328125, "learning_rate": 1.7993982423612941e-06, "loss": 1.7206, "step": 225 }, { "epoch": 0.6336088154269972, "grad_norm": 1.3359375, "learning_rate": 1.7906330478885174e-06, "loss": 1.7252, "step": 230 }, { "epoch": 0.6473829201101928, "grad_norm": 1.359375, "learning_rate": 1.7817027530361174e-06, "loss": 1.725, "step": 235 }, { "epoch": 0.6611570247933884, "grad_norm": 1.3359375, "learning_rate": 1.7726092226326315e-06, "loss": 1.7189, "step": 240 }, { "epoch": 0.6749311294765841, "grad_norm": 1.3359375, "learning_rate": 1.7633543555935245e-06, "loss": 1.719, "step": 245 }, { "epoch": 0.6887052341597796, "grad_norm": 1.3359375, "learning_rate": 1.7539400845246564e-06, "loss": 1.7121, "step": 250 }, { "epoch": 0.7024793388429752, "grad_norm": 1.3046875, "learning_rate": 1.744368375318715e-06, "loss": 1.7198, "step": 255 }, { "epoch": 0.7162534435261708, "grad_norm": 1.3359375, "learning_rate": 1.7346412267446958e-06, "loss": 1.7149, "step": 260 }, { "epoch": 0.7300275482093664, "grad_norm": 1.3515625, "learning_rate": 1.724760670030521e-06, "loss": 1.713, "step": 265 }, { "epoch": 0.743801652892562, "grad_norm": 1.3046875, "learning_rate": 1.7147287684388738e-06, "loss": 1.7192, "step": 270 }, { "epoch": 0.7575757575757576, "grad_norm": 1.359375, "learning_rate": 1.7045476168363498e-06, "loss": 1.721, "step": 275 }, { "epoch": 0.7713498622589532, "grad_norm": 1.3359375, "learning_rate": 1.6942193412560043e-06, "loss": 1.7102, "step": 280 }, { "epoch": 0.7851239669421488, "grad_norm": 1.34375, "learning_rate": 1.6837460984533934e-06, "loss": 1.7218, "step": 285 }, { "epoch": 0.7988980716253443, "grad_norm": 1.2890625, "learning_rate": 1.6731300754562008e-06, "loss": 1.7107, "step": 290 }, { "epoch": 0.8126721763085399, "grad_norm": 1.34375, "learning_rate": 1.6623734891075385e-06, "loss": 1.7138, "step": 295 }, { "epoch": 0.8264462809917356, "grad_norm": 1.3203125, "learning_rate": 1.6514785856030272e-06, "loss": 1.7112, "step": 300 }, { "epoch": 0.8264462809917356, "eval_loss": 1.7193405628204346, "eval_runtime": 8.403, "eval_samples_per_second": 83.779, "eval_steps_per_second": 2.618, "step": 300 }, { "epoch": 0.8402203856749312, "grad_norm": 1.359375, "learning_rate": 1.640447640021744e-06, "loss": 1.7211, "step": 305 }, { "epoch": 0.8539944903581267, "grad_norm": 1.2890625, "learning_rate": 1.6292829558511376e-06, "loss": 1.7142, "step": 310 }, { "epoch": 0.8677685950413223, "grad_norm": 1.328125, "learning_rate": 1.6179868645060162e-06, "loss": 1.7028, "step": 315 }, { "epoch": 0.8815426997245179, "grad_norm": 1.34375, "learning_rate": 1.6065617248416967e-06, "loss": 1.7091, "step": 320 }, { "epoch": 0.8953168044077136, "grad_norm": 1.3125, "learning_rate": 1.59500992266143e-06, "loss": 1.7143, "step": 325 }, { "epoch": 0.9090909090909091, "grad_norm": 1.3515625, "learning_rate": 1.5833338702181959e-06, "loss": 1.7105, "step": 330 }, { "epoch": 0.9228650137741047, "grad_norm": 1.40625, "learning_rate": 1.5715360057109744e-06, "loss": 1.7146, "step": 335 }, { "epoch": 0.9366391184573003, "grad_norm": 1.4140625, "learning_rate": 1.5596187927755993e-06, "loss": 1.7145, "step": 340 }, { "epoch": 0.9504132231404959, "grad_norm": 1.3828125, "learning_rate": 1.5475847199703033e-06, "loss": 1.7099, "step": 345 }, { "epoch": 0.9641873278236914, "grad_norm": 1.3359375, "learning_rate": 1.535436300256053e-06, "loss": 1.7143, "step": 350 }, { "epoch": 0.977961432506887, "grad_norm": 1.3828125, "learning_rate": 1.523176070471793e-06, "loss": 1.7131, "step": 355 }, { "epoch": 0.9917355371900827, "grad_norm": 1.34375, "learning_rate": 1.5108065908047014e-06, "loss": 1.7248, "step": 360 }, { "epoch": 1.0055096418732783, "grad_norm": 1.3203125, "learning_rate": 1.4983304442555698e-06, "loss": 1.7079, "step": 365 }, { "epoch": 1.019283746556474, "grad_norm": 1.3671875, "learning_rate": 1.4857502360994204e-06, "loss": 1.712, "step": 370 }, { "epoch": 1.0330578512396693, "grad_norm": 1.34375, "learning_rate": 1.4730685933414714e-06, "loss": 1.703, "step": 375 }, { "epoch": 1.046831955922865, "grad_norm": 1.34375, "learning_rate": 1.4602881641685643e-06, "loss": 1.6944, "step": 380 }, { "epoch": 1.0606060606060606, "grad_norm": 1.375, "learning_rate": 1.4474116173961668e-06, "loss": 1.7032, "step": 385 }, { "epoch": 1.0743801652892562, "grad_norm": 1.375, "learning_rate": 1.4344416419110728e-06, "loss": 1.7122, "step": 390 }, { "epoch": 1.0881542699724518, "grad_norm": 1.359375, "learning_rate": 1.4213809461099033e-06, "loss": 1.703, "step": 395 }, { "epoch": 1.1019283746556474, "grad_norm": 1.3515625, "learning_rate": 1.4082322573335422e-06, "loss": 1.703, "step": 400 }, { "epoch": 1.115702479338843, "grad_norm": 1.328125, "learning_rate": 1.394998321297608e-06, "loss": 1.7024, "step": 405 }, { "epoch": 1.1294765840220387, "grad_norm": 1.359375, "learning_rate": 1.3816819015190943e-06, "loss": 1.7058, "step": 410 }, { "epoch": 1.1432506887052343, "grad_norm": 1.34375, "learning_rate": 1.3682857787392905e-06, "loss": 1.6987, "step": 415 }, { "epoch": 1.1570247933884297, "grad_norm": 1.375, "learning_rate": 1.3548127503431038e-06, "loss": 1.7029, "step": 420 }, { "epoch": 1.1707988980716253, "grad_norm": 1.375, "learning_rate": 1.3412656297749135e-06, "loss": 1.6998, "step": 425 }, { "epoch": 1.184573002754821, "grad_norm": 1.3828125, "learning_rate": 1.327647245951058e-06, "loss": 1.7051, "step": 430 }, { "epoch": 1.1983471074380165, "grad_norm": 1.3359375, "learning_rate": 1.3139604426691072e-06, "loss": 1.7065, "step": 435 }, { "epoch": 1.2121212121212122, "grad_norm": 1.3359375, "learning_rate": 1.300208078014014e-06, "loss": 1.7019, "step": 440 }, { "epoch": 1.2258953168044078, "grad_norm": 1.359375, "learning_rate": 1.2863930237612896e-06, "loss": 1.6999, "step": 445 }, { "epoch": 1.2396694214876034, "grad_norm": 1.3203125, "learning_rate": 1.2725181647773174e-06, "loss": 1.7046, "step": 450 }, { "epoch": 1.2534435261707988, "grad_norm": 1.390625, "learning_rate": 1.2585863984169343e-06, "loss": 1.7069, "step": 455 }, { "epoch": 1.2672176308539944, "grad_norm": 1.359375, "learning_rate": 1.2446006339184035e-06, "loss": 1.7025, "step": 460 }, { "epoch": 1.28099173553719, "grad_norm": 1.40625, "learning_rate": 1.2305637917959058e-06, "loss": 1.7016, "step": 465 }, { "epoch": 1.2947658402203857, "grad_norm": 1.3203125, "learning_rate": 1.2164788032296755e-06, "loss": 1.6997, "step": 470 }, { "epoch": 1.3085399449035813, "grad_norm": 1.328125, "learning_rate": 1.2023486094539124e-06, "loss": 1.7003, "step": 475 }, { "epoch": 1.322314049586777, "grad_norm": 1.3671875, "learning_rate": 1.1881761611425888e-06, "loss": 1.6885, "step": 480 }, { "epoch": 1.3360881542699725, "grad_norm": 1.375, "learning_rate": 1.1739644177932907e-06, "loss": 1.7, "step": 485 }, { "epoch": 1.3498622589531681, "grad_norm": 1.34375, "learning_rate": 1.159716347109213e-06, "loss": 1.6989, "step": 490 }, { "epoch": 1.3636363636363638, "grad_norm": 1.3671875, "learning_rate": 1.1454349243794419e-06, "loss": 1.7009, "step": 495 }, { "epoch": 1.3774104683195592, "grad_norm": 1.359375, "learning_rate": 1.1311231318576545e-06, "loss": 1.7007, "step": 500 }, { "epoch": 1.3911845730027548, "grad_norm": 1.3828125, "learning_rate": 1.1167839581393628e-06, "loss": 1.6918, "step": 505 }, { "epoch": 1.4049586776859504, "grad_norm": 1.296875, "learning_rate": 1.1024203975378335e-06, "loss": 1.6975, "step": 510 }, { "epoch": 1.418732782369146, "grad_norm": 1.3203125, "learning_rate": 1.0880354494588138e-06, "loss": 1.6998, "step": 515 }, { "epoch": 1.4325068870523416, "grad_norm": 1.375, "learning_rate": 1.073632117774195e-06, "loss": 1.6973, "step": 520 }, { "epoch": 1.4462809917355373, "grad_norm": 1.296875, "learning_rate": 1.0592134101947417e-06, "loss": 1.7023, "step": 525 }, { "epoch": 1.4600550964187327, "grad_norm": 1.3125, "learning_rate": 1.0447823376420206e-06, "loss": 1.707, "step": 530 }, { "epoch": 1.4738292011019283, "grad_norm": 1.34375, "learning_rate": 1.0303419136196575e-06, "loss": 1.6916, "step": 535 }, { "epoch": 1.487603305785124, "grad_norm": 1.2890625, "learning_rate": 1.0158951535840576e-06, "loss": 1.6977, "step": 540 }, { "epoch": 1.5013774104683195, "grad_norm": 1.265625, "learning_rate": 1.0014450743147145e-06, "loss": 1.699, "step": 545 }, { "epoch": 1.5151515151515151, "grad_norm": 1.3046875, "learning_rate": 9.869946932842466e-07, "loss": 1.6907, "step": 550 }, { "epoch": 1.5289256198347108, "grad_norm": 1.3046875, "learning_rate": 9.725470280282855e-07, "loss": 1.7001, "step": 555 }, { "epoch": 1.5426997245179064, "grad_norm": 1.3125, "learning_rate": 9.581050955153545e-07, "loss": 1.7054, "step": 560 }, { "epoch": 1.556473829201102, "grad_norm": 1.296875, "learning_rate": 9.43671911516861e-07, "loss": 1.687, "step": 565 }, { "epoch": 1.5702479338842976, "grad_norm": 1.3203125, "learning_rate": 9.292504899773453e-07, "loss": 1.7055, "step": 570 }, { "epoch": 1.5840220385674932, "grad_norm": 1.28125, "learning_rate": 9.148438423851041e-07, "loss": 1.6975, "step": 575 }, { "epoch": 1.5977961432506889, "grad_norm": 1.2890625, "learning_rate": 9.00454977143331e-07, "loss": 1.6997, "step": 580 }, { "epoch": 1.6115702479338843, "grad_norm": 1.296875, "learning_rate": 8.860868989419017e-07, "loss": 1.6983, "step": 585 }, { "epoch": 1.6253443526170799, "grad_norm": 1.3515625, "learning_rate": 8.717426081299308e-07, "loss": 1.6995, "step": 590 }, { "epoch": 1.6391184573002755, "grad_norm": 1.296875, "learning_rate": 8.574251000892386e-07, "loss": 1.6948, "step": 595 }, { "epoch": 1.6528925619834711, "grad_norm": 1.3515625, "learning_rate": 8.431373646088549e-07, "loss": 1.6961, "step": 600 }, { "epoch": 1.6528925619834711, "eval_loss": 1.7039618492126465, "eval_runtime": 8.3736, "eval_samples_per_second": 84.073, "eval_steps_per_second": 2.627, "step": 600 }, { "epoch": 1.6666666666666665, "grad_norm": 1.2890625, "learning_rate": 8.288823852606882e-07, "loss": 1.6915, "step": 605 }, { "epoch": 1.6804407713498621, "grad_norm": 1.296875, "learning_rate": 8.14663138776496e-07, "loss": 1.7006, "step": 610 }, { "epoch": 1.6942148760330578, "grad_norm": 1.2890625, "learning_rate": 8.004825944262805e-07, "loss": 1.7029, "step": 615 }, { "epoch": 1.7079889807162534, "grad_norm": 1.375, "learning_rate": 7.863437133982471e-07, "loss": 1.6942, "step": 620 }, { "epoch": 1.721763085399449, "grad_norm": 1.3125, "learning_rate": 7.722494481804445e-07, "loss": 1.6927, "step": 625 }, { "epoch": 1.7355371900826446, "grad_norm": 1.3046875, "learning_rate": 7.582027419442268e-07, "loss": 1.6929, "step": 630 }, { "epoch": 1.7493112947658402, "grad_norm": 1.28125, "learning_rate": 7.442065279296578e-07, "loss": 1.702, "step": 635 }, { "epoch": 1.7630853994490359, "grad_norm": 1.28125, "learning_rate": 7.302637288329915e-07, "loss": 1.6971, "step": 640 }, { "epoch": 1.7768595041322315, "grad_norm": 1.2734375, "learning_rate": 7.163772561963519e-07, "loss": 1.6923, "step": 645 }, { "epoch": 1.790633608815427, "grad_norm": 1.28125, "learning_rate": 7.02550009799745e-07, "loss": 1.6883, "step": 650 }, { "epoch": 1.8044077134986227, "grad_norm": 1.3046875, "learning_rate": 6.887848770555234e-07, "loss": 1.6991, "step": 655 }, { "epoch": 1.8181818181818183, "grad_norm": 1.296875, "learning_rate": 6.750847324054374e-07, "loss": 1.7069, "step": 660 }, { "epoch": 1.8319559228650137, "grad_norm": 1.2890625, "learning_rate": 6.614524367203906e-07, "loss": 1.6979, "step": 665 }, { "epoch": 1.8457300275482094, "grad_norm": 1.3046875, "learning_rate": 6.478908367030338e-07, "loss": 1.6829, "step": 670 }, { "epoch": 1.859504132231405, "grad_norm": 1.3125, "learning_rate": 6.344027642933128e-07, "loss": 1.692, "step": 675 }, { "epoch": 1.8732782369146006, "grad_norm": 1.28125, "learning_rate": 6.209910360771033e-07, "loss": 1.703, "step": 680 }, { "epoch": 1.887052341597796, "grad_norm": 1.328125, "learning_rate": 6.076584526980484e-07, "loss": 1.706, "step": 685 }, { "epoch": 1.9008264462809916, "grad_norm": 1.2890625, "learning_rate": 5.944077982727285e-07, "loss": 1.6906, "step": 690 }, { "epoch": 1.9146005509641872, "grad_norm": 1.2734375, "learning_rate": 5.812418398092787e-07, "loss": 1.6982, "step": 695 }, { "epoch": 1.9283746556473829, "grad_norm": 1.3359375, "learning_rate": 5.681633266295834e-07, "loss": 1.6917, "step": 700 }, { "epoch": 1.9421487603305785, "grad_norm": 1.28125, "learning_rate": 5.551749897951582e-07, "loss": 1.7014, "step": 705 }, { "epoch": 1.955922865013774, "grad_norm": 1.2734375, "learning_rate": 5.422795415368518e-07, "loss": 1.6861, "step": 710 }, { "epoch": 1.9696969696969697, "grad_norm": 1.3046875, "learning_rate": 5.294796746884745e-07, "loss": 1.6953, "step": 715 }, { "epoch": 1.9834710743801653, "grad_norm": 1.3359375, "learning_rate": 5.167780621244801e-07, "loss": 1.6973, "step": 720 }, { "epoch": 1.997245179063361, "grad_norm": 1.2890625, "learning_rate": 5.041773562018135e-07, "loss": 1.7019, "step": 725 }, { "epoch": 2.0110192837465566, "grad_norm": 1.265625, "learning_rate": 4.91680188206047e-07, "loss": 1.7011, "step": 730 }, { "epoch": 2.024793388429752, "grad_norm": 1.265625, "learning_rate": 4.792891678019115e-07, "loss": 1.7013, "step": 735 }, { "epoch": 2.038567493112948, "grad_norm": 1.2734375, "learning_rate": 4.6700688248834664e-07, "loss": 1.6895, "step": 740 }, { "epoch": 2.0523415977961434, "grad_norm": 1.3125, "learning_rate": 4.548358970581757e-07, "loss": 1.7029, "step": 745 }, { "epoch": 2.0661157024793386, "grad_norm": 1.2734375, "learning_rate": 4.427787530625278e-07, "loss": 1.6931, "step": 750 }, { "epoch": 2.0798898071625342, "grad_norm": 1.2734375, "learning_rate": 4.3083796828010675e-07, "loss": 1.6886, "step": 755 }, { "epoch": 2.09366391184573, "grad_norm": 1.3046875, "learning_rate": 4.190160361914292e-07, "loss": 1.6907, "step": 760 }, { "epoch": 2.1074380165289255, "grad_norm": 1.296875, "learning_rate": 4.07315425458134e-07, "loss": 1.6924, "step": 765 }, { "epoch": 2.121212121212121, "grad_norm": 1.28125, "learning_rate": 3.9573857940747537e-07, "loss": 1.7019, "step": 770 }, { "epoch": 2.1349862258953167, "grad_norm": 1.28125, "learning_rate": 3.8428791552210594e-07, "loss": 1.6975, "step": 775 }, { "epoch": 2.1487603305785123, "grad_norm": 1.2890625, "learning_rate": 3.729658249352563e-07, "loss": 1.6986, "step": 780 }, { "epoch": 2.162534435261708, "grad_norm": 1.3125, "learning_rate": 3.6177467193141886e-07, "loss": 1.6893, "step": 785 }, { "epoch": 2.1763085399449036, "grad_norm": 1.2734375, "learning_rate": 3.5071679345263537e-07, "loss": 1.6833, "step": 790 }, { "epoch": 2.190082644628099, "grad_norm": 1.2734375, "learning_rate": 3.397944986104968e-07, "loss": 1.693, "step": 795 }, { "epoch": 2.203856749311295, "grad_norm": 1.3046875, "learning_rate": 3.290100682039516e-07, "loss": 1.6978, "step": 800 }, { "epoch": 2.2176308539944904, "grad_norm": 1.28125, "learning_rate": 3.1836575424303034e-07, "loss": 1.7019, "step": 805 }, { "epoch": 2.231404958677686, "grad_norm": 1.2890625, "learning_rate": 3.078637794785791e-07, "loss": 1.6977, "step": 810 }, { "epoch": 2.2451790633608817, "grad_norm": 1.28125, "learning_rate": 2.9750633693810224e-07, "loss": 1.6898, "step": 815 }, { "epoch": 2.2589531680440773, "grad_norm": 1.28125, "learning_rate": 2.872955894678153e-07, "loss": 1.6915, "step": 820 }, { "epoch": 2.2727272727272725, "grad_norm": 1.265625, "learning_rate": 2.7723366928099754e-07, "loss": 1.6922, "step": 825 }, { "epoch": 2.2865013774104685, "grad_norm": 1.3046875, "learning_rate": 2.673226775127422e-07, "loss": 1.6922, "step": 830 }, { "epoch": 2.3002754820936637, "grad_norm": 1.3046875, "learning_rate": 2.5756468378119533e-07, "loss": 1.6873, "step": 835 }, { "epoch": 2.3140495867768593, "grad_norm": 1.265625, "learning_rate": 2.4796172575537934e-07, "loss": 1.7068, "step": 840 }, { "epoch": 2.327823691460055, "grad_norm": 1.3203125, "learning_rate": 2.3851580872968435e-07, "loss": 1.6993, "step": 845 }, { "epoch": 2.3415977961432506, "grad_norm": 1.3046875, "learning_rate": 2.292289052051224e-07, "loss": 1.6992, "step": 850 }, { "epoch": 2.355371900826446, "grad_norm": 1.2578125, "learning_rate": 2.2010295447742743e-07, "loss": 1.6891, "step": 855 }, { "epoch": 2.369146005509642, "grad_norm": 1.296875, "learning_rate": 2.111398622320927e-07, "loss": 1.6968, "step": 860 }, { "epoch": 2.3829201101928374, "grad_norm": 1.296875, "learning_rate": 2.0234150014642305e-07, "loss": 1.6946, "step": 865 }, { "epoch": 2.396694214876033, "grad_norm": 1.265625, "learning_rate": 1.937097054986915e-07, "loss": 1.6892, "step": 870 }, { "epoch": 2.4104683195592287, "grad_norm": 1.2578125, "learning_rate": 1.8524628078447602e-07, "loss": 1.6915, "step": 875 }, { "epoch": 2.4242424242424243, "grad_norm": 1.2890625, "learning_rate": 1.769529933402637e-07, "loss": 1.6946, "step": 880 }, { "epoch": 2.43801652892562, "grad_norm": 1.2578125, "learning_rate": 1.6883157497439349e-07, "loss": 1.6975, "step": 885 }, { "epoch": 2.4517906336088156, "grad_norm": 1.2734375, "learning_rate": 1.6088372160541962e-07, "loss": 1.6871, "step": 890 }, { "epoch": 2.465564738292011, "grad_norm": 1.2734375, "learning_rate": 1.531110929079681e-07, "loss": 1.6909, "step": 895 }, { "epoch": 2.479338842975207, "grad_norm": 1.28125, "learning_rate": 1.4551531196616396e-07, "loss": 1.6908, "step": 900 }, { "epoch": 2.479338842975207, "eval_loss": 1.7025996446609497, "eval_runtime": 8.3873, "eval_samples_per_second": 83.936, "eval_steps_per_second": 2.623, "step": 900 }, { "epoch": 2.4931129476584024, "grad_norm": 1.265625, "learning_rate": 1.3809796493469728e-07, "loss": 1.6981, "step": 905 }, { "epoch": 2.5068870523415976, "grad_norm": 1.3046875, "learning_rate": 1.3086060070760196e-07, "loss": 1.6902, "step": 910 }, { "epoch": 2.5206611570247937, "grad_norm": 1.3046875, "learning_rate": 1.23804730594814e-07, "loss": 1.6964, "step": 915 }, { "epoch": 2.534435261707989, "grad_norm": 1.2890625, "learning_rate": 1.1693182800658042e-07, "loss": 1.6884, "step": 920 }, { "epoch": 2.5482093663911844, "grad_norm": 1.2890625, "learning_rate": 1.102433281457802e-07, "loss": 1.6969, "step": 925 }, { "epoch": 2.56198347107438, "grad_norm": 1.265625, "learning_rate": 1.0374062770822411e-07, "loss": 1.7003, "step": 930 }, { "epoch": 2.5757575757575757, "grad_norm": 1.2578125, "learning_rate": 9.742508459099707e-08, "loss": 1.7095, "step": 935 }, { "epoch": 2.5895316804407713, "grad_norm": 1.2890625, "learning_rate": 9.129801760890076e-08, "loss": 1.7026, "step": 940 }, { "epoch": 2.603305785123967, "grad_norm": 1.2578125, "learning_rate": 8.536070621905811e-08, "loss": 1.6964, "step": 945 }, { "epoch": 2.6170798898071626, "grad_norm": 1.3046875, "learning_rate": 7.961439025373617e-08, "loss": 1.6984, "step": 950 }, { "epoch": 2.630853994490358, "grad_norm": 1.28125, "learning_rate": 7.40602696614444e-08, "loss": 1.7022, "step": 955 }, { "epoch": 2.644628099173554, "grad_norm": 1.2734375, "learning_rate": 6.869950425636095e-08, "loss": 1.6955, "step": 960 }, { "epoch": 2.6584022038567494, "grad_norm": 1.265625, "learning_rate": 6.353321347613815e-08, "loss": 1.6962, "step": 965 }, { "epoch": 2.672176308539945, "grad_norm": 1.3828125, "learning_rate": 5.856247614814292e-08, "loss": 1.6914, "step": 970 }, { "epoch": 2.6859504132231407, "grad_norm": 1.296875, "learning_rate": 5.3788330264174506e-08, "loss": 1.6934, "step": 975 }, { "epoch": 2.6997245179063363, "grad_norm": 1.28125, "learning_rate": 4.921177276371069e-08, "loss": 1.6947, "step": 980 }, { "epoch": 2.7134986225895315, "grad_norm": 1.2890625, "learning_rate": 4.483375932572597e-08, "loss": 1.6929, "step": 985 }, { "epoch": 2.7272727272727275, "grad_norm": 1.2890625, "learning_rate": 4.0655204169127156e-08, "loss": 1.6944, "step": 990 }, { "epoch": 2.7410468319559227, "grad_norm": 1.2734375, "learning_rate": 3.667697986184526e-08, "loss": 1.6898, "step": 995 }, { "epoch": 2.7548209366391183, "grad_norm": 1.2578125, "learning_rate": 3.2899917138625055e-08, "loss": 1.7061, "step": 1000 }, { "epoch": 2.768595041322314, "grad_norm": 1.3125, "learning_rate": 2.9324804727551055e-08, "loss": 1.6974, "step": 1005 }, { "epoch": 2.7823691460055096, "grad_norm": 1.2734375, "learning_rate": 2.5952389185344925e-08, "loss": 1.6892, "step": 1010 }, { "epoch": 2.796143250688705, "grad_norm": 1.2421875, "learning_rate": 2.2783374741469186e-08, "loss": 1.696, "step": 1015 }, { "epoch": 2.809917355371901, "grad_norm": 1.28125, "learning_rate": 1.9818423151069406e-08, "loss": 1.6879, "step": 1020 }, { "epoch": 2.8236914600550964, "grad_norm": 1.3125, "learning_rate": 1.705815355678619e-08, "loss": 1.6943, "step": 1025 }, { "epoch": 2.837465564738292, "grad_norm": 1.3046875, "learning_rate": 1.4503142359465925e-08, "loss": 1.6919, "step": 1030 }, { "epoch": 2.8512396694214877, "grad_norm": 1.2421875, "learning_rate": 1.215392309779617e-08, "loss": 1.6907, "step": 1035 }, { "epoch": 2.8650137741046833, "grad_norm": 1.2734375, "learning_rate": 1.0010986336891458e-08, "loss": 1.704, "step": 1040 }, { "epoch": 2.878787878787879, "grad_norm": 1.2578125, "learning_rate": 8.074779565854117e-09, "loss": 1.691, "step": 1045 }, { "epoch": 2.8925619834710745, "grad_norm": 1.3203125, "learning_rate": 6.34570710432869e-09, "loss": 1.6975, "step": 1050 }, { "epoch": 2.90633608815427, "grad_norm": 1.2734375, "learning_rate": 4.824130018072026e-09, "loss": 1.6918, "step": 1055 }, { "epoch": 2.9201101928374653, "grad_norm": 1.2890625, "learning_rate": 3.5103660435551465e-09, "loss": 1.6933, "step": 1060 }, { "epoch": 2.9338842975206614, "grad_norm": 1.3046875, "learning_rate": 2.4046895216136563e-09, "loss": 1.6872, "step": 1065 }, { "epoch": 2.9476584022038566, "grad_norm": 1.28125, "learning_rate": 1.5073313401594568e-09, "loss": 1.696, "step": 1070 }, { "epoch": 2.9614325068870526, "grad_norm": 1.234375, "learning_rate": 8.184788859667557e-10, "loss": 1.6964, "step": 1075 }, { "epoch": 2.975206611570248, "grad_norm": 1.3203125, "learning_rate": 3.3827600554170444e-10, "loss": 1.6941, "step": 1080 }, { "epoch": 2.9889807162534434, "grad_norm": 1.3125, "learning_rate": 6.682297508464608e-11, "loss": 1.6993, "step": 1085 }, { "epoch": 3.0, "step": 1089, "total_flos": 6.427401199279931e+18, "train_loss": 1.711617823146263, "train_runtime": 5145.0339, "train_samples_per_second": 13.545, "train_steps_per_second": 0.212 } ], "logging_steps": 5, "max_steps": 1089, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 300, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 6.427401199279931e+18, "train_batch_size": 8, "trial_name": null, "trial_params": null }