diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,4160 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.9998993609419817, + "eval_steps": 150, + "global_step": 2484, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0040255623207366776, + "grad_norm": 38.58929443359375, + "learning_rate": 4.000000000000001e-06, + "loss": 9.0093, + "num_input_tokens_seen": 327680, + "step": 5 + }, + { + "epoch": 0.008051124641473355, + "grad_norm": 16.55693817138672, + "learning_rate": 8.000000000000001e-06, + "loss": 8.4074, + "num_input_tokens_seen": 655360, + "step": 10 + }, + { + "epoch": 0.012076686962210034, + "grad_norm": 10.222895622253418, + "learning_rate": 1.2e-05, + "loss": 7.9014, + "num_input_tokens_seen": 983040, + "step": 15 + }, + { + "epoch": 0.01610224928294671, + "grad_norm": 8.999139785766602, + "learning_rate": 1.6000000000000003e-05, + "loss": 7.6741, + "num_input_tokens_seen": 1310720, + "step": 20 + }, + { + "epoch": 0.02012781160368339, + "grad_norm": 8.544699668884277, + "learning_rate": 2e-05, + "loss": 7.3022, + "num_input_tokens_seen": 1638400, + "step": 25 + }, + { + "epoch": 0.024153373924420067, + "grad_norm": 7.598087310791016, + "learning_rate": 2.4e-05, + "loss": 6.7376, + "num_input_tokens_seen": 1966080, + "step": 30 + }, + { + "epoch": 0.028178936245156744, + "grad_norm": 5.1022233963012695, + "learning_rate": 2.8e-05, + "loss": 6.2707, + "num_input_tokens_seen": 2293760, + "step": 35 + }, + { + "epoch": 0.03220449856589342, + "grad_norm": 3.5978548526763916, + "learning_rate": 3.2000000000000005e-05, + "loss": 5.9112, + "num_input_tokens_seen": 2621440, + "step": 40 + }, + { + "epoch": 0.036230060886630104, + "grad_norm": 2.6466448307037354, + "learning_rate": 3.6e-05, + "loss": 5.6431, + "num_input_tokens_seen": 2949120, + "step": 45 + }, + { + "epoch": 0.04025562320736678, + "grad_norm": 1.9538609981536865, + "learning_rate": 4e-05, + "loss": 5.4321, + "num_input_tokens_seen": 3276800, + "step": 50 + }, + { + "epoch": 0.04428118552810346, + "grad_norm": 1.7290977239608765, + "learning_rate": 4.4000000000000006e-05, + "loss": 5.2876, + "num_input_tokens_seen": 3604480, + "step": 55 + }, + { + "epoch": 0.048306747848840134, + "grad_norm": 1.647463321685791, + "learning_rate": 4.8e-05, + "loss": 5.1749, + "num_input_tokens_seen": 3932160, + "step": 60 + }, + { + "epoch": 0.05233231016957681, + "grad_norm": 1.367295742034912, + "learning_rate": 5.2000000000000004e-05, + "loss": 5.0826, + "num_input_tokens_seen": 4259840, + "step": 65 + }, + { + "epoch": 0.05635787249031349, + "grad_norm": 1.4710743427276611, + "learning_rate": 5.6e-05, + "loss": 5.0053, + "num_input_tokens_seen": 4587520, + "step": 70 + }, + { + "epoch": 0.06038343481105017, + "grad_norm": 1.501865029335022, + "learning_rate": 6.000000000000001e-05, + "loss": 4.8943, + "num_input_tokens_seen": 4915200, + "step": 75 + }, + { + "epoch": 0.06440899713178684, + "grad_norm": 1.3167399168014526, + "learning_rate": 6.400000000000001e-05, + "loss": 4.9062, + "num_input_tokens_seen": 5242880, + "step": 80 + }, + { + "epoch": 0.06843455945252352, + "grad_norm": 1.6458464860916138, + "learning_rate": 6.8e-05, + "loss": 4.8432, + "num_input_tokens_seen": 5570560, + "step": 85 + }, + { + "epoch": 0.07246012177326021, + "grad_norm": 1.4131243228912354, + "learning_rate": 7.2e-05, + "loss": 4.8507, + "num_input_tokens_seen": 5898240, + "step": 90 + }, + { + "epoch": 0.07648568409399688, + "grad_norm": 1.6352394819259644, + "learning_rate": 7.6e-05, + "loss": 4.8214, + "num_input_tokens_seen": 6225920, + "step": 95 + }, + { + "epoch": 0.08051124641473356, + "grad_norm": 1.3868008852005005, + "learning_rate": 8e-05, + "loss": 4.7863, + "num_input_tokens_seen": 6553600, + "step": 100 + }, + { + "epoch": 0.08453680873547023, + "grad_norm": 1.5053304433822632, + "learning_rate": 7.807200583588265e-05, + "loss": 4.7478, + "num_input_tokens_seen": 6881280, + "step": 105 + }, + { + "epoch": 0.08856237105620691, + "grad_norm": 1.6993811130523682, + "learning_rate": 7.627700713964738e-05, + "loss": 4.7303, + "num_input_tokens_seen": 7208960, + "step": 110 + }, + { + "epoch": 0.0925879333769436, + "grad_norm": 1.4282785654067993, + "learning_rate": 7.460038465922511e-05, + "loss": 4.7332, + "num_input_tokens_seen": 7536640, + "step": 115 + }, + { + "epoch": 0.09661349569768027, + "grad_norm": 1.6601049900054932, + "learning_rate": 7.302967433402216e-05, + "loss": 4.7639, + "num_input_tokens_seen": 7864320, + "step": 120 + }, + { + "epoch": 0.10063905801841695, + "grad_norm": 1.4113882780075073, + "learning_rate": 7.155417527999328e-05, + "loss": 4.7138, + "num_input_tokens_seen": 8192000, + "step": 125 + }, + { + "epoch": 0.10466462033915362, + "grad_norm": 1.5578097105026245, + "learning_rate": 7.016464154456235e-05, + "loss": 4.7051, + "num_input_tokens_seen": 8519680, + "step": 130 + }, + { + "epoch": 0.1086901826598903, + "grad_norm": 1.4274314641952515, + "learning_rate": 6.885303726590964e-05, + "loss": 4.736, + "num_input_tokens_seen": 8847360, + "step": 135 + }, + { + "epoch": 0.11271574498062698, + "grad_norm": 1.354647159576416, + "learning_rate": 6.761234037828134e-05, + "loss": 4.6927, + "num_input_tokens_seen": 9175040, + "step": 140 + }, + { + "epoch": 0.11674130730136366, + "grad_norm": 1.402042031288147, + "learning_rate": 6.643638388299198e-05, + "loss": 4.703, + "num_input_tokens_seen": 9502720, + "step": 145 + }, + { + "epoch": 0.12076686962210034, + "grad_norm": 1.4121787548065186, + "learning_rate": 6.531972647421809e-05, + "loss": 4.6583, + "num_input_tokens_seen": 9830400, + "step": 150 + }, + { + "epoch": 0.12076686962210034, + "eval_accuracy": 0.3405695459290188, + "eval_loss": 4.505166053771973, + "eval_runtime": 6.5101, + "eval_samples_per_second": 46.082, + "eval_steps_per_second": 5.837, + "num_input_tokens_seen": 9830400, + "step": 150 + }, + { + "epoch": 0.12479243194283701, + "grad_norm": 1.5172020196914673, + "learning_rate": 6.425754631219992e-05, + "loss": 4.6715, + "num_input_tokens_seen": 10158080, + "step": 155 + }, + { + "epoch": 0.12881799426357368, + "grad_norm": 1.4244070053100586, + "learning_rate": 6.324555320336759e-05, + "loss": 4.6351, + "num_input_tokens_seen": 10485760, + "step": 160 + }, + { + "epoch": 0.13284355658431038, + "grad_norm": 1.352230191230774, + "learning_rate": 6.227991553292185e-05, + "loss": 4.6386, + "num_input_tokens_seen": 10813440, + "step": 165 + }, + { + "epoch": 0.13686911890504705, + "grad_norm": 1.3570430278778076, + "learning_rate": 6.135719910778964e-05, + "loss": 4.6563, + "num_input_tokens_seen": 11141120, + "step": 170 + }, + { + "epoch": 0.14089468122578372, + "grad_norm": 1.3318084478378296, + "learning_rate": 6.047431568147636e-05, + "loss": 4.6472, + "num_input_tokens_seen": 11468800, + "step": 175 + }, + { + "epoch": 0.14492024354652042, + "grad_norm": 1.4813034534454346, + "learning_rate": 5.96284793999944e-05, + "loss": 4.6493, + "num_input_tokens_seen": 11796480, + "step": 180 + }, + { + "epoch": 0.1489458058672571, + "grad_norm": 1.367981195449829, + "learning_rate": 5.881716976750463e-05, + "loss": 4.5858, + "num_input_tokens_seen": 12124160, + "step": 185 + }, + { + "epoch": 0.15297136818799376, + "grad_norm": 1.3320348262786865, + "learning_rate": 5.803810000880094e-05, + "loss": 4.5985, + "num_input_tokens_seen": 12451840, + "step": 190 + }, + { + "epoch": 0.15699693050873043, + "grad_norm": 1.233577847480774, + "learning_rate": 5.7289189923154636e-05, + "loss": 4.6574, + "num_input_tokens_seen": 12779520, + "step": 195 + }, + { + "epoch": 0.16102249282946712, + "grad_norm": 1.2315726280212402, + "learning_rate": 5.6568542494923805e-05, + "loss": 4.6412, + "num_input_tokens_seen": 13107200, + "step": 200 + }, + { + "epoch": 0.1650480551502038, + "grad_norm": 1.3596357107162476, + "learning_rate": 5.5874423661566265e-05, + "loss": 4.6032, + "num_input_tokens_seen": 13434880, + "step": 205 + }, + { + "epoch": 0.16907361747094046, + "grad_norm": 1.3829398155212402, + "learning_rate": 5.5205244747388335e-05, + "loss": 4.5954, + "num_input_tokens_seen": 13762560, + "step": 210 + }, + { + "epoch": 0.17309917979167716, + "grad_norm": 1.4083398580551147, + "learning_rate": 5.455954715763789e-05, + "loss": 4.5888, + "num_input_tokens_seen": 14090240, + "step": 215 + }, + { + "epoch": 0.17712474211241383, + "grad_norm": 1.467612862586975, + "learning_rate": 5.393598899705937e-05, + "loss": 4.6193, + "num_input_tokens_seen": 14417920, + "step": 220 + }, + { + "epoch": 0.1811503044331505, + "grad_norm": 1.2575774192810059, + "learning_rate": 5.333333333333333e-05, + "loss": 4.5579, + "num_input_tokens_seen": 14745600, + "step": 225 + }, + { + "epoch": 0.1851758667538872, + "grad_norm": 1.4239579439163208, + "learning_rate": 5.2750437871662975e-05, + "loss": 4.5607, + "num_input_tokens_seen": 15073280, + "step": 230 + }, + { + "epoch": 0.18920142907462387, + "grad_norm": 1.2363418340682983, + "learning_rate": 5.218624584427538e-05, + "loss": 4.5617, + "num_input_tokens_seen": 15400960, + "step": 235 + }, + { + "epoch": 0.19322699139536054, + "grad_norm": 1.242336392402649, + "learning_rate": 5.163977794943223e-05, + "loss": 4.5878, + "num_input_tokens_seen": 15728640, + "step": 240 + }, + { + "epoch": 0.1972525537160972, + "grad_norm": 1.2200363874435425, + "learning_rate": 5.11101251999952e-05, + "loss": 4.5518, + "num_input_tokens_seen": 16056320, + "step": 245 + }, + { + "epoch": 0.2012781160368339, + "grad_norm": 1.1616241931915283, + "learning_rate": 5.0596442562694074e-05, + "loss": 4.5944, + "num_input_tokens_seen": 16384000, + "step": 250 + }, + { + "epoch": 0.20530367835757057, + "grad_norm": 1.3502942323684692, + "learning_rate": 5.009794328681197e-05, + "loss": 4.5356, + "num_input_tokens_seen": 16711680, + "step": 255 + }, + { + "epoch": 0.20932924067830724, + "grad_norm": 1.1606285572052002, + "learning_rate": 4.961389383568338e-05, + "loss": 4.5341, + "num_input_tokens_seen": 17039360, + "step": 260 + }, + { + "epoch": 0.21335480299904394, + "grad_norm": 1.4353660345077515, + "learning_rate": 4.9143609346716104e-05, + "loss": 4.5766, + "num_input_tokens_seen": 17367040, + "step": 265 + }, + { + "epoch": 0.2173803653197806, + "grad_norm": 1.2260538339614868, + "learning_rate": 4.8686449556014764e-05, + "loss": 4.5123, + "num_input_tokens_seen": 17694720, + "step": 270 + }, + { + "epoch": 0.22140592764051728, + "grad_norm": 1.2337218523025513, + "learning_rate": 4.8241815132442184e-05, + "loss": 4.4989, + "num_input_tokens_seen": 18022400, + "step": 275 + }, + { + "epoch": 0.22543148996125395, + "grad_norm": 1.2576812505722046, + "learning_rate": 4.7809144373375745e-05, + "loss": 4.5362, + "num_input_tokens_seen": 18350080, + "step": 280 + }, + { + "epoch": 0.22945705228199065, + "grad_norm": 1.3166711330413818, + "learning_rate": 4.738791022072739e-05, + "loss": 4.5318, + "num_input_tokens_seen": 18677760, + "step": 285 + }, + { + "epoch": 0.23348261460272732, + "grad_norm": 1.353780746459961, + "learning_rate": 4.6977617561176284e-05, + "loss": 4.5475, + "num_input_tokens_seen": 19005440, + "step": 290 + }, + { + "epoch": 0.237508176923464, + "grad_norm": 1.2797174453735352, + "learning_rate": 4.657780077916657e-05, + "loss": 4.4992, + "num_input_tokens_seen": 19333120, + "step": 295 + }, + { + "epoch": 0.24153373924420068, + "grad_norm": 1.433968424797058, + "learning_rate": 4.618802153517007e-05, + "loss": 4.5365, + "num_input_tokens_seen": 19660800, + "step": 300 + }, + { + "epoch": 0.24153373924420068, + "eval_accuracy": 0.35246274229682784, + "eval_loss": 4.371213436126709, + "eval_runtime": 6.282, + "eval_samples_per_second": 47.755, + "eval_steps_per_second": 6.049, + "num_input_tokens_seen": 19660800, + "step": 300 + }, + { + "epoch": 0.24555930156493735, + "grad_norm": 1.3408712148666382, + "learning_rate": 4.580786674510946e-05, + "loss": 4.4919, + "num_input_tokens_seen": 19988480, + "step": 305 + }, + { + "epoch": 0.24958486388567402, + "grad_norm": 1.3724451065063477, + "learning_rate": 4.543694673976519e-05, + "loss": 4.498, + "num_input_tokens_seen": 20316160, + "step": 310 + }, + { + "epoch": 0.2536104262064107, + "grad_norm": 1.3110082149505615, + "learning_rate": 4.507489358552088e-05, + "loss": 4.4623, + "num_input_tokens_seen": 20643840, + "step": 315 + }, + { + "epoch": 0.25763598852714736, + "grad_norm": 1.2995046377182007, + "learning_rate": 4.47213595499958e-05, + "loss": 4.5249, + "num_input_tokens_seen": 20971520, + "step": 320 + }, + { + "epoch": 0.26166155084788406, + "grad_norm": 1.2765010595321655, + "learning_rate": 4.4376015698018335e-05, + "loss": 4.5525, + "num_input_tokens_seen": 21299200, + "step": 325 + }, + { + "epoch": 0.26568711316862076, + "grad_norm": 1.268876552581787, + "learning_rate": 4.403855060505443e-05, + "loss": 4.4876, + "num_input_tokens_seen": 21626880, + "step": 330 + }, + { + "epoch": 0.2697126754893574, + "grad_norm": 1.4470906257629395, + "learning_rate": 4.3708669176652244e-05, + "loss": 4.4901, + "num_input_tokens_seen": 21954560, + "step": 335 + }, + { + "epoch": 0.2737382378100941, + "grad_norm": 1.3335119485855103, + "learning_rate": 4.3386091563731235e-05, + "loss": 4.4488, + "num_input_tokens_seen": 22282240, + "step": 340 + }, + { + "epoch": 0.2777638001308308, + "grad_norm": 1.4562543630599976, + "learning_rate": 4.3070552164653247e-05, + "loss": 4.4705, + "num_input_tokens_seen": 22609920, + "step": 345 + }, + { + "epoch": 0.28178936245156744, + "grad_norm": 1.4459863901138306, + "learning_rate": 4.276179870598791e-05, + "loss": 4.4942, + "num_input_tokens_seen": 22937600, + "step": 350 + }, + { + "epoch": 0.28581492477230414, + "grad_norm": 1.193866491317749, + "learning_rate": 4.2459591394742024e-05, + "loss": 4.5002, + "num_input_tokens_seen": 23265280, + "step": 355 + }, + { + "epoch": 0.28984048709304083, + "grad_norm": 1.4731342792510986, + "learning_rate": 4.2163702135578394e-05, + "loss": 4.4794, + "num_input_tokens_seen": 23592960, + "step": 360 + }, + { + "epoch": 0.2938660494137775, + "grad_norm": 1.1683562994003296, + "learning_rate": 4.1873913807217094e-05, + "loss": 4.4578, + "num_input_tokens_seen": 23920640, + "step": 365 + }, + { + "epoch": 0.2978916117345142, + "grad_norm": 1.2372169494628906, + "learning_rate": 4.159001959280291e-05, + "loss": 4.5118, + "num_input_tokens_seen": 24248320, + "step": 370 + }, + { + "epoch": 0.30191717405525087, + "grad_norm": 1.4430092573165894, + "learning_rate": 4.131182235954578e-05, + "loss": 4.4861, + "num_input_tokens_seen": 24576000, + "step": 375 + }, + { + "epoch": 0.3059427363759875, + "grad_norm": 1.3162908554077148, + "learning_rate": 4.103913408340617e-05, + "loss": 4.4968, + "num_input_tokens_seen": 24903680, + "step": 380 + }, + { + "epoch": 0.3099682986967242, + "grad_norm": 1.238364815711975, + "learning_rate": 4.077177531501004e-05, + "loss": 4.4465, + "num_input_tokens_seen": 25231360, + "step": 385 + }, + { + "epoch": 0.31399386101746085, + "grad_norm": 1.2819691896438599, + "learning_rate": 4.050957468334667e-05, + "loss": 4.4656, + "num_input_tokens_seen": 25559040, + "step": 390 + }, + { + "epoch": 0.31801942333819755, + "grad_norm": 1.2802215814590454, + "learning_rate": 4.025236843413021e-05, + "loss": 4.4934, + "num_input_tokens_seen": 25886720, + "step": 395 + }, + { + "epoch": 0.32204498565893425, + "grad_norm": 1.232289433479309, + "learning_rate": 4e-05, + "loss": 4.4965, + "num_input_tokens_seen": 26214400, + "step": 400 + }, + { + "epoch": 0.3260705479796709, + "grad_norm": 1.3022078275680542, + "learning_rate": 3.975231959999627e-05, + "loss": 4.4332, + "num_input_tokens_seen": 26542080, + "step": 405 + }, + { + "epoch": 0.3300961103004076, + "grad_norm": 1.1387150287628174, + "learning_rate": 3.950918386598359e-05, + "loss": 4.4345, + "num_input_tokens_seen": 26869760, + "step": 410 + }, + { + "epoch": 0.3341216726211443, + "grad_norm": 1.1623753309249878, + "learning_rate": 3.9270455493905284e-05, + "loss": 4.4572, + "num_input_tokens_seen": 27197440, + "step": 415 + }, + { + "epoch": 0.3381472349418809, + "grad_norm": 1.1756770610809326, + "learning_rate": 3.9036002917941324e-05, + "loss": 4.4603, + "num_input_tokens_seen": 27525120, + "step": 420 + }, + { + "epoch": 0.3421727972626176, + "grad_norm": 1.243285059928894, + "learning_rate": 3.8805700005813275e-05, + "loss": 4.4576, + "num_input_tokens_seen": 27852800, + "step": 425 + }, + { + "epoch": 0.3461983595833543, + "grad_norm": 1.4163405895233154, + "learning_rate": 3.8579425773632976e-05, + "loss": 4.4693, + "num_input_tokens_seen": 28180480, + "step": 430 + }, + { + "epoch": 0.35022392190409096, + "grad_norm": 1.2970571517944336, + "learning_rate": 3.835706411883074e-05, + "loss": 4.4731, + "num_input_tokens_seen": 28508160, + "step": 435 + }, + { + "epoch": 0.35424948422482766, + "grad_norm": 1.2691751718521118, + "learning_rate": 3.813850356982369e-05, + "loss": 4.4548, + "num_input_tokens_seen": 28835840, + "step": 440 + }, + { + "epoch": 0.35827504654556436, + "grad_norm": 1.2060904502868652, + "learning_rate": 3.792363705119819e-05, + "loss": 4.5089, + "num_input_tokens_seen": 29163520, + "step": 445 + }, + { + "epoch": 0.362300608866301, + "grad_norm": 1.1918907165527344, + "learning_rate": 3.771236166328254e-05, + "loss": 4.4621, + "num_input_tokens_seen": 29491200, + "step": 450 + }, + { + "epoch": 0.362300608866301, + "eval_accuracy": 0.3575167675445771, + "eval_loss": 4.2810378074646, + "eval_runtime": 6.4142, + "eval_samples_per_second": 46.771, + "eval_steps_per_second": 5.924, + "num_input_tokens_seen": 29491200, + "step": 450 + }, + { + "epoch": 0.3663261711870377, + "grad_norm": 1.257439136505127, + "learning_rate": 3.750457847507964e-05, + "loss": 4.4083, + "num_input_tokens_seen": 29818880, + "step": 455 + }, + { + "epoch": 0.3703517335077744, + "grad_norm": 1.2655524015426636, + "learning_rate": 3.7300192329612555e-05, + "loss": 4.428, + "num_input_tokens_seen": 30146560, + "step": 460 + }, + { + "epoch": 0.37437729582851104, + "grad_norm": 1.3658162355422974, + "learning_rate": 3.709911166081346e-05, + "loss": 4.4843, + "num_input_tokens_seen": 30474240, + "step": 465 + }, + { + "epoch": 0.37840285814924773, + "grad_norm": 1.32870352268219, + "learning_rate": 3.6901248321155405e-05, + "loss": 4.449, + "num_input_tokens_seen": 30801920, + "step": 470 + }, + { + "epoch": 0.3824284204699844, + "grad_norm": 1.3358460664749146, + "learning_rate": 3.670651741928988e-05, + "loss": 4.435, + "num_input_tokens_seen": 31129600, + "step": 475 + }, + { + "epoch": 0.3864539827907211, + "grad_norm": 1.1790028810501099, + "learning_rate": 3.651483716701108e-05, + "loss": 4.426, + "num_input_tokens_seen": 31457280, + "step": 480 + }, + { + "epoch": 0.39047954511145777, + "grad_norm": 1.2213348150253296, + "learning_rate": 3.6326128734919986e-05, + "loss": 4.4451, + "num_input_tokens_seen": 31784960, + "step": 485 + }, + { + "epoch": 0.3945051074321944, + "grad_norm": 1.4075285196304321, + "learning_rate": 3.6140316116210055e-05, + "loss": 4.4582, + "num_input_tokens_seen": 32112640, + "step": 490 + }, + { + "epoch": 0.3985306697529311, + "grad_norm": 1.3073724508285522, + "learning_rate": 3.595732599803958e-05, + "loss": 4.4541, + "num_input_tokens_seen": 32440320, + "step": 495 + }, + { + "epoch": 0.4025562320736678, + "grad_norm": 1.3178701400756836, + "learning_rate": 3.577708763999664e-05, + "loss": 4.4466, + "num_input_tokens_seen": 32768000, + "step": 500 + }, + { + "epoch": 0.40658179439440445, + "grad_norm": 1.2765254974365234, + "learning_rate": 3.5599532759198786e-05, + "loss": 4.4159, + "num_input_tokens_seen": 33095680, + "step": 505 + }, + { + "epoch": 0.41060735671514115, + "grad_norm": 1.2380224466323853, + "learning_rate": 3.542459542160382e-05, + "loss": 4.4387, + "num_input_tokens_seen": 33423360, + "step": 510 + }, + { + "epoch": 0.41463291903587784, + "grad_norm": 1.4157094955444336, + "learning_rate": 3.525221193913854e-05, + "loss": 4.4529, + "num_input_tokens_seen": 33751040, + "step": 515 + }, + { + "epoch": 0.4186584813566145, + "grad_norm": 1.3274729251861572, + "learning_rate": 3.5082320772281174e-05, + "loss": 4.3776, + "num_input_tokens_seen": 34078720, + "step": 520 + }, + { + "epoch": 0.4226840436773512, + "grad_norm": 1.2400481700897217, + "learning_rate": 3.4914862437758787e-05, + "loss": 4.4103, + "num_input_tokens_seen": 34406400, + "step": 525 + }, + { + "epoch": 0.4267096059980879, + "grad_norm": 1.2797058820724487, + "learning_rate": 3.4749779421045556e-05, + "loss": 4.4005, + "num_input_tokens_seen": 34734080, + "step": 530 + }, + { + "epoch": 0.4307351683188245, + "grad_norm": 1.4183690547943115, + "learning_rate": 3.458701609336936e-05, + "loss": 4.4181, + "num_input_tokens_seen": 35061760, + "step": 535 + }, + { + "epoch": 0.4347607306395612, + "grad_norm": 1.2432236671447754, + "learning_rate": 3.442651863295482e-05, + "loss": 4.4037, + "num_input_tokens_seen": 35389440, + "step": 540 + }, + { + "epoch": 0.4387862929602979, + "grad_norm": 1.19386887550354, + "learning_rate": 3.426823495024955e-05, + "loss": 4.4211, + "num_input_tokens_seen": 35717120, + "step": 545 + }, + { + "epoch": 0.44281185528103456, + "grad_norm": 1.24849534034729, + "learning_rate": 3.4112114616897666e-05, + "loss": 4.3953, + "num_input_tokens_seen": 36044800, + "step": 550 + }, + { + "epoch": 0.44683741760177126, + "grad_norm": 1.230302333831787, + "learning_rate": 3.395810879824072e-05, + "loss": 4.4187, + "num_input_tokens_seen": 36372480, + "step": 555 + }, + { + "epoch": 0.4508629799225079, + "grad_norm": 1.293820858001709, + "learning_rate": 3.380617018914067e-05, + "loss": 4.3944, + "num_input_tokens_seen": 36700160, + "step": 560 + }, + { + "epoch": 0.4548885422432446, + "grad_norm": 1.3907512426376343, + "learning_rate": 3.36562529529337e-05, + "loss": 4.4285, + "num_input_tokens_seen": 37027840, + "step": 565 + }, + { + "epoch": 0.4589141045639813, + "grad_norm": 1.4436863660812378, + "learning_rate": 3.350831266333564e-05, + "loss": 4.4322, + "num_input_tokens_seen": 37355520, + "step": 570 + }, + { + "epoch": 0.46293966688471794, + "grad_norm": 1.3175897598266602, + "learning_rate": 3.336230624913197e-05, + "loss": 4.4196, + "num_input_tokens_seen": 37683200, + "step": 575 + }, + { + "epoch": 0.46696522920545463, + "grad_norm": 1.1326998472213745, + "learning_rate": 3.321819194149599e-05, + "loss": 4.4478, + "num_input_tokens_seen": 38010880, + "step": 580 + }, + { + "epoch": 0.47099079152619133, + "grad_norm": 1.3267349004745483, + "learning_rate": 3.307592922378893e-05, + "loss": 4.4068, + "num_input_tokens_seen": 38338560, + "step": 585 + }, + { + "epoch": 0.475016353846928, + "grad_norm": 1.2397940158843994, + "learning_rate": 3.2935478783704735e-05, + "loss": 4.3565, + "num_input_tokens_seen": 38666240, + "step": 590 + }, + { + "epoch": 0.47904191616766467, + "grad_norm": 1.290412187576294, + "learning_rate": 3.279680246763151e-05, + "loss": 4.3957, + "num_input_tokens_seen": 38993920, + "step": 595 + }, + { + "epoch": 0.48306747848840137, + "grad_norm": 1.3627963066101074, + "learning_rate": 3.2659863237109046e-05, + "loss": 4.4116, + "num_input_tokens_seen": 39321600, + "step": 600 + }, + { + "epoch": 0.48306747848840137, + "eval_accuracy": 0.36151503242287286, + "eval_loss": 4.246631622314453, + "eval_runtime": 6.3292, + "eval_samples_per_second": 47.4, + "eval_steps_per_second": 6.004, + "num_input_tokens_seen": 39321600, + "step": 600 + }, + { + "epoch": 0.487093040809138, + "grad_norm": 1.2708163261413574, + "learning_rate": 3.2524625127269674e-05, + "loss": 4.3632, + "num_input_tokens_seen": 39649280, + "step": 605 + }, + { + "epoch": 0.4911186031298747, + "grad_norm": 1.1186234951019287, + "learning_rate": 3.2391053207156646e-05, + "loss": 4.3974, + "num_input_tokens_seen": 39976960, + "step": 610 + }, + { + "epoch": 0.4951441654506114, + "grad_norm": 1.2380205392837524, + "learning_rate": 3.2259113541820474e-05, + "loss": 4.3891, + "num_input_tokens_seen": 40304640, + "step": 615 + }, + { + "epoch": 0.49916972777134805, + "grad_norm": 1.220374584197998, + "learning_rate": 3.212877315609996e-05, + "loss": 4.3911, + "num_input_tokens_seen": 40632320, + "step": 620 + }, + { + "epoch": 0.5031952900920847, + "grad_norm": 1.273556113243103, + "learning_rate": 3.2000000000000005e-05, + "loss": 4.3715, + "num_input_tokens_seen": 40960000, + "step": 625 + }, + { + "epoch": 0.5072208524128214, + "grad_norm": 1.2880346775054932, + "learning_rate": 3.187276291558383e-05, + "loss": 4.4068, + "num_input_tokens_seen": 41287680, + "step": 630 + }, + { + "epoch": 0.5112464147335581, + "grad_norm": 1.3127074241638184, + "learning_rate": 3.1747031605301803e-05, + "loss": 4.3636, + "num_input_tokens_seen": 41615360, + "step": 635 + }, + { + "epoch": 0.5152719770542947, + "grad_norm": 1.2499847412109375, + "learning_rate": 3.1622776601683795e-05, + "loss": 4.384, + "num_input_tokens_seen": 41943040, + "step": 640 + }, + { + "epoch": 0.5192975393750314, + "grad_norm": 1.28691828250885, + "learning_rate": 3.1499969238326315e-05, + "loss": 4.3927, + "num_input_tokens_seen": 42270720, + "step": 645 + }, + { + "epoch": 0.5233231016957681, + "grad_norm": 1.221285104751587, + "learning_rate": 3.137858162210945e-05, + "loss": 4.3894, + "num_input_tokens_seen": 42598400, + "step": 650 + }, + { + "epoch": 0.5273486640165048, + "grad_norm": 1.2766237258911133, + "learning_rate": 3.125858660658254e-05, + "loss": 4.3689, + "num_input_tokens_seen": 42926080, + "step": 655 + }, + { + "epoch": 0.5313742263372415, + "grad_norm": 1.3662595748901367, + "learning_rate": 3.113995776646092e-05, + "loss": 4.3842, + "num_input_tokens_seen": 43253760, + "step": 660 + }, + { + "epoch": 0.5353997886579782, + "grad_norm": 1.264195203781128, + "learning_rate": 3.102266937317925e-05, + "loss": 4.3665, + "num_input_tokens_seen": 43581440, + "step": 665 + }, + { + "epoch": 0.5394253509787148, + "grad_norm": 1.4342658519744873, + "learning_rate": 3.0906696371450235e-05, + "loss": 4.3987, + "num_input_tokens_seen": 43909120, + "step": 670 + }, + { + "epoch": 0.5434509132994515, + "grad_norm": 1.2556202411651611, + "learning_rate": 3.0792014356780046e-05, + "loss": 4.3837, + "num_input_tokens_seen": 44236800, + "step": 675 + }, + { + "epoch": 0.5474764756201882, + "grad_norm": 1.3614763021469116, + "learning_rate": 3.067859955389482e-05, + "loss": 4.339, + "num_input_tokens_seen": 44564480, + "step": 680 + }, + { + "epoch": 0.5515020379409249, + "grad_norm": 1.3664342164993286, + "learning_rate": 3.0566428796034804e-05, + "loss": 4.3543, + "num_input_tokens_seen": 44892160, + "step": 685 + }, + { + "epoch": 0.5555276002616616, + "grad_norm": 1.2612766027450562, + "learning_rate": 3.045547950507524e-05, + "loss": 4.3698, + "num_input_tokens_seen": 45219840, + "step": 690 + }, + { + "epoch": 0.5595531625823982, + "grad_norm": 1.3563127517700195, + "learning_rate": 3.0345729672435268e-05, + "loss": 4.3561, + "num_input_tokens_seen": 45547520, + "step": 695 + }, + { + "epoch": 0.5635787249031349, + "grad_norm": 1.15888249874115, + "learning_rate": 3.023715784073818e-05, + "loss": 4.4154, + "num_input_tokens_seen": 45875200, + "step": 700 + }, + { + "epoch": 0.5676042872238716, + "grad_norm": 1.2462992668151855, + "learning_rate": 3.0129743086188382e-05, + "loss": 4.4011, + "num_input_tokens_seen": 46202880, + "step": 705 + }, + { + "epoch": 0.5716298495446083, + "grad_norm": 1.2652802467346191, + "learning_rate": 3.0023465001632065e-05, + "loss": 4.4053, + "num_input_tokens_seen": 46530560, + "step": 710 + }, + { + "epoch": 0.575655411865345, + "grad_norm": 1.4011398553848267, + "learning_rate": 2.991830368027063e-05, + "loss": 4.3539, + "num_input_tokens_seen": 46858240, + "step": 715 + }, + { + "epoch": 0.5796809741860817, + "grad_norm": 1.248789668083191, + "learning_rate": 2.98142396999972e-05, + "loss": 4.3758, + "num_input_tokens_seen": 47185920, + "step": 720 + }, + { + "epoch": 0.5837065365068183, + "grad_norm": 1.218556523323059, + "learning_rate": 2.97112541083283e-05, + "loss": 4.3695, + "num_input_tokens_seen": 47513600, + "step": 725 + }, + { + "epoch": 0.587732098827555, + "grad_norm": 1.28030526638031, + "learning_rate": 2.960932840790421e-05, + "loss": 4.3676, + "num_input_tokens_seen": 47841280, + "step": 730 + }, + { + "epoch": 0.5917576611482916, + "grad_norm": 1.1834255456924438, + "learning_rate": 2.95084445425327e-05, + "loss": 4.3445, + "num_input_tokens_seen": 48168960, + "step": 735 + }, + { + "epoch": 0.5957832234690283, + "grad_norm": 1.324885606765747, + "learning_rate": 2.9408584883752315e-05, + "loss": 4.3495, + "num_input_tokens_seen": 48496640, + "step": 740 + }, + { + "epoch": 0.599808785789765, + "grad_norm": 1.3806536197662354, + "learning_rate": 2.930973221789247e-05, + "loss": 4.3369, + "num_input_tokens_seen": 48824320, + "step": 745 + }, + { + "epoch": 0.6038343481105017, + "grad_norm": 1.239935278892517, + "learning_rate": 2.9211869733608858e-05, + "loss": 4.3487, + "num_input_tokens_seen": 49152000, + "step": 750 + }, + { + "epoch": 0.6038343481105017, + "eval_accuracy": 0.3661367484834038, + "eval_loss": 4.179458141326904, + "eval_runtime": 6.3307, + "eval_samples_per_second": 47.388, + "eval_steps_per_second": 6.003, + "num_input_tokens_seen": 49152000, + "step": 750 + }, + { + "epoch": 0.6078599104312383, + "grad_norm": 1.294979214668274, + "learning_rate": 2.911498100987356e-05, + "loss": 4.3559, + "num_input_tokens_seen": 49479680, + "step": 755 + }, + { + "epoch": 0.611885472751975, + "grad_norm": 1.2415525913238525, + "learning_rate": 2.901905000440047e-05, + "loss": 4.3862, + "num_input_tokens_seen": 49807360, + "step": 760 + }, + { + "epoch": 0.6159110350727117, + "grad_norm": 1.2934507131576538, + "learning_rate": 2.8924061042487487e-05, + "loss": 4.3778, + "num_input_tokens_seen": 50135040, + "step": 765 + }, + { + "epoch": 0.6199365973934484, + "grad_norm": 1.3969470262527466, + "learning_rate": 2.8829998806257887e-05, + "loss": 4.3623, + "num_input_tokens_seen": 50462720, + "step": 770 + }, + { + "epoch": 0.6239621597141851, + "grad_norm": 1.3450722694396973, + "learning_rate": 2.873684832428399e-05, + "loss": 4.3819, + "num_input_tokens_seen": 50790400, + "step": 775 + }, + { + "epoch": 0.6279877220349217, + "grad_norm": 1.4024951457977295, + "learning_rate": 2.8644594961577318e-05, + "loss": 4.3577, + "num_input_tokens_seen": 51118080, + "step": 780 + }, + { + "epoch": 0.6320132843556584, + "grad_norm": 1.4493414163589478, + "learning_rate": 2.855322440992999e-05, + "loss": 4.3402, + "num_input_tokens_seen": 51445760, + "step": 785 + }, + { + "epoch": 0.6360388466763951, + "grad_norm": 1.12810480594635, + "learning_rate": 2.8462722678592803e-05, + "loss": 4.3095, + "num_input_tokens_seen": 51773440, + "step": 790 + }, + { + "epoch": 0.6400644089971318, + "grad_norm": 1.2812013626098633, + "learning_rate": 2.837307608527635e-05, + "loss": 4.3664, + "num_input_tokens_seen": 52101120, + "step": 795 + }, + { + "epoch": 0.6440899713178685, + "grad_norm": 1.269395112991333, + "learning_rate": 2.8284271247461902e-05, + "loss": 4.3699, + "num_input_tokens_seen": 52428800, + "step": 800 + }, + { + "epoch": 0.6481155336386052, + "grad_norm": 1.2164851427078247, + "learning_rate": 2.819629507400966e-05, + "loss": 4.3331, + "num_input_tokens_seen": 52756480, + "step": 805 + }, + { + "epoch": 0.6521410959593418, + "grad_norm": 1.249258279800415, + "learning_rate": 2.8109134757052262e-05, + "loss": 4.3384, + "num_input_tokens_seen": 53084160, + "step": 810 + }, + { + "epoch": 0.6561666582800785, + "grad_norm": 1.2432383298873901, + "learning_rate": 2.8022777764162353e-05, + "loss": 4.3628, + "num_input_tokens_seen": 53411840, + "step": 815 + }, + { + "epoch": 0.6601922206008152, + "grad_norm": 1.2008548974990845, + "learning_rate": 2.7937211830783133e-05, + "loss": 4.3751, + "num_input_tokens_seen": 53739520, + "step": 820 + }, + { + "epoch": 0.6642177829215519, + "grad_norm": 1.3261739015579224, + "learning_rate": 2.785242495291166e-05, + "loss": 4.3405, + "num_input_tokens_seen": 54067200, + "step": 825 + }, + { + "epoch": 0.6682433452422886, + "grad_norm": 1.3359473943710327, + "learning_rate": 2.7768405380024934e-05, + "loss": 4.3384, + "num_input_tokens_seen": 54394880, + "step": 830 + }, + { + "epoch": 0.6722689075630253, + "grad_norm": 1.3322173357009888, + "learning_rate": 2.7685141608239315e-05, + "loss": 4.3532, + "num_input_tokens_seen": 54722560, + "step": 835 + }, + { + "epoch": 0.6762944698837619, + "grad_norm": 1.2270052433013916, + "learning_rate": 2.7602622373694168e-05, + "loss": 4.3314, + "num_input_tokens_seen": 55050240, + "step": 840 + }, + { + "epoch": 0.6803200322044985, + "grad_norm": 1.2989898920059204, + "learning_rate": 2.752083664615126e-05, + "loss": 4.3486, + "num_input_tokens_seen": 55377920, + "step": 845 + }, + { + "epoch": 0.6843455945252352, + "grad_norm": 1.2281562089920044, + "learning_rate": 2.7439773622801412e-05, + "loss": 4.3269, + "num_input_tokens_seen": 55705600, + "step": 850 + }, + { + "epoch": 0.6883711568459719, + "grad_norm": 1.3172951936721802, + "learning_rate": 2.7359422722270782e-05, + "loss": 4.3333, + "num_input_tokens_seen": 56033280, + "step": 855 + }, + { + "epoch": 0.6923967191667086, + "grad_norm": 1.3179869651794434, + "learning_rate": 2.7279773578818943e-05, + "loss": 4.3657, + "num_input_tokens_seen": 56360960, + "step": 860 + }, + { + "epoch": 0.6964222814874452, + "grad_norm": 1.277402400970459, + "learning_rate": 2.720081603672184e-05, + "loss": 4.3581, + "num_input_tokens_seen": 56688640, + "step": 865 + }, + { + "epoch": 0.7004478438081819, + "grad_norm": 1.2060281038284302, + "learning_rate": 2.712254014483242e-05, + "loss": 4.372, + "num_input_tokens_seen": 57016320, + "step": 870 + }, + { + "epoch": 0.7044734061289186, + "grad_norm": 1.3300620317459106, + "learning_rate": 2.704493615131253e-05, + "loss": 4.3463, + "num_input_tokens_seen": 57344000, + "step": 875 + }, + { + "epoch": 0.7084989684496553, + "grad_norm": 1.2079811096191406, + "learning_rate": 2.6967994498529685e-05, + "loss": 4.3365, + "num_input_tokens_seen": 57671680, + "step": 880 + }, + { + "epoch": 0.712524530770392, + "grad_norm": 1.2995448112487793, + "learning_rate": 2.6891705818112575e-05, + "loss": 4.3474, + "num_input_tokens_seen": 57999360, + "step": 885 + }, + { + "epoch": 0.7165500930911287, + "grad_norm": 1.3099631071090698, + "learning_rate": 2.681606092615964e-05, + "loss": 4.3505, + "num_input_tokens_seen": 58327040, + "step": 890 + }, + { + "epoch": 0.7205756554118653, + "grad_norm": 1.1766512393951416, + "learning_rate": 2.6741050818595026e-05, + "loss": 4.3314, + "num_input_tokens_seen": 58654720, + "step": 895 + }, + { + "epoch": 0.724601217732602, + "grad_norm": 1.2225860357284546, + "learning_rate": 2.6666666666666667e-05, + "loss": 4.338, + "num_input_tokens_seen": 58982400, + "step": 900 + }, + { + "epoch": 0.724601217732602, + "eval_accuracy": 0.3663363114127162, + "eval_loss": 4.187371730804443, + "eval_runtime": 6.3051, + "eval_samples_per_second": 47.581, + "eval_steps_per_second": 6.027, + "num_input_tokens_seen": 58982400, + "step": 900 + }, + { + "epoch": 0.7286267800533387, + "grad_norm": 1.1610972881317139, + "learning_rate": 2.6592899812581347e-05, + "loss": 4.3445, + "num_input_tokens_seen": 59310080, + "step": 905 + }, + { + "epoch": 0.7326523423740754, + "grad_norm": 1.3581031560897827, + "learning_rate": 2.6519741765271837e-05, + "loss": 4.3591, + "num_input_tokens_seen": 59637760, + "step": 910 + }, + { + "epoch": 0.7366779046948121, + "grad_norm": 1.255743145942688, + "learning_rate": 2.6447184196291452e-05, + "loss": 4.342, + "num_input_tokens_seen": 59965440, + "step": 915 + }, + { + "epoch": 0.7407034670155488, + "grad_norm": 1.4238628149032593, + "learning_rate": 2.6375218935831487e-05, + "loss": 4.3019, + "num_input_tokens_seen": 60293120, + "step": 920 + }, + { + "epoch": 0.7447290293362854, + "grad_norm": 1.2544608116149902, + "learning_rate": 2.6303837968857168e-05, + "loss": 4.3089, + "num_input_tokens_seen": 60620800, + "step": 925 + }, + { + "epoch": 0.7487545916570221, + "grad_norm": 1.1759463548660278, + "learning_rate": 2.623303343135812e-05, + "loss": 4.3121, + "num_input_tokens_seen": 60948480, + "step": 930 + }, + { + "epoch": 0.7527801539777588, + "grad_norm": 1.3340330123901367, + "learning_rate": 2.616279760670915e-05, + "loss": 4.3074, + "num_input_tokens_seen": 61276160, + "step": 935 + }, + { + "epoch": 0.7568057162984955, + "grad_norm": 1.2744321823120117, + "learning_rate": 2.609312292213769e-05, + "loss": 4.3436, + "num_input_tokens_seen": 61603840, + "step": 940 + }, + { + "epoch": 0.7608312786192322, + "grad_norm": 1.1901994943618774, + "learning_rate": 2.602400194529422e-05, + "loss": 4.2908, + "num_input_tokens_seen": 61931520, + "step": 945 + }, + { + "epoch": 0.7648568409399688, + "grad_norm": 1.202046513557434, + "learning_rate": 2.595542738092201e-05, + "loss": 4.3347, + "num_input_tokens_seen": 62259200, + "step": 950 + }, + { + "epoch": 0.7688824032607054, + "grad_norm": 1.2437232732772827, + "learning_rate": 2.5887392067622943e-05, + "loss": 4.3434, + "num_input_tokens_seen": 62586880, + "step": 955 + }, + { + "epoch": 0.7729079655814421, + "grad_norm": 1.2383480072021484, + "learning_rate": 2.5819888974716114e-05, + "loss": 4.3144, + "num_input_tokens_seen": 62914560, + "step": 960 + }, + { + "epoch": 0.7769335279021788, + "grad_norm": 1.2294889688491821, + "learning_rate": 2.575291119918602e-05, + "loss": 4.2971, + "num_input_tokens_seen": 63242240, + "step": 965 + }, + { + "epoch": 0.7809590902229155, + "grad_norm": 1.235829472541809, + "learning_rate": 2.5686451962717426e-05, + "loss": 4.3006, + "num_input_tokens_seen": 63569920, + "step": 970 + }, + { + "epoch": 0.7849846525436522, + "grad_norm": 1.2483339309692383, + "learning_rate": 2.5620504608813944e-05, + "loss": 4.2846, + "num_input_tokens_seen": 63897600, + "step": 975 + }, + { + "epoch": 0.7890102148643888, + "grad_norm": 1.2668383121490479, + "learning_rate": 2.55550625999976e-05, + "loss": 4.3256, + "num_input_tokens_seen": 64225280, + "step": 980 + }, + { + "epoch": 0.7930357771851255, + "grad_norm": 1.2814356088638306, + "learning_rate": 2.549011951508665e-05, + "loss": 4.2869, + "num_input_tokens_seen": 64552960, + "step": 985 + }, + { + "epoch": 0.7970613395058622, + "grad_norm": 1.3956332206726074, + "learning_rate": 2.5425669046549127e-05, + "loss": 4.3438, + "num_input_tokens_seen": 64880640, + "step": 990 + }, + { + "epoch": 0.8010869018265989, + "grad_norm": 1.3514597415924072, + "learning_rate": 2.536170499792966e-05, + "loss": 4.3132, + "num_input_tokens_seen": 65208320, + "step": 995 + }, + { + "epoch": 0.8051124641473356, + "grad_norm": 1.198144793510437, + "learning_rate": 2.5298221281347037e-05, + "loss": 4.291, + "num_input_tokens_seen": 65536000, + "step": 1000 + }, + { + "epoch": 0.8091380264680723, + "grad_norm": 1.2150394916534424, + "learning_rate": 2.523521191506047e-05, + "loss": 4.3226, + "num_input_tokens_seen": 65863680, + "step": 1005 + }, + { + "epoch": 0.8131635887888089, + "grad_norm": 1.3794989585876465, + "learning_rate": 2.5172671021102106e-05, + "loss": 4.331, + "num_input_tokens_seen": 66191360, + "step": 1010 + }, + { + "epoch": 0.8171891511095456, + "grad_norm": 1.2272028923034668, + "learning_rate": 2.511059282297378e-05, + "loss": 4.3232, + "num_input_tokens_seen": 66519040, + "step": 1015 + }, + { + "epoch": 0.8212147134302823, + "grad_norm": 1.275801420211792, + "learning_rate": 2.5048971643405984e-05, + "loss": 4.3274, + "num_input_tokens_seen": 66846720, + "step": 1020 + }, + { + "epoch": 0.825240275751019, + "grad_norm": 1.3816701173782349, + "learning_rate": 2.4987801902176975e-05, + "loss": 4.3331, + "num_input_tokens_seen": 67174400, + "step": 1025 + }, + { + "epoch": 0.8292658380717557, + "grad_norm": 1.2978571653366089, + "learning_rate": 2.4927078113990235e-05, + "loss": 4.2978, + "num_input_tokens_seen": 67502080, + "step": 1030 + }, + { + "epoch": 0.8332914003924923, + "grad_norm": 1.3196457624435425, + "learning_rate": 2.486679488640837e-05, + "loss": 4.2904, + "num_input_tokens_seen": 67829760, + "step": 1035 + }, + { + "epoch": 0.837316962713229, + "grad_norm": 1.3695876598358154, + "learning_rate": 2.480694691784169e-05, + "loss": 4.3261, + "num_input_tokens_seen": 68157440, + "step": 1040 + }, + { + "epoch": 0.8413425250339657, + "grad_norm": 1.3481625318527222, + "learning_rate": 2.474752899558984e-05, + "loss": 4.2861, + "num_input_tokens_seen": 68485120, + "step": 1045 + }, + { + "epoch": 0.8453680873547024, + "grad_norm": 1.1765189170837402, + "learning_rate": 2.4688535993934706e-05, + "loss": 4.342, + "num_input_tokens_seen": 68812800, + "step": 1050 + }, + { + "epoch": 0.8453680873547024, + "eval_accuracy": 0.3695067117398011, + "eval_loss": 4.1475090980529785, + "eval_runtime": 6.6416, + "eval_samples_per_second": 45.17, + "eval_steps_per_second": 5.722, + "num_input_tokens_seen": 68812800, + "step": 1050 + }, + { + "epoch": 0.8493936496754391, + "grad_norm": 1.3343596458435059, + "learning_rate": 2.4629962872283168e-05, + "loss": 4.2779, + "num_input_tokens_seen": 69140480, + "step": 1055 + }, + { + "epoch": 0.8534192119961758, + "grad_norm": 1.31355619430542, + "learning_rate": 2.4571804673358052e-05, + "loss": 4.2672, + "num_input_tokens_seen": 69468160, + "step": 1060 + }, + { + "epoch": 0.8574447743169124, + "grad_norm": 1.2236064672470093, + "learning_rate": 2.4514056521435823e-05, + "loss": 4.3232, + "num_input_tokens_seen": 69795840, + "step": 1065 + }, + { + "epoch": 0.861470336637649, + "grad_norm": 1.2435745000839233, + "learning_rate": 2.4456713620629726e-05, + "loss": 4.3067, + "num_input_tokens_seen": 70123520, + "step": 1070 + }, + { + "epoch": 0.8654958989583857, + "grad_norm": 1.2577111721038818, + "learning_rate": 2.439977125321675e-05, + "loss": 4.3227, + "num_input_tokens_seen": 70451200, + "step": 1075 + }, + { + "epoch": 0.8695214612791224, + "grad_norm": 1.1730388402938843, + "learning_rate": 2.4343224778007382e-05, + "loss": 4.2894, + "num_input_tokens_seen": 70778880, + "step": 1080 + }, + { + "epoch": 0.8735470235998591, + "grad_norm": 1.2375553846359253, + "learning_rate": 2.4287069628756655e-05, + "loss": 4.3004, + "num_input_tokens_seen": 71106560, + "step": 1085 + }, + { + "epoch": 0.8775725859205958, + "grad_norm": 1.3142080307006836, + "learning_rate": 2.4231301312615306e-05, + "loss": 4.3036, + "num_input_tokens_seen": 71434240, + "step": 1090 + }, + { + "epoch": 0.8815981482413324, + "grad_norm": 1.218998670578003, + "learning_rate": 2.4175915408619977e-05, + "loss": 4.3097, + "num_input_tokens_seen": 71761920, + "step": 1095 + }, + { + "epoch": 0.8856237105620691, + "grad_norm": 1.312283992767334, + "learning_rate": 2.4120907566221092e-05, + "loss": 4.2724, + "num_input_tokens_seen": 72089600, + "step": 1100 + }, + { + "epoch": 0.8896492728828058, + "grad_norm": 1.273376703262329, + "learning_rate": 2.4066273503847476e-05, + "loss": 4.2896, + "num_input_tokens_seen": 72417280, + "step": 1105 + }, + { + "epoch": 0.8936748352035425, + "grad_norm": 1.2696881294250488, + "learning_rate": 2.4012009007506573e-05, + "loss": 4.3627, + "num_input_tokens_seen": 72744960, + "step": 1110 + }, + { + "epoch": 0.8977003975242792, + "grad_norm": 1.3189187049865723, + "learning_rate": 2.3958109929419195e-05, + "loss": 4.2899, + "num_input_tokens_seen": 73072640, + "step": 1115 + }, + { + "epoch": 0.9017259598450158, + "grad_norm": 1.1770976781845093, + "learning_rate": 2.3904572186687872e-05, + "loss": 4.2978, + "num_input_tokens_seen": 73400320, + "step": 1120 + }, + { + "epoch": 0.9057515221657525, + "grad_norm": 1.1813838481903076, + "learning_rate": 2.385139175999776e-05, + "loss": 4.2919, + "num_input_tokens_seen": 73728000, + "step": 1125 + }, + { + "epoch": 0.9097770844864892, + "grad_norm": 1.2608633041381836, + "learning_rate": 2.3798564692349184e-05, + "loss": 4.3118, + "num_input_tokens_seen": 74055680, + "step": 1130 + }, + { + "epoch": 0.9138026468072259, + "grad_norm": 1.3048661947250366, + "learning_rate": 2.3746087087820993e-05, + "loss": 4.286, + "num_input_tokens_seen": 74383360, + "step": 1135 + }, + { + "epoch": 0.9178282091279626, + "grad_norm": 1.1029573678970337, + "learning_rate": 2.3693955110363694e-05, + "loss": 4.2935, + "num_input_tokens_seen": 74711040, + "step": 1140 + }, + { + "epoch": 0.9218537714486993, + "grad_norm": 1.2217323780059814, + "learning_rate": 2.3642164982621672e-05, + "loss": 4.3049, + "num_input_tokens_seen": 75038720, + "step": 1145 + }, + { + "epoch": 0.9258793337694359, + "grad_norm": 1.2029818296432495, + "learning_rate": 2.3590712984783544e-05, + "loss": 4.3105, + "num_input_tokens_seen": 75366400, + "step": 1150 + }, + { + "epoch": 0.9299048960901726, + "grad_norm": 1.3691645860671997, + "learning_rate": 2.353959545345999e-05, + "loss": 4.2643, + "num_input_tokens_seen": 75694080, + "step": 1155 + }, + { + "epoch": 0.9339304584109093, + "grad_norm": 1.441178321838379, + "learning_rate": 2.3488808780588142e-05, + "loss": 4.2654, + "num_input_tokens_seen": 76021760, + "step": 1160 + }, + { + "epoch": 0.937956020731646, + "grad_norm": 1.2421964406967163, + "learning_rate": 2.3438349412361906e-05, + "loss": 4.3056, + "num_input_tokens_seen": 76349440, + "step": 1165 + }, + { + "epoch": 0.9419815830523827, + "grad_norm": 1.349487543106079, + "learning_rate": 2.338821384818745e-05, + "loss": 4.281, + "num_input_tokens_seen": 76677120, + "step": 1170 + }, + { + "epoch": 0.9460071453731194, + "grad_norm": 1.3027498722076416, + "learning_rate": 2.3338398639663132e-05, + "loss": 4.2827, + "num_input_tokens_seen": 77004800, + "step": 1175 + }, + { + "epoch": 0.950032707693856, + "grad_norm": 1.313684344291687, + "learning_rate": 2.3288900389583283e-05, + "loss": 4.3015, + "num_input_tokens_seen": 77332480, + "step": 1180 + }, + { + "epoch": 0.9540582700145926, + "grad_norm": 1.1965482234954834, + "learning_rate": 2.3239715750965074e-05, + "loss": 4.2841, + "num_input_tokens_seen": 77660160, + "step": 1185 + }, + { + "epoch": 0.9580838323353293, + "grad_norm": 1.2446861267089844, + "learning_rate": 2.3190841426097943e-05, + "loss": 4.3269, + "num_input_tokens_seen": 77987840, + "step": 1190 + }, + { + "epoch": 0.962109394656066, + "grad_norm": 1.218593955039978, + "learning_rate": 2.314227416561487e-05, + "loss": 4.3048, + "num_input_tokens_seen": 78315520, + "step": 1195 + }, + { + "epoch": 0.9661349569768027, + "grad_norm": 1.1979824304580688, + "learning_rate": 2.3094010767585035e-05, + "loss": 4.268, + "num_input_tokens_seen": 78643200, + "step": 1200 + }, + { + "epoch": 0.9661349569768027, + "eval_accuracy": 0.37141276761892367, + "eval_loss": 4.121544361114502, + "eval_runtime": 6.4083, + "eval_samples_per_second": 46.814, + "eval_steps_per_second": 5.93, + "num_input_tokens_seen": 78643200, + "step": 1200 + }, + { + "epoch": 0.9701605192975393, + "grad_norm": 1.2462401390075684, + "learning_rate": 2.3046048076627097e-05, + "loss": 4.2736, + "num_input_tokens_seen": 78970880, + "step": 1205 + }, + { + "epoch": 0.974186081618276, + "grad_norm": 1.3557130098342896, + "learning_rate": 2.2998382983042763e-05, + "loss": 4.2592, + "num_input_tokens_seen": 79298560, + "step": 1210 + }, + { + "epoch": 0.9782116439390127, + "grad_norm": 1.342148780822754, + "learning_rate": 2.2951012421969877e-05, + "loss": 4.2947, + "num_input_tokens_seen": 79626240, + "step": 1215 + }, + { + "epoch": 0.9822372062597494, + "grad_norm": 1.269107699394226, + "learning_rate": 2.290393337255473e-05, + "loss": 4.2789, + "num_input_tokens_seen": 79953920, + "step": 1220 + }, + { + "epoch": 0.9862627685804861, + "grad_norm": 1.189437747001648, + "learning_rate": 2.2857142857142858e-05, + "loss": 4.2793, + "num_input_tokens_seen": 80281600, + "step": 1225 + }, + { + "epoch": 0.9902883309012228, + "grad_norm": 1.217842698097229, + "learning_rate": 2.2810637940488043e-05, + "loss": 4.2606, + "num_input_tokens_seen": 80609280, + "step": 1230 + }, + { + "epoch": 0.9943138932219594, + "grad_norm": 1.1911648511886597, + "learning_rate": 2.2764415728978892e-05, + "loss": 4.2786, + "num_input_tokens_seen": 80936960, + "step": 1235 + }, + { + "epoch": 0.9983394555426961, + "grad_norm": 1.2993587255477905, + "learning_rate": 2.2718473369882594e-05, + "loss": 4.2791, + "num_input_tokens_seen": 81264640, + "step": 1240 + }, + { + "epoch": 1.002365017863433, + "grad_norm": 1.2911006212234497, + "learning_rate": 2.2672808050605434e-05, + "loss": 4.2619, + "num_input_tokens_seen": 81591296, + "step": 1245 + }, + { + "epoch": 1.0063905801841695, + "grad_norm": 1.4278241395950317, + "learning_rate": 2.2627416997969522e-05, + "loss": 4.2724, + "num_input_tokens_seen": 81918976, + "step": 1250 + }, + { + "epoch": 1.010416142504906, + "grad_norm": 1.3427342176437378, + "learning_rate": 2.2582297477505467e-05, + "loss": 4.2481, + "num_input_tokens_seen": 82246656, + "step": 1255 + }, + { + "epoch": 1.0144417048256429, + "grad_norm": 1.337169885635376, + "learning_rate": 2.253744679276044e-05, + "loss": 4.291, + "num_input_tokens_seen": 82574336, + "step": 1260 + }, + { + "epoch": 1.0184672671463795, + "grad_norm": 1.2553439140319824, + "learning_rate": 2.249286228462133e-05, + "loss": 4.2706, + "num_input_tokens_seen": 82902016, + "step": 1265 + }, + { + "epoch": 1.0224928294671163, + "grad_norm": 1.2947965860366821, + "learning_rate": 2.244854133065255e-05, + "loss": 4.21, + "num_input_tokens_seen": 83229696, + "step": 1270 + }, + { + "epoch": 1.0265183917878529, + "grad_norm": 1.327486515045166, + "learning_rate": 2.2404481344448157e-05, + "loss": 4.2847, + "num_input_tokens_seen": 83557376, + "step": 1275 + }, + { + "epoch": 1.0305439541085895, + "grad_norm": 1.2696667909622192, + "learning_rate": 2.23606797749979e-05, + "loss": 4.2864, + "num_input_tokens_seen": 83885056, + "step": 1280 + }, + { + "epoch": 1.0345695164293263, + "grad_norm": 1.3726356029510498, + "learning_rate": 2.2317134106066828e-05, + "loss": 4.265, + "num_input_tokens_seen": 84212736, + "step": 1285 + }, + { + "epoch": 1.0385950787500629, + "grad_norm": 1.250604271888733, + "learning_rate": 2.2273841855588186e-05, + "loss": 4.2894, + "num_input_tokens_seen": 84540416, + "step": 1290 + }, + { + "epoch": 1.0426206410707997, + "grad_norm": 1.3346271514892578, + "learning_rate": 2.223080057506914e-05, + "loss": 4.2751, + "num_input_tokens_seen": 84868096, + "step": 1295 + }, + { + "epoch": 1.0466462033915362, + "grad_norm": 1.272316575050354, + "learning_rate": 2.2188007849009167e-05, + "loss": 4.2783, + "num_input_tokens_seen": 85195776, + "step": 1300 + }, + { + "epoch": 1.0506717657122728, + "grad_norm": 1.37924063205719, + "learning_rate": 2.214546129433066e-05, + "loss": 4.2534, + "num_input_tokens_seen": 85523456, + "step": 1305 + }, + { + "epoch": 1.0546973280330096, + "grad_norm": 1.4074842929840088, + "learning_rate": 2.2103158559821507e-05, + "loss": 4.2809, + "num_input_tokens_seen": 85851136, + "step": 1310 + }, + { + "epoch": 1.0587228903537462, + "grad_norm": 1.3106921911239624, + "learning_rate": 2.206109732558935e-05, + "loss": 4.2831, + "num_input_tokens_seen": 86178816, + "step": 1315 + }, + { + "epoch": 1.062748452674483, + "grad_norm": 1.1723382472991943, + "learning_rate": 2.2019275302527215e-05, + "loss": 4.2617, + "num_input_tokens_seen": 86506496, + "step": 1320 + }, + { + "epoch": 1.0667740149952196, + "grad_norm": 1.2937371730804443, + "learning_rate": 2.1977690231790248e-05, + "loss": 4.2897, + "num_input_tokens_seen": 86834176, + "step": 1325 + }, + { + "epoch": 1.0707995773159564, + "grad_norm": 1.2420538663864136, + "learning_rate": 2.193633988428327e-05, + "loss": 4.2304, + "num_input_tokens_seen": 87161856, + "step": 1330 + }, + { + "epoch": 1.074825139636693, + "grad_norm": 1.2572234869003296, + "learning_rate": 2.1895222060158936e-05, + "loss": 4.2599, + "num_input_tokens_seen": 87489536, + "step": 1335 + }, + { + "epoch": 1.0788507019574296, + "grad_norm": 1.302157998085022, + "learning_rate": 2.1854334588326122e-05, + "loss": 4.2394, + "num_input_tokens_seen": 87817216, + "step": 1340 + }, + { + "epoch": 1.0828762642781664, + "grad_norm": 1.2681870460510254, + "learning_rate": 2.1813675325968476e-05, + "loss": 4.3021, + "num_input_tokens_seen": 88144896, + "step": 1345 + }, + { + "epoch": 1.086901826598903, + "grad_norm": 1.1405740976333618, + "learning_rate": 2.1773242158072697e-05, + "loss": 4.2185, + "num_input_tokens_seen": 88472576, + "step": 1350 + }, + { + "epoch": 1.086901826598903, + "eval_accuracy": 0.37249442300490127, + "eval_loss": 4.103240489959717, + "eval_runtime": 6.6226, + "eval_samples_per_second": 45.3, + "eval_steps_per_second": 5.738, + "num_input_tokens_seen": 88472576, + "step": 1350 + }, + { + "epoch": 1.0909273889196398, + "grad_norm": 1.144243836402893, + "learning_rate": 2.1733032996966454e-05, + "loss": 4.2665, + "num_input_tokens_seen": 88800256, + "step": 1355 + }, + { + "epoch": 1.0949529512403764, + "grad_norm": 1.158898115158081, + "learning_rate": 2.1693045781865617e-05, + "loss": 4.2608, + "num_input_tokens_seen": 89127936, + "step": 1360 + }, + { + "epoch": 1.098978513561113, + "grad_norm": 1.2325701713562012, + "learning_rate": 2.165327847843067e-05, + "loss": 4.2529, + "num_input_tokens_seen": 89455616, + "step": 1365 + }, + { + "epoch": 1.1030040758818498, + "grad_norm": 1.3270164728164673, + "learning_rate": 2.161372907833197e-05, + "loss": 4.2576, + "num_input_tokens_seen": 89783296, + "step": 1370 + }, + { + "epoch": 1.1070296382025864, + "grad_norm": 1.2680999040603638, + "learning_rate": 2.157439559882375e-05, + "loss": 4.2411, + "num_input_tokens_seen": 90110976, + "step": 1375 + }, + { + "epoch": 1.1110552005233232, + "grad_norm": 1.37815260887146, + "learning_rate": 2.1535276082326623e-05, + "loss": 4.3014, + "num_input_tokens_seen": 90438656, + "step": 1380 + }, + { + "epoch": 1.1150807628440598, + "grad_norm": 1.165544867515564, + "learning_rate": 2.149636859601836e-05, + "loss": 4.2405, + "num_input_tokens_seen": 90766336, + "step": 1385 + }, + { + "epoch": 1.1191063251647964, + "grad_norm": 1.184525966644287, + "learning_rate": 2.1457671231432803e-05, + "loss": 4.2646, + "num_input_tokens_seen": 91094016, + "step": 1390 + }, + { + "epoch": 1.1231318874855332, + "grad_norm": 1.216493010520935, + "learning_rate": 2.1419182104066638e-05, + "loss": 4.2586, + "num_input_tokens_seen": 91421696, + "step": 1395 + }, + { + "epoch": 1.1271574498062698, + "grad_norm": 1.4095522165298462, + "learning_rate": 2.1380899352993955e-05, + "loss": 4.2576, + "num_input_tokens_seen": 91749376, + "step": 1400 + }, + { + "epoch": 1.1311830121270066, + "grad_norm": 1.1573916673660278, + "learning_rate": 2.134282114048833e-05, + "loss": 4.2463, + "num_input_tokens_seen": 92077056, + "step": 1405 + }, + { + "epoch": 1.1352085744477431, + "grad_norm": 1.213179111480713, + "learning_rate": 2.13049456516523e-05, + "loss": 4.2402, + "num_input_tokens_seen": 92404736, + "step": 1410 + }, + { + "epoch": 1.1392341367684797, + "grad_norm": 1.197209119796753, + "learning_rate": 2.1267271094054026e-05, + "loss": 4.237, + "num_input_tokens_seen": 92732416, + "step": 1415 + }, + { + "epoch": 1.1432596990892165, + "grad_norm": 1.2352854013442993, + "learning_rate": 2.1229795697371012e-05, + "loss": 4.248, + "num_input_tokens_seen": 93060096, + "step": 1420 + }, + { + "epoch": 1.1472852614099531, + "grad_norm": 1.196413516998291, + "learning_rate": 2.1192517713040704e-05, + "loss": 4.272, + "num_input_tokens_seen": 93387776, + "step": 1425 + }, + { + "epoch": 1.15131082373069, + "grad_norm": 1.2881050109863281, + "learning_rate": 2.11554354139178e-05, + "loss": 4.258, + "num_input_tokens_seen": 93715456, + "step": 1430 + }, + { + "epoch": 1.1553363860514265, + "grad_norm": 1.2045120000839233, + "learning_rate": 2.1118547093938186e-05, + "loss": 4.2621, + "num_input_tokens_seen": 94043136, + "step": 1435 + }, + { + "epoch": 1.1593619483721633, + "grad_norm": 1.1793156862258911, + "learning_rate": 2.1081851067789197e-05, + "loss": 4.2386, + "num_input_tokens_seen": 94370816, + "step": 1440 + }, + { + "epoch": 1.1633875106929, + "grad_norm": 1.3340944051742554, + "learning_rate": 2.1045345670586257e-05, + "loss": 4.2837, + "num_input_tokens_seen": 94698496, + "step": 1445 + }, + { + "epoch": 1.1674130730136365, + "grad_norm": 1.2929213047027588, + "learning_rate": 2.100902925755561e-05, + "loss": 4.2334, + "num_input_tokens_seen": 95026176, + "step": 1450 + }, + { + "epoch": 1.1714386353343733, + "grad_norm": 1.3204522132873535, + "learning_rate": 2.0972900203723056e-05, + "loss": 4.2522, + "num_input_tokens_seen": 95353856, + "step": 1455 + }, + { + "epoch": 1.17546419765511, + "grad_norm": 1.305235743522644, + "learning_rate": 2.0936956903608547e-05, + "loss": 4.2534, + "num_input_tokens_seen": 95681536, + "step": 1460 + }, + { + "epoch": 1.1794897599758467, + "grad_norm": 1.2224024534225464, + "learning_rate": 2.0901197770926567e-05, + "loss": 4.2439, + "num_input_tokens_seen": 96009216, + "step": 1465 + }, + { + "epoch": 1.1835153222965833, + "grad_norm": 1.249813199043274, + "learning_rate": 2.0865621238292045e-05, + "loss": 4.2009, + "num_input_tokens_seen": 96336896, + "step": 1470 + }, + { + "epoch": 1.18754088461732, + "grad_norm": 1.2258191108703613, + "learning_rate": 2.0830225756931824e-05, + "loss": 4.2517, + "num_input_tokens_seen": 96664576, + "step": 1475 + }, + { + "epoch": 1.1915664469380567, + "grad_norm": 1.2485374212265015, + "learning_rate": 2.0795009796401456e-05, + "loss": 4.2254, + "num_input_tokens_seen": 96992256, + "step": 1480 + }, + { + "epoch": 1.1955920092587933, + "grad_norm": 1.2434406280517578, + "learning_rate": 2.0759971844307282e-05, + "loss": 4.208, + "num_input_tokens_seen": 97319936, + "step": 1485 + }, + { + "epoch": 1.19961757157953, + "grad_norm": 1.212769865989685, + "learning_rate": 2.072511040603359e-05, + "loss": 4.2253, + "num_input_tokens_seen": 97647616, + "step": 1490 + }, + { + "epoch": 1.2036431339002667, + "grad_norm": 1.1651945114135742, + "learning_rate": 2.0690424004474868e-05, + "loss": 4.2694, + "num_input_tokens_seen": 97975296, + "step": 1495 + }, + { + "epoch": 1.2076686962210035, + "grad_norm": 1.2737419605255127, + "learning_rate": 2.065591117977289e-05, + "loss": 4.2645, + "num_input_tokens_seen": 98302976, + "step": 1500 + }, + { + "epoch": 1.2076686962210035, + "eval_accuracy": 0.37566890863567187, + "eval_loss": 4.08589506149292, + "eval_runtime": 6.6303, + "eval_samples_per_second": 45.247, + "eval_steps_per_second": 5.731, + "num_input_tokens_seen": 98302976, + "step": 1500 + }, + { + "epoch": 1.21169425854174, + "grad_norm": 1.2633168697357178, + "learning_rate": 2.0621570489058684e-05, + "loss": 4.2746, + "num_input_tokens_seen": 98630656, + "step": 1505 + }, + { + "epoch": 1.2157198208624767, + "grad_norm": 1.2094192504882812, + "learning_rate": 2.0587400506199153e-05, + "loss": 4.2431, + "num_input_tokens_seen": 98958336, + "step": 1510 + }, + { + "epoch": 1.2197453831832135, + "grad_norm": 1.2098276615142822, + "learning_rate": 2.0553399821548317e-05, + "loss": 4.2635, + "num_input_tokens_seen": 99286016, + "step": 1515 + }, + { + "epoch": 1.22377094550395, + "grad_norm": 1.1902521848678589, + "learning_rate": 2.0519567041703087e-05, + "loss": 4.283, + "num_input_tokens_seen": 99613696, + "step": 1520 + }, + { + "epoch": 1.2277965078246869, + "grad_norm": 1.2712699174880981, + "learning_rate": 2.048590078926336e-05, + "loss": 4.2737, + "num_input_tokens_seen": 99941376, + "step": 1525 + }, + { + "epoch": 1.2318220701454234, + "grad_norm": 1.2726564407348633, + "learning_rate": 2.0452399702596545e-05, + "loss": 4.2715, + "num_input_tokens_seen": 100269056, + "step": 1530 + }, + { + "epoch": 1.23584763246616, + "grad_norm": 1.2632936239242554, + "learning_rate": 2.0419062435606238e-05, + "loss": 4.2266, + "num_input_tokens_seen": 100596736, + "step": 1535 + }, + { + "epoch": 1.2398731947868968, + "grad_norm": 1.470897912979126, + "learning_rate": 2.038588765750502e-05, + "loss": 4.235, + "num_input_tokens_seen": 100924416, + "step": 1540 + }, + { + "epoch": 1.2438987571076334, + "grad_norm": 1.3083152770996094, + "learning_rate": 2.035287405259138e-05, + "loss": 4.233, + "num_input_tokens_seen": 101252096, + "step": 1545 + }, + { + "epoch": 1.2479243194283702, + "grad_norm": 1.2998907566070557, + "learning_rate": 2.032002032003048e-05, + "loss": 4.22, + "num_input_tokens_seen": 101579776, + "step": 1550 + }, + { + "epoch": 1.2519498817491068, + "grad_norm": 1.3198977708816528, + "learning_rate": 2.0287325173638952e-05, + "loss": 4.236, + "num_input_tokens_seen": 101907456, + "step": 1555 + }, + { + "epoch": 1.2559754440698434, + "grad_norm": 1.3605294227600098, + "learning_rate": 2.0254787341673334e-05, + "loss": 4.2318, + "num_input_tokens_seen": 102235136, + "step": 1560 + }, + { + "epoch": 1.2600010063905802, + "grad_norm": 1.2605648040771484, + "learning_rate": 2.0222405566622346e-05, + "loss": 4.232, + "num_input_tokens_seen": 102562816, + "step": 1565 + }, + { + "epoch": 1.2640265687113168, + "grad_norm": 1.2441262006759644, + "learning_rate": 2.019017860500275e-05, + "loss": 4.1917, + "num_input_tokens_seen": 102890496, + "step": 1570 + }, + { + "epoch": 1.2680521310320536, + "grad_norm": 1.3246742486953735, + "learning_rate": 2.0158105227158783e-05, + "loss": 4.2272, + "num_input_tokens_seen": 103218176, + "step": 1575 + }, + { + "epoch": 1.2720776933527902, + "grad_norm": 1.2322101593017578, + "learning_rate": 2.0126184217065105e-05, + "loss": 4.1989, + "num_input_tokens_seen": 103545856, + "step": 1580 + }, + { + "epoch": 1.2761032556735268, + "grad_norm": 1.2290526628494263, + "learning_rate": 2.0094414372133134e-05, + "loss": 4.2389, + "num_input_tokens_seen": 103873536, + "step": 1585 + }, + { + "epoch": 1.2801288179942636, + "grad_norm": 1.3795405626296997, + "learning_rate": 2.006279450302077e-05, + "loss": 4.1988, + "num_input_tokens_seen": 104201216, + "step": 1590 + }, + { + "epoch": 1.2841543803150002, + "grad_norm": 1.3456906080245972, + "learning_rate": 2.003132343344538e-05, + "loss": 4.2586, + "num_input_tokens_seen": 104528896, + "step": 1595 + }, + { + "epoch": 1.288179942635737, + "grad_norm": 1.2667040824890137, + "learning_rate": 2e-05, + "loss": 4.2505, + "num_input_tokens_seen": 104856576, + "step": 1600 + }, + { + "epoch": 1.2922055049564736, + "grad_norm": 1.2646143436431885, + "learning_rate": 1.9968823051972716e-05, + "loss": 4.2537, + "num_input_tokens_seen": 105184256, + "step": 1605 + }, + { + "epoch": 1.2962310672772102, + "grad_norm": 1.2952028512954712, + "learning_rate": 1.9937791451169073e-05, + "loss": 4.2439, + "num_input_tokens_seen": 105511936, + "step": 1610 + }, + { + "epoch": 1.300256629597947, + "grad_norm": 1.346574068069458, + "learning_rate": 1.9906904071737584e-05, + "loss": 4.24, + "num_input_tokens_seen": 105839616, + "step": 1615 + }, + { + "epoch": 1.3042821919186838, + "grad_norm": 1.2765649557113647, + "learning_rate": 1.9876159799998135e-05, + "loss": 4.2229, + "num_input_tokens_seen": 106167296, + "step": 1620 + }, + { + "epoch": 1.3083077542394204, + "grad_norm": 1.3729643821716309, + "learning_rate": 1.9845557534273358e-05, + "loss": 4.201, + "num_input_tokens_seen": 106494976, + "step": 1625 + }, + { + "epoch": 1.312333316560157, + "grad_norm": 1.2387956380844116, + "learning_rate": 1.98150961847228e-05, + "loss": 4.2235, + "num_input_tokens_seen": 106822656, + "step": 1630 + }, + { + "epoch": 1.3163588788808938, + "grad_norm": 1.283973217010498, + "learning_rate": 1.978477467317992e-05, + "loss": 4.2616, + "num_input_tokens_seen": 107150336, + "step": 1635 + }, + { + "epoch": 1.3203844412016303, + "grad_norm": 1.3440768718719482, + "learning_rate": 1.9754591932991796e-05, + "loss": 4.2215, + "num_input_tokens_seen": 107478016, + "step": 1640 + }, + { + "epoch": 1.3244100035223672, + "grad_norm": 1.292891263961792, + "learning_rate": 1.9724546908861517e-05, + "loss": 4.1927, + "num_input_tokens_seen": 107805696, + "step": 1645 + }, + { + "epoch": 1.3284355658431037, + "grad_norm": 1.3360971212387085, + "learning_rate": 1.9694638556693238e-05, + "loss": 4.2542, + "num_input_tokens_seen": 108133376, + "step": 1650 + }, + { + "epoch": 1.3284355658431037, + "eval_accuracy": 0.37502241786907964, + "eval_loss": 4.072964668273926, + "eval_runtime": 6.6283, + "eval_samples_per_second": 45.26, + "eval_steps_per_second": 5.733, + "num_input_tokens_seen": 108133376, + "step": 1650 + }, + { + "epoch": 1.3324611281638403, + "grad_norm": 1.2728638648986816, + "learning_rate": 1.9664865843439752e-05, + "loss": 4.2287, + "num_input_tokens_seen": 108461056, + "step": 1655 + }, + { + "epoch": 1.3364866904845771, + "grad_norm": 1.242002248764038, + "learning_rate": 1.9635227746952642e-05, + "loss": 4.2355, + "num_input_tokens_seen": 108788736, + "step": 1660 + }, + { + "epoch": 1.3405122528053137, + "grad_norm": 1.2821120023727417, + "learning_rate": 1.9605723255834874e-05, + "loss": 4.2256, + "num_input_tokens_seen": 109116416, + "step": 1665 + }, + { + "epoch": 1.3445378151260505, + "grad_norm": 1.380469560623169, + "learning_rate": 1.9576351369295855e-05, + "loss": 4.2284, + "num_input_tokens_seen": 109444096, + "step": 1670 + }, + { + "epoch": 1.3485633774467871, + "grad_norm": 1.241410493850708, + "learning_rate": 1.9547111097008838e-05, + "loss": 4.2122, + "num_input_tokens_seen": 109771776, + "step": 1675 + }, + { + "epoch": 1.3525889397675237, + "grad_norm": 1.2036563158035278, + "learning_rate": 1.9518001458970662e-05, + "loss": 4.2312, + "num_input_tokens_seen": 110099456, + "step": 1680 + }, + { + "epoch": 1.3566145020882605, + "grad_norm": 1.3382368087768555, + "learning_rate": 1.9489021485363838e-05, + "loss": 4.2168, + "num_input_tokens_seen": 110427136, + "step": 1685 + }, + { + "epoch": 1.360640064408997, + "grad_norm": 1.2150115966796875, + "learning_rate": 1.9460170216420796e-05, + "loss": 4.2322, + "num_input_tokens_seen": 110754816, + "step": 1690 + }, + { + "epoch": 1.364665626729734, + "grad_norm": 1.2400802373886108, + "learning_rate": 1.9431446702290413e-05, + "loss": 4.2018, + "num_input_tokens_seen": 111082496, + "step": 1695 + }, + { + "epoch": 1.3686911890504705, + "grad_norm": 1.3871190547943115, + "learning_rate": 1.9402850002906638e-05, + "loss": 4.2346, + "num_input_tokens_seen": 111410176, + "step": 1700 + }, + { + "epoch": 1.372716751371207, + "grad_norm": 1.2957957983016968, + "learning_rate": 1.9374379187859312e-05, + "loss": 4.2078, + "num_input_tokens_seen": 111737856, + "step": 1705 + }, + { + "epoch": 1.3767423136919439, + "grad_norm": 1.2237967252731323, + "learning_rate": 1.934603333626698e-05, + "loss": 4.2472, + "num_input_tokens_seen": 112065536, + "step": 1710 + }, + { + "epoch": 1.3807678760126805, + "grad_norm": 1.2709431648254395, + "learning_rate": 1.9317811536651808e-05, + "loss": 4.2679, + "num_input_tokens_seen": 112393216, + "step": 1715 + }, + { + "epoch": 1.3847934383334173, + "grad_norm": 1.3072236776351929, + "learning_rate": 1.9289712886816488e-05, + "loss": 4.2672, + "num_input_tokens_seen": 112720896, + "step": 1720 + }, + { + "epoch": 1.3888190006541539, + "grad_norm": 1.2216880321502686, + "learning_rate": 1.9261736493723075e-05, + "loss": 4.2124, + "num_input_tokens_seen": 113048576, + "step": 1725 + }, + { + "epoch": 1.3928445629748905, + "grad_norm": 1.2216253280639648, + "learning_rate": 1.9233881473373802e-05, + "loss": 4.2168, + "num_input_tokens_seen": 113376256, + "step": 1730 + }, + { + "epoch": 1.3968701252956273, + "grad_norm": 1.15086030960083, + "learning_rate": 1.9206146950693745e-05, + "loss": 4.2097, + "num_input_tokens_seen": 113703936, + "step": 1735 + }, + { + "epoch": 1.4008956876163638, + "grad_norm": 1.2962745428085327, + "learning_rate": 1.917853205941537e-05, + "loss": 4.2328, + "num_input_tokens_seen": 114031616, + "step": 1740 + }, + { + "epoch": 1.4049212499371007, + "grad_norm": 1.3710294961929321, + "learning_rate": 1.915103594196486e-05, + "loss": 4.2205, + "num_input_tokens_seen": 114359296, + "step": 1745 + }, + { + "epoch": 1.4089468122578372, + "grad_norm": 1.290256142616272, + "learning_rate": 1.91236577493503e-05, + "loss": 4.2239, + "num_input_tokens_seen": 114686976, + "step": 1750 + }, + { + "epoch": 1.4129723745785738, + "grad_norm": 1.1622158288955688, + "learning_rate": 1.9096396641051548e-05, + "loss": 4.2077, + "num_input_tokens_seen": 115014656, + "step": 1755 + }, + { + "epoch": 1.4169979368993106, + "grad_norm": 1.2251081466674805, + "learning_rate": 1.9069251784911845e-05, + "loss": 4.2479, + "num_input_tokens_seen": 115342336, + "step": 1760 + }, + { + "epoch": 1.4210234992200472, + "grad_norm": 1.2181967496871948, + "learning_rate": 1.9042222357031166e-05, + "loss": 4.2252, + "num_input_tokens_seen": 115670016, + "step": 1765 + }, + { + "epoch": 1.425049061540784, + "grad_norm": 1.3037039041519165, + "learning_rate": 1.9015307541661134e-05, + "loss": 4.22, + "num_input_tokens_seen": 115997696, + "step": 1770 + }, + { + "epoch": 1.4290746238615206, + "grad_norm": 1.2290278673171997, + "learning_rate": 1.8988506531101655e-05, + "loss": 4.22, + "num_input_tokens_seen": 116325376, + "step": 1775 + }, + { + "epoch": 1.4331001861822572, + "grad_norm": 1.287455677986145, + "learning_rate": 1.8961818525599093e-05, + "loss": 4.2533, + "num_input_tokens_seen": 116653056, + "step": 1780 + }, + { + "epoch": 1.437125748502994, + "grad_norm": 1.1899845600128174, + "learning_rate": 1.8935242733246034e-05, + "loss": 4.2289, + "num_input_tokens_seen": 116980736, + "step": 1785 + }, + { + "epoch": 1.4411513108237308, + "grad_norm": 1.2166035175323486, + "learning_rate": 1.8908778369882623e-05, + "loss": 4.218, + "num_input_tokens_seen": 117308416, + "step": 1790 + }, + { + "epoch": 1.4451768731444674, + "grad_norm": 1.2712666988372803, + "learning_rate": 1.888242465899932e-05, + "loss": 4.1944, + "num_input_tokens_seen": 117636096, + "step": 1795 + }, + { + "epoch": 1.449202435465204, + "grad_norm": 1.1924896240234375, + "learning_rate": 1.885618083164127e-05, + "loss": 4.2614, + "num_input_tokens_seen": 117963776, + "step": 1800 + }, + { + "epoch": 1.449202435465204, + "eval_accuracy": 0.37485385814497274, + "eval_loss": 4.068163871765137, + "eval_runtime": 7.4672, + "eval_samples_per_second": 40.176, + "eval_steps_per_second": 5.089, + "num_input_tokens_seen": 117963776, + "step": 1800 + }, + { + "epoch": 1.4532279977859408, + "grad_norm": 1.2685041427612305, + "learning_rate": 1.883004612631402e-05, + "loss": 4.236, + "num_input_tokens_seen": 118291456, + "step": 1805 + }, + { + "epoch": 1.4572535601066774, + "grad_norm": 1.1977604627609253, + "learning_rate": 1.880401978889074e-05, + "loss": 4.258, + "num_input_tokens_seen": 118619136, + "step": 1810 + }, + { + "epoch": 1.4612791224274142, + "grad_norm": 1.337408423423767, + "learning_rate": 1.877810107252081e-05, + "loss": 4.2549, + "num_input_tokens_seen": 118946816, + "step": 1815 + }, + { + "epoch": 1.4653046847481508, + "grad_norm": 1.348046898841858, + "learning_rate": 1.875228923753982e-05, + "loss": 4.2513, + "num_input_tokens_seen": 119274496, + "step": 1820 + }, + { + "epoch": 1.4693302470688874, + "grad_norm": 1.2587336301803589, + "learning_rate": 1.8726583551380893e-05, + "loss": 4.2251, + "num_input_tokens_seen": 119602176, + "step": 1825 + }, + { + "epoch": 1.4733558093896242, + "grad_norm": 1.3538405895233154, + "learning_rate": 1.8700983288487377e-05, + "loss": 4.2219, + "num_input_tokens_seen": 119929856, + "step": 1830 + }, + { + "epoch": 1.4773813717103608, + "grad_norm": 1.2679195404052734, + "learning_rate": 1.8675487730226835e-05, + "loss": 4.2026, + "num_input_tokens_seen": 120257536, + "step": 1835 + }, + { + "epoch": 1.4814069340310976, + "grad_norm": 1.2712976932525635, + "learning_rate": 1.8650096164806278e-05, + "loss": 4.2277, + "num_input_tokens_seen": 120585216, + "step": 1840 + }, + { + "epoch": 1.4854324963518342, + "grad_norm": 1.2575690746307373, + "learning_rate": 1.862480788718875e-05, + "loss": 4.2329, + "num_input_tokens_seen": 120912896, + "step": 1845 + }, + { + "epoch": 1.4894580586725708, + "grad_norm": 1.2137782573699951, + "learning_rate": 1.8599622199011086e-05, + "loss": 4.2076, + "num_input_tokens_seen": 121240576, + "step": 1850 + }, + { + "epoch": 1.4934836209933076, + "grad_norm": 1.2346251010894775, + "learning_rate": 1.8574538408502883e-05, + "loss": 4.2357, + "num_input_tokens_seen": 121568256, + "step": 1855 + }, + { + "epoch": 1.4975091833140441, + "grad_norm": 1.2532144784927368, + "learning_rate": 1.854955583040673e-05, + "loss": 4.2616, + "num_input_tokens_seen": 121895936, + "step": 1860 + }, + { + "epoch": 1.501534745634781, + "grad_norm": 1.1912769079208374, + "learning_rate": 1.8524673785899573e-05, + "loss": 4.1788, + "num_input_tokens_seen": 122223616, + "step": 1865 + }, + { + "epoch": 1.5055603079555175, + "grad_norm": 1.2742645740509033, + "learning_rate": 1.849989160251521e-05, + "loss": 4.2119, + "num_input_tokens_seen": 122551296, + "step": 1870 + }, + { + "epoch": 1.5095858702762541, + "grad_norm": 1.264236569404602, + "learning_rate": 1.847520861406802e-05, + "loss": 4.2219, + "num_input_tokens_seen": 122878976, + "step": 1875 + }, + { + "epoch": 1.513611432596991, + "grad_norm": 1.2975112199783325, + "learning_rate": 1.8450624160577702e-05, + "loss": 4.2227, + "num_input_tokens_seen": 123206656, + "step": 1880 + }, + { + "epoch": 1.5176369949177275, + "grad_norm": 1.231823205947876, + "learning_rate": 1.842613758819515e-05, + "loss": 4.2185, + "num_input_tokens_seen": 123534336, + "step": 1885 + }, + { + "epoch": 1.5216625572384643, + "grad_norm": 1.3271822929382324, + "learning_rate": 1.8401748249129447e-05, + "loss": 4.2174, + "num_input_tokens_seen": 123862016, + "step": 1890 + }, + { + "epoch": 1.525688119559201, + "grad_norm": 1.2804478406906128, + "learning_rate": 1.8377455501575864e-05, + "loss": 4.1906, + "num_input_tokens_seen": 124189696, + "step": 1895 + }, + { + "epoch": 1.5297136818799375, + "grad_norm": 1.2299883365631104, + "learning_rate": 1.835325870964494e-05, + "loss": 4.1957, + "num_input_tokens_seen": 124517376, + "step": 1900 + }, + { + "epoch": 1.5337392442006743, + "grad_norm": 1.2697046995162964, + "learning_rate": 1.8329157243292555e-05, + "loss": 4.2115, + "num_input_tokens_seen": 124845056, + "step": 1905 + }, + { + "epoch": 1.5377648065214111, + "grad_norm": 1.20834219455719, + "learning_rate": 1.8305150478251023e-05, + "loss": 4.2224, + "num_input_tokens_seen": 125172736, + "step": 1910 + }, + { + "epoch": 1.5417903688421477, + "grad_norm": 1.2465049028396606, + "learning_rate": 1.8281237795961206e-05, + "loss": 4.1757, + "num_input_tokens_seen": 125500416, + "step": 1915 + }, + { + "epoch": 1.5458159311628843, + "grad_norm": 1.2020412683486938, + "learning_rate": 1.825741858350554e-05, + "loss": 4.2047, + "num_input_tokens_seen": 125828096, + "step": 1920 + }, + { + "epoch": 1.5498414934836209, + "grad_norm": 1.384621024131775, + "learning_rate": 1.8233692233542074e-05, + "loss": 4.2553, + "num_input_tokens_seen": 126155776, + "step": 1925 + }, + { + "epoch": 1.5538670558043577, + "grad_norm": 1.212430477142334, + "learning_rate": 1.8210058144239417e-05, + "loss": 4.1952, + "num_input_tokens_seen": 126483456, + "step": 1930 + }, + { + "epoch": 1.5578926181250945, + "grad_norm": 1.2624080181121826, + "learning_rate": 1.8186515719212627e-05, + "loss": 4.2208, + "num_input_tokens_seen": 126811136, + "step": 1935 + }, + { + "epoch": 1.561918180445831, + "grad_norm": 1.2795370817184448, + "learning_rate": 1.8163064367459993e-05, + "loss": 4.1835, + "num_input_tokens_seen": 127138816, + "step": 1940 + }, + { + "epoch": 1.5659437427665677, + "grad_norm": 1.3003804683685303, + "learning_rate": 1.813970350330073e-05, + "loss": 4.2235, + "num_input_tokens_seen": 127466496, + "step": 1945 + }, + { + "epoch": 1.5699693050873043, + "grad_norm": 1.328814148902893, + "learning_rate": 1.8116432546313533e-05, + "loss": 4.1928, + "num_input_tokens_seen": 127794176, + "step": 1950 + }, + { + "epoch": 1.5699693050873043, + "eval_accuracy": 0.37578634243079373, + "eval_loss": 4.0595831871032715, + "eval_runtime": 7.0878, + "eval_samples_per_second": 42.326, + "eval_steps_per_second": 5.361, + "num_input_tokens_seen": 127794176, + "step": 1950 + }, + { + "epoch": 1.573994867408041, + "grad_norm": 1.258589506149292, + "learning_rate": 1.8093250921276027e-05, + "loss": 4.2336, + "num_input_tokens_seen": 128121856, + "step": 1955 + }, + { + "epoch": 1.5780204297287779, + "grad_norm": 1.226361632347107, + "learning_rate": 1.8070158058105027e-05, + "loss": 4.1945, + "num_input_tokens_seen": 128449536, + "step": 1960 + }, + { + "epoch": 1.5820459920495145, + "grad_norm": 1.2249149084091187, + "learning_rate": 1.8047153391797662e-05, + "loss": 4.1916, + "num_input_tokens_seen": 128777216, + "step": 1965 + }, + { + "epoch": 1.586071554370251, + "grad_norm": 1.3121646642684937, + "learning_rate": 1.8024236362373317e-05, + "loss": 4.1833, + "num_input_tokens_seen": 129104896, + "step": 1970 + }, + { + "epoch": 1.5900971166909876, + "grad_norm": 1.2635945081710815, + "learning_rate": 1.8001406414816385e-05, + "loss": 4.208, + "num_input_tokens_seen": 129432576, + "step": 1975 + }, + { + "epoch": 1.5941226790117244, + "grad_norm": 1.2306846380233765, + "learning_rate": 1.797866299901979e-05, + "loss": 4.1832, + "num_input_tokens_seen": 129760256, + "step": 1980 + }, + { + "epoch": 1.5981482413324613, + "grad_norm": 1.2278416156768799, + "learning_rate": 1.7956005569729342e-05, + "loss": 4.1765, + "num_input_tokens_seen": 130087936, + "step": 1985 + }, + { + "epoch": 1.6021738036531978, + "grad_norm": 1.2951785326004028, + "learning_rate": 1.7933433586488815e-05, + "loss": 4.2123, + "num_input_tokens_seen": 130415616, + "step": 1990 + }, + { + "epoch": 1.6061993659739344, + "grad_norm": 1.341288447380066, + "learning_rate": 1.79109465135858e-05, + "loss": 4.1865, + "num_input_tokens_seen": 130743296, + "step": 1995 + }, + { + "epoch": 1.610224928294671, + "grad_norm": 1.269637942314148, + "learning_rate": 1.788854381999832e-05, + "loss": 4.2004, + "num_input_tokens_seen": 131070976, + "step": 2000 + }, + { + "epoch": 1.6142504906154078, + "grad_norm": 1.3138514757156372, + "learning_rate": 1.7866224979342134e-05, + "loss": 4.2172, + "num_input_tokens_seen": 131398656, + "step": 2005 + }, + { + "epoch": 1.6182760529361446, + "grad_norm": 1.1714235544204712, + "learning_rate": 1.7843989469818822e-05, + "loss": 4.1903, + "num_input_tokens_seen": 131726336, + "step": 2010 + }, + { + "epoch": 1.6223016152568812, + "grad_norm": 1.3786474466323853, + "learning_rate": 1.7821836774164523e-05, + "loss": 4.2, + "num_input_tokens_seen": 132054016, + "step": 2015 + }, + { + "epoch": 1.6263271775776178, + "grad_norm": 1.3901410102844238, + "learning_rate": 1.7799766379599393e-05, + "loss": 4.1928, + "num_input_tokens_seen": 132381696, + "step": 2020 + }, + { + "epoch": 1.6303527398983544, + "grad_norm": 1.3432561159133911, + "learning_rate": 1.7777777777777777e-05, + "loss": 4.1983, + "num_input_tokens_seen": 132709376, + "step": 2025 + }, + { + "epoch": 1.6343783022190912, + "grad_norm": 1.2629408836364746, + "learning_rate": 1.7755870464739014e-05, + "loss": 4.2136, + "num_input_tokens_seen": 133037056, + "step": 2030 + }, + { + "epoch": 1.638403864539828, + "grad_norm": 1.2436974048614502, + "learning_rate": 1.7734043940858906e-05, + "loss": 4.1695, + "num_input_tokens_seen": 133364736, + "step": 2035 + }, + { + "epoch": 1.6424294268605646, + "grad_norm": 1.2254369258880615, + "learning_rate": 1.771229771080191e-05, + "loss": 4.1636, + "num_input_tokens_seen": 133692416, + "step": 2040 + }, + { + "epoch": 1.6464549891813012, + "grad_norm": 1.3272947072982788, + "learning_rate": 1.769063128347386e-05, + "loss": 4.1979, + "num_input_tokens_seen": 134020096, + "step": 2045 + }, + { + "epoch": 1.650480551502038, + "grad_norm": 1.2953760623931885, + "learning_rate": 1.7669044171975447e-05, + "loss": 4.2026, + "num_input_tokens_seen": 134347776, + "step": 2050 + }, + { + "epoch": 1.6545061138227746, + "grad_norm": 1.2235223054885864, + "learning_rate": 1.764753589355622e-05, + "loss": 4.1674, + "num_input_tokens_seen": 134675456, + "step": 2055 + }, + { + "epoch": 1.6585316761435114, + "grad_norm": 1.1912940740585327, + "learning_rate": 1.762610596956927e-05, + "loss": 4.1811, + "num_input_tokens_seen": 135003136, + "step": 2060 + }, + { + "epoch": 1.662557238464248, + "grad_norm": 1.3255517482757568, + "learning_rate": 1.7604753925426462e-05, + "loss": 4.2094, + "num_input_tokens_seen": 135330816, + "step": 2065 + }, + { + "epoch": 1.6665828007849846, + "grad_norm": 1.2693239450454712, + "learning_rate": 1.758347929055432e-05, + "loss": 4.2113, + "num_input_tokens_seen": 135658496, + "step": 2070 + }, + { + "epoch": 1.6706083631057214, + "grad_norm": 1.2487210035324097, + "learning_rate": 1.7562281598350458e-05, + "loss": 4.2341, + "num_input_tokens_seen": 135986176, + "step": 2075 + }, + { + "epoch": 1.6746339254264582, + "grad_norm": 1.3119174242019653, + "learning_rate": 1.7541160386140587e-05, + "loss": 4.1846, + "num_input_tokens_seen": 136313856, + "step": 2080 + }, + { + "epoch": 1.6786594877471948, + "grad_norm": 1.2278319597244263, + "learning_rate": 1.7520115195136116e-05, + "loss": 4.1637, + "num_input_tokens_seen": 136641536, + "step": 2085 + }, + { + "epoch": 1.6826850500679313, + "grad_norm": 1.2323991060256958, + "learning_rate": 1.7499145570392288e-05, + "loss": 4.1943, + "num_input_tokens_seen": 136969216, + "step": 2090 + }, + { + "epoch": 1.686710612388668, + "grad_norm": 1.3337785005569458, + "learning_rate": 1.747825106076687e-05, + "loss": 4.2297, + "num_input_tokens_seen": 137296896, + "step": 2095 + }, + { + "epoch": 1.6907361747094047, + "grad_norm": 1.1806713342666626, + "learning_rate": 1.7457431218879393e-05, + "loss": 4.1971, + "num_input_tokens_seen": 137624576, + "step": 2100 + }, + { + "epoch": 1.6907361747094047, + "eval_accuracy": 0.3776899715899814, + "eval_loss": 4.050466537475586, + "eval_runtime": 6.6452, + "eval_samples_per_second": 45.145, + "eval_steps_per_second": 5.718, + "num_input_tokens_seen": 137624576, + "step": 2100 + }, + { + "epoch": 1.6947617370301415, + "grad_norm": 1.1816240549087524, + "learning_rate": 1.7436685601070914e-05, + "loss": 4.1944, + "num_input_tokens_seen": 137952256, + "step": 2105 + }, + { + "epoch": 1.6987872993508781, + "grad_norm": 1.258262276649475, + "learning_rate": 1.7416013767364324e-05, + "loss": 4.1668, + "num_input_tokens_seen": 138279936, + "step": 2110 + }, + { + "epoch": 1.7028128616716147, + "grad_norm": 1.2995336055755615, + "learning_rate": 1.7395415281425125e-05, + "loss": 4.2091, + "num_input_tokens_seen": 138607616, + "step": 2115 + }, + { + "epoch": 1.7068384239923513, + "grad_norm": 1.341255784034729, + "learning_rate": 1.7374889710522778e-05, + "loss": 4.1956, + "num_input_tokens_seen": 138935296, + "step": 2120 + }, + { + "epoch": 1.7108639863130881, + "grad_norm": 1.3196258544921875, + "learning_rate": 1.7354436625492496e-05, + "loss": 4.1887, + "num_input_tokens_seen": 139262976, + "step": 2125 + }, + { + "epoch": 1.714889548633825, + "grad_norm": 1.3612850904464722, + "learning_rate": 1.7334055600697583e-05, + "loss": 4.1805, + "num_input_tokens_seen": 139590656, + "step": 2130 + }, + { + "epoch": 1.7189151109545615, + "grad_norm": 1.259021282196045, + "learning_rate": 1.73137462139922e-05, + "loss": 4.2013, + "num_input_tokens_seen": 139918336, + "step": 2135 + }, + { + "epoch": 1.722940673275298, + "grad_norm": 1.226407766342163, + "learning_rate": 1.729350804668468e-05, + "loss": 4.1963, + "num_input_tokens_seen": 140246016, + "step": 2140 + }, + { + "epoch": 1.7269662355960347, + "grad_norm": 1.278818130493164, + "learning_rate": 1.727334068350122e-05, + "loss": 4.1719, + "num_input_tokens_seen": 140573696, + "step": 2145 + }, + { + "epoch": 1.7309917979167715, + "grad_norm": 1.2456446886062622, + "learning_rate": 1.7253243712550147e-05, + "loss": 4.2157, + "num_input_tokens_seen": 140901376, + "step": 2150 + }, + { + "epoch": 1.7350173602375083, + "grad_norm": 1.249311089515686, + "learning_rate": 1.723321672528655e-05, + "loss": 4.1854, + "num_input_tokens_seen": 141229056, + "step": 2155 + }, + { + "epoch": 1.7390429225582449, + "grad_norm": 1.3078937530517578, + "learning_rate": 1.721325931647741e-05, + "loss": 4.1911, + "num_input_tokens_seen": 141556736, + "step": 2160 + }, + { + "epoch": 1.7430684848789815, + "grad_norm": 1.2260226011276245, + "learning_rate": 1.7193371084167158e-05, + "loss": 4.1945, + "num_input_tokens_seen": 141884416, + "step": 2165 + }, + { + "epoch": 1.747094047199718, + "grad_norm": 1.2438477277755737, + "learning_rate": 1.7173551629643676e-05, + "loss": 4.1877, + "num_input_tokens_seen": 142212096, + "step": 2170 + }, + { + "epoch": 1.7511196095204549, + "grad_norm": 1.3562901020050049, + "learning_rate": 1.7153800557404717e-05, + "loss": 4.1662, + "num_input_tokens_seen": 142539776, + "step": 2175 + }, + { + "epoch": 1.7551451718411917, + "grad_norm": 1.2909834384918213, + "learning_rate": 1.7134117475124774e-05, + "loss": 4.1718, + "num_input_tokens_seen": 142867456, + "step": 2180 + }, + { + "epoch": 1.7591707341619283, + "grad_norm": 1.3816372156143188, + "learning_rate": 1.7114501993622322e-05, + "loss": 4.187, + "num_input_tokens_seen": 143195136, + "step": 2185 + }, + { + "epoch": 1.7631962964826648, + "grad_norm": 1.1970911026000977, + "learning_rate": 1.7094953726827533e-05, + "loss": 4.2293, + "num_input_tokens_seen": 143522816, + "step": 2190 + }, + { + "epoch": 1.7672218588034014, + "grad_norm": 1.300299882888794, + "learning_rate": 1.707547229175031e-05, + "loss": 4.1583, + "num_input_tokens_seen": 143850496, + "step": 2195 + }, + { + "epoch": 1.7712474211241382, + "grad_norm": 1.3158056735992432, + "learning_rate": 1.7056057308448833e-05, + "loss": 4.1986, + "num_input_tokens_seen": 144178176, + "step": 2200 + }, + { + "epoch": 1.775272983444875, + "grad_norm": 1.3262073993682861, + "learning_rate": 1.7036708399998398e-05, + "loss": 4.1911, + "num_input_tokens_seen": 144505856, + "step": 2205 + }, + { + "epoch": 1.7792985457656116, + "grad_norm": 1.2565948963165283, + "learning_rate": 1.7017425192460684e-05, + "loss": 4.1982, + "num_input_tokens_seen": 144833536, + "step": 2210 + }, + { + "epoch": 1.7833241080863482, + "grad_norm": 1.3026916980743408, + "learning_rate": 1.699820731485341e-05, + "loss": 4.1539, + "num_input_tokens_seen": 145161216, + "step": 2215 + }, + { + "epoch": 1.787349670407085, + "grad_norm": 1.1916109323501587, + "learning_rate": 1.697905439912036e-05, + "loss": 4.1723, + "num_input_tokens_seen": 145488896, + "step": 2220 + }, + { + "epoch": 1.7913752327278216, + "grad_norm": 1.2661770582199097, + "learning_rate": 1.6959966080101762e-05, + "loss": 4.1575, + "num_input_tokens_seen": 145816576, + "step": 2225 + }, + { + "epoch": 1.7954007950485584, + "grad_norm": 1.1987600326538086, + "learning_rate": 1.694094199550507e-05, + "loss": 4.1784, + "num_input_tokens_seen": 146144256, + "step": 2230 + }, + { + "epoch": 1.799426357369295, + "grad_norm": 1.2095016241073608, + "learning_rate": 1.6921981785876064e-05, + "loss": 4.1995, + "num_input_tokens_seen": 146471936, + "step": 2235 + }, + { + "epoch": 1.8034519196900316, + "grad_norm": 1.220253586769104, + "learning_rate": 1.6903085094570334e-05, + "loss": 4.1876, + "num_input_tokens_seen": 146799616, + "step": 2240 + }, + { + "epoch": 1.8074774820107684, + "grad_norm": 1.3598766326904297, + "learning_rate": 1.6884251567725093e-05, + "loss": 4.1952, + "num_input_tokens_seen": 147127296, + "step": 2245 + }, + { + "epoch": 1.8115030443315052, + "grad_norm": 1.2614084482192993, + "learning_rate": 1.6865480854231357e-05, + "loss": 4.1966, + "num_input_tokens_seen": 147454976, + "step": 2250 + }, + { + "epoch": 1.8115030443315052, + "eval_accuracy": 0.3787434709282334, + "eval_loss": 4.016251564025879, + "eval_runtime": 6.8672, + "eval_samples_per_second": 43.686, + "eval_steps_per_second": 5.534, + "num_input_tokens_seen": 147454976, + "step": 2250 + }, + { + "epoch": 1.8155286066522418, + "grad_norm": 1.1849647760391235, + "learning_rate": 1.6846772605706453e-05, + "loss": 4.1795, + "num_input_tokens_seen": 147782656, + "step": 2255 + }, + { + "epoch": 1.8195541689729784, + "grad_norm": 1.217572569847107, + "learning_rate": 1.682812647646685e-05, + "loss": 4.2173, + "num_input_tokens_seen": 148110336, + "step": 2260 + }, + { + "epoch": 1.823579731293715, + "grad_norm": 1.3434370756149292, + "learning_rate": 1.6809542123501345e-05, + "loss": 4.1826, + "num_input_tokens_seen": 148438016, + "step": 2265 + }, + { + "epoch": 1.8276052936144518, + "grad_norm": 1.2767517566680908, + "learning_rate": 1.6791019206444543e-05, + "loss": 4.1785, + "num_input_tokens_seen": 148765696, + "step": 2270 + }, + { + "epoch": 1.8316308559351886, + "grad_norm": 1.1992509365081787, + "learning_rate": 1.6772557387550694e-05, + "loss": 4.1884, + "num_input_tokens_seen": 149093376, + "step": 2275 + }, + { + "epoch": 1.8356564182559252, + "grad_norm": 1.1988239288330078, + "learning_rate": 1.675415633166782e-05, + "loss": 4.1726, + "num_input_tokens_seen": 149421056, + "step": 2280 + }, + { + "epoch": 1.8396819805766618, + "grad_norm": 1.3448524475097656, + "learning_rate": 1.673581570621216e-05, + "loss": 4.1467, + "num_input_tokens_seen": 149748736, + "step": 2285 + }, + { + "epoch": 1.8437075428973984, + "grad_norm": 1.2856345176696777, + "learning_rate": 1.6717535181142915e-05, + "loss": 4.1613, + "num_input_tokens_seen": 150076416, + "step": 2290 + }, + { + "epoch": 1.8477331052181352, + "grad_norm": 1.357901692390442, + "learning_rate": 1.669931442893732e-05, + "loss": 4.1587, + "num_input_tokens_seen": 150404096, + "step": 2295 + }, + { + "epoch": 1.851758667538872, + "grad_norm": 1.295551061630249, + "learning_rate": 1.6681153124565983e-05, + "loss": 4.1588, + "num_input_tokens_seen": 150731776, + "step": 2300 + }, + { + "epoch": 1.8557842298596086, + "grad_norm": 1.2811824083328247, + "learning_rate": 1.6663050945468536e-05, + "loss": 4.1839, + "num_input_tokens_seen": 151059456, + "step": 2305 + }, + { + "epoch": 1.8598097921803451, + "grad_norm": 1.3775153160095215, + "learning_rate": 1.6645007571529582e-05, + "loss": 4.2016, + "num_input_tokens_seen": 151387136, + "step": 2310 + }, + { + "epoch": 1.8638353545010817, + "grad_norm": 1.3014804124832153, + "learning_rate": 1.6627022685054907e-05, + "loss": 4.2082, + "num_input_tokens_seen": 151714816, + "step": 2315 + }, + { + "epoch": 1.8678609168218185, + "grad_norm": 1.2765071392059326, + "learning_rate": 1.6609095970747994e-05, + "loss": 4.1453, + "num_input_tokens_seen": 152042496, + "step": 2320 + }, + { + "epoch": 1.8718864791425553, + "grad_norm": 1.4745502471923828, + "learning_rate": 1.6591227115686806e-05, + "loss": 4.1856, + "num_input_tokens_seen": 152370176, + "step": 2325 + }, + { + "epoch": 1.875912041463292, + "grad_norm": 1.2726967334747314, + "learning_rate": 1.6573415809300835e-05, + "loss": 4.1977, + "num_input_tokens_seen": 152697856, + "step": 2330 + }, + { + "epoch": 1.8799376037840285, + "grad_norm": 1.286971092224121, + "learning_rate": 1.6555661743348445e-05, + "loss": 4.1962, + "num_input_tokens_seen": 153025536, + "step": 2335 + }, + { + "epoch": 1.883963166104765, + "grad_norm": 1.2794855833053589, + "learning_rate": 1.6537964611894465e-05, + "loss": 4.1606, + "num_input_tokens_seen": 153353216, + "step": 2340 + }, + { + "epoch": 1.887988728425502, + "grad_norm": 1.2264660596847534, + "learning_rate": 1.652032411128802e-05, + "loss": 4.1724, + "num_input_tokens_seen": 153680896, + "step": 2345 + }, + { + "epoch": 1.8920142907462387, + "grad_norm": 1.3459347486495972, + "learning_rate": 1.6502739940140696e-05, + "loss": 4.1663, + "num_input_tokens_seen": 154008576, + "step": 2350 + }, + { + "epoch": 1.8960398530669753, + "grad_norm": 1.284248948097229, + "learning_rate": 1.6485211799304873e-05, + "loss": 4.1927, + "num_input_tokens_seen": 154336256, + "step": 2355 + }, + { + "epoch": 1.900065415387712, + "grad_norm": 1.273288369178772, + "learning_rate": 1.6467739391852368e-05, + "loss": 4.1442, + "num_input_tokens_seen": 154663936, + "step": 2360 + }, + { + "epoch": 1.9040909777084485, + "grad_norm": 1.3548243045806885, + "learning_rate": 1.6450322423053303e-05, + "loss": 4.153, + "num_input_tokens_seen": 154991616, + "step": 2365 + }, + { + "epoch": 1.9081165400291853, + "grad_norm": 1.2603306770324707, + "learning_rate": 1.6432960600355222e-05, + "loss": 4.1805, + "num_input_tokens_seen": 155319296, + "step": 2370 + }, + { + "epoch": 1.912142102349922, + "grad_norm": 1.3440006971359253, + "learning_rate": 1.6415653633362467e-05, + "loss": 4.1954, + "num_input_tokens_seen": 155646976, + "step": 2375 + }, + { + "epoch": 1.9161676646706587, + "grad_norm": 1.385733723640442, + "learning_rate": 1.6398401233815756e-05, + "loss": 4.1677, + "num_input_tokens_seen": 155974656, + "step": 2380 + }, + { + "epoch": 1.9201932269913953, + "grad_norm": 1.408675193786621, + "learning_rate": 1.638120311557203e-05, + "loss": 4.1724, + "num_input_tokens_seen": 156302336, + "step": 2385 + }, + { + "epoch": 1.924218789312132, + "grad_norm": 1.3172813653945923, + "learning_rate": 1.6364058994584528e-05, + "loss": 4.1658, + "num_input_tokens_seen": 156630016, + "step": 2390 + }, + { + "epoch": 1.9282443516328687, + "grad_norm": 1.4357519149780273, + "learning_rate": 1.6346968588883048e-05, + "loss": 4.1538, + "num_input_tokens_seen": 156957696, + "step": 2395 + }, + { + "epoch": 1.9322699139536055, + "grad_norm": 1.3051438331604004, + "learning_rate": 1.6329931618554523e-05, + "loss": 4.16, + "num_input_tokens_seen": 157285376, + "step": 2400 + }, + { + "epoch": 1.9322699139536055, + "eval_accuracy": 0.3773649363114314, + "eval_loss": 4.035182952880859, + "eval_runtime": 7.0048, + "eval_samples_per_second": 42.828, + "eval_steps_per_second": 5.425, + "num_input_tokens_seen": 157285376, + "step": 2400 + }, + { + "epoch": 1.936295476274342, + "grad_norm": 1.229457139968872, + "learning_rate": 1.6312947805723712e-05, + "loss": 4.1306, + "num_input_tokens_seen": 157613056, + "step": 2405 + }, + { + "epoch": 1.9403210385950787, + "grad_norm": 1.3241009712219238, + "learning_rate": 1.629601687453421e-05, + "loss": 4.1803, + "num_input_tokens_seen": 157940736, + "step": 2410 + }, + { + "epoch": 1.9443466009158155, + "grad_norm": 1.462347149848938, + "learning_rate": 1.6279138551129596e-05, + "loss": 4.1708, + "num_input_tokens_seen": 158268416, + "step": 2415 + }, + { + "epoch": 1.9483721632365523, + "grad_norm": 1.261893630027771, + "learning_rate": 1.6262312563634837e-05, + "loss": 4.1739, + "num_input_tokens_seen": 158596096, + "step": 2420 + }, + { + "epoch": 1.9523977255572889, + "grad_norm": 1.2686680555343628, + "learning_rate": 1.624553864213791e-05, + "loss": 4.1601, + "num_input_tokens_seen": 158923776, + "step": 2425 + }, + { + "epoch": 1.9564232878780254, + "grad_norm": 1.291801929473877, + "learning_rate": 1.622881651867159e-05, + "loss": 4.1971, + "num_input_tokens_seen": 159251456, + "step": 2430 + }, + { + "epoch": 1.960448850198762, + "grad_norm": 1.2599908113479614, + "learning_rate": 1.6212145927195505e-05, + "loss": 4.1876, + "num_input_tokens_seen": 159579136, + "step": 2435 + }, + { + "epoch": 1.9644744125194988, + "grad_norm": 1.2517198324203491, + "learning_rate": 1.6195526603578323e-05, + "loss": 4.2083, + "num_input_tokens_seen": 159906816, + "step": 2440 + }, + { + "epoch": 1.9684999748402356, + "grad_norm": 1.407639741897583, + "learning_rate": 1.6178958285580194e-05, + "loss": 4.1763, + "num_input_tokens_seen": 160234496, + "step": 2445 + }, + { + "epoch": 1.9725255371609722, + "grad_norm": 1.3766093254089355, + "learning_rate": 1.6162440712835372e-05, + "loss": 4.1656, + "num_input_tokens_seen": 160562176, + "step": 2450 + }, + { + "epoch": 1.9765510994817088, + "grad_norm": 1.201467514038086, + "learning_rate": 1.6145973626835027e-05, + "loss": 4.1317, + "num_input_tokens_seen": 160889856, + "step": 2455 + }, + { + "epoch": 1.9805766618024454, + "grad_norm": 1.245469093322754, + "learning_rate": 1.6129556770910237e-05, + "loss": 4.188, + "num_input_tokens_seen": 161217536, + "step": 2460 + }, + { + "epoch": 1.9846022241231822, + "grad_norm": 1.360424518585205, + "learning_rate": 1.611318989021522e-05, + "loss": 4.1817, + "num_input_tokens_seen": 161545216, + "step": 2465 + }, + { + "epoch": 1.988627786443919, + "grad_norm": 1.2917958498001099, + "learning_rate": 1.6096872731710677e-05, + "loss": 4.1912, + "num_input_tokens_seen": 161872896, + "step": 2470 + }, + { + "epoch": 1.9926533487646556, + "grad_norm": 1.311262607574463, + "learning_rate": 1.6080605044147395e-05, + "loss": 4.1897, + "num_input_tokens_seen": 162200576, + "step": 2475 + }, + { + "epoch": 1.9966789110853922, + "grad_norm": 1.233081340789795, + "learning_rate": 1.606438657804998e-05, + "loss": 4.1536, + "num_input_tokens_seen": 162528256, + "step": 2480 + }, + { + "epoch": 1.9998993609419817, + "num_input_tokens_seen": 162790400, + "step": 2484, + "total_flos": 1.059416318592e+16, + "train_loss": 4.377665276304727, + "train_runtime": 2972.3443, + "train_samples_per_second": 106.974, + "train_steps_per_second": 0.836, + "train_tokens_per_second": 54770.764 + } + ], + "logging_steps": 5, + "max_steps": 2484, + "num_input_tokens_seen": 162790400, + "num_train_epochs": 2, + "save_steps": 200, + "total_flos": 1.059416318592e+16, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +}